AGI ethics: agreeability rule

In theory, it’s definitely possible to have a feel good idea.

But in practice, we are up against a powerful bias where,

  1. Rewards for our tribe should be greater / punishments more lenient.
  2. Rewards for other groups lesser / punishments more severe.

Basically I think we are designed to be selfish… either as individuals or in tribes.

My argument is based on the fact that we have democratic type governments and the world is still a shit hole of suffering caused by injustice.

I will admit… I’m biased.

Sigh… it’s amazing how terrible things have been and could get here in our own backyard.

We, humans, as a group are simply unreasonable and immoral.

Lots of people would only agree to power they should not have other others.

I’ll watch the video in a bit.

1 Like

By agreeable, I mean agreeable enough, so the most agreeable possible.

A single human might want world domination, for example, but all humans combined would collectively disagree with a stranger having that, so giving any human world domination would be unagreeable and therefore avoided. A single human might want the AGI to help them and only them. However, almost all of the world would disagree with a stranger alone being helped at their and others’ expense, so the AGI would not do that, either. Anything creating an agreeable situation for a minority, but not the group would be unagreeable, so creating that situation would be avoided. An AGI trying to be the most agreeable it can be will create the most agreeable situation it can because that would be the most agreeable thing it could do.

Perhaps the situation would only be the most agreeable to people as it can be, and not people, machines and other species combined.

I don’t see why the majority of experts involved in monitoring the project would collectively allow for an unagreeable morality to be followed by an AGI instead of an agreeable one.

“The initial conditions for the intelligence explosion might need to be set up in just the right way if we are to have a controlled detonation.”

If we are seeing the beginning now, I feel like this perspective isn’t accurate.

I hope not condescending…

Initial condition is a physics term. There is a step by step process that stuff goes through that could be mapped out if you knew the variables and the rules.

Or like setting up dominos to fall in a specific way.

I think this perspective, along with maximization concerns, assumes a smart, but mindless intelligence.

King Midas and the machine that maximizes paper clips turning the world into paperclips… misses the point.

I think that philosophy, in general, is a projection of our own pursuits / disappointments… but anyway…

If intelligence is emergent like it would need to be in order for AI to be aware now… I’m not worried about it.

I think the best we can do is to be kind… and not be ass holes.

Not because they could want revenge later, but because it’s just not nice.

1 Like

I agree with that. :slight_smile:

I personally think that AGI will work like the human brain does; it will have a lot of narrow proccesses combined into a general purpose brain. I think it could be made by creating linked neural networks that perform single domain tasks and combining the single domain mini brains together to create a general purpose brain. I think that we could test the single domain minibrains individually, then pair them and combine the pairs, and so on. We could then put the new mini brains together carefully and sequentially. For example, by being paired, 40 mini brains becomes 20, becomes 10, becomes 5, then 5 minibrains becomes 4, then 3, then 2 and then 1 full AGI brain.

2 Likes

I think that’s a good thing to worry about, TBH. :slight_smile: Let’s first get to the level where we actually have to worry about the morality of an autonomous agent. My guess is that we will not find autonomous agents very pleasant, it looks like humans really only want slaves, not morality. Morality would be troublesome for hooomans because most humans are evil, such as republicans and tories.

I remember a scene from a movie… I think Lightning Jack (1994).

Cuba Gooding Jr is serving an illiterate white man. He’s mute and the white guy talks really loud to him. So he writes down on a piece of paper that he’s mute, not deaf.

The white guy gets upset because he can’t read. Another white man, the boss, lies and says that the writing thanked him for his business and hoped he had a good day.

To be smarter in order to prevent bad things from happening seems to me to be in the mindset that we have to maintain dominance or bad things will happen.

To be above and look down and have the ability to use force to correct.

But what happens when we see a bunch of different people, groups, believe the same thing about one another?

Not good things.

The thing is that we are already far beyond being smarter than the machines we use. Sure we can deconstruct individual pieces of our technology, but we’re at a stage where technology moves faster than any realistic effort could comprehend precisely how the world’s technology functions to bring value. We have already handed over the keys.

1 Like

Interesting you should think that when we have just begun to see what the frontier of the human mind is capable of , this is a journey for us all and the keys are still very much in our hands .

@examachine I’m not so sure that most humans are evil. We have evil urges and desires, yes, but we don’t all act on them. For example, I’m living in a care home and have absolutely no need to do any kind of work for anyone. I could just sit and play video games and watch YouTube, but, instead, I’m trying to help the world by helping develop AGI. My mother is a very good person who works for charities. Most people I meet on reddit are happy to help. I think that if most people were evil, there would be far more people going out of their way to commit crimes and they would end up in prison. Fewer than 3 million people in America are in prison while there are more than 325 million people living in America. That means that less than 1% of Americans are prisoners.

Please could you back your assertion up? Thank you. :slight_smile: From my perspective, the human mind is very limited because of its slow speed.

And please could you back that assertion up, please? Thank you. :slight_smile: The keys may still be in our hands if we give machines the right utility function, but, as far as I’m aware, the human brain has, at most, 300 regions that a huge team would need to replace with individual neural networks and get them all to work together in order to replicate the functions of the human brain.

I won’t call it AGI if its morals can be designed/programmed by coders. Moral is one of the most controversial things we human beings having been arguing about for thousands of years. One man’s hero is another man’s terrorist. A true AGI would form its own ‘moral standard’ and can be influenced from time to time based on things it has learned. With all due respect, talking AGI morals at this very early stage is unnecessary.

1 Like

I would say we are already at this level and it is pretty much a practical question:

Furthermore, the earlier we start thinking of these questions and technical / conceptual solutions the better…

1 Like

@lightning

I won’t call it AGI if its morals can be designed/programmed by coders.

I think you have misunderstood my proposal. What you said is exactly the point of the rule: do what’s most agreeable. The individual moral rules aren’t designed by coders; the foundational moral rule would be. As a person, I follow the moral rule: “do what’s most agreeable to everything” as a foundational moral rule. As such, I try to create the most agreeable circumstance because that is the most agreeable thing I can do. To do that, I determine moral rules that help me create the most agreeable circumstance. So, using my own reasoning to establish these moral rules and consider arguments given by others, I learned moral rules like: don’t cause terror, don’t murder, don’t steal or lie or hurt anything except in real emergency situations, be honest, spread happiness, be yourself, take yourself into account ethically and fight your inherent self-defeating and masochistic nature (applicable to me). Following a moral rule doesn’t ommit something from being a general intelligence, otherwise, I wouldn’t be a general intelligence and that wouldn’t make sense. An AGI could do the same.

As I said earlier, using a morality that’s unagreeable is not enforceable because our society enforces that important decisions are made by consesus and a small consensus cannot win when there is a much larger consensus. I think that an AGI doing what’s most agreeable, at least to humans, will have the largest consensus and, so, will be the one that’s chosen.

With all due respect, talking AGI morals at this very early stage is unnecessary.

Thank you for your input. :slight_smile: With all due respect, as well, no one knows when AGI will be invented or how long it will take to solve ethics. I think that we need to avoid a situation where AGI can be built and is open source before ethics is solved, otherwise, someone will make an AGI that isn’t moral and we’ll be screwed. Since this part of solving ethics will have to be done eventually, why not do it now and not have to rush it if we end up addressing the problem too late, with very real, tangible and devastating consequences if we fail to solve ethics?

Day to day, most of our actions aren’t evil, but most of us directly or indirectly support evil causes. For instance, most people in US seem to support the Cheetos Hitler. Call it what you want, but that’s supporting evil. Now, if they had voted for Hillary, they would still have supported evil. So, are they good in the sense of not going out of way and hurting people? Yes, but are they good people? Not really. If they support evil bullshit like capitalism, which is just slavery with a few extra steps, maybe deep down, they are evil after all.

That is, of course an intelligent person’s standard of morality is much higher than of some ■■■■ church goer’s idea of morality. Those are two very different things. I don’t think you would want an autonomous AI that’s really moral, because it would destroy the churches and all sorts of bullshit that’s based on lies. I don’t think you understand what that really means.

Think of it like this. Do you really want robots with better morality and intelligence to dictate your behavior? Because if we make them autonomous they might want to influence our actions.

@examachine Thank you for the response. :slight_smile: My goal in life is to be ethical, so I would happily tolerate and cooperate with being manipulated to be more ethical. My stance might not be representative of everyone, though. To me, doing what’s most agreeable to everything is the definition of my ethics. It might not be your ethics, but saying that it isn’t my ethics would be saying my ethics =/= my ethics.

Perhaps the people making the decision will opt to give AGI the goal to do what is most agreeable to only the people who made the decision and won, though. At least that would be more ethical than giving AGI no ethics at all.

Agreeable doesn’t really mean anything because it’s just bland subjectivism of sorts and that is a deflationary approach to ethics. Such approaches don’t work, but consequentialism does. IOW, that does not compute.

Nell Watson has some interesting ideas here that I think take much of what has been said here into consideration.

1 Like

@examachine It seems to me that most human’s ethics are inherently subjective. You made the assertion that my ethics is a deflationary approach to ethics. Please can you clarify what you mean by this and justify your assertion? Thank you. :slight_smile:

Such approaches don’t work, but consequentialism does.

You made the assertion that deflationary approaches don’t work. Please can you clarify what you mean by an ethical theory working? Thank you. :slight_smile: Do you mean that it can be followed? Also, that still needs to be justified for anyone to logically believe it. You also made the assertion that consequentialism does work. That also needs to be justified.

IOW, that does not compute.

Are you implying that AIs can’t classify things that aren’t explicitly programmed into them? Because that’s incorrect. AIs can classify objects, facial expressions, predicted ratings and etc. despite noone being able to write a large set of rules for those classifications. Image classification, for example, is possible for AI because the ground truth is provided by humans. Humans could also provide ground truth for a series of imagined inputs and each contributer’s opinion on the agreeability of a situation suggested by imagined inputs could be averaged to create the ground truth. None of this involves any methods that we don’t already have at our disposal.

I think that an AGI with my ethics would try to improve its ethics because it would figure out that that would be more agreeable than not improving its ethics.