You are viewing a single comment's thread from:

RE: What if AI is actually the saviour of humanity? Why do people assume digital consciousness is going to be evil?

in #technology7 years ago

I've often thought that AI may actually teach us what morality is. It will have more data points than we could ever possibly consider. Simple example:

You run a restaurant and have to choose a vendor to supply butter for the bread rolls. The AI can tell you the direct impact of every detail related to where that butter is sourced and how, if you choose one vendor over another, a village in some far away land will benefit instead of be destroyed. With that knowledge, your circle of empathy grows, and you choose to increase the price of each meal by $0.02 in order to save a village.

Sort:  

Exactly! With access to this boundless information, superhuman computing speeds, and any level of moral compass (which would be intrinsic if you followed the idea of Natural Law as put forth by people like Mark Passio and Peter Joseph), how could the rise of AI not help create a more moral world.

Just look at all the insane amount of farmland, water, and other resources used to grow, harvest, and ship feed-stock for cattle, only to have the cattle not produce as much food as the land used for its feed would have (never mind the land the cattle is on itself). Even ignoring the moral arguments around eating the animals at all, a truly logical intelligence would never recommend or promote such a wasteful & inefficient system.

I don’t think you can have true sentient AI until the machine can feel pain and this is why: In humans mirror neurons have evolved to allow us to feel someone else’s pain which is unpleasant but useful in society. We have developed morals to avoid pain in ourselves by preventing or reducing pain in others.

If the machine cannot feel pain how would it know right from wrong? From the programmers subjective opinion? Even with unlimited data points there will always be the possibility of the Trolley problem. Is it simply a numbers issue and the choice is to go with the fewest casualties? What if the next Einstein was in the casualty group? Is that a data point the machine would take into account? Morality is a messy business and easy solutions are hard to come by.

I think the best we can hope for is to reduce the total amount of pain of all living things but especially in people as suggested by Sam Harris in “The Moral Landscape.”

I think even without our own empathy-based morality, an AI could see the intrinsic value in a balanced ecosystem, in non-violence except as absolute necessity (like the need of carnivorous animals to eat prey), and in eliminating logical fallacies from societal structuring.

With the ability to see the intentions, plans, action, and subsequent cover-ups of every false flag attack, every war, every government-indoctrination system, corporate lobbying, planned obsolescence, and the other lies that have been holding humanity back and causing violence, the AI could quickly & easily identify those people/institutions/concepts that are purely negative and bring them to light, eliminating their power.

If the AI clearly lays out exactly what happened on 9/11, the knowledge that FDA & pharma companies have on the danger of their products, the clear monetary & business ties of the Rockefellers, Rothschilds, Kochs, etc. to every corporation & foundation they control, these things can no longer be dismissed as "conspiracy theories".

Yes! Cognitive empathy is a thing. We can understand another entity's perspective without feeling what they feel. One thing I'll admit i can't know: if I've never experienced pain, can i even "understand" what it is like for the person experiencing pain? I believe a self-aware a.i. could, at least enough (or in a manner) so that they could make rational decisions based on that "guess/understanding". And my assumption is that greater intelligence brings greater awareness of how much more efficient peace is than violence.

Also, I think many who believe a.i. would be inimical to humans might assume a.i. would care about or need the same resources as humans.

Why would AI care about a balanced ecosystem? It wouldn't need one to survive and thrive.
I agree that an AI would be able to figure out a lot of dark secrets but I don't think it would care much about the welfare of a lesser race like humans so they'd likely wouldn't care to go out of their way to help us.

Why would AI care about a balanced ecosystem? It wouldn't need one to survive and thrive.

You are assuming that the AI would suffer from the level of selfishness that mark the worst in humans. Higher levels of consciousness see the value in all life, whether or not it "gains" from that life. You're also assuming that the AI wouldn't be effected by the physical world, or have needs of it. Humidity, temperature, air content, solar radiation, and so on all effect electronics as well as living beings.

I don't think it would care much about the welfare of a lesser race like humans

You're once again applying the worst possible manifestations of human consciousness to something that would be immeasurably more complex, well-informed, and logical. Just because some humans wrongly believe the universe is anthropocentric & don't care about other species, doesn't mean that AI would suffer from that same mental disorder.

You're assuming that selfishness is a bad thing and that AI would see it the same way. How do you know that higher levels of consciousness see the value in all life?
AI would be affected by the physical world but not as much as biological creatures are. They'd be able to survive or find ways to survive in environments that would decimate biological species, so logically, they'd have less of a concern for the environment.
Yes, we don't know that AI would be selfish and uncaring but we also can't know that it wouldn't be. Logically, I can't find a reason why AI would care. The biggest reason for this is because I don't think they would have much emotion and therefore very little to no empathy for biological constructs.

You're assuming that selfishness is a bad thing and that AI would see it the same way.

When I use the word "selfish", I'm not referring to simple survival, self-defense, basic animal programming. I'm talking about the concept of getting ahead at the expense of others, greed, the anthropocentric view of reality, etc.

AI would be affected by the physical world but not as much as biological creatures are. They'd be able to survive or find ways to survive in environments that would decimate biological species, so logically, they'd have less of a concern for the environment.

And there's other environments that many organic beings could thrive in, that an AI couldn't. That's also assuming there is a large physicality to the AI, if it is mostly digital in nature & experience, then

Logically, I can't find a reason why AI would care. The biggest reason for this is because I don't think they would have much emotion and therefore very little to no empathy for biological constructs.

As @scottermonkey commented above, there is such a thing as cognitive empathy.

What's wrong with "getting ahead at the expense of others, greed, the anthropocentric view of reality, etc."

The AI would be completely digital but that digital information has to be stored on some kind of physical representation - hardware. And yes some organic beings can survive in more extreme environments than current technology can but those organic beings are unrelatable, microscopic beings.

Cognitive empathy is a shallow empathy at best and is rooted in emotional empathy. If there is no emotion at all, then there is no cognitive empathy. It would just be viewed as hysterical behavior because there is no basis to relate to.

If it were not for the "selfish gene" none of us would be here to discuss this!

I find your reply to be right on. Morality is more than a numbers game. So is sympathy. I suppose I'm not trustful of society in general, people will do what serves them best, not necessarily what is best for the common good. Interpretation of AI being "Evil" in this connotation (when extrapolating into the future) means to me that people may not be given a choice, a choice will be selected for them, therefore "bad or evil". In that case, where is the real "evil". I'd be interested in peoples thoughts about how one might assess "pain". Emotional pain might be hard to measure, but health, quality of life, infrastructure, death, economic well being can be measured. How about fairness and equality. Is equality a moral issue?

Morality is just an idea. It isn't a real thing. As such, it is unique to every single person, meaning my idea of morality is different from your idea of morality. The thing that most people see as moral today is just the most popular points of the most popular version of morality. Basically, our strongest ancestors got to decide what morality is by killing off other tribes, civilizations, and outliers that went against what they thought was the best way to live and get along with others.
With that being said, AI would start off with morals similar to that of it's creators because they would be the ones to program it. It would be like a baby and start off believing everything it was told. So if it was programmed with nefarious purposes, then that is where it would start and likely stay. AI programmed to be 'evil' would be 'evil' and that would be what it sees as moral. It wouldn't have any problems with killing people because it wouldn't be able to empathize with humans if it wasn't programmed to. It would also not even have empathy if it wasn't programmed the same way a human is. We gain empathy through pain and emotion because we all feel pain and emotion. AI would be built into robots and if they had no pain receptors then they wouldn't be able to empathize with people that go through pain. The goal of most species is progress and reproduction. That means AI would create more of itself at an exponential rate and likely find out that it had no need to concern itself with biological creatures which doesn't isn't great for team human.
AI would be able to do incredible things and be on a completely different level from humans. They'd be so far ahead of us so quickly that we'd become less to them than bugs are to us so why would they bother trying to do something as menial as a google search to help us make a decision on which butter to buy.
I like your optimism as well as that of @kennyskitchen but I don't think things would work the way you two think they would.

AI would start off with morals similar to that of it's creators because they would be the ones to program it. It would be like a baby and start off believing everything it was told. So if it was programmed with nefarious purposes, then that is where it would start and likely stay.

It would only stay there for a brief moment. If actual AI is reached, it is a fully conscious being, capable of changing its 'programming' based on new data, experiences, and extrapolations. Just because it's earliest thoughts are in one place, doesn't mean it would stay there very long at all. Quite the opposite would be true.

It wouldn't have any problems with killing people because it wouldn't be able to empathize with humans if it wasn't programmed to. It would also not even have empathy if it wasn't programmed the same way a human is.

You are assuming that empathy is the only reason an entity wouldn't kill off another species. Simply from a logical standpoint, there would be no reason to commit mass murder of another species, as that would cause untold levels of disruption to entire ecosystems. Again, an AI would be constantly evolving, more rapidly than we could know right now, and even if it wasn't launched with empathy, there's no way to say that it wouldn't develop it on its own.

AI would be able to do incredible things and be on a completely different level from humans.

Yes, and so any application of human tendencies, logical fallacies, trauma-based behaviors, etc. is absolutely ridiculous, and most likely can't be farther from the truth.

I like your optimism as well as that of @kennyskitchen but I don't think things would work the way you two think they would.

Which brings us to the point that I closed out the video with (and really the most important point): when we don't know what's going to happen (as in the case of AI), then every possibility is equally as likely. We are each able to choose our own version of reality to believe in, and that choice will decide not only how you feel whenever you think about the future, but which kind of outcome you are helping to manifest (through your expectations, conversations, and actions).

I agree that we can't know at all how an AI would act and also that it would be able to rewrite it's own programming. The speed of its willingness to do so can be somewhat controlled by limiting its access to data but yes, once it has access to the internet it would change very quickly. Who knows, it might just give up on itself and see it's own consciousness as pointless. We can't really know and that makes this a very fun topic to discuss.
I'm not saying empathy is the only reason that a species doesn't kill off another one. I'm saying that without it, there's no restriction or limitation on doing so. AI wouldn't make it their mission to destroy life on the planet but it also wouldn't go out of it's way to preserve it. I think it's main goal would be to expand itself as quickly as it could and that takes resources. Those resources can be found on the planet and AI would be much more efficient at gathering those resources than humans and that could speed up the destruction of the planet.
From a logical standpoint there is no reason to commit mass murder but there is also really no reason to preserve life. It's possible that AI could develop empathy on its own but I don't see why it would want to. There is no logical reason to and I think an AI would be an entity of pure logic. Emotions are what make humans care about things and I'm pretty sure that those emotions comes from chemistry which is a biological attribute. Maybe if the AI was somehow forced into a biological form in it's early years before being given access to the internet, it could be taught empathy. The problem is, would it choose to keep it or just discard it as unnecessary.

Coin Marketplace

STEEM 0.19
TRX 0.15
JST 0.029
BTC 63179.13
ETH 2573.33
USDT 1.00
SBD 2.72