We all know and see the issue with morals being subjective, with there being no objective and absolute moral pinnacle/source.
On the TV show, Marvel's Agents of Shield, the plot line has become very interesting:
An AI, AIDA, has become fully human ( with the help of a book that taps into the energy of the universe and other realms) and this latest episode addressed the issue of Morals.
As an AI she was programmed not to hurt humans but that subjectivity of that view meant that she could do bad things in the short term if the long term end result was, according to her programming, better for them.
So she did a lot of crap.
Now, as human, she was exposed to empathy and one of the characters tried to make her do the right thing ( save lives even though she was sacred of the consequences) by using her FEELINGS of empathy to motivate her to save them.
Which she did, her feelings of empathy and compassion ( very new feelings for her) overruled her self-preservation.
For a BIT.
When things didn't go her way emotionally ( the character she loved spurned her advances because he loves another), her EMOTIONAL pain overrode her EMOTIONAL empathy and, well, you can guess what happend after that ( deaths).
See, without an objective and absolute moral then right and wrong is entirely subjective to either the reason of the moment or the emotion of the moment, which means that when those things chance ( as the always do) the moral code changes.
On Morals
-
- Board Moderator
- Posts: 9224
- Joined: Fri Mar 25, 2011 12:29 pm
- Christian: Yes
- Sex: Male
- Creation Position: Theistic Evolution
- Location: Ontario, Canada
- Nessa
- Ultimate Member
- Posts: 3593
- Joined: Sun Jul 05, 2015 7:10 pm
- Christian: Yes
- Creation Position: Undecided
Re: On Morals
Well, according to Sam Harris there is a 'moral landscape' where science can answer these questions...PaulSacramento wrote:We all know and see the issue with morals being subjective, with there being no objective and absolute moral pinnacle/source.
On the TV show, Marvel's Agents of Shield, the plot line has become very interesting:
An AI, AIDA, has become fully human ( with the help of a book that taps into the energy of the universe and other realms) and this latest episode addressed the issue of Morals.
As an AI she was programmed not to hurt humans but that subjectivity of that view meant that she could do bad things in the short term if the long term end result was, according to her programming, better for them.
So she did a lot of crap.
Now, as human, she was exposed to empathy and one of the characters tried to make her do the right thing ( save lives even though she was sacred of the consequences) by using her FEELINGS of empathy to motivate her to save them.
Which she did, her feelings of empathy and compassion ( very new feelings for her) overruled her self-preservation.
For a BIT.
When things didn't go her way emotionally ( the character she loved spurned her advances because he loves another), her EMOTIONAL pain overrode her EMOTIONAL empathy and, well, you can guess what happend after that ( deaths).
See, without an objective and absolute moral then right and wrong is entirely subjective to either the reason of the moment or the emotion of the moment, which means that when those things chance ( as the always do) the moral code changes.
https://www.samharris.org/the-moral-landscape
maybe we can dance in the fields like Maria did from Sound of music
-
- Board Moderator
- Posts: 9224
- Joined: Fri Mar 25, 2011 12:29 pm
- Christian: Yes
- Sex: Male
- Creation Position: Theistic Evolution
- Location: Ontario, Canada
Re: On Morals
Sam tries and I understand why, but the evidence is not there for anything BUT a subjective worldview.