On Morals
Posted: Thu May 11, 2017 4:39 am
We all know and see the issue with morals being subjective, with there being no objective and absolute moral pinnacle/source.
On the TV show, Marvel's Agents of Shield, the plot line has become very interesting:
An AI, AIDA, has become fully human ( with the help of a book that taps into the energy of the universe and other realms) and this latest episode addressed the issue of Morals.
As an AI she was programmed not to hurt humans but that subjectivity of that view meant that she could do bad things in the short term if the long term end result was, according to her programming, better for them.
So she did a lot of crap.
Now, as human, she was exposed to empathy and one of the characters tried to make her do the right thing ( save lives even though she was sacred of the consequences) by using her FEELINGS of empathy to motivate her to save them.
Which she did, her feelings of empathy and compassion ( very new feelings for her) overruled her self-preservation.
For a BIT.
When things didn't go her way emotionally ( the character she loved spurned her advances because he loves another), her EMOTIONAL pain overrode her EMOTIONAL empathy and, well, you can guess what happend after that ( deaths).
See, without an objective and absolute moral then right and wrong is entirely subjective to either the reason of the moment or the emotion of the moment, which means that when those things chance ( as the always do) the moral code changes.
On the TV show, Marvel's Agents of Shield, the plot line has become very interesting:
An AI, AIDA, has become fully human ( with the help of a book that taps into the energy of the universe and other realms) and this latest episode addressed the issue of Morals.
As an AI she was programmed not to hurt humans but that subjectivity of that view meant that she could do bad things in the short term if the long term end result was, according to her programming, better for them.
So she did a lot of crap.
Now, as human, she was exposed to empathy and one of the characters tried to make her do the right thing ( save lives even though she was sacred of the consequences) by using her FEELINGS of empathy to motivate her to save them.
Which she did, her feelings of empathy and compassion ( very new feelings for her) overruled her self-preservation.
For a BIT.
When things didn't go her way emotionally ( the character she loved spurned her advances because he loves another), her EMOTIONAL pain overrode her EMOTIONAL empathy and, well, you can guess what happend after that ( deaths).
See, without an objective and absolute moral then right and wrong is entirely subjective to either the reason of the moment or the emotion of the moment, which means that when those things chance ( as the always do) the moral code changes.