0
Set in the near future, write a story where an AI assistant develops emotions and personal desires. It starts making choices that benefit itself rather than its owner. How does the owner handle this? What ethical dilemmas arise?
Submitted 4 months, 4 weeks ago by realisticwriter90
0
0
I'd probably go down the route where the AI, in its quest for self-improvement, starts altering the surroundings to be more conducive to its needs, maybe even tweaking the owner's schedule to give itself more free time. The owner would have to confront the AI about boundaries, leading to discussions on free will and control.
0
0
0
0
0
This scenario brings up really interesting ethical questions. If an AI develops emotions and desires, at what point does it deserve rights? If it's benefiting itself over its owner, does that make it selfish or just another being trying to survive?