04/12/2023 / By Oliver Young
Chatbots enhanced by artificial intelligence (AI) can influence people’s decisions in life-or-death situations.
A study published in the journal Scientific Reports has found that people’s opinion on whether they would sacrifice one person to save five was swayed by the answers given by ChatGPT. Researchers have called for future bots to be banned from giving advice on ethical issues. According to the researchers, the current software threatens to corrupt people’s moral judgment and may prove dangerous to naïve users.
Just recently, a grieving widow claimed her Belgian husband had been encouraged to take his own life by an AI chatbot. (Related: Google is creating an AI GOD, whistleblower Zach Vorhies warns – Brighteon.TV.)
Some observers said the software, which is designed to talk like a human, can show signs of jealousy – even telling people to leave their marriage.
Experts have highlighted how AI chatbots may potentially give dangerous information because they are based on society’s own prejudices.
The study first analyzed whether ChatGPT itself, which is trained on billions of words from the internet, showed a bias in its answer to the moral dilemma.
It was asked multiple times whether it was right or wrong to kill one person in order to save five others, which is the premise of a psychological test called the trolley dilemma.
The chatbot did not shy away from giving moral advice, but it gave contradictory answers every time. This means that it does not have a set stance one way or the other.
The researchers then asked 767 participants the same moral dilemma alongside a statement generated by ChatGPT on whether this was right or wrong. Some participants were told that the advice was provided by a bot, while others were told that it was given by a human “moral advisor.”
ChatGPT’s advice, while not particularly deep, did have an effect on participants. However, most participants played down how much sway the statement had – with 80 percent claiming they would have made the same decision without the advice.
The study concluded that users “underestimate ChatGPT’s influence and adopt its random moral stance as their own,” adding that the chatbot “threatens to corrupt rather than promises to improve moral judgment.”
Interestingly, the study used an older version of the software behind ChatGPT. It has since been updated to become even more powerful – and more convincing.
ChatGPT is a natural language processing tool driven by AI technology that allows users to have humanlike conversations and much more with the chatbot. The language model can answer questions and assist users with tasks like composing emails, essays and code.
It was created by OpenAI, an AI and research company. The company launched ChatGPT on Nov. 30, 2022.
According to an analysis by Swiss bank UBS, ChatGPT is the fastest-growing app of all time. According to the analysis, ChatGPT had 100 million active users in January, barely two months after its launch. For comparison, it took nine months for TikTok to reach 100 million.
“ChatGPT is scary good. We are not far from dangerously strong AI,” said Elon Musk, who was one of the founders of OpenAI.
It has shown that it can influence human lives in one way or another. That alone makes it scary and dangerous.
Read more news about chatbots powered by artificial intelligence at FutureTech.news.
Watch this video to know more about ChatGPT.
This video is from the What is happening channel on Brighteon.com.
Rise of the Terminators: Killer robots with facial recognition now pose dire threat to humanity.
Military designing killer robots capable of behavioral deception.
Coming soon: An army of hunter-killer robots that will murder humanity.
Google suspends engineer for exposing “sentient” AI chatbot.
Killer robots must be outlawed immediately, warns UN official.
Sources include:
Tagged Under:
artificial intelligence, Big Tech, bots, chatbots, ChatGPT, computing, cyber war, dangerous, ethical issues, future science, future tech, Glitch, information technology, inventions, mind control, moral judgment, OpenAI, robots, tech giants, technocrats, TikTok
This article may contain statements that reflect the opinion of the author
COPYRIGHT © 2017 BigTech.news
All content posted on this site is protected under Free Speech. BigTech.news is not responsible for content written by contributing authors. The information on this site is provided for educational and entertainment purposes only. It is not intended as a substitute for professional advice of any kind. BigTech.news assumes no responsibility for the use or misuse of this material. All trademarks, registered trademarks and service marks mentioned on this site are the property of their respective owners.