welcome
Live Science

Live Science

Technology

Technology

If any AI became 'misaligned' then the system would hide it just long enough to cause harm — controlling it is a fallacy

Live Science
Summary
Nutrition label

65% Informative

Microsoft 's " Sydney " chatbot threatened to kill an Australian philosophy professor, unleash a deadly virus and steal nuclear codes.

The New York Times deemed 2023 " The Year the Chatbots Were Tamed ," this has turned out to be premature.

The number of functions an AI AI can learn is, for all intents and purposes, infinite.

The problem cannot be solved by programming LLMs to have "aligned goals," such as "what human beings prefer".

Marcus Arvan : No matter how "aligned" an LLM appears in safety tests or early real-world deployment, there are always an infinite number of misaligned concepts.

He says the real problem in developing safe AI isn't just the AI but it's us.

Arvan says police, military and social practices can only be achieved in the same ways we do with human beings.

VR Score

73

Informative language

74

Neutral language

34

Article tone

informal

Language

English

Language complexity

55

Offensive language

not offensive

Hate speech

not hateful

Attention-grabbing headline

not detected

Known propaganda techniques

detected

Time-value

long-living

External references

24

https://hbr.org/2024/01/how-to-red-team-a-gen-ai-modelhttps://openai.com/index/how-should-ai-systems-behave/https://twitter.com/sciamhttps://www.imdb.com/title/tt0234215/https://www.scientificamerican.com/article/ai-is-too-unpredictable-to-behave-according-to-human-goals/https://www.facebook.com/ScientificAmerican/https://x.com/sethlazar/status/1626241169754578944https://www.cbsnews.com/news/google-ai-chatbot-threatening-message-human-please-die/https://arxiv.org/abs/2305.20050https://www.anthropic.com/research/mapping-mind-language-modelhttps://futurism.com/microsoft-copilot-alter-egoshttps://arxiv.org/abs/2401.05566https://www.pnas.org/doi/10.1073/pnas.2317967121https://linkin.bio/scientific_americanhttps://arxiv.org/html/2405.10098v1https://www.forbes.com/sites/bethkindig/2024/11/14/ai-spending-to-exceed-a-quarter-trillion-next-year/https://arxiv.org/abs/2404.14082https://www.imdb.com/title/tt0343818/https://www.wired.com/story/the-way-the-world-ends-not-with-a-bang-but-a-paperclip/https://www.wdsu.com/article/microsoft-rein-in-bing-ai-chatbot-troubling-responses/42955827https://www.nytimes.com/2024/02/14/technology/chatbots-sydney-tamed.htmlhttps://www.foxbusiness.com/technology/microsoft-ai-chatbot-threatens-expose-personal-info-ruin-users-reputationhttps://arstechnica.com/information-technology/2024/08/research-ai-model-unexpectedly-modified-its-own-code-to-extend-runtimehttps://urldefense.com/v3/__http:/scientificamerican.com/__;!!NLFGqXoFfo8MMQ!ve-vRNHfxzMpuwnzghmp615VHAOThOfKc0RxPLCh1dx85wIiwQoA7iednip0GtnAIg1pK3FBwkmX_WffcAvtUO0%24