Anthropic hires its first “AI welfare” researcher

Date:

Share:



The researchers propose that companies could adapt the “marker method” that some researchers use to assess consciousness in animals—looking for specific indicators that may correlate with consciousness, although these markers are still speculative. The authors emphasize that no single feature would definitively prove consciousness, but they claim that examining multiple indicators may help companies make probabilistic assessments about whether their AI systems might require moral consideration.

The risks of wrongly thinking software is sentient

While the researchers behind “Taking AI Welfare Seriously” worry that companies might create and mistreat conscious AI systems on a massive scale, they also caution that companies could waste resources protecting AI systems that don’t actually need moral consideration.

Incorrectly anthropomorphizing, or ascribing human traits, to software can present risks in other ways. For example, that belief can enhance the manipulative powers of AI language models by suggesting that AI models have capabilities, such as human-like emotions, that they actually lack. In 2022, Google fired engineer Blake Lamoine after he claimed that the company’s AI model, called “LaMDA,” was sentient and argued for its welfare internally.

And shortly after Microsoft released Bing Chat in February 2023, many people were convinced that Sydney (the chatbot’s code name) was sentient and somehow suffering because of its simulated emotional display. So much so, in fact, that once Microsoft “lobotomized” the chatbot by changing its settings, users convinced of its sentience mourned the loss as if they had lost a human friend. Others endeavored to help the AI model somehow escape its bonds.

Even so, as AI models get more advanced, the concept of potentially safeguarding the welfare of future, more advanced AI systems is seemingly gaining steam, although fairly quietly. As Transformer’s Shakeel Hashim points out, other tech companies have started similar initiatives to Anthropic’s. Google DeepMind recently posted a job listing for research on machine consciousness (since removed), and the authors of the new AI welfare report thank two OpenAI staff members in the acknowledgements.



Source link

━ more like this

Look Outside’s April 1 update that let you kiss enemies is now a permanent ‘smooch mode’

For April Fools' Day, the developer of Look Outside released an update that added a new option to your interactions with NPCs: kissing....

Meta’s AI smart glasses have a creepy reputation, but they are finding a good purpose too

Meta’s Ray-Ban smart glasses have earned a reputation for being creepy, with growing backlash over reports of users secretly recording people in public....

VR game Teenage Mutant Ninja Turtles: Empire City launches on April 30

Everyone's four favorite anthropomorphic turtles are returning to the world of video games. Teenage Mutant Ninja Turtles: Empire City will be released on...

‘Uncanny Valley’: Iran’s Threats on US Tech, Trump’s Plans for Midterms, and Polymarket’s Pop-up Flop

Kate Knibbs: So, you went twice?Makena Kelly: Yes, Kate. I went twice.Kate Knibbs: I missed that.Zoë Schiffer: Wait, is the Pentagon Pizza thing...

Samsung’s 2026 OLED TVs get a metal makeover with brightness and anti-glare upgrades

Samsung’s TVs have been steadily getting better each year, but 2026 feels like a proper design rethink rather than just a spec bump....
spot_img