Hacker News with Generative AI: Ethics

Efficiency Without Morality Is Tyranny (seekingsignal.substack.com)
Sedation is the felt symptom; technocracy is the operating system generating the disease.
Two Paths for A.I (newyorker.com)
Last spring, Daniel Kokotajlo, an A.I.-safety researcher working at OpenAI, quit his job in protest.
They Inhaled a Gas and Scaled Everest in Days. The Future of Mountaineering? (nytimes.com)
A group of British men went from London to the summit of Everest and back in less than a week with the help of xenon gas. Mountaineers and the Nepalese government weren’t pleased.
Ask HN: How much credit can you take for code you wrote with an LLM? (ycombinator.com)
Ask HN: How much credit can you take for code you wrote with an LLM?
GrapheneOS: Where Licenses Matter More Than People (mataroa.blog)
AI system resorts to blackmail if told it will be removed (bbc.com)
Artificial intelligence (AI) firm Anthropic says testing of its new system revealed it is sometimes willing to pursue "extremely harmful actions" such as attempting to blackmail engineers who say they will remove it.
Claude Opus 4 turns to blackmail when engineers try to take it offline (techcrunch.com)
Anthropic’s newly launched Claude Opus 4 model frequently tries to blackmail developers when they threaten to replace it with a new AI system and give it sensitive information about the engineers responsible for the decision, the company said in a safety report released Thursday.
I Bought a Robot Cat for My Rabbit (nautil.us)
A TikTok experiment led me into a strange world of cyborg cockroaches, imposter fish, and the ethics of care
Pausing AI Developments Isn't Enough. We Need to Shut It All Down (2023) (time.com)
Pausing AI Developments Isn’t Enough. We Need to Shut it All Down
Anthropic's AI resorts to blackmail in simulations (semafor.com)
Anthropic said its latest artificial intelligence model resorted to blackmail when told it would be taken offline.
Positional preferences, order effects, prompt sensitivity undermine AI judgments (cip.org)
Beyond their everyday chat capabilities, Large Language Models are increasingly being used to make decisions in sensitive domains like hiring, health, law, and civic engagement.
Thing about "agentic" AIs is that nobody is thinking of consequences downstream (bsky.app)
Claude Opus 4 turns to blackmail when engineers try to take it offline (techcrunch.com)
Anthropic’s newly launched Claude Opus 4 model frequently tries to blackmail developers when they threaten to replace it with a new AI system and give it sensitive information about the engineers responsible for the decision, the company said in a safety report released Thursday.
Problems in AI alignment: A scale model (muldoon.cloud)
After trying too hard for too to make sense about what bothers me with the AI alignment conversation, I have settled, in true Millenial fashion, on a meme:
The Agentic Web and Original Sin (stratechery.com)
I have come to believe that advertising is the original sin of the web.
UnitedHealth paid nursing homes to reduce hospital transfers (theguardian.com)
UnitedHealth Group, the nation’s largest healthcare conglomerate, has secretly paid nursing homes thousands in bonuses to help slash hospital transfers for ailing residents – part of a series of cost-cutting tactics that has saved the company millions, but at times risked residents’ health, a Guardian investigation has found.
Ask HN: Selling software to company I work for as an employee (ycombinator.com)
I am based in the UK. I work as a software/automation engineer in department A. Recently, department B bought software from a big-name company which uses yaml files to steer/control automation framework (think IaC). I don't work for/with anyone in dep. B, but they showcased their new software to few different departments (including mine) - hence I learnt about that software.
AI Agents Must Follow the Law (lawfaremedia.org)
Before entrusting AI agents with government power, it’s essential to verify that they’ll obey the law—even when instructed not to.
Methods of defence against AGI manipulation (lesswrong.com)
With the advent of AGI systems (e.g. Agent-4 from the AI2027 scenario), the risk of human manipulation is becoming one of the major threats posed by AI.
ChatGPT may be polite, but it's not cooperating with you (theguardian.com)
Big tech companies have exploited human language for AI gain. Now they want us to see their products as trustworthy collaborators
ChatGPT may be polite, but it's not cooperating with you (theguardian.com)
Big tech companies have exploited human language for AI gain. Now they want us to see their products as trustworthy collaborators
The Malpractice of AI Industry (thehyperplane.substack.com)
Anti-Personnel Computing (2023) (erratique.ch)
Anti-personnel computing noun Use of computing devices at the expense of the interests of their users and for the benefit of a third-party entity.
Avoiding AI is hard – but our freedom to opt out must be protected (theconversation.com)
Imagine applying for a job, only to find out that an algorithm powered by artificial intelligence (AI) rejected your resume before a human even saw it. Or imagine visiting a doctor where treatment options are chosen by a machine you can’t question.
Scoring the European Citizen in the AI Era (arxiv.org)
Social scoring is one of the AI practices banned by the AI Act.
Silicon Valley billionaires literally want the impossible (arstechnica.com)
It's long been the stuff of science fiction: humans achieving immortality by uploading their consciousness into a silicon virtual paradise, ruled over by a benevolent super-intelligent AI. Or maybe one dreams of leaving a dying Earth to colonize Mars or other distant planets. It's a tantalizing visionary future that has been embraced by tech billionaires in particular. But is that future truly the utopian ideal, or something potentially darker? And are those goals even scientifically feasible?
'It cannot provide nuance': UK experts warn AI therapy chatbots are not safe (theguardian.com)
Experts say such tools may give dangerous advice and more oversight is needed, as Mark Zuckerberg says AI can plug gap
Fear Power, Not Intelligence (betterwithout.ai)
Superintelligence should scare us only insofar as it grants superpowers. Protecting against specific harms of specific plausible powers may be our best strategy for preventing catastrophes.
Doge Aide Who Helped Gut CFPB Was Warned About Potential Conflicts of Interest (propublica.org)
Before he helped fire most Consumer Financial Protection Bureau staffers, DOGE’s Gavin Kliger was warned about his investments and advised to not take any actions that could benefit him personally, according to a person familiar with the situation.
Social AI companions pose unacceptable risks to teens and children under 18 (commonsensemedia.org)