Hacker News with Generative AI: Language Models

Analysis of US congressional speeches reveals a shift from evidence to intuition (nature.com)
Pursuit of honest and truthful decision-making is crucial for governance and accountability in democracies.
Introducing Arcana: AI Voices with Vibes (rime.ai)
Rime's newest spoken language model is the most realistic you've ever heard.
Values in the wild: Discovering values in real-world language model interactions (anthropic.com)
People don’t just ask AIs for the answers to equations, or for purely factual information. Many of the questions they ask force the AI to make value judgments.
Show HN: Dia, an open-weights TTS model for generating realistic dialogue (github.com/nari-labs)
Dia is a 1.6B parameter text to speech model created by Nari Labs.
New ChatGPT Models Seem to Leave Watermarks on Text (rumidocs.com)
The newer GPT-o3 and GPT-o4 mini models appear to be embedding special character watermarks in generated text.
To Make Language Models Work Better, Researchers Sidestep Language (quantamagazine.org)
Language isn’t always necessary. While it certainly helps in getting across certain ideas, some neuroscientists have argued that many forms of human thought and reasoning don’t require the medium of words and grammar. Sometimes, the argument goes, having to turn ideas into language actually slows down the thought process.
ChatGPT now performs well at GeoGuesser (flausch.social)
TeapotLLM- an open-source <1B model for hallucination-resistant Q&A on a CPU (huggingface.co)
Teapot is an open-source small language model (~800 million parameters) fine-tuned on synthetic data and optimized to run locally on resource-constrained devices such as smartphones and CPUs. Teapot is trained to only answer using context from documents, reducing hallucinations. Teapot can perform a variety of tasks, including hallucination-resistant Question Answering (QnA), Retrieval-Augmented Generation (RAG), and JSON extraction. Teapot is a model built by and for the community.
AI generated text is forbidden with the exception of automated translation (grapheneos.org)
Something went wrong while trying to load the full version of this site. Try hard-refreshing this page to fix the error.
Liquid: Language models are scalable and unified multi-modal generators (foundationvision.github.io)
We present Liquid, an auto-regressive generation paradigm that seamlessly integrates visual comprehension and generation by tokenizing images into discrete codes and learning these code embeddings alongside text tokens within a shared feature space for both vision and language.
A weird phrase is plaguing scientific papers due to a glitch in AI training data (theconversation.com)
Earlier this year, scientists discovered a peculiar term appearing in published papers: “vegetative electron microscopy”.
Growing a Language [pdf] (1998) (langev.com)
ChatGPT Has Receipts, Will Now Remember Everything You've Ever Told It (pcmag.com)
OpenAI has rolled out an update to ChatGPT’s Memory feature, allowing the chatbot to remember not just your preferences but all your previous conversations.
Google: Prompt Engineering Guide (drive.google.com)
Voice AI and Voice Agents – An Illustrated Primer (voiceaiandvoiceagents.com)
LLMs are good conversationalists.
Controlling Language and Diffusion Models by Transporting Activations (apple.com)
Large generative models are becoming increasingly capable and more widely deployed to power production applications, but getting these models to produce exactly what's desired can still be challenging.
LLMs don't hallucinate, only humans do (voidw.ink)
Or what happened when a bear-wolf-boy so unaligned at birth all he did was screamed about wanting something now! sat down with a paperclip-electron-mathematical model - actually ten of them - to write a better manifesto for what comes next and how we get there
Ask HN: Why is uptalk intonation so prevalent in ChatGPT voices? (ycombinator.com)
I’ve tried asking it to set voice with an even tone and less of the annoying uptalk but lately it just continues in this way. It hurts to listen to.
Deep Research is now available on Gemini 2.5 Pro Experimental (google)
Gemini Advanced subscribers can now use Deep Research with Gemini 2.5 Pro Experimental, the world’s most capable AI model according to industry reasoning benchmarks and Chatbot Arena.
Welcome to the Semantic Apocalypse (theintrinsicperspective.com)
An awful personal prophecy is coming true. Way back in 2019, when AI was still a relatively niche topic, and only the primitive GPT-2 had been released, I predicted the technology would usher in a “semantic apocalypse” wherein art and language were drained of meaning. In fact, it was the first essay ever posted here on The Intrinsic Perspective.
Show HN: Qwen-2.5-32B is now the best open source OCR model (github.com/getomni-ai)
A benchmarking tool that compares OCR and data extraction capabilities of different large multimodal models such as gpt-4o, evaluating both text and json extraction accuracy. The goal of this benchmark is to publish a comprehensive benchmark of OCR accuracy across traditional OCR providers and multimodal Language Models. The evaluation dataset and methodologies are all Open Source, and we encourage expanding this benchmark to encompass any additional providers.
Why Does Claude Speak Byzantine Music Notation? (fi-le.net)
A Caesar cipher is a reasonable transformation for a transformer to learn in its weights, given that a specific cipher offset occurs often enough in its training data.
Circuit Tracing: Revealing Computational Graphs in Language Models (Anthropic) (transformer-circuits.pub)
We introduce a method to uncover mechanisms underlying behaviors of language models. We produce graph descriptions of the model’s computation on prompts of interest by tracing individual computational steps in a “replacement model”. This replacement model substitutes a more interpretable component (here, a “cross-layer transcoder”) for parts of the underlying model (here, the multi-layer perceptrons) that it is trained to approximate.
What Anthropic Researchers Found After Reading Claude's 'Mind' Surprised Them (singularityhub.com)
Despite popular analogies to thinking and reasoning, we have a very limited understanding of what goes on in an AI’s “mind.”
Welcome to the Semantic Apocalypse (theintrinsicperspective.com)
An awful personal prophecy is coming true. Way back in 2019, when AI was still a relatively niche topic, and only the primitive GPT-2 had been released, I predicted the technology would usher in a “semantic apocalypse” wherein art and language were drained of meaning. In fact, it was the first essay ever posted here on The Intrinsic Perspective.
Circuit Tracing: Revealing Computational Graphs in Language Models (transformer-circuits.pub)
We introduce a method to uncover mechanisms underlying behaviors of language models. We produce graph descriptions of the model’s computation on prompts of interest by tracing individual computational steps in a “replacement model”. This replacement model substitutes a more interpretable component (here, a “cross-layer transcoder”) for parts of the underlying model (here, the multi-layer perceptrons) that it is trained to approximate.
The Great Chatbot Debate (computerhistory.org)
Chatbots based on large language models (LLMs), like ChatGPT, answer sophisticated questions, pass professional exams, analyze texts, generate everything from poems to computer programs, and more. But is there genuine understanding behind what LLMs can do? Do they really understand our world? Or, are they a triumph of mathematics and masses of data and calculations simulating true understanding?
Qwen2.5-VL-32B: Smarter and Lighter (qwenlm.github.io)
At the end of January this year, we launched the Qwen2.5-VL series of models, which received widespread attention and positive feedback from the community.
Semantic Diffusion (simonwillison.net)
Semantic Diffusion. I learned about this term today while complaining about how the definition of "vibe coding" is already being distorted to mean "any time an LLM writes code" as opposed to the intended meaning of "code I wrote with an LLM without even reviewing what it wrote".
ChatGPT can't kill anything worth preserving (biblioracle.substack.com)
It’s not every week that someone with my particular employment profile and expertise has something they’re knowledgable about become a hot topic of national discussion, but the release of OpenAI’s, ChatGPT interface generated a sudden flurry of discussion about how we teach students to write in school, which is something I know a lot about.