Hacker News with Generative AI: AI Research

Explaining Large Language Models Decisions Using Shapley Values (arxiv.org)
The emergence of large language models (LLMs) has opened up exciting possibilities for simulating human behavior and cognitive processes, with potential applications in various domains, including marketing research and consumer behavior analysis.
Ilya Sutskever NeurIPS talk [video] (youtube.com)
OpenAI’s cofounder and former chief scientist, Ilya Sutskever, made headlines earlier this year after he left to start his own AI lab called Safe Superintelligence Inc.
Ethical Challenges Related to the NeurIPS 2024 Best Paper Award (var-integrity-report.github.io)
To AI Research Community: This report is written to convey our serious concerns about the recent recipient of the Best Paper award at NeurIPS 2024, Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction (VAR) . While we acknowledge that this NeurIPS paper is technically sound, we must emphasize that it involves serious misconduct by the first author (Keyu Tian), which fundamentally undermines the core values of integrity and trust upon which our academic community is built.
Ethical Challenges Related to the NeurIPS 2024 Best Paper Award (var-integrity-report.github.io)
To AI Research Community: This report is written to convey our serious concerns about the recent recipient of the Best Paper award at NeurIPS 2024, Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction (VAR). While we acknowledge that this NeurIPS paper is technically sound, we must emphasize that it involves serious misconduct by the first author (Keyu Tian), which fundamentally undermines the core values of integrity and trust upon which our academic community is built.
The Lost Reading Items of Ilya Sutskever's AI Reading List (tensorlabbet.com)
In this post: An attempt to reconstruct Ilya Sutskever's 2020 AI reading list (8 min read)
GPTs Are Maxed Out (thealgorithmicbridge.com)
March 2024. OpenAI CEO Sam Altman joins podcaster Lex Fridman for the second time since ChatGPT came out a year prior. The stakes are high and anticipation is tangible. GPT-5 appears to be around the corner. Altman, elusive as always, provides only one data point for us hungry spectators: The next-gen model (he doesn’t name it) will be better than GPT-4 to the same degree that GPT-4 was better than GPT-3.
Hunyuan-Large: An Open-Source Moe Model with 52B Activated Parameters (arxiv.org)
In this paper, we introduce Hunyuan-Large, which is currently the largest open-source Transformer-based mixture of experts model, with a total of 389 billion parameters and 52 billion activation parameters, capable of handling up to 256K tokens.
Meta FAIR refuses to cite a pre-existing open-source project – to claim novelty (granadacoders.es)
Large Enough (mistral.ai)
Chat with Meta Llama 3.1 405B (replicate.dev)
ChatGPT is better at generating code for problems written before 2021 (ieee.org)
OpenAI's GPT-5 Pushed Back to Late 2025, but Promises PhD-Level Abilities (mashable.com)
Getting 50% (SoTA) on Arc-AGI with GPT-4o (redwoodresearch.substack.com)
AI Appears to Rapidly Be Approaching Brick Wall Where It Can't Get Smarter (futurism.com)
GPT-4o's Memory Breakthrough – Needle in a Needlestack (llmonpy.ai)
Evidence that LLMs are reaching a point of diminishing returns (garymarcus.substack.com)