Hacker News with Generative AI: Computer Science

Cursed fire or #define black magic (ssloy.github.io)
Have you ever wondered whether it is possible to write fully functional code using only the #define directive in C?
Relaxed Radix Balanced Trees (2024) (horne-khan.com)
I’m adding immutable vectors to my language, Ivan, and needed to pick a suitable data structure to implement them.
Tensor evolution: A framework for fast tensor computations using recurrences (arxiv.org)
This paper introduces a new mathematical framework for analysis and optimization of tensor expressions within an enclosing loop.
Catalytic computing taps the full power of a full hard drive (quantamagazine.org)
Ten years ago, researchers proved that adding full memory can theoretically aid computation. They’re just now beginning to understand the implications.
XOR (greenend.org.uk)
Recently I was called on to explain the ‘XOR’ operator to somebody who vaguely knew of its existence, but didn’t have a good intuition for what it was useful for and how it behaved.
This New Algorithm for Sorting Books or Files Is Close to Perfection (wired.com)
Computer scientists often deal with abstract problems that are hard to comprehend, but an exciting new algorithm matters to anyone who owns books and at least one shelf.
Native Sparse Attention: Hardware-Aligned, Natively Trainable Sparse Attention (arxiv.org)
Long-context modeling is crucial for next-generation language models, yet the high computational cost of standard attention mechanisms poses significant computational challenges.
Ask HN: What fiction/non-fiction book should everyone read on the topic of CS? (ycombinator.com)
This should be interpreted in the broadest sense to include books like both American Kingpin and Practical Vim as examples.
Reasoning Without Hesitating: Efficient Cot Through Certainty Probing (hao-ai-lab.github.io)
We observe reasoning models often exhibit poor token efficiency: they waste many tokens second-guessing themselves. We develop Dynasor-CoT, a certainty-based approach for dynamically allocating inference compute for reasoning models.
EnigmaEval: A Benchmark of Long Multimodal Reasoning Challenges (arxiv.org)
As language models master existing reasoning benchmarks, we need new challenges to evaluate their cognitive frontiers.
Books by Niklaus Wirth (hansotten.com)
Niklaus Wirth is a gifted writer. His easy to read style and the simplicity that must have taken so much effort to achieve makes his books jewels in the often so obscure computer science world.
(Ab)using general search algorithms on dynamic optimization problems (2023) (dubovik.eu)
In retrospect, my most ambitious blog yet. As it goes, I was reading “Artificial Intelligence. A Modern Approach” the other day. In one of the earlier chapters the authors discuss general search algorithms: breadth-first search, depth-first search, uniform-cost search (Dijkstra), and variations of those. A bit later they also cover Monte Carlo tree search as a way of finding approximate solutions in big state spaces.
My Time at MIT (blogspot.com)
Twenty years ago, in 2004-2005, I spent a year at MIT’s Computer Science department as a postdoc working with Professor Nancy Lynch.
Show HN: A GPU-accelerated binary vector index (rlafuente.com)
Are LLMs able to play the card game Set? (github.com/vnglst)
The Joy of Nand2Tetris (tristanrhodes.com)
For the last month, I have been working through Nand2Tetris: The Elements of Computing Systems; Building a Modern Computer from First Principles.
Ask HN: Seeking Book (ycombinator.com)
Hello All,<p>Many years ago I came across a book from a female Researcher which was basically using low-level mathematics to model low-level type 'circuits' or 'functions'. Not circuits in the electrical sense but I do know this researcher eventually did work for some Intelligence Agencies as part of extensions to her work. Problem I have is I have searched high and low via various search engines though with proliferation of AI texts it has increased the noise level.
How do modern compilers choose which variables to put in registers? (stackexchange.com)
How do modern compilers choose which variables to put in registers?
What if Eye...? (eyes.mit.edu)
We created a virtual petri dish where digital creatures evolve eyes from scratch, replaying millions of years of evolution.
Why cryptography is not based on NP-complete problems (blintzbase.com)
Cryptography is not based on NP-complete problems - let me explain why.
Questioning the Criteria for Evaluating Non-Cryptographic Hash Functions (cacm.acm.org)
Computing practitioners encounter hash functions almost every day, although they may not necessarily be the center of attention.
Basis of the Kalman Filter [pdf] (github.com/tpn)
Understanding the Basis of the Kalman Filter Via a Simple and Intuitive Derivation (2012).pdf
Tiny Pointers (arxiv.org)
This paper introduces a new data-structural object that we call the tiny pointer.
The biggest microcode attack in our history is underway (theregister.com)
When your state machines are vulnerable, all bets are off
Scaling up test-time compute with latent reasoning: A recurrent depth approach (arxiv.org)
We study a novel language model architecture that is capable of scaling test-time computation by implicitly reasoning in latent space.
Undergraduate shows that searches within hash tables can be much faster (quantamagazine.org)
A young computer scientist and two colleagues show that searches within data structures called hash tables can be much faster than previously deemed possible.
An Infinitely Large Napkin [pdf] (2019) (venhance.github.io)
Baffled by generational garbage collection – wingolog (wingolog.org)
Usually in this space I like to share interesting things that I find out; you might call it a research-epistle-publish loop. Today, though, I come not with answers, but with questions, or rather one question, but with fractal surface area: what is the value proposition of generational garbage collection?
Regex is almost all you need (lookingatcomputer.substack.com)
When it comes to secret detection, regex is all you need. Almost.
Demystifying Long Chain-of-Thought Reasoning in LLMs (arxiv.org)
Scaling inference compute enhances reasoning in large language models (LLMs), with long chains-of-thought (CoTs) enabling strategies like backtracking and error correction.