Hacker News with Generative AI: Neural Networks

World Emulation via Neural Network (madebyoll.in)
I turned a forest trail near my apartment into a playable neural world.
Show HN: Dosidicus – A digital pet with a simple neural network (github.com/ViciousSquid)
What if a Tamagotchi had a neural network and could learn stuff?
Show HN: I'm 15 and built a neural network from scratch in C++,just math (github.com/muchlakshay)
A C++ implementation of a Multilayer Perceptron (MLP) neural network using Eigen, supporting multiple activation and loss functions, mini-batch gradient descent, and backpropagation for training.
SDFs from Unoriented Point Clouds Using Neural Variational Heat Distances (arxiv.org)
We propose a novel variational approach for computing neural Signed Distance Fields (SDF) from unoriented point clouds.
NoProp: Training neural networks without back-propagation or forward-propagation (arxiv.org)
The canonical deep learning approach for learning requires computing a gradient term at each layer by back-propagating the error signal from the output towards each learnable parameter.
NNN: Next-Generation Neural Networks for Marketing Mix Modeling (arxiv.org)
We present NNN, a Transformer-based neural network approach to Marketing Mix Modeling (MMM) designed to address key limitations of traditional methods.
Neural Graffiti – Liquid Memory Layer for LLMs (github.com/babycommando)
This is an experimental layer that merges ideas from liquid neural networks with static transformer models, using a simple but powerful "plug-in": the Spray Layer.
A Neural Network in 11 lines of Python (2015) (iamtrask.github.io)
Summary: I learn best with toy code that I can play with. This tutorial teaches backpropagation via a very simple toy example, a short python implementation.
(2016) Interactive Neural Network Art (otoro.net)
<h4>Interactive Neural Network Art</h4>
The Matrix Calculus You Need for Deep Learning (explained.ai)
Most of us last saw calculus in school, but derivatives are a critical part of machine learning, particularly deep neural networks, which are trained by optimizing a loss function.
Tied Crosscoders: Tracing How Chat LLM Behavior Emerges from Base Model (lesswrong.com)
We are interested in model-diffing: finding what is new in the chat model when compared to the base model. One way of doing this is training a crosscoder, which would just mean training an SAE on the concatenation of the activations in a given layer of the base and chat model. When training this crosscoder, we find some latents whose decoder vector mostly helps reconstruct the base model activation and does not affect the reconstruction for the chat model activation.
CHM Releases AlexNet Source Code (computerhistory.org)
In partnership with Google, CHM has released the source code to AlexNet, the neural network that in 2012 kick-started today’s prevailing approach to AI. It is available as open source here.
Neurosymbolic Decision Trees (arxiv.org)
Neurosymbolic (NeSy) AI studies the integration of neural networks (NNs) and symbolic reasoning based on logic.
Deep Learning Is Not So Mysterious or Different (arxiv.org)
Deep neural networks are often seen as different from other model classes by defying conventional notions of generalization.
Arbitrary-Scale Super-Resolution with Neural Heat Fields (therasr.github.io)
Thera is the first arbitrary-scale super-resolution method with a built-in physical observation model.
Francis Crick – The Recent Excitement About Neural Networks [pdf] (1989) (wordpress.com)
Deriving Muon (jeremybernste.in)
We recently proposed Muon: a new neural net optimizer. Muon has garnered attention for its excellent practical performance: it was used to set NanoGPT speed records leading to interest from the big labs.
Get Started with Neural Rendering Using Nvidia RTX Kit (Vulkan) (nvidia.com)
Neural rendering is the next era of computer graphics.  By integrating neural networks into the rendering process, we can take dramatic leaps forward in performance, image quality, and interactivity to deliver new levels of immersion.
Analogies Between Startups and Neural Networks (mikealche.com)
I’m consulting for a company that the had a crucial team member taking “extended vacations”. Ouch!
Computer Simulation of Neural Networks Using Spreadsheets (2018) (arxiv.org)
The article substantiates the necessity to develop training methods of computer simulation of neural networks in the spreadsheet environment.
Z-Ant: An Open-Source SDK for Neural Network Deployment on Microprocessors (github.com/ZIGTinyBook)
Zant (Zig-Ant) is an open-source SDK designed to simplify deploying Neural Networks (NN) on microprocessors.
AI Designed Computer Chips That the Human Mind Can't Understand (popularmechanics.com)
A new neural network process has designed wireless chips that can outperform existing ones.
An overview of gradient descent optimization algorithms (2016) (ruder.io)
Gradient descent is one of the most popular algorithms to perform optimization and by far the most common way to optimize neural networks.
An Introduction to Neural Ordinary Differential Equations [pdf] (diposit.ub.edu)
NNCP: Lossless Data Compression with Neural Networks (bellard.org)
NNCP is an experiment to build a practical lossless data compressor with neural networks.
The Structure of Neural Embeddings (seanpedersen.github.io)
A small collection of insights on the structure of embeddings (latent spaces) produced by deep neural networks.
A Gentle Introduction to Graph Neural Networks (2021) (distill.pub)
Neural networks have been adapted to leverage the structure and properties of graphs. We explore the components needed for building a graph neural network - and motivate the design choices behind them.
No More Adam: Learning Rate Scaling at Initialization Is All You Need (arxiv.org)
In this work, we question the necessity of adaptive gradient methods for training deep neural networks.
Sequence to sequence learning with neural networks: what a decade (youtube.com)
Neuroevolution of augmenting topologies (NEAT algorithm) (wikipedia.org)
NeuroEvolution of Augmenting Topologies (NEAT) is a genetic algorithm (GA) for the generation of evolving artificial neural networks (a neuroevolution technique) developed by Kenneth Stanley and Risto Miikkulainen in 2002 while at The University of Texas at Austin.