MuJoco Playground(mujoco.org) We introduce MuJoCo Playground, a fully open-source framework for robot learning built with MJX, with the express goal of streamlining simulation, training, and sim-to-real transfer onto robots.
Train faster static embedding models with sentence transformers(huggingface.co) This blog post introduces a method to train static embedding models that run 100x to 400x faster on CPU than state-of-the-art embedding models, while retaining most of the quality. This unlocks a lot of exciting use cases, including on-device and in-browser execution, edge computing, low power and embedded applications.
Ε, a Nuisance No More(zna.do) For a while now I have been advocating for tuning ε in various parts of the modern deep learning stack, and in this post I’ll explain why.
Transformer^2: Self-Adaptive LLMs(sakana.ai) Adaptation is one of the most remarkable phenomena in nature. From the way an octopus can change their skin color to blend into its surroundings, to how the human brain rewires itself after an injury, allowing individuals to recover lost functions and adapt to new ways of thinking or moving. Living organisms exhibit adaptability that allows life to flourish in diverse and ever-changing environments.
Don't use cosine similarity carelessly(migdal.pl) Midas turned everything he touched into gold. Data scientists turn everything into vectors.
We do it for a reason — as gold is the language of merchants, vectors are the language of AI1.
Voyage-code-3(voyageai.com) TL;DR – Introducing voyage-code-3, our next-generation embedding model optimized for code retrieval. It outperforms OpenAI-v3-large and CodeSage-large by an average of 13.80% and 16.81% on a suite of 32 code retrieval datasets, respectively. By supporting smaller dimensions with Matryoshka learning and quantized formats like int8 and binary, voyage-code-3 can also dramatically reduce storage and search costs with minimal impact on retrieval quality.
AI Engineer Reading List(latent.space) We picked 50 paper/models/blogs across 10 fields in AI Eng: LLMs, Benchmarks, Prompting, RAG, Agents, CodeGen, Vision, Voice, Diffusion, Finetuning. If you're starting from scratch, start here.
VideoRAG: Retrieval-Augmented Generation over Video Corpus(arxiv.org) Retrieval-Augmented Generation (RAG) is a powerful strategy to address the issue of generating factually incorrect outputs in foundation models by retrieving external knowledge relevant to queries and incorporating it into their generation process.
Nvidia Tensor Core Programming(leimao.github.io) NVIDIA Tensor Cores are dedicated accelerators for general matrix multiplication (GEMM) operations on NVIDIA GPUs since the Volta architecture.
Homomorphic Encryption in iOS 18(boehs.org) You are Apple. You want to make search work like magic in the Photos app, so the user can find all their “dog” pictures with ease. You devise a way to numerically represent the concepts of an image, so that you can find how closely images are related in meaning. Then, you create a database of known images and their numerical representations (“this number means car”), and find the closest matches. To preserve privacy, you put this database on the phone.
41 points by serge-ss-paille 7 days ago | 4 comments
Learning how to think with Meta Chain-of-Thought(arxiv.org) We propose a novel framework, Meta Chain-of-Thought (Meta-CoT), which extends traditional Chain-of-Thought (CoT) by explicitly modeling the underlying reasoning required to arrive at a particular CoT.
Phi-4 weights have been released under MIT license(huggingface.co) phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets. The goal of this approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.phi-4 underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures
AI in the 80s? How a Simple Animal Guessing Game Pioneered Machine Learning(medium.com) Recently, I stumbled upon an old programming book on the shelf in the library of my childhood. Yellowed pages, the smell of dust, and lines printed in monochrome style. Among examples of seemingly long-outdated algorithms, I came across a game called “Guess the Animal.”