Links for 2025-02-13
AI:
1. Training Deep Learning Models with Norm-Constrained LMOs—has the potential to significantly improve the efficiency and speed of training LLMs, allowing for the training of even larger and more complex models.
https://arxiv.org/abs/2502.075292. LLM Pretraining with Continuous Concepts
https://arxiv.org/abs/2502.085243. Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving —iteratively refines the prover through expert iteration, dramatically increasing the number of solved problems (e.g., 29.7K solved in Lean Workbook) and securing top rankings on benchmarks like PutnamBench.
https://arxiv.org/abs/2502.076404. RAGEN: A General-Purpose Reasoning Agent Training Framework
https://github.com/ZihanWang314/ragen/tree/main5. Unsupervised Predictive Memory in a Goal-Directed Agent [published in 2018]
https://arxiv.org/abs/1803.107606. CodeI/O: Condensing Reasoning Patterns via Code Input-Output Prediction
https://codei-o.github.io/7. Elon Musk says Grok 3 will be released in "a week or two" and it is "scary smart", displaying reasoning skills that outperform any other AI model that has been released
https://www.youtube.com/live/eV396ioBs3g?si=KOAokGapPj_Cb666&t=8118. Noam Shazeer, co-lead on Google's Gemini, says by 2030 there will be AI assistants in glasses that provide advice and solve problems for you in real time, as well as turning programmers into 10,000,000x engineers
https://youtu.be/v0gjI__RyCY?si=QHw1hrywgBvBnieQ&t=53909. Studies of Human Error Rate: "…skeptics often gesture to hallucinations, errors. An ideal symbolic system never makes such errors, therefore LLMs cannot truly "understand" even simple concepts like addition. See e.g. Evaluating the World Model Implicit in a Generative Model for this argument in the literature. However, such arguments reliably rule out human "understanding" as well! Studies within Human Reliability Analysis find startlingly high rates even for basic tasks, and even with double checking. Generally, the human reference class is too often absent (or assumed ideal) in AI discussions, and many LLM oddities have close parallels in psychology. If you're willing to look!"
https://www.lesswrong.com/posts/9unBWgRXFT5BpeSdb/studies-of-human-error-rate10. Rogo scales AI-driven financial research with OpenAI o1
https://openai.com/index/rogo/AI politics and safety:
1. Tell me about yourself: LLMs are aware of their learned behaviors
https://arxiv.org/abs/2501.111202. Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models
https://arxiv.org/abs/2411.142573. OpenAI hides chain-of-thought reasoning because it may include unaligned content. From “Model Spec—a document which defines how we want our models to behave.”
https://model-spec.openai.com/2025-02-12.html4. Meta Starts Eliminating Jobs in Shift to Find AI Talent
https://www.bloomberg.com/news/articles/2025-02-10/meta-starts-eliminating-jobs-in-shift-to-find-ai-talent [no paywall:
https://archive.is/T7Kog]
Science and Technology:
1. Learning produces an orthogonalized state machine in the hippocampus
https://www.nature.com/articles/s41586-024-08548-w2. Rarely categorical, always high-dimensional: how the neural code changes along the cortical hierarchy
https://www.biorxiv.org/content/10.1101/2024.11.15.623878v33. "Dozens of new obesity drugs are coming: these are ones to watch; next-generation obesity drugs will work differently from Ozempic & Wegovy—aiming to deliver greater weight loss with fewer side effects"
https://www.nature.com/articles/d41586-025-00404-9 [no paywall:
https://archive.is/X9CW3]
4. A single human zygote contains all the information you need to develop into an adult human and at the same time contains within it, the evolutionary history of our species. The Genomic Code: the genome instantiates a generative model of the organism
https://www.cell.com/trends/genetics/fulltext/S0168-9525(25)00008-3