Links for 2024-12-30
AI:
1. “By default, capital will matter more than ever after AGI”
https://www.lesswrong.com/posts/KFFaKu27FNugCHFmh/by-default-capital-will-matter-more-than-ever-after-agi2. China’s Deepseek V3, trained on 2048 GPUs for 2 months—a low-bandwidth-NVLink version of H100, available for China before 2023.10 export restrictions. It’s allegedly on par with GPT-4o and Claude 3.5 Sonnet. That would be a >10X cost reduction.
https://github.com/deepseek-ai/DeepSeek-V3/blob/main/DeepSeek_V3.pdf3. There's now a clear path to super-intelligent scientists with data-driven experiment simulation and super-scaling RL. OmniPred: Language Models as Universal Regressors
https://arxiv.org/abs/2402.145474. Byte Latent Transformer scales efficiently with byte-level encoding, outperforming tokenized models.
https://arxiv.org/abs/2412.09871v15. How Hallucinatory A.I. Helps Science Dream Up Big Breakthroughs
https://www.nytimes.com/2024/12/23/science/ai-hallucinations-science.html [no paywall:
https://archive.is/zFtgu]
6. Reversible molecular simulation for training classical and machine learning force fields
https://arxiv.org/abs/2412.043747. Let your LLM generate a few tokens and you will reduce the need for retrieval
https://www.arxiv.org/abs/2412.115368. RLEF: Grounding Code LLMs in Execution Feedback with Reinforcement Learning
https://arxiv.org/abs/2410.020899. “We instructed o1-preview to play to win against Stockfish. Without explicit prompting, o1 figured out it could edit the game state to win against a stronger opponent. GPT-4o and Claude 3.5 required more nudging to figure this out”
https://x.com/JeffLadish/status/187280545322444820810. Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey
https://arxiv.org/abs/2412.1861911. HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs
https://arxiv.org/abs/2412.1892512. Why backprop was resisted for 20 years: assumption of discretely spiking neurons, goal of synthesizing Boolean logic, fear of local optima, and bad luck. Werbos has the best claim for invention.
https://yuxi-liu-wired.github.io/essays/posts/backstory-of-backpropagation/13. OpenAI lays out its for-profit transition plans
https://openai.com/index/why-our-structure-must-evolve-to-advance-our-mission/14. 321 real-world gen AI use cases from the world's leading organizations
https://cloud.google.com/transform/101-real-world-generative-ai-use-cases-from-industry-leadersBiology:
1. Genetically edited mosquitoes haven't scaled yet. Why?
https://eryney.substack.com/p/genetically-edited-mosquitoes-havent2. BAAI aims to develop such closed-loop biophysically detailed models ('life models' that accurately simulate an organism's complex behavioral repertoire from its neural, biomechanical and environmental interactions).
https://www.nature.com/articles/s43588-024-00738-w3. Genetic Sequences of Highly Pathogenic Avian Influenza A(H5N1) Viruses Identified in a Person in Louisiana — The H5 virus mutated inside the single patient to gain an ability to bind human receptors in the upper respiratory tract.
https://www.cdc.gov/bird-flu/spotlights/h5n1-response-12232024.html4. Considerations on orca intelligence
https://www.lesswrong.com/posts/dzLwCBvwC4hWytnus/considerations-on-orca-intelligenceMiscellaneous:
1. Adam Brown (a lead of Blueshift at DeepMind & theoretical physicist at Stanford) – How Future Civilizations Could Change The Laws of Physics
https://www.dwarkeshpatel.com/p/adam-brown2. How to fix computing's AI energy problem: run everything backwards — About reversible computing, thermodynamics as the natural language for computation, and how we will get to 1000x reduction in costs in