New Paradigm: AI Research Summaries
Activity Overview
Episode publication activity over the past year
Episodes
How OpenAI is Advancing AI Competitive Programming with Reinforcement Learning
23 Feb 2025
Contributed by Lukas
This episode analyzes the study "Competitive Programming with Large Reasoning Models," conducted by researchers from OpenAI, DeepSeek-R1, and Kimi k1....
Examining Stanford's ZebraLogic Study: AI's Struggles with Complex Logical Reasoning
18 Feb 2025
Contributed by Lukas
This episode analyzes the study "ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning," conducted by Bill Yuchen Lin, Ronan Le Bras, Kyle R...
A Summary of Stanford's "s1: Simple test-time scaling" AI Research Paper
15 Feb 2025
Contributed by Lukas
This episode analyzes "s1: Simple test-time scaling," a research study conducted by Niklas Muennighoff, Zitong Yang, Weijia Shi, Xiang Lisa Li, Li Fei...
The Impact of AI Tools On Critical Thinking
13 Feb 2025
Contributed by Lukas
This episode analyzes "AI Tools in Society: Impacts on Cognitive Offloading and the Future of Critical Thinking," a study conducted by Michael Gerlich...
Examining Microsoft Research’s 'Multimodal Visualization-of-Thought'
11 Feb 2025
Contributed by Lukas
This episode analyzes the "Multimodal Visualization-of-Thought" (MVoT) study conducted by Chengzu Li, Wenshan Wu, Huanyu Zhang, Yan Xia, Shaoguang Mao...
A Summary of 'Increased Compute Efficiency and the Diffusion of AI Capabilities'
10 Feb 2025
Contributed by Lukas
This episode analyzes the research paper titled "Increased Compute Efficiency and the Diffusion of AI Capabilities," authored by Konstantin Pilz, Lenn...
Insights from Tencent AI Lab: Overcoming Underthinking in AI with Token Efficiency
07 Feb 2025
Contributed by Lukas
This episode analyzes the research paper "Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs," authored by Yue Wang and colleagues ...
Can Tencent AI Lab's O1 Models Streamline Reasoning and Boost Efficiency?
05 Feb 2025
Contributed by Lukas
This episode analyzes the study "On the Overthinking of o1-Like Models" conducted by researchers Xingyu Chen, Jiahao Xu, Tian Liang, Zhiwei He, Jianhu...
Harvard Research: What if AI Could Redefine Its Understanding with New Contexts?
03 Feb 2025
Contributed by Lukas
This episode analyzes the research paper titled "In-Context Learning of Representations," authored by Core Francisco Park, Andrew Lee, Ekdeep Singh Lu...
A summary of Agent Laboratory: Leveraging AI to Revolutionize Research
29 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled "Agent Laboratory: Using LLM Agents as Research Assistants," authored by Samuel Schmidgall, Yusheng Su...
Can Google's Mind Evolution Approach Unlock Deeper Thinking in Large Language Models?
28 Jan 2025
Contributed by Lukas
This episode analyzes the research paper "Evolving Deeper LLM Thinking" by Kuang-Huei Lee, Ian Fischer, Yueh-Hua Wu, Dave Marwood, Shumeet Baluja, Dal...
What might The University of Sydney's Transformers Unlock in Predicting Human Brain States?
26 Jan 2025
Contributed by Lukas
This episode analyzes the study "Predicting Human Brain States with Transformer" conducted by Yifei Sun, Mariano Cabezas, Jiah Lee, Chenyu Wang, Wei Z...
How might DeepSeek-R1 Revolutionize Reasoning in AI Language Models?
25 Jan 2025
Contributed by Lukas
This episode analyzes "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning," a study conducted by Daya Guo and colleagu...
Remember the Titans: Google Research’s Breakthrough in Enhancing AI Memory
22 Jan 2025
Contributed by Lukas
This episode analyzes the study "Titans: Learning to Memorize at Test Time" by Ali Behrouz, Peilin Zhong, and Vahab Mirrokni from Google Research. It ...
How Does Search-o1 Revolutionize Large Reasoning Models with Autonomous Search?
20 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled **"Search-o1: Agentic Search-Enhanced Large Reasoning Models,"** authored by Xiaoxi Li, Guanting Dong,...
How Is Transformer2 Transforming Real-Time Language Model Adaptation? (ENHANCED)
19 Jan 2025
Contributed by Lukas
This episode analyzes the research paper "TRANSFORMER2: SELF-ADAPTIVE LLM S" by Qi Sun, Edoardo Cetin, and Yujin Tang from Sakana AI and the Institute...
Simulating One Million Agents For Social Media With OASIS
16 Jan 2025
Contributed by Lukas
This episode analyzes "OASIS: OpenAgent Social Interaction Simulations with One Million Agents," a research initiative conducted by a diverse team fro...
Insights from NVIDIA on Generative AI Pricing and Market Competition Strategies
14 Jan 2025
Contributed by Lukas
This episode analyzes Rafid Mahmood's paper, "Pricing and Competition for Generative AI," authored by Mahmood from NVIDIA and the University of Ottawa...
Insights from NVIDIA: Creating Compact Language Models through Pruning and Knowledge Distillation
12 Jan 2025
Contributed by Lukas
This episode analyzes the research paper "**Compact Language Models via Pruning and Knowledge Distillation**" authored by Saurav Muralidharan, Sharath...
Success with synthetic data - a summary of the Microsoft's Phi-4 AI model technical report
09 Jan 2025
Contributed by Lukas
This episode analyzes the "Phi-4 Technical Report," published on December 12, 2024, by a team of researchers from Microsoft Research, including Marah ...
What makes Microsoft's rStar-Math a breakthrough small AI reasoning model
09 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled "rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking," authored by Xinyu ...
Google DeepMind's paradigm shift to scaling AI model test time compute
09 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled **"Scaling LLM Test-Time Compute Optimally can be More Effective Than Scaling Model Parameters,"** aut...
Exploring NVIDIA’s Cosmos: advancing physical AI through digital twins and robotics
09 Jan 2025
Contributed by Lukas
This episode analyzes NVIDIA's "Cosmos World Foundation Model Platform for Physical AI," released on January 7, 2025. Based on research by NVIDIA, the...
How might Meta AI's Mender transform personalized recommendations with LLM-enhanced retrieval?
09 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled "Preference Discerning with LLM-Enhanced Generative Retrieval," authored by Fabian Paischer, Liu Yang,...
How does Meta FAIR's Ewe give AI models working memory?
08 Jan 2025
Contributed by Lukas
This episode analyzes the research paper "Improving Factuality with Explicit Working Memory" by Mingda Chen, Yang Li, Karthik Padthe, Rulin Shao, Alic...
Should You Use CAG (Cache-Augmented Generation) Instead of RAG for LLM Knowledge Retrieval
07 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled "Don’t Do RAG: When Cache-Augmented Generation is All You Need for Knowledge Tasks," authored by Bri...
Could GitHub Inc.’s Copilot Boost Developer Productivity and Transform Work Dynamics
06 Jan 2025
Contributed by Lukas
This episode analyzes the study "Generative AI and the Nature of Work," conducted by Manuel Hoffmann, Sam Boysel, Frank Nagle, Sida Peng, and Kevin Xu...
What role does Stanford's Putnam-AXIOM play in evaluating AI's mathematical reasoning?
04 Jan 2025
Contributed by Lukas
This episode reviews "Putnam-AXIOM: A Functional and Static Benchmark for Measuring Higher Level Mathematical Reasoning," a study conducted in 2024 by...
What Does Google DeepMind's Research Reveal About Machine Unlearning’s Limitations Protecting Privacy and Copyright in Generative AI?
03 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled *"Machine Unlearning Doesn’t Do What You Think: Lessons for Generative AI Policy, Research, and Prac...
Understanding of The Inner Workings of AI Models With MONET's Advanced Mechanistic Interpretability?
01 Jan 2025
Contributed by Lukas
This episode analyzes the research paper titled "MONET: Mixture of Monosemantic Experts for Transformers," authored by Jungwoo Park, Young Jin Ahn, Ke...
Key insights from Google DeepMind's PaliGemma 2: Transforming Vision-Language AI
30 Dec 2024
Contributed by Lukas
This episode analyzes "PaliGemma 2: A Family of Versatile Vision-Language Models for Transfer," a December 2024 study by Andreas Steiner, André Susan...
Could DeepSeek-V3 Revolutionize Language Modeling?
30 Dec 2024
Contributed by Lukas
This episode analyzes the "DeepSeek-V3 Technical Report," authored by Aixin Liu and colleagues from DeepSeek-AI and published on December 27, 2024. It...
Understanding The Roadmap to Reproduce o1 Reasoning AI Models
30 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "OpenMOSS Scaling of Search and Learning: A Roadmap to Reproduce o1 from a Reinforcement Learning Pers...
Investigating Deceptive AI Behaviors: UC Berkeley’s Analysis of User Feedback Optimization in LLMs
27 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Untargeted Manipulation and Deception When Optimizing LLMs for User Feedback" authored by Marcus Williams, M...
Exploring the DeMo Optimizer by Nous Research: Enhancing Large Neural Network Training
27 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "DeMo: Decoupled Momentum Optimization" by Bowen Peng, Jeffrey Quesnelle, and Diederik P. Kingma from Nous Re...
Can the Tsinghua University AI Lab Prevent Model Collapse in Synthetic Data?
24 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "HOW TO SYNTHESIZE TEXT DATA WITHOUT MODEL COLLAPSE?" authored by Xuekai Zhu, Daixuan Cheng, Hengli Li...
Can Salesforce AI Research's LaTRO Unlock Hidden Reasoning in Language Models?
24 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Language Models Are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding," au...
A Summary of Netflix's Research on Cosine Similarity Unreliability in Semantic Embeddings
23 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Is Cosine-Similarity of Embeddings Really About Similarity?" by Harald Steck, Chaitanya Ekanadham, an...
Key insights from Salesforce Research: Enhancing LLMs with Offline Reinforcement Learning
23 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Offline Reinforcement Learning for LLM Multi-Step Reasoning" authored by Huaijie Wang, Shibo Hao, Hanze Dong...
Breaking down Johns Hopkins University's GenEx: AI Transforms Images into Immersive 3D Worlds
23 Dec 2024
Contributed by Lukas
This episode analyzes **'GenEx: Generating an Explorable World'**, a research project conducted by Taiming Lu, Tianmin Shu, Junfei Xiao, Luoxin Ye, Ji...
What Makes Anthropic's Sparse Autoencoders and Metrics Revolutionize AI Interpretability
21 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Evaluating Sparse Autoencoders on Targeted Concept Erasure Tasks" by Adam Karvonen, Can Rager, Samuel Marks,...
How Can Google DeepMind’s Models Reveal Hidden Biases in Feature Representations
21 Dec 2024
Contributed by Lukas
This episode analyzes the research conducted by Andrew Kyle Lampinen, Stephanie C. Y. Chan, and Katherine Hermann at Google DeepMind, as presented in ...
Breaking down OpenAI’s Deliberative Alignment: A New Approach to Safer Language Models
20 Dec 2024
Contributed by Lukas
This episode analyzes OpenAI's research paper titled "Deliberative Alignment: Reasoning Enables Safer Language Models," authored by Melody Y. Guan and...
How does Bytedance Inc's Liquid Revolutionize Scalable Multi-modal AI Systems
20 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Liquid: Language Models are Scalable Multi-modal Generators" by Junfeng Wu, Yi Jiang, Chuofan Ma, Yuliang Li...
What does OpenAI's Sparse Autoencoder Reveal About GPT-4’s Inner Workings
20 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"Scaling and Evaluating Sparse Autoencoders"** authored by Leo Gao, Tom Dupré la Tour, Henk Tillman...
Oxford University Research: How Do Sparse Auto-Encoders Reveal Universal Feature Similarities in Large Language Models
19 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"Sparse Autoencoders Reveal Universal Feature Spaces Across Large Language Models"** by Michael Lan, Philip...
Understanding How Google Research Uses Process Reward Models to Improve LLM Reasoning
19 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning"** by Amrith Setlur, Chirag Nagp...
Examining the Alibaba Group's Multi-Agent Planning Framework for Enhanced Collaboration and Performance
19 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Agent-Oriented Planning in Multi-Agent Systems" by Ao Li, Yuexiang Xie, Songze Li, Fugee Tsung, Bolin Ding, ...
According to Google DeepMind Can Language Models Perform Multi-Hop Reasoning Without Shortcuts?
18 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?" by Sohee Y...
Breaking down Google DeepMind's AI Planning Strategies to Achieve Grandmaster-Level Chess
18 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"Mastering Board Games by External and Internal Planning with Language Models"**, authored by John S...
Rethinking Transformer Efficiency: The University of Maryland Unveils Attention Layer Pruning
18 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "WHAT MATTERS IN TRANSFORMERS? NOT ALL ATTENTION IS NEEDED," authored by Shwai He, Guoheng Sun, Zhenyu Shen, ...
What Might Google DeepMind's Language Models Reveal About AI Cooperation Evolution
17 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Cultural Evolution of Cooperation among LLM Agents" by Aron Vallinder and Edward Hughes, affiliated with Ind...
Exploring the UC Berkeley TEMPERA Approach to Dynamic AI Prompt Optimization
17 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"TEMPERA: Test-Time Prompt Editing via Reinforcement Learning,"** authored by Tianjun Zhang, Xuezhi ...
What Does Harvard Kennedy School Research Reveal About Generative AI’s Rapid Adoption?
17 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "The Rapid Adoption of Generative AI," authored by Alexander Bick, Adam Blandin, and David J. Deming f...
Breaking down Harvard's Insights into Hidden Capabilities and Concept Spaces in Generative Models
16 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space,"** authored by Core Franci...
Can the Socratic Learning Approach from Google DeepMind Unlock AI Autonomy?
16 Dec 2024
Contributed by Lukas
This episode analyzes Tom Schaul's research paper, "Boundless Socratic Learning with Language Games," authored on November 25, 2024, under the affilia...
Investigating Google DeepMind's Gemini 2.0: Next-Gen Multimodal AI and Applications
16 Dec 2024
Contributed by Lukas
This episode analyzes the research paper “Introducing Gemini 2.0: our new AI model for the agentic era” authored by Demis Hassabis and Koray Kavuk...
How can Google DeepMind's Genie 2 revolutionize AI training and virtual interactions?
16 Dec 2024
Contributed by Lukas
This episode reviews "Genie 2: A Large-Scale Foundation World Model," a research publication dated December 4, 2024, authored by a team from Google De...
How Can Google DeepMind's OmegaPRM Revolutionize AI Mathematical Reasoning?
15 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"Improve Mathematical Reasoning in Language Models by Automated Process Supervision"** authored by L...
A summary of Microsoft Research's Phi-4: Transforming Language Models with Advanced Training Techniques
15 Dec 2024
Contributed by Lukas
This episode analyzes the "Phi-4 Technical Report" authored by Marah Abdin, Jyoti Aneja, Harkirat Behl, Sébastien Bubeck, Ronen Eldan, and colleagues...
What if FAIR at Meta Replaces Tokens with Concepts in Language Modeling
14 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"Language Modeling in a Sentence Representation Space"** authored by Loïc Barrault, Paul-Ambroise Duquenne...
Insights from Stanford: Precision Scaling Laws Enhance Language Model Efficiency and Accuracy
14 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"Scaling Laws for Precision,"** authored by Tanishq Kumar, Zachary Ankner, Benjamin F. Spector, Blake Borde...
Exploring FAIR at Meta’s Byte Latent Transformer: Enhancing AI Efficiency with Byte Patches
14 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"Byte Latent Transformer: Patches Scale Better Than Tokens,"** authored by Artidoro Pagnoni, Ram Pas...
How Can NVIDIA's LLaMA-Mesh Transform Content Creation with AI-Generated 3D Models
14 Dec 2024
Contributed by Lukas
This episode analyzes the research paper **"LLaMA-Mesh: Unifying 3D Mesh Generation with Language Models,"** authored by Zhengyi Wang, Jonathan Lorrai...
How does Apollo Research Reveal AI Models' Potential for Deceptive Scheming Behaviors?
13 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Frontier Models are Capable of In-context Scheming" authored by Alexander Meinke, Bronson Schoen, Jérémy S...
Can AI Models Solve Proportional Analogies Through Knowledge-Enhanced Prompting? (Research by Stanford)
13 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"Exploring the Abilities of Large Language Models to Solve Proportional Analogies via Knowledge-Enha...
Can LLMs Hide Hallucinations in Their Internal Truth Representations? (Research by Google)
13 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled **"LLM Know More Than They Show: On the Intrinsic Representation of LLM Hallucinations,"** authored by...
Can Google DeepMind's AlphaQubit Achieve High-Accuracy Quantum Error Correction?
12 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Learning High-Accuracy Error Decoding for Quantum Processors," authored by Johannes Bausch, Andrew W....
Proven Scaling Laws to Boost LLM Reliability and Accuracy
12 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "A Simple and Provable Scaling Law for the Test-Time Compute of Large Language Models," authored by Ya...
Evaluating the SIFT Algorithm: Enhancing Large Language Model Fine-Tuning at Test-Time
12 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs," authored by Jonas Hübotter, Sascha Bongni, ...
Can Advanced Machine Unlearning Techniques Enable Greater Privacy and Model Accuracy?
12 Dec 2024
Contributed by Lukas
This episode analyzes the study titled "Improved Localized Machine Unlearning Through the Lens of Memorization," authored by Reihaneh Torkzadehmahani,...
Breaking down HiAR-ICL: Revolutionizing AI Reasoning with Monte Carlo Tree Search
11 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Beyond Examples: High-level Automated Reasoning Paradigm in In-Context Learning via MCTS," authored b...
Could Agent Workflow Memory Transform AI's Ability to Navigate and Solve Complex Web Tasks?
11 Dec 2024
Contributed by Lukas
This episode analyzes "Agent Workflow Memory," a study conducted by Zora Zhiruo Wang, Jiayuan Mao, Daniel Fried, and Graham Neubig from Carnegie Mello...
Key insights from Apple Ferret-UI 2: Mastering Cross-Platform User Interface Understanding
11 Dec 2024
Contributed by Lukas
This episode analyzes the study titled "FERRET-UI 2: Mastering Universal User Interface Understanding Across Platforms," authored by Zhangheng Li, Kee...
How Does AGORA BENCH Compare Language Models in Synthetic Data Generation?
10 Dec 2024
Contributed by Lukas
This episode analyzes the study "Evaluating Language Models as Synthetic Data Generators" by Seungone Kim, Juyoung Suk, Xiang Yue, Vijay Viswanathan, ...
What if AI Wins Short Rounds but Humans Excel in Long-Term Research
10 Dec 2024
Contributed by Lukas
This episode analyzes the study titled "RE-Bench: Evaluating Frontier AI R&D Capabilities of Language Model Agents Against Human Experts," authored by...
Understanding the Coconut Method: Enhancing AI Reasoning with a Continuous Latent Space Approach
10 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Training Large Language Models to Reason in a Continuous Latent Space" by Shibo Hao, Sainbayar Sukhbaatar, D...
Can the Shift to Process Reward Models Revolutionize Large Language Model Reasoning?
10 Dec 2024
Contributed by Lukas
This episode analyzes the research paper "Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning" by Amrith Setlur, Chirag Nagpal, ...
Could CoALA’s Cognitive Architecture Transform Intelligent Language Agents?
10 Dec 2024
Contributed by Lukas
This episode analyzes "Cognitive Architectures for Language Agents," a paper authored by Theodore R. Sumers, Shunyu Yao, Karthik Narasimhan, and Thoma...
A summary of REVTHINK: Reverse Thinking Enhances LLM Reasoning
09 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Reverse Thinking Makes LLMs Stronger Reasoners," authored by Justin Chih-Yao Chen, Zifeng Wang, Hamid...
Can a Neural Model Achieve Human-Level Abstract Reasoning?
09 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning" by Ekin Akyürek, Mehul Da...
Exploring ARC Prize 2024: Breakthroughs in AGI
09 Dec 2024
Contributed by Lukas
This episode analyzes the ARC Prize 2024 technical report authored by François Chollet, Mike Knoop, Gregory Kamradt, and Bryan Landers from Lab42, da...
Can a Domain-Specific Language Boost AI's Reasoning?
09 Dec 2024
Contributed by Lukas
This episode analyzes Martin Andrews' paper, "Capturing Sparks of Abstraction for the ARC Challenge," published on November 17, 2024, by Red Dragon AI...
Can a Tiny Subset of Super-Weights Control Large Language Models?
09 Dec 2024
Contributed by Lukas
This episode analyzes the concept of super weights in Large Language Models, drawing on research by Mengxia Yu, De Wang, Qi Shan, Colorado Reed, and A...
Key Insights from Grokked Transformers: Implicit Reasoning
08 Dec 2024
Contributed by Lukas
This episode analyzes the research paper titled "Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization," au...
What does NVILA Bring to Visual Language Models?
08 Dec 2024
Contributed by Lukas
This episode analyzes the research presented in "NVILA: Efficient Frontier Visual Language Models," authored by Zhijian Liu and colleagues from instit...
Can the Densing Law Revolutionize AI Efficiency?
08 Dec 2024
Contributed by Lukas
This episode analyzes the research titled "Densing Law of LLMs" by Chaojun Xiao, Jie Cai, Weilin Zhao, Guoyang Zeng, Biyuan Lin, Jie Zhou, Xu Han, Zhi...
A Summary of 'Scaling Synthetic Data Creation with One Billion Personas' by Tencent AI Lab
15 Jul 2024
Contributed by Lukas
A Summary of Tencent AI Lab's 'Scaling Synthetic Data Creation with One Billion Personas' Available at: https://arxiv.org/abs/2406.20094 This summar...
A Summary of 'Improving Alignment and Robustness with Circuit Breakers' by Black Swan AI, Carnegie Mellon University, & the Center for AI Sa
10 Jul 2024
Contributed by Lukas
A Summary of Black Swan AI, Carnegie Mellon University, & the Center for AI Safety's 'Improving Alignment and Robustness with Circuit Breakers' A...
A Summary of 'Refusal in Language Models Is Mediated by a Single Direction' by Anthropic, MIT, ETH Zürich & The University of Maryland
05 Jul 2024
Contributed by Lukas
A Summary of Anthropic, MIT, ETH Zürich & The University of Maryland's 'Refusal in Language Models Is Mediated by a Single Direction' Available ...
A Summary of 'LLMs achieve adult human performance on higher-order theory of mind tasks' by Google DeepMind, Johns Hopkins University & The
06 Jun 2024
Contributed by Lukas
A Summary of Google DeepMind, Johns Hopkins University & The University of Oxford's 'LLMs achieve adult human performance on higher-order theory o...
A Summary of 'LoRA Learns Less and Forgets Less' by Databricks Mosaic AI & Columbia University
04 Jun 2024
Contributed by Lukas
A Summary of Databricks Mosaic AI & Columbia University's 'LoRA Learns Less and Forgets Less' Available at: https://arxiv.org/abs/2405.09673 Thi...
A Summary of 'Mastering Diverse Domains through World Models' by Google DeepMind & The University of Toronto
01 Jun 2024
Contributed by Lukas
A Summary of Google DeepMind & The University of Toronto's 'Mastering Diverse Domains through World Models' Available at: https://arxiv.org/abs/2...
A Summary of 'Let’s Think Dot by Dot: Hidden Computation in Transformer Language Models' by CDS at New York University
12 May 2024
Contributed by Lukas
A Summary of CDS at New York University's 'Let’s Think Dot by Dot: Hidden Computation in Transformer Language Models' Available at: https://arxiv....
A Summary of Predibase's 'LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report'
11 May 2024
Contributed by Lukas
A Summary of Predibase's 'LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report' Available at: https://arxiv.org/abs/2405.00732 This...
A Summary of 'Creative Problem Solving in Large Language and Vision Models – What Would it Take?' by Georgia Institute of Technology & Tufts
06 May 2024
Contributed by Lukas
A Summary of Georgia Institute of Technology & Tufts University, Medford's 'Creative Problem Solving in Large Language and Vision Models – What ...
A Summary of 'KAN: Kolmogorov–Arnold Networks' by MIT, CALTECH & Others
04 May 2024
Contributed by Lukas
A Summary of MIT, CALTECH & Other's 'KAN: Kolmogorov–Arnold Networks' Available at: https://arxiv.org/abs/2404.19756 This summary is AI gene...
A Summary of Stanford University, MIT & Sequoia Capital's 'Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Rea
03 May 2024
Contributed by Lukas
A Summary of Stanford University, MIT & Sequoia Capital's 'Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and ...
A Summary of FAIR at Meta's 'Better & Faster Large Language Models via Multi-token Prediction'
01 May 2024
Contributed by Lukas
A Summary of FAIR at Meta's 'Better & Faster Large Language Models via Multi-token Prediction' Available at: https://arxiv.org/abs/2404.19737 ...
A Summary of Apple's 'Ferret-UI: Grounded Mobile UI Understanding with Multimodal LLMs'
29 Apr 2024
Contributed by Lukas
A Summary of Apple's 'Ferret-UI: Grounded Mobile UI Understanding with Multimodal LLMs' Available at: https://arxiv.org/pdf/2404.05719 This summar...