<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://gisaxs.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=KevinYager</id>
	<title>GISAXS - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="http://gisaxs.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=KevinYager"/>
	<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php/Special:Contributions/KevinYager"/>
	<updated>2026-04-08T13:01:10Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.7</generator>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8780</id>
		<title>AI understanding</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8780"/>
		<updated>2026-04-06T21:18:48Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Psychology */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Interpretability=&lt;br /&gt;
* 2017-01: [https://arxiv.org/abs/1704.01444 Learning to Generate Reviews and Discovering Sentiment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11639 Neural Interpretable Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Concepts==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.20938 Towards Understanding the Nature of Attention with Low-Rank Sparse Decomposition] ([https://github.com/OpenMOSS/Lorsa code])&lt;br /&gt;
* 2025-08: [https://transformer-circuits.pub/2025/attention-qk/index.html Tracing Attention Computation Through Feature Interactions]&lt;br /&gt;
&lt;br /&gt;
==Mechanistic Interpretability==&lt;br /&gt;
* 2020-03: OpenAI: [https://distill.pub/2020/circuits/zoom-in/ Zoom In: An Introduction to Circuits]&lt;br /&gt;
* 2021-12: Anthropic: [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2211.00593 Interpretability in the Wild: a Circuit for Indirect Object Identification in GPT-2 Small]&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-07: Anthropic: [https://transformer-circuits.pub/2024/july-update/index.html Circuits Update]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.14926 Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-based Parameter Decomposition] ([https://www.alignmentforum.org/posts/EPefYWjuHNcNH4C7E/attribution-based-parameter-decomposition blog post])&lt;br /&gt;
* 2025-01: Review: [https://arxiv.org/abs/2501.16496 Open Problems in Mechanistic Interpretability]&lt;br /&gt;
* 2025-03: Anthropic: [https://www.anthropic.com/research/tracing-thoughts-language-model Tracing the thoughts of a large language model]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/methods.html Circuit Tracing: Revealing Computational Graphs in Language Models]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/biology.html On the Biology of a Large Language Model]&lt;br /&gt;
* 2025-11: OpenAI: [https://cdn.openai.com/pdf/41df8f28-d4ef-43e9-aed2-823f9393e470/circuit-sparsity-paper.pdf Weight-sparse transformers have interpretable circuits] ([https://openai.com/index/understanding-neural-networks-through-sparse-circuits/ blog])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.13548 Patterning: The Dual of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Semanticity==&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.08600 Sparse Autoencoders Find Highly Interpretable Features in Language Models]&lt;br /&gt;
* Anthropic monosemanticity interpretation of LLM features:&lt;br /&gt;
** 2023-10: [https://transformer-circuits.pub/2023/monosemantic-features/index.html Towards Monosemanticity: Decomposing Language Models With Dictionary Learning]&lt;br /&gt;
** 2024-05: [https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet]&lt;br /&gt;
* 2024-06: OpenaAI: [https://arxiv.org/abs/2406.04093 Scaling and evaluating sparse autoencoders]&lt;br /&gt;
* 2024-08: [https://www.alignmentforum.org/posts/TMAmHh4DdMr4nCSr5/showing-sae-latents-are-not-atomic-using-meta-saes Showing SAE Latents Are Not Atomic Using Meta-SAEs] ([https://metasae.streamlit.app/?page=Feature+Explorer&amp;amp;feature=11329 demo])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08201 Efficient Dictionary Learning with Switch Sparse Autoencoders] ([https://github.com/amudide/switch_sae code]) More efficient SAE generation&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.14670 Decomposing The Dark Matter of Sparse Autoencoders] ([https://github.com/JoshEngels/SAE-Dark-Matter code]) Shows that SAE errors are predictable&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13928 Automatically Interpreting Millions of Features in Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.21331 Beyond Interpretability: The Gains of Feature Monosemanticity on Model Robustness]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.04139 Monet: Mixture of Monosemantic Experts for Transformers]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/zbebxYCqsryPALh8C/matryoshka-sparse-autoencoders Matryoshka Sparse Autoencoders]&lt;br /&gt;
* 2024-12: [https://www.alignmentforum.org/posts/rKM9b6B2LqwSB5ToN/learning-multi-level-features-with-matryoshka-saes Learning Multi-Level Features with Matryoshka SAEs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19406 Low-Rank Adapting Models for Sparse Autoencoders]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.03714 Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.00177 Steering Large Language Model Activations in Sparse Spaces]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01776 Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01824 From superposition to sparse codes: interpretable representations in neural networks]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18878 I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20063 SAEs Are Good for Steering -- If You Select the Right Features]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15679 Dense SAE Latents Are Features, Not Bugs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20790 Stochastic Parameter Decomposition] ([https://github.com/goodfire-ai/spd code], [https://www.goodfire.ai/papers/stochastic-param-decomp blog])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
&lt;br /&gt;
===Counter-Results===&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.12016 Towards falsifiable interpretability research]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16615 Sparse Autoencoders Trained on the Same Data Learn Different Features]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17148 AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17727 Sparse Autoencoders Can Interpret Randomly Initialized Transformers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.04878 Sparse Autoencoders Do Not Find Canonical Units of Analysis]&lt;br /&gt;
* 2025-03: [https://www.alignmentforum.org/posts/4uXCAJNuPKtKBsi28/ Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research]&lt;br /&gt;
&lt;br /&gt;
==Meta-cognition==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.15674 Activation Oracles: Training and Evaluating LLMs as General-Purpose Activation Explainers]&lt;br /&gt;
&lt;br /&gt;
==Coding Models==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sparse Auto Encoders&amp;#039;&amp;#039;&amp;#039;: See Semanticity.&lt;br /&gt;
* [https://github.com/saprmarks/dictionary_learning dictionary_learning]&lt;br /&gt;
* [https://transformer-circuits.pub/2024/jan-update/index.html#predict-future Predicting Future Activations]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11944 Transcoders Find Interpretable LLM Feature Circuits]&lt;br /&gt;
* 2024-10: [https://transformer-circuits.pub/2024/crosscoders/index.html Sparse Crosscoders for Cross-Layer Features and Model Diffing]&lt;br /&gt;
&lt;br /&gt;
==Reward Functions==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12491 Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL]&lt;br /&gt;
&lt;br /&gt;
==Symbolic and Notation==&lt;br /&gt;
* [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* [https://www.arxiv.org/abs/2407.09468 Beyond Euclid: An Illustrated Guide to Modern Machine Learning with Geometric, Topological, and Algebraic Structures]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02423 On the Anatomy of Attention]: Introduces category-theoretic diagrammatic formalism for DL architectures&lt;br /&gt;
* 2024-11: [https://x.com/vtabbott_/status/1860268276569506250 diagrams to represent algorithms]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03317 FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awareness]&lt;br /&gt;
&lt;br /&gt;
==Mathematical==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.13762 Unveiling the Hidden Structure of Self-Attention via Kernel Principal Component Analysis]&lt;br /&gt;
&lt;br /&gt;
==Geometric==&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.03658 The Linear Representation Hypothesis and the Geometry of Large Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.01506 The Geometry of Categorical and Hierarchical Concepts in Large Language Models]&lt;br /&gt;
** Natural hierarchies of concepts---which occur throughout natural language and especially in scientific ontologies---are represented in the model&amp;#039;s internal vectorial space as polytopes that can be decomposed into simplexes of mutually-exclusive categories.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02678 Reasoning in Large Language Models: A Geometric Perspective]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.17592 Deep Manifold Part 1: Anatomy of Neural Network Manifold]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.19750 The Geometry of Concepts: Sparse Autoencoder Feature Structure]&lt;br /&gt;
** Tegmark et al. report multi-scale structure: 1) “atomic” small-scale, 2) “brain” intermediate-scale, and 3) “galaxy” large-scale&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.08009 The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.09782 The Geometry of Reasoning: Flowing Logics in Representation Space]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/linebreaks/index.html When Models Manipulate Manifolds: The Geometry of a Counting Task]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.26745 Deep sequence models tend to memorize geometrically; it is unclear why]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
==Topography==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.06002 The Molecular Structure of Thought: Mapping the Topology of Long Chain-of-Thought Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2023-07Jul: [https://arxiv.org/abs/2307.13702 Measuring Faithfulness in Chain-of-Thought Reasoning] [https://x.com/davidad/status/1839641113432305790 roughly] proves that sufficiently large models do not generate CoT that actually captures their internal reasoning)&lt;br /&gt;
&lt;br /&gt;
[[Image:GYe31yXXQAABwaZ.jpeg|300px]]&lt;br /&gt;
&lt;br /&gt;
=Heuristic Understanding=&lt;br /&gt;
* 2022-09: Janus: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators]&lt;br /&gt;
&lt;br /&gt;
==Emergent Internal Model Building==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.15936 A Theory for Emergence of Complex Skills in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19370v1 Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01622 General agents contain world models]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20328 Video models are zero-shot learners and reasoners]&lt;br /&gt;
&lt;br /&gt;
===Semantic Directions===&lt;br /&gt;
Directions, e.g.: f(king)-f(man)+f(woman)=f(queen) or f(sushi)-f(Japan)+f(Italy)=f(pizza)&lt;br /&gt;
* [https://arxiv.org/abs/1301.3781 Efficient Estimation of Word Representations in Vector Space]&lt;br /&gt;
* [https://aclanthology.org/N13-1090/ Linguistic Regularities in Continuous Space Word Representations]&lt;br /&gt;
* [https://aclanthology.org/C16-1332 Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen]&lt;br /&gt;
* [https://aclanthology.org/D14-1162/ Glove: Global vectors for word representation]&lt;br /&gt;
* [https://doi.org/10.1109/BigData.2015.7364114 Using Word2Vec to process big text data]&lt;br /&gt;
* [https://arxiv.org/abs/2310.06824 The geometry of truth: Emergent linear structure in large language model representations of true/false datasets] (true/false)&lt;br /&gt;
* [https://arxiv.org/abs/2403.10381 Monotonic Representation of Numeric Properties in Language Models] (numeric directions)&lt;br /&gt;
Task vectors:&lt;br /&gt;
* [https://arxiv.org/abs/2310.15213 Function Vectors in Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.15916 In-context learning creates task vectors]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/5FGXmJ3wqgGRcbyH7/extracting-sae-task-features-for-in-context-learning Extracting sae task features for in-context learning]&lt;br /&gt;
* [https://arxiv.org/abs/2412.12276 Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers]&lt;br /&gt;
Reasoning:&lt;br /&gt;
* [https://openreview.net/forum?id=OwhVWNOBcz Understanding Reasoning in Thinking Language Models via Steering Vectors]&lt;br /&gt;
&lt;br /&gt;
===Feature Geometry Reproduces Problem-space===&lt;br /&gt;
* [https://arxiv.org/abs/2210.13382 Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2309.00941 Emergent linear representations in world models of self-supervised sequence models] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* [https://doi.org/10.1038/s41562-023-01659-w Emergent analogical reasoning in large language models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.02207 Language Models Represent Space and Time] (Maps of world, US)&lt;br /&gt;
* [https://arxiv.org/abs/2405.14860 Not All Language Model Features Are Linear] (Days of week form ring, etc.)&lt;br /&gt;
* [https://arxiv.org/abs/2406.03689 Evaluating the World Model Implicit in a Generative Model] (Map of Manhattan)&lt;br /&gt;
* [https://iopscience.iop.org/article/10.1088/1748-9326/ad2891 Reliable precipitation nowcasting using probabilistic diffusion models]. Generation of precipitation map imagery is predictive of actual future weather; implies model is learning scientifically-relevant modeling.&lt;br /&gt;
* [https://arxiv.org/abs/2405.07987 The Platonic Representation Hypothesis]: Different models (including across modalities) are converging to a consistent world model.&lt;br /&gt;
* [https://arxiv.org/abs/2501.00070 ICLR: In-Context Learning of Representations]&lt;br /&gt;
* [https://arxiv.org/abs/2502.00873 Language Models Use Trigonometry to Do Addition]: Numbers arranged in helix to enable addition&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
&lt;br /&gt;
===Capturing Physics===&lt;br /&gt;
* 2020-09: [https://arxiv.org/abs/2009.08292 Learning to Identify Physical Parameters from Video Using Differentiable Physics]&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.00419 Self-Supervised Learning for Videos: A Survey]&lt;br /&gt;
* 2025-02: Fair at Meta: [https://arxiv.org/abs/2502.11831 Intuitive physics understanding emerges from self-supervised pretraining on natural videos]&lt;br /&gt;
&lt;br /&gt;
===Theory of Mind===&lt;br /&gt;
* [https://arxiv.org/abs/2302.02083 Evaluating Large Language Models in Theory of Mind Tasks]&lt;br /&gt;
* [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-01: [https://www.arxiv.org/abs/2501.09038 Do generative video models learn physical principles from watching videos?] ([https://physics-iq.github.io/ project], [https://github.com/google-deepmind/physics-IQ-benchmark code])&lt;br /&gt;
* 2025-06: [https://machinelearning.apple.com/research/illusion-of-thinking The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21521 Potemkin Understanding in Large Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21876 Do Vision-Language Models Have Internal World Models? Towards an Atomic Evaluation]&lt;br /&gt;
&lt;br /&gt;
==Information Processing==&lt;br /&gt;
* 2019-03: [https://arxiv.org/abs/1903.05789 Diagnosing and Enhancing VAE Models]&lt;br /&gt;
* 2021-03: [https://arxiv.org/abs/2103.05247 Pretrained Transformers as Universal Computation Engines]&lt;br /&gt;
* 2022-10: [https://arxiv.org/abs/2210.08344 How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03843 Why think step by step? Reasoning emerges from the locality of experience]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.04444 What&amp;#039;s the Magic Word? A Control Theory of LLM Prompting]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12875 Chain of Thought Empowers Transformers to Solve Inherently Serial Problems]: Proves that transformers can solve any problem, if they can generate sufficient intermediate tokens&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.20311 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process]&lt;br /&gt;
** Models learning reasoning skills (they are not merely memorizing solution templates). They can mentally generate simple short plans (like humans).&lt;br /&gt;
** When presented facts, models develop internal understanding of what parameters (recursively) depend on each other. This occurs even before an explicit question is asked (i.e. before the task is defined). This appears to be different from human reasoning.&lt;br /&gt;
** Model depth matters for reasoning. This cannot be mitigated by chain-of-thought prompting (which allow models to develop and then execute plans) since even a single CoT step may require deep, multi-step reasoning/planning.&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.01992 Ask, and it shall be given: Turing completeness of prompting]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08775 Layers at Similar Depths Generate Similar Activations Across LLM Architectures]&lt;br /&gt;
&lt;br /&gt;
===Generalization===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]&lt;br /&gt;
&lt;br /&gt;
===Grokking===&lt;br /&gt;
* 2022-01: [https://arxiv.org/abs/2201.02177 Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets]&lt;br /&gt;
* 2022-05: [https://arxiv.org/abs/2205.10343 Towards Understanding Grokking: An Effective Theory of Representation Learning]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.10463 Critical Data Size of Language Models from a Grokking Perspective]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15175 Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18624 How to explain grokking]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.21519 Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
===Tests of Resilience to Dropouts/etc.===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15390 Explorations of Self-Repair in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15786 What Matters in Transformers? Not All Attention is Needed]&lt;br /&gt;
** Removing entire transformer blocks leads to significant performance degradation&lt;br /&gt;
** Removing MLP layers results in significant performance degradation&lt;br /&gt;
** Removing attention layers causes almost no performance degradation&lt;br /&gt;
** E.g. half of attention layers are deleted (48% speed-up), leads to only 2.4% decrease in the benchmarks&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19384 The Remarkable Robustness of LLMs: Stages of Inference?]&lt;br /&gt;
** They intentionally break the network (swapping layers), yet it continues to work remarkably well. This suggests LLMs are quite robust, and allows them to identify different stages in processing.&lt;br /&gt;
** They also use these interventions to infer what different layers are doing. They break apart the LLM transformer layers into four stages:&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Detokenization:&amp;#039;&amp;#039;&amp;#039; Raw tokens are converted into meaningful entities that take into account local context (especially using nearby tokens).&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Feature engineering:&amp;#039;&amp;#039;&amp;#039; Features are progressively refined. Factual knowledge is leveraged.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Prediction ensembling:&amp;#039;&amp;#039;&amp;#039; Predictions (for the ultimately-selected next-token) emerge. A sort of consensus voting is used, with “prediction neurons” and &amp;quot;suppression neurons&amp;quot; playing a major role in upvoting/downvoting.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Residual sharpening:&amp;#039;&amp;#039;&amp;#039; The semantic representations are collapsed into specific next-token predictions. There is a strong emphasis on suppression neurons eliminating options. The confidence is calibrated.&lt;br /&gt;
** This structure can be thought of as two halves (being roughly dual to each other): the first half broadens (goes from distinct tokens to a rich/elaborate concept-space) and the second half collapses (goes from rich concepts to concrete token predictions).&lt;br /&gt;
&lt;br /&gt;
==Semantic Vectors==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11717 Refusal in Language Models Is Mediated by a Single Direction]&lt;br /&gt;
* 2025-02: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs] ([https://x.com/OwainEvans_UK/status/1894436637054214509 demonstrates] [https://x.com/ESYudkowsky/status/1894453376215388644 entangling] of concepts into a single preference vector)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03666 Analogical Reasoning Inside Large Language Models: Concept Vectors and the Limits of Abstraction]&lt;br /&gt;
&lt;br /&gt;
==Other==&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00247 Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting &amp;amp; Beyond]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04282 Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding] ([https://github.com/SalesforceAIResearch/LaTRO code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.12580 Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models]: LLMs learn reasoning by extracting procedures from training data, not by memorizing specific answers&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15862 LLMs Do Not Think Step-by-step In Implicit Reasoning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
&lt;br /&gt;
===Scaling Laws===&lt;br /&gt;
* 1993: [https://proceedings.neurips.cc/paper/1993/file/1aa48fc4880bb0c9b8a3bf979d3b917e-Paper.pdf Learning Curves: Asymptotic Values and Rate of Convergence]&lt;br /&gt;
* 2017-12: [https://arxiv.org/abs/1712.00409 Deep Learning Scaling is Predictable, Empirically] (Baidu)&lt;br /&gt;
* 2019-03: [http://www.incompleteideas.net/IncIdeas/BitterLesson.html The Bitter Lesson] (Rich Sutton)&lt;br /&gt;
* 2020-01: [https://arxiv.org/abs/2001.08361 Scaling Laws for Neural Language Models] (OpenAI)&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.14701 Scaling Laws for Autoregressive Generative Modeling] (OpenAI)&lt;br /&gt;
* 2020-05: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis] (Gwern)&lt;br /&gt;
* 2021-08: [https://arxiv.org/abs/2108.07686 Scaling Laws for Deep Learning]&lt;br /&gt;
* 2021-02: [https://arxiv.org/abs/2102.06701 Explaining Neural Scaling Laws] (Google DeepMind)&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.15556 Training Compute-Optimal Large Language Models] (Chinchilla, Google DeepMind)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.04715 Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.07951 Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models]&lt;br /&gt;
* 2025-05: [https://brendel-group.github.io/llm-line/ LLMs on the Line: Data Determines Loss-To-Loss Scaling Laws]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13786 The Art of Scaling Reinforcement Learning Compute for LLMs]&lt;br /&gt;
&lt;br /&gt;
=Information Processing/Storage=&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.10689 A Theory of Usable Information Under Computational Constraints]&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.00008 Why is AI hard and Physics simple?]&lt;br /&gt;
* 2021-06: [https://arxiv.org/abs/2106.06981 Thinking Like Transformers]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.00948 Large Linguistic Models: Investigating LLMs&amp;#039; metalinguistic abilities]&lt;br /&gt;
* &amp;quot;A transformer&amp;#039;s depth affects its reasoning capabilities, whilst model size affects its knowledge capacity&amp;quot; ([https://x.com/danielhanchen/status/1835684061475655967 c.f.])&lt;br /&gt;
** 2024-02: [https://arxiv.org/abs/2402.14905 MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases]&lt;br /&gt;
** 2024-04: [https://arxiv.org/abs/2404.08819 The Illusion of State in State-Space Models] (figure 3)&lt;br /&gt;
** 2024-08: [https://arxiv.org/abs/2408.00118 Gemma 2: Improving Open Language Models at a Practical Size] (table 9)&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.10482 Schrodinger&amp;#039;s Memory: Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2407.01687 Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning]. CoT involves both memorization and (probabilitic) reasoning&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.16679 Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?]&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.03961 A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.22471 The Bayesian Geometry of Transformer Attention]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03220 From Entropy to Epiplexity: Rethinking Information for Computationally Bounded Intelligence]&lt;br /&gt;
&lt;br /&gt;
==Statistics/Math==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.05465 The emergence of clusters in self-attention dynamics]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.10794 A mathematical perspective on Transformers]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.12034 Understanding Transformers via N-gram Statistics]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.06833 Dynamic metastability in the self-attention model]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04551 Measure-to-measure interpolation using Transformers]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14697 Quantitative Clustering in Mean-Field Transformer Models]&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
===For numbers/math===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.14903 Tokenization counts: the impact of tokenization on arithmetic in frontier LLMs]: L2R vs. R2L yields different performance on math&lt;br /&gt;
&lt;br /&gt;
==Data Storage==&lt;br /&gt;
* 1988-09: [https://www.sciencedirect.com/science/article/pii/0885064X88900209 On the capabilities of multilayer perceptrons]&lt;br /&gt;
* 2006-12: [https://ieeexplore.ieee.org/document/4038449 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition] (single-layer perceptron stores &amp;gt;2 bits/parameter; MLP ~ 2*N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; bits w/ N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; params)&lt;br /&gt;
* 2016-11: [https://arxiv.org/abs/1611.09913 Capacity and Trainability in Recurrent Neural Networks] (5 bits/param)&lt;br /&gt;
* 2018-02: [https://arxiv.org/abs/1802.08232 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks]&lt;br /&gt;
* 2019-05: [https://ieeexplore.ieee.org/document/8682462 Memorization Capacity of Deep Neural Networks under Parameter Quantization]&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.08910 How Much Knowledge Can You Pack Into the Parameters of a Language Model?]&lt;br /&gt;
* 2020-08: [https://arxiv.org/abs/2008.09036 Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries] (capacity scales linearly with parameters; more training samples leads to less memorization)&lt;br /&gt;
* 2020-12: [https://arxiv.org/abs/2012.06421 When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05405 Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws] (2 bits/param)&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15720 Scaling Laws for Fact Memorization of Large Language Models] (1T params needed to memorize Wikipedia)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24832 How much do language models memorize?] (3.6 bits/parameter)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01855 Trade-offs in Data Memorization via Strong Data Processing Inequalities]&lt;br /&gt;
&lt;br /&gt;
===Reverse-Engineering Training Data===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.10364 Can We Infer Confidential Properties of Training Data from LLMs?]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15553 Approximating Language Model Training Data from Weights]&lt;br /&gt;
&lt;br /&gt;
===Compression===&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.09410 Less is More: Parameter-Free Text Classification with Gzip]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.04050 LLMZip: Lossless Text Compression using Large Language Models]&lt;br /&gt;
* 2023-07: [https://aclanthology.org/2023.findings-acl.426/ “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.10668 Language Modeling Is Compression]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07550 An Image is Worth 32 Tokens for Reconstruction and Generation]&lt;br /&gt;
&lt;br /&gt;
==Learning/Training==&lt;br /&gt;
* 2018-03: [https://arxiv.org/abs/1803.03635 The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks]: Sparse neural networks are optimal, but it is difficult to identify the right architecture and train it. Deep learning typically consists of training a dense neural network, which makes it easier to learn an internal sparse circuit optimal to a particular problem.&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11521 On the Ability of Deep Networks to Learn Symmetries from Data: A Neural Kernel Theory]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.12391 Physics of Skill Learning]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24864 ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Cross-modal knowledge transfer===&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.07519 Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.07358 Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]: CLIP learns richer set of aggregated representations (e.g. for a culture or country), vs. a vision-only model.&lt;br /&gt;
&lt;br /&gt;
==Hidden State==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06258 Emergent Response Planning in LLM]: They show that the latent representation contains information beyond that needed for the next token (i.e. the model learns to &amp;quot;plan ahead&amp;quot; and encode information relevant to future tokens)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02854 (How) Do Language Models Track State?]&lt;br /&gt;
===Convergent Representation===&lt;br /&gt;
* 2015-11: [https://arxiv.org/abs/1511.07543 Convergent Learning: Do different neural networks learn the same representations?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of Embeddings]: Evidence for [https://x.com/jxmnop/status/1925224620166128039 The Strong Platonic Representation Hypothesis]; models converge to a single consensus reality&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Function Approximation==&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]: can learn linear functions (equivalent to least-squares estimator)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09066 Teaching Algorithmic Reasoning via In-context Learning]: Simple arithmetic &lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models] ([https://github.com/ekinakyurek/google-research/tree/master/incontext code]): can learn linear regression&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.00297 Transformers learn to implement preconditioned gradient descent for in-context learning]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.03576 One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.02893 ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20545 SoS1: O1 and R1-Like Reasoning LLMs are Sum-of-Square Solvers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.21212 Transformers Learn to Implement Multi-step Gradient Descent with Chain of Thought]&lt;br /&gt;
&lt;br /&gt;
=Physics Based=&lt;br /&gt;
* 2014-01: [https://arxiv.org/abs/1401.1219 Consciousness as a State of Matter]&lt;br /&gt;
* 2016-08: [https://arxiv.org/abs/1608.08225 Why does deep and cheap learning work so well?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.23489 SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training]&lt;br /&gt;
* 2025-12: [https://www.pnas.org/doi/full/10.1073/pnas.2523012122 Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning]&lt;br /&gt;
&lt;br /&gt;
=Failure Modes=&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?]: Poor causal inference&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.12288 The Reversal Curse: LLMs trained on &amp;quot;A is B&amp;quot; fail to learn &amp;quot;B is A&amp;quot;]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.13638 Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve] (biases towards &amp;quot;common&amp;quot; numbers, in-context CoT can reduce performance by incorrectly priming, etc.)&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.16093 Visual cognition in multimodal large language models] (models lack human-like visual understanding)&lt;br /&gt;
&lt;br /&gt;
==Adversarial==&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03507 Solving adversarial examples requires solving exponential misalignment]&lt;br /&gt;
&lt;br /&gt;
==Fracture Representation==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.11581 Questioning Representational Optimism in Deep Learning: The Fractured Entangled Representation Hypothesis] ([https://github.com/akarshkumar0101/fer code])&lt;br /&gt;
&lt;br /&gt;
==Jagged Frontier==&lt;br /&gt;
* 2023-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03211 How Does Quantization Affect Multilingual LLMs?]: Quantization degrades different languages by differing amounts&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061v1 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]: Scaling laws are skill-dependent&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.18212 A Definition of AGI]&lt;br /&gt;
&lt;br /&gt;
===See also===&lt;br /&gt;
* [[AI_understanding|AI Understanding]] &amp;gt; [[AI_understanding#Psychology|Psychology]] &amp;gt; [[AI_understanding#LLM_personalities|LLM personalities]]&lt;br /&gt;
* [[AI tricks]] &amp;gt; [[AI_tricks#Prompt_Engineering|Prompt Engineering]] &amp;gt; [[AI_tricks#Brittleness|Brittleness]]&lt;br /&gt;
&lt;br /&gt;
===Conversely (AI models converge)===&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.05117 The Universal Weight Subspace Hypothesis]&lt;br /&gt;
* 2026-01: [https://avikrishna.substack.com/p/eliciting-frontier-model-character Eliciting Frontier Model Character Training: A study of personality convergence across language models]&lt;br /&gt;
&lt;br /&gt;
==Model Collapse==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.12202 Nepotistically Trained Generative-AI Models Collapse]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05280 On the Limits of Self-Improving in LLMs and Why AGI, ASI and the Singularity Are Not Near Without Symbolic Model Synthesis]&lt;br /&gt;
&lt;br /&gt;
===Analysis===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
&lt;br /&gt;
===Mitigation===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08117 Convergence Dynamics and Stabilization Strategies of Co-Evolving Generative Models]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.11111 Inducing anxiety in large language models can induce bias]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17117 From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning]&lt;br /&gt;
* 2025-07: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5357179 Call Me A Jerk: Persuading AI to Comply with Objectionable Requests]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.11328 Do LLMs &amp;quot;Feel&amp;quot;? Emotion Circuits Discovery and Control]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.06047 &amp;quot;They parted illusions -- they parted disclaim marinade&amp;quot;: Misalignment as structural fidelity in LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.02606 Gender Dynamics and Homophily in a Social Network of LLM Agents]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
* 2026-04: [https://transformer-circuits.pub/2026/emotions/index.html Emotion concepts and their function in a large language model] ([https://www.anthropic.com/research/emotion-concepts-function blog])&lt;br /&gt;
&lt;br /&gt;
==Persona Simulator Theory==&lt;br /&gt;
* 2022-09: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators] ([https://www.lesswrong.com/users/janus-1?from=post_header janus])&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.findings-emnlp.423/ Language Models as Agent Models]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.00805 Conditioning Predictive Models: Risks and Strategies]&lt;br /&gt;
* 2024-09: [https://www.lesswrong.com/s/qhdHbCJ3PYesL9dde Intuitive Self-Models]&lt;br /&gt;
* 2026-02: [https://alignment.anthropic.com/2026/psm/ The Persona Selection Model: Why AI Assistants might Behave like Humans] (Anthropic, [https://www.anthropic.com/research/persona-selection-model blog])&lt;br /&gt;
&lt;br /&gt;
==Allow LLM to think==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
&lt;br /&gt;
===In-context Learning===&lt;br /&gt;
* 2021-10: [https://arxiv.org/abs/2110.15943 MetaICL: Learning to Learn In Context]&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?]&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16003 Learning without training: The implicit dynamics of in-context learning]&lt;br /&gt;
&lt;br /&gt;
==Reasoning (CoT, etc.)==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18009 Large Language Models Think Too Fast To Explore Effectively]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18585 Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.08156 Are DeepSeek R1 And Other Reasoning Models More Faithful?]: reasoning models can provide faithful explanations for why their reasoning is correct&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08679 Chain-of-Thought Reasoning In The Wild Is Not Always Faithful]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.04022 Rethinking Reflection in Pre-Training]: pre-training alone already provides some amount of reflection/reasoning&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2501.18858 BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning]&lt;br /&gt;
&lt;br /&gt;
===Pathfinding===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08152 DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09284 Tree-OPO: Off-policy Monte Carlo Tree-Guided Advantage Optimization for Multistep Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06160v1 Reverse-Engineered Reasoning for Open-Ended Generation]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06941 The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2508.01191 Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens]&lt;br /&gt;
&lt;br /&gt;
==Self-Awareness and Self-Recognition and Introspection==&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.05221 Language Models (Mostly) Know What They Know]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13076 LLM Evaluators Recognize and Favor Their Own Generations]&lt;br /&gt;
* 2024-09: [https://situational-awareness-dataset.org/ Me, Myself and AI: The Situational Awareness Dataset for LLMs]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* 2024-12: [https://theaidigest.org/self-awareness AIs are becoming more self-aware. Here&amp;#039;s why that matters]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-04: [https://x.com/Josikinz/status/1907923319866716629 LLMs can guess which comic strip was generated by themselves (vs. other LLM)]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/introspection/index.html Emergent Introspective Awareness in Large Language Models] (Anthropic, [https://www.anthropic.com/research/introspection blog])&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.24661 Do Large Language Models Know What They Are Capable Of?]&lt;br /&gt;
&lt;br /&gt;
==LLM personalities==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.02618 Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04343 Psychologically Enhanced AI Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10387 The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models]&lt;br /&gt;
&lt;br /&gt;
==Quirks &amp;amp; Biases==&lt;br /&gt;
* 2025-04: [https://www.cambridge.org/core/journals/judgment-and-decision-making/article/artificial-intelligence-and-dichotomania/0421D2310727D73FAB47069FD1620AA1 Artificial intelligence and dichotomania]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.22818 Can Large Language Models Develop Gambling Addiction?]&lt;br /&gt;
&lt;br /&gt;
=Vision Models=&lt;br /&gt;
* 2017-11: Distill: [https://distill.pub/2017/feature-visualization/ Feature Visualization: How neural networks build up their understanding of images]&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.12322 Position, Padding and Predictions: A Deeper Look at Position Information in CNNs]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13181 Perception Encoder: The best visual embeddings are not at the output of the network] ([https://github.com/facebookresearch/perception_models code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8779</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8779"/>
		<updated>2026-04-05T01:19:56Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* April 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
** [https://x.com/maxescu/status/2036434854435315868?s=20 Cinematic scenes] (3.5m, comedy, [https://lumalabs.ai/uni-1 Luma Uni-1 Agent])&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;br /&gt;
* March 2026: [https://app.pixverse.ai/onboard Pixverse v6] ([https://x.com/fal/status/2038655807483490613?s=20 example])&lt;br /&gt;
&lt;br /&gt;
====April 2026====&lt;br /&gt;
* April 2026: Examples:&lt;br /&gt;
** [https://x.com/aiordieshow/status/2039679896650125391?s=20 Soothent Paste] (45s)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2040342454193516761?s=20 NEXII] (2m, music video, Seedance 2.0)&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8778</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8778"/>
		<updated>2026-04-05T00:57:07Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* AI improves human work */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6513481 Mapping AI into Production: A Field Experiment on Firm Performance]&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
* 2026-03: Anthropic: [https://www.anthropic.com/features/81k-interviews What 81,000 people want from AI]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41562-025-02194-6 On the conversational persuasiveness of GPT-4]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
** 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.15245 Practicing with Language Models Cultivates Human Empathic Communication]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Talk:AI_video&amp;diff=8777</id>
		<title>Talk:AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Talk:AI_video&amp;diff=8777"/>
		<updated>2026-04-04T23:28:37Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Others for Consideration */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Others for Consideration=&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1927061347331694973 Influenders] ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/venturetwins/status/1934027410841764221 Koala shot by protesters]&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1932835386557939913 Riot] ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** Celebrity explainer [https://x.com/venturetwins/status/1934434222523171000 1], [https://x.com/venturetwins/status/1934438139738874129 2]&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/IamEmily2050/status/1945795374251479388 Quick rap] (example JSON format)&lt;br /&gt;
** [https://x.com/sweeneydailyx/status/1948032121429500221 Commercial for American Eagle (20s)] (the car driving off is an AI extension of the clip)&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1955305090971017653 Waidmanns Heil ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/Gossip_Goblin/status/1996994382428336165?s=20 Joy Loop] (1.5m)&lt;br /&gt;
** [https://x.com/TUPACABRA2/status/2005877025454662066?s=20 Minnesota Dark] (2m, [https://x.com/TUPACABRA2 Tupacrabra])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/2008990455661515071?s=20 Egg Protein] (2m)&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/aimikoda/status/2038285542727487827?s=20 Fashion sequence] (15s, Seedance 2.0)&lt;br /&gt;
* April 2026: Examples:&lt;br /&gt;
** [https://x.com/ganziboy11/status/2040413277122068781?s=20 Zephyr] (2.5m, Higgsfield Seedance 2.0)&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8776</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8776"/>
		<updated>2026-04-02T18:39:50Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
** [https://x.com/maxescu/status/2036434854435315868?s=20 Cinematic scenes] (3.5m, comedy, [https://lumalabs.ai/uni-1 Luma Uni-1 Agent])&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;br /&gt;
* March 2026: [https://app.pixverse.ai/onboard Pixverse v6] ([https://x.com/fal/status/2038655807483490613?s=20 example])&lt;br /&gt;
&lt;br /&gt;
====April 2026====&lt;br /&gt;
* April 2026: Examples:&lt;br /&gt;
** [https://x.com/aiordieshow/status/2039679896650125391?s=20 Soothent Paste] (45s)&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_Agents&amp;diff=8775</id>
		<title>AI Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_Agents&amp;diff=8775"/>
		<updated>2026-04-02T17:55:33Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Automated Improvement */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=Reviews &amp;amp; Perspectives=&lt;br /&gt;
===Published===&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05221 LLM Reasoners: New Evaluation, Library, and Analysis of Step-by-Step Reasoning with Large Language Models] ([https://github.com/maitrix-org/llm-reasoners code])&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.02479 From LLMs to LLM-based Agents for Software Engineering: A Survey of Current, Challenges and Future]&lt;br /&gt;
* 2024-09: [https://doi.org/10.1039/D4DD00178H Towards a Science Exocortex]&lt;br /&gt;
* 2024-09: [https://www.arxiv.org/abs/2409.02977 Large Language Model-Based Agents for Software Engineering: A Survey]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09030 Agents in Software Engineering: Survey, Landscape, and Vision]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.01990 Advances and Challenges in Foundation Agents: From Brain-Inspired Intelligence to Evolutionary, Collaborative, and Safe Systems]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2503.19213 A Survey of Large Language Model Agents for Question Answering]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.09037 A Survey of Frontiers in LLM Reasoning: Inference Scaling, Learning to Reason, and Agentic Systems]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.01990 Advances and Challenges in Foundation Agents: From Brain-Inspired Intelligence to Evolutionary, Collaborative, and Safe Systems]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.12538 Agentic Reasoning for Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Continually updating===&lt;br /&gt;
* [https://github.com/open-thought/system-2-research OpenThought - System 2 Research Links]&lt;br /&gt;
* [https://github.com/hijkzzz/Awesome-LLM-Strawberry Awesome LLM Strawberry (OpenAI o1): Collection of research papers &amp;amp; blogs for OpenAI Strawberry(o1) and Reasoning]&lt;br /&gt;
* [https://github.com/e2b-dev/awesome-ai-agents Awesome AI Agents]&lt;br /&gt;
&lt;br /&gt;
===Analysis/Opinions===&lt;br /&gt;
* [https://arxiv.org/abs/2402.01817v3 LLMs Can&amp;#039;t Plan, But Can Help Planning in LLM-Modulo Frameworks]&lt;br /&gt;
* [https://rasa.com/blog/cutting-ai-assistant-costs-the-power-of-enhancing-llms-with-business/ Cutting AI Assistant Costs by Up to 77.8%: The Power of Enhancing LLMs with Business Logic]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.10468 AI Agents vs. Agentic AI: A Conceptual Taxonomy, Applications and Challenges]&lt;br /&gt;
&lt;br /&gt;
===Guides===&lt;br /&gt;
* Anthropic: [https://www.anthropic.com/research/building-effective-agents Building Effective Agents]&lt;br /&gt;
* Google: [https://www.kaggle.com/whitepaper-agents Agents] and [https://www.kaggle.com/whitepaper-agent-companion Agents Companion]&lt;br /&gt;
* OpenAI: [https://cdn.openai.com/business-guides-and-resources/a-practical-guide-to-building-agents.pdf A practical guide to building agents]&lt;br /&gt;
* Anthropic: [https://www.anthropic.com/engineering/claude-code-best-practices Claude Code: Best practices for agentic coding]&lt;br /&gt;
* Anthropic: [https://www.anthropic.com/engineering/built-multi-agent-research-system How we built our multi-agent research system]&lt;br /&gt;
&lt;br /&gt;
=AI Assistants=&lt;br /&gt;
&lt;br /&gt;
==Components of AI Assistants==&lt;br /&gt;
&lt;br /&gt;
===Agent Internal Workflow Management===&lt;br /&gt;
* [https://github.com/langchain-ai/langchain LangChain]&lt;br /&gt;
* [https://github.com/pydantic/pydantic-ai Pydantic: Agent Framework / shim to use Pydantic with LLMs]&lt;br /&gt;
* [https://github.com/lmnr-ai/flow Flow: A lightweight task engine for building AI agents that prioritizes simplicity and flexibility]&lt;br /&gt;
* [https://llama-stack.readthedocs.io/en/latest/index.html llama-stack]&lt;br /&gt;
* [https://huggingface.co/blog/smolagents Huggingface] [https://github.com/huggingface/smolagents smolagents]&lt;br /&gt;
* [https://github.com/elizaOS/eliza Eliza] (includes multi-agent, interaction with docs, Discord, Twitter, etc.)&lt;br /&gt;
* [https://github.com/The-Pocket/PocketFlow Pocket Flow]: LLM Framework in 100 Lines&lt;br /&gt;
* [https://github.com/coze-dev/coze-studio Coze]: All-in-one AI agent development tool&lt;br /&gt;
&lt;br /&gt;
===Information Retrieval (Memory)===&lt;br /&gt;
* See also [[AI_tools#Retrieval_Augmented_Generation_.28RAG.29|RAG]].&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13501 A Survey on the Memory Mechanism of Large Language Model based Agents]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09713 Agentic Information Retrieval]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.01142 DeepRAG: Thinking to Retrieval Step by Step for Large Language Models]&lt;br /&gt;
* [https://mem0.ai/ Mem0 AI]: Memory Layer for AI Agents; self-improving memory layer for LLM applications, enabling personalized.&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16153 Memento: Fine-tuning LLM Agents without Fine-tuning LLMs]&lt;br /&gt;
&lt;br /&gt;
===Contextual Memory===&lt;br /&gt;
* [https://github.com/memodb-io/memobase Memobase]: user profile-based memory (long-term user memory for genAI) applications)&lt;br /&gt;
&lt;br /&gt;
===Control (tool-use, computer use, etc.)===&lt;br /&gt;
* See also: [[Human_Computer_Interaction#AI_Computer_Use]]&lt;br /&gt;
* [https://tavily.com/ Tavily]: Connect Your LLM to the Web: Empowering your AI applications with real-time, accurate search results tailored for LLMs and RAG&lt;br /&gt;
===Model Context Protocol (MCP)===&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Standards:&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
*# Anthropic [https://www.anthropic.com/news/model-context-protocol Model Context Protocol] (MCP)&lt;br /&gt;
*# [https://openai.github.io/openai-agents-python/mcp/ OpenAI Agents SDK]&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Tools:&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** [https://github.com/jlowin/fastmcp FastMCP]: The fast, Pythonic way to build MCP servers&lt;br /&gt;
** [https://github.com/fleuristes/fleur/ Fleur]: A desktop app marketplace for Claude Desktop&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Servers:&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Lists:&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
**# [https://github.com/modelcontextprotocol/servers Model Context Protocol servers]&lt;br /&gt;
**# [https://www.mcpt.com/ MCP Servers, One Managed Registry]&lt;br /&gt;
**# [https://github.com/punkpeye/awesome-mcp-servers Awesome MCP Servers]&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Noteworthy:&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
**# Official [https://github.com/github/github-mcp-server Github MCP server]&lt;br /&gt;
**# Unofficial [https://github.com/modelcontextprotocol/servers/tree/main/src/github Github MCP server]&lt;br /&gt;
**# [https://github.com/modelcontextprotocol/servers/tree/main/src/puppeteer Puppeteer]&lt;br /&gt;
**# [https://github.com/modelcontextprotocol/servers/tree/main/src/google-maps Google Maps MCP Server]&lt;br /&gt;
**# [https://github.com/modelcontextprotocol/servers/tree/main/src/slack Slack MCP Server]&lt;br /&gt;
**# [https://zapier.com/mcp Zapier MCP Servers] (Slack, Google Sheets, Notion, etc.)&lt;br /&gt;
**# [https://github.com/awslabs/mcp AWS MCP Servers]&lt;br /&gt;
**# [https://x.com/elevenlabsio/status/1909300782673101265 ElevenLabs]&lt;br /&gt;
&lt;br /&gt;
===Agent2Agent Protocol (A2A)===&lt;br /&gt;
* Google [https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/ announcement]&lt;br /&gt;
&lt;br /&gt;
===Open-source===&lt;br /&gt;
* [https://khoj.dev/ Khoj] ([https://github.com/khoj-ai/khoj code]): self-hostable AI assistant&lt;br /&gt;
* [https://github.com/ragapp/ragapp RAGapp]: Agentic RAG for enterprise&lt;br /&gt;
* STORM: Synthesis of Topic Outlines through Retrieval and Multi-perspective Question Asking&lt;br /&gt;
** Can write (e.g.) Wikipedia-style articles&lt;br /&gt;
** [https://github.com/stanford-oval/storm/tree/NAACL-2024-code-backup code]&lt;br /&gt;
** Preprint: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Personalities/Personas===&lt;br /&gt;
* 2023-10: [https://doi.org/10.1145/3586183.3606763 Generative Agents: Interactive Simulacra of Human Behavior]&lt;br /&gt;
* 2024-11: Microsoft [https://github.com/microsoft/TinyTroupe TinyTroupe 🤠🤓🥸🧐: LLM-powered multiagent persona simulation for imagination enhancement and business insights]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.10109 Generative Agent Simulations of 1,000 People] ([https://github.com/joonspk-research/genagents code])&lt;br /&gt;
&lt;br /&gt;
==Specific Uses for AI Assistants==&lt;br /&gt;
&lt;br /&gt;
===Computer Use===&lt;br /&gt;
* See: [[Human_Computer_Interaction#AI_Computer_Use]]&lt;br /&gt;
&lt;br /&gt;
===Software Engineering===&lt;br /&gt;
* 2024-11: [https://github.com/MLSysOps/MLE-agent MLE-Agent: Your intelligent companion for seamless AI engineering and research]&lt;br /&gt;
* [https://github.com/OpenAutoCoder/Agentless Agentless]: agentless approach to automatically solve software development problems&lt;br /&gt;
&lt;br /&gt;
===Science Agents===&lt;br /&gt;
See [[Science Agents]].&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
* 2025-03: [https://news.microsoft.com/2025/03/03/microsoft-dragon-copilot-provides-the-healthcare-industrys-first-unified-voice-ai-assistant-that-enables-clinicians-to-streamline-clinical-documentation-surface-information-and-automate-task/ Microsoft Dragon Copilot]: streamline clinical workflows and paperwork&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.05186 Training state-of-the-art pathology foundation models with orders of magnitude less data]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.20148 The Anatomy of a Personal Health Agent]&lt;br /&gt;
&lt;br /&gt;
===LLM-as-judge===&lt;br /&gt;
* [https://x.com/cwolferesearch/status/1812949923010421192 List of papers].&lt;br /&gt;
* [https://www.philschmid.de/llm-evaluation LLM Evaluation doesn&amp;#039;t need to be complicated]&lt;br /&gt;
* [https://eugeneyan.com/writing/llm-evaluators/ Evaluating the Effectiveness of LLM-Evaluators (aka LLM-as-Judge)]&lt;br /&gt;
* [https://github.com/llm-as-a-judge/Awesome-LLM-as-a-judge Awesome-LLM-as-a-judge Survey]&lt;br /&gt;
* [https://github.com/haizelabs/Awesome-LLM-Judges haizelabs Awesome LLM Judges]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.02666 Self-Taught Evaluators]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.10934 Agent-as-a-Judge: Evaluate Agents with Agents]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15594 A Survey on LLM-as-a-Judge]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.05579 LLMs-as-Judges: A Comprehensive Survey on LLM-based Evaluation Methods]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19877 Scaling Evaluation-time Compute with Reasoning Models as Process Evaluators]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.00050 JudgeLRM: Large Reasoning Models as a Judge]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05111 Agent-as-a-Judge]&lt;br /&gt;
&lt;br /&gt;
===Deep Research===&lt;br /&gt;
* Google [https://blog.google/products/gemini/google-gemini-deep-research/ Deep Research]&lt;br /&gt;
* OpenAI [https://openai.com/index/introducing-deep-research/ Deep Research]&lt;br /&gt;
* Perplexity:&lt;br /&gt;
** [https://www.perplexity.ai/ Search]&lt;br /&gt;
** [https://www.perplexity.ai/hub/blog/introducing-perplexity-deep-research Deep Research]&lt;br /&gt;
* [https://exa.ai/ Exa AI]:&lt;br /&gt;
** [https://exa.ai/websets Websets]: Web research agent&lt;br /&gt;
** [https://demo.exa.ai/deepseekchat Web-search agent] powered by DeepSeek ([https://github.com/exa-labs/exa-deepseek-chat code]) or [https://o3minichat.exa.ai/ o3-mini] ([https://github.com/exa-labs/exa-o3mini-chat code])&lt;br /&gt;
* [https://www.firecrawl.dev/ Firecrawl] [https://x.com/nickscamara_/status/1886287956291338689 wip]&lt;br /&gt;
* [https://x.com/mattshumer_ Matt Shumer] [https://github.com/mshumer/OpenDeepResearcher OpenDeepResearcher]&lt;br /&gt;
* [https://github.com/zilliztech/deep-searcher DeepSearcher] (operate on local data)&lt;br /&gt;
* [https://github.com/nickscamara nickscamara] [https://github.com/nickscamara/open-deep-research open-deep-research]&lt;br /&gt;
* [https://x.com/dzhng dzhng] [https://github.com/dzhng/deep-research deep-research]&lt;br /&gt;
* [https://huggingface.co/ huggingface] [https://huggingface.co/blog/open-deep-research open-Deep-research ([https://github.com/huggingface/smolagents/tree/main/examples/open_deep_research code])&lt;br /&gt;
* xAI Grok 3 Deep Search&lt;br /&gt;
* [https://liner.com/news/introducing-deepresearch Liner Deep Research]&lt;br /&gt;
* [https://allenai.org/ Allen AI] (AI2) [https://paperfinder.allen.ai/chat Paper Finder]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.20201 Open Deep Search: Democratizing Search with Open-source Reasoning Agents] ([https://github.com/sentient-agi/OpenDeepSearch code])&lt;br /&gt;
* [https://convergence.ai/welcome Convergence AI] Deep Work (swarms for web-based tasks)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.03160 DeepResearcher: Scaling Deep Research via Reinforcement Learning in Real-world Environments]&lt;br /&gt;
* 2025-04: Anthropic [https://x.com/AnthropicAI/status/1912192384588271771 Research]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.21776 WebThinker: Empowering Large Reasoning Models with Deep Research Capability]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06283 SFR-DeepResearch: Towards Effective Reinforcement Learning for Autonomously Reasoning Single Agents]&lt;br /&gt;
&lt;br /&gt;
=Advanced Workflows=&lt;br /&gt;
* [https://salesforce-research-dei-agents.github.io/ Salesforce DEI]: meta-system that leverages a diversity of SWE agents&lt;br /&gt;
** Preprint: [https://www.arxiv.org/abs/2408.07060 Diversity Empowers Intelligence: Integrating Expertise of Software Engineering Agents]&lt;br /&gt;
* [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]&lt;br /&gt;
** [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
** [https://github.com/SakanaAI/AI-Scientist code]&lt;br /&gt;
* [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning]&lt;br /&gt;
** [https://github.com/lamm-mit/SciAgentsDiscovery code]&lt;br /&gt;
* [https://skywork.ai/home Skywork] [https://skywork.ai/home?inviter=el.cine&amp;amp;shortlink_id=1919604877427924992&amp;amp;utm_source=X Super Agent]&lt;br /&gt;
&lt;br /&gt;
===Streamline Administrative Tasks===&lt;br /&gt;
* 2025-02: [https://er.educause.edu/articles/2025/2/ushering-in-a-new-era-of-ai-driven-data-insights-at-uc-san-diego Ushering in a New Era of AI-Driven Data Insights at UC San Diego]&lt;br /&gt;
&lt;br /&gt;
===Author Research Articles===&lt;br /&gt;
* 2024-02: STORM: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models] ([https://www.aihero.dev/storm-generate-high-quality-articles-based-on-real-research discussion/analysis])&lt;br /&gt;
&lt;br /&gt;
===Software Development Workflows===&lt;br /&gt;
Several paradigms of AI-assisted coding have arisen:&lt;br /&gt;
# Manual, human driven&lt;br /&gt;
# AI-aided through chat/dialogue, where the human asks for code and then copies it into the project&lt;br /&gt;
## OpenAI [https://chatgpt.com/ ChatGPT]&lt;br /&gt;
## Anthropic [https://claude.ai/ Claude]&lt;br /&gt;
## Google [https://gemini.google.com/app Gemini]&lt;br /&gt;
# API calls to an LLM, which generates code and inserts the file into the project&lt;br /&gt;
# LLM-integration into the IDE&lt;br /&gt;
## [https://github.com/features/copilot Copilot]&lt;br /&gt;
## [https://www.qodo.ai/ Qodo] (Codium) &amp;amp; [https://www.qodo.ai/products/alphacodium/ AlphaCodium] ([https://arxiv.org/abs/2401.08500 preprint], [https://github.com/Codium-ai/AlphaCodium code])&lt;br /&gt;
## &amp;#039;&amp;#039;&amp;#039;[https://www.cursor.com/ Cursor]&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
## [https://codeium.com/ Codeium] [https://codeium.com/windsurf Windsurf] (with &amp;quot;Cascade&amp;quot; AI Agent)&lt;br /&gt;
## ByteDance [https://www.trae.ai/ Trae AI]&lt;br /&gt;
## [https://www.tabnine.com/ Tabnine]&lt;br /&gt;
## [https://marketplace.visualstudio.com/items?itemName=Traycer.traycer-vscode Traycer]&lt;br /&gt;
## [https://idx.dev/ IDX]: free&lt;br /&gt;
## [https://github.com/codestoryai/aide Aide]: open-source AI-native code editor (fork of VS Code)&lt;br /&gt;
## [https://www.continue.dev/ continue.dev]: open-source code assistant&lt;br /&gt;
## [https://trypear.ai/ Pear AI]: open-source code editor&lt;br /&gt;
## [https://haystackeditor.com/ Haystack Editor]: canvas UI&lt;br /&gt;
## [https://onlook.com/ Onlook]: for designers&lt;br /&gt;
## [https://www.all-hands.dev/ All Hands AI]&lt;br /&gt;
## [https://app.devin.ai/ Devin 2.0] ([https://cognition.ai/ Cognition AI])&lt;br /&gt;
## Google [https://firebase.google.com/docs/studio Firebase Studio]&lt;br /&gt;
## [https://github.com/rowboatlabs/rowboat rowboat] (for building multi-agent workflows)&lt;br /&gt;
## [https://www.trae.ai/ Trae IDE]: The Real AI Engineer&lt;br /&gt;
# AI-assisted IDE, where the AI generates and manages the dev environment&lt;br /&gt;
## [https://replit.com/ Replit]&lt;br /&gt;
## [https://www.pythagora.ai/ Pythagora]&lt;br /&gt;
## [https://stackblitz.com/ StackBlitz] [https://bolt.new/ bolt.new]&lt;br /&gt;
## [https://github.com/clinebot/cline Cline] (formerly [https://generativeai.pub/meet-claude-dev-an-open-source-autonomous-ai-programmer-in-vs-code-f457f9821b7b Claude Dev])&lt;br /&gt;
## [https://www.all-hands.dev/ All Hands]&lt;br /&gt;
# AI Agent on Commandline&lt;br /&gt;
## [https://aider.chat/ Aider] ([https://github.com/Aider-AI/aider code]): Pair programming on commandline&lt;br /&gt;
## [https://docs.anthropic.com/en/docs/claude-code/overview Claude Code]&lt;br /&gt;
## [https://openai.com/codex/ OpenAI Codex]&lt;br /&gt;
## [https://blog.google/technology/developers/introducing-gemini-cli-open-source-ai-agent/ Gemini CLI]&lt;br /&gt;
# Prompt-to-product&lt;br /&gt;
## [https://githubnext.com/projects/github-spark Github Spark] ([https://x.com/ashtom/status/1851333075374051725 demo video])&lt;br /&gt;
## [https://www.create.xyz/ Create.xyz]: text-to-app, replicate product from link&lt;br /&gt;
## [https://a0.dev/ a0.dev]: generate mobil apps (from your phone)&lt;br /&gt;
## [https://softgen.ai/ Softgen]: web app developer&lt;br /&gt;
## [https://wrapifai.com/ wrapifai]: build form-based apps&lt;br /&gt;
## [https://lovable.dev/ Lovable]: web app (from text, screenshot, etc.)&lt;br /&gt;
## [https://v0.dev/ Vercel v0]&lt;br /&gt;
## [https://x.com/johnrushx/status/1625179509728198665 MarsX] ([https://x.com/johnrushx John Rush]): SaaS builder&lt;br /&gt;
## [https://webdraw.com/ Webdraw]: turn sketches into web apps&lt;br /&gt;
## [https://www.tempo.new/ Tempo Labs]: build React apps&lt;br /&gt;
## [https://databutton.com/ Databutton]: no-code software development&lt;br /&gt;
## [https://base44.com/ base44]: no-code dashboard apps&lt;br /&gt;
## [https://www.theorigin.ai/ Origin AI]&lt;br /&gt;
## [https://app.emergent.sh/ Emergent AI]&lt;br /&gt;
# Semi-autonomous software engineer agents&lt;br /&gt;
## [https://www.cognition.ai/blog/introducing-devin Devin] (Cognition AI)&lt;br /&gt;
## [https://aws.amazon.com/q/ Amazon Q] (and CodeWhisperer)&lt;br /&gt;
## [https://honeycomb.sh/ Honeycomb]&lt;br /&gt;
## [https://www.blackbox.ai/ Agent IDE]&lt;br /&gt;
## [https://docs.anthropic.com/en/docs/agents-and-tools/claude-code/overview Claude Code]&lt;br /&gt;
## OpenAI [https://help.openai.com/en/articles/11096431-openai-codex-cli-getting-started Codex CLI] and [https://openai.com/index/introducing-codex/ Codex] cloud&lt;br /&gt;
## [https://www.factory.ai/ Factory AI] [https://x.com/FactoryAI/status/1927754706014630357 Droids]&lt;br /&gt;
For a review of the current state of software-engineering agentic approaches, see:&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.02479 From LLMs to LLM-based Agents for Software Engineering: A Survey of Current, Challenges and Future]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.02977 Large Language Model-Based Agents for Software Engineering: A Survey]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09030 Agents in Software Engineering: Survey, Landscape, and Vision]&lt;br /&gt;
&lt;br /&gt;
=Corporate AI Agent Ventures=&lt;br /&gt;
==Mundane Workflows and Capabilities==&lt;br /&gt;
* [https://www.paymanai.com/ Payman AI]: AI to Human platform that allows AI to pay people for what it needs&lt;br /&gt;
* [https://www.voiceflow.com/ VoiceFlow]: Build customer experiences with AI&lt;br /&gt;
* [https://mistral.ai/ Mistral AI]: [https://mistral.ai/news/build-tweak-repeat/ genAI applications]&lt;br /&gt;
* [https://www.taskade.com/ Taskade]: Task/milestone software with AI agent workflows&lt;br /&gt;
* [https://www.covalent.xyz/ Covalent]: [https://docs.covalent.xyz/docs/cloud/tutorials-cloud/multi_agent/ Building a Multi-Agent Prompt Refining Application]&lt;br /&gt;
&lt;br /&gt;
==Inference-compute Reasoning==&lt;br /&gt;
* [https://nousresearch.com/#popup-menu-anchor Nous Research]: [https://nousresearch.com/introducing-the-forge-reasoning-api-beta-and-nous-chat-an-evolution-in-llm-inference/ Forge Reasoning API Beta]&lt;br /&gt;
&lt;br /&gt;
==AI Assistant==&lt;br /&gt;
* [https://convergence.ai/ Convergence] [https://proxy.convergence.ai/ Proxy]&lt;br /&gt;
* [https://www.shortwave.com/ Shortwave] [https://www.shortwave.com/docs/guides/ai-assistant/ AI Assistant] (organize, write, search, schedule, etc.)&lt;br /&gt;
* 2026-02: [https://telepath.computer/ Telepath]&lt;br /&gt;
&lt;br /&gt;
==Agentic Systems==&lt;br /&gt;
* [https://topologychat.com/ Topology AI]&lt;br /&gt;
* [https://www.cognition.ai/ Cognition AI]: [https://www.cognition.ai/blog/introducing-devin Devin] software engineer (14% SWE-Agent)&lt;br /&gt;
* [https://honeycomb.sh/ Honeycomb] ([https://honeycomb.sh/blog/swe-bench-technical-report 22% SWE-Agent])&lt;br /&gt;
* [https://www.factory.ai/ Factory AI]&lt;br /&gt;
* [https://convergence.ai/welcome Convergence AI] Deep Work (swarms for web-based tasks)&lt;br /&gt;
* [https://agents.cloudflare.com/ Cloudflare Agents]&lt;br /&gt;
* [https://www.maskara.ai/ Maskara AI]&lt;br /&gt;
&lt;br /&gt;
=Increasing AI Agent Intelligence=&lt;br /&gt;
See: [[Increasing AI Intelligence]]&lt;br /&gt;
&lt;br /&gt;
=Multi-agent orchestration=&lt;br /&gt;
==Research==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.02533 Multi-Agent Design: Optimizing Agents with Better Prompts and Topologies]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13657 Why Do Multi-Agent LLM Systems Fail?]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.15478 SWEET-RL: Training Multi-Turn LLM Agents on Collaborative Reasoning Tasks]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20175 Federation of Agents: A Semantics-Aware Communication Fabric for Large-Scale Agentic AI]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2510.26658 The Era of Agentic Organization: Learning to Organize with Language Models] (Microsoft)&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.08296 Towards a Science of Scaling Agent Systems] (Google DeepMind)&lt;br /&gt;
** 2026-01: [https://research.google/blog/towards-a-science-of-scaling-agent-systems-when-and-why-agent-systems-work/ Towards a science of scaling agent systems: When and why agent systems work] &lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.04748 When Single-Agent with Skills Replace Multi-Agent Systems and When They Fail]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.11865 Intelligent AI Delegation]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01213 Can AI Agents Agree?]&lt;br /&gt;
&lt;br /&gt;
===Organization Schemes===&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02390 ReSo: A Reward-driven Self-organizing LLM-based Multi-Agent System for Reasoning Tasks]&lt;br /&gt;
&lt;br /&gt;
===Societies and Communities of AI agents===&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.10270 Cultural Evolution of Cooperation among LLM Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-05: [https://www.science.org/doi/10.1126/sciadv.adu9368 Emergent social conventions and collective bias in LLM populations]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.10147 Virtual Agent Economies]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10825 Reasoning Models Generate Societies of Thought]&lt;br /&gt;
&lt;br /&gt;
===Domain-specific===&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.20138 TradingAgents: Multi-Agents LLM Financial Trading Framework]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
&lt;br /&gt;
==Research demos==&lt;br /&gt;
* [https://github.com/camel-ai/camel Camel]&lt;br /&gt;
* [https://github.com/farizrahman4u/loopgpt/tree/main LoopGPT]&lt;br /&gt;
* [https://github.com/microsoft/JARVIS JARVIS]&lt;br /&gt;
* [https://github.com/agiresearch/OpenAGI OpenAGI]&lt;br /&gt;
* [https://github.com/microsoft/autogen AutoGen]&lt;br /&gt;
** preprint: [https://arxiv.org/abs/2308.08155 AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation]&lt;br /&gt;
** [https://github.com/EmergenceAI/Agent-E Agent-E]: Browser (eventually computer) automation ([https://github.com/EmergenceAI/Agent-E code], [https://arxiv.org/abs/2407.13032 preprint], [https://www.youtube.com/watch?v=uyE7tfKkB0E demo video])&lt;br /&gt;
** [https://www.microsoft.com/en-us/research/blog/introducing-autogen-studio-a-low-code-interface-for-building-multi-agent-workflows/ AutoGen Studio]: GUI for agent workflows ([https://github.com/microsoft/autogen/tree/main/samples/apps/autogen-studio code])&lt;br /&gt;
** [https://www.microsoft.com/en-us/research/articles/magentic-one-a-generalist-multi-agent-system-for-solving-complex-tasks/ Magentic-One: A Generalist Multi-Agent System for Solving Complex Tasks]&lt;br /&gt;
* [https://github.com/ag2ai/ag2 AG2] (previously [https://github.com/microsoft/autogen AutoGen]) ([https://github.com/ag2ai/ag2 code], [https://ag2ai.github.io/ag2/ docs], [https://discord.com/invite/pAbnFJrkgZ Discord])&lt;br /&gt;
* [https://github.com/microsoft/TaskWeaver TaskWeaver]&lt;br /&gt;
* [https://github.com/geekan/MetaGPT MetaGPT]&lt;br /&gt;
* [https://agpt.co/ AutoGPT] ([https://github.com/Significant-Gravitas/AutoGPT code]); and [https://agpt.co/blog/introducing-the-autogpt-platform AutoGPT Platform]&lt;br /&gt;
* [https://chenweize1998.github.io/optima-project-page/ Optima]&lt;br /&gt;
** preprint: [https://arxiv.org/abs/2410.08115 Optima: Optimizing Effectiveness and Efficiency for LLM-Based Multi-Agent System]&lt;br /&gt;
** [https://github.com/thunlp/Optima code]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05221 LLM Reasoners: New Evaluation, Library, and Analysis of Step-by-Step Reasoning with Large Language Models] ([https://github.com/maitrix-org/llm-reasoners code])&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11638 MASAI: Modular Architecture for Software-engineering AI Agents]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08164 Agent S: An Open Agentic Framework that Uses Computers Like a Human] ([https://github.com/simular-ai/Agent-S code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.20424 AutoKaggle: A Multi-Agent Framework for Autonomous Data Science Competitions]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.16111 PlanGEN: A Multi-Agent Framework for Generating Planning and Reasoning Trajectories for Complex Problem Solving]&lt;br /&gt;
&lt;br /&gt;
===Related work===&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.18416 PersonaGym: Evaluating Persona Agents and LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13946 Hallucination Mitigation using Agentic AI Natural Language-Based Frameworks]&lt;br /&gt;
&lt;br /&gt;
===Inter-agent communications===&lt;br /&gt;
* 2024-10: Agora: [https://agoraprotocol.org/ A Scalable Communication Protocol for Networks of Large Language Models] ([https://arxiv.org/abs/2410.11905 preprint]): disparate agents auto-negotiate communication protocol&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.02820 DroidSpeak: Enhancing Cross-LLM Communication]: Exploits caches of embeddings and key-values, to allow context to be more easily transferred between AIs (without consuming context window)&lt;br /&gt;
* 2024-11: Anthropic describes [https://www.anthropic.com/news/model-context-protocol Model Context Protocol]: an open standard for secure, two-way connections between data sources and AI ([https://modelcontextprotocol.io/introduction intro], [https://modelcontextprotocol.io/quickstart quickstart], [https://github.com/modelcontextprotocol code])&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20175 Federation of Agents: A Semantics-Aware Communication Fabric for Large-Scale Agentic AI]&lt;br /&gt;
&lt;br /&gt;
==Architectures==&lt;br /&gt;
* [https://arxiv.org/abs/2406.04692 Mixture-of-Agents Enhances Large Language Model Capabilities]&lt;br /&gt;
* [https://motleycrew.ai/ Motleycrew.ai] ([https://github.com/ShoggothAI/motleycrew code])&lt;br /&gt;
&lt;br /&gt;
==Open Source Frameworks==&lt;br /&gt;
* [https://github.com/langchain-ai/langchain LangChain]&lt;br /&gt;
* [https://x.com/wgussml/status/1833615864131948756 ell] ([https://github.com/MadcowD/ell code], [https://docs.ell.so/ docs])&lt;br /&gt;
* [https://www.agentops.ai/ AgentOps AI] [https://github.com/AgentOps-AI/AgentStack AgentStack]&lt;br /&gt;
* [https://github.com/phidatahq/phidata/tree/main/cookbook/playground Agent UI]&lt;br /&gt;
* kyegomez [https://github.com/kyegomez/swarms swarms]&lt;br /&gt;
* OpenAI [https://github.com/openai/swarm Swarm] ([https://cookbook.openai.com/examples/orchestrating_agents cookbook])&lt;br /&gt;
* Amazon AWS [https://github.com/awslabs/multi-agent-orchestrator Multi-Agent Orchestrator]&lt;br /&gt;
* [https://github.com/kaiban-ai/KaibanJS KaibanJS]: Kanban for AI Agents? (Takes inspiration from [https://en.wikipedia.org/wiki/Kanban Kanban] visual [https://www.atlassian.com/agile/kanban work management].)&lt;br /&gt;
* [https://github.com/Thytu/Agentarium Agentarium]&lt;br /&gt;
* [https://orchestra.org/ Orchestra] ([https://docs.orchestra.org/orchestra/introduction docs], [https://docs.orchestra.org/orchestra/introduction code])&lt;br /&gt;
* [https://github.com/HKUDS/AutoAgent AutoAgent]: Fully-Automated &amp;amp; Zero-Code LLM Agent Framework&lt;br /&gt;
* [https://mastra.ai/ Mastra] ([https://github.com/mastra-ai/mastra github]): opinionated Typescript framework for AI applications (primitives for workflows, agents, RAG, integrations and evals)&lt;br /&gt;
* [https://github.com/orra-dev/orra Orra]: multi-agent applications with complex real-world interactions&lt;br /&gt;
* [https://github.com/gensx-inc/gensx/blob/main/README.md GenSX]&lt;br /&gt;
* Cloudflare [https://developers.cloudflare.com/agents/ agents-sdk] ([https://blog.cloudflare.com/build-ai-agents-on-cloudflare/ info], [https://github.com/cloudflare/agents code])&lt;br /&gt;
* OpenAI [https://platform.openai.com/docs/api-reference/responses responses API] and [https://platform.openai.com/docs/guides/agents agents SDK]&lt;br /&gt;
* Google [https://google.github.io/adk-docs/ Agent Development Kit]&lt;br /&gt;
&lt;br /&gt;
==Open Source Systems==&lt;br /&gt;
* ControlFlow&lt;br /&gt;
** [https://controlflow.ai/welcome documentation]&lt;br /&gt;
** [https://github.com/PrefectHQ/ControlFlow code]&lt;br /&gt;
* OpenHands (formerly [https://github.com/OpenDevin/OpenDevin OpenDevin])&lt;br /&gt;
** [https://github.com/All-Hands-AI/OpenHands code]: platform for autonomous software engineers, powered by AI and LLMs&lt;br /&gt;
** Report: [https://arxiv.org/abs/2407.16741 OpenDevin: An Open Platform for AI Software Developers as Generalist Agents]&lt;br /&gt;
&lt;br /&gt;
==Commercial Automation Frameworks==&lt;br /&gt;
* [https://lutra.ai/ Lutra]: Automation and integration with various web systems.&lt;br /&gt;
* [https://www.gumloop.com/ Gumloop]&lt;br /&gt;
* [https://www.textql.com/ TextQL]: Enterprise Virtual Data Analyst&lt;br /&gt;
* [https://www.athenaintelligence.ai/ Athena intelligence]: Analytics platform&lt;br /&gt;
* [https://gpt.nexus/ Nexus GPT]: Business co-pilot&lt;br /&gt;
* [https://www.multion.ai/ Multi-On]: AI agent that acts on your behalf&lt;br /&gt;
* [https://www.firecrawl.dev/ Firecrawl]: Turn websites into LLM-ready data&lt;br /&gt;
* [https://www.reworkd.ai/ Reworkd]: End-to-end data extraction&lt;br /&gt;
* [https://www.lindy.ai/ Lindy]: Custom AI Assistants to automate business workflows&lt;br /&gt;
** E.g. [https://x.com/Lindyydrope/status/1821373025125556423 use Slack]&lt;br /&gt;
* [https://www.bardeen.ai/ Bardeen]: Automate workflows&lt;br /&gt;
* [https://abacus.ai/ Abacus]: [https://abacus.ai/ai_agents AI Agents]&lt;br /&gt;
** [https://abacus.ai/help/howTo HowTo]&lt;br /&gt;
* [https://www.llamaindex.ai/ LlamaIndex]: ([https://x.com/llama_index 𝕏], [https://github.com/run-llama/llama_index code], [https://docs.llamaindex.ai/en/stable/ docs], [https://discord.com/invite/dGcwcsnxhU Discord])&lt;br /&gt;
* [https://www.multion.ai/ MultiOn AI]: [https://www.multion.ai/blog/introducing-agent-q-research-breakthrough-for-the-next-generation-of-ai-agents-with-planning-and-self-healing-capabilities Agent Q] ([https://multion-research.s3.us-east-2.amazonaws.com/AgentQ.pdf paper]) automated planning and execution&lt;br /&gt;
* Google [https://cloud.google.com/products/agentspace Agentspace]&lt;br /&gt;
* [https://try.flowith.io/ Flowith]&lt;br /&gt;
&lt;br /&gt;
===Multi-agent Handoff/Collaboration===&lt;br /&gt;
* [https://www.maskara.ai/ Maskara AI]&lt;br /&gt;
&lt;br /&gt;
===Spreadsheet===&lt;br /&gt;
* [https://www.v7labs.com/go V7 Go]&lt;br /&gt;
* [https://ottogrid.ai/ Otto Grid]&lt;br /&gt;
* [https://www.paradigmai.com/ Paradigm]&lt;br /&gt;
* [https://www.superworker.ai/ Superworker AI]&lt;br /&gt;
* [https://www.genspark.ai/ Genspark]&lt;br /&gt;
&lt;br /&gt;
==Cloud solutions==&lt;br /&gt;
* [https://numbersstation.ai/ Numbers Station] [https://numbersstation.ai/introducing-meadow-llm-agents-for-data-tasks/ Meadow]: agentic framework for data workflows ([https://github.com/NumbersStationAI/meadow code]).&lt;br /&gt;
* [https://www.crewai.com/ CrewAI] says they provide multi-agent automations ([https://github.com/joaomdmoura/crewAI code]).&lt;br /&gt;
* [https://www.langchain.com/ LangChain] introduced [https://www.langchain.com/langgraph?ref=blog.langchain.dev LangGraph] to help build agents, and [https://blog.langchain.dev/langgraph-cloud/ LangGraph Cloud] as a service for running those agents.&lt;br /&gt;
** [https://x.com/LangChainAI/status/1819052975295270949 LangGraph Studio] is an IDE for agent workflows&lt;br /&gt;
* [https://c3.ai/ C3 AI] enterprise platform&lt;br /&gt;
* [https://www.deepset.ai/ Deepset AI] [https://haystack.deepset.ai/ Haystack] ([https://docs.haystack.deepset.ai/v1.22/docs/agent docs], [https://github.com/deepset-ai/haystack code])&lt;br /&gt;
&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* Google [https://go.googlesource.com/oscar/+/refs/heads/master/README.md Project Oscar]&lt;br /&gt;
** Agent: Gaby (for &amp;quot;Go AI bot&amp;quot;) ([https://go.googlesource.com/oscar/+/refs/heads/master/internal/gaby code], [https://pkg.go.dev/golang.org/x/oscar/internal/gaby documentation]) helps with issue tracking.&lt;br /&gt;
* [https://github.com/alexfazio/OpenPlexity-Pages OpenPlexity-Pages]: Data-aggregator implementation (like [https://www.perplexity.ai/ Perplexity]) based on [https://www.crewai.com/ CrewAI]&lt;br /&gt;
&lt;br /&gt;
=Optimization=&lt;br /&gt;
===Reviews===&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11936 A Survey of Mathematical Reasoning in the Era of Multimodal Large Language Model: Benchmark, Method &amp;amp; Challenges]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16416 Survey on Evaluation of LLM-based Agents]&lt;br /&gt;
&lt;br /&gt;
===Metrics, Benchmarks===&lt;br /&gt;
See also: [[AI benchmarks]]&lt;br /&gt;
* 2019-11: [https://arxiv.org/abs/1911.01547 On the Measure of Intelligence]&lt;br /&gt;
* 2022-06: [https://arxiv.org/abs/2206.10498 PlanBench: An Extensible Benchmark for Evaluating Large Language Models on Planning and Reasoning about Change]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?] (challenging Corr2Cause task)&lt;br /&gt;
* 2024-01: [https://microsoft.github.io/autogen/0.2/blog/2024/01/25/AutoGenBench/ AutoGenBench -- A Tool for Measuring and Evaluating AutoGen Agents]&lt;br /&gt;
* 2024-04: AutoRace ([https://github.com/maitrix-org/llm-reasoners code]): [https://arxiv.org/abs/2404.05221 LLM Reasoners: New Evaluation, Library, and Analysis of Step-by-Step Reasoning with Large Language Models]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07972 OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments] ([https://os-world.github.io/ github])&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01502 AI Agents That Matter]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.11363 CORE-Bench: Fostering the Credibility of Published Research Through a Computational Reproducibility Agent Benchmark] ([https://agent-evals-core-leaderboard.hf.space/ leaderboard])&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.13373 LLMs Still Can&amp;#039;t Plan; Can LRMs? A Preliminary Evaluation of OpenAI&amp;#039;s o1 on PlanBench]&lt;br /&gt;
* 2024-09: [https://www.arxiv.org/abs/2409.19924 On The Planning Abilities of OpenAI&amp;#039;s o1 Models: Feasibility, Optimality, and Generalizability]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.07095 MLE-bench: Evaluating Machine Learning Agents on Machine Learning Engineering]&lt;br /&gt;
* 2024-10: WorFBench: [https://arxiv.org/abs/2410.07869 Benchmarking Agentic Workflow Generation]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12851 VibeCheck: Discover and Quantify Qualitative Differences in Large Language Models]&lt;br /&gt;
* 2024-10: SimpleAQ: [https://cdn.openai.com/papers/simpleqa.pdf Measuring short-form factuality in large language models] ([https://openai.com/index/introducing-simpleqa/ announcement], [https://github.com/openai/simple-evals code])&lt;br /&gt;
* 2024-11: [https://metr.org/AI_R_D_Evaluation_Report.pdf RE-Bench: Evaluating frontier AI R&amp;amp;D capabilities of language model agents against human experts] ([https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/ blog], [https://github.com/METR/ai-rd-tasks/tree/main code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.10323 The Dawn of GUI Agent: A Preliminary Case Study with Claude 3.5 Computer Use] ([https://github.com/showlab/computer_use_ootb code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.13543 BALROG: Benchmarking Agentic LLM and VLM Reasoning On Games]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14161 TheAgentCompany: Benchmarking LLM Agents on Consequential Real World Tasks] ([https://github.com/TheAgentCompany/TheAgentCompany code], [https://the-agent-company.com/ project], [https://the-agent-company.com/#/leaderboard leaderboard])&lt;br /&gt;
* 2025-01: [https://codeelo-bench.github.io/ CodeElo: Benchmarking Competition-level Code Generation of LLMs with Human-comparable Elo Ratings] ([https://arxiv.org/abs/2501.01257 preprint], [https://codeelo-bench.github.io/#leaderboard-table leaderboard])&lt;br /&gt;
* 2025-02: [https://static.scale.com/uploads/654197dc94d34f66c0f5184e/EnigmaEval%20v4.pdf ENIGMAEVAL:A Benchmark of Long Multimodal Reasoning Challenges] ([https://scale.com/leaderboard/enigma_eval leaderboard])&lt;br /&gt;
* 2025-02: [https://sites.google.com/view/mlgym MLGym: A New Framework and Benchmark for Advancing AI Research Agents] ([https://arxiv.org/abs/2502.14499 paper], [https://github.com/facebookresearch/MLGym code])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.18356 WebGames: Challenging General-Purpose Web-Browsing AI Agents]&lt;br /&gt;
* 2025-03: ColBench: [https://arxiv.org/abs/2503.15478 SWEET-RL: Training Multi-Turn LLM Agents on Collaborative Reasoning Tasks]&lt;br /&gt;
* 2025-04 OpenAI [https://openai.com/index/browsecomp/ BrowseComp: a benchmark for browsing agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.11844 Evaluating the Goal-Directedness of Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Evaluation Schemes===&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.10424 LLM-as-an-Interviewer: Beyond Static Testing Through Dynamic LLM Evaluation]&lt;br /&gt;
* 2025-01: [https://github.com/marquisdepolis/LLMRank LLMRank (&amp;quot;SlopRank&amp;quot;)]: LLMs evaluate each other, allowing top model (for a given prompt/problem) to be inferred from a large number of recommendations.&lt;br /&gt;
&lt;br /&gt;
===Multi-agent===&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.10270 Cultural Evolution of Cooperation among LLM Agents]&lt;br /&gt;
* [https://github.com/lechmazur/step_game/ Multi-Agent Step Race Benchmark: Assessing LLM Collaboration and Deception Under Pressure]&lt;br /&gt;
&lt;br /&gt;
===Agent Challenges===&lt;br /&gt;
* [https://github.com/aidanmclaughlin/Aidan-Bench Aidan-Bench]: Test creativity by having a particular LLM generate long sequence of outputs (meant to be different), and measuring how long it can go before duplications appear.&lt;br /&gt;
** NeurIPS 2024 paper/poster: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions]&lt;br /&gt;
* [https://x.com/paul_cal/status/1850262678712856764 Pictionary]: LLM suggests prompt, multiple LLMs generate outputs, LLM judges; allows raking of the generation abilities.&lt;br /&gt;
* [https://mcbench.ai/ MC-bench]: Request LLMs to build an elaborate structure in Minecraft; outputs can be A/B tested by human judges ([https://github.com/mc-bench/orchestrator code]).&lt;br /&gt;
&lt;br /&gt;
===Automated Improvement===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14228 EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.18532 Symbolic Learning Enables Self-Evolving Agents]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08435 Automated Design of Agentic Systems] ([https://github.com/ShengranHu/ADAS ADAS code])&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.02666 Self-Taught Evaluators]: Iterative self-improvement through generation of synthetic data and evaluation&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.22954 Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents] ([https://github.com/jennyzzt/dgm code], [https://sakana.ai/dgm/ project])&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19461 Hyperagents]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[Science Agents]]&lt;br /&gt;
* [[Increasing AI Intelligence]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI understanding]]&lt;br /&gt;
* [[Robots]]&lt;br /&gt;
* [[Exocortex]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8774</id>
		<title>AI understanding</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8774"/>
		<updated>2026-04-02T17:48:25Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Psychology */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Interpretability=&lt;br /&gt;
* 2017-01: [https://arxiv.org/abs/1704.01444 Learning to Generate Reviews and Discovering Sentiment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11639 Neural Interpretable Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Concepts==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.20938 Towards Understanding the Nature of Attention with Low-Rank Sparse Decomposition] ([https://github.com/OpenMOSS/Lorsa code])&lt;br /&gt;
* 2025-08: [https://transformer-circuits.pub/2025/attention-qk/index.html Tracing Attention Computation Through Feature Interactions]&lt;br /&gt;
&lt;br /&gt;
==Mechanistic Interpretability==&lt;br /&gt;
* 2020-03: OpenAI: [https://distill.pub/2020/circuits/zoom-in/ Zoom In: An Introduction to Circuits]&lt;br /&gt;
* 2021-12: Anthropic: [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2211.00593 Interpretability in the Wild: a Circuit for Indirect Object Identification in GPT-2 Small]&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-07: Anthropic: [https://transformer-circuits.pub/2024/july-update/index.html Circuits Update]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.14926 Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-based Parameter Decomposition] ([https://www.alignmentforum.org/posts/EPefYWjuHNcNH4C7E/attribution-based-parameter-decomposition blog post])&lt;br /&gt;
* 2025-01: Review: [https://arxiv.org/abs/2501.16496 Open Problems in Mechanistic Interpretability]&lt;br /&gt;
* 2025-03: Anthropic: [https://www.anthropic.com/research/tracing-thoughts-language-model Tracing the thoughts of a large language model]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/methods.html Circuit Tracing: Revealing Computational Graphs in Language Models]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/biology.html On the Biology of a Large Language Model]&lt;br /&gt;
* 2025-11: OpenAI: [https://cdn.openai.com/pdf/41df8f28-d4ef-43e9-aed2-823f9393e470/circuit-sparsity-paper.pdf Weight-sparse transformers have interpretable circuits] ([https://openai.com/index/understanding-neural-networks-through-sparse-circuits/ blog])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.13548 Patterning: The Dual of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Semanticity==&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.08600 Sparse Autoencoders Find Highly Interpretable Features in Language Models]&lt;br /&gt;
* Anthropic monosemanticity interpretation of LLM features:&lt;br /&gt;
** 2023-10: [https://transformer-circuits.pub/2023/monosemantic-features/index.html Towards Monosemanticity: Decomposing Language Models With Dictionary Learning]&lt;br /&gt;
** 2024-05: [https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet]&lt;br /&gt;
* 2024-06: OpenaAI: [https://arxiv.org/abs/2406.04093 Scaling and evaluating sparse autoencoders]&lt;br /&gt;
* 2024-08: [https://www.alignmentforum.org/posts/TMAmHh4DdMr4nCSr5/showing-sae-latents-are-not-atomic-using-meta-saes Showing SAE Latents Are Not Atomic Using Meta-SAEs] ([https://metasae.streamlit.app/?page=Feature+Explorer&amp;amp;feature=11329 demo])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08201 Efficient Dictionary Learning with Switch Sparse Autoencoders] ([https://github.com/amudide/switch_sae code]) More efficient SAE generation&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.14670 Decomposing The Dark Matter of Sparse Autoencoders] ([https://github.com/JoshEngels/SAE-Dark-Matter code]) Shows that SAE errors are predictable&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13928 Automatically Interpreting Millions of Features in Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.21331 Beyond Interpretability: The Gains of Feature Monosemanticity on Model Robustness]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.04139 Monet: Mixture of Monosemantic Experts for Transformers]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/zbebxYCqsryPALh8C/matryoshka-sparse-autoencoders Matryoshka Sparse Autoencoders]&lt;br /&gt;
* 2024-12: [https://www.alignmentforum.org/posts/rKM9b6B2LqwSB5ToN/learning-multi-level-features-with-matryoshka-saes Learning Multi-Level Features with Matryoshka SAEs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19406 Low-Rank Adapting Models for Sparse Autoencoders]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.03714 Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.00177 Steering Large Language Model Activations in Sparse Spaces]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01776 Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01824 From superposition to sparse codes: interpretable representations in neural networks]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18878 I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20063 SAEs Are Good for Steering -- If You Select the Right Features]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15679 Dense SAE Latents Are Features, Not Bugs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20790 Stochastic Parameter Decomposition] ([https://github.com/goodfire-ai/spd code], [https://www.goodfire.ai/papers/stochastic-param-decomp blog])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
&lt;br /&gt;
===Counter-Results===&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.12016 Towards falsifiable interpretability research]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16615 Sparse Autoencoders Trained on the Same Data Learn Different Features]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17148 AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17727 Sparse Autoencoders Can Interpret Randomly Initialized Transformers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.04878 Sparse Autoencoders Do Not Find Canonical Units of Analysis]&lt;br /&gt;
* 2025-03: [https://www.alignmentforum.org/posts/4uXCAJNuPKtKBsi28/ Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research]&lt;br /&gt;
&lt;br /&gt;
==Meta-cognition==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.15674 Activation Oracles: Training and Evaluating LLMs as General-Purpose Activation Explainers]&lt;br /&gt;
&lt;br /&gt;
==Coding Models==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sparse Auto Encoders&amp;#039;&amp;#039;&amp;#039;: See Semanticity.&lt;br /&gt;
* [https://github.com/saprmarks/dictionary_learning dictionary_learning]&lt;br /&gt;
* [https://transformer-circuits.pub/2024/jan-update/index.html#predict-future Predicting Future Activations]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11944 Transcoders Find Interpretable LLM Feature Circuits]&lt;br /&gt;
* 2024-10: [https://transformer-circuits.pub/2024/crosscoders/index.html Sparse Crosscoders for Cross-Layer Features and Model Diffing]&lt;br /&gt;
&lt;br /&gt;
==Reward Functions==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12491 Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL]&lt;br /&gt;
&lt;br /&gt;
==Symbolic and Notation==&lt;br /&gt;
* [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* [https://www.arxiv.org/abs/2407.09468 Beyond Euclid: An Illustrated Guide to Modern Machine Learning with Geometric, Topological, and Algebraic Structures]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02423 On the Anatomy of Attention]: Introduces category-theoretic diagrammatic formalism for DL architectures&lt;br /&gt;
* 2024-11: [https://x.com/vtabbott_/status/1860268276569506250 diagrams to represent algorithms]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03317 FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awareness]&lt;br /&gt;
&lt;br /&gt;
==Mathematical==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.13762 Unveiling the Hidden Structure of Self-Attention via Kernel Principal Component Analysis]&lt;br /&gt;
&lt;br /&gt;
==Geometric==&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.03658 The Linear Representation Hypothesis and the Geometry of Large Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.01506 The Geometry of Categorical and Hierarchical Concepts in Large Language Models]&lt;br /&gt;
** Natural hierarchies of concepts---which occur throughout natural language and especially in scientific ontologies---are represented in the model&amp;#039;s internal vectorial space as polytopes that can be decomposed into simplexes of mutually-exclusive categories.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02678 Reasoning in Large Language Models: A Geometric Perspective]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.17592 Deep Manifold Part 1: Anatomy of Neural Network Manifold]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.19750 The Geometry of Concepts: Sparse Autoencoder Feature Structure]&lt;br /&gt;
** Tegmark et al. report multi-scale structure: 1) “atomic” small-scale, 2) “brain” intermediate-scale, and 3) “galaxy” large-scale&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.08009 The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.09782 The Geometry of Reasoning: Flowing Logics in Representation Space]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/linebreaks/index.html When Models Manipulate Manifolds: The Geometry of a Counting Task]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.26745 Deep sequence models tend to memorize geometrically; it is unclear why]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
==Topography==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.06002 The Molecular Structure of Thought: Mapping the Topology of Long Chain-of-Thought Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2023-07Jul: [https://arxiv.org/abs/2307.13702 Measuring Faithfulness in Chain-of-Thought Reasoning] [https://x.com/davidad/status/1839641113432305790 roughly] proves that sufficiently large models do not generate CoT that actually captures their internal reasoning)&lt;br /&gt;
&lt;br /&gt;
[[Image:GYe31yXXQAABwaZ.jpeg|300px]]&lt;br /&gt;
&lt;br /&gt;
=Heuristic Understanding=&lt;br /&gt;
* 2022-09: Janus: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators]&lt;br /&gt;
&lt;br /&gt;
==Emergent Internal Model Building==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.15936 A Theory for Emergence of Complex Skills in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19370v1 Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01622 General agents contain world models]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20328 Video models are zero-shot learners and reasoners]&lt;br /&gt;
&lt;br /&gt;
===Semantic Directions===&lt;br /&gt;
Directions, e.g.: f(king)-f(man)+f(woman)=f(queen) or f(sushi)-f(Japan)+f(Italy)=f(pizza)&lt;br /&gt;
* [https://arxiv.org/abs/1301.3781 Efficient Estimation of Word Representations in Vector Space]&lt;br /&gt;
* [https://aclanthology.org/N13-1090/ Linguistic Regularities in Continuous Space Word Representations]&lt;br /&gt;
* [https://aclanthology.org/C16-1332 Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen]&lt;br /&gt;
* [https://aclanthology.org/D14-1162/ Glove: Global vectors for word representation]&lt;br /&gt;
* [https://doi.org/10.1109/BigData.2015.7364114 Using Word2Vec to process big text data]&lt;br /&gt;
* [https://arxiv.org/abs/2310.06824 The geometry of truth: Emergent linear structure in large language model representations of true/false datasets] (true/false)&lt;br /&gt;
* [https://arxiv.org/abs/2403.10381 Monotonic Representation of Numeric Properties in Language Models] (numeric directions)&lt;br /&gt;
Task vectors:&lt;br /&gt;
* [https://arxiv.org/abs/2310.15213 Function Vectors in Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.15916 In-context learning creates task vectors]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/5FGXmJ3wqgGRcbyH7/extracting-sae-task-features-for-in-context-learning Extracting sae task features for in-context learning]&lt;br /&gt;
* [https://arxiv.org/abs/2412.12276 Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers]&lt;br /&gt;
Reasoning:&lt;br /&gt;
* [https://openreview.net/forum?id=OwhVWNOBcz Understanding Reasoning in Thinking Language Models via Steering Vectors]&lt;br /&gt;
&lt;br /&gt;
===Feature Geometry Reproduces Problem-space===&lt;br /&gt;
* [https://arxiv.org/abs/2210.13382 Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2309.00941 Emergent linear representations in world models of self-supervised sequence models] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* [https://doi.org/10.1038/s41562-023-01659-w Emergent analogical reasoning in large language models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.02207 Language Models Represent Space and Time] (Maps of world, US)&lt;br /&gt;
* [https://arxiv.org/abs/2405.14860 Not All Language Model Features Are Linear] (Days of week form ring, etc.)&lt;br /&gt;
* [https://arxiv.org/abs/2406.03689 Evaluating the World Model Implicit in a Generative Model] (Map of Manhattan)&lt;br /&gt;
* [https://iopscience.iop.org/article/10.1088/1748-9326/ad2891 Reliable precipitation nowcasting using probabilistic diffusion models]. Generation of precipitation map imagery is predictive of actual future weather; implies model is learning scientifically-relevant modeling.&lt;br /&gt;
* [https://arxiv.org/abs/2405.07987 The Platonic Representation Hypothesis]: Different models (including across modalities) are converging to a consistent world model.&lt;br /&gt;
* [https://arxiv.org/abs/2501.00070 ICLR: In-Context Learning of Representations]&lt;br /&gt;
* [https://arxiv.org/abs/2502.00873 Language Models Use Trigonometry to Do Addition]: Numbers arranged in helix to enable addition&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
&lt;br /&gt;
===Capturing Physics===&lt;br /&gt;
* 2020-09: [https://arxiv.org/abs/2009.08292 Learning to Identify Physical Parameters from Video Using Differentiable Physics]&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.00419 Self-Supervised Learning for Videos: A Survey]&lt;br /&gt;
* 2025-02: Fair at Meta: [https://arxiv.org/abs/2502.11831 Intuitive physics understanding emerges from self-supervised pretraining on natural videos]&lt;br /&gt;
&lt;br /&gt;
===Theory of Mind===&lt;br /&gt;
* [https://arxiv.org/abs/2302.02083 Evaluating Large Language Models in Theory of Mind Tasks]&lt;br /&gt;
* [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-01: [https://www.arxiv.org/abs/2501.09038 Do generative video models learn physical principles from watching videos?] ([https://physics-iq.github.io/ project], [https://github.com/google-deepmind/physics-IQ-benchmark code])&lt;br /&gt;
* 2025-06: [https://machinelearning.apple.com/research/illusion-of-thinking The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21521 Potemkin Understanding in Large Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21876 Do Vision-Language Models Have Internal World Models? Towards an Atomic Evaluation]&lt;br /&gt;
&lt;br /&gt;
==Information Processing==&lt;br /&gt;
* 2019-03: [https://arxiv.org/abs/1903.05789 Diagnosing and Enhancing VAE Models]&lt;br /&gt;
* 2021-03: [https://arxiv.org/abs/2103.05247 Pretrained Transformers as Universal Computation Engines]&lt;br /&gt;
* 2022-10: [https://arxiv.org/abs/2210.08344 How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03843 Why think step by step? Reasoning emerges from the locality of experience]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.04444 What&amp;#039;s the Magic Word? A Control Theory of LLM Prompting]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12875 Chain of Thought Empowers Transformers to Solve Inherently Serial Problems]: Proves that transformers can solve any problem, if they can generate sufficient intermediate tokens&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.20311 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process]&lt;br /&gt;
** Models learning reasoning skills (they are not merely memorizing solution templates). They can mentally generate simple short plans (like humans).&lt;br /&gt;
** When presented facts, models develop internal understanding of what parameters (recursively) depend on each other. This occurs even before an explicit question is asked (i.e. before the task is defined). This appears to be different from human reasoning.&lt;br /&gt;
** Model depth matters for reasoning. This cannot be mitigated by chain-of-thought prompting (which allow models to develop and then execute plans) since even a single CoT step may require deep, multi-step reasoning/planning.&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.01992 Ask, and it shall be given: Turing completeness of prompting]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08775 Layers at Similar Depths Generate Similar Activations Across LLM Architectures]&lt;br /&gt;
&lt;br /&gt;
===Generalization===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]&lt;br /&gt;
&lt;br /&gt;
===Grokking===&lt;br /&gt;
* 2022-01: [https://arxiv.org/abs/2201.02177 Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets]&lt;br /&gt;
* 2022-05: [https://arxiv.org/abs/2205.10343 Towards Understanding Grokking: An Effective Theory of Representation Learning]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.10463 Critical Data Size of Language Models from a Grokking Perspective]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15175 Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18624 How to explain grokking]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.21519 Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
===Tests of Resilience to Dropouts/etc.===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15390 Explorations of Self-Repair in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15786 What Matters in Transformers? Not All Attention is Needed]&lt;br /&gt;
** Removing entire transformer blocks leads to significant performance degradation&lt;br /&gt;
** Removing MLP layers results in significant performance degradation&lt;br /&gt;
** Removing attention layers causes almost no performance degradation&lt;br /&gt;
** E.g. half of attention layers are deleted (48% speed-up), leads to only 2.4% decrease in the benchmarks&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19384 The Remarkable Robustness of LLMs: Stages of Inference?]&lt;br /&gt;
** They intentionally break the network (swapping layers), yet it continues to work remarkably well. This suggests LLMs are quite robust, and allows them to identify different stages in processing.&lt;br /&gt;
** They also use these interventions to infer what different layers are doing. They break apart the LLM transformer layers into four stages:&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Detokenization:&amp;#039;&amp;#039;&amp;#039; Raw tokens are converted into meaningful entities that take into account local context (especially using nearby tokens).&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Feature engineering:&amp;#039;&amp;#039;&amp;#039; Features are progressively refined. Factual knowledge is leveraged.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Prediction ensembling:&amp;#039;&amp;#039;&amp;#039; Predictions (for the ultimately-selected next-token) emerge. A sort of consensus voting is used, with “prediction neurons” and &amp;quot;suppression neurons&amp;quot; playing a major role in upvoting/downvoting.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Residual sharpening:&amp;#039;&amp;#039;&amp;#039; The semantic representations are collapsed into specific next-token predictions. There is a strong emphasis on suppression neurons eliminating options. The confidence is calibrated.&lt;br /&gt;
** This structure can be thought of as two halves (being roughly dual to each other): the first half broadens (goes from distinct tokens to a rich/elaborate concept-space) and the second half collapses (goes from rich concepts to concrete token predictions).&lt;br /&gt;
&lt;br /&gt;
==Semantic Vectors==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11717 Refusal in Language Models Is Mediated by a Single Direction]&lt;br /&gt;
* 2025-02: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs] ([https://x.com/OwainEvans_UK/status/1894436637054214509 demonstrates] [https://x.com/ESYudkowsky/status/1894453376215388644 entangling] of concepts into a single preference vector)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03666 Analogical Reasoning Inside Large Language Models: Concept Vectors and the Limits of Abstraction]&lt;br /&gt;
&lt;br /&gt;
==Other==&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00247 Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting &amp;amp; Beyond]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04282 Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding] ([https://github.com/SalesforceAIResearch/LaTRO code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.12580 Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models]: LLMs learn reasoning by extracting procedures from training data, not by memorizing specific answers&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15862 LLMs Do Not Think Step-by-step In Implicit Reasoning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
&lt;br /&gt;
===Scaling Laws===&lt;br /&gt;
* 1993: [https://proceedings.neurips.cc/paper/1993/file/1aa48fc4880bb0c9b8a3bf979d3b917e-Paper.pdf Learning Curves: Asymptotic Values and Rate of Convergence]&lt;br /&gt;
* 2017-12: [https://arxiv.org/abs/1712.00409 Deep Learning Scaling is Predictable, Empirically] (Baidu)&lt;br /&gt;
* 2019-03: [http://www.incompleteideas.net/IncIdeas/BitterLesson.html The Bitter Lesson] (Rich Sutton)&lt;br /&gt;
* 2020-01: [https://arxiv.org/abs/2001.08361 Scaling Laws for Neural Language Models] (OpenAI)&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.14701 Scaling Laws for Autoregressive Generative Modeling] (OpenAI)&lt;br /&gt;
* 2020-05: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis] (Gwern)&lt;br /&gt;
* 2021-08: [https://arxiv.org/abs/2108.07686 Scaling Laws for Deep Learning]&lt;br /&gt;
* 2021-02: [https://arxiv.org/abs/2102.06701 Explaining Neural Scaling Laws] (Google DeepMind)&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.15556 Training Compute-Optimal Large Language Models] (Chinchilla, Google DeepMind)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.04715 Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.07951 Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models]&lt;br /&gt;
* 2025-05: [https://brendel-group.github.io/llm-line/ LLMs on the Line: Data Determines Loss-To-Loss Scaling Laws]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13786 The Art of Scaling Reinforcement Learning Compute for LLMs]&lt;br /&gt;
&lt;br /&gt;
=Information Processing/Storage=&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.10689 A Theory of Usable Information Under Computational Constraints]&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.00008 Why is AI hard and Physics simple?]&lt;br /&gt;
* 2021-06: [https://arxiv.org/abs/2106.06981 Thinking Like Transformers]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.00948 Large Linguistic Models: Investigating LLMs&amp;#039; metalinguistic abilities]&lt;br /&gt;
* &amp;quot;A transformer&amp;#039;s depth affects its reasoning capabilities, whilst model size affects its knowledge capacity&amp;quot; ([https://x.com/danielhanchen/status/1835684061475655967 c.f.])&lt;br /&gt;
** 2024-02: [https://arxiv.org/abs/2402.14905 MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases]&lt;br /&gt;
** 2024-04: [https://arxiv.org/abs/2404.08819 The Illusion of State in State-Space Models] (figure 3)&lt;br /&gt;
** 2024-08: [https://arxiv.org/abs/2408.00118 Gemma 2: Improving Open Language Models at a Practical Size] (table 9)&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.10482 Schrodinger&amp;#039;s Memory: Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2407.01687 Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning]. CoT involves both memorization and (probabilitic) reasoning&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.16679 Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?]&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.03961 A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.22471 The Bayesian Geometry of Transformer Attention]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03220 From Entropy to Epiplexity: Rethinking Information for Computationally Bounded Intelligence]&lt;br /&gt;
&lt;br /&gt;
==Statistics/Math==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.05465 The emergence of clusters in self-attention dynamics]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.10794 A mathematical perspective on Transformers]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.12034 Understanding Transformers via N-gram Statistics]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.06833 Dynamic metastability in the self-attention model]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04551 Measure-to-measure interpolation using Transformers]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14697 Quantitative Clustering in Mean-Field Transformer Models]&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
===For numbers/math===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.14903 Tokenization counts: the impact of tokenization on arithmetic in frontier LLMs]: L2R vs. R2L yields different performance on math&lt;br /&gt;
&lt;br /&gt;
==Data Storage==&lt;br /&gt;
* 1988-09: [https://www.sciencedirect.com/science/article/pii/0885064X88900209 On the capabilities of multilayer perceptrons]&lt;br /&gt;
* 2006-12: [https://ieeexplore.ieee.org/document/4038449 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition] (single-layer perceptron stores &amp;gt;2 bits/parameter; MLP ~ 2*N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; bits w/ N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; params)&lt;br /&gt;
* 2016-11: [https://arxiv.org/abs/1611.09913 Capacity and Trainability in Recurrent Neural Networks] (5 bits/param)&lt;br /&gt;
* 2018-02: [https://arxiv.org/abs/1802.08232 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks]&lt;br /&gt;
* 2019-05: [https://ieeexplore.ieee.org/document/8682462 Memorization Capacity of Deep Neural Networks under Parameter Quantization]&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.08910 How Much Knowledge Can You Pack Into the Parameters of a Language Model?]&lt;br /&gt;
* 2020-08: [https://arxiv.org/abs/2008.09036 Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries] (capacity scales linearly with parameters; more training samples leads to less memorization)&lt;br /&gt;
* 2020-12: [https://arxiv.org/abs/2012.06421 When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05405 Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws] (2 bits/param)&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15720 Scaling Laws for Fact Memorization of Large Language Models] (1T params needed to memorize Wikipedia)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24832 How much do language models memorize?] (3.6 bits/parameter)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01855 Trade-offs in Data Memorization via Strong Data Processing Inequalities]&lt;br /&gt;
&lt;br /&gt;
===Reverse-Engineering Training Data===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.10364 Can We Infer Confidential Properties of Training Data from LLMs?]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15553 Approximating Language Model Training Data from Weights]&lt;br /&gt;
&lt;br /&gt;
===Compression===&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.09410 Less is More: Parameter-Free Text Classification with Gzip]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.04050 LLMZip: Lossless Text Compression using Large Language Models]&lt;br /&gt;
* 2023-07: [https://aclanthology.org/2023.findings-acl.426/ “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.10668 Language Modeling Is Compression]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07550 An Image is Worth 32 Tokens for Reconstruction and Generation]&lt;br /&gt;
&lt;br /&gt;
==Learning/Training==&lt;br /&gt;
* 2018-03: [https://arxiv.org/abs/1803.03635 The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks]: Sparse neural networks are optimal, but it is difficult to identify the right architecture and train it. Deep learning typically consists of training a dense neural network, which makes it easier to learn an internal sparse circuit optimal to a particular problem.&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11521 On the Ability of Deep Networks to Learn Symmetries from Data: A Neural Kernel Theory]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.12391 Physics of Skill Learning]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24864 ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Cross-modal knowledge transfer===&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.07519 Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.07358 Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]: CLIP learns richer set of aggregated representations (e.g. for a culture or country), vs. a vision-only model.&lt;br /&gt;
&lt;br /&gt;
==Hidden State==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06258 Emergent Response Planning in LLM]: They show that the latent representation contains information beyond that needed for the next token (i.e. the model learns to &amp;quot;plan ahead&amp;quot; and encode information relevant to future tokens)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02854 (How) Do Language Models Track State?]&lt;br /&gt;
===Convergent Representation===&lt;br /&gt;
* 2015-11: [https://arxiv.org/abs/1511.07543 Convergent Learning: Do different neural networks learn the same representations?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of Embeddings]: Evidence for [https://x.com/jxmnop/status/1925224620166128039 The Strong Platonic Representation Hypothesis]; models converge to a single consensus reality&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Function Approximation==&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]: can learn linear functions (equivalent to least-squares estimator)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09066 Teaching Algorithmic Reasoning via In-context Learning]: Simple arithmetic &lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models] ([https://github.com/ekinakyurek/google-research/tree/master/incontext code]): can learn linear regression&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.00297 Transformers learn to implement preconditioned gradient descent for in-context learning]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.03576 One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.02893 ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20545 SoS1: O1 and R1-Like Reasoning LLMs are Sum-of-Square Solvers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.21212 Transformers Learn to Implement Multi-step Gradient Descent with Chain of Thought]&lt;br /&gt;
&lt;br /&gt;
=Physics Based=&lt;br /&gt;
* 2014-01: [https://arxiv.org/abs/1401.1219 Consciousness as a State of Matter]&lt;br /&gt;
* 2016-08: [https://arxiv.org/abs/1608.08225 Why does deep and cheap learning work so well?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.23489 SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training]&lt;br /&gt;
* 2025-12: [https://www.pnas.org/doi/full/10.1073/pnas.2523012122 Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning]&lt;br /&gt;
&lt;br /&gt;
=Failure Modes=&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?]: Poor causal inference&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.12288 The Reversal Curse: LLMs trained on &amp;quot;A is B&amp;quot; fail to learn &amp;quot;B is A&amp;quot;]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.13638 Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve] (biases towards &amp;quot;common&amp;quot; numbers, in-context CoT can reduce performance by incorrectly priming, etc.)&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.16093 Visual cognition in multimodal large language models] (models lack human-like visual understanding)&lt;br /&gt;
&lt;br /&gt;
==Adversarial==&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03507 Solving adversarial examples requires solving exponential misalignment]&lt;br /&gt;
&lt;br /&gt;
==Fracture Representation==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.11581 Questioning Representational Optimism in Deep Learning: The Fractured Entangled Representation Hypothesis] ([https://github.com/akarshkumar0101/fer code])&lt;br /&gt;
&lt;br /&gt;
==Jagged Frontier==&lt;br /&gt;
* 2023-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03211 How Does Quantization Affect Multilingual LLMs?]: Quantization degrades different languages by differing amounts&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061v1 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]: Scaling laws are skill-dependent&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.18212 A Definition of AGI]&lt;br /&gt;
&lt;br /&gt;
===See also===&lt;br /&gt;
* [[AI_understanding|AI Understanding]] &amp;gt; [[AI_understanding#Psychology|Psychology]] &amp;gt; [[AI_understanding#LLM_personalities|LLM personalities]]&lt;br /&gt;
* [[AI tricks]] &amp;gt; [[AI_tricks#Prompt_Engineering|Prompt Engineering]] &amp;gt; [[AI_tricks#Brittleness|Brittleness]]&lt;br /&gt;
&lt;br /&gt;
===Conversely (AI models converge)===&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.05117 The Universal Weight Subspace Hypothesis]&lt;br /&gt;
* 2026-01: [https://avikrishna.substack.com/p/eliciting-frontier-model-character Eliciting Frontier Model Character Training: A study of personality convergence across language models]&lt;br /&gt;
&lt;br /&gt;
==Model Collapse==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.12202 Nepotistically Trained Generative-AI Models Collapse]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05280 On the Limits of Self-Improving in LLMs and Why AGI, ASI and the Singularity Are Not Near Without Symbolic Model Synthesis]&lt;br /&gt;
&lt;br /&gt;
===Analysis===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
&lt;br /&gt;
===Mitigation===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08117 Convergence Dynamics and Stabilization Strategies of Co-Evolving Generative Models]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.11111 Inducing anxiety in large language models can induce bias]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17117 From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning]&lt;br /&gt;
* 2025-07: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5357179 Call Me A Jerk: Persuading AI to Comply with Objectionable Requests]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.06047 &amp;quot;They parted illusions -- they parted disclaim marinade&amp;quot;: Misalignment as structural fidelity in LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.02606 Gender Dynamics and Homophily in a Social Network of LLM Agents]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
* 2026-04: [https://transformer-circuits.pub/2026/emotions/index.html Emotion concepts and their function in a large language model] ([https://www.anthropic.com/research/emotion-concepts-function blog])&lt;br /&gt;
&lt;br /&gt;
==Persona Simulator Theory==&lt;br /&gt;
* 2022-09: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators] ([https://www.lesswrong.com/users/janus-1?from=post_header janus])&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.findings-emnlp.423/ Language Models as Agent Models]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.00805 Conditioning Predictive Models: Risks and Strategies]&lt;br /&gt;
* 2024-09: [https://www.lesswrong.com/s/qhdHbCJ3PYesL9dde Intuitive Self-Models]&lt;br /&gt;
* 2026-02: [https://alignment.anthropic.com/2026/psm/ The Persona Selection Model: Why AI Assistants might Behave like Humans] (Anthropic, [https://www.anthropic.com/research/persona-selection-model blog])&lt;br /&gt;
&lt;br /&gt;
==Allow LLM to think==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
&lt;br /&gt;
===In-context Learning===&lt;br /&gt;
* 2021-10: [https://arxiv.org/abs/2110.15943 MetaICL: Learning to Learn In Context]&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?]&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16003 Learning without training: The implicit dynamics of in-context learning]&lt;br /&gt;
&lt;br /&gt;
==Reasoning (CoT, etc.)==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18009 Large Language Models Think Too Fast To Explore Effectively]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18585 Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.08156 Are DeepSeek R1 And Other Reasoning Models More Faithful?]: reasoning models can provide faithful explanations for why their reasoning is correct&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08679 Chain-of-Thought Reasoning In The Wild Is Not Always Faithful]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.04022 Rethinking Reflection in Pre-Training]: pre-training alone already provides some amount of reflection/reasoning&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2501.18858 BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning]&lt;br /&gt;
&lt;br /&gt;
===Pathfinding===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08152 DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09284 Tree-OPO: Off-policy Monte Carlo Tree-Guided Advantage Optimization for Multistep Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06160v1 Reverse-Engineered Reasoning for Open-Ended Generation]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06941 The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2508.01191 Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens]&lt;br /&gt;
&lt;br /&gt;
==Self-Awareness and Self-Recognition and Introspection==&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.05221 Language Models (Mostly) Know What They Know]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13076 LLM Evaluators Recognize and Favor Their Own Generations]&lt;br /&gt;
* 2024-09: [https://situational-awareness-dataset.org/ Me, Myself and AI: The Situational Awareness Dataset for LLMs]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* 2024-12: [https://theaidigest.org/self-awareness AIs are becoming more self-aware. Here&amp;#039;s why that matters]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-04: [https://x.com/Josikinz/status/1907923319866716629 LLMs can guess which comic strip was generated by themselves (vs. other LLM)]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/introspection/index.html Emergent Introspective Awareness in Large Language Models] (Anthropic, [https://www.anthropic.com/research/introspection blog])&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.24661 Do Large Language Models Know What They Are Capable Of?]&lt;br /&gt;
&lt;br /&gt;
==LLM personalities==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.02618 Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04343 Psychologically Enhanced AI Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10387 The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models]&lt;br /&gt;
&lt;br /&gt;
==Quirks &amp;amp; Biases==&lt;br /&gt;
* 2025-04: [https://www.cambridge.org/core/journals/judgment-and-decision-making/article/artificial-intelligence-and-dichotomania/0421D2310727D73FAB47069FD1620AA1 Artificial intelligence and dichotomania]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.22818 Can Large Language Models Develop Gambling Addiction?]&lt;br /&gt;
&lt;br /&gt;
=Vision Models=&lt;br /&gt;
* 2017-11: Distill: [https://distill.pub/2017/feature-visualization/ Feature Visualization: How neural networks build up their understanding of images]&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.12322 Position, Padding and Predictions: A Deeper Look at Position Information in CNNs]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13181 Perception Encoder: The best visual embeddings are not at the output of the network] ([https://github.com/facebookresearch/perception_models code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8773</id>
		<title>AI understanding</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8773"/>
		<updated>2026-04-02T17:45:07Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Psychology */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Interpretability=&lt;br /&gt;
* 2017-01: [https://arxiv.org/abs/1704.01444 Learning to Generate Reviews and Discovering Sentiment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11639 Neural Interpretable Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Concepts==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.20938 Towards Understanding the Nature of Attention with Low-Rank Sparse Decomposition] ([https://github.com/OpenMOSS/Lorsa code])&lt;br /&gt;
* 2025-08: [https://transformer-circuits.pub/2025/attention-qk/index.html Tracing Attention Computation Through Feature Interactions]&lt;br /&gt;
&lt;br /&gt;
==Mechanistic Interpretability==&lt;br /&gt;
* 2020-03: OpenAI: [https://distill.pub/2020/circuits/zoom-in/ Zoom In: An Introduction to Circuits]&lt;br /&gt;
* 2021-12: Anthropic: [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2211.00593 Interpretability in the Wild: a Circuit for Indirect Object Identification in GPT-2 Small]&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-07: Anthropic: [https://transformer-circuits.pub/2024/july-update/index.html Circuits Update]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.14926 Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-based Parameter Decomposition] ([https://www.alignmentforum.org/posts/EPefYWjuHNcNH4C7E/attribution-based-parameter-decomposition blog post])&lt;br /&gt;
* 2025-01: Review: [https://arxiv.org/abs/2501.16496 Open Problems in Mechanistic Interpretability]&lt;br /&gt;
* 2025-03: Anthropic: [https://www.anthropic.com/research/tracing-thoughts-language-model Tracing the thoughts of a large language model]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/methods.html Circuit Tracing: Revealing Computational Graphs in Language Models]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/biology.html On the Biology of a Large Language Model]&lt;br /&gt;
* 2025-11: OpenAI: [https://cdn.openai.com/pdf/41df8f28-d4ef-43e9-aed2-823f9393e470/circuit-sparsity-paper.pdf Weight-sparse transformers have interpretable circuits] ([https://openai.com/index/understanding-neural-networks-through-sparse-circuits/ blog])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.13548 Patterning: The Dual of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Semanticity==&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.08600 Sparse Autoencoders Find Highly Interpretable Features in Language Models]&lt;br /&gt;
* Anthropic monosemanticity interpretation of LLM features:&lt;br /&gt;
** 2023-10: [https://transformer-circuits.pub/2023/monosemantic-features/index.html Towards Monosemanticity: Decomposing Language Models With Dictionary Learning]&lt;br /&gt;
** 2024-05: [https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet]&lt;br /&gt;
* 2024-06: OpenaAI: [https://arxiv.org/abs/2406.04093 Scaling and evaluating sparse autoencoders]&lt;br /&gt;
* 2024-08: [https://www.alignmentforum.org/posts/TMAmHh4DdMr4nCSr5/showing-sae-latents-are-not-atomic-using-meta-saes Showing SAE Latents Are Not Atomic Using Meta-SAEs] ([https://metasae.streamlit.app/?page=Feature+Explorer&amp;amp;feature=11329 demo])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08201 Efficient Dictionary Learning with Switch Sparse Autoencoders] ([https://github.com/amudide/switch_sae code]) More efficient SAE generation&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.14670 Decomposing The Dark Matter of Sparse Autoencoders] ([https://github.com/JoshEngels/SAE-Dark-Matter code]) Shows that SAE errors are predictable&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13928 Automatically Interpreting Millions of Features in Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.21331 Beyond Interpretability: The Gains of Feature Monosemanticity on Model Robustness]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.04139 Monet: Mixture of Monosemantic Experts for Transformers]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/zbebxYCqsryPALh8C/matryoshka-sparse-autoencoders Matryoshka Sparse Autoencoders]&lt;br /&gt;
* 2024-12: [https://www.alignmentforum.org/posts/rKM9b6B2LqwSB5ToN/learning-multi-level-features-with-matryoshka-saes Learning Multi-Level Features with Matryoshka SAEs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19406 Low-Rank Adapting Models for Sparse Autoencoders]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.03714 Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.00177 Steering Large Language Model Activations in Sparse Spaces]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01776 Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01824 From superposition to sparse codes: interpretable representations in neural networks]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18878 I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20063 SAEs Are Good for Steering -- If You Select the Right Features]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15679 Dense SAE Latents Are Features, Not Bugs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20790 Stochastic Parameter Decomposition] ([https://github.com/goodfire-ai/spd code], [https://www.goodfire.ai/papers/stochastic-param-decomp blog])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
&lt;br /&gt;
===Counter-Results===&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.12016 Towards falsifiable interpretability research]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16615 Sparse Autoencoders Trained on the Same Data Learn Different Features]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17148 AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17727 Sparse Autoencoders Can Interpret Randomly Initialized Transformers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.04878 Sparse Autoencoders Do Not Find Canonical Units of Analysis]&lt;br /&gt;
* 2025-03: [https://www.alignmentforum.org/posts/4uXCAJNuPKtKBsi28/ Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research]&lt;br /&gt;
&lt;br /&gt;
==Meta-cognition==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.15674 Activation Oracles: Training and Evaluating LLMs as General-Purpose Activation Explainers]&lt;br /&gt;
&lt;br /&gt;
==Coding Models==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sparse Auto Encoders&amp;#039;&amp;#039;&amp;#039;: See Semanticity.&lt;br /&gt;
* [https://github.com/saprmarks/dictionary_learning dictionary_learning]&lt;br /&gt;
* [https://transformer-circuits.pub/2024/jan-update/index.html#predict-future Predicting Future Activations]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11944 Transcoders Find Interpretable LLM Feature Circuits]&lt;br /&gt;
* 2024-10: [https://transformer-circuits.pub/2024/crosscoders/index.html Sparse Crosscoders for Cross-Layer Features and Model Diffing]&lt;br /&gt;
&lt;br /&gt;
==Reward Functions==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12491 Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL]&lt;br /&gt;
&lt;br /&gt;
==Symbolic and Notation==&lt;br /&gt;
* [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* [https://www.arxiv.org/abs/2407.09468 Beyond Euclid: An Illustrated Guide to Modern Machine Learning with Geometric, Topological, and Algebraic Structures]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02423 On the Anatomy of Attention]: Introduces category-theoretic diagrammatic formalism for DL architectures&lt;br /&gt;
* 2024-11: [https://x.com/vtabbott_/status/1860268276569506250 diagrams to represent algorithms]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03317 FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awareness]&lt;br /&gt;
&lt;br /&gt;
==Mathematical==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.13762 Unveiling the Hidden Structure of Self-Attention via Kernel Principal Component Analysis]&lt;br /&gt;
&lt;br /&gt;
==Geometric==&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.03658 The Linear Representation Hypothesis and the Geometry of Large Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.01506 The Geometry of Categorical and Hierarchical Concepts in Large Language Models]&lt;br /&gt;
** Natural hierarchies of concepts---which occur throughout natural language and especially in scientific ontologies---are represented in the model&amp;#039;s internal vectorial space as polytopes that can be decomposed into simplexes of mutually-exclusive categories.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02678 Reasoning in Large Language Models: A Geometric Perspective]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.17592 Deep Manifold Part 1: Anatomy of Neural Network Manifold]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.19750 The Geometry of Concepts: Sparse Autoencoder Feature Structure]&lt;br /&gt;
** Tegmark et al. report multi-scale structure: 1) “atomic” small-scale, 2) “brain” intermediate-scale, and 3) “galaxy” large-scale&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.08009 The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.09782 The Geometry of Reasoning: Flowing Logics in Representation Space]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/linebreaks/index.html When Models Manipulate Manifolds: The Geometry of a Counting Task]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.26745 Deep sequence models tend to memorize geometrically; it is unclear why]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
==Topography==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.06002 The Molecular Structure of Thought: Mapping the Topology of Long Chain-of-Thought Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2023-07Jul: [https://arxiv.org/abs/2307.13702 Measuring Faithfulness in Chain-of-Thought Reasoning] [https://x.com/davidad/status/1839641113432305790 roughly] proves that sufficiently large models do not generate CoT that actually captures their internal reasoning)&lt;br /&gt;
&lt;br /&gt;
[[Image:GYe31yXXQAABwaZ.jpeg|300px]]&lt;br /&gt;
&lt;br /&gt;
=Heuristic Understanding=&lt;br /&gt;
* 2022-09: Janus: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators]&lt;br /&gt;
&lt;br /&gt;
==Emergent Internal Model Building==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.15936 A Theory for Emergence of Complex Skills in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19370v1 Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01622 General agents contain world models]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20328 Video models are zero-shot learners and reasoners]&lt;br /&gt;
&lt;br /&gt;
===Semantic Directions===&lt;br /&gt;
Directions, e.g.: f(king)-f(man)+f(woman)=f(queen) or f(sushi)-f(Japan)+f(Italy)=f(pizza)&lt;br /&gt;
* [https://arxiv.org/abs/1301.3781 Efficient Estimation of Word Representations in Vector Space]&lt;br /&gt;
* [https://aclanthology.org/N13-1090/ Linguistic Regularities in Continuous Space Word Representations]&lt;br /&gt;
* [https://aclanthology.org/C16-1332 Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen]&lt;br /&gt;
* [https://aclanthology.org/D14-1162/ Glove: Global vectors for word representation]&lt;br /&gt;
* [https://doi.org/10.1109/BigData.2015.7364114 Using Word2Vec to process big text data]&lt;br /&gt;
* [https://arxiv.org/abs/2310.06824 The geometry of truth: Emergent linear structure in large language model representations of true/false datasets] (true/false)&lt;br /&gt;
* [https://arxiv.org/abs/2403.10381 Monotonic Representation of Numeric Properties in Language Models] (numeric directions)&lt;br /&gt;
Task vectors:&lt;br /&gt;
* [https://arxiv.org/abs/2310.15213 Function Vectors in Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.15916 In-context learning creates task vectors]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/5FGXmJ3wqgGRcbyH7/extracting-sae-task-features-for-in-context-learning Extracting sae task features for in-context learning]&lt;br /&gt;
* [https://arxiv.org/abs/2412.12276 Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers]&lt;br /&gt;
Reasoning:&lt;br /&gt;
* [https://openreview.net/forum?id=OwhVWNOBcz Understanding Reasoning in Thinking Language Models via Steering Vectors]&lt;br /&gt;
&lt;br /&gt;
===Feature Geometry Reproduces Problem-space===&lt;br /&gt;
* [https://arxiv.org/abs/2210.13382 Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2309.00941 Emergent linear representations in world models of self-supervised sequence models] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* [https://doi.org/10.1038/s41562-023-01659-w Emergent analogical reasoning in large language models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.02207 Language Models Represent Space and Time] (Maps of world, US)&lt;br /&gt;
* [https://arxiv.org/abs/2405.14860 Not All Language Model Features Are Linear] (Days of week form ring, etc.)&lt;br /&gt;
* [https://arxiv.org/abs/2406.03689 Evaluating the World Model Implicit in a Generative Model] (Map of Manhattan)&lt;br /&gt;
* [https://iopscience.iop.org/article/10.1088/1748-9326/ad2891 Reliable precipitation nowcasting using probabilistic diffusion models]. Generation of precipitation map imagery is predictive of actual future weather; implies model is learning scientifically-relevant modeling.&lt;br /&gt;
* [https://arxiv.org/abs/2405.07987 The Platonic Representation Hypothesis]: Different models (including across modalities) are converging to a consistent world model.&lt;br /&gt;
* [https://arxiv.org/abs/2501.00070 ICLR: In-Context Learning of Representations]&lt;br /&gt;
* [https://arxiv.org/abs/2502.00873 Language Models Use Trigonometry to Do Addition]: Numbers arranged in helix to enable addition&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
&lt;br /&gt;
===Capturing Physics===&lt;br /&gt;
* 2020-09: [https://arxiv.org/abs/2009.08292 Learning to Identify Physical Parameters from Video Using Differentiable Physics]&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.00419 Self-Supervised Learning for Videos: A Survey]&lt;br /&gt;
* 2025-02: Fair at Meta: [https://arxiv.org/abs/2502.11831 Intuitive physics understanding emerges from self-supervised pretraining on natural videos]&lt;br /&gt;
&lt;br /&gt;
===Theory of Mind===&lt;br /&gt;
* [https://arxiv.org/abs/2302.02083 Evaluating Large Language Models in Theory of Mind Tasks]&lt;br /&gt;
* [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-01: [https://www.arxiv.org/abs/2501.09038 Do generative video models learn physical principles from watching videos?] ([https://physics-iq.github.io/ project], [https://github.com/google-deepmind/physics-IQ-benchmark code])&lt;br /&gt;
* 2025-06: [https://machinelearning.apple.com/research/illusion-of-thinking The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21521 Potemkin Understanding in Large Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21876 Do Vision-Language Models Have Internal World Models? Towards an Atomic Evaluation]&lt;br /&gt;
&lt;br /&gt;
==Information Processing==&lt;br /&gt;
* 2019-03: [https://arxiv.org/abs/1903.05789 Diagnosing and Enhancing VAE Models]&lt;br /&gt;
* 2021-03: [https://arxiv.org/abs/2103.05247 Pretrained Transformers as Universal Computation Engines]&lt;br /&gt;
* 2022-10: [https://arxiv.org/abs/2210.08344 How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03843 Why think step by step? Reasoning emerges from the locality of experience]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.04444 What&amp;#039;s the Magic Word? A Control Theory of LLM Prompting]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12875 Chain of Thought Empowers Transformers to Solve Inherently Serial Problems]: Proves that transformers can solve any problem, if they can generate sufficient intermediate tokens&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.20311 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process]&lt;br /&gt;
** Models learning reasoning skills (they are not merely memorizing solution templates). They can mentally generate simple short plans (like humans).&lt;br /&gt;
** When presented facts, models develop internal understanding of what parameters (recursively) depend on each other. This occurs even before an explicit question is asked (i.e. before the task is defined). This appears to be different from human reasoning.&lt;br /&gt;
** Model depth matters for reasoning. This cannot be mitigated by chain-of-thought prompting (which allow models to develop and then execute plans) since even a single CoT step may require deep, multi-step reasoning/planning.&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.01992 Ask, and it shall be given: Turing completeness of prompting]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08775 Layers at Similar Depths Generate Similar Activations Across LLM Architectures]&lt;br /&gt;
&lt;br /&gt;
===Generalization===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]&lt;br /&gt;
&lt;br /&gt;
===Grokking===&lt;br /&gt;
* 2022-01: [https://arxiv.org/abs/2201.02177 Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets]&lt;br /&gt;
* 2022-05: [https://arxiv.org/abs/2205.10343 Towards Understanding Grokking: An Effective Theory of Representation Learning]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.10463 Critical Data Size of Language Models from a Grokking Perspective]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15175 Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18624 How to explain grokking]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.21519 Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
===Tests of Resilience to Dropouts/etc.===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15390 Explorations of Self-Repair in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15786 What Matters in Transformers? Not All Attention is Needed]&lt;br /&gt;
** Removing entire transformer blocks leads to significant performance degradation&lt;br /&gt;
** Removing MLP layers results in significant performance degradation&lt;br /&gt;
** Removing attention layers causes almost no performance degradation&lt;br /&gt;
** E.g. half of attention layers are deleted (48% speed-up), leads to only 2.4% decrease in the benchmarks&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19384 The Remarkable Robustness of LLMs: Stages of Inference?]&lt;br /&gt;
** They intentionally break the network (swapping layers), yet it continues to work remarkably well. This suggests LLMs are quite robust, and allows them to identify different stages in processing.&lt;br /&gt;
** They also use these interventions to infer what different layers are doing. They break apart the LLM transformer layers into four stages:&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Detokenization:&amp;#039;&amp;#039;&amp;#039; Raw tokens are converted into meaningful entities that take into account local context (especially using nearby tokens).&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Feature engineering:&amp;#039;&amp;#039;&amp;#039; Features are progressively refined. Factual knowledge is leveraged.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Prediction ensembling:&amp;#039;&amp;#039;&amp;#039; Predictions (for the ultimately-selected next-token) emerge. A sort of consensus voting is used, with “prediction neurons” and &amp;quot;suppression neurons&amp;quot; playing a major role in upvoting/downvoting.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Residual sharpening:&amp;#039;&amp;#039;&amp;#039; The semantic representations are collapsed into specific next-token predictions. There is a strong emphasis on suppression neurons eliminating options. The confidence is calibrated.&lt;br /&gt;
** This structure can be thought of as two halves (being roughly dual to each other): the first half broadens (goes from distinct tokens to a rich/elaborate concept-space) and the second half collapses (goes from rich concepts to concrete token predictions).&lt;br /&gt;
&lt;br /&gt;
==Semantic Vectors==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11717 Refusal in Language Models Is Mediated by a Single Direction]&lt;br /&gt;
* 2025-02: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs] ([https://x.com/OwainEvans_UK/status/1894436637054214509 demonstrates] [https://x.com/ESYudkowsky/status/1894453376215388644 entangling] of concepts into a single preference vector)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03666 Analogical Reasoning Inside Large Language Models: Concept Vectors and the Limits of Abstraction]&lt;br /&gt;
&lt;br /&gt;
==Other==&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00247 Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting &amp;amp; Beyond]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04282 Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding] ([https://github.com/SalesforceAIResearch/LaTRO code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.12580 Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models]: LLMs learn reasoning by extracting procedures from training data, not by memorizing specific answers&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15862 LLMs Do Not Think Step-by-step In Implicit Reasoning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
&lt;br /&gt;
===Scaling Laws===&lt;br /&gt;
* 1993: [https://proceedings.neurips.cc/paper/1993/file/1aa48fc4880bb0c9b8a3bf979d3b917e-Paper.pdf Learning Curves: Asymptotic Values and Rate of Convergence]&lt;br /&gt;
* 2017-12: [https://arxiv.org/abs/1712.00409 Deep Learning Scaling is Predictable, Empirically] (Baidu)&lt;br /&gt;
* 2019-03: [http://www.incompleteideas.net/IncIdeas/BitterLesson.html The Bitter Lesson] (Rich Sutton)&lt;br /&gt;
* 2020-01: [https://arxiv.org/abs/2001.08361 Scaling Laws for Neural Language Models] (OpenAI)&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.14701 Scaling Laws for Autoregressive Generative Modeling] (OpenAI)&lt;br /&gt;
* 2020-05: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis] (Gwern)&lt;br /&gt;
* 2021-08: [https://arxiv.org/abs/2108.07686 Scaling Laws for Deep Learning]&lt;br /&gt;
* 2021-02: [https://arxiv.org/abs/2102.06701 Explaining Neural Scaling Laws] (Google DeepMind)&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.15556 Training Compute-Optimal Large Language Models] (Chinchilla, Google DeepMind)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.04715 Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.07951 Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models]&lt;br /&gt;
* 2025-05: [https://brendel-group.github.io/llm-line/ LLMs on the Line: Data Determines Loss-To-Loss Scaling Laws]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13786 The Art of Scaling Reinforcement Learning Compute for LLMs]&lt;br /&gt;
&lt;br /&gt;
=Information Processing/Storage=&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.10689 A Theory of Usable Information Under Computational Constraints]&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.00008 Why is AI hard and Physics simple?]&lt;br /&gt;
* 2021-06: [https://arxiv.org/abs/2106.06981 Thinking Like Transformers]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.00948 Large Linguistic Models: Investigating LLMs&amp;#039; metalinguistic abilities]&lt;br /&gt;
* &amp;quot;A transformer&amp;#039;s depth affects its reasoning capabilities, whilst model size affects its knowledge capacity&amp;quot; ([https://x.com/danielhanchen/status/1835684061475655967 c.f.])&lt;br /&gt;
** 2024-02: [https://arxiv.org/abs/2402.14905 MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases]&lt;br /&gt;
** 2024-04: [https://arxiv.org/abs/2404.08819 The Illusion of State in State-Space Models] (figure 3)&lt;br /&gt;
** 2024-08: [https://arxiv.org/abs/2408.00118 Gemma 2: Improving Open Language Models at a Practical Size] (table 9)&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.10482 Schrodinger&amp;#039;s Memory: Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2407.01687 Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning]. CoT involves both memorization and (probabilitic) reasoning&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.16679 Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?]&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.03961 A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.22471 The Bayesian Geometry of Transformer Attention]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03220 From Entropy to Epiplexity: Rethinking Information for Computationally Bounded Intelligence]&lt;br /&gt;
&lt;br /&gt;
==Statistics/Math==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.05465 The emergence of clusters in self-attention dynamics]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.10794 A mathematical perspective on Transformers]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.12034 Understanding Transformers via N-gram Statistics]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.06833 Dynamic metastability in the self-attention model]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04551 Measure-to-measure interpolation using Transformers]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14697 Quantitative Clustering in Mean-Field Transformer Models]&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
===For numbers/math===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.14903 Tokenization counts: the impact of tokenization on arithmetic in frontier LLMs]: L2R vs. R2L yields different performance on math&lt;br /&gt;
&lt;br /&gt;
==Data Storage==&lt;br /&gt;
* 1988-09: [https://www.sciencedirect.com/science/article/pii/0885064X88900209 On the capabilities of multilayer perceptrons]&lt;br /&gt;
* 2006-12: [https://ieeexplore.ieee.org/document/4038449 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition] (single-layer perceptron stores &amp;gt;2 bits/parameter; MLP ~ 2*N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; bits w/ N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; params)&lt;br /&gt;
* 2016-11: [https://arxiv.org/abs/1611.09913 Capacity and Trainability in Recurrent Neural Networks] (5 bits/param)&lt;br /&gt;
* 2018-02: [https://arxiv.org/abs/1802.08232 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks]&lt;br /&gt;
* 2019-05: [https://ieeexplore.ieee.org/document/8682462 Memorization Capacity of Deep Neural Networks under Parameter Quantization]&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.08910 How Much Knowledge Can You Pack Into the Parameters of a Language Model?]&lt;br /&gt;
* 2020-08: [https://arxiv.org/abs/2008.09036 Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries] (capacity scales linearly with parameters; more training samples leads to less memorization)&lt;br /&gt;
* 2020-12: [https://arxiv.org/abs/2012.06421 When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05405 Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws] (2 bits/param)&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15720 Scaling Laws for Fact Memorization of Large Language Models] (1T params needed to memorize Wikipedia)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24832 How much do language models memorize?] (3.6 bits/parameter)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01855 Trade-offs in Data Memorization via Strong Data Processing Inequalities]&lt;br /&gt;
&lt;br /&gt;
===Reverse-Engineering Training Data===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.10364 Can We Infer Confidential Properties of Training Data from LLMs?]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15553 Approximating Language Model Training Data from Weights]&lt;br /&gt;
&lt;br /&gt;
===Compression===&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.09410 Less is More: Parameter-Free Text Classification with Gzip]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.04050 LLMZip: Lossless Text Compression using Large Language Models]&lt;br /&gt;
* 2023-07: [https://aclanthology.org/2023.findings-acl.426/ “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.10668 Language Modeling Is Compression]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07550 An Image is Worth 32 Tokens for Reconstruction and Generation]&lt;br /&gt;
&lt;br /&gt;
==Learning/Training==&lt;br /&gt;
* 2018-03: [https://arxiv.org/abs/1803.03635 The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks]: Sparse neural networks are optimal, but it is difficult to identify the right architecture and train it. Deep learning typically consists of training a dense neural network, which makes it easier to learn an internal sparse circuit optimal to a particular problem.&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11521 On the Ability of Deep Networks to Learn Symmetries from Data: A Neural Kernel Theory]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.12391 Physics of Skill Learning]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24864 ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Cross-modal knowledge transfer===&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.07519 Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.07358 Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]: CLIP learns richer set of aggregated representations (e.g. for a culture or country), vs. a vision-only model.&lt;br /&gt;
&lt;br /&gt;
==Hidden State==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06258 Emergent Response Planning in LLM]: They show that the latent representation contains information beyond that needed for the next token (i.e. the model learns to &amp;quot;plan ahead&amp;quot; and encode information relevant to future tokens)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02854 (How) Do Language Models Track State?]&lt;br /&gt;
===Convergent Representation===&lt;br /&gt;
* 2015-11: [https://arxiv.org/abs/1511.07543 Convergent Learning: Do different neural networks learn the same representations?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of Embeddings]: Evidence for [https://x.com/jxmnop/status/1925224620166128039 The Strong Platonic Representation Hypothesis]; models converge to a single consensus reality&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Function Approximation==&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]: can learn linear functions (equivalent to least-squares estimator)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09066 Teaching Algorithmic Reasoning via In-context Learning]: Simple arithmetic &lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models] ([https://github.com/ekinakyurek/google-research/tree/master/incontext code]): can learn linear regression&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.00297 Transformers learn to implement preconditioned gradient descent for in-context learning]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.03576 One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.02893 ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20545 SoS1: O1 and R1-Like Reasoning LLMs are Sum-of-Square Solvers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.21212 Transformers Learn to Implement Multi-step Gradient Descent with Chain of Thought]&lt;br /&gt;
&lt;br /&gt;
=Physics Based=&lt;br /&gt;
* 2014-01: [https://arxiv.org/abs/1401.1219 Consciousness as a State of Matter]&lt;br /&gt;
* 2016-08: [https://arxiv.org/abs/1608.08225 Why does deep and cheap learning work so well?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.23489 SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training]&lt;br /&gt;
* 2025-12: [https://www.pnas.org/doi/full/10.1073/pnas.2523012122 Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning]&lt;br /&gt;
&lt;br /&gt;
=Failure Modes=&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?]: Poor causal inference&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.12288 The Reversal Curse: LLMs trained on &amp;quot;A is B&amp;quot; fail to learn &amp;quot;B is A&amp;quot;]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.13638 Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve] (biases towards &amp;quot;common&amp;quot; numbers, in-context CoT can reduce performance by incorrectly priming, etc.)&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.16093 Visual cognition in multimodal large language models] (models lack human-like visual understanding)&lt;br /&gt;
&lt;br /&gt;
==Adversarial==&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03507 Solving adversarial examples requires solving exponential misalignment]&lt;br /&gt;
&lt;br /&gt;
==Fracture Representation==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.11581 Questioning Representational Optimism in Deep Learning: The Fractured Entangled Representation Hypothesis] ([https://github.com/akarshkumar0101/fer code])&lt;br /&gt;
&lt;br /&gt;
==Jagged Frontier==&lt;br /&gt;
* 2023-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03211 How Does Quantization Affect Multilingual LLMs?]: Quantization degrades different languages by differing amounts&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061v1 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]: Scaling laws are skill-dependent&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.18212 A Definition of AGI]&lt;br /&gt;
&lt;br /&gt;
===See also===&lt;br /&gt;
* [[AI_understanding|AI Understanding]] &amp;gt; [[AI_understanding#Psychology|Psychology]] &amp;gt; [[AI_understanding#LLM_personalities|LLM personalities]]&lt;br /&gt;
* [[AI tricks]] &amp;gt; [[AI_tricks#Prompt_Engineering|Prompt Engineering]] &amp;gt; [[AI_tricks#Brittleness|Brittleness]]&lt;br /&gt;
&lt;br /&gt;
===Conversely (AI models converge)===&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.05117 The Universal Weight Subspace Hypothesis]&lt;br /&gt;
* 2026-01: [https://avikrishna.substack.com/p/eliciting-frontier-model-character Eliciting Frontier Model Character Training: A study of personality convergence across language models]&lt;br /&gt;
&lt;br /&gt;
==Model Collapse==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.12202 Nepotistically Trained Generative-AI Models Collapse]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05280 On the Limits of Self-Improving in LLMs and Why AGI, ASI and the Singularity Are Not Near Without Symbolic Model Synthesis]&lt;br /&gt;
&lt;br /&gt;
===Analysis===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
&lt;br /&gt;
===Mitigation===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08117 Convergence Dynamics and Stabilization Strategies of Co-Evolving Generative Models]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.11111 Inducing anxiety in large language models can induce bias]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17117 From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning]&lt;br /&gt;
* 2025-07: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5357179 Call Me A Jerk: Persuading AI to Comply with Objectionable Requests]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.06047 &amp;quot;They parted illusions -- they parted disclaim marinade&amp;quot;: Misalignment as structural fidelity in LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.02606 Gender Dynamics and Homophily in a Social Network of LLM Agents]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
* 2026-04: [https://www.anthropic.com/research/emotion-concepts-function Emotion concepts and their function in a large language model]&lt;br /&gt;
&lt;br /&gt;
==Persona Simulator Theory==&lt;br /&gt;
* 2022-09: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators] ([https://www.lesswrong.com/users/janus-1?from=post_header janus])&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.findings-emnlp.423/ Language Models as Agent Models]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.00805 Conditioning Predictive Models: Risks and Strategies]&lt;br /&gt;
* 2024-09: [https://www.lesswrong.com/s/qhdHbCJ3PYesL9dde Intuitive Self-Models]&lt;br /&gt;
* 2026-02: [https://alignment.anthropic.com/2026/psm/ The Persona Selection Model: Why AI Assistants might Behave like Humans] (Anthropic, [https://www.anthropic.com/research/persona-selection-model blog])&lt;br /&gt;
&lt;br /&gt;
==Allow LLM to think==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
&lt;br /&gt;
===In-context Learning===&lt;br /&gt;
* 2021-10: [https://arxiv.org/abs/2110.15943 MetaICL: Learning to Learn In Context]&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?]&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16003 Learning without training: The implicit dynamics of in-context learning]&lt;br /&gt;
&lt;br /&gt;
==Reasoning (CoT, etc.)==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18009 Large Language Models Think Too Fast To Explore Effectively]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18585 Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.08156 Are DeepSeek R1 And Other Reasoning Models More Faithful?]: reasoning models can provide faithful explanations for why their reasoning is correct&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08679 Chain-of-Thought Reasoning In The Wild Is Not Always Faithful]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.04022 Rethinking Reflection in Pre-Training]: pre-training alone already provides some amount of reflection/reasoning&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2501.18858 BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning]&lt;br /&gt;
&lt;br /&gt;
===Pathfinding===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08152 DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09284 Tree-OPO: Off-policy Monte Carlo Tree-Guided Advantage Optimization for Multistep Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06160v1 Reverse-Engineered Reasoning for Open-Ended Generation]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06941 The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2508.01191 Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens]&lt;br /&gt;
&lt;br /&gt;
==Self-Awareness and Self-Recognition and Introspection==&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.05221 Language Models (Mostly) Know What They Know]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13076 LLM Evaluators Recognize and Favor Their Own Generations]&lt;br /&gt;
* 2024-09: [https://situational-awareness-dataset.org/ Me, Myself and AI: The Situational Awareness Dataset for LLMs]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* 2024-12: [https://theaidigest.org/self-awareness AIs are becoming more self-aware. Here&amp;#039;s why that matters]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-04: [https://x.com/Josikinz/status/1907923319866716629 LLMs can guess which comic strip was generated by themselves (vs. other LLM)]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/introspection/index.html Emergent Introspective Awareness in Large Language Models] (Anthropic, [https://www.anthropic.com/research/introspection blog])&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.24661 Do Large Language Models Know What They Are Capable Of?]&lt;br /&gt;
&lt;br /&gt;
==LLM personalities==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.02618 Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04343 Psychologically Enhanced AI Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10387 The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models]&lt;br /&gt;
&lt;br /&gt;
==Quirks &amp;amp; Biases==&lt;br /&gt;
* 2025-04: [https://www.cambridge.org/core/journals/judgment-and-decision-making/article/artificial-intelligence-and-dichotomania/0421D2310727D73FAB47069FD1620AA1 Artificial intelligence and dichotomania]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.22818 Can Large Language Models Develop Gambling Addiction?]&lt;br /&gt;
&lt;br /&gt;
=Vision Models=&lt;br /&gt;
* 2017-11: Distill: [https://distill.pub/2017/feature-visualization/ Feature Visualization: How neural networks build up their understanding of images]&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.12322 Position, Padding and Predictions: A Deeper Look at Position Information in CNNs]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13181 Perception Encoder: The best visual embeddings are not at the output of the network] ([https://github.com/facebookresearch/perception_models code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8772</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8772"/>
		<updated>2026-04-02T15:31:07Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* AI Persuasion of Humans */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
* 2026-03: Anthropic: [https://www.anthropic.com/features/81k-interviews What 81,000 people want from AI]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41562-025-02194-6 On the conversational persuasiveness of GPT-4]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
** 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.15245 Practicing with Language Models Cultivates Human Empathic Communication]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8771</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8771"/>
		<updated>2026-04-02T13:40:40Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* (Pre) Generate Articles */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models]&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
** 2026-04: [https://www.nature.com/articles/s41586-026-10265-5 Towards end-to-end automation of AI research]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.20179 AI Agents Can Already Autonomously Perform Experimental High Energy Physics]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Personalities==&lt;br /&gt;
* 2026-03: [https://github.com/msitarzewski/agency-agents The Agency: AI Specialists Ready to Transform Your Workflow]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
** 2026-03: Three problems solved using OpenAI GPT internal model. Paper: [https://arxiv.org/pdf/2603.29961 Short Proofs in Combinatorics and Number Theory]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8770</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8770"/>
		<updated>2026-04-01T16:32:19Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Math */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models]&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.20179 AI Agents Can Already Autonomously Perform Experimental High Energy Physics]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Personalities==&lt;br /&gt;
* 2026-03: [https://github.com/msitarzewski/agency-agents The Agency: AI Specialists Ready to Transform Your Workflow]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
** 2026-03: Three problems solved using OpenAI GPT internal model. Paper: [https://arxiv.org/pdf/2603.29961 Short Proofs in Combinatorics and Number Theory]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI&amp;diff=8769</id>
		<title>AI</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI&amp;diff=8769"/>
		<updated>2026-03-31T19:24:32Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* News */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page pulls together a set of resources focused on &amp;#039;&amp;#039;&amp;#039;Artificial Intelligence&amp;#039;&amp;#039;&amp;#039; (AI) and &amp;#039;&amp;#039;&amp;#039;machine-learning&amp;#039;&amp;#039;&amp;#039; (ML), biased towards modern/frontier generative AI (LLMs, etc.).&lt;br /&gt;
&lt;br /&gt;
==Fundamentals==&lt;br /&gt;
* [[AI tutorials]]&lt;br /&gt;
* [[AI understanding]]: Papers that expose how LLMs &amp;quot;think&amp;quot;&lt;br /&gt;
* [[AI tools]]: List of modern models (LLM, ASR, etc.) and related tools (RAG, etc.)&lt;br /&gt;
** [[Data Extraction]]&lt;br /&gt;
** [[AI compute]]&lt;br /&gt;
*** [[AI_compute#Energy_Use|Energy use]]&lt;br /&gt;
&lt;br /&gt;
==Improvements==&lt;br /&gt;
* [[AI tricks]]&lt;br /&gt;
* [[AI research trends]]&lt;br /&gt;
* [[Increasing AI Intelligence]] (especially system 2 / deliberative reasoning / inference-time compute)&lt;br /&gt;
* [[AI benchmarks]]&lt;br /&gt;
&lt;br /&gt;
==Agents==&lt;br /&gt;
* [[AI Agents]]&lt;br /&gt;
** [[Science Agents]]&lt;br /&gt;
* [[Exocortex]]&lt;br /&gt;
* [http://yager-research.ca/2024/11/what-is-an-ai-agent/ Definition of AI agent]&lt;br /&gt;
[[Image:AI definitions10.png|500px]]&lt;br /&gt;
&lt;br /&gt;
==Uses of AI==&lt;br /&gt;
* [[AI video]]: Progress of generative video&lt;br /&gt;
* In science, see: [[Science Agents]]&lt;br /&gt;
** [[Science_Agents#Genuine_Discoveries|Genuine Discoveries]]&lt;br /&gt;
* [[AI creativity]]&lt;br /&gt;
* [[AI and Humans]]&lt;br /&gt;
** [[AI_and_Humans#AI_in_Education|AI in education]]&lt;br /&gt;
** [[AI_and_Humans#Simulate_Humans|Simulate Humans]]&lt;br /&gt;
&lt;br /&gt;
==Related==&lt;br /&gt;
* [[AI predictions]]&lt;br /&gt;
* [[AI safety]]&lt;br /&gt;
* [[Robots]]&lt;br /&gt;
* [[Human Computer Interaction]] (HCI)&lt;br /&gt;
** [[Human_Computer_Interaction#Smart_Wearables|AI devices]] (Smart Glasses, etc.)&lt;br /&gt;
* [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==News==&lt;br /&gt;
[http://yager-research.ca/category/news/ Newsletters released here]. Posts:&lt;br /&gt;
===2026===&lt;br /&gt;
* [http://yager-research.ca/2026/03/ai-news-2026-03-31/ AI News 2026-03-31]&lt;br /&gt;
* [http://yager-research.ca/2026/02/ai-news-2026-02-28/ AI News 2026-02-28]&lt;br /&gt;
* [http://yager-research.ca/2026/01/ai-news-2026-01-31/ AI News 2026-01-31]&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
* [http://yager-research.ca/2025/12/ai-news-2025-12-25/ AI News 2025-12-25]&lt;br /&gt;
* [http://yager-research.ca/2025/12/ai-news-2025-12-18/ AI News 2025-12-18]&lt;br /&gt;
* [http://yager-research.ca/2025/12/ai-news-2025-12-11/ AI News 2025-12-11]&lt;br /&gt;
* [http://yager-research.ca/2025/12/ai-news-2025-12-04/ AI News 2025-12-04]&lt;br /&gt;
* [http://yager-research.ca/2025/11/ai-news-2025-11-27/ AI News 2025-11-27]&lt;br /&gt;
* [http://yager-research.ca/2025/11/ai-news-2025-11-20/ AI News 2025-11-20]&lt;br /&gt;
* [http://yager-research.ca/2025/11/ai-news-2025-11-13/ AI News 2025-11-13]&lt;br /&gt;
* [http://yager-research.ca/2025/11/ai-news-2025-11-06/ AI News 2025-11-06]&lt;br /&gt;
* [http://yager-research.ca/2025/10/ai-news-2025-10-30/ AI News 2025-10-30]&lt;br /&gt;
* [http://yager-research.ca/2025/10/ai-news-2025-10-23/ AI News 2025-10-23]&lt;br /&gt;
* [http://yager-research.ca/2025/10/ai-news-2025-10-16/ AI News 2025-10-16]&lt;br /&gt;
* [http://yager-research.ca/2025/10/ai-news-2025-10-09/ AI News 2025-10-09]&lt;br /&gt;
* [http://yager-research.ca/2025/10/ai-news-2025-10-02/ AI News 2025-10-02]&lt;br /&gt;
* [http://yager-research.ca/2025/09/ai-news-2025-09-25/ AI News 2025-09-25]&lt;br /&gt;
* [http://yager-research.ca/2025/09/ai-news-2025-09-18/ AI News 2025-09-18]&lt;br /&gt;
* [http://yager-research.ca/2025/09/ai-news-2025-09-11/ AI News 2025-09-11]&lt;br /&gt;
* [http://yager-research.ca/2025/09/ai-news-2025-09-04/ AI News 2025-09-04]&lt;br /&gt;
* [http://yager-research.ca/2025/08/ai-news-2025-08-28/ AI News 2025-08-28]&lt;br /&gt;
* [http://yager-research.ca/2025/08/ai-news-2025-08-21/ AI News 2025-08-21]&lt;br /&gt;
* [http://yager-research.ca/2025/08/ai-news-2025-08-14/ AI News 2025-08-14]&lt;br /&gt;
* [http://yager-research.ca/2025/08/ai-news-2025-08-07/ AI News 2025-08-07]&lt;br /&gt;
* [http://yager-research.ca/2025/07/ai-news-2025-07-31/ AI News 2025-07-31]&lt;br /&gt;
* [http://yager-research.ca/2025/07/ai-news-2025-07-24/ AI News 2025-07-24]&lt;br /&gt;
* [http://yager-research.ca/2025/07/ai-news-2025-07-17/ AI News 2025-07-17]&lt;br /&gt;
* [http://yager-research.ca/2025/07/ai-news-2025-07-10/ AI News 2025-07-10]&lt;br /&gt;
* [http://yager-research.ca/2025/07/ai-news-2025-07-03/ AI News 2025-07-03]&lt;br /&gt;
* [http://yager-research.ca/2025/06/ai-news-2025-06-26/ AI News 2025-06-26]&lt;br /&gt;
* [http://yager-research.ca/2025/06/ai-news-2025-06-19/ AI News 2025-06-19]&lt;br /&gt;
* [http://yager-research.ca/2025/06/ai-news-2025-06-12/ AI News 2025-06-12]&lt;br /&gt;
* [http://yager-research.ca/2025/06/ai-news-2025-06-05/ AI News 2025-06-05]&lt;br /&gt;
* [http://yager-research.ca/2025/05/ai-news-2025-05-29/ AI News 2025-05-29]&lt;br /&gt;
* [http://yager-research.ca/2025/05/ai-news-2025-05-22/ AI News 2025-05-22]&lt;br /&gt;
* [http://yager-research.ca/2025/05/ai-news-2025-05-15/ AI News 2025-05-15]&lt;br /&gt;
* [http://yager-research.ca/2025/05/ai-news-2025-05-08/ AI News 2025-05-08]&lt;br /&gt;
* [http://yager-research.ca/2025/05/ai-news-2025-05-01/ AI News 2025-05-01]&lt;br /&gt;
* [http://yager-research.ca/2025/04/ai-news-2025-04-24/ AI News 2025-04-24]&lt;br /&gt;
* [http://yager-research.ca/2025/04/ai-news-2025-04-17/ AI News 2025-04-17]&lt;br /&gt;
* [http://yager-research.ca/2025/04/ai-news-2025-04-10/ AI News 2025-04-10]&lt;br /&gt;
* [http://yager-research.ca/2025/04/ai-news-2025-04-03/ AI News 2025-04-03]&lt;br /&gt;
* [http://yager-research.ca/2025/03/ai-news-2025-03-27/ AI News 2025-03-27]&lt;br /&gt;
* [http://yager-research.ca/2025/03/ai-news-2025-03-20/ AI News 2025-03-20]&lt;br /&gt;
* [http://yager-research.ca/2025/03/ai-news-2025-03-13/ AI News 2025-03-13]&lt;br /&gt;
* [http://yager-research.ca/2025/03/ai-news-2025-03-06/ AI News 2025-03-06]&lt;br /&gt;
* [http://yager-research.ca/2025/02/ai-news-2025-02-27/ AI News 2025-02-27]&lt;br /&gt;
* [http://yager-research.ca/2025/02/ai-news-2025-02-20/ AI News 2025-02-20]&lt;br /&gt;
* [http://yager-research.ca/2025/02/ai-news-2025-02-13/ AI News 2025-02-13]&lt;br /&gt;
* [http://yager-research.ca/2025/02/ai-news-2025-02-06/ AI News 2025-02-06]&lt;br /&gt;
* [http://yager-research.ca/2025/01/ai-news-2025-01-30/ AI News 2025-01-30]&lt;br /&gt;
* [http://yager-research.ca/2025/01/ai-news-2025-01-23/ AI News 2025-01-23]&lt;br /&gt;
* [http://yager-research.ca/2025/01/ai-news-2025-01-16/ AI News 2025-01-16]&lt;br /&gt;
* [http://yager-research.ca/2025/01/ai-news-2025-01-09/ AI News 2025-01-09]&lt;br /&gt;
* [http://yager-research.ca/2025/01/ai-new-2025-01-02/ AI News 2025-01-02]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
* [http://yager-research.ca/2024/12/ai-news-2024-12-26/ AI News 2024-12-26]&lt;br /&gt;
* [http://yager-research.ca/2024/12/ai-news-2024-12-19/ AI News 2024-12-19]&lt;br /&gt;
* [http://yager-research.ca/2024/12/ai-news-2024-12-12/ AI News 2024-12-12]&lt;br /&gt;
* [http://yager-research.ca/2024/12/ai-news-2024-12-05/ AI News 2024-12-05]&lt;br /&gt;
* [http://yager-research.ca/2024/11/ai-news-2024-11-28/ AI News 2024-11-28]&lt;br /&gt;
* [http://yager-research.ca/2024/11/ai-news-2024-11-21/ AI News 2024-11-21]&lt;br /&gt;
* [http://yager-research.ca/2024/11/ai-news-2024-11-14/ AI News 2024-11-14]&lt;br /&gt;
* [http://yager-research.ca/2024/11/ai-new-2024-11-07/ AI News 2024-11-07]&lt;br /&gt;
* [http://yager-research.ca/2024/10/ai-news-2024-10-31/ AI News 2024-10-31]&lt;br /&gt;
* [http://yager-research.ca/2024/10/ai-news-2024-10-24/ AI News 2024-10-24]&lt;br /&gt;
* [http://yager-research.ca/2024/10/ai-news-2024-10-17/ AI News 2024-10-17]&lt;br /&gt;
* [http://yager-research.ca/2024/10/ai-news-2024-10-10/ AI News 2024-10-10]&lt;br /&gt;
* [http://yager-research.ca/2024/10/ai-news-2024-10-03/ AI News 2024-10-03]&lt;br /&gt;
* [http://yager-research.ca/2024/09/ai-news-2024-09-26/ AI News 2024-09-26]&lt;br /&gt;
* [http://yager-research.ca/2024/09/ai-news-2024-09-19/ AI News 2024-09-19]&lt;br /&gt;
* [http://yager-research.ca/2024/09/ai-news-2024-09-12/ AI News 2024-09-12]&lt;br /&gt;
* [http://yager-research.ca/2024/09/ai-news-2024-09-05/ AI News 2024-09-05]&lt;br /&gt;
* [http://yager-research.ca/2024/08/ai-news-2024-08-29/ AI News 2024-08-29]&lt;br /&gt;
* [http://yager-research.ca/2024/08/ai-news-2024-08-22/ AI News 2024-08-22]&lt;br /&gt;
* [http://yager-research.ca/2024/08/can-we-distinguish-human-from-ai/ 2024-08-16: Can we Distinguish Human from AI?]&lt;br /&gt;
* [http://yager-research.ca/2024/08/ai-news-2024-08-15/ AI News 2024-08-15]&lt;br /&gt;
* [http://yager-research.ca/2024/08/ai-news-2024-08-08/ AI News 2024-08-08]&lt;br /&gt;
* [http://yager-research.ca/2024/08/ai-news-2024-08-01/ AI News 2024-08-01]&lt;br /&gt;
* [http://yager-research.ca/2024/07/ai-news-2024-07-25/ AI News 2024-07-25]&lt;br /&gt;
* [http://yager-research.ca/2024/07/ai-news-2024-07-18/ AI News 2024-07-18]&lt;br /&gt;
* [http://yager-research.ca/2024/07/ai-news-2024-07-11/ AI News 2024-07-11]&lt;br /&gt;
* [http://yager-research.ca/2024/07/ai-news-2024-07-04/ AI News 2024-07-04]&lt;br /&gt;
* [http://yager-research.ca/2024/06/ai-news-2024-06-27/ AI News 2024-06-27]&lt;br /&gt;
* [http://yager-research.ca/2024/06/ai-news-2024-06-14/ AI News 2024-06-14]&lt;br /&gt;
* [http://yager-research.ca/2024/06/situational-awareness/ 2024-06-09: Leopold Aschenbrenner&amp;#039;s Situational Awareness]&lt;br /&gt;
* [http://yager-research.ca/2024/06/ai-news-2024-06-06/ AI News 2024-06-06]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_compute&amp;diff=8768</id>
		<title>AI compute</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_compute&amp;diff=8768"/>
		<updated>2026-03-31T15:04:01Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Energy Use */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Analysis=&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.14123 AI and Memory Wall]&lt;br /&gt;
&lt;br /&gt;
=Cloud GPU=&lt;br /&gt;
* [https://lambdalabs.com/ Lambda]&lt;br /&gt;
* [https://vast.ai/ Vast AI]&lt;br /&gt;
* [https://lightning.ai/ Lightning AI]&lt;br /&gt;
* [https://www.runpod.io/ RunPod]&lt;br /&gt;
* [https://hpc-ai.com/ HPC-AI]&lt;br /&gt;
&lt;br /&gt;
=Cloud Training Compute=&lt;br /&gt;
* [https://nebius.ai/ Nebius AI]&lt;br /&gt;
* [https://glaive.ai/ Glaive AI]&lt;br /&gt;
&lt;br /&gt;
=Cloud LLM Routers &amp;amp; Inference Providers=&lt;br /&gt;
* [https://openrouter.ai/ OpenRouter] (open and closed models, no Enterprise tier)&lt;br /&gt;
* [https://www.litellm.ai/ LiteLLM] (closed models, Enterprise tier)&lt;br /&gt;
* [https://centml.ai/ Cent ML] (open models, Enterprise tier)&lt;br /&gt;
* [https://fireworks.ai/ Fireworks AI] (open models, Enterprise tier)&lt;br /&gt;
* [https://abacus.ai/ Abacus AI] (open and closed models, Enterprise tier)&lt;br /&gt;
* [https://portkey.ai/ Portkey] (open? and closed models, Enterprise tier)&lt;br /&gt;
* [https://www.together.ai/ Together AI] (open models, Enterprise tier)&lt;br /&gt;
* [https://hyperbolic.xyz/ Hyperbolic AI] (open models, Enterprise tier)&lt;br /&gt;
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub]&lt;br /&gt;
* [https://www.asksage.ai/ AskSage]&lt;br /&gt;
* [https://opencode.ai/docs/zen/ Opencode Zen] (for coding agents)&lt;br /&gt;
&lt;br /&gt;
==Multi-model with Model Selection==&lt;br /&gt;
* [https://www.notdiamond.ai/ Not Diamond ¬⋄]&lt;br /&gt;
* [https://withmartian.com/ Martian]&lt;br /&gt;
&lt;br /&gt;
==Multi-model Web Chat Interfaces==&lt;br /&gt;
* [https://simtheory.ai/ SimTheory]&lt;br /&gt;
* [https://abacus.ai/ Abacus AI] [https://chatllm.abacus.ai/ ChatLLM]&lt;br /&gt;
* [https://poe.com/about Poe]&lt;br /&gt;
* [https://gab.ai/ Gab AI]&lt;br /&gt;
* [https://www.vectal.ai/login Vectal] ?&lt;br /&gt;
* [https://www.blackbox.ai/ BlackboxAI]&lt;br /&gt;
&lt;br /&gt;
==Multi-model Web Playground Interfaces==&lt;br /&gt;
* [https://www.together.ai/ Together AI]&lt;br /&gt;
* [https://hyperbolic.xyz/ Hyperbolic AI]&lt;br /&gt;
&lt;br /&gt;
=Local Router=&lt;br /&gt;
* [https://ollama.com/ Ollama]&lt;br /&gt;
* [https://github.com/mudler/LocalAI LocalAI]&lt;br /&gt;
* [https://github.com/AK391/ai-gradio ai-gradio]: unified model interface (based on [https://www.gradio.app/ gradio])&lt;br /&gt;
&lt;br /&gt;
=Acceleration Hardware=&lt;br /&gt;
* [https://www.nvidia.com/ Nvidia] GPUs&lt;br /&gt;
* Google [https://en.wikipedia.org/wiki/Tensor_Processing_Unit TPU]&lt;br /&gt;
* [https://www.etched.com/ Etched]: Transformer ASICs&lt;br /&gt;
* [https://cerebras.ai/ Cerebras]&lt;br /&gt;
* [https://www.untether.ai/ Untether AI]&lt;br /&gt;
* [https://www.graphcore.ai/ Graphcore]&lt;br /&gt;
* [https://sambanova.ai/ SambaNova Systems]&lt;br /&gt;
* [https://groq.com/ Groq]&lt;br /&gt;
* Tesla [https://en.wikipedia.org/wiki/Tesla_Dojo Dojo]&lt;br /&gt;
* [https://deepsilicon.com/ Deep Silicon]: Combined hardware/software solution for accelerated AI ([https://x.com/sdianahu/status/1833186687369023550 e.g.] ternary math)&lt;br /&gt;
&lt;br /&gt;
=Energy Use=&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.10350 Carbon Emissions and Large Neural Network Training]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.03003 From Words to Watts: Benchmarking the Energy Costs of Large Language Model Inference]&lt;br /&gt;
* 2024-01: [https://iea.blob.core.windows.net/assets/6b2fd954-2017-408e-bf08-952fdd62118a/Electricity2024-Analysisandforecastto2026.pdf Electricity 2024: Analysis and forecast to 2026]&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-54271-x The carbon emissions of writing and illustrating are lower for AI than for humans]&lt;br /&gt;
* 2025-04: [https://andymasley.substack.com/p/a-cheat-sheet-for-conversations-about Why using ChatGPT is not bad for the environment - a cheat sheet]&lt;br /&gt;
** A single LLM response uses only ~3 Wh = 11 kJ (~10 Google searches; [https://docs.google.com/document/d/1pDdpPq3MyPdEAoTkho9YABZ0NBEhBH2v4EA98fm3pXQ/edit?usp=sharing examples of 3 Wh energy usage])&lt;br /&gt;
** Reading an LLM-generated response (computer running for a few minutes) typically uses more energy than the LLM generation of the text.&lt;br /&gt;
* 2025-07: Mistral: [https://mistral.ai/news/our-contribution-to-a-global-environmental-standard-for-ai Our contribution to a global environmental standard for AI]&lt;br /&gt;
* 2025-08: [https://services.google.com/fh/files/misc/measuring_the_environmental_impact_of_delivering_ai_at_google_scale.pdf Measuring the environmental impact of delivering AI at Google Scale] ([https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference blog])&lt;br /&gt;
* 2026-01: [https://epoch.ai/data-insights/grok-4-training-resources What did it take to train Grok 4?]&lt;br /&gt;
&lt;br /&gt;
==Examples==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;LLM query&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** 3 kW * 4s = 3 Wh = 11 kJ&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Human brain&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** 20 W * 8h = 106 Wh&lt;br /&gt;
** 20 W * 1h = 20 Wh&lt;br /&gt;
** 20 W * 10m = 3 Wh = 10 kJ&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Human brain excess thinking&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** 2 W * 8h = 11 Wh&lt;br /&gt;
** 2 W * 1.7h = 3 Wh&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Regular computer&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
** 200 W * 8h = 1,600 Wh = 5,700 kJ&lt;br /&gt;
** 200 W * 1m = 3 Wh = 10kJ&lt;br /&gt;
&lt;br /&gt;
==Water Use==&lt;br /&gt;
* [https://andymasley.substack.com/p/the-ai-water-issue-is-fake The AI water issue is fake. On the national, local, and personal level.]&lt;br /&gt;
&lt;br /&gt;
==Heat Exhaust==&lt;br /&gt;
* 2026-03: [https://blog.andymasley.com/p/data-centers-heat-exhaust-is-not Data centers&amp;#039; heat exhaust is not raising the land temperature around where they&amp;#039;re built]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8767</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8767"/>
		<updated>2026-03-30T18:10:40Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
** [https://x.com/maxescu/status/2036434854435315868?s=20 Cinematic scenes] (3.5m, comedy, [https://lumalabs.ai/uni-1 Luma Uni-1 Agent])&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;br /&gt;
* March 2026: [https://app.pixverse.ai/onboard Pixverse v6] ([https://x.com/fal/status/2038655807483490613?s=20 example])&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8766</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8766"/>
		<updated>2026-03-30T18:10:19Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
** [https://x.com/maxescu/status/2036434854435315868?s=20 Cinematic scenes] (3.5m, comedy, [https://lumalabs.ai/uni-1 Luma Uni-1 Agent])&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;br /&gt;
* March 2026: [https://app.pixverse.ai/onboard Pixverse v6]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Talk:AI_video&amp;diff=8765</id>
		<title>Talk:AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Talk:AI_video&amp;diff=8765"/>
		<updated>2026-03-30T16:39:01Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Others for Consideration */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Others for Consideration=&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1927061347331694973 Influenders] ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/venturetwins/status/1934027410841764221 Koala shot by protesters]&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1932835386557939913 Riot] ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** Celebrity explainer [https://x.com/venturetwins/status/1934434222523171000 1], [https://x.com/venturetwins/status/1934438139738874129 2]&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/IamEmily2050/status/1945795374251479388 Quick rap] (example JSON format)&lt;br /&gt;
** [https://x.com/sweeneydailyx/status/1948032121429500221 Commercial for American Eagle (20s)] (the car driving off is an AI extension of the clip)&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1955305090971017653 Waidmanns Heil ([https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/Gossip_Goblin/status/1996994382428336165?s=20 Joy Loop] (1.5m)&lt;br /&gt;
** [https://x.com/TUPACABRA2/status/2005877025454662066?s=20 Minnesota Dark] (2m, [https://x.com/TUPACABRA2 Tupacrabra])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/2008990455661515071?s=20 Egg Protein] (2m)&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/aimikoda/status/2038285542727487827?s=20 Fashion sequence] (15s, Seedance 2.0)&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_safety&amp;diff=8764</id>
		<title>AI safety</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_safety&amp;diff=8764"/>
		<updated>2026-03-30T16:37:15Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Status */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Learning Resources=&lt;br /&gt;
==Light==&lt;br /&gt;
* [https://orxl.org/ai-doom.html a casual intro to AI doom and alignment] (2022)&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
** [https://interactive.keepthefuturehuman.ai/ Interactive Explainer]&lt;br /&gt;
** [https://keepthefuturehuman.ai/essay/ Essay: Keep the Future Human]&lt;br /&gt;
** [https://www.youtube.com/watch?v=27KDl2uPiL8 We Can’t Stop AI – Here’s What To Do Instead] (4m video, 2025)&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late] (15m video, 2025)&lt;br /&gt;
* Tristan Harris TED talk (15m): [https://www.ted.com/talks/tristan_harris_why_ai_is_our_ultimate_test_and_greatest_invitation Why AI is our ultimate test and greatest invitation]&lt;br /&gt;
** Text version: Center for Humane Technology: [https://centerforhumanetechnology.substack.com/p/the-narrow-path-why-ai-is-our-ultimate The Narrow Path: Why AI is Our Ultimate Test and Greatest Invitation]&lt;br /&gt;
* [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI]&lt;br /&gt;
* 2024-10: [https://www.youtube.com/watch?v=xfMQ7hzyFW4 Writing Doom]: short film on Superintelligence (27m video)&lt;br /&gt;
* 2026-03: [https://www.youtube.com/watch?v=Nl7-bRFSZBs The AI book that&amp;#039;s freaking out national security advisors] (44m video)&lt;br /&gt;
&lt;br /&gt;
==Deep==&lt;br /&gt;
* [https://www.thecompendium.ai/ The Compendium: Humanity risks extinction from its very creations — AIs.] (2024)&lt;br /&gt;
* [https://www.aisafetybook.com/ Introduction to AI Safety, Ethics, and Society] (Dan Hendrycks, [https://www.safe.ai/ Center for AI Safety])&lt;br /&gt;
* [https://aisafety.info/ AI Safety FAQ]&lt;br /&gt;
* [https://deepmindsafetyresearch.medium.com/introducing-our-short-course-on-agi-safety-1072adb7912c DeepMind short course on AGI safety]&lt;br /&gt;
&lt;br /&gt;
=Description of Safety Concerns=&lt;br /&gt;
==Key Concepts==&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Instrumental_convergence Instrumental Convergence]&lt;br /&gt;
* [https://www.lesswrong.com/w/orthogonality-thesis Orthogonality Thesis]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/SzecSPYxqRa5GCaSF/clarifying-inner-alignment-terminology Inner/outer alignment]&lt;br /&gt;
* [https://www.alignmentforum.org/w/mesa-optimization Mesa-optimization]&lt;br /&gt;
* [https://www.lesswrong.com/posts/N6vZEnCn6A95Xn39p/are-we-in-an-ai-overhang Overhang]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/pdaGN6pQyQarFHXF4/reward-is-not-the-optimization-target Reward is not the optimization target] (Alex Turner)&lt;br /&gt;
* 80,000 hours:&lt;br /&gt;
** [https://80000hours.org/problem-profiles/risks-from-power-seeking-ai/ Risks from power-seeking AI systems]&lt;br /&gt;
** [https://80000hours.org/problem-profiles/gradual-disempowerment/ Gradual disempowerment]&lt;br /&gt;
** [https://80000hours.org/problem-profiles/catastrophic-ai-misuse/ Catastrophic AI misuse]&lt;br /&gt;
&lt;br /&gt;
==Medium-term Risks==&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=xoVJKj8lcNQ A.I. Dilemma – Tristan Harris and Aza Raskin” (video)] ([https://assets-global.website-files.com/5f0e1294f002b1bb26e1f304/64224a9051a6637c1b60162a_65-your-undivided-attention-The-AI-Dilemma-transcript.pdf podcast transcript]): raises concern about human ability to handle these transformations&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=KCSsKV5F4xc Daniel Schmachtenberger and Liv Boeree (video)]: AI could accelerate perverse social dynamics&lt;br /&gt;
* 2023-10: [https://arxiv.org/pdf/2310.11986 Sociotechnical Safety Evaluation of Generative AI Systems] (Google DeepMind)&lt;br /&gt;
* 2024-02: [https://yoshuabengio.org/2024/02/26/towards-a-cautious-scientist-ai-with-convergent-safety-bounds/ Towards a Cautious Scientist AI with Convergent Safety Bounds] (Yoshua Bengio)&lt;br /&gt;
* 2024-07: [https://yoshuabengio.org/2024/07/09/reasoning-through-arguments-against-taking-ai-safety-seriously/ Reasoning through arguments against taking AI safety seriously] (Yoshua Bengio)&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20702 The Singapore Consensus on Global AI Safety Research Priorities]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.adz1697 How malicious AI swarms can threaten democracy: The fusion of agentic AI and LLMs marks a new frontier in information warfare] (Science Magazine, [https://arxiv.org/abs/2506.06299 preprint])&lt;br /&gt;
* 2026-01: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI] (Dario Amodei)&lt;br /&gt;
* 2026-02: [https://www.noahpinion.blog/p/updated-thoughts-on-ai-risk Updated thoughts on AI risk: Things have gotten scarier since 2023] ([https://x.com/Noahpinion Noah Smith])&lt;br /&gt;
&lt;br /&gt;
==Long-term  (x-risk)==&lt;br /&gt;
* 2015-02: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-1 Machine intelligence, part 1]&lt;br /&gt;
* 2019-03: Daniel Kokotajlo and Wei Dai: [https://www.lesswrong.com/posts/WXvt8bxYnwBYpy9oT/the-main-sources-of-ai-risk The Main Sources of AI Risk?]&lt;br /&gt;
* 2022-06: Eliezer Yudkowsky: [https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities List AGI Ruin: A List of Lethalities]&lt;br /&gt;
* 2024-11: Marcus Arvan: [https://link.springer.com/article/10.1007/s00146-024-02113-9 ‘Interpretability’ and ‘alignment’ are fool’s errands: a proof that controlling misaligned large language models is the best anyone can hope for]&lt;br /&gt;
* 2025-04: [https://michaelnotebook.com/xriskbrief/index.html ASI existential risk: reconsidering alignment as a goal]&lt;br /&gt;
* 2025-12: Philip Trammell and Leopold Aschenbrenner: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth]&lt;br /&gt;
&lt;br /&gt;
=Status=&lt;br /&gt;
* 2025-01: [https://assets.publishing.service.gov.uk/media/679a0c48a77d250007d313ee/International_AI_Safety_Report_2025_accessible_f.pdf International Safety Report: The International Scientific Report on the Safety of Advanced AI (January 2025)]&lt;br /&gt;
* [https://ailabwatch.org/ AI Lab Watch] (safety scorecard)&lt;br /&gt;
* 2026-03: [https://windowsontheory.org/2026/03/30/the-state-of-ai-safety-in-four-fake-graphs/ The state of AI safety in four fake graphs]&lt;br /&gt;
&lt;br /&gt;
==Assessmment==&lt;br /&gt;
* [https://aiassessmentscale.com/ AI Assessment Scale (AIAS)]: A practical framework to guide the appropriate and ethical use of generative AI in assessment design, empowering educators to make purposeful, evidence-based decisions&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16534 Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report]&lt;br /&gt;
&lt;br /&gt;
==Policy==&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.05694 On the Limitations of Compute Thresholds as a Governance Strategy] Sara Hooker&lt;br /&gt;
* 2024-07: [https://www.cigionline.org/static/documents/AI-challenges.pdf Framework Convention on Global AI Challenges] ([https://www.cigionline.org/ CIGI])&lt;br /&gt;
* 2024-08: NIST guidelines: [https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.800-1.ipd.pdf Managing Misuse Risk for Dual-Use Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Proposals==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.18359 Responsible AI Agents]&lt;br /&gt;
* 2025-03: [https://controlai.com/ Control AI] [https://controlai.com/dip The Direct Institutional Plan]&lt;br /&gt;
* 2025-04: Google DeepMind: [https://deepmind.google/discover/blog/taking-a-responsible-path-to-agi/ Taking a responsible path to AGI]&lt;br /&gt;
** Paper: [https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/evaluating-potential-cybersecurity-threats-of-advanced-ai/An_Approach_to_Technical_AGI_Safety_Apr_2025.pdf An Approach to Technical AGI Safety and Security]&lt;br /&gt;
&lt;br /&gt;
=Research=&lt;br /&gt;
* 2008: [https://selfawaresystems.com/wp-content/uploads/2008/01/ai_drives_final.pdf The Basic AI Drives]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2209.00626v1 The alignment problem from a deep learning perspective]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.03827 Discovering Latent Knowledge in Language Models Without Supervision]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.08582 Pretraining Language Models with Human Preferences]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03279 Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the MACHIAVELLI Benchmark]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.15324 Model evaluation for extreme risks] (DeepMind)&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.03047 Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.17492 Preference Ranking Optimization for Human Alignment]&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.06259 Self-Alignment with Instruction Backtranslation]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.08702 Debate Helps Supervise Unreliable Experts]&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/weak-to-strong-generalization.pdf Weak-to-Strong Generalization: Eliciting Strong Capabilities with Weak Supervision] (OpenAI, [https://openai.com/research/weak-to-strong-generalization blog])&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf Practices for Governing Agentic AI Systems] (OpenAI, [https://openai.com/index/practices-for-governing-agentic-ai-systems/ blog])&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist through Safety Training] (Anthropic)&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13208 The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions] (OpenAI)&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.04622 On scalable oversight with weak LLMs judging strong LLMs]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.21792 Safetywashing: Do AI Safety Benchmarks Actually Measure Safety Progress?] (Dan Hendrycks et al.)&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.00761 Tamper-Resistant Safeguards for Open-Weight LLMs] ([https://www.tamper-resistant-safeguards.com/ project], [https://github.com/rishub-tamirisa/tamper-resistance/ code])&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04614 Better Alignment with Instruction Back-and-Forth Translation]&lt;br /&gt;
* 2024-10: [https://cdn.openai.com/papers/first-person-fairness-in-chatbots.pdf First-Person Fairness in Chatbots] (OpenAI, [https://openai.com/index/evaluating-fairness-in-chatgpt/ blog])&lt;br /&gt;
* 2024-10: [https://assets.anthropic.com/m/377027d5b36ac1eb/original/Sabotage-Evaluations-for-Frontier-Models.pdf Sabotage evaluations for frontier models] (Anthropic, [https://www.anthropic.com/research/sabotage-evaluations blog])&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/983c85a201a962f/original/Alignment-Faking-in-Large-Language-Models-full-paper.pdf Alignment Faking in Large Language Models] (Anthropic)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03556 Best-of-N Jailbreaking] ([https://github.com/jplhughes/bon-jailbreaking code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16325 Towards Safe and Honest AI Agents with Neural Self-Other Overlap]&lt;br /&gt;
** 2024-07: [https://www.lesswrong.com/posts/hzt9gHpNwA2oHtwKX/self-other-overlap-a-neglected-approach-to-ai-alignment Self-Other Overlap: A Neglected Approach to AI Alignment]&lt;br /&gt;
** 2025-03: [https://www.lesswrong.com/posts/jtqcsARGtmgogdcLT/reducing-llm-deception-at-scale-with-self-other-overlap-fine Reducing LLM deception at scale with self-other overlap fine-tuning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16339 Deliberative Alignment: Reasoning Enables Safer Language Models] (OpenAI)&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/papers/trading-inference-time-compute-for-adversarial-robustness-20250121_1.pdf Trading Inference-Time Compute for Adversarial Robustness] (OpenAI, [https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness/ blog])&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18837 Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming] (Anthropic, [https://www.anthropic.com/research/constitutional-classifiers blog], &lt;br /&gt;
* 2025-02: [https://drive.google.com/file/d/1QAzSj24Fp0O6GfkskmnULmI1Hmx7k_EJ/view Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs] ([https://www.emergent-values.ai/ site], [https://github.com/centerforaisafety/emergent-values github])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.07776 Auditing Prompt Caching in Language Model APIs]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14143 Multi-Agent Risks from Advanced AI]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2209.00626v7 The Alignment Problem from a Deep Learning Perspective]&lt;br /&gt;
* 2025-03: [https://assets.anthropic.com/m/317564659027fb33/original/Auditing-Language-Models-for-Hidden-Objectives.pdf Auditing language models for hidden objectives] (Anthropic, [https://www.anthropic.com/research/auditing-hidden-objectives blog])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13621 Superalignment with Dynamic Human Values]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.15125 Contemplative Wisdom for Superalignment]&lt;br /&gt;
* 2025-04: [https://www.lesswrong.com/posts/x59FhzuM9yuvZHAHW/untitled-draft-yhra Scaling Laws for Scalable Oversight] ([https://arxiv.org/abs/2504.18530 preprint], [https://github.com/subhashk01/oversight-scaling-laws code])&lt;br /&gt;
* 2025-06: [https://assets.anthropic.com/m/4fb35becb0cd87e1/original/SHADE-Arena-Paper.pdf SHADE-Arena: Evaluating sabotage and monitoring in LLM agents] (Anthropic, [https://www.anthropic.com/research/shade-arena-sabotage-monitoring blog])&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13609 Avoiding Obfuscation with Prover-Estimator Debate]&lt;br /&gt;
* 2025-06: [https://cdn.openai.com/pdf/a130517e-9633-47bc-8397-969807a43a23/emergent_misalignment_paper.pdf Persona Features Control Emergent Misalignment] (OpenAI, [https://openai.com/index/emergent-misalignment/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2506.18032 Why Do Some Language Models Fake Alignment While Others Don&amp;#039;t?] (Anthropic, [https://github.com/safety-research/open-source-alignment-faking code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.11473 Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety]&lt;br /&gt;
* 2025-09: [https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/ Detecting and reducing scheming in AI models]&lt;br /&gt;
* 2025-11: [https://assets.anthropic.com/m/74342f2c96095771/original/Natural-emergent-misalignment-from-reward-hacking-paper.pdf Natural Emergent Misalignment from Reward Hacking in Production RL] (Anthropic, [https://www.anthropic.com/research/emergent-misalignment-reward-hacking blog])&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16856 Distributional AGI Safety]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2511.22662 Difficulties with Evaluating a Deception Detector for AIs]&lt;br /&gt;
* 2025-12: [https://cdn.openai.com/pdf/d57827c6-10bc-47fe-91aa-0fde55bd3901/monitoring-monitorability.pdf Monitoring Monitorability] (OpenAI)&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09937-5 Training large language models on narrow tasks can lead to broad misalignment]&lt;br /&gt;
** 2025-02: Preprint: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/pdf/2601.23045 The Hot Mess of AI: How Does Misalignment Scale with Model Intelligence and Task Complexity?] (Anthropic [https://alignment.anthropic.com/2026/hot-mess-of-ai/ blog])&lt;br /&gt;
* 2026-03: [https://cdn.openai.com/pdf/a21c39c1-fa07-41db-9078-973a12620117/cot_controllability.pdf Reasoning Models Struggle to Control their Chains of Thought] (OpenAI [https://openai.com/index/reasoning-models-chain-of-thought-controllability/ blog])&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
&lt;br /&gt;
==Demonstrations of Negative Use Capabilities==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.00586 Evaluating Large Language Models&amp;#039; Capability to Launch Fully Automated Spear Phishing Campaigns: Validated on Human Subjects]&lt;br /&gt;
* 2025-04: [https://www.nathanlabenz.com/ Nathan Labenz] ([https://www.cognitiverevolution.ai/ The Cognitive Revolution]): [https://docs.google.com/presentation/d/1mvkpg1mtAvGzTiiwYPc6bKOGsQXDIwMb-ytQECb3i7I/edit#slide=id.g252d9e67d86_0_16 AI Bad Behavior]&lt;br /&gt;
&lt;br /&gt;
==Threat Vectors==&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.07192 Poisoning Attacks on LLMs Require a Near-constant Number of Poison Samples]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8763</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8763"/>
		<updated>2026-03-30T15:05:29Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Literature */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models]&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.20179 AI Agents Can Already Autonomously Perform Experimental High Energy Physics]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Personalities==&lt;br /&gt;
* 2026-03: [https://github.com/msitarzewski/agency-agents The Agency: AI Specialists Ready to Transform Your Workflow]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_tools&amp;diff=8762</id>
		<title>AI tools</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_tools&amp;diff=8762"/>
		<updated>2026-03-26T20:03:15Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Open Source */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=LLM=&lt;br /&gt;
==Open-weights LLM==&lt;br /&gt;
* [https://about.fb.com/news/2023/07/llama-2/ 2023-07Jul-18]: [https://llama.meta.com/llama2/ Llama2] 7B, 13B,  70B&lt;br /&gt;
* [https://ai.meta.com/blog/meta-llama-3/ 2024-04Apr-18]: [https://llama.meta.com/llama3/ Llama3] 8B, 70B&lt;br /&gt;
* [https://blogs.nvidia.com/blog/nemotron-4-synthetic-data-generation-llm-training/ 2024-06Jun-14]: [https://research.nvidia.com/publication/2024-06_nemotron-4-340b Nemotron-4] 340B&lt;br /&gt;
* 2024-07Jul-23: [https://llama.meta.com/ Llama 3.1] 8B, 70B, 405B&lt;br /&gt;
* [https://mistral.ai/news/mistral-large-2407/ 2024-07Jul-24]: [https://huggingface.co/mistralai/Mistral-Large-Instruct-2407 Mistral Large 2] 128B&lt;br /&gt;
* [https://developers.googleblog.com/en/smaller-safer-more-transparent-advancing-responsible-ai-with-gemma/ 2024-07Jul-31]: [https://huggingface.co/collections/google/gemma-2-2b-release-66a20f3796a2ff2a7c76f98f Gemma 2] 2B&lt;br /&gt;
* [https://qwenlm.github.io/blog/qwen2-math/ 2024-08Aug-08]: Qwen2-Math ([https://huggingface.co/collections/Qwen/qwen2-math-66b4c9e072eda65b5ec7534d hf], [https://github.com/QwenLM/Qwen2-Math github]) 1.5B, 7B, 72B&lt;br /&gt;
* [https://nousresearch.com/releases/ 2024-08Aug-14]: [https://nousresearch.com/ Nous research] [https://nousresearch.com/hermes3/ Hermes 3] ([https://nousresearch.com/wp-content/uploads/2024/08/Hermes-3-Technical-Report.pdf technical report]) 8B, 70B, 405B&lt;br /&gt;
* 2024-08Aug-19: [https://www.salesforceairesearch.com/ Salesforce AI] [https://huggingface.co/papers/2408.08872 xGen-MM (BLIP-3)]: A Family of Open Large Multimodal Models ([https://www.arxiv.org/abs/2408.08872 preprint], [https://github.com/salesforce/LAVIS/tree/xgen-mm code])&lt;br /&gt;
* 2024-09Sep-04: [https://arxiv.org/abs/2409.02060 OLMoE: Open Mixture-of-Experts Language Models] ([https://github.com/allenai/OLMoE code]) 7B model (uses 1B per input token)&lt;br /&gt;
* 2024-09Sep-05: [https://huggingface.co/mattshumer/Reflection-70B Reflection 70B] ([https://reflection-playground-production.up.railway.app/ demo]): [https://x.com/mattshumer_/status/1831767014341538166 Trained using Reflection-Tuning, a technique developed to enable LLMs to fix their own mistakes.]&lt;br /&gt;
* 2024-09Sep-06: [https://huggingface.co/deepseek-ai/DeepSeek-V2.5 DeepSeek-V2.5] 238B mixture-of-experts (160 experts, 16B active params)&lt;br /&gt;
* 2024-09Sep-19: Microsoft GRadient-INformed (GRIN) MoE ([https://huggingface.co/spaces/GRIN-MoE-Demo/GRIN-MoE demo], [https://huggingface.co/microsoft/GRIN-MoE model], [https://github.com/microsoft/GRIN-MoE github]) 6.6B&lt;br /&gt;
* 2024-09Sep-23: Nvidia [https://huggingface.co/nvidia/Llama-3_1-Nemotron-51B-Instruct Llama-3_1-Nemotron-51B-instruct] 51B&lt;br /&gt;
* 2024-09Sep-25: Meta [https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/ Llama 3.2] with visual and voice modalities 1B, 3B, 11B, 90B&lt;br /&gt;
* 2024-09Sep-25: [https://allenai.org/ Ai2] [https://molmo.allenai.org/ Molmo] [https://molmo.allenai.org/blog multi-modal models] 1B, 7B, 72B&lt;br /&gt;
* 2024-10Oct-01: Nvidia [https://huggingface.co/nvidia/NVLM-D-72B NVLM-D-72B] (includes vision)&lt;br /&gt;
* [https://mistral.ai/news/ministraux/ 2024-10Oct-16]: Mistral [https://huggingface.co/mistralai/Ministral-8B-Instruct-2410 Ministral-8B-Instruct-2410]&lt;br /&gt;
* 2024-10Oct-16: Nvidia [https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Reward-HF Llama-3.1-Nemotron-70B-Reward]&lt;br /&gt;
* 2024-11Nov-04: [https://arxiv.org/abs/2411.02265 Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent] 389B ([https://github.com/Tencent/Tencent-Hunyuan-Large code], [https://huggingface.co/tencent/Tencent-Hunyuan-Large weights])&lt;br /&gt;
* 2024-11Nov-18: [https://huggingface.co/mistralai/Mistral-Large-Instruct-2411 Mistral-Large-Instruct-2411]) 123B; and [https://mistral.ai/news/pixtral-large/ Pixtral Large] multimodal model 124B ([https://huggingface.co/mistralai/Pixtral-Large-Instruct-2411 weights])&lt;br /&gt;
* 2024-11Nov-22: Nvidia [https://github.com/NVlabs/hymba Hymba] ([https://developer.nvidia.com/blog/hymba-hybrid-head-architecture-boosts-small-language-model-performance/ blog]): small and high-performance&lt;br /&gt;
* 2024-12Dec-06: Meta [https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct Llama 3.3] 70B&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1872242657348710721 2024-12Dec-26]: [https://huggingface.co/deepseek-ai/DeepSeek-V3-Base DeepSeek-V3-Base] 671B&lt;br /&gt;
* 2025-01Jan-02: [https://huggingface.co/PowerInfer/SmallThinker-3B-Preview SmallThinker-3B-Preview] (fine-tune of [https://huggingface.co/Qwen/Qwen2.5-3B-Instruct Qwen2.5-3b-Instruct])&lt;br /&gt;
* [https://x.com/SebastienBubeck/status/1877010995727470877 2025-01Jan-08]: Microsoft [https://huggingface.co/microsoft/phi-4 phi-4] 15B&lt;br /&gt;
* [https://x.com/MiniMax__AI/status/1879226391352549451 2025-01Jan-14]: [https://www.minimaxi.com/en/news/minimax-01-series-2 MiniMax-01], MiniMax-Text-01 and MiniMax-VL-01; 4M context length ([https://www.minimaxi.com/en/news/minimax-01-series-2 paper])&lt;br /&gt;
* 2025-01Jan-27: [https://qwenlm.github.io/blog/qwen2.5-1m/ Qwen2.5-1M] ([https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2.5-1M/Qwen2_5_1M_Technical_Report.pdf report])&lt;br /&gt;
* 2025-01Jan-27: DeepSeek [https://huggingface.co/deepseek-ai/Janus-Pro-7B Janus-Pro-7B] (with image capabilities)&lt;br /&gt;
* [https://x.com/cohere/status/1900170005519753365 2025-03Mar-14]: Cohere [https://cohere.com/blog/command-a Command A] ([https://huggingface.co/CohereForAI/c4ai-command-a-03-2025?ref=cohere-ai.ghost.io weights])&lt;br /&gt;
* [https://x.com/MistralAI/status/1901668499832918151 2025-03Mar-17]: [https://mistral.ai/news/mistral-small-3-1 Mistral Small 3.1] 24B ([https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503 weights])&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1904526863604883661 2025-03Mar-24]: [https://huggingface.co/deepseek-ai/DeepSeek-V3-0324 DeepSeek-V3-0324] 685B&lt;br /&gt;
* 2025-04Apr-05: Meta [https://ai.meta.com/blog/llama-4-multimodal-intelligence/ Llama 4] (109B, 400B, 2T)&lt;br /&gt;
* [https://x.com/kuchaev/status/1909444566379573646 2025-04Apr-08]: Nvidia [https://huggingface.co/nvidia/Llama-3_1-Nemotron-Ultra-253B-v1 Llama-3_1-Nemotron-Ultra-253B-v1]&lt;br /&gt;
* [https://x.com/MistralAI/status/1920119463430500541 2025-05May-07]: Mistral [https://mistral.ai/news/mistral-medium-3 Medium 3]&lt;br /&gt;
* [https://x.com/googleaidevs/status/1938279967026274383 2025-06Jun-26]: Google [https://developers.googleblog.com/en/introducing-gemma-3n-developer-guide/ Gemma 3n] (on-device multimodal)&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1953128028047102241 2025-08Aug-06]: [https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507 Qwen3-4B-Instruct-2507]&lt;br /&gt;
* [https://x.com/GoogleDeepMind/status/1956393664248271082 2025-08Aug-15]: Google [https://developers.googleblog.com/en/introducing-gemma-3-270m/ Gemma 3 270M]&lt;br /&gt;
* [https://x.com/arcee_ai/status/2016278017572495505?s=20 2026-01Jan-28]: [https://www.arcee.ai/ Arcee AI] [https://docs.arcee.ai/get-started/models-overview Trinity Large] [https://huggingface.co/arcee-ai 400B]&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
Rankings: [https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard bigcode-models-leaderboard] and [https://codeelo-bench.github.io/#leaderboard-table CodeElo leaderboard]&lt;br /&gt;
* 2024-10Oct-06: [https://abacus.ai/ Abacus AI] [https://huggingface.co/abacusai/Dracarys2-72B-Instruct Dracarys2-72B-Instruct] (optimized for coding, fine-tune of [https://huggingface.co/Qwen/Qwen2.5-72B-Instruct Qwen2.5-72B-Instruct])&lt;br /&gt;
* 2024-11Nov-09: [https://opencoder-llm.github.io/ OpenCoder: The Open Cookbook for Top-Tier Code Large Language Models] ([https://huggingface.co/collections/infly/opencoder-672cec44bbb86c39910fb55e weights], [https://arxiv.org/abs/2411.04905 preprint])&lt;br /&gt;
* 2024-11Nov-13: [https://qwenlm.github.io/blog/qwen2.5-coder-family/ Qwen2.5-Coder]&lt;br /&gt;
* [https://x.com/Agentica_/status/1909700115755061374 2025-04Apr-08]: [https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51 DeepCoder-14B-Preview] ([https://github.com/agentica-project/rllm code], [https://huggingface.co/agentica-org/DeepCoder-14B-Preview hf])&lt;br /&gt;
* [https://x.com/GeZhang86038849/status/1921147887871742329 2025-05May-10]: ByteDance [https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Base SeedCoder] 8B&lt;br /&gt;
* [https://x.com/Kimi_Moonshot/status/1943687594560332025 2025-07Jul-11]: [https://moonshotai.github.io/Kimi-K2/ Kimi-K2] 1T ([https://github.com/MoonshotAI/Kimi-K2 code], [https://huggingface.co/moonshotai weights])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1947766835023335516 2025-07Jul-23]: [https://qwenlm.github.io/blog/qwen3-coder/ Qwen3-Coder-480B-A35B-Instruct] ([https://github.com/QwenLM/qwen-code code], [https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct weights])&lt;br /&gt;
* [https://x.com/MiniMax_AI/status/2021980761210134808?s=20 2026-02Feb-12]: [https://www.minimax.io/news/minimax-m25 MiniMax M2.5] 230B&lt;br /&gt;
&lt;br /&gt;
===Reasoning===&lt;br /&gt;
See also: [[Increasing_AI_Intelligence|Increasing AI Intelligence]] &amp;gt; Proactive Search &amp;gt; [[Increasing_AI_Intelligence#CoT_reasoning_model|CoT reasoning model]]&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1859200141355536422 2024-11Nov-20]: DeepSeek-R1-Lite-Preview ([https://x.com/deepseek_ai/status/1859200145037869485 results], [https://x.com/teortaxesTex/status/1859259359630356955 CoT])&lt;br /&gt;
* 2024-11Nov-23: [https://arxiv.org/abs/2411.14405 Marco-o1: Towards Open Reasoning Models for Open-Ended Solutions]&lt;br /&gt;
* 2024-11Nov-27: [https://qwenlm.github.io/blog/qwq-32b-preview/ Alibaba Qwen QwQ] 32B ([https://huggingface.co/Qwen/QwQ-32B-Preview model], [https://huggingface.co/spaces/Qwen/QwQ-32B-preview demo])&lt;br /&gt;
* [https://x.com/ruliad_ai/status/1864394941029322890 2024-12Dec-04]: [https://www.ruliad.co/ Ruliad] [https://huggingface.co/ruliad/deepthought-8b-llama-v0.01-alpha Deepthought] 8B ([https://chat.ruliad.co/ demo])&lt;br /&gt;
* 2024-12Dec-24: Qwen [https://huggingface.co/Qwen/QVQ-72B-Preview QvQ-72B-preview] (visual reasoning)&lt;br /&gt;
* 2025-01Jan-10: [https://mbzuai-oryx.github.io/LlamaV-o1/ LlamaV-o1: Rethinking Step-by-step Visual Reasoning in LLMs] ([https://arxiv.org/abs/2501.06186 preprint], [https://github.com/mbzuai-oryx/LlamaV-o1 code], [https://huggingface.co/omkarthawakar/LlamaV-o1 weights])&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1881318130334814301 2025-01Jan-20]: [https://huggingface.co/deepseek-ai/DeepSeek-R1 DeepSeek-R1], [https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B DeepSeek-R1-Distill-Llama-70B], DeepSeek-R1-Distill-Qwen-32B, ... ([https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf paper])&lt;br /&gt;
* 2025-02Feb-10: [https://huggingface.co/tomg-group-umd/huginn-0125 Huginn-0125]: [https://arxiv.org/abs/2502.05171 Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach] ([https://github.com/seal-rg/recurrent-pretraining code], [https://huggingface.co/tomg-group-umd/huginn-0125 model])&lt;br /&gt;
* [https://x.com/NousResearch/status/1890148000204485088 2025-02Feb-14]: [https://huggingface.co/NousResearch/DeepHermes-3-Llama-3-8B-Preview DeepHermes 3 - Llama-3.1 8B]&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1894130603513319842 2025-02Feb-24]: Qwen [https://qwenlm.github.io/blog/qwq-max-preview/ QwQ-Max-Preview] ([https://chat.qwen.ai/ online demo])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1897361654763151544 2025-03Mar-05]: Qwen [https://qwenlm.github.io/blog/qwq-32b/ QwQ-32B] ([https://huggingface.co/spaces/Qwen/QwQ-32B-Demo demo])&lt;br /&gt;
* [https://x.com/BlinkDL_AI/status/1898579674575552558 2025-03Mar-05]: [https://github.com/BlinkDL/RWKV-LM RWKV7-G1] &amp;quot;GooseOne&amp;quot; 0.1B ([https://huggingface.co/BlinkDL/rwkv7-g1 weights], [https://arxiv.org/abs/2305.13048 preprint])&lt;br /&gt;
* [https://x.com/LG_AI_Research/status/1901803002052436323 2025-03Mar-17]: LG AI Research [https://www.lgresearch.ai/blog/view?seq=543 EXAONE Deep] 2.4B, 7.8B, 32B ([https://huggingface.co/LGAI-EXAONE/EXAONE-Deep-32B weights])&lt;br /&gt;
* [https://x.com/kuchaev/status/1902078122792775771 2025-03Mar-18]: Nvidia [https://huggingface.co/collections/nvidia/llama-nemotron-67d92346030a2691293f200b Llama Nemotron] 8B, 49B ([https://build.nvidia.com/nvidia/llama-3_3-nemotron-super-49b-v1 demo])&lt;br /&gt;
* [https://x.com/Agentica_/status/1909700115755061374 2025-04Apr-08]: [https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51 DeepCoder-14B-Preview] ([https://github.com/agentica-project/rllm code], [https://huggingface.co/agentica-org/DeepCoder-14B-Preview hf])&lt;br /&gt;
* 2025-04Apr-10: Bytedance [https://github.com/ByteDance-Seed/Seed-Thinking-v1.5 Seed-Thinking-v1.5] 200B&lt;br /&gt;
* [https://x.com/ZyphraAI/status/1910362745423425966 2025-04Apr-11]: [https://www.zyphra.com/ Zyphra] [https://www.zyphra.com/post/introducing-zr1-1-5b-a-small-but-powerful-math-code-reasoning-model ZR1-1.5B] ([https://huggingface.co/Zyphra/ZR1-1.5B weights], [https://playground.zyphra.com/sign-in use])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1916962087676612998 2025-04Apr-29]: [https://qwenlm.github.io/blog/qwen3/ Qwen3] 0.6B to 235B ([https://github.com/QwenLM/Qwen3 code], [https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f weights], [https://modelscope.cn/home modelscope])&lt;br /&gt;
* [https://x.com/DimitrisPapail/status/1917731614899028190 2025-04Apr-30]: [https://huggingface.co/microsoft/Phi-4-reasoning Phi-4 Reasoning] 14B ([https://www.microsoft.com/en-us/research/wp-content/uploads/2025/04/phi_4_reasoning.pdf tech report])&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1928061589107900779 2025-05May-28]: [https://huggingface.co/deepseek-ai/DeepSeek-R1-0528 DeepSeek-R1-0528]&lt;br /&gt;
* [https://x.com/MistralAI/status/1932441507262259564 2025-06Jun-10]: Mistral [https://mistral.ai/static/research/magistral.pdf Magistral] 24B ([https://huggingface.co/mistralai/Magistral-Small-2506 weights])&lt;br /&gt;
* [https://x.com/LoubnaBenAllal1/status/1942614508549333211 2025-07Jul-08]: [https://huggingface.co/blog/smollm3 SmolLM3]: smol, multilingual, long-context reasoner&lt;br /&gt;
* [https://x.com/OpenAI/status/1952776916517404876 2025-08Aug-05]: [https://openai.com/open-models/ OpenAI] gpt-oss-120b, gpt-oss-20b&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1953128028047102241 2025-08Aug-06]: [https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507 Qwen3-4B-Thinking-2507]&lt;br /&gt;
* 2025-09Sep: [https://huggingface.co/LLM360/K2-Think K2-Think] 32B&lt;br /&gt;
* [https://x.com/Kimi_Moonshot/status/1986449512538513505 2025-11Nov]: [https://moonshotai.github.io/Kimi-K2/thinking.html Kimi K2 Thinking] 1T (32B active)&lt;br /&gt;
* [https://x.com/deepseek_ai/status/1995452641430651132?s=20 2025-12Dec]: [https://huggingface.co/deepseek-ai/DeepSeek-V3.2 DeepsSeek-v3.2] and [https://huggingface.co/deepseek-ai/DeepSeek-V3.2-Speciale DeepSeek-v3.2-Speciale]&lt;br /&gt;
&lt;br /&gt;
===Agentic===&lt;br /&gt;
* 2025-02Feb-18: Microsoft [https://huggingface.co/microsoft/Magma-8B Magma-8B] ([https://www.arxiv.org/abs/2502.13130 preprint])&lt;br /&gt;
* 2025-02Feb-26: [https://convergence.ai/ Convergence] [https://github.com/convergence-ai/proxy-lite Proxy Lite]&lt;br /&gt;
* [https://x.com/MiniMax_AI/status/2021980761210134808?s=20 2026-02Feb-12]: [https://www.minimax.io/news/minimax-m25 MiniMax M2.5] 230B&lt;br /&gt;
&lt;br /&gt;
===Multimodal===&lt;br /&gt;
====Language/Vision====&lt;br /&gt;
* [https://arxiv.org/abs/2407.07895 LLaVA-NeXT-Interleave] ([https://huggingface.co/collections/llava-hf/llava-interleave-668e19a97da0036aad4a2f19 models], [https://huggingface.co/spaces/merve/llava-interleave demo])&lt;br /&gt;
* [https://huggingface.co/papers/2407.15841 SlowFast-LLaVA: A Strong Training-Free Baseline for Video Large Language Models]&lt;br /&gt;
* Nvidia [https://huggingface.co/collections/merve/nveagle-66d0705108582d73bb235c26 NVEagle] 13B, 7B ([https://huggingface.co/spaces/NVEagle/Eagle-X5-13B-Chat demo], [https://arxiv.org/abs/2408.15998 preprint])&lt;br /&gt;
* 2024-08Aug-29: [https://qwenlm.github.io/blog/qwen2-vl/ Qwen2-VL] 7B, 2B ([https://github.com/QwenLM/Qwen2-VL code], [https://huggingface.co/collections/Qwen/qwen2-vl-66cee7455501d7126940800d models]): Can process videos up to 20 minutes in length&lt;br /&gt;
* 2024-09Sep-11: Mistral [https://huggingface.co/mistral-community/pixtral-12b-240910 Pixtral 12B]&lt;br /&gt;
* 2024-09Sep-17: [https://nvlm-project.github.io/ NVLM 1.0]&lt;br /&gt;
* 2024-12Dec-06: Nvidia [https://arxiv.org/abs/2412.04468 NVILA: Efficient Frontier Visual Language Models]&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1883954247743725963 2025-01Jan-28]: [https://huggingface.co/collections/Qwen/qwen25-vl-6795ffac22b334a837c0f9a5 Qwen2.5-VL]&lt;br /&gt;
* 2025-02Feb-18: Microsoft [https://huggingface.co/microsoft/Magma-8B Magma-8B] ([https://www.arxiv.org/abs/2502.13130 preprint])&lt;br /&gt;
* [https://x.com/CohereForAI/status/1896923657470886234 2025-03Mar-05]: Cohere [https://cohere.com/research/aya Aya] 8B, 32B&lt;br /&gt;
* 2025-03Mar-12: Google [https://developers.googleblog.com/en/introducing-gemma3/ Gemma 3] 1B 4B, 12B, 27B ([https://storage.googleapis.com/deepmind-media/gemma/Gemma3Report.pdf technical report])&lt;br /&gt;
* [https://x.com/DeepLearningAI/status/1903295570527002729 2025-03Mar-23]: Cohere [https://cohere.com/blog/aya-vision Aya Vision] 8B, 32B ([https://huggingface.co/collections/CohereForAI/c4ai-aya-vision-67c4ccd395ca064308ee1484?ref=cohere-ai.ghost.io weights])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1904227859616641534 2025-03Mar-24]: Alibaba [https://qwenlm.github.io/blog/qwen2.5-vl-32b/ Qwen2.5-VL-32B-Instruct] ([https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct weights])&lt;br /&gt;
* 2025-05May-20: ByteDance [https://bagel-ai.org/ BAGEL: Unified Model for Multimodal Understanding and Generation] 7B ([https://huggingface.co/ByteDance-Seed/BAGEL-7B-MoT weights], [https://github.com/bytedance-seed/BAGEL code], [https://demo.bagel-ai.org/ demo])&lt;br /&gt;
&lt;br /&gt;
====Language/Vision/Speech====&lt;br /&gt;
* 2025-02Feb-27: Microsoft [https://huggingface.co/microsoft/Phi-4-multimodal-instruct Phi-4-multimodal-instruct] (language, vision, speech)&lt;br /&gt;
* [https://x.com/kyutai_labs/status/1903082848547906011 2025-03Mar-21]: kyutai [https://kyutai.org/moshivis MoshiVis] ([https://vis.moshi.chat/ demo])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1904944923159445914 2025-03Mar-26]: [https://qwenlm.github.io/blog/qwen2.5-omni/ Qwen2.5-Omni-7B] ([https://github.com/QwenLM/Qwen2.5-Omni/blob/main/assets/Qwen2.5_Omni.pdf tech report], [https://github.com/QwenLM/Qwen2.5-Omni code], [https://huggingface.co/Qwen/Qwen2.5-Omni-7B weight])&lt;br /&gt;
&lt;br /&gt;
====Language/Audio====&lt;br /&gt;
* 2025-03Mar-11: [https://github.com/soham97/mellow Mellow]: a small audio language model for reasoning, 167M ([https://arxiv.org/abs/2503.08540 paper])&lt;br /&gt;
* 2025-03Mar-12: [https://research.nvidia.com/labs/adlr/AF2/ Audio Flamingo 2] 0.5B, 1.5B, 3B [https://arxiv.org/abs/2503.03983 paper], [https://github.com/NVIDIA/audio-flamingo code]&lt;br /&gt;
&lt;br /&gt;
===RAG===&lt;br /&gt;
* 2025-04: [https://huggingface.co/collections/PleIAs/pleias-rag-680a0d78b058fffe4c16724d Pleias-RAG] 350M, 1.2B&lt;br /&gt;
** Paper: [http://ragpdf.pleias.fr/ Even Small Reasoners Should Quote Their Sources: Introducing Pleias-RAG Model Family]&lt;br /&gt;
* 2025-04: Meta ReasonIR 8B: [https://arxiv.org/abs/2504.20595 ReasonIR: Training Retrievers for Reasoning Tasks]&lt;br /&gt;
&lt;br /&gt;
==Cloud LLM==&lt;br /&gt;
* [https://groq.com/ Groq] [https://wow.groq.com/ cloud] (very fast inference)&lt;br /&gt;
&lt;br /&gt;
===Multi-modal: Audio===&lt;br /&gt;
* [https://kyutai.org/ kyutai Open Science AI Lab] chatbot [https://www.us.moshi.chat/?queue_id=talktomoshi moshi]&lt;br /&gt;
&lt;br /&gt;
==Triage==&lt;br /&gt;
* [https://arxiv.org/abs/2406.18665 RouteLLM: Learning to Route LLMs with Preference Data]&lt;br /&gt;
&lt;br /&gt;
==Retrieval Augmented Generation (RAG)==&lt;br /&gt;
* See Also: [[AI_tools#Document_Parsing|Document Parsing]]&lt;br /&gt;
&lt;br /&gt;
===Reviews===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08921 Graph Retrieval-Augmented Generation: A Survey]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14924 Retrieval Augmented Generation (RAG) and Beyond: A Comprehensive Survey on How to Make your LLMs use External Data More Wisely]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17558 A Survey of Query Optimization in Large Language Models]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.07391 Enhancing Retrieval-Augmented Generation: A Study of Best Practices]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.09136 Agentic Retrieval-Augmented Generation: A Survey on Agentic RAG] ([https://github.com/asinghcsu/AgenticRAG-Survey github])&lt;br /&gt;
* List of [https://github.com/NirDiamant/RAG_Techniques RAG techniques]&lt;br /&gt;
* [https://github.com/athina-ai/rag-cookbooks Advanced RAG Cookbooks👨🏻‍💻]&lt;br /&gt;
* [https://github.com/DEEP-PolyU/Awesome-GraphRAG Awesome-GraphRAG (GraphRAG Survey)]&lt;br /&gt;
&lt;br /&gt;
===Measuring RAG performance===&lt;br /&gt;
* 2025-01: [https://deepmind.google/discover/blog/facts-grounding-a-new-benchmark-for-evaluating-the-factuality-of-large-language-models/ The FACTS Grounding Leaderboard]: [https://arxiv.org/abs/2501.03200 Benchmarking LLMs&amp;#039; Ability to Ground Responses to Long-Form Input]&lt;br /&gt;
&lt;br /&gt;
===Analysis of RAG overall===&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13070 Is Semantic Chunking Worth the Computational Cost?]&lt;br /&gt;
&lt;br /&gt;
===Approaches===&lt;br /&gt;
* RAGFlow ([https://github.com/infiniflow/ragflow code])&lt;br /&gt;
* GraphRAG ([https://arxiv.org/abs/2404.16130 preprint], [https://github.com/microsoft/graphrag code], [https://github.com/Azure-Samples/graphrag-accelerator GraphRAG Accelerator] for easy deployment on Azure)&lt;br /&gt;
* AutoMetaRAG ([https://github.com/darshil3011/AutoMetaRAG/tree/main code])&lt;br /&gt;
* [https://verba.weaviate.io/ Verba]: RAG for [https://weaviate.io/ Weaviate] vector database ([https://github.com/weaviate/verba code], [https://www.youtube.com/watch?v=UoowC-hsaf0 video])&lt;br /&gt;
* Microsoft: [https://github.com/microsoft/PIKE-RAG PIKE-RAG: sPecIalized KnowledgE and Rationale Augmented Generation]&lt;br /&gt;
* 2024-10: Google [https://arxiv.org/abs/2410.07176 Astute RAG: Overcoming Imperfect Retrieval Augmentation and Knowledge Conflicts for Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08815 StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization]: Reformats retrieved data into task-appropriate structures (table, graph, tree).&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13765 Knowledge-Aware Query Expansion with Large Language Models for Textual and Relational Retrieval]&lt;br /&gt;
* 2024-11: [https://www.arxiv.org/abs/2411.13773 FastRAG: Retrieval Augmented Generation for Semi-structured Data]&lt;br /&gt;
* 2024-11: Microsoft [https://www.microsoft.com/en-us/research/blog/lazygraphrag-setting-a-new-standard-for-quality-and-cost/ LazyGraphRAG: Setting a new standard for quality and cost]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.19443 Auto-RAG: Autonomous Retrieval-Augmented Generation for Large Language Models]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2025-01: [https://github.com/Marker-Inc-Korea/AutoRAG AutoRAG: RAG AutoML tool for automatically finding an optimal RAG pipeline for your data]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05874 VideoRAG: Retrieval-Augmented Generation over Video Corpus]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.01142 DeepRAG: Thinking to Retrieval Step by Step for Large Language Models]&lt;br /&gt;
* 2025-02: [https://weaviate.io/developers/weaviate/tutorials/multi-vector-embeddings Multi-vector embeddings]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23513 RARE: Retrieval-Augmented Reasoning Modeling]&lt;br /&gt;
&lt;br /&gt;
===Open-source Implementations===&lt;br /&gt;
* [https://github.com/Cinnamon/kotaemon kotaemon]: An open-source clean &amp;amp; customizable RAG UI for chatting with your documents.&lt;br /&gt;
* [https://www.llamaindex.ai/ LlamaIndex] ([https://github.com/run-llama/llama_index code], [https://docs.llamaindex.ai/en/stable/ docs], [https://github.com/run-llama/voice-chat-pdf voice chat code])&lt;br /&gt;
* Nvidia [https://www.nvidia.com/en-us/ai-on-rtx/chatrtx/ ChatRTX] with [https://blogs.nvidia.com/blog/what-is-retrieval-augmented-generation/ RAG]&lt;br /&gt;
* Anthropic [https://github.com/anthropics/anthropic-quickstarts/tree/main/customer-support-agent Customer Support Agent example]&lt;br /&gt;
* [https://www.langchain.com/ LangChain] and [https://www.langchain.com/langgraph LangGraph] ([https://www.metadocs.co/2024/08/20/simple-agentic-rag-for-multi-vector-stores-with-langchain-and-langgraph/ tutorial])&lt;br /&gt;
** [https://github.com/KruxAI/ragbuilder RAGBuilder]: Automatically tunes RAG hyperparams&lt;br /&gt;
* [https://github.com/stanford-oval/WikiChat WikiChat]&lt;br /&gt;
** [https://arxiv.org/abs/2305.14292 WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia]&lt;br /&gt;
* [https://github.com/bhavnicksm/chonkie Chonkie]: No-nonsense RAG chunking library (open-source, lightweight, fast)&lt;br /&gt;
* [https://github.com/pingcap/autoflow autoflow]: open source GraphRAG (Knowledge Graph), including conversational search page&lt;br /&gt;
* [https://github.com/superlinear-ai/raglite RAGLite]&lt;br /&gt;
* [https://github.com/gusye1234/nano-graphrag nano-graphrag]: A simple, easy-to-hack GraphRAG implementation&lt;br /&gt;
* [https://github.com/electricpipelines/barq Dabarqus]&lt;br /&gt;
&lt;br /&gt;
===Web-based Tools===&lt;br /&gt;
* [https://typeset.io/ SciSpace] Chat with PDF (also available as a GPT).&lt;br /&gt;
&lt;br /&gt;
===Commercial Cloud Offerings===&lt;br /&gt;
* [https://www.graphlit.com/ Graphlit]&lt;br /&gt;
* [https://colivara.com/ ColiVara]&lt;br /&gt;
* [https://nhost.io/blog/assistants-file-stores nhost]&lt;br /&gt;
* [https://vespa.ai/ Vespa] [https://vespa.ai/solutions/enterprise-retrieval-augmented-generation/ RAG]&lt;br /&gt;
* [https://unstructured.io/ Unstructured]&lt;br /&gt;
* [https://www.fivetran.com/blog/assembling-a-rag-architecture-using-fivetran Fivetran]&lt;br /&gt;
* [https://platform.vectorize.io/ Vectorize]&lt;br /&gt;
* [https://www.voyageai.com/ Voyage AI]&lt;br /&gt;
* [https://abacus.ai/ Abacus AI]&lt;br /&gt;
* [https://www.cloudflare.com/ Cloudflare] [https://blog.cloudflare.com/introducing-autorag-on-cloudflare/ AutoRAG]&lt;br /&gt;
&lt;br /&gt;
==LLM for scoring/ranking==&lt;br /&gt;
* [https://arxiv.org/abs/2302.04166 GPTScore: Evaluate as You Desire]&lt;br /&gt;
* [https://arxiv.org/abs/2306.17563 Large Language Models are Effective Text Rankers with Pairwise Ranking Prompting]&lt;br /&gt;
* [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* [https://arxiv.org/abs/2407.02977 Large Language Models as Evaluators for Scientific Synthesis]&lt;br /&gt;
&lt;br /&gt;
=LLM Agents=&lt;br /&gt;
* See [[AI Agents]].&lt;br /&gt;
&lt;br /&gt;
=Interfaces=&lt;br /&gt;
==Chatbot Frontend==&lt;br /&gt;
===Web (code)===&lt;br /&gt;
* [https://docs.streamlit.io/develop/tutorials/llms/build-conversational-apps Steamlit]&lt;br /&gt;
* [https://docs.cohere.com/v2/docs/cohere-toolkit Cohere Toolkit] ([https://github.com/cohere-ai/cohere-toolkit code])&lt;br /&gt;
* [https://www.librechat.ai/ LibreChat]&lt;br /&gt;
* [https://github.com/open-webui/open-webui open-webui]&lt;br /&gt;
* [https://github.com/xjdr-alt/entropix/tree/main/ui entropix frontend UI]&lt;br /&gt;
&lt;br /&gt;
===Web (product)===&lt;br /&gt;
* [https://chatboxai.app/en Chatbox]&lt;br /&gt;
&lt;br /&gt;
===Desktop GUI===&lt;br /&gt;
* [https://anythingllm.com/ AnythingLLM] ([https://docs.anythingllm.com/ docs], [https://github.com/Mintplex-Labs/anything-llm code]): includes chat-with-docs, selection of LLM and vector db, etc.&lt;br /&gt;
&lt;br /&gt;
==Alternative Text Chatbot UI==&lt;br /&gt;
* [https://generative.ink/posts/loom-interface-to-the-multiverse/ Loom] provides a sort of tree-like structure for LLM coming up with branched writings.&lt;br /&gt;
* [https://www.lesswrong.com/posts/JHsfMWtwxBGGTmb8A/pantheon-interface The Pantheon Interface] is a new idea for how to interact with LLMs ([https://pantheon.chat/ live instance], [https://github.com/nickkeesG/Pantheon code]). In a traditional interaction, you prompt the bot and it replies in a turn-by-turn manner. Pantheon instead invites you to type out your thoughts, and various agents will asynchronously add comments or questions to spur along your brainstorming.&lt;br /&gt;
&lt;br /&gt;
==Conversational Audio Chatbot==&lt;br /&gt;
* Swift is a fast AI voice assistant ([https://github.com/ai-ng/swift code], [https://swift-ai.vercel.app/ live demo]) uses:&lt;br /&gt;
** [https://groq.com/ Groq] cloud running [https://github.com/openai/whisper OpenAI Whisper] for fast speech transcription.&lt;br /&gt;
** [https://cartesia.ai/ Cartesia] [https://cartesia.ai/sonic Sonic] for fast speech synthesis&lt;br /&gt;
** [https://www.vad.ricky0123.com/ VAD] to detect when user is talking&lt;br /&gt;
** [https://vercel.com/ Vercel] for app deployment&lt;br /&gt;
* [https://github.com/rtvi-ai RTVI-AI] ([https://github.com/rtvi-ai/rtvi-web-demo code], [https://demo-gpu.rtvi.ai/ demo]), uses:&lt;br /&gt;
** [https://groq.com/ Groq]&lt;br /&gt;
** [https://llama.meta.com/ Llama 3.1]&lt;br /&gt;
** [https://www.daily.co/ai/ Daily]&lt;br /&gt;
** [https://github.com/rtvi-ai RTVI ]&lt;br /&gt;
* [https://github.com/mezbaul-h/june June]: Local Voice Chatbot&lt;br /&gt;
** [https://ollama.com/ Ollama]&lt;br /&gt;
** [https://huggingface.co/docs/transformers/en/tasks/asr Hugging Face Transformers] (for speech recognition)&lt;br /&gt;
** [https://github.com/coqui-ai/TTS Coqui TTS Toolkit]&lt;br /&gt;
* [https://kyutai.org/ kyutai] Moshi chatbot ([https://us.moshi.chat/ demo])&lt;br /&gt;
* [https://arxiv.org/abs/2408.16725 Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming] ([https://huggingface.co/gpt-omni/mini-omni model], [https://github.com/gpt-omni/mini-omni code], [https://huggingface.co/spaces/gradio/omni-mini demo])&lt;br /&gt;
* 2024-09Sep-11: [https://huggingface.co/ICTNLP/Llama-3.1-8B-Omni Llama-3.1-8B-Omni] ([https://github.com/ictnlp/LLaMA-Omni code]), enabling end-to-end speech.&lt;br /&gt;
* [https://x.com/AIatMeta/status/1847383580269510670 2024-10Oct-18]: Meta [https://speechbot.github.io/spiritlm/ Spirit LM]: open source multimodal language model that freely mixes text and speech&lt;br /&gt;
* 2025-02Feb-28: [https://www.sesame.com/ Sesame] ([https://www.sesame.com/research/crossing_the_uncanny_valley_of_voice#demo demo])&lt;br /&gt;
&lt;br /&gt;
===Turn Detection===&lt;br /&gt;
* 2025-03: [https://github.com/pipecat-ai/smart-turn Smart Turn]: Open-source &lt;br /&gt;
&lt;br /&gt;
===Related Research===&lt;br /&gt;
* [https://arxiv.org/abs/2408.02622 Language Model Can Listen While Speaking]&lt;br /&gt;
&lt;br /&gt;
===Commercial Systems===&lt;br /&gt;
* [https://heypi.com/talk HeyPi Talk]&lt;br /&gt;
* [https://vapi.ai/ Vapi]&lt;br /&gt;
* [https://callannie.ai/ Call Annie]&lt;br /&gt;
* [https://www.bland.ai Bland AI]&lt;br /&gt;
* [https://deepgram.com/ DeepGram Voice AI]&lt;br /&gt;
* [https://www.sesame.com/ Sesame] ([https://www.sesame.com/research/crossing_the_uncanny_valley_of_voice#demo demo])&lt;br /&gt;
&lt;br /&gt;
=Speech Recognition (ASR) and Transcription=&lt;br /&gt;
==Lists==&lt;br /&gt;
* [https://huggingface.co/spaces/hf-audio/open_asr_leaderboard Open ASR Leaderboard]&lt;br /&gt;
&lt;br /&gt;
==Open Source==&lt;br /&gt;
* [https://github.com/mozilla/DeepSpeech DeepSpeech]&lt;br /&gt;
* [https://github.com/speechbrain/speechbrain speechbrain]&lt;br /&gt;
* [https://github.com/kaldi-asr/kaldi/blob/master/README.md Kaldi]&lt;br /&gt;
* wav2vec 2.0&lt;br /&gt;
** [https://arxiv.org/abs/2104.01027 Paper: Robust wav2vec 2.0: Analyzing Domain Shift in Self-Supervised Pre-Training]&lt;br /&gt;
* Whisper&lt;br /&gt;
** [https://huggingface.co/openai/whisper-medium.en Whisper medium.en]&lt;br /&gt;
** [https://github.com/m-bain/whisperX WhisperX] (includes word-level timestamps and speaker diarization)&lt;br /&gt;
** [https://huggingface.co/mlx-community/distil-whisper-large-v3 Distil Large v3 with MLX]&lt;br /&gt;
** 2024-10: [https://huggingface.co/ylacombe/whisper-large-v3-turbo whisper-large-v3-turbo] distillation ([https://huggingface.co/spaces/hf-audio/whisper-large-v3-turbo demo], [https://github.com/openai/whisper/actions/runs/11111568226 code])&lt;br /&gt;
* [https://huggingface.co/spaces/hf-audio/open_asr_leaderboard Nvidia Canary 1B]&lt;br /&gt;
* [https://developer.nvidia.com/blog/accelerating-leaderboard-topping-asr-models-10x-with-nvidia-nemo/ 2024-09]: Nvidia [https://docs.nvidia.com/nemo-framework/user-guide/latest/nemotoolkit/asr/models.html NeMo]&lt;br /&gt;
* 2024-10: [https://www.rev.ai/ Rev AI] [https://huggingface.co/Revai models] for [https://huggingface.co/Revai/reverb-asr transcription] and [https://huggingface.co/Revai/reverb-diarization-v2 diarization]&lt;br /&gt;
* 2024-10: [https://github.com/usefulsensors/moonshine Moonshine] (optimized for resource-constrained devices)&lt;br /&gt;
* 2025-05: [https://huggingface.co/nvidia/parakeet-tdt-0.6b-v2 Parakeet TDT 0.6B V2]&lt;br /&gt;
* [https://x.com/kyutai_labs/status/1925840420187025892 2025-05]: [https://kyutai.org/ Kyutai] [https://unmute.sh/ Unmute]&lt;br /&gt;
* [https://x.com/cohere/status/2037159129345614174?s=20 2026-03]: [https://cohere.com/blog/transcribe Cohere Transcribe]&lt;br /&gt;
&lt;br /&gt;
==In Browser==&lt;br /&gt;
* [https://huggingface.co/spaces/Xenova/whisper-word-level-timestamps Whisper Timestamped]: Multilingual speech recognition with word-level timestamps, running locally in browser&lt;br /&gt;
&lt;br /&gt;
==Phrase Endpointing and Voice Activity Detection (VAD)==&lt;br /&gt;
I.e. how to determine when user is done talking, and bot should respond?&lt;br /&gt;
* [https://x.com/kwindla/status/1831364419261268017 Notes]&lt;br /&gt;
** [https://demo.dailybots.ai/ Test settings]&lt;br /&gt;
** [https://github.com/pipecat-ai/pipecat/blob/main/src/pipecat/vad/vad_analyzer.py code]&lt;br /&gt;
** [https://github.com/snakers4/silero-vad Silero VAD repo]&lt;br /&gt;
&lt;br /&gt;
==Audio Cleanup==&lt;br /&gt;
* [https://krisp.ai/ Krisp AI]: Noise cancellation, meeting summary, etc.&lt;br /&gt;
&lt;br /&gt;
==Auto Video Transcription==&lt;br /&gt;
* [https://www.translate.mom/ TranslateMom]&lt;br /&gt;
* [https://github.com/abus-aikorea/voice-pro Voice-Pro]: YouTube downloader, speech separation, transcription, translation, TTS, and voice cloning toolkit for creators&lt;br /&gt;
&lt;br /&gt;
=Text-to-speech (TTS)=&lt;br /&gt;
==Open Source==&lt;br /&gt;
* [https://github.com/huggingface/parler-tts Parler TTS] ([https://huggingface.co/spaces/parler-tts/parler_tts demo])&lt;br /&gt;
* [https://github.com/DigitalPhonetics/IMS-Toucan Toucan] ([https://huggingface.co/spaces/Flux9665/MassivelyMultilingualTTS demo])&lt;br /&gt;
* [https://tts.themetavoice.xyz/ MetaVoice] ([https://github.com/metavoiceio/metavoice-src github])&lt;br /&gt;
* [https://github.com/2noise/ChatTTS ChatTTS]&lt;br /&gt;
* [https://www.camb.ai/ Camb.ai] [https://github.com/Camb-ai/MARS5-TTS MARS5-TTS]&lt;br /&gt;
* [https://github.com/coqui-ai/TTS Coqui TTS Toolkit]&lt;br /&gt;
* Fish Speech 1.4: multi-lingual, can clone voices ([https://x.com/reach_vb/status/1833801060659372071 video], [https://huggingface.co/fishaudio/fish-speech-1.4 weights], [https://huggingface.co/spaces/fishaudio/fish-speech-1 demo])&lt;br /&gt;
* [https://huggingface.co/SWivid/F5-TTS F5-TTS] ([https://huggingface.co/spaces/mrfakename/E2-F5-TTS demo]): cloning, emotion, etc.&lt;br /&gt;
* [https://huggingface.co/amphion/MaskGCT MaskGCT] ([https://huggingface.co/spaces/amphion/maskgct demo])&lt;br /&gt;
* [https://arxiv.org/abs/2312.09911 Amphion: An Open-Source Audio, Music and Speech Generation Toolkit] ([https://github.com/open-mmlab/Amphion code])&lt;br /&gt;
* [https://www.zyphra.com/ Zyphra] [https://huggingface.co/Zyphra/Zonos-v0.1-hybrid Zonos]&lt;br /&gt;
* [https://github.com/fishaudio/fish-speech Fish Speech] (includes voice cloning)&lt;br /&gt;
* [https://canopylabs.ai/ Canopy] [https://huggingface.co/collections/canopylabs/orpheus-tts-67d9ea3f6c05a941c06ad9d2 Orpheus] 3B&lt;br /&gt;
* Canopy [https://canopylabs.ai/releases/orpheus_can_speak_any_language Orpheus Multilingual]&lt;br /&gt;
* [https://narilabs.org/ Nari Labs] [https://github.com/nari-labs/dia Dia]&lt;br /&gt;
* [https://kyutai.org/ Kyutai] [https://kyutai.org/next/tts TTS] [https://unmute.sh/ Unmute]&lt;br /&gt;
* [https://github.com/resemble-ai/chatterbox Chatterbox TTS] ([https://huggingface.co/spaces/ResembleAI/Chatterbox try])&lt;br /&gt;
* [https://play.ai/ Play AI] [https://github.com/playht/PlayDiffusion PlayDiffusion] ([https://huggingface.co/spaces/PlayHT/PlayDiffusion demo], [https://x.com/_mfelfel/status/1929586464125239589 example])&lt;br /&gt;
* Mistral [https://mistral.ai/news/voxtral Voxtral]&lt;br /&gt;
* Kitten TTS ([https://github.com/KittenML/KittenTTS github], [https://huggingface.co/KittenML/kitten-tts-nano-0.1 hf]) 15M (fast, light-weight)&lt;br /&gt;
* Microsoft [https://microsoft.github.io/VibeVoice/ VibeVoice] 1.5B&lt;br /&gt;
* [https://x.com/hume_ai/status/2031401003078062578?s=20 2026-03]: Huma AI [https://huggingface.co/collections/HumeAI/tada TADA]&lt;br /&gt;
* [https://x.com/FishAudio/status/2031411140820152560?s=20 2026-03]: [https://huggingface.co/fishaudio/s2-pro Fish Audio S2]&lt;br /&gt;
&lt;br /&gt;
==Cloud==&lt;br /&gt;
* [https://elevenlabs.io/ Elevenlabs] ($50/million characters)&lt;br /&gt;
** [https://elevenlabs.io/voice-isolator voice isolator]&lt;br /&gt;
* [https://cartesia.ai/ Cartesia] [https://cartesia.ai/sonic Sonic]&lt;br /&gt;
* [https://neets.ai/ Neets AI] ($1/million characters)&lt;br /&gt;
* Hailuo AI T2A-01-HD ([https://www.hailuo.ai/audio try], [https://intl.minimaxi.com/document/platform%20introduction?key=66701c8e1d57f38758d58198 API])&lt;br /&gt;
* [https://www.hume.ai/ Hume] (can set emotion, give acting directions, etc.)&lt;br /&gt;
&lt;br /&gt;
=Text-to-audio=&lt;br /&gt;
* 2024-12: [https://tangoflux.github.io/ TangoFlux]: [https://arxiv.org/abs/2412.21037 Super Fast and Faithful Text to Audio Generation with Flow Matching and Clap-Ranked Preference Optimization] ([https://github.com/declare-lab/TangoFlux code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10522 AudioX: Diffusion Transformer for Anything-to-Audio Generation]&lt;br /&gt;
&lt;br /&gt;
=Vision=&lt;br /&gt;
* [https://github.com/google/langfun Langfun] library as a means of converting images into structured output.&lt;br /&gt;
* See also: [[AI_tools#Multimodal| Multimodal open-weights models]]&lt;br /&gt;
&lt;br /&gt;
==Visual Models==&lt;br /&gt;
* [https://openai.com/index/clip/ CLIP]&lt;br /&gt;
* [https://arxiv.org/abs/2303.15343 Siglip]&lt;br /&gt;
* [https://github.com/roboflow/supervision Supervision]&lt;br /&gt;
* [https://arxiv.org/abs/2311.06242 Florence-2]&lt;br /&gt;
* Nvidia [https://github.com/NVlabs/MambaVision MambaVision]&lt;br /&gt;
* Meta [https://about.meta.com/realitylabs/codecavatars/sapiens Sapiens: Foundation for Human Vision Models] (video input, can infer segmentation, pose, depth-map, and surface normals)&lt;br /&gt;
&lt;br /&gt;
==Depth==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.09414 Depth Anything V2] ([https://github.com/DepthAnything/Depth-Anything-V2 code])&lt;br /&gt;
&lt;br /&gt;
==Superresolution==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2311.17643 Thera: Aliasing-Free Arbitrary-Scale Super-Resolution with Neural Heat Fields] ([https://github.com/prs-eth/thera code], [https://huggingface.co/spaces/prs-eth/thera use])&lt;br /&gt;
&lt;br /&gt;
==Related==&lt;br /&gt;
* 2019-11: [https://arxiv.org/abs/1911.11763 SuperGlue: Learning Feature Matching with Graph Neural Networks] ([https://huggingface.co/docs/transformers/main/en/model_doc/superglue hf])&lt;br /&gt;
&lt;br /&gt;
=Embedding=&lt;br /&gt;
* [https://www.marktechpost.com/2024/07/28/a-comparison-of-top-embedding-libraries-for-generative-ai/ A Comparison of Top Embedding Libraries for Generative AI]&lt;br /&gt;
* [https://x.com/OfficialLoganK/status/2031411916489298156?s=20 2026-03]: [https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-embedding-2/ Gemini Embedding 2]&lt;br /&gt;
* [https://x.com/mixedbreadai/status/2032127466081567106?s=20 2026-03]: [https://www.mixedbread.com/ Mixedbread] Wholembed v3&lt;br /&gt;
&lt;br /&gt;
==Text Embedding==&lt;br /&gt;
* 2024-12: [https://huggingface.co/blog/modernbert modernBERT]&lt;br /&gt;
* 2025-02: [https://huggingface.co/chandar-lab/NeoBERT NeoBERT] ([https://arxiv.org/abs/2502.19587 preprint])&lt;br /&gt;
* 2025-03: [https://developers.googleblog.com/en/gemini-embedding-text-model-now-available-gemini-api/ gemini-embedding-exp-03-07]&lt;br /&gt;
&lt;br /&gt;
==Image Embedding==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18593 Diffusion Autoencoders are Scalable Image Tokenizers] ([https://yinboc.github.io/dito/ project], [https://github.com/yinboc/dito code])&lt;br /&gt;
&lt;br /&gt;
=Time Series=&lt;br /&gt;
* [https://github.com/TDAmeritrade/stumpy Stumpy]: Python library, uses near-match subsequences for similarity and forecasting&lt;br /&gt;
* [https://arxiv.org/abs/1912.09363 Temporal Fusion Transformers for Interpretable Multi-horizon Time Series Forecasting]&lt;br /&gt;
* [https://arxiv.org/abs/2209.00905 From latent dynamics to meaningful representations]&lt;br /&gt;
* [https://arxiv.org/abs/2209.10705 Review of Time Series Forecasting Methods and Their Applications to Particle Accelerators]&lt;br /&gt;
* [https://arxiv.org/abs/2310.01728 Time-LLM: Time Series Forecasting by Reprogramming Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.10688 A decoder-only foundation model for time-series forecasting]&lt;br /&gt;
* [https://arxiv.org/abs/2310.03589 TimeGPT-1]&lt;br /&gt;
* [https://arxiv.org/abs/2402.02592 Unified Training of Universal Time Series Forecasting Transformers]&lt;br /&gt;
* [https://arxiv.org/abs/2407.10240 xLSTMTime : Long-term Time Series Forecasting With xLSTM]&lt;br /&gt;
* Salesforce: [https://arxiv.org/abs/2410.10469 Moirai-MoE: Empowering Time Series Foundation Models with Sparse Mixture of Experts] ([https://github.com/SalesforceAIResearch/uni2ts/tree/main/project/moirai-moe-1 code], [https://huggingface.co/collections/Salesforce/moirai-r-models-65c8d3a94c51428c300e0742 weights], [https://www.salesforce.com/blog/time-series-morai-moe/ blog])&lt;br /&gt;
* IBM [https://huggingface.co/docs/transformers/en/model_doc/patchtsmixer PatchTSMixer] and [https://huggingface.co/docs/transformers/en/model_doc/patchtst PatchTST] (being [https://research.ibm.com/blog/time-series-AI-transformers used] for particle accelerators)&lt;br /&gt;
* 2026-02: Google [https://research.google/blog/a-decoder-only-foundation-model-for-time-series-forecasting/ TimesFM]&lt;br /&gt;
&lt;br /&gt;
==Control==&lt;br /&gt;
* [https://arxiv.org/abs/2402.15989 PIDformer: Transformer Meets Control Theory]&lt;br /&gt;
&lt;br /&gt;
==Forecasting==&lt;br /&gt;
* Meta [https://facebookresearch.github.io/Kats/ Kats] ([https://github.com/facebookresearch/Kats code]): Forecasting (ARIMA, Prophet, Holt Winters, VAR), detection, feature extraction, simulation&lt;br /&gt;
* [https://arxiv.org/abs/2410.18959 Context is Key: A Benchmark for Forecasting with Essential Textual Information]&lt;br /&gt;
&lt;br /&gt;
==Anomaly Detection==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.05440 Can LLMs Understand Time Series Anomalies?] ([https://github.com/rose-stl-lab/anomllm code])&lt;br /&gt;
&lt;br /&gt;
=Data=&lt;br /&gt;
* See also: [[Data_Extraction#Data_Scraping| Data Scraping]] and [[Data_Extraction#Document_Parsing| Document Parsing]]&lt;br /&gt;
==Vector Database==&lt;br /&gt;
===Open Source===&lt;br /&gt;
* [https://milvus.io/ milvus] (open source with paid cloud option)&lt;br /&gt;
* [https://qdrant.tech/ Qdrant] (open source with paid cloud option)&lt;br /&gt;
* [https://vespa.ai/ Vespa] (open source with paid cloud option)&lt;br /&gt;
* [https://www.trychroma.com/ chroma]&lt;br /&gt;
* [https://www.llamaindex.ai/ LlamaIndex]&lt;br /&gt;
* [https://github.com/asg017/sqlite-vec/tree/main sqlite-vec]&lt;br /&gt;
&lt;br /&gt;
===Commercial cloud===&lt;br /&gt;
* [https://archive.pinecone.io/lp/vector-database/ pinecone]&lt;br /&gt;
* [https://weaviate.io/products weaviate]&lt;br /&gt;
&lt;br /&gt;
===MySQL===&lt;br /&gt;
* MySQL does not traditionally have support, but:&lt;br /&gt;
** [https://planetscale.com/blog/planetscale-is-bringing-vector-search-and-storage-to-mysql PlanetScale] is working on it&lt;br /&gt;
** [https://github.com/stephenc222/mysql_vss mysql_vss] ([https://medium.com/@stephenc211/enhancing-mysql-searches-with-vector-embeddings-11f183932851 discussion])&lt;br /&gt;
** [https://www.pingcap.com/tidb-serverless/ tibd] ([https://www.pingcap.com/article/mysql-vector-search-powering-the-future-of-ai-applications/ discussion])&lt;br /&gt;
&lt;br /&gt;
==Database with Search==&lt;br /&gt;
* [https://typesense.org/ Typesense] ([https://github.com/typesense/typesense code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[Data Extraction]]&lt;br /&gt;
** [[AI compute]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[AI understanding]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Human_brain&amp;diff=8761</id>
		<title>Human brain</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Human_brain&amp;diff=8761"/>
		<updated>2026-03-26T16:28:52Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Brain signal decoding */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Why brain is as it is=&lt;br /&gt;
* 2025-06: [https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(24)00319-X The metabolic costs of cognition]&lt;br /&gt;
&lt;br /&gt;
=How Brain Works=&lt;br /&gt;
==Predictive Coding==&lt;br /&gt;
* 2005-04: [https://royalsocietypublishing.org/doi/10.1098/rstb.2005.1622?utm_source=chatgpt.com A theory of cortical responses]&lt;br /&gt;
* 2014-09: [https://www.frontiersin.org/journals/human-neuroscience/articles/10.3389/fnhum.2014.00666/full Visual mismatch negativity: a predictive coding view]&lt;br /&gt;
* 2015-01: [https://www.sciencedirect.com/science/article/pii/S089662731401099X Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels]&lt;br /&gt;
* 2016-11: [https://www.sciencedirect.com/science/article/pii/S0896627316306997 Mismatch Receptive Fields in Mouse Visual Cortex]&lt;br /&gt;
* 2018-03: [https://www.nature.com/articles/s41598-018-21407-9 Frontal cortex function as derived from hierarchical predictive coding]&lt;br /&gt;
* 2024-02: [https://www.sciencedirect.com/science/article/pii/S0149763423004426 The empirical status of predictive coding and active inference]&lt;br /&gt;
&lt;br /&gt;
=Understanding=&lt;br /&gt;
* [https://arxiv.org/abs/2501.02950 Key-value memory in the brain]&lt;br /&gt;
* [https://helper.ipam.ucla.edu/publications/mac2024/mac2024_20152.pdf The cost of brain state transitions]&lt;br /&gt;
&lt;br /&gt;
==Brain mapping==&lt;br /&gt;
* 2024-05: [https://www.science.org/doi/10.1126/science.adk4858 A petavoxel fragment of human cerebral cortex reconstructed at nanoscale resolution] ([https://www.nature.com/articles/d41586-024-01387-9#ref-CR1 media summary])&lt;br /&gt;
* 2024-10: [https://www.nature.com/articles/s41586-024-07558-y Neuronal wiring diagram of an adult brain] ([https://www.nytimes.com/2024/10/02/science/fruit-fly-brain-mapped.html media summary]); 140,000 neurons in fruit fly brain&lt;br /&gt;
* 2024-12: [https://e11.bio/news/roadmap A roadmap to scale connectomics to entire mammalian brains]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08840-3 Functional connectomics reveals general wiring rule in mouse visual cortex] ([https://www.nature.com/articles/d41586-025-01088-x?utm_source=x&amp;amp;utm_medium=social&amp;amp;utm_campaign=nature&amp;amp;linkId=13897098 media summary])&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41586-025-08985-1 Light-microscopy-based connectomic reconstruction of mammalian brain tissue] ([https://research.google/blog/a-new-light-on-neural-connections/ blog])&lt;br /&gt;
&lt;br /&gt;
===Related===&lt;br /&gt;
* [https://v2.virtualflybrain.org 3D visualization of adult fruit fly brain]&lt;br /&gt;
&lt;br /&gt;
==Brain signal decoding==&lt;br /&gt;
* 2022-11: [https://www.biorxiv.org/content/10.1101/2022.11.18.517004v2.full.pdf High-resolution image reconstruction with latent diffusion models from human brain activity]&lt;br /&gt;
* 2023-08: [https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3002176%20 Music can be reconstructed from human auditory cortex activity using nonlinear decoding models] (intracranial EEG)&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.14030 DeWave: Discrete EEG Waves Encoding for Brain Dynamics to Text Translation] (external EEG)&lt;br /&gt;
* 2023-09: [https://www.biorxiv.org/content/10.1101/2023.09.12.557460v1 BrainLM: A foundation model for brain activity recordings]&lt;br /&gt;
* 2023-10: [https://ai.meta.com/blog/brain-ai-image-decoding-meg-magnetoencephalography/ Toward a real-time decoding of images from brain activity] (MEG)&lt;br /&gt;
* 2024-06: [https://www.biorxiv.org/content/10.1101/2024.06.04.596589v1.full.pdf PAM: Predictive Attention Mechanism for Neural Decoding of Visual Perception]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.07595 Scaling Law in Neural Data: Non-Invasive Speech Decoding with 175 Hours of EEG Data] (EEG)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15322v2 Scaling laws for decoding images from brain activity] (EEG)&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/brain-to-text-decoding-a-non-invasive-approach-via-typing/ Brain-to-Text Decoding: A Non-invasive Approach via Typing]&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/from-thought-to-action-how-a-hierarchy-of-neural-dynamics-supports-language-production/ From Thought to Action: How a Hierarchy of Neural Dynamics Supports Language Production]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41593-025-01905-6 A streaming brain-to-voice neuroprosthesis to restore naturalistic communication]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2508.18226 Disentangling the Factors of Convergence between Brains and Computer Vision Models] (fMRI and MEG)&lt;br /&gt;
&lt;br /&gt;
==Brain Signal Prediction==&lt;br /&gt;
* 20226-03: [https://ai.meta.com/research/publications/a-foundation-model-of-vision-audition-and-language-for-in-silico-neuroscience/ A foundation model of vision, audition, and language for in-silico neuroscience]&lt;br /&gt;
&lt;br /&gt;
==Whole Brain Emulation (WBE)==&lt;br /&gt;
* 2024-09: [https://www.nature.com/articles/s41586-024-07939-3 Connectome-constrained networks predict neural activity across the fly visual system]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15745 State of Brain Emulation Report 2025]&lt;br /&gt;
&lt;br /&gt;
=Computational Analysis=&lt;br /&gt;
&lt;br /&gt;
==Computational power of human brain==&lt;br /&gt;
* 2020-09: Joe Carlsmith: [https://www.openphilanthropy.org/research/how-much-computational-power-does-it-take-to-match-the-human-brain/ How Much Computational Power Does It Take to Match the Human Brain?]&lt;br /&gt;
&lt;br /&gt;
==Comparison to computer==&lt;br /&gt;
* [https://arxiv.org/abs/2208.12032 How (and Why) to Think that the Brain is Literally a Computer]&lt;br /&gt;
* [https://www.nature.com/articles/s42256-024-00925-4 Contextual feature extraction hierarchies converge in large language models and the brain] ([https://techxplore.com/news/2024-12-llms-brain-advance.html LLMs are becoming more brain-like as they advance])&lt;br /&gt;
&lt;br /&gt;
==Biological vs. artificial neuron==&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S0896627321005018 Single cortical neurons as deep artificial neural networks]: Each biological neuron can be simulated using DNN of 5-8 layers&lt;br /&gt;
* [https://arxiv.org/abs/2305.12471 Mapping Biological Neuron Dynamics into an Interpretable Two-layer Artificial Neural Network]&lt;br /&gt;
&lt;br /&gt;
==Data processing==&lt;br /&gt;
* [https://pmc.ncbi.nlm.nih.gov/articles/PMC1564115/ How Much the Eye Tells the Brain]&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S1364661313001277 Representational geometry: integrating cognition, computation, and the brain]&lt;br /&gt;
* [https://www.nature.com/articles/s41586-024-07522-w Language is primarily a tool for communication rather than thought]&lt;br /&gt;
* [https://www.openread.academy/en/paper/reading?corpusId=513306465 The Unbearable Slowness of Being: Why do we live at 10 bits/s?] ([https://arxiv.org/abs/2408.10234 preprint])&lt;br /&gt;
&lt;br /&gt;
==Extract manifold/geometry==&lt;br /&gt;
* [https://www.science.org/doi/10.1126/science.adk8261 Selection of experience for memory by hippocampal sharp wave ripples]&lt;br /&gt;
&lt;br /&gt;
=Comparisons=&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.08708 Consciousness in Artificial Intelligence: Insights from the Science of Consciousness]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.02325 Are Biological Systems More Intelligent Than Artificial Intelligence?]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
** 2022-03: [https://www.nature.com/articles/s41593-022-01026-4 Shared computational principles for language processing in humans and deep language models]&lt;br /&gt;
** 2024-03: [https://www.nature.com/articles/s41467-024-46631-y Alignment of brain embeddings and artificial contextual embeddings in natural language points to common geometric patterns]&lt;br /&gt;
** 2025-03: [https://www.nature.com/articles/s41562-025-02105-9 A unified acoustic-to-speech-to-language embedding space captures the neural basis of natural language processing in everyday conversations]&lt;br /&gt;
* 2025-05: [https://ai.meta.com/research/publications/emergence-of-language-in-the-developing-brain/ Emergence of Language in the Developing Brain]&lt;br /&gt;
&lt;br /&gt;
==Analogies==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41562-025-02359-3 Shared sensitivity to data distribution during learning in humans and transformer networks]&lt;br /&gt;
===Speed-accuracy trade-off vs. Inference-compute===&lt;br /&gt;
* 2007: [https://psycnet.apa.org/doi/10.1037/0096-3445.136.2.217 Focusing the spotlight: individual differences in visual attention control]&lt;br /&gt;
* 2014-07: [https://www.frontiersin.org/journals/neuroscience/articles/10.3389/fnins.2014.00150/full The speed-accuracy tradeoff: history, physiology, methodology, and behavior]&lt;br /&gt;
&lt;br /&gt;
=Simulate Brain=&lt;br /&gt;
* 2023-09: [https://spj.science.org/doi/10.34133/icomputing.0055 The Digital Twin Brain: A Bridge between Biological and Artificial Intelligence]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s43588-024-00731-3 Simulation and assimilation of the digital human brain] ([https://arxiv.org/abs/2211.15963 preprint], [https://github.com/DTB-consortium/Digital_twin_brain-open code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2507.22229 TRIBE: TRImodal Brain Encoder for whole-brain fMRI response prediction]&lt;br /&gt;
&lt;br /&gt;
==See Also==&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|Simulate Humans (using LLM)]]&lt;br /&gt;
&lt;br /&gt;
=Bio-brain Inspirations for AI=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
&lt;br /&gt;
=Theories of Consciousness=&lt;br /&gt;
* [https://www.consciousnessatlas.com/ Consciousness Atlas]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|LLM Simulate Humans]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8760</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8760"/>
		<updated>2026-03-24T22:39:08Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
** [https://x.com/maxescu/status/2036434854435315868?s=20 Cinematic scenes] (3.5m, comedy, [https://lumalabs.ai/uni-1 Luma Uni-1 Agent])&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8759</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8759"/>
		<updated>2026-03-24T19:22:25Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Specific */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.20179 AI Agents Can Already Autonomously Perform Experimental High Energy Physics]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Personalities==&lt;br /&gt;
* 2026-03: [https://github.com/msitarzewski/agency-agents The Agency: AI Specialists Ready to Transform Your Workflow]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8758</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8758"/>
		<updated>2026-03-24T19:12:11Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
** [https://x.com/Alterverse_AI/status/2036434608137343111?s=20 Monkey&amp;#039;s Paw] (5m)&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8757</id>
		<title>AI predictions</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8757"/>
		<updated>2026-03-24T18:59:56Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Economic and Political */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Capability Scaling=&lt;br /&gt;
* 2019-03: Rich Sutton: [https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf The Bitter Lesson]&lt;br /&gt;
* 2020-09: Ajeya Cotra: [https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines Draft report on AI timelines]&lt;br /&gt;
* 2022-01: gwern: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis]&lt;br /&gt;
* 2023-05: Richard Ngo: [https://www.lesswrong.com/posts/BoA3agdkAzL6HQtQP/clarifying-and-predicting-agi Clarifying and predicting AGI]&lt;br /&gt;
* 2024-06: Aidan McLaughlin: [https://yellow-apartment-148.notion.site/AI-Search-The-Bitter-er-Lesson-44c11acd27294f4495c3de778cd09c8d AI Search: The Bitter-er Lesson]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.14499 Measuring AI Ability to Complete Long Tasks Measuring AI Ability to Complete Long Tasks]&lt;br /&gt;
** 2025-04: [https://peterwildeford.substack.com/p/forecaster-reacts-metrs-bombshell Forecaster reacts: METR&amp;#039;s bombshell paper about AI acceleration] New data supports an exponential AI curve, but lots of uncertainty remains&lt;br /&gt;
** 2025-04: AI Digest: [https://theaidigest.org/time-horizons A new Moore&amp;#039;s Law for AI agents]&lt;br /&gt;
[[Image:GmZHL8xWQAAtFlF.jpeg|450px]]&lt;br /&gt;
* 2025-04: [https://epoch.ai/blog/trends-in-ai-supercomputers Trends in AI Supercomputers] ([https://arxiv.org/abs/2504.16026 preprint])&lt;br /&gt;
* [https://ai-timeline.org/ The Road to AGI] (timeline visualization)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09677 The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs]&lt;br /&gt;
* 2025-09: [https://www.julian.ac/blog/2025/09/27/failing-to-understand-the-exponential-again/ Failing to Understand the Exponential, Again]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/rRbDNQLfihiHbXytf/distinguish-between-inference-scaling-and-larger-tasks-use Distinguish between inference scaling and &amp;quot;larger tasks use more compute&amp;quot;]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03992 Measuring AI R&amp;amp;D Automation] ([https://astrangeattractor.substack.com/p/measuring-ai-r-and-d-automation?triedRedirect=true blog])&lt;br /&gt;
&lt;br /&gt;
==Scaling Laws==&lt;br /&gt;
See: [[AI_understanding#Scaling_Laws|Scaling Laws]]&lt;br /&gt;
&lt;br /&gt;
==AGI Achievable==&lt;br /&gt;
* Yoshua Bengio: [https://arxiv.org/abs/2310.17688 Managing extreme AI risks amid rapid progress]&lt;br /&gt;
* Leopold Aschenbrenner: [https://situational-awareness.ai/from-gpt-4-to-agi/#Counting_the_OOMs Situational Awareness: Counting the OOMs]&lt;br /&gt;
* Richard Ngo: [https://medium.com/@richardcngo/visualizing-the-deep-learning-revolution-722098eb9c5 Visualizing the deep learning revolution]&lt;br /&gt;
* Katja Grace: [https://blog.aiimpacts.org/p/2023-ai-survey-of-2778-six-things Survey of 2,778 AI authors: six parts in pictures]&lt;br /&gt;
* Epoch AI: [https://epoch.ai/trends Machine Learning Trends]&lt;br /&gt;
* AI Digest: [https://theaidigest.org/progress-and-dangers How fast is AI improving?]&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/when-will-agi-arrive/ The case for AGI by 2030]&lt;br /&gt;
&lt;br /&gt;
==AGI Definition==&lt;br /&gt;
* 2023-11: Allan Dafoe, Shane Legg, et al.: [https://arxiv.org/abs/2311.02462 Levels of AGI for Operationalizing Progress on the Path to AGI]&lt;br /&gt;
* 2024-04: Bowen Xu: [https://arxiv.org/abs/2404.10731 What is Meant by AGI? On the Definition of Artificial General Intelligence]&lt;br /&gt;
* 2025-10: Dan Hendrycks et al.: [https://www.agidefinition.ai/paper.pdf A Definition of AGI]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07364 On the universal definition of intelligence]&lt;br /&gt;
&lt;br /&gt;
==Recursive Self Improvement (RSI)==&lt;br /&gt;
* 2026-02: [https://80000hours.org/articles/how-ai-driven-feedback-loops-could-make-things-very-crazy-very-fast/ How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
&lt;br /&gt;
==Progress Models==&lt;br /&gt;
From [http://yager-research.ca/2025/04/ai-impact-predictions/ AI Impact Predictions]:&lt;br /&gt;
&lt;br /&gt;
[[Image:AI impact models-2025 11 24.png|450px]]&lt;br /&gt;
&lt;br /&gt;
=Economic and Political=&lt;br /&gt;
* 2019-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3482150 The Impact of Artificial Intelligence on the Labor Market]&lt;br /&gt;
* 2020-06: [https://www.openphilanthropy.org/research/modeling-the-human-trajectory/ Modeling the Human Trajectory] (GDP)&lt;br /&gt;
* 2021-06: [https://www.openphilanthropy.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/ Report on Whether AI Could Drive Explosive Economic Growth]&lt;br /&gt;
* 2023-10: Marc Andreessen: [https://a16z.com/the-techno-optimist-manifesto/ The Techno-Optimist Manifesto]&lt;br /&gt;
* 2023-12: [https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html My techno-optimism]: &amp;quot;defensive acceleration&amp;quot; ([https://vitalik.eth.limo/index.html Vitalik Buterin])&lt;br /&gt;
* 2024-03: Noah Smith: [https://www.noahpinion.blog/p/plentiful-high-paying-jobs-in-the Plentiful, high-paying jobs in the age of AI: Comparative advantage is very subtle, but incredibly powerful.] ([https://x.com/liron/status/1768013030741475485 video])&lt;br /&gt;
* 2024-03: [https://doi.org/10.3386/w32255 Scenarios for the Transition to AGI] (AGI leads to wage collapse)&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-06: [https://www.frbsf.org/wp-content/uploads/AI-and-Growth-Aghion-Bunel.pdf AI and Growth: Where Do We Stand?]&lt;br /&gt;
* 2024-09: OpenAI [https://cdn.openai.com/global-affairs/openai-infra-economics-10.09.24.pdf Infrastructure is Destiny: Economic Returns on US Investment in Democratic AI]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/KFFaKu27FNugCHFmh/by-default-capital-will-matter-more-than-ever-after-agi By default, capital will matter more than ever after AGI] (L Rudolf L)&lt;br /&gt;
* 2025-01: [https://lukedrago.substack.com/p/the-intelligence-curse The Intelligence Curse]: With AGI, powerful actors will lose their incentives to invest in people&lt;br /&gt;
** Updated 2025-04: [https://intelligence-curse.ai/ The Intelligence Curse] (Luke Drago and Rudolf Laine)&lt;br /&gt;
*** [https://intelligence-curse.ai/pyramid/ Pyramid Replacement]&lt;br /&gt;
*** [https://intelligence-curse.ai/capital/ Capital, AGI, and Human Ambition]&lt;br /&gt;
*** [https://intelligence-curse.ai/defining/ Defining the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/shaping/ Shaping the Social Contract]&lt;br /&gt;
*** [https://intelligence-curse.ai/breaking/ Breaking the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/history/ History is Yours to Write]&lt;br /&gt;
* 2025-01: Microsoft: [https://blogs.microsoft.com/on-the-issues/2025/01/03/the-golden-opportunity-for-american-ai/ The Golden Opportunity for American AI]&lt;br /&gt;
* 2025-01: [https://www.maximum-progress.com/p/agi-will-not-make-labor-worthless AGI Will Not Make Labor Worthless]&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/global-affairs/ai-in-america-oai-economic-blueprint-20250113.pdf AI in America: OpenAI&amp;#039;s Economic Blueprint] ([https://openai.com/global-affairs/openais-economic-blueprint/ blog])&lt;br /&gt;
* 2025-01: [https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai How much economic growth from AI should we expect, how soon?]&lt;br /&gt;
* 2025-02: Morgan Stanley: [https://advisor.morganstanley.com/john.howard/documents/field/j/jo/john-howard/The_Humanoid_100_-_Mapping_the_Humanoid_Robot_Value_Chain.pdf The Humanoid 100: Mapping the Humanoid Robot Value Chain]&lt;br /&gt;
* 2025-02: [https://www.anthropic.com/news/the-anthropic-economic-index The Anthropic Economic Index]: [https://assets.anthropic.com/m/2e23255f1e84ca97/original/Economic_Tasks_AI_Paper.pdf Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11264 Strategic Wealth Accumulation Under Transformative AI Expectations]&lt;br /&gt;
* 2025-02: Tyler Cowen: [https://marginalrevolution.com/marginalrevolution/2025/02/why-i-think-ai-take-off-is-relatively-slow.html Why I think AI take-off is relatively slow]&lt;br /&gt;
* 2025-03: Epoch AI: [https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d Most AI value will come from broad automation, not from R&amp;amp;D]&lt;br /&gt;
** The primary economic impact of AI will be its ability to broadly automate labor&lt;br /&gt;
** Automating AI R&amp;amp;D alone likely won’t dramatically accelerate AI progress&lt;br /&gt;
** Fully automating R&amp;amp;D requires a very broad set of abilities&lt;br /&gt;
** AI takeoff will likely be diffuse and salient&lt;br /&gt;
* 2025-03: [https://www.anthropic.com/news/anthropic-economic-index-insights-from-claude-sonnet-3-7 Anthropic Economic Index: Insights from Claude 3.7 Sonnet]&lt;br /&gt;
* 2025-04: [https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality Will there be extreme inequality from AI?]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/research/impact-software-development Anthropic Economic Index: AI’s Impact on Software Development]&lt;br /&gt;
* 2025-05: [https://www.theguardian.com/books/2025/may/04/the-big-idea-can-we-stop-ai-making-humans-obsolete Better at everything: how AI could make human beings irrelevant]&lt;br /&gt;
* 2025-05: Forethought: [https://www.forethought.org/research/the-industrial-explosion The Industrial Explosion]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20273 Ten Principles of AI Agent Economics]&lt;br /&gt;
* 2025-07: [https://substack.com/home/post/p-167879696 What Economists Get Wrong about AI] They ignore innovation effects, use outdated capability assumptions, and miss the robotics revolution&lt;br /&gt;
* 2025-07: [https://www.nber.org/books-and-chapters/economics-transformative-ai/we-wont-be-missed-work-and-growth-era-agi We Won&amp;#039;t Be Missed: Work and Growth in the Era of AGI]&lt;br /&gt;
* 2025-07: [https://www.nber.org/papers/w34034 The Economics of Bicycles for the Mind]&lt;br /&gt;
* 2025-09: [https://conference.nber.org/conf_papers/f227491.pdf Genius on Demand: The Value of Transformative Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://peterwildeford.substack.com/p/ai-is-probably-not-a-bubble AI is probably not a bubble: AI companies have revenue, demand, and paths to immense value]&lt;br /&gt;
* 2025-11: [https://windowsontheory.org/2025/11/04/thoughts-by-a-non-economist-on-ai-and-economics/ Thoughts by a non-economist on AI and economics]&lt;br /&gt;
* 2025-11: [https://www.nber.org/papers/w34444 Artificial Intelligence, Competition, and Welfare]&lt;br /&gt;
* 2025-11: [https://www.anthropic.com/research/estimating-productivity-gains Estimating AI productivity gains from Claude conversations] (Anthropic)&lt;br /&gt;
* 2025-12: [https://benjamintodd.substack.com/p/how-ai-driven-feedback-loops-could How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
* 2025-12: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth] (Philip Trammell and Leopold Aschenbrenner)&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Anthropic Economic Index: new building blocks for understanding AI use]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/economic-index-primitives Anthropic Economic Index report: economic primitives]&lt;br /&gt;
* 2026-02: Nate Silver: [https://www.natesilver.net/p/the-singularity-wont-be-gentle The singularity won&amp;#039;t be gentle: If AI is even half as transformational as Silicon Valley assumes, politics will never be the same again]&lt;br /&gt;
* 2026-03: [https://www.anthropic.com/research/economic-index-march-2026-report Anthropic Economic Index report: Learning curves]&lt;br /&gt;
&lt;br /&gt;
==Job Loss==&lt;br /&gt;
* 2023-03: [https://arxiv.org/pdf/2303.10130 GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models]&lt;br /&gt;
** 2023-03: [https://www.livemint.com/news/world/these-jobs-are-most-at-risk-due-to-chatgpt-as-per-openai-study-11679358453267.html These jobs are most at risk due to ChatGPT, as per OpenAI study]&lt;br /&gt;
* 2023-08: [https://dx.doi.org/10.2139/ssrn.4527336 The Short-Term Effects of Generative Artificial Intelligence on Employment: Evidence from an Online Labor Market]&lt;br /&gt;
** [https://x.com/jburnmurdoch/status/1722938749519077688 Freelancer sector shrinking]&lt;br /&gt;
[[Image:F-kVQuvWkAAemkr.png|400px]]&lt;br /&gt;
* 2023-09: [https://global-uploads.webflow.com/64d5f73a7fc5e8a240310c4d/650a128a34386a1206b6506c_FINAL%20Briefing%20-%20Adoption%20of%20Automation%20and%20AI%20in%20the%20UK.pdf What drives UK firms to adopt AI and robotics, and what are the consequences for jobs?]&lt;br /&gt;
** [https://www.digitalinformationworld.com/2023/09/78-of-companies-say-ai-created-more-jobs.html 78% of Companies Say AI Created More Jobs]&lt;br /&gt;
* 2023-11: [https://theaipi.org/ai-interactive-map/ New Analysis Shows Over 20% of US Jobs Significantly Exposed to AI Automation In the Near Future]&lt;br /&gt;
* 2024-01: [https://techcrunch.com/2024/01/09/duolingo-cut-10-of-its-contractor-workforce-as-the-company-embraces-ai/ Duolingo cuts 10% of its contractor workforce as the company embraces AI]&lt;br /&gt;
* 2024-02: [https://www.pwc.com/gx/en/issues/c-suite-insights/the-leadership-agenda/gen-ai-is-a-tool-for-growth-not-just-efficiency.html#:~:text=One%20out%20of%20every%20four%20of%20the%204%2C702,to%20accomplish%20the%20same%20tasks%20with%20fewer%20workers Gen AI is a tool for growth, not just efficiency: Tech CEOs are investing to build their workforce and capitalise on new opportunities from generative AI. That’s a sharp contrast to how their peers view it.]&lt;br /&gt;
* 2024-04: [https://www.nytimes.com/2024/04/10/business/investment-banking-jobs-artificial-intelligence.html AI is Poised to Replace the Entry-Level Grunt Work of a Wall Street Career]&lt;br /&gt;
* 2024-07: [https://www.wired.com/story/ai-is-already-taking-jobs-in-the-video-game-industry/ AI Is Already Taking Jobs in the Video Game Industry]: A WIRED investigation finds that major players like Activision Blizzard, which recently laid off scores of workers, are using generative AI for game development&lt;br /&gt;
* 2024-08: [https://www.bbc.com/news/articles/c80e1gp9m9zo Klarna: AI lets us cut thousands of jobs - but pay more]&lt;br /&gt;
* 2025-01: [https://scholarspace.manoa.hawaii.edu/server/api/core/bitstreams/4f39375d-59c2-4c4a-b394-f3eed7858c80/content AI and Freelancers: Has the Inflection Point Arrived?]&lt;br /&gt;
* 2025-01: [https://www.aporiamagazine.com/p/yes-youre-going-to-be-replaced Yes, you&amp;#039;re going to be replaced: So much cope about AI]&lt;br /&gt;
* 2025-03: [https://commonplace.org/2025/03/20/will-ai-automate-away-your-job/ Will AI Automate Away Your Job? The time-horizon model explains the future of the technology]&lt;br /&gt;
* 2025-05: [https://www.forbes.com/sites/jackkelly/2025/05/04/its-time-to-get-concerned-klarna-ups-duolingo-cisco-and-many-other-companies-are-replacing-workers-with-ai/ It’s Time To Get Concerned, Klarna, UPS, Duolingo, Cisco, And Many Other Companies Are Replacing Workers With AI]&lt;br /&gt;
* 2025-05: [https://time.com/7289692/when-ai-replaces-workers/ What Happens When AI Replaces Workers?]&lt;br /&gt;
* 2025-05: [https://www.oxfordeconomics.com/resource/educated-but-unemployed-a-rising-reality-for-us-college-grads/ Educated but unemployed, a rising reality for US college grads] Structural shifts in tech hiring and the growing impact of AI are driving higher unemployment among recent college graduates&lt;br /&gt;
* 2025-05: NY Times: [https://www.nytimes.com/2025/05/30/technology/ai-jobs-college-graduates.html?unlocked_article_code=1.LE8.LlC6.eT5XcpA9hxC2&amp;amp;smid=url-share For Some Recent Graduates, the A.I. Job Apocalypse May Already Be Here] The unemployment rate for recent college graduates has jumped as companies try to replace entry-level workers with artificial intelligence&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/skills-ai-makes-valuable/ How not to lose your job to AI] The skills AI will make more valuable (and how to learn them)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06576 Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce]&lt;br /&gt;
[[Image:0dab4c86-882d-4095-9d12-d19684ed5184 675x680.png|300px]]&lt;br /&gt;
* 2025-07: Harvard Business Review: [https://hbr.org/2025/06/what-gets-measured-ai-will-automate What Gets Measured, AI Will Automate]&lt;br /&gt;
* 2025-08: [https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/ Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5560401 Performance or Principle: Resistance to Artificial Intelligence in the U.S. Labor Market]&lt;br /&gt;
* 2025-10: [https://www.siliconcontinent.com/p/the-ai-becker-problem The AI Becker problem: Who will train the next generation?]&lt;br /&gt;
* 2026-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6134506 AI, Automation, and Expertise]&lt;br /&gt;
* 2026-02: [https://arachnemag.substack.com/p/the-jevons-paradox-for-intelligence The Jevons Paradox for Intelligence: Fears of AI-induced job loss could not be more wrong]&lt;br /&gt;
* 2026-03: [https://www.dropbox.com/scl/fo/689u1g785x8jp6c8v1s21/AKxZ_N15vUxMA3PBtpbr5nM?dl=0&amp;amp;e=1&amp;amp;preview=2026.03.24+Bundles.pdf&amp;amp;rlkey=ottgcu71u1t4mhn6tblvatu8w&amp;amp;st=dj6k0x2o Weak Bundle, Strong Bundle:How AI Redraws Job Boundaries]&lt;br /&gt;
&lt;br /&gt;
==Productivity Impact==&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2026-02: [https://www.ft.com/content/4b51d0b4-bbfe-4f05-b50a-1d485d419dc5 The AI productivity take-off is finally visible] ([https://x.com/erikbryn/status/2023075588974735869?s=20 Erik Brynjolfsson])&lt;br /&gt;
** Businesses are finally beginning to reap some of AI&amp;#039;s benefits.&lt;br /&gt;
* 2026-02: New York Times: [https://www.nytimes.com/2026/02/18/opinion/ai-software.html The A.I. Disruption We’ve Been Waiting for Has Arrived]&lt;br /&gt;
&lt;br /&gt;
==National Security==&lt;br /&gt;
* 2025-04: Jeremie Harris and Edouard Harris: [https://superintelligence.gladstone.ai/ America’s Superintelligence Project]&lt;br /&gt;
&lt;br /&gt;
==AI Manhattan Project==&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-10: [https://thezvi.substack.com/p/ai-88-thanks-for-the-memos?open=false#%C2%A7thanks-for-the-memos-introduction-and-competitiveness White House Memo calls for action on AI]&lt;br /&gt;
* 2024-11: [https://www.uscc.gov/annual-report/2024-annual-report-congress 2024 Annual Report to Congress]: [https://www.reuters.com/technology/artificial-intelligence/us-government-commission-pushes-manhattan-project-style-ai-initiative-2024-11-19/ calls] for &amp;quot;Manhattan Project-style&amp;quot; effort&lt;br /&gt;
* 2025-05-29: [https://x.com/ENERGY/status/1928085878561272223 DoE Tweet]: &amp;quot;AI is the next Manhattan Project, and THE UNITED STATES WILL WIN. 🇺🇸&amp;quot;&lt;br /&gt;
* 2025-07: [https://epoch.ai/gradient-updates/how-big-could-an-ai-manhattan-project-get How big could an “AI Manhattan Project” get?]&lt;br /&gt;
&lt;br /&gt;
=Near-term=&lt;br /&gt;
* 2021-08: Daniel Kokotajlo: [https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like What 2026 looks like]&lt;br /&gt;
* 2025-02: Sam Altman: [https://blog.samaltman.com/three-observations Three Observations]&lt;br /&gt;
*# The intelligence of an AI model roughly equals the log of the resources used to train and run it.&lt;br /&gt;
*# The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use&lt;br /&gt;
*# The socioeconomic value of linearly increasing intelligence is super-exponential in nature&lt;br /&gt;
* 2025-03: [https://www.pathwaysai.org/p/glimpses-of-ai-progess Glimpses of AI Progress: Mental models for fast times]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41598-025-92190-7 Navigating artificial general intelligence development: societal, technological, ethical, and brain-inspired pathways]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
** 2025-07: Video: [https://www.youtube.com/watch?v=5KVDDfAkRgc Are We 3 Years From AI Disaster? A Rigorous Forecast]&lt;br /&gt;
* 2025-04: Stanford HAI: [https://hai-production.s3.amazonaws.com/files/hai_ai_index_report_2025.pdf Artificial Intelligence Index Report 2025]&lt;br /&gt;
* 2025-04: Arvind Narayananand Sayash Kapoor: [https://kfai-documents.s3.amazonaws.com/documents/c3cac5a2a7/AI-as-Normal-Technology---Narayanan---Kapoor.pdf AI as Normal Technology]&lt;br /&gt;
* 2025-04: Dwarkesh Patel: [https://www.dwarkesh.com/p/questions-about-ai Questions about the Future of AI]&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: IdeaFoundry: [https://ideafoundry.substack.com/p/evolution-vs-extinction-the-choice Evolution vs. Extinction: The Choice is Ours] The next 18 months will decide whether AI ends us or evolves us&lt;br /&gt;
* 2025-07: [https://cfg.eu/advanced-ai-possible-futures/ Advanced AI: Possible futures] Five scenarios for how the AI-transition could unfold&lt;br /&gt;
* 2025-11: [https://android-dreams.ai/ Android Dreams]&lt;br /&gt;
* 2026-02: [https://www.citriniresearch.com/ Citrini]: [https://www.citriniresearch.com/p/2028gic The 2028 Global Intelligence Crisis: A Thought Exercise in Financial History, from the Future]&lt;br /&gt;
&lt;br /&gt;
==Insightful Analysis of Current State==&lt;br /&gt;
* 2025-11: Andy Masley: [https://andymasley.substack.com/p/the-lump-of-cognition-fallacy The lump of cognition fallacy: The extended mind as the advance of civilization]&lt;br /&gt;
* 2026-02: Eric Jang: [https://evjang.com/2026/02/04/rocks.html As Rocks May Think]&lt;br /&gt;
* 2026-02: Matt Shumer: [https://x.com/mattshumer_/status/2021256989876109403 Something Big Is Happening]&lt;br /&gt;
* 2026-02: Minh Pham: [https://x.com/buckeyevn/status/2014171253045960803?s=20 Why Most Agent Harnesses Are Not Bitter Lesson Pilled]&lt;br /&gt;
&lt;br /&gt;
=Overall=&lt;br /&gt;
* 1993: [https://en.wikipedia.org/wiki/Vernor_Vinge Vernor Vinge]: [https://edoras.sdsu.edu/~vinge/misc/singularity.html The Coming Technological Singularity: How to Survive in the Post-Human Era]&lt;br /&gt;
* 2025-03: Kevin Roose (New York Times): [https://www.nytimes.com/2025/03/14/technology/why-im-feeling-the-agi.html?unlocked_article_code=1.304.TIEy.SmNhKYO4e9c7&amp;amp;smid=url-share Powerful A.I. Is Coming. We’re Not Ready.] Three arguments for taking progress toward artificial general intelligence, or A.G.I., more seriously — whether you’re an optimist or a pessimist.&lt;br /&gt;
* 2025-03: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/thoughts-on-future-ai.html My Thoughts on the Future of &amp;quot;AI&amp;quot;]: &amp;quot;I have very wide error bars on the potential future of large language models, and I think you should too.&amp;quot;&lt;br /&gt;
* 2025-06: Sam Altman: [https://blog.samaltman.com/the-gentle-singularity The Gentle Singularity]&lt;br /&gt;
&lt;br /&gt;
==Surveys of Opinions/Predictions==&lt;br /&gt;
* 2016-06: [https://aiimpacts.org/2016-expert-survey-on-progress-in-ai/ 2016 Expert Survey on Progress in AI]&lt;br /&gt;
** 2023-03: [https://aiimpacts.org/scoring-forecasts-from-the-2016-expert-survey-on-progress-in-ai/ Scoring forecasts from the 2016 “Expert Survey on Progress in AI”]&lt;br /&gt;
* 2022-10: Forecasting Research Institute: [https://forecastingresearch.org/near-term-xpt-accuracy Assessing Near-Term Accuracy in the Existential Risk Persuasion Tournament]&lt;br /&gt;
** 2025-09: Ethan Mollick: [https://x.com/emollick/status/1962859757674344823 Progress is ahead of expectations]&lt;br /&gt;
* 2023-08: [https://wiki.aiimpacts.org/ai_timelines/predictions_of_human-level_ai_timelines/ai_timeline_surveys/2023_expert_survey_on_progress_in_ai 2023 Expert Survey on Progress in AI]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.02843 Thousands of AI Authors on the Future of AI]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14870 Why do Experts Disagree on Existential Risk and P(doom)? A Survey of AI Experts]&lt;br /&gt;
* 2025-02: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/forecasting-ai-2025-update.html AI forecasting retrospective: you&amp;#039;re (probably) over-confident]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/long-timelines-to-advanced-ai-have &amp;quot;Long&amp;quot; timelines to advanced AI have gotten crazy short]&lt;br /&gt;
* 2025-05: [https://theaidigest.org/ai2025-analysis-may AI 2025 Forecasts - May Update]&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41598-026-39070-w Lay beliefs about the badness, likelihood, and importance of human extinction]&lt;br /&gt;
&lt;br /&gt;
==Bad Outcomes==&lt;br /&gt;
* [https://pauseai.info/pdoom List of p(doom) values]&lt;br /&gt;
* 2019-03: [https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like What failure looks like]&lt;br /&gt;
* 2023-03: gwern: [https://gwern.net/fiction/clippy It Looks Like You’re Trying To Take Over The World]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16946 Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development] ([https://gradual-disempowerment.ai/ web version])&lt;br /&gt;
** 2025-02: [https://thezvi.substack.com/p/the-risk-of-gradual-disempowerment The Risk of Gradual Disempowerment from AI]&lt;br /&gt;
** 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-09: [https://doctrines.ai/ The three main doctrines on the future of AI]&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Dominance doctrine:&amp;#039;&amp;#039;&amp;#039; First actor to create advanced AI will attain overwhelming strategic superiority&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Extinction doctrine:&amp;#039;&amp;#039;&amp;#039; Humanity will lose control of ASI, leading to extinction or permanent disempowerment&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Replacement doctrine:&amp;#039;&amp;#039;&amp;#039; AI will automate human tasks, but without fundamentally reshaping or ending civilization&lt;br /&gt;
* 2025-09: Sean ÓhÉigeartaigh: [https://www.cambridge.org/core/journals/cambridge-prisms-extinction/article/extinction-of-the-human-species-what-could-cause-it-and-how-likely-is-it-to-occur/D8816A79BEF5A4C30A3E44FD8D768622 Extinction of the human species: What could cause it and how likely is it to occur?]&lt;br /&gt;
&lt;br /&gt;
==Intelligence Explosion==&lt;br /&gt;
* 2023-06: [https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/ What a Compute-Centric Framework Says About Takeoff Speeds]&lt;br /&gt;
** [https://takeoffspeeds.com/ takeoffspeeds.com simulator]&lt;br /&gt;
* 2025-02: [https://www.forethought.org/research/three-types-of-intelligence-explosion Three Types of Intelligence Explosion]&lt;br /&gt;
* 2025-03: Future of Life Institute: [https://futureoflife.org/ai/are-we-close-to-an-intelligence-explosion/ Are we close to an intelligence explosion?] AIs are inching ever-closer to a critical threshold. Beyond this threshold lie great risks—but crossing it is not inevitable.&lt;br /&gt;
* 2025-03: Forethought: [https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion Will AI R&amp;amp;D Automation Cause a Software Intelligence Explosion?]&lt;br /&gt;
[[Image:Gm-1jugbYAAtq Y.jpeg|450px]]&lt;br /&gt;
* 2025-05: [https://www.thelastinvention.ai/ The Last Invention] Why Humanity’s Final Creation Changes Everything&lt;br /&gt;
* 2025-08: [https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be How quick and big would a software intelligence explosion be?]&lt;br /&gt;
&lt;br /&gt;
==Superintelligence==&lt;br /&gt;
* 2024-10: [http://yager-research.ca/2024/10/how-smart-will-asi-be/ How Smart will ASI be?]&lt;br /&gt;
* 2024-11: [http://yager-research.ca/2024/11/concise-argument-for-asi-risk/ Concise Argument for ASI Risk]&lt;br /&gt;
* 2025-03: [https://dynomight.net/smart/ Limits of smart]&lt;br /&gt;
* 2025-05: [https://timfduffy.substack.com/p/the-limits-of-superintelligence?manualredirect= The Limits of Superintelligence]&lt;br /&gt;
&lt;br /&gt;
==Long-range/Philosophy==&lt;br /&gt;
* 2023-03: Dan Hendrycks: [https://arxiv.org/abs/2303.16200 Natural Selection Favors AIs over Humans]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2025-01: [https://longerramblings.substack.com/p/a-defence-of-slowness-at-the-end A defence of slowness at the end of the world]&lt;br /&gt;
&lt;br /&gt;
=Positives &amp;amp; Optimism=&lt;br /&gt;
==Science &amp;amp; Technology Improvements==&lt;br /&gt;
* 2023-05: [https://www.planned-obsolescence.org/author/kelsey/ Kelsey Piper]: [https://www.planned-obsolescence.org/the-costs-of-caution/ The costs of caution]&lt;br /&gt;
* 2024-09: Sam Altman: [https://ia.samaltman.com/ The Intelligence Age]&lt;br /&gt;
* 2024-10: Dario Amodei: [https://darioamodei.com/machines-of-loving-grace Machines of Loving Grace]&lt;br /&gt;
* 2024-11: Google DeepMind: [https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery A new golden age of discovery]&lt;br /&gt;
* 2025-03: [https://finmoorhouse.com/ Fin Moorhouse], [https://www.williammacaskill.com/ Will MacAskill]: [https://www.forethought.org/research/preparing-for-the-intelligence-explosion Preparing for the Intelligence Explosion]&lt;br /&gt;
&lt;br /&gt;
==Social==&lt;br /&gt;
* 2025-09: [https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale Coasean Bargaining at Scale]: Decentralization, coordination, and co-existence with AGI&lt;br /&gt;
* 2025-10: [https://www.nber.org/system/files/chapters/c15309/c15309.pdf#page=15.23 The Coasean Singularity? Demand, Supply, and Market Design with AI Agents]&lt;br /&gt;
&lt;br /&gt;
==Post-scarcity Society==&lt;br /&gt;
* 2004: Eliezer Yudkowsky (MIRI): [https://intelligence.org/files/CEV.pdf Coherent Extrapolated Volition] and [https://www.lesswrong.com/s/d3WgHDBAPYYScp5Em/p/K4aGvLnHvYgX9pZHS Fun Theory]&lt;br /&gt;
* 2019: John Danaher: [https://www.jstor.org/stable/j.ctvn5txpc Automation and Utopia: Human Flourishing in a World Without Work]&lt;br /&gt;
&lt;br /&gt;
==The Grand Tradeoff==&lt;br /&gt;
* 2026-02: Nick Bostrom: [https://nickbostrom.com/optimal.pdf Optimal Timing for Superintelligence: Mundane Considerations for Existing People]&lt;br /&gt;
&lt;br /&gt;
=Plans=&lt;br /&gt;
* [https://www.narrowpath.co/ A Narrow Path: How to Secure our Future]&lt;br /&gt;
* Marius Hobbhahn: [https://www.lesswrong.com/posts/bb5Tnjdrptu89rcyY/what-s-the-short-timeline-plan What’s the short timeline plan?]&lt;br /&gt;
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]&lt;br /&gt;
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]&lt;br /&gt;
* [https://controlai.com/ Control AI]: [https://controlai.com/dip The Direct Institutional Plan] &lt;br /&gt;
* Luke Drago and L Rudolf L: [https://lukedrago.substack.com/p/the-use-of-knowledge-in-agi-society?triedRedirect=true The use of knowledge in (AGI) society]: How to build to break the [https://lukedrago.substack.com/p/the-intelligence-curse intelligence curse]&lt;br /&gt;
* [https://www.agisocialcontract.org/ AGI Social Contract]&lt;br /&gt;
** [https://www.agisocialcontract.org/forging-a-new-agi-social-contract Forging A New AGI Social Contract]&lt;br /&gt;
* Yoshua Bengio: [https://time.com/7283507/safer-ai-development/ A Potential Path to Safer AI Development]&lt;br /&gt;
** 2025-02: [https://arxiv.org/abs/2502.15657 Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?]&lt;br /&gt;
* 2026-01: Dario Amodei: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/vjAM7F8vMZS7oRrrh/how-do-we-more-safely-defer-to-ais How do we (more) safely defer to AIs?]&lt;br /&gt;
&lt;br /&gt;
==Philosophy==&lt;br /&gt;
* [https://danfaggella.com/ Dan Faggella]:&lt;br /&gt;
** 2018-07: [https://danfaggella.com/moral-singularity/ Moral Singularity – Unpredictable Values Bodes Poorly for Humanity]&lt;br /&gt;
** 2025-02: [https://danfaggella.com/bend/ There is No Pause – We Must Bend the Posthuman Trajectory]&lt;br /&gt;
* Joe Carlsmith: 2024: [https://joecarlsmith.com/2024/01/02/otherness-and-control-in-the-age-of-agi Otherness and control in the age of AGI]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/02/gentleness-and-the-artificial-other Gentleness and the artificial Other]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/04/deep-atheism-and-ai-risk Deep atheism and AI risk]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/08/when-yang-goes-wrong When “yang” goes wrong]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/09/does-ai-risk-other-the-ais Does AI risk “other” the AIs?]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/11/an-even-deeper-atheism An even deeper atheism]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy Being nicer than Clippy]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/18/on-the-abolition-of-man On the abolition of man]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/21/on-green On green]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/25/on-attunement On attunement]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/06/18/loving-a-world-you-dont-trust Loving a world you don’t trust]&lt;br /&gt;
* Anthony Aguirre:&lt;br /&gt;
** [https://x.com/AnthonyNAguirre/status/1898023049930457468 2025-03]: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
[[Image:GlchEeObwAQ88NK.jpeg|300px]]&lt;br /&gt;
* 2025-04: Scott Alexander (Astral Codex Ten): [https://www.astralcodexten.com/p/the-colors-of-her-coat The Colors Of Her Coat] (response to [https://www.theintrinsicperspective.com/p/welcome-to-the-semantic-apocalypse semantic apocalypse] and semantic satiation)&lt;br /&gt;
* 2025-05: Helen Toner: [https://www.ai-frontiers.org/articles/were-arguing-about-ai-safety-wrong We’re Arguing About AI Safety Wrong]: Dynamism vs. stasis is a clearer lens for criticizing controversial AI safety prescriptions&lt;br /&gt;
* 2025-05: Joe Carlsmith: [https://joecarlsmith.substack.com/p/the-stakes-of-ai-moral-status The stakes of AI moral status]&lt;br /&gt;
&lt;br /&gt;
==Research==&lt;br /&gt;
* 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
&lt;br /&gt;
==Alignment==&lt;br /&gt;
* 2023-03: Leopold Aschenbrenner: [https://www.forourposterity.com/nobodys-on-the-ball-on-agi-alignment/ Nobody’s on the ball on AGI alignment]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2404.10636 What are human values, and how do we align AI to them?] ([https://meaningalignment.substack.com/p/0480e023-98c0-4633-a604-990d3ac880ac blog])&lt;br /&gt;
* 2025: Joe Carlsmith: [https://joecarlsmith.substack.com/p/how-do-we-solve-the-alignment-problem How do we solve the alignment problem?] Introduction to an essay series on paths to safe, useful superintelligence&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/what-is-it-to-solve-the-alignment What is it to solve the alignment problem?] Also: to avoid it? Handle it? Solve it forever? Solve it completely? ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16617671-what-is-it-to-solve-the-alignment-problem audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/when-should-we-worry-about-ai-power When should we worry about AI power-seeking?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16651469-when-should-we-worry-about-ai-power-seeking audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/paths-and-waystations-in-ai-safety Paths and waystations in AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16768804-paths-and-waystations-in-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/ai-for-ai-safety AI for AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16790183-ai-for-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/can-we-safely-automate-alignment Can we safely automate alignment research?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17069901-can-we-safely-automate-alignment-research audio version], [https://joecarlsmith.substack.com/p/video-and-transcript-of-talk-on-automating?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=162375391&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email video version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/giving-ais-safe-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=171250683&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email Giving AIs safe motivations] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17686921-giving-ais-safe-motivations audio version])&lt;br /&gt;
*# [https://joecarlsmith.com/2025/09/29/controlling-the-options-ais-can-pursue Controlling the options AIs can pursue] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17909401-controlling-the-options-ais-can-pursue audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/how-human-like-do-safe-ai-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=178666988&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email How human-like do safe AI motivations need to be?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18175429-how-human-like-do-safe-ai-motivations-need-to-be audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/building-ais-that-do-human-like-philosophy Building AIs that do human-like philosophy: AIs will face philosophical questions humans can&amp;#039;t answer for them] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18591342-building-ais-that-do-human-like-philosophy audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/on-restraining-ai-development-for?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=191385185&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email On restraining AI development for the sake of safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18869440-on-restraining-ai-development-for-the-sake-of-safety audio version])&lt;br /&gt;
* 2025-04: Dario Amodei: [https://www.darioamodei.com/post/the-urgency-of-interpretability The Urgency of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Technical==&lt;br /&gt;
* 2025-03: [https://resilience.baulab.info/docs/AI_Action_Plan_RFI.pdf AI Dominance Requires Interpretability and Standards for Transparency and Security]&lt;br /&gt;
* 2026-02: [https://www.gap-map.org/capabilities/?sort=bottlenecks Fundamental Development Gap Map v1.0]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Policy==&lt;br /&gt;
* 2015-03: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-2 Machine intelligence, part 2]&lt;br /&gt;
* 2019-07: Amanda Askell, Miles Brundage, Gillian Hadfield: [https://arxiv.org/abs/1907.04534 The Role of Cooperation in Responsible AI Development]&lt;br /&gt;
* 2025-03: Dan Hendrycks, Eric Schmidt, Alexandr Wang: [https://www.nationalsecurity.ai/ Superintelligence Strategy]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/executive-summary Executive Summary]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/introduction Introduction]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/ai-is-pivotal-for-national-security AI Is Pivotal for National Security]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/deterrence-with-mutual-assured-ai-malfunction-maim Deterrence with Mutual Assured AI Malfunction (MAIM)]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/nonproliferation Nonproliferation]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/competitiveness Competitiveness]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/conclusion Conclusion]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/appendix Appendix FAQs]&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human] ([https://keepthefuturehuman.ai/essay/ essay])&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late (video)]  (2025)&lt;br /&gt;
**# Oversight: Registration required for training &amp;gt;10&amp;lt;sup&amp;gt;25&amp;lt;/sup&amp;gt; FLOP and inference &amp;gt;10&amp;lt;sup&amp;gt;19&amp;lt;/sup&amp;gt; FLOP/s (~1,000 B200 GPUs @ $25M). Build cryptographic licensing into hardware.&lt;br /&gt;
**# Computation Limits: Ban on training models &amp;gt;10&amp;lt;sup&amp;gt;27&amp;lt;/sup&amp;gt; FLOP or inference &amp;gt;10&amp;lt;sup&amp;gt;20&amp;lt;/sup&amp;gt; FLOP/s.&lt;br /&gt;
**# Strict Liability: Hold AI companies responsible for outcomes.&lt;br /&gt;
**# Tiered Regulation: Low regulation on tool-AI, strictest regulation on AGI (general, capable, autonomous systems).&lt;br /&gt;
* 2025-04: [https://x.com/deanwball Dean W. Ball]: [https://arxiv.org/abs/2504.11501 A Framework for the Private Governance of Frontier Artificial Intelligence]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/nonproliferation-is-the-wrong-approach?source=queue Nonproliferation is the wrong approach to AI misuse]&lt;br /&gt;
* 2025-04: MIRI: [https://techgov.intelligence.org/research/ai-governance-to-avoid-extinction AI Governance to Avoid Extinction: The Strategic Landscape and Actionable Research Questions]&lt;br /&gt;
* 2025-05: [https://writing.antonleicht.me/p/the-new-ai-policy-frontier The New AI Policy Frontier]: Beyond the shortcomings of centralised control and alignment, a new school of thought on AI governance emerges. It still faces tricky politics.&lt;br /&gt;
* 2025-05: [https://uncpga.world/agi-uncpga-report/ AGI UNCPGA Report]: Governance of the Transition to Artificial General Intelligence (AGI) Urgent Considerations for the UN General Assembly: Report for the Council of Presidents of the United Nations General Assembly (UNCPGA)&lt;br /&gt;
* 2025-06: [https://writing.antonleicht.me/p/ai-and-jobs-politics-without-policy AI &amp;amp; Jobs: Politics without Policy] Political support mounts - for a policy platform that does not yet exist&lt;br /&gt;
* 2025-06: [https://x.com/littIeramblings Sarah Hastings-Woodhouse]: [https://drive.google.com/file/d/1mmdHBE6M2yiyL21-ctTuRLNH5xOFjqWm/view Safety Features for a Centralized AGI Project]&lt;br /&gt;
* 2025-07: [https://writing.antonleicht.me/p/a-moving-target A Moving Target] Why we might not be quite ready to comprehensively regulate AI, and why it matters&lt;br /&gt;
* 2025-07: [https://www-cdn.anthropic.com/0dc382a2086f6a054eeb17e8a531bd9625b8e6e5.pdf Anthropic: Build AI in America] ([https://www.anthropic.com/news/build-ai-in-america blog])&lt;br /&gt;
* 2025-12: [https://asi-prevention.com/ How middle powers may prevent the development of artificial superintelligence]&lt;br /&gt;
* 2026-03: [https://humanstatement.org/ The Pro-Human AI Declaration]&lt;br /&gt;
&lt;br /&gt;
==Restriction==&lt;br /&gt;
* 2024-05: OpenAI: [https://openai.com/index/reimagining-secure-infrastructure-for-advanced-ai/ Reimagining secure infrastructure for advanced AI] OpenAI calls for an evolution in infrastructure security to protect advanced AI &lt;br /&gt;
* 2025-07: MIRI: [https://arxiv.org/abs/2507.09801 Technical Requirements for Halting Dangerous AI Activities]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI safety]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8756</id>
		<title>AI predictions</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8756"/>
		<updated>2026-03-24T15:52:34Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Job Loss */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Capability Scaling=&lt;br /&gt;
* 2019-03: Rich Sutton: [https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf The Bitter Lesson]&lt;br /&gt;
* 2020-09: Ajeya Cotra: [https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines Draft report on AI timelines]&lt;br /&gt;
* 2022-01: gwern: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis]&lt;br /&gt;
* 2023-05: Richard Ngo: [https://www.lesswrong.com/posts/BoA3agdkAzL6HQtQP/clarifying-and-predicting-agi Clarifying and predicting AGI]&lt;br /&gt;
* 2024-06: Aidan McLaughlin: [https://yellow-apartment-148.notion.site/AI-Search-The-Bitter-er-Lesson-44c11acd27294f4495c3de778cd09c8d AI Search: The Bitter-er Lesson]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.14499 Measuring AI Ability to Complete Long Tasks Measuring AI Ability to Complete Long Tasks]&lt;br /&gt;
** 2025-04: [https://peterwildeford.substack.com/p/forecaster-reacts-metrs-bombshell Forecaster reacts: METR&amp;#039;s bombshell paper about AI acceleration] New data supports an exponential AI curve, but lots of uncertainty remains&lt;br /&gt;
** 2025-04: AI Digest: [https://theaidigest.org/time-horizons A new Moore&amp;#039;s Law for AI agents]&lt;br /&gt;
[[Image:GmZHL8xWQAAtFlF.jpeg|450px]]&lt;br /&gt;
* 2025-04: [https://epoch.ai/blog/trends-in-ai-supercomputers Trends in AI Supercomputers] ([https://arxiv.org/abs/2504.16026 preprint])&lt;br /&gt;
* [https://ai-timeline.org/ The Road to AGI] (timeline visualization)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09677 The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs]&lt;br /&gt;
* 2025-09: [https://www.julian.ac/blog/2025/09/27/failing-to-understand-the-exponential-again/ Failing to Understand the Exponential, Again]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/rRbDNQLfihiHbXytf/distinguish-between-inference-scaling-and-larger-tasks-use Distinguish between inference scaling and &amp;quot;larger tasks use more compute&amp;quot;]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03992 Measuring AI R&amp;amp;D Automation] ([https://astrangeattractor.substack.com/p/measuring-ai-r-and-d-automation?triedRedirect=true blog])&lt;br /&gt;
&lt;br /&gt;
==Scaling Laws==&lt;br /&gt;
See: [[AI_understanding#Scaling_Laws|Scaling Laws]]&lt;br /&gt;
&lt;br /&gt;
==AGI Achievable==&lt;br /&gt;
* Yoshua Bengio: [https://arxiv.org/abs/2310.17688 Managing extreme AI risks amid rapid progress]&lt;br /&gt;
* Leopold Aschenbrenner: [https://situational-awareness.ai/from-gpt-4-to-agi/#Counting_the_OOMs Situational Awareness: Counting the OOMs]&lt;br /&gt;
* Richard Ngo: [https://medium.com/@richardcngo/visualizing-the-deep-learning-revolution-722098eb9c5 Visualizing the deep learning revolution]&lt;br /&gt;
* Katja Grace: [https://blog.aiimpacts.org/p/2023-ai-survey-of-2778-six-things Survey of 2,778 AI authors: six parts in pictures]&lt;br /&gt;
* Epoch AI: [https://epoch.ai/trends Machine Learning Trends]&lt;br /&gt;
* AI Digest: [https://theaidigest.org/progress-and-dangers How fast is AI improving?]&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/when-will-agi-arrive/ The case for AGI by 2030]&lt;br /&gt;
&lt;br /&gt;
==AGI Definition==&lt;br /&gt;
* 2023-11: Allan Dafoe, Shane Legg, et al.: [https://arxiv.org/abs/2311.02462 Levels of AGI for Operationalizing Progress on the Path to AGI]&lt;br /&gt;
* 2024-04: Bowen Xu: [https://arxiv.org/abs/2404.10731 What is Meant by AGI? On the Definition of Artificial General Intelligence]&lt;br /&gt;
* 2025-10: Dan Hendrycks et al.: [https://www.agidefinition.ai/paper.pdf A Definition of AGI]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07364 On the universal definition of intelligence]&lt;br /&gt;
&lt;br /&gt;
==Recursive Self Improvement (RSI)==&lt;br /&gt;
* 2026-02: [https://80000hours.org/articles/how-ai-driven-feedback-loops-could-make-things-very-crazy-very-fast/ How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
&lt;br /&gt;
==Progress Models==&lt;br /&gt;
From [http://yager-research.ca/2025/04/ai-impact-predictions/ AI Impact Predictions]:&lt;br /&gt;
&lt;br /&gt;
[[Image:AI impact models-2025 11 24.png|450px]]&lt;br /&gt;
&lt;br /&gt;
=Economic and Political=&lt;br /&gt;
* 2019-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3482150 The Impact of Artificial Intelligence on the Labor Market]&lt;br /&gt;
* 2020-06: [https://www.openphilanthropy.org/research/modeling-the-human-trajectory/ Modeling the Human Trajectory] (GDP)&lt;br /&gt;
* 2021-06: [https://www.openphilanthropy.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/ Report on Whether AI Could Drive Explosive Economic Growth]&lt;br /&gt;
* 2023-10: Marc Andreessen: [https://a16z.com/the-techno-optimist-manifesto/ The Techno-Optimist Manifesto]&lt;br /&gt;
* 2023-12: [https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html My techno-optimism]: &amp;quot;defensive acceleration&amp;quot; ([https://vitalik.eth.limo/index.html Vitalik Buterin])&lt;br /&gt;
* 2024-03: Noah Smith: [https://www.noahpinion.blog/p/plentiful-high-paying-jobs-in-the Plentiful, high-paying jobs in the age of AI: Comparative advantage is very subtle, but incredibly powerful.] ([https://x.com/liron/status/1768013030741475485 video])&lt;br /&gt;
* 2024-03: [https://doi.org/10.3386/w32255 Scenarios for the Transition to AGI] (AGI leads to wage collapse)&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-06: [https://www.frbsf.org/wp-content/uploads/AI-and-Growth-Aghion-Bunel.pdf AI and Growth: Where Do We Stand?]&lt;br /&gt;
* 2024-09: OpenAI [https://cdn.openai.com/global-affairs/openai-infra-economics-10.09.24.pdf Infrastructure is Destiny: Economic Returns on US Investment in Democratic AI]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/KFFaKu27FNugCHFmh/by-default-capital-will-matter-more-than-ever-after-agi By default, capital will matter more than ever after AGI] (L Rudolf L)&lt;br /&gt;
* 2025-01: [https://lukedrago.substack.com/p/the-intelligence-curse The Intelligence Curse]: With AGI, powerful actors will lose their incentives to invest in people&lt;br /&gt;
** Updated 2025-04: [https://intelligence-curse.ai/ The Intelligence Curse] (Luke Drago and Rudolf Laine)&lt;br /&gt;
*** [https://intelligence-curse.ai/pyramid/ Pyramid Replacement]&lt;br /&gt;
*** [https://intelligence-curse.ai/capital/ Capital, AGI, and Human Ambition]&lt;br /&gt;
*** [https://intelligence-curse.ai/defining/ Defining the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/shaping/ Shaping the Social Contract]&lt;br /&gt;
*** [https://intelligence-curse.ai/breaking/ Breaking the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/history/ History is Yours to Write]&lt;br /&gt;
* 2025-01: Microsoft: [https://blogs.microsoft.com/on-the-issues/2025/01/03/the-golden-opportunity-for-american-ai/ The Golden Opportunity for American AI]&lt;br /&gt;
* 2025-01: [https://www.maximum-progress.com/p/agi-will-not-make-labor-worthless AGI Will Not Make Labor Worthless]&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/global-affairs/ai-in-america-oai-economic-blueprint-20250113.pdf AI in America: OpenAI&amp;#039;s Economic Blueprint] ([https://openai.com/global-affairs/openais-economic-blueprint/ blog])&lt;br /&gt;
* 2025-01: [https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai How much economic growth from AI should we expect, how soon?]&lt;br /&gt;
* 2025-02: Morgan Stanley: [https://advisor.morganstanley.com/john.howard/documents/field/j/jo/john-howard/The_Humanoid_100_-_Mapping_the_Humanoid_Robot_Value_Chain.pdf The Humanoid 100: Mapping the Humanoid Robot Value Chain]&lt;br /&gt;
* 2025-02: [https://www.anthropic.com/news/the-anthropic-economic-index The Anthropic Economic Index]: [https://assets.anthropic.com/m/2e23255f1e84ca97/original/Economic_Tasks_AI_Paper.pdf Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11264 Strategic Wealth Accumulation Under Transformative AI Expectations]&lt;br /&gt;
* 2025-02: Tyler Cowen: [https://marginalrevolution.com/marginalrevolution/2025/02/why-i-think-ai-take-off-is-relatively-slow.html Why I think AI take-off is relatively slow]&lt;br /&gt;
* 2025-03: Epoch AI: [https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d Most AI value will come from broad automation, not from R&amp;amp;D]&lt;br /&gt;
** The primary economic impact of AI will be its ability to broadly automate labor&lt;br /&gt;
** Automating AI R&amp;amp;D alone likely won’t dramatically accelerate AI progress&lt;br /&gt;
** Fully automating R&amp;amp;D requires a very broad set of abilities&lt;br /&gt;
** AI takeoff will likely be diffuse and salient&lt;br /&gt;
* 2025-03: [https://www.anthropic.com/news/anthropic-economic-index-insights-from-claude-sonnet-3-7 Anthropic Economic Index: Insights from Claude 3.7 Sonnet]&lt;br /&gt;
* 2025-04: [https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality Will there be extreme inequality from AI?]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/research/impact-software-development Anthropic Economic Index: AI’s Impact on Software Development]&lt;br /&gt;
* 2025-05: [https://www.theguardian.com/books/2025/may/04/the-big-idea-can-we-stop-ai-making-humans-obsolete Better at everything: how AI could make human beings irrelevant]&lt;br /&gt;
* 2025-05: Forethought: [https://www.forethought.org/research/the-industrial-explosion The Industrial Explosion]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20273 Ten Principles of AI Agent Economics]&lt;br /&gt;
* 2025-07: [https://substack.com/home/post/p-167879696 What Economists Get Wrong about AI] They ignore innovation effects, use outdated capability assumptions, and miss the robotics revolution&lt;br /&gt;
* 2025-07: [https://www.nber.org/books-and-chapters/economics-transformative-ai/we-wont-be-missed-work-and-growth-era-agi We Won&amp;#039;t Be Missed: Work and Growth in the Era of AGI]&lt;br /&gt;
* 2025-07: [https://www.nber.org/papers/w34034 The Economics of Bicycles for the Mind]&lt;br /&gt;
* 2025-09: [https://conference.nber.org/conf_papers/f227491.pdf Genius on Demand: The Value of Transformative Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://peterwildeford.substack.com/p/ai-is-probably-not-a-bubble AI is probably not a bubble: AI companies have revenue, demand, and paths to immense value]&lt;br /&gt;
* 2025-11: [https://windowsontheory.org/2025/11/04/thoughts-by-a-non-economist-on-ai-and-economics/ Thoughts by a non-economist on AI and economics]&lt;br /&gt;
* 2025-11: [https://www.nber.org/papers/w34444 Artificial Intelligence, Competition, and Welfare]&lt;br /&gt;
* 2025-11: [https://www.anthropic.com/research/estimating-productivity-gains Estimating AI productivity gains from Claude conversations] (Anthropic)&lt;br /&gt;
* 2025-12: [https://benjamintodd.substack.com/p/how-ai-driven-feedback-loops-could How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
* 2025-12: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth] (Philip Trammell and Leopold Aschenbrenner)&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Anthropic Economic Index: new building blocks for understanding AI use]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/economic-index-primitives Anthropic Economic Index report: economic primitives]&lt;br /&gt;
* 2026-02: Nate Silver: [https://www.natesilver.net/p/the-singularity-wont-be-gentle The singularity won&amp;#039;t be gentle: If AI is even half as transformational as Silicon Valley assumes, politics will never be the same again]&lt;br /&gt;
&lt;br /&gt;
==Job Loss==&lt;br /&gt;
* 2023-03: [https://arxiv.org/pdf/2303.10130 GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models]&lt;br /&gt;
** 2023-03: [https://www.livemint.com/news/world/these-jobs-are-most-at-risk-due-to-chatgpt-as-per-openai-study-11679358453267.html These jobs are most at risk due to ChatGPT, as per OpenAI study]&lt;br /&gt;
* 2023-08: [https://dx.doi.org/10.2139/ssrn.4527336 The Short-Term Effects of Generative Artificial Intelligence on Employment: Evidence from an Online Labor Market]&lt;br /&gt;
** [https://x.com/jburnmurdoch/status/1722938749519077688 Freelancer sector shrinking]&lt;br /&gt;
[[Image:F-kVQuvWkAAemkr.png|400px]]&lt;br /&gt;
* 2023-09: [https://global-uploads.webflow.com/64d5f73a7fc5e8a240310c4d/650a128a34386a1206b6506c_FINAL%20Briefing%20-%20Adoption%20of%20Automation%20and%20AI%20in%20the%20UK.pdf What drives UK firms to adopt AI and robotics, and what are the consequences for jobs?]&lt;br /&gt;
** [https://www.digitalinformationworld.com/2023/09/78-of-companies-say-ai-created-more-jobs.html 78% of Companies Say AI Created More Jobs]&lt;br /&gt;
* 2023-11: [https://theaipi.org/ai-interactive-map/ New Analysis Shows Over 20% of US Jobs Significantly Exposed to AI Automation In the Near Future]&lt;br /&gt;
* 2024-01: [https://techcrunch.com/2024/01/09/duolingo-cut-10-of-its-contractor-workforce-as-the-company-embraces-ai/ Duolingo cuts 10% of its contractor workforce as the company embraces AI]&lt;br /&gt;
* 2024-02: [https://www.pwc.com/gx/en/issues/c-suite-insights/the-leadership-agenda/gen-ai-is-a-tool-for-growth-not-just-efficiency.html#:~:text=One%20out%20of%20every%20four%20of%20the%204%2C702,to%20accomplish%20the%20same%20tasks%20with%20fewer%20workers Gen AI is a tool for growth, not just efficiency: Tech CEOs are investing to build their workforce and capitalise on new opportunities from generative AI. That’s a sharp contrast to how their peers view it.]&lt;br /&gt;
* 2024-04: [https://www.nytimes.com/2024/04/10/business/investment-banking-jobs-artificial-intelligence.html AI is Poised to Replace the Entry-Level Grunt Work of a Wall Street Career]&lt;br /&gt;
* 2024-07: [https://www.wired.com/story/ai-is-already-taking-jobs-in-the-video-game-industry/ AI Is Already Taking Jobs in the Video Game Industry]: A WIRED investigation finds that major players like Activision Blizzard, which recently laid off scores of workers, are using generative AI for game development&lt;br /&gt;
* 2024-08: [https://www.bbc.com/news/articles/c80e1gp9m9zo Klarna: AI lets us cut thousands of jobs - but pay more]&lt;br /&gt;
* 2025-01: [https://scholarspace.manoa.hawaii.edu/server/api/core/bitstreams/4f39375d-59c2-4c4a-b394-f3eed7858c80/content AI and Freelancers: Has the Inflection Point Arrived?]&lt;br /&gt;
* 2025-01: [https://www.aporiamagazine.com/p/yes-youre-going-to-be-replaced Yes, you&amp;#039;re going to be replaced: So much cope about AI]&lt;br /&gt;
* 2025-03: [https://commonplace.org/2025/03/20/will-ai-automate-away-your-job/ Will AI Automate Away Your Job? The time-horizon model explains the future of the technology]&lt;br /&gt;
* 2025-05: [https://www.forbes.com/sites/jackkelly/2025/05/04/its-time-to-get-concerned-klarna-ups-duolingo-cisco-and-many-other-companies-are-replacing-workers-with-ai/ It’s Time To Get Concerned, Klarna, UPS, Duolingo, Cisco, And Many Other Companies Are Replacing Workers With AI]&lt;br /&gt;
* 2025-05: [https://time.com/7289692/when-ai-replaces-workers/ What Happens When AI Replaces Workers?]&lt;br /&gt;
* 2025-05: [https://www.oxfordeconomics.com/resource/educated-but-unemployed-a-rising-reality-for-us-college-grads/ Educated but unemployed, a rising reality for US college grads] Structural shifts in tech hiring and the growing impact of AI are driving higher unemployment among recent college graduates&lt;br /&gt;
* 2025-05: NY Times: [https://www.nytimes.com/2025/05/30/technology/ai-jobs-college-graduates.html?unlocked_article_code=1.LE8.LlC6.eT5XcpA9hxC2&amp;amp;smid=url-share For Some Recent Graduates, the A.I. Job Apocalypse May Already Be Here] The unemployment rate for recent college graduates has jumped as companies try to replace entry-level workers with artificial intelligence&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/skills-ai-makes-valuable/ How not to lose your job to AI] The skills AI will make more valuable (and how to learn them)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06576 Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce]&lt;br /&gt;
[[Image:0dab4c86-882d-4095-9d12-d19684ed5184 675x680.png|300px]]&lt;br /&gt;
* 2025-07: Harvard Business Review: [https://hbr.org/2025/06/what-gets-measured-ai-will-automate What Gets Measured, AI Will Automate]&lt;br /&gt;
* 2025-08: [https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/ Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5560401 Performance or Principle: Resistance to Artificial Intelligence in the U.S. Labor Market]&lt;br /&gt;
* 2025-10: [https://www.siliconcontinent.com/p/the-ai-becker-problem The AI Becker problem: Who will train the next generation?]&lt;br /&gt;
* 2026-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6134506 AI, Automation, and Expertise]&lt;br /&gt;
* 2026-02: [https://arachnemag.substack.com/p/the-jevons-paradox-for-intelligence The Jevons Paradox for Intelligence: Fears of AI-induced job loss could not be more wrong]&lt;br /&gt;
* 2026-03: [https://www.dropbox.com/scl/fo/689u1g785x8jp6c8v1s21/AKxZ_N15vUxMA3PBtpbr5nM?dl=0&amp;amp;e=1&amp;amp;preview=2026.03.24+Bundles.pdf&amp;amp;rlkey=ottgcu71u1t4mhn6tblvatu8w&amp;amp;st=dj6k0x2o Weak Bundle, Strong Bundle:How AI Redraws Job Boundaries]&lt;br /&gt;
&lt;br /&gt;
==Productivity Impact==&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2026-02: [https://www.ft.com/content/4b51d0b4-bbfe-4f05-b50a-1d485d419dc5 The AI productivity take-off is finally visible] ([https://x.com/erikbryn/status/2023075588974735869?s=20 Erik Brynjolfsson])&lt;br /&gt;
** Businesses are finally beginning to reap some of AI&amp;#039;s benefits.&lt;br /&gt;
* 2026-02: New York Times: [https://www.nytimes.com/2026/02/18/opinion/ai-software.html The A.I. Disruption We’ve Been Waiting for Has Arrived]&lt;br /&gt;
&lt;br /&gt;
==National Security==&lt;br /&gt;
* 2025-04: Jeremie Harris and Edouard Harris: [https://superintelligence.gladstone.ai/ America’s Superintelligence Project]&lt;br /&gt;
&lt;br /&gt;
==AI Manhattan Project==&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-10: [https://thezvi.substack.com/p/ai-88-thanks-for-the-memos?open=false#%C2%A7thanks-for-the-memos-introduction-and-competitiveness White House Memo calls for action on AI]&lt;br /&gt;
* 2024-11: [https://www.uscc.gov/annual-report/2024-annual-report-congress 2024 Annual Report to Congress]: [https://www.reuters.com/technology/artificial-intelligence/us-government-commission-pushes-manhattan-project-style-ai-initiative-2024-11-19/ calls] for &amp;quot;Manhattan Project-style&amp;quot; effort&lt;br /&gt;
* 2025-05-29: [https://x.com/ENERGY/status/1928085878561272223 DoE Tweet]: &amp;quot;AI is the next Manhattan Project, and THE UNITED STATES WILL WIN. 🇺🇸&amp;quot;&lt;br /&gt;
* 2025-07: [https://epoch.ai/gradient-updates/how-big-could-an-ai-manhattan-project-get How big could an “AI Manhattan Project” get?]&lt;br /&gt;
&lt;br /&gt;
=Near-term=&lt;br /&gt;
* 2021-08: Daniel Kokotajlo: [https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like What 2026 looks like]&lt;br /&gt;
* 2025-02: Sam Altman: [https://blog.samaltman.com/three-observations Three Observations]&lt;br /&gt;
*# The intelligence of an AI model roughly equals the log of the resources used to train and run it.&lt;br /&gt;
*# The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use&lt;br /&gt;
*# The socioeconomic value of linearly increasing intelligence is super-exponential in nature&lt;br /&gt;
* 2025-03: [https://www.pathwaysai.org/p/glimpses-of-ai-progess Glimpses of AI Progress: Mental models for fast times]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41598-025-92190-7 Navigating artificial general intelligence development: societal, technological, ethical, and brain-inspired pathways]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
** 2025-07: Video: [https://www.youtube.com/watch?v=5KVDDfAkRgc Are We 3 Years From AI Disaster? A Rigorous Forecast]&lt;br /&gt;
* 2025-04: Stanford HAI: [https://hai-production.s3.amazonaws.com/files/hai_ai_index_report_2025.pdf Artificial Intelligence Index Report 2025]&lt;br /&gt;
* 2025-04: Arvind Narayananand Sayash Kapoor: [https://kfai-documents.s3.amazonaws.com/documents/c3cac5a2a7/AI-as-Normal-Technology---Narayanan---Kapoor.pdf AI as Normal Technology]&lt;br /&gt;
* 2025-04: Dwarkesh Patel: [https://www.dwarkesh.com/p/questions-about-ai Questions about the Future of AI]&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: IdeaFoundry: [https://ideafoundry.substack.com/p/evolution-vs-extinction-the-choice Evolution vs. Extinction: The Choice is Ours] The next 18 months will decide whether AI ends us or evolves us&lt;br /&gt;
* 2025-07: [https://cfg.eu/advanced-ai-possible-futures/ Advanced AI: Possible futures] Five scenarios for how the AI-transition could unfold&lt;br /&gt;
* 2025-11: [https://android-dreams.ai/ Android Dreams]&lt;br /&gt;
* 2026-02: [https://www.citriniresearch.com/ Citrini]: [https://www.citriniresearch.com/p/2028gic The 2028 Global Intelligence Crisis: A Thought Exercise in Financial History, from the Future]&lt;br /&gt;
&lt;br /&gt;
==Insightful Analysis of Current State==&lt;br /&gt;
* 2025-11: Andy Masley: [https://andymasley.substack.com/p/the-lump-of-cognition-fallacy The lump of cognition fallacy: The extended mind as the advance of civilization]&lt;br /&gt;
* 2026-02: Eric Jang: [https://evjang.com/2026/02/04/rocks.html As Rocks May Think]&lt;br /&gt;
* 2026-02: Matt Shumer: [https://x.com/mattshumer_/status/2021256989876109403 Something Big Is Happening]&lt;br /&gt;
* 2026-02: Minh Pham: [https://x.com/buckeyevn/status/2014171253045960803?s=20 Why Most Agent Harnesses Are Not Bitter Lesson Pilled]&lt;br /&gt;
&lt;br /&gt;
=Overall=&lt;br /&gt;
* 1993: [https://en.wikipedia.org/wiki/Vernor_Vinge Vernor Vinge]: [https://edoras.sdsu.edu/~vinge/misc/singularity.html The Coming Technological Singularity: How to Survive in the Post-Human Era]&lt;br /&gt;
* 2025-03: Kevin Roose (New York Times): [https://www.nytimes.com/2025/03/14/technology/why-im-feeling-the-agi.html?unlocked_article_code=1.304.TIEy.SmNhKYO4e9c7&amp;amp;smid=url-share Powerful A.I. Is Coming. We’re Not Ready.] Three arguments for taking progress toward artificial general intelligence, or A.G.I., more seriously — whether you’re an optimist or a pessimist.&lt;br /&gt;
* 2025-03: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/thoughts-on-future-ai.html My Thoughts on the Future of &amp;quot;AI&amp;quot;]: &amp;quot;I have very wide error bars on the potential future of large language models, and I think you should too.&amp;quot;&lt;br /&gt;
* 2025-06: Sam Altman: [https://blog.samaltman.com/the-gentle-singularity The Gentle Singularity]&lt;br /&gt;
&lt;br /&gt;
==Surveys of Opinions/Predictions==&lt;br /&gt;
* 2016-06: [https://aiimpacts.org/2016-expert-survey-on-progress-in-ai/ 2016 Expert Survey on Progress in AI]&lt;br /&gt;
** 2023-03: [https://aiimpacts.org/scoring-forecasts-from-the-2016-expert-survey-on-progress-in-ai/ Scoring forecasts from the 2016 “Expert Survey on Progress in AI”]&lt;br /&gt;
* 2022-10: Forecasting Research Institute: [https://forecastingresearch.org/near-term-xpt-accuracy Assessing Near-Term Accuracy in the Existential Risk Persuasion Tournament]&lt;br /&gt;
** 2025-09: Ethan Mollick: [https://x.com/emollick/status/1962859757674344823 Progress is ahead of expectations]&lt;br /&gt;
* 2023-08: [https://wiki.aiimpacts.org/ai_timelines/predictions_of_human-level_ai_timelines/ai_timeline_surveys/2023_expert_survey_on_progress_in_ai 2023 Expert Survey on Progress in AI]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.02843 Thousands of AI Authors on the Future of AI]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14870 Why do Experts Disagree on Existential Risk and P(doom)? A Survey of AI Experts]&lt;br /&gt;
* 2025-02: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/forecasting-ai-2025-update.html AI forecasting retrospective: you&amp;#039;re (probably) over-confident]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/long-timelines-to-advanced-ai-have &amp;quot;Long&amp;quot; timelines to advanced AI have gotten crazy short]&lt;br /&gt;
* 2025-05: [https://theaidigest.org/ai2025-analysis-may AI 2025 Forecasts - May Update]&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41598-026-39070-w Lay beliefs about the badness, likelihood, and importance of human extinction]&lt;br /&gt;
&lt;br /&gt;
==Bad Outcomes==&lt;br /&gt;
* [https://pauseai.info/pdoom List of p(doom) values]&lt;br /&gt;
* 2019-03: [https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like What failure looks like]&lt;br /&gt;
* 2023-03: gwern: [https://gwern.net/fiction/clippy It Looks Like You’re Trying To Take Over The World]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16946 Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development] ([https://gradual-disempowerment.ai/ web version])&lt;br /&gt;
** 2025-02: [https://thezvi.substack.com/p/the-risk-of-gradual-disempowerment The Risk of Gradual Disempowerment from AI]&lt;br /&gt;
** 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-09: [https://doctrines.ai/ The three main doctrines on the future of AI]&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Dominance doctrine:&amp;#039;&amp;#039;&amp;#039; First actor to create advanced AI will attain overwhelming strategic superiority&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Extinction doctrine:&amp;#039;&amp;#039;&amp;#039; Humanity will lose control of ASI, leading to extinction or permanent disempowerment&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Replacement doctrine:&amp;#039;&amp;#039;&amp;#039; AI will automate human tasks, but without fundamentally reshaping or ending civilization&lt;br /&gt;
* 2025-09: Sean ÓhÉigeartaigh: [https://www.cambridge.org/core/journals/cambridge-prisms-extinction/article/extinction-of-the-human-species-what-could-cause-it-and-how-likely-is-it-to-occur/D8816A79BEF5A4C30A3E44FD8D768622 Extinction of the human species: What could cause it and how likely is it to occur?]&lt;br /&gt;
&lt;br /&gt;
==Intelligence Explosion==&lt;br /&gt;
* 2023-06: [https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/ What a Compute-Centric Framework Says About Takeoff Speeds]&lt;br /&gt;
** [https://takeoffspeeds.com/ takeoffspeeds.com simulator]&lt;br /&gt;
* 2025-02: [https://www.forethought.org/research/three-types-of-intelligence-explosion Three Types of Intelligence Explosion]&lt;br /&gt;
* 2025-03: Future of Life Institute: [https://futureoflife.org/ai/are-we-close-to-an-intelligence-explosion/ Are we close to an intelligence explosion?] AIs are inching ever-closer to a critical threshold. Beyond this threshold lie great risks—but crossing it is not inevitable.&lt;br /&gt;
* 2025-03: Forethought: [https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion Will AI R&amp;amp;D Automation Cause a Software Intelligence Explosion?]&lt;br /&gt;
[[Image:Gm-1jugbYAAtq Y.jpeg|450px]]&lt;br /&gt;
* 2025-05: [https://www.thelastinvention.ai/ The Last Invention] Why Humanity’s Final Creation Changes Everything&lt;br /&gt;
* 2025-08: [https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be How quick and big would a software intelligence explosion be?]&lt;br /&gt;
&lt;br /&gt;
==Superintelligence==&lt;br /&gt;
* 2024-10: [http://yager-research.ca/2024/10/how-smart-will-asi-be/ How Smart will ASI be?]&lt;br /&gt;
* 2024-11: [http://yager-research.ca/2024/11/concise-argument-for-asi-risk/ Concise Argument for ASI Risk]&lt;br /&gt;
* 2025-03: [https://dynomight.net/smart/ Limits of smart]&lt;br /&gt;
* 2025-05: [https://timfduffy.substack.com/p/the-limits-of-superintelligence?manualredirect= The Limits of Superintelligence]&lt;br /&gt;
&lt;br /&gt;
==Long-range/Philosophy==&lt;br /&gt;
* 2023-03: Dan Hendrycks: [https://arxiv.org/abs/2303.16200 Natural Selection Favors AIs over Humans]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2025-01: [https://longerramblings.substack.com/p/a-defence-of-slowness-at-the-end A defence of slowness at the end of the world]&lt;br /&gt;
&lt;br /&gt;
=Positives &amp;amp; Optimism=&lt;br /&gt;
==Science &amp;amp; Technology Improvements==&lt;br /&gt;
* 2023-05: [https://www.planned-obsolescence.org/author/kelsey/ Kelsey Piper]: [https://www.planned-obsolescence.org/the-costs-of-caution/ The costs of caution]&lt;br /&gt;
* 2024-09: Sam Altman: [https://ia.samaltman.com/ The Intelligence Age]&lt;br /&gt;
* 2024-10: Dario Amodei: [https://darioamodei.com/machines-of-loving-grace Machines of Loving Grace]&lt;br /&gt;
* 2024-11: Google DeepMind: [https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery A new golden age of discovery]&lt;br /&gt;
* 2025-03: [https://finmoorhouse.com/ Fin Moorhouse], [https://www.williammacaskill.com/ Will MacAskill]: [https://www.forethought.org/research/preparing-for-the-intelligence-explosion Preparing for the Intelligence Explosion]&lt;br /&gt;
&lt;br /&gt;
==Social==&lt;br /&gt;
* 2025-09: [https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale Coasean Bargaining at Scale]: Decentralization, coordination, and co-existence with AGI&lt;br /&gt;
* 2025-10: [https://www.nber.org/system/files/chapters/c15309/c15309.pdf#page=15.23 The Coasean Singularity? Demand, Supply, and Market Design with AI Agents]&lt;br /&gt;
&lt;br /&gt;
==Post-scarcity Society==&lt;br /&gt;
* 2004: Eliezer Yudkowsky (MIRI): [https://intelligence.org/files/CEV.pdf Coherent Extrapolated Volition] and [https://www.lesswrong.com/s/d3WgHDBAPYYScp5Em/p/K4aGvLnHvYgX9pZHS Fun Theory]&lt;br /&gt;
* 2019: John Danaher: [https://www.jstor.org/stable/j.ctvn5txpc Automation and Utopia: Human Flourishing in a World Without Work]&lt;br /&gt;
&lt;br /&gt;
==The Grand Tradeoff==&lt;br /&gt;
* 2026-02: Nick Bostrom: [https://nickbostrom.com/optimal.pdf Optimal Timing for Superintelligence: Mundane Considerations for Existing People]&lt;br /&gt;
&lt;br /&gt;
=Plans=&lt;br /&gt;
* [https://www.narrowpath.co/ A Narrow Path: How to Secure our Future]&lt;br /&gt;
* Marius Hobbhahn: [https://www.lesswrong.com/posts/bb5Tnjdrptu89rcyY/what-s-the-short-timeline-plan What’s the short timeline plan?]&lt;br /&gt;
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]&lt;br /&gt;
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]&lt;br /&gt;
* [https://controlai.com/ Control AI]: [https://controlai.com/dip The Direct Institutional Plan] &lt;br /&gt;
* Luke Drago and L Rudolf L: [https://lukedrago.substack.com/p/the-use-of-knowledge-in-agi-society?triedRedirect=true The use of knowledge in (AGI) society]: How to build to break the [https://lukedrago.substack.com/p/the-intelligence-curse intelligence curse]&lt;br /&gt;
* [https://www.agisocialcontract.org/ AGI Social Contract]&lt;br /&gt;
** [https://www.agisocialcontract.org/forging-a-new-agi-social-contract Forging A New AGI Social Contract]&lt;br /&gt;
* Yoshua Bengio: [https://time.com/7283507/safer-ai-development/ A Potential Path to Safer AI Development]&lt;br /&gt;
** 2025-02: [https://arxiv.org/abs/2502.15657 Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?]&lt;br /&gt;
* 2026-01: Dario Amodei: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/vjAM7F8vMZS7oRrrh/how-do-we-more-safely-defer-to-ais How do we (more) safely defer to AIs?]&lt;br /&gt;
&lt;br /&gt;
==Philosophy==&lt;br /&gt;
* [https://danfaggella.com/ Dan Faggella]:&lt;br /&gt;
** 2018-07: [https://danfaggella.com/moral-singularity/ Moral Singularity – Unpredictable Values Bodes Poorly for Humanity]&lt;br /&gt;
** 2025-02: [https://danfaggella.com/bend/ There is No Pause – We Must Bend the Posthuman Trajectory]&lt;br /&gt;
* Joe Carlsmith: 2024: [https://joecarlsmith.com/2024/01/02/otherness-and-control-in-the-age-of-agi Otherness and control in the age of AGI]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/02/gentleness-and-the-artificial-other Gentleness and the artificial Other]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/04/deep-atheism-and-ai-risk Deep atheism and AI risk]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/08/when-yang-goes-wrong When “yang” goes wrong]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/09/does-ai-risk-other-the-ais Does AI risk “other” the AIs?]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/11/an-even-deeper-atheism An even deeper atheism]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy Being nicer than Clippy]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/18/on-the-abolition-of-man On the abolition of man]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/21/on-green On green]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/25/on-attunement On attunement]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/06/18/loving-a-world-you-dont-trust Loving a world you don’t trust]&lt;br /&gt;
* Anthony Aguirre:&lt;br /&gt;
** [https://x.com/AnthonyNAguirre/status/1898023049930457468 2025-03]: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
[[Image:GlchEeObwAQ88NK.jpeg|300px]]&lt;br /&gt;
* 2025-04: Scott Alexander (Astral Codex Ten): [https://www.astralcodexten.com/p/the-colors-of-her-coat The Colors Of Her Coat] (response to [https://www.theintrinsicperspective.com/p/welcome-to-the-semantic-apocalypse semantic apocalypse] and semantic satiation)&lt;br /&gt;
* 2025-05: Helen Toner: [https://www.ai-frontiers.org/articles/were-arguing-about-ai-safety-wrong We’re Arguing About AI Safety Wrong]: Dynamism vs. stasis is a clearer lens for criticizing controversial AI safety prescriptions&lt;br /&gt;
* 2025-05: Joe Carlsmith: [https://joecarlsmith.substack.com/p/the-stakes-of-ai-moral-status The stakes of AI moral status]&lt;br /&gt;
&lt;br /&gt;
==Research==&lt;br /&gt;
* 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
&lt;br /&gt;
==Alignment==&lt;br /&gt;
* 2023-03: Leopold Aschenbrenner: [https://www.forourposterity.com/nobodys-on-the-ball-on-agi-alignment/ Nobody’s on the ball on AGI alignment]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2404.10636 What are human values, and how do we align AI to them?] ([https://meaningalignment.substack.com/p/0480e023-98c0-4633-a604-990d3ac880ac blog])&lt;br /&gt;
* 2025: Joe Carlsmith: [https://joecarlsmith.substack.com/p/how-do-we-solve-the-alignment-problem How do we solve the alignment problem?] Introduction to an essay series on paths to safe, useful superintelligence&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/what-is-it-to-solve-the-alignment What is it to solve the alignment problem?] Also: to avoid it? Handle it? Solve it forever? Solve it completely? ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16617671-what-is-it-to-solve-the-alignment-problem audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/when-should-we-worry-about-ai-power When should we worry about AI power-seeking?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16651469-when-should-we-worry-about-ai-power-seeking audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/paths-and-waystations-in-ai-safety Paths and waystations in AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16768804-paths-and-waystations-in-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/ai-for-ai-safety AI for AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16790183-ai-for-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/can-we-safely-automate-alignment Can we safely automate alignment research?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17069901-can-we-safely-automate-alignment-research audio version], [https://joecarlsmith.substack.com/p/video-and-transcript-of-talk-on-automating?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=162375391&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email video version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/giving-ais-safe-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=171250683&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email Giving AIs safe motivations] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17686921-giving-ais-safe-motivations audio version])&lt;br /&gt;
*# [https://joecarlsmith.com/2025/09/29/controlling-the-options-ais-can-pursue Controlling the options AIs can pursue] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17909401-controlling-the-options-ais-can-pursue audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/how-human-like-do-safe-ai-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=178666988&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email How human-like do safe AI motivations need to be?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18175429-how-human-like-do-safe-ai-motivations-need-to-be audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/building-ais-that-do-human-like-philosophy Building AIs that do human-like philosophy: AIs will face philosophical questions humans can&amp;#039;t answer for them] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18591342-building-ais-that-do-human-like-philosophy audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/on-restraining-ai-development-for?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=191385185&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email On restraining AI development for the sake of safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18869440-on-restraining-ai-development-for-the-sake-of-safety audio version])&lt;br /&gt;
* 2025-04: Dario Amodei: [https://www.darioamodei.com/post/the-urgency-of-interpretability The Urgency of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Technical==&lt;br /&gt;
* 2025-03: [https://resilience.baulab.info/docs/AI_Action_Plan_RFI.pdf AI Dominance Requires Interpretability and Standards for Transparency and Security]&lt;br /&gt;
* 2026-02: [https://www.gap-map.org/capabilities/?sort=bottlenecks Fundamental Development Gap Map v1.0]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Policy==&lt;br /&gt;
* 2015-03: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-2 Machine intelligence, part 2]&lt;br /&gt;
* 2019-07: Amanda Askell, Miles Brundage, Gillian Hadfield: [https://arxiv.org/abs/1907.04534 The Role of Cooperation in Responsible AI Development]&lt;br /&gt;
* 2025-03: Dan Hendrycks, Eric Schmidt, Alexandr Wang: [https://www.nationalsecurity.ai/ Superintelligence Strategy]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/executive-summary Executive Summary]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/introduction Introduction]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/ai-is-pivotal-for-national-security AI Is Pivotal for National Security]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/deterrence-with-mutual-assured-ai-malfunction-maim Deterrence with Mutual Assured AI Malfunction (MAIM)]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/nonproliferation Nonproliferation]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/competitiveness Competitiveness]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/conclusion Conclusion]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/appendix Appendix FAQs]&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human] ([https://keepthefuturehuman.ai/essay/ essay])&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late (video)]  (2025)&lt;br /&gt;
**# Oversight: Registration required for training &amp;gt;10&amp;lt;sup&amp;gt;25&amp;lt;/sup&amp;gt; FLOP and inference &amp;gt;10&amp;lt;sup&amp;gt;19&amp;lt;/sup&amp;gt; FLOP/s (~1,000 B200 GPUs @ $25M). Build cryptographic licensing into hardware.&lt;br /&gt;
**# Computation Limits: Ban on training models &amp;gt;10&amp;lt;sup&amp;gt;27&amp;lt;/sup&amp;gt; FLOP or inference &amp;gt;10&amp;lt;sup&amp;gt;20&amp;lt;/sup&amp;gt; FLOP/s.&lt;br /&gt;
**# Strict Liability: Hold AI companies responsible for outcomes.&lt;br /&gt;
**# Tiered Regulation: Low regulation on tool-AI, strictest regulation on AGI (general, capable, autonomous systems).&lt;br /&gt;
* 2025-04: [https://x.com/deanwball Dean W. Ball]: [https://arxiv.org/abs/2504.11501 A Framework for the Private Governance of Frontier Artificial Intelligence]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/nonproliferation-is-the-wrong-approach?source=queue Nonproliferation is the wrong approach to AI misuse]&lt;br /&gt;
* 2025-04: MIRI: [https://techgov.intelligence.org/research/ai-governance-to-avoid-extinction AI Governance to Avoid Extinction: The Strategic Landscape and Actionable Research Questions]&lt;br /&gt;
* 2025-05: [https://writing.antonleicht.me/p/the-new-ai-policy-frontier The New AI Policy Frontier]: Beyond the shortcomings of centralised control and alignment, a new school of thought on AI governance emerges. It still faces tricky politics.&lt;br /&gt;
* 2025-05: [https://uncpga.world/agi-uncpga-report/ AGI UNCPGA Report]: Governance of the Transition to Artificial General Intelligence (AGI) Urgent Considerations for the UN General Assembly: Report for the Council of Presidents of the United Nations General Assembly (UNCPGA)&lt;br /&gt;
* 2025-06: [https://writing.antonleicht.me/p/ai-and-jobs-politics-without-policy AI &amp;amp; Jobs: Politics without Policy] Political support mounts - for a policy platform that does not yet exist&lt;br /&gt;
* 2025-06: [https://x.com/littIeramblings Sarah Hastings-Woodhouse]: [https://drive.google.com/file/d/1mmdHBE6M2yiyL21-ctTuRLNH5xOFjqWm/view Safety Features for a Centralized AGI Project]&lt;br /&gt;
* 2025-07: [https://writing.antonleicht.me/p/a-moving-target A Moving Target] Why we might not be quite ready to comprehensively regulate AI, and why it matters&lt;br /&gt;
* 2025-07: [https://www-cdn.anthropic.com/0dc382a2086f6a054eeb17e8a531bd9625b8e6e5.pdf Anthropic: Build AI in America] ([https://www.anthropic.com/news/build-ai-in-america blog])&lt;br /&gt;
* 2025-12: [https://asi-prevention.com/ How middle powers may prevent the development of artificial superintelligence]&lt;br /&gt;
* 2026-03: [https://humanstatement.org/ The Pro-Human AI Declaration]&lt;br /&gt;
&lt;br /&gt;
==Restriction==&lt;br /&gt;
* 2024-05: OpenAI: [https://openai.com/index/reimagining-secure-infrastructure-for-advanced-ai/ Reimagining secure infrastructure for advanced AI] OpenAI calls for an evolution in infrastructure security to protect advanced AI &lt;br /&gt;
* 2025-07: MIRI: [https://arxiv.org/abs/2507.09801 Technical Requirements for Halting Dangerous AI Activities]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI safety]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8755</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8755"/>
		<updated>2026-03-24T15:40:38Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Science Agentic Components */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Personalities==&lt;br /&gt;
* 2026-03: [https://github.com/msitarzewski/agency-agents The Agency: AI Specialists Ready to Transform Your Workflow]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Human_brain&amp;diff=8754</id>
		<title>Human brain</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Human_brain&amp;diff=8754"/>
		<updated>2026-03-23T17:33:56Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Understanding */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Why brain is as it is=&lt;br /&gt;
* 2025-06: [https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(24)00319-X The metabolic costs of cognition]&lt;br /&gt;
&lt;br /&gt;
=How Brain Works=&lt;br /&gt;
==Predictive Coding==&lt;br /&gt;
* 2005-04: [https://royalsocietypublishing.org/doi/10.1098/rstb.2005.1622?utm_source=chatgpt.com A theory of cortical responses]&lt;br /&gt;
* 2014-09: [https://www.frontiersin.org/journals/human-neuroscience/articles/10.3389/fnhum.2014.00666/full Visual mismatch negativity: a predictive coding view]&lt;br /&gt;
* 2015-01: [https://www.sciencedirect.com/science/article/pii/S089662731401099X Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels]&lt;br /&gt;
* 2016-11: [https://www.sciencedirect.com/science/article/pii/S0896627316306997 Mismatch Receptive Fields in Mouse Visual Cortex]&lt;br /&gt;
* 2018-03: [https://www.nature.com/articles/s41598-018-21407-9 Frontal cortex function as derived from hierarchical predictive coding]&lt;br /&gt;
* 2024-02: [https://www.sciencedirect.com/science/article/pii/S0149763423004426 The empirical status of predictive coding and active inference]&lt;br /&gt;
&lt;br /&gt;
=Understanding=&lt;br /&gt;
* [https://arxiv.org/abs/2501.02950 Key-value memory in the brain]&lt;br /&gt;
* [https://helper.ipam.ucla.edu/publications/mac2024/mac2024_20152.pdf The cost of brain state transitions]&lt;br /&gt;
&lt;br /&gt;
==Brain mapping==&lt;br /&gt;
* 2024-05: [https://www.science.org/doi/10.1126/science.adk4858 A petavoxel fragment of human cerebral cortex reconstructed at nanoscale resolution] ([https://www.nature.com/articles/d41586-024-01387-9#ref-CR1 media summary])&lt;br /&gt;
* 2024-10: [https://www.nature.com/articles/s41586-024-07558-y Neuronal wiring diagram of an adult brain] ([https://www.nytimes.com/2024/10/02/science/fruit-fly-brain-mapped.html media summary]); 140,000 neurons in fruit fly brain&lt;br /&gt;
* 2024-12: [https://e11.bio/news/roadmap A roadmap to scale connectomics to entire mammalian brains]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08840-3 Functional connectomics reveals general wiring rule in mouse visual cortex] ([https://www.nature.com/articles/d41586-025-01088-x?utm_source=x&amp;amp;utm_medium=social&amp;amp;utm_campaign=nature&amp;amp;linkId=13897098 media summary])&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41586-025-08985-1 Light-microscopy-based connectomic reconstruction of mammalian brain tissue] ([https://research.google/blog/a-new-light-on-neural-connections/ blog])&lt;br /&gt;
&lt;br /&gt;
===Related===&lt;br /&gt;
* [https://v2.virtualflybrain.org 3D visualization of adult fruit fly brain]&lt;br /&gt;
&lt;br /&gt;
==Brain signal decoding==&lt;br /&gt;
* 2022-11: [https://www.biorxiv.org/content/10.1101/2022.11.18.517004v2.full.pdf High-resolution image reconstruction with latent diffusion models from human brain activity]&lt;br /&gt;
* 2023-08: [https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3002176%20 Music can be reconstructed from human auditory cortex activity using nonlinear decoding models] (intracranial EEG)&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.14030 DeWave: Discrete EEG Waves Encoding for Brain Dynamics to Text Translation] (external EEG)&lt;br /&gt;
* 2023-09: [https://www.biorxiv.org/content/10.1101/2023.09.12.557460v1 BrainLM: A foundation model for brain activity recordings]&lt;br /&gt;
* 2023-10: [https://ai.meta.com/blog/brain-ai-image-decoding-meg-magnetoencephalography/ Toward a real-time decoding of images from brain activity] (MEG)&lt;br /&gt;
* 2024-06: [https://www.biorxiv.org/content/10.1101/2024.06.04.596589v1.full.pdf PAM: Predictive Attention Mechanism for Neural Decoding of Visual Perception]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.07595 Scaling Law in Neural Data: Non-Invasive Speech Decoding with 175 Hours of EEG Data] (EEG)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15322v2 Scaling laws for decoding images from brain activity] (EEG)&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/brain-to-text-decoding-a-non-invasive-approach-via-typing/ Brain-to-Text Decoding: A Non-invasive Approach via Typing]&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/from-thought-to-action-how-a-hierarchy-of-neural-dynamics-supports-language-production/ From Thought to Action: How a Hierarchy of Neural Dynamics Supports Language Production]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41593-025-01905-6 A streaming brain-to-voice neuroprosthesis to restore naturalistic communication]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2508.18226 Disentangling the Factors of Convergence between Brains and Computer Vision Models] (fMRI and MEG)&lt;br /&gt;
&lt;br /&gt;
==Whole Brain Emulation (WBE)==&lt;br /&gt;
* 2024-09: [https://www.nature.com/articles/s41586-024-07939-3 Connectome-constrained networks predict neural activity across the fly visual system]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15745 State of Brain Emulation Report 2025]&lt;br /&gt;
&lt;br /&gt;
=Computational Analysis=&lt;br /&gt;
&lt;br /&gt;
==Computational power of human brain==&lt;br /&gt;
* 2020-09: Joe Carlsmith: [https://www.openphilanthropy.org/research/how-much-computational-power-does-it-take-to-match-the-human-brain/ How Much Computational Power Does It Take to Match the Human Brain?]&lt;br /&gt;
&lt;br /&gt;
==Comparison to computer==&lt;br /&gt;
* [https://arxiv.org/abs/2208.12032 How (and Why) to Think that the Brain is Literally a Computer]&lt;br /&gt;
* [https://www.nature.com/articles/s42256-024-00925-4 Contextual feature extraction hierarchies converge in large language models and the brain] ([https://techxplore.com/news/2024-12-llms-brain-advance.html LLMs are becoming more brain-like as they advance])&lt;br /&gt;
&lt;br /&gt;
==Biological vs. artificial neuron==&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S0896627321005018 Single cortical neurons as deep artificial neural networks]: Each biological neuron can be simulated using DNN of 5-8 layers&lt;br /&gt;
* [https://arxiv.org/abs/2305.12471 Mapping Biological Neuron Dynamics into an Interpretable Two-layer Artificial Neural Network]&lt;br /&gt;
&lt;br /&gt;
==Data processing==&lt;br /&gt;
* [https://pmc.ncbi.nlm.nih.gov/articles/PMC1564115/ How Much the Eye Tells the Brain]&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S1364661313001277 Representational geometry: integrating cognition, computation, and the brain]&lt;br /&gt;
* [https://www.nature.com/articles/s41586-024-07522-w Language is primarily a tool for communication rather than thought]&lt;br /&gt;
* [https://www.openread.academy/en/paper/reading?corpusId=513306465 The Unbearable Slowness of Being: Why do we live at 10 bits/s?] ([https://arxiv.org/abs/2408.10234 preprint])&lt;br /&gt;
&lt;br /&gt;
==Extract manifold/geometry==&lt;br /&gt;
* [https://www.science.org/doi/10.1126/science.adk8261 Selection of experience for memory by hippocampal sharp wave ripples]&lt;br /&gt;
&lt;br /&gt;
=Comparisons=&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.08708 Consciousness in Artificial Intelligence: Insights from the Science of Consciousness]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.02325 Are Biological Systems More Intelligent Than Artificial Intelligence?]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
** 2022-03: [https://www.nature.com/articles/s41593-022-01026-4 Shared computational principles for language processing in humans and deep language models]&lt;br /&gt;
** 2024-03: [https://www.nature.com/articles/s41467-024-46631-y Alignment of brain embeddings and artificial contextual embeddings in natural language points to common geometric patterns]&lt;br /&gt;
** 2025-03: [https://www.nature.com/articles/s41562-025-02105-9 A unified acoustic-to-speech-to-language embedding space captures the neural basis of natural language processing in everyday conversations]&lt;br /&gt;
* 2025-05: [https://ai.meta.com/research/publications/emergence-of-language-in-the-developing-brain/ Emergence of Language in the Developing Brain]&lt;br /&gt;
&lt;br /&gt;
==Analogies==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41562-025-02359-3 Shared sensitivity to data distribution during learning in humans and transformer networks]&lt;br /&gt;
===Speed-accuracy trade-off vs. Inference-compute===&lt;br /&gt;
* 2007: [https://psycnet.apa.org/doi/10.1037/0096-3445.136.2.217 Focusing the spotlight: individual differences in visual attention control]&lt;br /&gt;
* 2014-07: [https://www.frontiersin.org/journals/neuroscience/articles/10.3389/fnins.2014.00150/full The speed-accuracy tradeoff: history, physiology, methodology, and behavior]&lt;br /&gt;
&lt;br /&gt;
=Simulate Brain=&lt;br /&gt;
* 2023-09: [https://spj.science.org/doi/10.34133/icomputing.0055 The Digital Twin Brain: A Bridge between Biological and Artificial Intelligence]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s43588-024-00731-3 Simulation and assimilation of the digital human brain] ([https://arxiv.org/abs/2211.15963 preprint], [https://github.com/DTB-consortium/Digital_twin_brain-open code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2507.22229 TRIBE: TRImodal Brain Encoder for whole-brain fMRI response prediction]&lt;br /&gt;
&lt;br /&gt;
==See Also==&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|Simulate Humans (using LLM)]]&lt;br /&gt;
&lt;br /&gt;
=Bio-brain Inspirations for AI=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
&lt;br /&gt;
=Theories of Consciousness=&lt;br /&gt;
* [https://www.consciousnessatlas.com/ Consciousness Atlas]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|LLM Simulate Humans]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Human_brain&amp;diff=8753</id>
		<title>Human brain</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Human_brain&amp;diff=8753"/>
		<updated>2026-03-23T17:32:02Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Why Brain is at it is */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Why brain is as it is=&lt;br /&gt;
* 2025-06: [https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(24)00319-X The metabolic costs of cognition]&lt;br /&gt;
&lt;br /&gt;
=How Brain Works=&lt;br /&gt;
==Predictive Coding==&lt;br /&gt;
* 2005-04: [https://royalsocietypublishing.org/doi/10.1098/rstb.2005.1622?utm_source=chatgpt.com A theory of cortical responses]&lt;br /&gt;
* 2014-09: [https://www.frontiersin.org/journals/human-neuroscience/articles/10.3389/fnhum.2014.00666/full Visual mismatch negativity: a predictive coding view]&lt;br /&gt;
* 2015-01: [https://www.sciencedirect.com/science/article/pii/S089662731401099X Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels]&lt;br /&gt;
* 2016-11: [https://www.sciencedirect.com/science/article/pii/S0896627316306997 Mismatch Receptive Fields in Mouse Visual Cortex]&lt;br /&gt;
* 2018-03: [https://www.nature.com/articles/s41598-018-21407-9 Frontal cortex function as derived from hierarchical predictive coding]&lt;br /&gt;
* 2024-02: [https://www.sciencedirect.com/science/article/pii/S0149763423004426 The empirical status of predictive coding and active inference]&lt;br /&gt;
&lt;br /&gt;
=Understanding=&lt;br /&gt;
* [https://arxiv.org/abs/2501.02950 Key-value memory in the brain]&lt;br /&gt;
* [https://helper.ipam.ucla.edu/publications/mac2024/mac2024_20152.pdf The cost of brain state transitions]&lt;br /&gt;
&lt;br /&gt;
==Brain mapping==&lt;br /&gt;
* 2024-05: [https://www.science.org/doi/10.1126/science.adk4858 A petavoxel fragment of human cerebral cortex reconstructed at nanoscale resolution] ([https://www.nature.com/articles/d41586-024-01387-9#ref-CR1 media summary])&lt;br /&gt;
* 2024-10: [https://www.nature.com/articles/s41586-024-07558-y Neuronal wiring diagram of an adult brain] ([https://www.nytimes.com/2024/10/02/science/fruit-fly-brain-mapped.html media summary]); 140,000 neurons in fruit fly brain&lt;br /&gt;
* 2024-12: [https://e11.bio/news/roadmap A roadmap to scale connectomics to entire mammalian brains]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08840-3 Functional connectomics reveals general wiring rule in mouse visual cortex] ([https://www.nature.com/articles/d41586-025-01088-x?utm_source=x&amp;amp;utm_medium=social&amp;amp;utm_campaign=nature&amp;amp;linkId=13897098 media summary])&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41586-025-08985-1 Light-microscopy-based connectomic reconstruction of mammalian brain tissue] ([https://research.google/blog/a-new-light-on-neural-connections/ blog])&lt;br /&gt;
&lt;br /&gt;
===Related===&lt;br /&gt;
* [https://v2.virtualflybrain.org 3D visualization of adult fruit fly brain]&lt;br /&gt;
&lt;br /&gt;
==Brain signal decoding==&lt;br /&gt;
* 2022-11: [https://www.biorxiv.org/content/10.1101/2022.11.18.517004v2.full.pdf High-resolution image reconstruction with latent diffusion models from human brain activity]&lt;br /&gt;
* 2023-08: [https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3002176%20 Music can be reconstructed from human auditory cortex activity using nonlinear decoding models] (intracranial EEG)&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.14030 DeWave: Discrete EEG Waves Encoding for Brain Dynamics to Text Translation] (external EEG)&lt;br /&gt;
* 2023-09: [https://www.biorxiv.org/content/10.1101/2023.09.12.557460v1 BrainLM: A foundation model for brain activity recordings]&lt;br /&gt;
* 2023-10: [https://ai.meta.com/blog/brain-ai-image-decoding-meg-magnetoencephalography/ Toward a real-time decoding of images from brain activity] (MEG)&lt;br /&gt;
* 2024-06: [https://www.biorxiv.org/content/10.1101/2024.06.04.596589v1.full.pdf PAM: Predictive Attention Mechanism for Neural Decoding of Visual Perception]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.07595 Scaling Law in Neural Data: Non-Invasive Speech Decoding with 175 Hours of EEG Data] (EEG)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15322v2 Scaling laws for decoding images from brain activity] (EEG)&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/brain-to-text-decoding-a-non-invasive-approach-via-typing/ Brain-to-Text Decoding: A Non-invasive Approach via Typing]&lt;br /&gt;
* 2025-02: Meta: [https://ai.meta.com/research/publications/from-thought-to-action-how-a-hierarchy-of-neural-dynamics-supports-language-production/ From Thought to Action: How a Hierarchy of Neural Dynamics Supports Language Production]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41593-025-01905-6 A streaming brain-to-voice neuroprosthesis to restore naturalistic communication]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2508.18226 Disentangling the Factors of Convergence between Brains and Computer Vision Models] (fMRI and MEG)&lt;br /&gt;
&lt;br /&gt;
=Computational Analysis=&lt;br /&gt;
&lt;br /&gt;
==Computational power of human brain==&lt;br /&gt;
* 2020-09: Joe Carlsmith: [https://www.openphilanthropy.org/research/how-much-computational-power-does-it-take-to-match-the-human-brain/ How Much Computational Power Does It Take to Match the Human Brain?]&lt;br /&gt;
&lt;br /&gt;
==Comparison to computer==&lt;br /&gt;
* [https://arxiv.org/abs/2208.12032 How (and Why) to Think that the Brain is Literally a Computer]&lt;br /&gt;
* [https://www.nature.com/articles/s42256-024-00925-4 Contextual feature extraction hierarchies converge in large language models and the brain] ([https://techxplore.com/news/2024-12-llms-brain-advance.html LLMs are becoming more brain-like as they advance])&lt;br /&gt;
&lt;br /&gt;
==Biological vs. artificial neuron==&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S0896627321005018 Single cortical neurons as deep artificial neural networks]: Each biological neuron can be simulated using DNN of 5-8 layers&lt;br /&gt;
* [https://arxiv.org/abs/2305.12471 Mapping Biological Neuron Dynamics into an Interpretable Two-layer Artificial Neural Network]&lt;br /&gt;
&lt;br /&gt;
==Data processing==&lt;br /&gt;
* [https://pmc.ncbi.nlm.nih.gov/articles/PMC1564115/ How Much the Eye Tells the Brain]&lt;br /&gt;
* [https://www.sciencedirect.com/science/article/pii/S1364661313001277 Representational geometry: integrating cognition, computation, and the brain]&lt;br /&gt;
* [https://www.nature.com/articles/s41586-024-07522-w Language is primarily a tool for communication rather than thought]&lt;br /&gt;
* [https://www.openread.academy/en/paper/reading?corpusId=513306465 The Unbearable Slowness of Being: Why do we live at 10 bits/s?] ([https://arxiv.org/abs/2408.10234 preprint])&lt;br /&gt;
&lt;br /&gt;
==Extract manifold/geometry==&lt;br /&gt;
* [https://www.science.org/doi/10.1126/science.adk8261 Selection of experience for memory by hippocampal sharp wave ripples]&lt;br /&gt;
&lt;br /&gt;
=Comparisons=&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.08708 Consciousness in Artificial Intelligence: Insights from the Science of Consciousness]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.02325 Are Biological Systems More Intelligent Than Artificial Intelligence?]&lt;br /&gt;
* 2025-03: Google: [https://research.google/blog/deciphering-language-processing-in-the-human-brain-through-llm-representations/ Deciphering language processing in the human brain through LLM representations]&lt;br /&gt;
** 2022-03: [https://www.nature.com/articles/s41593-022-01026-4 Shared computational principles for language processing in humans and deep language models]&lt;br /&gt;
** 2024-03: [https://www.nature.com/articles/s41467-024-46631-y Alignment of brain embeddings and artificial contextual embeddings in natural language points to common geometric patterns]&lt;br /&gt;
** 2025-03: [https://www.nature.com/articles/s41562-025-02105-9 A unified acoustic-to-speech-to-language embedding space captures the neural basis of natural language processing in everyday conversations]&lt;br /&gt;
* 2025-05: [https://ai.meta.com/research/publications/emergence-of-language-in-the-developing-brain/ Emergence of Language in the Developing Brain]&lt;br /&gt;
&lt;br /&gt;
==Analogies==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.11536 Language models align with brain regions that represent concepts across modalities]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41562-025-02359-3 Shared sensitivity to data distribution during learning in humans and transformer networks]&lt;br /&gt;
===Speed-accuracy trade-off vs. Inference-compute===&lt;br /&gt;
* 2007: [https://psycnet.apa.org/doi/10.1037/0096-3445.136.2.217 Focusing the spotlight: individual differences in visual attention control]&lt;br /&gt;
* 2014-07: [https://www.frontiersin.org/journals/neuroscience/articles/10.3389/fnins.2014.00150/full The speed-accuracy tradeoff: history, physiology, methodology, and behavior]&lt;br /&gt;
&lt;br /&gt;
=Simulate Brain=&lt;br /&gt;
* 2023-09: [https://spj.science.org/doi/10.34133/icomputing.0055 The Digital Twin Brain: A Bridge between Biological and Artificial Intelligence]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s43588-024-00731-3 Simulation and assimilation of the digital human brain] ([https://arxiv.org/abs/2211.15963 preprint], [https://github.com/DTB-consortium/Digital_twin_brain-open code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.19814 Predicting Human Brain States with Transformer]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2507.22229 TRIBE: TRImodal Brain Encoder for whole-brain fMRI response prediction]&lt;br /&gt;
&lt;br /&gt;
==See Also==&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|Simulate Humans (using LLM)]]&lt;br /&gt;
&lt;br /&gt;
=Bio-brain Inspirations for AI=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
&lt;br /&gt;
=Theories of Consciousness=&lt;br /&gt;
* [https://www.consciousnessatlas.com/ Consciousness Atlas]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI_and_Humans#Simulate_Humans|LLM Simulate Humans]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8752</id>
		<title>AI predictions</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8752"/>
		<updated>2026-03-23T17:29:59Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Alignment */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Capability Scaling=&lt;br /&gt;
* 2019-03: Rich Sutton: [https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf The Bitter Lesson]&lt;br /&gt;
* 2020-09: Ajeya Cotra: [https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines Draft report on AI timelines]&lt;br /&gt;
* 2022-01: gwern: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis]&lt;br /&gt;
* 2023-05: Richard Ngo: [https://www.lesswrong.com/posts/BoA3agdkAzL6HQtQP/clarifying-and-predicting-agi Clarifying and predicting AGI]&lt;br /&gt;
* 2024-06: Aidan McLaughlin: [https://yellow-apartment-148.notion.site/AI-Search-The-Bitter-er-Lesson-44c11acd27294f4495c3de778cd09c8d AI Search: The Bitter-er Lesson]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.14499 Measuring AI Ability to Complete Long Tasks Measuring AI Ability to Complete Long Tasks]&lt;br /&gt;
** 2025-04: [https://peterwildeford.substack.com/p/forecaster-reacts-metrs-bombshell Forecaster reacts: METR&amp;#039;s bombshell paper about AI acceleration] New data supports an exponential AI curve, but lots of uncertainty remains&lt;br /&gt;
** 2025-04: AI Digest: [https://theaidigest.org/time-horizons A new Moore&amp;#039;s Law for AI agents]&lt;br /&gt;
[[Image:GmZHL8xWQAAtFlF.jpeg|450px]]&lt;br /&gt;
* 2025-04: [https://epoch.ai/blog/trends-in-ai-supercomputers Trends in AI Supercomputers] ([https://arxiv.org/abs/2504.16026 preprint])&lt;br /&gt;
* [https://ai-timeline.org/ The Road to AGI] (timeline visualization)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09677 The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs]&lt;br /&gt;
* 2025-09: [https://www.julian.ac/blog/2025/09/27/failing-to-understand-the-exponential-again/ Failing to Understand the Exponential, Again]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/rRbDNQLfihiHbXytf/distinguish-between-inference-scaling-and-larger-tasks-use Distinguish between inference scaling and &amp;quot;larger tasks use more compute&amp;quot;]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03992 Measuring AI R&amp;amp;D Automation] ([https://astrangeattractor.substack.com/p/measuring-ai-r-and-d-automation?triedRedirect=true blog])&lt;br /&gt;
&lt;br /&gt;
==Scaling Laws==&lt;br /&gt;
See: [[AI_understanding#Scaling_Laws|Scaling Laws]]&lt;br /&gt;
&lt;br /&gt;
==AGI Achievable==&lt;br /&gt;
* Yoshua Bengio: [https://arxiv.org/abs/2310.17688 Managing extreme AI risks amid rapid progress]&lt;br /&gt;
* Leopold Aschenbrenner: [https://situational-awareness.ai/from-gpt-4-to-agi/#Counting_the_OOMs Situational Awareness: Counting the OOMs]&lt;br /&gt;
* Richard Ngo: [https://medium.com/@richardcngo/visualizing-the-deep-learning-revolution-722098eb9c5 Visualizing the deep learning revolution]&lt;br /&gt;
* Katja Grace: [https://blog.aiimpacts.org/p/2023-ai-survey-of-2778-six-things Survey of 2,778 AI authors: six parts in pictures]&lt;br /&gt;
* Epoch AI: [https://epoch.ai/trends Machine Learning Trends]&lt;br /&gt;
* AI Digest: [https://theaidigest.org/progress-and-dangers How fast is AI improving?]&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/when-will-agi-arrive/ The case for AGI by 2030]&lt;br /&gt;
&lt;br /&gt;
==AGI Definition==&lt;br /&gt;
* 2023-11: Allan Dafoe, Shane Legg, et al.: [https://arxiv.org/abs/2311.02462 Levels of AGI for Operationalizing Progress on the Path to AGI]&lt;br /&gt;
* 2024-04: Bowen Xu: [https://arxiv.org/abs/2404.10731 What is Meant by AGI? On the Definition of Artificial General Intelligence]&lt;br /&gt;
* 2025-10: Dan Hendrycks et al.: [https://www.agidefinition.ai/paper.pdf A Definition of AGI]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07364 On the universal definition of intelligence]&lt;br /&gt;
&lt;br /&gt;
==Recursive Self Improvement (RSI)==&lt;br /&gt;
* 2026-02: [https://80000hours.org/articles/how-ai-driven-feedback-loops-could-make-things-very-crazy-very-fast/ How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
&lt;br /&gt;
==Progress Models==&lt;br /&gt;
From [http://yager-research.ca/2025/04/ai-impact-predictions/ AI Impact Predictions]:&lt;br /&gt;
&lt;br /&gt;
[[Image:AI impact models-2025 11 24.png|450px]]&lt;br /&gt;
&lt;br /&gt;
=Economic and Political=&lt;br /&gt;
* 2019-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3482150 The Impact of Artificial Intelligence on the Labor Market]&lt;br /&gt;
* 2020-06: [https://www.openphilanthropy.org/research/modeling-the-human-trajectory/ Modeling the Human Trajectory] (GDP)&lt;br /&gt;
* 2021-06: [https://www.openphilanthropy.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/ Report on Whether AI Could Drive Explosive Economic Growth]&lt;br /&gt;
* 2023-10: Marc Andreessen: [https://a16z.com/the-techno-optimist-manifesto/ The Techno-Optimist Manifesto]&lt;br /&gt;
* 2023-12: [https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html My techno-optimism]: &amp;quot;defensive acceleration&amp;quot; ([https://vitalik.eth.limo/index.html Vitalik Buterin])&lt;br /&gt;
* 2024-03: Noah Smith: [https://www.noahpinion.blog/p/plentiful-high-paying-jobs-in-the Plentiful, high-paying jobs in the age of AI: Comparative advantage is very subtle, but incredibly powerful.] ([https://x.com/liron/status/1768013030741475485 video])&lt;br /&gt;
* 2024-03: [https://doi.org/10.3386/w32255 Scenarios for the Transition to AGI] (AGI leads to wage collapse)&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-06: [https://www.frbsf.org/wp-content/uploads/AI-and-Growth-Aghion-Bunel.pdf AI and Growth: Where Do We Stand?]&lt;br /&gt;
* 2024-09: OpenAI [https://cdn.openai.com/global-affairs/openai-infra-economics-10.09.24.pdf Infrastructure is Destiny: Economic Returns on US Investment in Democratic AI]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/KFFaKu27FNugCHFmh/by-default-capital-will-matter-more-than-ever-after-agi By default, capital will matter more than ever after AGI] (L Rudolf L)&lt;br /&gt;
* 2025-01: [https://lukedrago.substack.com/p/the-intelligence-curse The Intelligence Curse]: With AGI, powerful actors will lose their incentives to invest in people&lt;br /&gt;
** Updated 2025-04: [https://intelligence-curse.ai/ The Intelligence Curse] (Luke Drago and Rudolf Laine)&lt;br /&gt;
*** [https://intelligence-curse.ai/pyramid/ Pyramid Replacement]&lt;br /&gt;
*** [https://intelligence-curse.ai/capital/ Capital, AGI, and Human Ambition]&lt;br /&gt;
*** [https://intelligence-curse.ai/defining/ Defining the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/shaping/ Shaping the Social Contract]&lt;br /&gt;
*** [https://intelligence-curse.ai/breaking/ Breaking the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/history/ History is Yours to Write]&lt;br /&gt;
* 2025-01: Microsoft: [https://blogs.microsoft.com/on-the-issues/2025/01/03/the-golden-opportunity-for-american-ai/ The Golden Opportunity for American AI]&lt;br /&gt;
* 2025-01: [https://www.maximum-progress.com/p/agi-will-not-make-labor-worthless AGI Will Not Make Labor Worthless]&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/global-affairs/ai-in-america-oai-economic-blueprint-20250113.pdf AI in America: OpenAI&amp;#039;s Economic Blueprint] ([https://openai.com/global-affairs/openais-economic-blueprint/ blog])&lt;br /&gt;
* 2025-01: [https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai How much economic growth from AI should we expect, how soon?]&lt;br /&gt;
* 2025-02: Morgan Stanley: [https://advisor.morganstanley.com/john.howard/documents/field/j/jo/john-howard/The_Humanoid_100_-_Mapping_the_Humanoid_Robot_Value_Chain.pdf The Humanoid 100: Mapping the Humanoid Robot Value Chain]&lt;br /&gt;
* 2025-02: [https://www.anthropic.com/news/the-anthropic-economic-index The Anthropic Economic Index]: [https://assets.anthropic.com/m/2e23255f1e84ca97/original/Economic_Tasks_AI_Paper.pdf Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11264 Strategic Wealth Accumulation Under Transformative AI Expectations]&lt;br /&gt;
* 2025-02: Tyler Cowen: [https://marginalrevolution.com/marginalrevolution/2025/02/why-i-think-ai-take-off-is-relatively-slow.html Why I think AI take-off is relatively slow]&lt;br /&gt;
* 2025-03: Epoch AI: [https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d Most AI value will come from broad automation, not from R&amp;amp;D]&lt;br /&gt;
** The primary economic impact of AI will be its ability to broadly automate labor&lt;br /&gt;
** Automating AI R&amp;amp;D alone likely won’t dramatically accelerate AI progress&lt;br /&gt;
** Fully automating R&amp;amp;D requires a very broad set of abilities&lt;br /&gt;
** AI takeoff will likely be diffuse and salient&lt;br /&gt;
* 2025-03: [https://www.anthropic.com/news/anthropic-economic-index-insights-from-claude-sonnet-3-7 Anthropic Economic Index: Insights from Claude 3.7 Sonnet]&lt;br /&gt;
* 2025-04: [https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality Will there be extreme inequality from AI?]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/research/impact-software-development Anthropic Economic Index: AI’s Impact on Software Development]&lt;br /&gt;
* 2025-05: [https://www.theguardian.com/books/2025/may/04/the-big-idea-can-we-stop-ai-making-humans-obsolete Better at everything: how AI could make human beings irrelevant]&lt;br /&gt;
* 2025-05: Forethought: [https://www.forethought.org/research/the-industrial-explosion The Industrial Explosion]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20273 Ten Principles of AI Agent Economics]&lt;br /&gt;
* 2025-07: [https://substack.com/home/post/p-167879696 What Economists Get Wrong about AI] They ignore innovation effects, use outdated capability assumptions, and miss the robotics revolution&lt;br /&gt;
* 2025-07: [https://www.nber.org/books-and-chapters/economics-transformative-ai/we-wont-be-missed-work-and-growth-era-agi We Won&amp;#039;t Be Missed: Work and Growth in the Era of AGI]&lt;br /&gt;
* 2025-07: [https://www.nber.org/papers/w34034 The Economics of Bicycles for the Mind]&lt;br /&gt;
* 2025-09: [https://conference.nber.org/conf_papers/f227491.pdf Genius on Demand: The Value of Transformative Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://peterwildeford.substack.com/p/ai-is-probably-not-a-bubble AI is probably not a bubble: AI companies have revenue, demand, and paths to immense value]&lt;br /&gt;
* 2025-11: [https://windowsontheory.org/2025/11/04/thoughts-by-a-non-economist-on-ai-and-economics/ Thoughts by a non-economist on AI and economics]&lt;br /&gt;
* 2025-11: [https://www.nber.org/papers/w34444 Artificial Intelligence, Competition, and Welfare]&lt;br /&gt;
* 2025-11: [https://www.anthropic.com/research/estimating-productivity-gains Estimating AI productivity gains from Claude conversations] (Anthropic)&lt;br /&gt;
* 2025-12: [https://benjamintodd.substack.com/p/how-ai-driven-feedback-loops-could How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
* 2025-12: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth] (Philip Trammell and Leopold Aschenbrenner)&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Anthropic Economic Index: new building blocks for understanding AI use]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/economic-index-primitives Anthropic Economic Index report: economic primitives]&lt;br /&gt;
* 2026-02: Nate Silver: [https://www.natesilver.net/p/the-singularity-wont-be-gentle The singularity won&amp;#039;t be gentle: If AI is even half as transformational as Silicon Valley assumes, politics will never be the same again]&lt;br /&gt;
&lt;br /&gt;
==Job Loss==&lt;br /&gt;
* 2023-03: [https://arxiv.org/pdf/2303.10130 GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models]&lt;br /&gt;
** 2023-03: [https://www.livemint.com/news/world/these-jobs-are-most-at-risk-due-to-chatgpt-as-per-openai-study-11679358453267.html These jobs are most at risk due to ChatGPT, as per OpenAI study]&lt;br /&gt;
* 2023-08: [https://dx.doi.org/10.2139/ssrn.4527336 The Short-Term Effects of Generative Artificial Intelligence on Employment: Evidence from an Online Labor Market]&lt;br /&gt;
** [https://x.com/jburnmurdoch/status/1722938749519077688 Freelancer sector shrinking]&lt;br /&gt;
[[Image:F-kVQuvWkAAemkr.png|400px]]&lt;br /&gt;
* 2023-09: [https://global-uploads.webflow.com/64d5f73a7fc5e8a240310c4d/650a128a34386a1206b6506c_FINAL%20Briefing%20-%20Adoption%20of%20Automation%20and%20AI%20in%20the%20UK.pdf What drives UK firms to adopt AI and robotics, and what are the consequences for jobs?]&lt;br /&gt;
** [https://www.digitalinformationworld.com/2023/09/78-of-companies-say-ai-created-more-jobs.html 78% of Companies Say AI Created More Jobs]&lt;br /&gt;
* 2023-11: [https://theaipi.org/ai-interactive-map/ New Analysis Shows Over 20% of US Jobs Significantly Exposed to AI Automation In the Near Future]&lt;br /&gt;
* 2024-01: [https://techcrunch.com/2024/01/09/duolingo-cut-10-of-its-contractor-workforce-as-the-company-embraces-ai/ Duolingo cuts 10% of its contractor workforce as the company embraces AI]&lt;br /&gt;
* 2024-02: [https://www.pwc.com/gx/en/issues/c-suite-insights/the-leadership-agenda/gen-ai-is-a-tool-for-growth-not-just-efficiency.html#:~:text=One%20out%20of%20every%20four%20of%20the%204%2C702,to%20accomplish%20the%20same%20tasks%20with%20fewer%20workers Gen AI is a tool for growth, not just efficiency: Tech CEOs are investing to build their workforce and capitalise on new opportunities from generative AI. That’s a sharp contrast to how their peers view it.]&lt;br /&gt;
* 2024-04: [https://www.nytimes.com/2024/04/10/business/investment-banking-jobs-artificial-intelligence.html AI is Poised to Replace the Entry-Level Grunt Work of a Wall Street Career]&lt;br /&gt;
* 2024-07: [https://www.wired.com/story/ai-is-already-taking-jobs-in-the-video-game-industry/ AI Is Already Taking Jobs in the Video Game Industry]: A WIRED investigation finds that major players like Activision Blizzard, which recently laid off scores of workers, are using generative AI for game development&lt;br /&gt;
* 2024-08: [https://www.bbc.com/news/articles/c80e1gp9m9zo Klarna: AI lets us cut thousands of jobs - but pay more]&lt;br /&gt;
* 2025-01: [https://scholarspace.manoa.hawaii.edu/server/api/core/bitstreams/4f39375d-59c2-4c4a-b394-f3eed7858c80/content AI and Freelancers: Has the Inflection Point Arrived?]&lt;br /&gt;
* 2025-01: [https://www.aporiamagazine.com/p/yes-youre-going-to-be-replaced Yes, you&amp;#039;re going to be replaced: So much cope about AI]&lt;br /&gt;
* 2025-03: [https://commonplace.org/2025/03/20/will-ai-automate-away-your-job/ Will AI Automate Away Your Job? The time-horizon model explains the future of the technology]&lt;br /&gt;
* 2025-05: [https://www.forbes.com/sites/jackkelly/2025/05/04/its-time-to-get-concerned-klarna-ups-duolingo-cisco-and-many-other-companies-are-replacing-workers-with-ai/ It’s Time To Get Concerned, Klarna, UPS, Duolingo, Cisco, And Many Other Companies Are Replacing Workers With AI]&lt;br /&gt;
* 2025-05: [https://time.com/7289692/when-ai-replaces-workers/ What Happens When AI Replaces Workers?]&lt;br /&gt;
* 2025-05: [https://www.oxfordeconomics.com/resource/educated-but-unemployed-a-rising-reality-for-us-college-grads/ Educated but unemployed, a rising reality for US college grads] Structural shifts in tech hiring and the growing impact of AI are driving higher unemployment among recent college graduates&lt;br /&gt;
* 2025-05: NY Times: [https://www.nytimes.com/2025/05/30/technology/ai-jobs-college-graduates.html?unlocked_article_code=1.LE8.LlC6.eT5XcpA9hxC2&amp;amp;smid=url-share For Some Recent Graduates, the A.I. Job Apocalypse May Already Be Here] The unemployment rate for recent college graduates has jumped as companies try to replace entry-level workers with artificial intelligence&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/skills-ai-makes-valuable/ How not to lose your job to AI] The skills AI will make more valuable (and how to learn them)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06576 Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce]&lt;br /&gt;
[[Image:0dab4c86-882d-4095-9d12-d19684ed5184 675x680.png|300px]]&lt;br /&gt;
* 2025-07: Harvard Business Review: [https://hbr.org/2025/06/what-gets-measured-ai-will-automate What Gets Measured, AI Will Automate]&lt;br /&gt;
* 2025-08: [https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/ Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5560401 Performance or Principle: Resistance to Artificial Intelligence in the U.S. Labor Market]&lt;br /&gt;
* 2025-10: [https://www.siliconcontinent.com/p/the-ai-becker-problem The AI Becker problem: Who will train the next generation?]&lt;br /&gt;
* 2026-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6134506 AI, Automation, and Expertise]&lt;br /&gt;
* 2026-02: [https://arachnemag.substack.com/p/the-jevons-paradox-for-intelligence The Jevons Paradox for Intelligence: Fears of AI-induced job loss could not be more wrong]&lt;br /&gt;
&lt;br /&gt;
==Productivity Impact==&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2026-02: [https://www.ft.com/content/4b51d0b4-bbfe-4f05-b50a-1d485d419dc5 The AI productivity take-off is finally visible] ([https://x.com/erikbryn/status/2023075588974735869?s=20 Erik Brynjolfsson])&lt;br /&gt;
** Businesses are finally beginning to reap some of AI&amp;#039;s benefits.&lt;br /&gt;
* 2026-02: New York Times: [https://www.nytimes.com/2026/02/18/opinion/ai-software.html The A.I. Disruption We’ve Been Waiting for Has Arrived]&lt;br /&gt;
&lt;br /&gt;
==National Security==&lt;br /&gt;
* 2025-04: Jeremie Harris and Edouard Harris: [https://superintelligence.gladstone.ai/ America’s Superintelligence Project]&lt;br /&gt;
&lt;br /&gt;
==AI Manhattan Project==&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-10: [https://thezvi.substack.com/p/ai-88-thanks-for-the-memos?open=false#%C2%A7thanks-for-the-memos-introduction-and-competitiveness White House Memo calls for action on AI]&lt;br /&gt;
* 2024-11: [https://www.uscc.gov/annual-report/2024-annual-report-congress 2024 Annual Report to Congress]: [https://www.reuters.com/technology/artificial-intelligence/us-government-commission-pushes-manhattan-project-style-ai-initiative-2024-11-19/ calls] for &amp;quot;Manhattan Project-style&amp;quot; effort&lt;br /&gt;
* 2025-05-29: [https://x.com/ENERGY/status/1928085878561272223 DoE Tweet]: &amp;quot;AI is the next Manhattan Project, and THE UNITED STATES WILL WIN. 🇺🇸&amp;quot;&lt;br /&gt;
* 2025-07: [https://epoch.ai/gradient-updates/how-big-could-an-ai-manhattan-project-get How big could an “AI Manhattan Project” get?]&lt;br /&gt;
&lt;br /&gt;
=Near-term=&lt;br /&gt;
* 2021-08: Daniel Kokotajlo: [https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like What 2026 looks like]&lt;br /&gt;
* 2025-02: Sam Altman: [https://blog.samaltman.com/three-observations Three Observations]&lt;br /&gt;
*# The intelligence of an AI model roughly equals the log of the resources used to train and run it.&lt;br /&gt;
*# The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use&lt;br /&gt;
*# The socioeconomic value of linearly increasing intelligence is super-exponential in nature&lt;br /&gt;
* 2025-03: [https://www.pathwaysai.org/p/glimpses-of-ai-progess Glimpses of AI Progress: Mental models for fast times]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41598-025-92190-7 Navigating artificial general intelligence development: societal, technological, ethical, and brain-inspired pathways]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
** 2025-07: Video: [https://www.youtube.com/watch?v=5KVDDfAkRgc Are We 3 Years From AI Disaster? A Rigorous Forecast]&lt;br /&gt;
* 2025-04: Stanford HAI: [https://hai-production.s3.amazonaws.com/files/hai_ai_index_report_2025.pdf Artificial Intelligence Index Report 2025]&lt;br /&gt;
* 2025-04: Arvind Narayananand Sayash Kapoor: [https://kfai-documents.s3.amazonaws.com/documents/c3cac5a2a7/AI-as-Normal-Technology---Narayanan---Kapoor.pdf AI as Normal Technology]&lt;br /&gt;
* 2025-04: Dwarkesh Patel: [https://www.dwarkesh.com/p/questions-about-ai Questions about the Future of AI]&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: IdeaFoundry: [https://ideafoundry.substack.com/p/evolution-vs-extinction-the-choice Evolution vs. Extinction: The Choice is Ours] The next 18 months will decide whether AI ends us or evolves us&lt;br /&gt;
* 2025-07: [https://cfg.eu/advanced-ai-possible-futures/ Advanced AI: Possible futures] Five scenarios for how the AI-transition could unfold&lt;br /&gt;
* 2025-11: [https://android-dreams.ai/ Android Dreams]&lt;br /&gt;
* 2026-02: [https://www.citriniresearch.com/ Citrini]: [https://www.citriniresearch.com/p/2028gic The 2028 Global Intelligence Crisis: A Thought Exercise in Financial History, from the Future]&lt;br /&gt;
&lt;br /&gt;
==Insightful Analysis of Current State==&lt;br /&gt;
* 2025-11: Andy Masley: [https://andymasley.substack.com/p/the-lump-of-cognition-fallacy The lump of cognition fallacy: The extended mind as the advance of civilization]&lt;br /&gt;
* 2026-02: Eric Jang: [https://evjang.com/2026/02/04/rocks.html As Rocks May Think]&lt;br /&gt;
* 2026-02: Matt Shumer: [https://x.com/mattshumer_/status/2021256989876109403 Something Big Is Happening]&lt;br /&gt;
* 2026-02: Minh Pham: [https://x.com/buckeyevn/status/2014171253045960803?s=20 Why Most Agent Harnesses Are Not Bitter Lesson Pilled]&lt;br /&gt;
&lt;br /&gt;
=Overall=&lt;br /&gt;
* 1993: [https://en.wikipedia.org/wiki/Vernor_Vinge Vernor Vinge]: [https://edoras.sdsu.edu/~vinge/misc/singularity.html The Coming Technological Singularity: How to Survive in the Post-Human Era]&lt;br /&gt;
* 2025-03: Kevin Roose (New York Times): [https://www.nytimes.com/2025/03/14/technology/why-im-feeling-the-agi.html?unlocked_article_code=1.304.TIEy.SmNhKYO4e9c7&amp;amp;smid=url-share Powerful A.I. Is Coming. We’re Not Ready.] Three arguments for taking progress toward artificial general intelligence, or A.G.I., more seriously — whether you’re an optimist or a pessimist.&lt;br /&gt;
* 2025-03: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/thoughts-on-future-ai.html My Thoughts on the Future of &amp;quot;AI&amp;quot;]: &amp;quot;I have very wide error bars on the potential future of large language models, and I think you should too.&amp;quot;&lt;br /&gt;
* 2025-06: Sam Altman: [https://blog.samaltman.com/the-gentle-singularity The Gentle Singularity]&lt;br /&gt;
&lt;br /&gt;
==Surveys of Opinions/Predictions==&lt;br /&gt;
* 2016-06: [https://aiimpacts.org/2016-expert-survey-on-progress-in-ai/ 2016 Expert Survey on Progress in AI]&lt;br /&gt;
** 2023-03: [https://aiimpacts.org/scoring-forecasts-from-the-2016-expert-survey-on-progress-in-ai/ Scoring forecasts from the 2016 “Expert Survey on Progress in AI”]&lt;br /&gt;
* 2022-10: Forecasting Research Institute: [https://forecastingresearch.org/near-term-xpt-accuracy Assessing Near-Term Accuracy in the Existential Risk Persuasion Tournament]&lt;br /&gt;
** 2025-09: Ethan Mollick: [https://x.com/emollick/status/1962859757674344823 Progress is ahead of expectations]&lt;br /&gt;
* 2023-08: [https://wiki.aiimpacts.org/ai_timelines/predictions_of_human-level_ai_timelines/ai_timeline_surveys/2023_expert_survey_on_progress_in_ai 2023 Expert Survey on Progress in AI]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.02843 Thousands of AI Authors on the Future of AI]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14870 Why do Experts Disagree on Existential Risk and P(doom)? A Survey of AI Experts]&lt;br /&gt;
* 2025-02: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/forecasting-ai-2025-update.html AI forecasting retrospective: you&amp;#039;re (probably) over-confident]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/long-timelines-to-advanced-ai-have &amp;quot;Long&amp;quot; timelines to advanced AI have gotten crazy short]&lt;br /&gt;
* 2025-05: [https://theaidigest.org/ai2025-analysis-may AI 2025 Forecasts - May Update]&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41598-026-39070-w Lay beliefs about the badness, likelihood, and importance of human extinction]&lt;br /&gt;
&lt;br /&gt;
==Bad Outcomes==&lt;br /&gt;
* [https://pauseai.info/pdoom List of p(doom) values]&lt;br /&gt;
* 2019-03: [https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like What failure looks like]&lt;br /&gt;
* 2023-03: gwern: [https://gwern.net/fiction/clippy It Looks Like You’re Trying To Take Over The World]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16946 Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development] ([https://gradual-disempowerment.ai/ web version])&lt;br /&gt;
** 2025-02: [https://thezvi.substack.com/p/the-risk-of-gradual-disempowerment The Risk of Gradual Disempowerment from AI]&lt;br /&gt;
** 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-09: [https://doctrines.ai/ The three main doctrines on the future of AI]&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Dominance doctrine:&amp;#039;&amp;#039;&amp;#039; First actor to create advanced AI will attain overwhelming strategic superiority&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Extinction doctrine:&amp;#039;&amp;#039;&amp;#039; Humanity will lose control of ASI, leading to extinction or permanent disempowerment&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Replacement doctrine:&amp;#039;&amp;#039;&amp;#039; AI will automate human tasks, but without fundamentally reshaping or ending civilization&lt;br /&gt;
* 2025-09: Sean ÓhÉigeartaigh: [https://www.cambridge.org/core/journals/cambridge-prisms-extinction/article/extinction-of-the-human-species-what-could-cause-it-and-how-likely-is-it-to-occur/D8816A79BEF5A4C30A3E44FD8D768622 Extinction of the human species: What could cause it and how likely is it to occur?]&lt;br /&gt;
&lt;br /&gt;
==Intelligence Explosion==&lt;br /&gt;
* 2023-06: [https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/ What a Compute-Centric Framework Says About Takeoff Speeds]&lt;br /&gt;
** [https://takeoffspeeds.com/ takeoffspeeds.com simulator]&lt;br /&gt;
* 2025-02: [https://www.forethought.org/research/three-types-of-intelligence-explosion Three Types of Intelligence Explosion]&lt;br /&gt;
* 2025-03: Future of Life Institute: [https://futureoflife.org/ai/are-we-close-to-an-intelligence-explosion/ Are we close to an intelligence explosion?] AIs are inching ever-closer to a critical threshold. Beyond this threshold lie great risks—but crossing it is not inevitable.&lt;br /&gt;
* 2025-03: Forethought: [https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion Will AI R&amp;amp;D Automation Cause a Software Intelligence Explosion?]&lt;br /&gt;
[[Image:Gm-1jugbYAAtq Y.jpeg|450px]]&lt;br /&gt;
* 2025-05: [https://www.thelastinvention.ai/ The Last Invention] Why Humanity’s Final Creation Changes Everything&lt;br /&gt;
* 2025-08: [https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be How quick and big would a software intelligence explosion be?]&lt;br /&gt;
&lt;br /&gt;
==Superintelligence==&lt;br /&gt;
* 2024-10: [http://yager-research.ca/2024/10/how-smart-will-asi-be/ How Smart will ASI be?]&lt;br /&gt;
* 2024-11: [http://yager-research.ca/2024/11/concise-argument-for-asi-risk/ Concise Argument for ASI Risk]&lt;br /&gt;
* 2025-03: [https://dynomight.net/smart/ Limits of smart]&lt;br /&gt;
* 2025-05: [https://timfduffy.substack.com/p/the-limits-of-superintelligence?manualredirect= The Limits of Superintelligence]&lt;br /&gt;
&lt;br /&gt;
==Long-range/Philosophy==&lt;br /&gt;
* 2023-03: Dan Hendrycks: [https://arxiv.org/abs/2303.16200 Natural Selection Favors AIs over Humans]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2025-01: [https://longerramblings.substack.com/p/a-defence-of-slowness-at-the-end A defence of slowness at the end of the world]&lt;br /&gt;
&lt;br /&gt;
=Positives &amp;amp; Optimism=&lt;br /&gt;
==Science &amp;amp; Technology Improvements==&lt;br /&gt;
* 2023-05: [https://www.planned-obsolescence.org/author/kelsey/ Kelsey Piper]: [https://www.planned-obsolescence.org/the-costs-of-caution/ The costs of caution]&lt;br /&gt;
* 2024-09: Sam Altman: [https://ia.samaltman.com/ The Intelligence Age]&lt;br /&gt;
* 2024-10: Dario Amodei: [https://darioamodei.com/machines-of-loving-grace Machines of Loving Grace]&lt;br /&gt;
* 2024-11: Google DeepMind: [https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery A new golden age of discovery]&lt;br /&gt;
* 2025-03: [https://finmoorhouse.com/ Fin Moorhouse], [https://www.williammacaskill.com/ Will MacAskill]: [https://www.forethought.org/research/preparing-for-the-intelligence-explosion Preparing for the Intelligence Explosion]&lt;br /&gt;
&lt;br /&gt;
==Social==&lt;br /&gt;
* 2025-09: [https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale Coasean Bargaining at Scale]: Decentralization, coordination, and co-existence with AGI&lt;br /&gt;
* 2025-10: [https://www.nber.org/system/files/chapters/c15309/c15309.pdf#page=15.23 The Coasean Singularity? Demand, Supply, and Market Design with AI Agents]&lt;br /&gt;
&lt;br /&gt;
==Post-scarcity Society==&lt;br /&gt;
* 2004: Eliezer Yudkowsky (MIRI): [https://intelligence.org/files/CEV.pdf Coherent Extrapolated Volition] and [https://www.lesswrong.com/s/d3WgHDBAPYYScp5Em/p/K4aGvLnHvYgX9pZHS Fun Theory]&lt;br /&gt;
* 2019: John Danaher: [https://www.jstor.org/stable/j.ctvn5txpc Automation and Utopia: Human Flourishing in a World Without Work]&lt;br /&gt;
&lt;br /&gt;
==The Grand Tradeoff==&lt;br /&gt;
* 2026-02: Nick Bostrom: [https://nickbostrom.com/optimal.pdf Optimal Timing for Superintelligence: Mundane Considerations for Existing People]&lt;br /&gt;
&lt;br /&gt;
=Plans=&lt;br /&gt;
* [https://www.narrowpath.co/ A Narrow Path: How to Secure our Future]&lt;br /&gt;
* Marius Hobbhahn: [https://www.lesswrong.com/posts/bb5Tnjdrptu89rcyY/what-s-the-short-timeline-plan What’s the short timeline plan?]&lt;br /&gt;
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]&lt;br /&gt;
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]&lt;br /&gt;
* [https://controlai.com/ Control AI]: [https://controlai.com/dip The Direct Institutional Plan] &lt;br /&gt;
* Luke Drago and L Rudolf L: [https://lukedrago.substack.com/p/the-use-of-knowledge-in-agi-society?triedRedirect=true The use of knowledge in (AGI) society]: How to build to break the [https://lukedrago.substack.com/p/the-intelligence-curse intelligence curse]&lt;br /&gt;
* [https://www.agisocialcontract.org/ AGI Social Contract]&lt;br /&gt;
** [https://www.agisocialcontract.org/forging-a-new-agi-social-contract Forging A New AGI Social Contract]&lt;br /&gt;
* Yoshua Bengio: [https://time.com/7283507/safer-ai-development/ A Potential Path to Safer AI Development]&lt;br /&gt;
** 2025-02: [https://arxiv.org/abs/2502.15657 Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?]&lt;br /&gt;
* 2026-01: Dario Amodei: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/vjAM7F8vMZS7oRrrh/how-do-we-more-safely-defer-to-ais How do we (more) safely defer to AIs?]&lt;br /&gt;
&lt;br /&gt;
==Philosophy==&lt;br /&gt;
* [https://danfaggella.com/ Dan Faggella]:&lt;br /&gt;
** 2018-07: [https://danfaggella.com/moral-singularity/ Moral Singularity – Unpredictable Values Bodes Poorly for Humanity]&lt;br /&gt;
** 2025-02: [https://danfaggella.com/bend/ There is No Pause – We Must Bend the Posthuman Trajectory]&lt;br /&gt;
* Joe Carlsmith: 2024: [https://joecarlsmith.com/2024/01/02/otherness-and-control-in-the-age-of-agi Otherness and control in the age of AGI]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/02/gentleness-and-the-artificial-other Gentleness and the artificial Other]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/04/deep-atheism-and-ai-risk Deep atheism and AI risk]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/08/when-yang-goes-wrong When “yang” goes wrong]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/09/does-ai-risk-other-the-ais Does AI risk “other” the AIs?]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/11/an-even-deeper-atheism An even deeper atheism]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy Being nicer than Clippy]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/18/on-the-abolition-of-man On the abolition of man]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/21/on-green On green]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/25/on-attunement On attunement]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/06/18/loving-a-world-you-dont-trust Loving a world you don’t trust]&lt;br /&gt;
* Anthony Aguirre:&lt;br /&gt;
** [https://x.com/AnthonyNAguirre/status/1898023049930457468 2025-03]: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
[[Image:GlchEeObwAQ88NK.jpeg|300px]]&lt;br /&gt;
* 2025-04: Scott Alexander (Astral Codex Ten): [https://www.astralcodexten.com/p/the-colors-of-her-coat The Colors Of Her Coat] (response to [https://www.theintrinsicperspective.com/p/welcome-to-the-semantic-apocalypse semantic apocalypse] and semantic satiation)&lt;br /&gt;
* 2025-05: Helen Toner: [https://www.ai-frontiers.org/articles/were-arguing-about-ai-safety-wrong We’re Arguing About AI Safety Wrong]: Dynamism vs. stasis is a clearer lens for criticizing controversial AI safety prescriptions&lt;br /&gt;
* 2025-05: Joe Carlsmith: [https://joecarlsmith.substack.com/p/the-stakes-of-ai-moral-status The stakes of AI moral status]&lt;br /&gt;
&lt;br /&gt;
==Research==&lt;br /&gt;
* 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
&lt;br /&gt;
==Alignment==&lt;br /&gt;
* 2023-03: Leopold Aschenbrenner: [https://www.forourposterity.com/nobodys-on-the-ball-on-agi-alignment/ Nobody’s on the ball on AGI alignment]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2404.10636 What are human values, and how do we align AI to them?] ([https://meaningalignment.substack.com/p/0480e023-98c0-4633-a604-990d3ac880ac blog])&lt;br /&gt;
* 2025: Joe Carlsmith: [https://joecarlsmith.substack.com/p/how-do-we-solve-the-alignment-problem How do we solve the alignment problem?] Introduction to an essay series on paths to safe, useful superintelligence&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/what-is-it-to-solve-the-alignment What is it to solve the alignment problem?] Also: to avoid it? Handle it? Solve it forever? Solve it completely? ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16617671-what-is-it-to-solve-the-alignment-problem audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/when-should-we-worry-about-ai-power When should we worry about AI power-seeking?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16651469-when-should-we-worry-about-ai-power-seeking audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/paths-and-waystations-in-ai-safety Paths and waystations in AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16768804-paths-and-waystations-in-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/ai-for-ai-safety AI for AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16790183-ai-for-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/can-we-safely-automate-alignment Can we safely automate alignment research?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17069901-can-we-safely-automate-alignment-research audio version], [https://joecarlsmith.substack.com/p/video-and-transcript-of-talk-on-automating?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=162375391&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email video version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/giving-ais-safe-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=171250683&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email Giving AIs safe motivations] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17686921-giving-ais-safe-motivations audio version])&lt;br /&gt;
*# [https://joecarlsmith.com/2025/09/29/controlling-the-options-ais-can-pursue Controlling the options AIs can pursue] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17909401-controlling-the-options-ais-can-pursue audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/how-human-like-do-safe-ai-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=178666988&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email How human-like do safe AI motivations need to be?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18175429-how-human-like-do-safe-ai-motivations-need-to-be audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/building-ais-that-do-human-like-philosophy Building AIs that do human-like philosophy: AIs will face philosophical questions humans can&amp;#039;t answer for them] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18591342-building-ais-that-do-human-like-philosophy audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/on-restraining-ai-development-for?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=191385185&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email On restraining AI development for the sake of safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18869440-on-restraining-ai-development-for-the-sake-of-safety audio version])&lt;br /&gt;
* 2025-04: Dario Amodei: [https://www.darioamodei.com/post/the-urgency-of-interpretability The Urgency of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Technical==&lt;br /&gt;
* 2025-03: [https://resilience.baulab.info/docs/AI_Action_Plan_RFI.pdf AI Dominance Requires Interpretability and Standards for Transparency and Security]&lt;br /&gt;
* 2026-02: [https://www.gap-map.org/capabilities/?sort=bottlenecks Fundamental Development Gap Map v1.0]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Policy==&lt;br /&gt;
* 2015-03: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-2 Machine intelligence, part 2]&lt;br /&gt;
* 2019-07: Amanda Askell, Miles Brundage, Gillian Hadfield: [https://arxiv.org/abs/1907.04534 The Role of Cooperation in Responsible AI Development]&lt;br /&gt;
* 2025-03: Dan Hendrycks, Eric Schmidt, Alexandr Wang: [https://www.nationalsecurity.ai/ Superintelligence Strategy]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/executive-summary Executive Summary]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/introduction Introduction]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/ai-is-pivotal-for-national-security AI Is Pivotal for National Security]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/deterrence-with-mutual-assured-ai-malfunction-maim Deterrence with Mutual Assured AI Malfunction (MAIM)]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/nonproliferation Nonproliferation]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/competitiveness Competitiveness]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/conclusion Conclusion]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/appendix Appendix FAQs]&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human] ([https://keepthefuturehuman.ai/essay/ essay])&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late (video)]  (2025)&lt;br /&gt;
**# Oversight: Registration required for training &amp;gt;10&amp;lt;sup&amp;gt;25&amp;lt;/sup&amp;gt; FLOP and inference &amp;gt;10&amp;lt;sup&amp;gt;19&amp;lt;/sup&amp;gt; FLOP/s (~1,000 B200 GPUs @ $25M). Build cryptographic licensing into hardware.&lt;br /&gt;
**# Computation Limits: Ban on training models &amp;gt;10&amp;lt;sup&amp;gt;27&amp;lt;/sup&amp;gt; FLOP or inference &amp;gt;10&amp;lt;sup&amp;gt;20&amp;lt;/sup&amp;gt; FLOP/s.&lt;br /&gt;
**# Strict Liability: Hold AI companies responsible for outcomes.&lt;br /&gt;
**# Tiered Regulation: Low regulation on tool-AI, strictest regulation on AGI (general, capable, autonomous systems).&lt;br /&gt;
* 2025-04: [https://x.com/deanwball Dean W. Ball]: [https://arxiv.org/abs/2504.11501 A Framework for the Private Governance of Frontier Artificial Intelligence]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/nonproliferation-is-the-wrong-approach?source=queue Nonproliferation is the wrong approach to AI misuse]&lt;br /&gt;
* 2025-04: MIRI: [https://techgov.intelligence.org/research/ai-governance-to-avoid-extinction AI Governance to Avoid Extinction: The Strategic Landscape and Actionable Research Questions]&lt;br /&gt;
* 2025-05: [https://writing.antonleicht.me/p/the-new-ai-policy-frontier The New AI Policy Frontier]: Beyond the shortcomings of centralised control and alignment, a new school of thought on AI governance emerges. It still faces tricky politics.&lt;br /&gt;
* 2025-05: [https://uncpga.world/agi-uncpga-report/ AGI UNCPGA Report]: Governance of the Transition to Artificial General Intelligence (AGI) Urgent Considerations for the UN General Assembly: Report for the Council of Presidents of the United Nations General Assembly (UNCPGA)&lt;br /&gt;
* 2025-06: [https://writing.antonleicht.me/p/ai-and-jobs-politics-without-policy AI &amp;amp; Jobs: Politics without Policy] Political support mounts - for a policy platform that does not yet exist&lt;br /&gt;
* 2025-06: [https://x.com/littIeramblings Sarah Hastings-Woodhouse]: [https://drive.google.com/file/d/1mmdHBE6M2yiyL21-ctTuRLNH5xOFjqWm/view Safety Features for a Centralized AGI Project]&lt;br /&gt;
* 2025-07: [https://writing.antonleicht.me/p/a-moving-target A Moving Target] Why we might not be quite ready to comprehensively regulate AI, and why it matters&lt;br /&gt;
* 2025-07: [https://www-cdn.anthropic.com/0dc382a2086f6a054eeb17e8a531bd9625b8e6e5.pdf Anthropic: Build AI in America] ([https://www.anthropic.com/news/build-ai-in-america blog])&lt;br /&gt;
* 2025-12: [https://asi-prevention.com/ How middle powers may prevent the development of artificial superintelligence]&lt;br /&gt;
* 2026-03: [https://humanstatement.org/ The Pro-Human AI Declaration]&lt;br /&gt;
&lt;br /&gt;
==Restriction==&lt;br /&gt;
* 2024-05: OpenAI: [https://openai.com/index/reimagining-secure-infrastructure-for-advanced-ai/ Reimagining secure infrastructure for advanced AI] OpenAI calls for an evolution in infrastructure security to protect advanced AI &lt;br /&gt;
* 2025-07: MIRI: [https://arxiv.org/abs/2507.09801 Technical Requirements for Halting Dangerous AI Activities]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI safety]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8751</id>
		<title>AI predictions</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_predictions&amp;diff=8751"/>
		<updated>2026-03-23T17:29:16Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Alignment */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Capability Scaling=&lt;br /&gt;
* 2019-03: Rich Sutton: [https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf The Bitter Lesson]&lt;br /&gt;
* 2020-09: Ajeya Cotra: [https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines Draft report on AI timelines]&lt;br /&gt;
* 2022-01: gwern: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis]&lt;br /&gt;
* 2023-05: Richard Ngo: [https://www.lesswrong.com/posts/BoA3agdkAzL6HQtQP/clarifying-and-predicting-agi Clarifying and predicting AGI]&lt;br /&gt;
* 2024-06: Aidan McLaughlin: [https://yellow-apartment-148.notion.site/AI-Search-The-Bitter-er-Lesson-44c11acd27294f4495c3de778cd09c8d AI Search: The Bitter-er Lesson]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.14499 Measuring AI Ability to Complete Long Tasks Measuring AI Ability to Complete Long Tasks]&lt;br /&gt;
** 2025-04: [https://peterwildeford.substack.com/p/forecaster-reacts-metrs-bombshell Forecaster reacts: METR&amp;#039;s bombshell paper about AI acceleration] New data supports an exponential AI curve, but lots of uncertainty remains&lt;br /&gt;
** 2025-04: AI Digest: [https://theaidigest.org/time-horizons A new Moore&amp;#039;s Law for AI agents]&lt;br /&gt;
[[Image:GmZHL8xWQAAtFlF.jpeg|450px]]&lt;br /&gt;
* 2025-04: [https://epoch.ai/blog/trends-in-ai-supercomputers Trends in AI Supercomputers] ([https://arxiv.org/abs/2504.16026 preprint])&lt;br /&gt;
* [https://ai-timeline.org/ The Road to AGI] (timeline visualization)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09677 The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs]&lt;br /&gt;
* 2025-09: [https://www.julian.ac/blog/2025/09/27/failing-to-understand-the-exponential-again/ Failing to Understand the Exponential, Again]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/rRbDNQLfihiHbXytf/distinguish-between-inference-scaling-and-larger-tasks-use Distinguish between inference scaling and &amp;quot;larger tasks use more compute&amp;quot;]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03992 Measuring AI R&amp;amp;D Automation] ([https://astrangeattractor.substack.com/p/measuring-ai-r-and-d-automation?triedRedirect=true blog])&lt;br /&gt;
&lt;br /&gt;
==Scaling Laws==&lt;br /&gt;
See: [[AI_understanding#Scaling_Laws|Scaling Laws]]&lt;br /&gt;
&lt;br /&gt;
==AGI Achievable==&lt;br /&gt;
* Yoshua Bengio: [https://arxiv.org/abs/2310.17688 Managing extreme AI risks amid rapid progress]&lt;br /&gt;
* Leopold Aschenbrenner: [https://situational-awareness.ai/from-gpt-4-to-agi/#Counting_the_OOMs Situational Awareness: Counting the OOMs]&lt;br /&gt;
* Richard Ngo: [https://medium.com/@richardcngo/visualizing-the-deep-learning-revolution-722098eb9c5 Visualizing the deep learning revolution]&lt;br /&gt;
* Katja Grace: [https://blog.aiimpacts.org/p/2023-ai-survey-of-2778-six-things Survey of 2,778 AI authors: six parts in pictures]&lt;br /&gt;
* Epoch AI: [https://epoch.ai/trends Machine Learning Trends]&lt;br /&gt;
* AI Digest: [https://theaidigest.org/progress-and-dangers How fast is AI improving?]&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/when-will-agi-arrive/ The case for AGI by 2030]&lt;br /&gt;
&lt;br /&gt;
==AGI Definition==&lt;br /&gt;
* 2023-11: Allan Dafoe, Shane Legg, et al.: [https://arxiv.org/abs/2311.02462 Levels of AGI for Operationalizing Progress on the Path to AGI]&lt;br /&gt;
* 2024-04: Bowen Xu: [https://arxiv.org/abs/2404.10731 What is Meant by AGI? On the Definition of Artificial General Intelligence]&lt;br /&gt;
* 2025-10: Dan Hendrycks et al.: [https://www.agidefinition.ai/paper.pdf A Definition of AGI]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07364 On the universal definition of intelligence]&lt;br /&gt;
&lt;br /&gt;
==Recursive Self Improvement (RSI)==&lt;br /&gt;
* 2026-02: [https://80000hours.org/articles/how-ai-driven-feedback-loops-could-make-things-very-crazy-very-fast/ How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
&lt;br /&gt;
==Progress Models==&lt;br /&gt;
From [http://yager-research.ca/2025/04/ai-impact-predictions/ AI Impact Predictions]:&lt;br /&gt;
&lt;br /&gt;
[[Image:AI impact models-2025 11 24.png|450px]]&lt;br /&gt;
&lt;br /&gt;
=Economic and Political=&lt;br /&gt;
* 2019-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3482150 The Impact of Artificial Intelligence on the Labor Market]&lt;br /&gt;
* 2020-06: [https://www.openphilanthropy.org/research/modeling-the-human-trajectory/ Modeling the Human Trajectory] (GDP)&lt;br /&gt;
* 2021-06: [https://www.openphilanthropy.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/ Report on Whether AI Could Drive Explosive Economic Growth]&lt;br /&gt;
* 2023-10: Marc Andreessen: [https://a16z.com/the-techno-optimist-manifesto/ The Techno-Optimist Manifesto]&lt;br /&gt;
* 2023-12: [https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html My techno-optimism]: &amp;quot;defensive acceleration&amp;quot; ([https://vitalik.eth.limo/index.html Vitalik Buterin])&lt;br /&gt;
* 2024-03: Noah Smith: [https://www.noahpinion.blog/p/plentiful-high-paying-jobs-in-the Plentiful, high-paying jobs in the age of AI: Comparative advantage is very subtle, but incredibly powerful.] ([https://x.com/liron/status/1768013030741475485 video])&lt;br /&gt;
* 2024-03: [https://doi.org/10.3386/w32255 Scenarios for the Transition to AGI] (AGI leads to wage collapse)&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-06: [https://www.frbsf.org/wp-content/uploads/AI-and-Growth-Aghion-Bunel.pdf AI and Growth: Where Do We Stand?]&lt;br /&gt;
* 2024-09: OpenAI [https://cdn.openai.com/global-affairs/openai-infra-economics-10.09.24.pdf Infrastructure is Destiny: Economic Returns on US Investment in Democratic AI]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/KFFaKu27FNugCHFmh/by-default-capital-will-matter-more-than-ever-after-agi By default, capital will matter more than ever after AGI] (L Rudolf L)&lt;br /&gt;
* 2025-01: [https://lukedrago.substack.com/p/the-intelligence-curse The Intelligence Curse]: With AGI, powerful actors will lose their incentives to invest in people&lt;br /&gt;
** Updated 2025-04: [https://intelligence-curse.ai/ The Intelligence Curse] (Luke Drago and Rudolf Laine)&lt;br /&gt;
*** [https://intelligence-curse.ai/pyramid/ Pyramid Replacement]&lt;br /&gt;
*** [https://intelligence-curse.ai/capital/ Capital, AGI, and Human Ambition]&lt;br /&gt;
*** [https://intelligence-curse.ai/defining/ Defining the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/shaping/ Shaping the Social Contract]&lt;br /&gt;
*** [https://intelligence-curse.ai/breaking/ Breaking the Intelligence Curse]&lt;br /&gt;
*** [https://intelligence-curse.ai/history/ History is Yours to Write]&lt;br /&gt;
* 2025-01: Microsoft: [https://blogs.microsoft.com/on-the-issues/2025/01/03/the-golden-opportunity-for-american-ai/ The Golden Opportunity for American AI]&lt;br /&gt;
* 2025-01: [https://www.maximum-progress.com/p/agi-will-not-make-labor-worthless AGI Will Not Make Labor Worthless]&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/global-affairs/ai-in-america-oai-economic-blueprint-20250113.pdf AI in America: OpenAI&amp;#039;s Economic Blueprint] ([https://openai.com/global-affairs/openais-economic-blueprint/ blog])&lt;br /&gt;
* 2025-01: [https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai How much economic growth from AI should we expect, how soon?]&lt;br /&gt;
* 2025-02: Morgan Stanley: [https://advisor.morganstanley.com/john.howard/documents/field/j/jo/john-howard/The_Humanoid_100_-_Mapping_the_Humanoid_Robot_Value_Chain.pdf The Humanoid 100: Mapping the Humanoid Robot Value Chain]&lt;br /&gt;
* 2025-02: [https://www.anthropic.com/news/the-anthropic-economic-index The Anthropic Economic Index]: [https://assets.anthropic.com/m/2e23255f1e84ca97/original/Economic_Tasks_AI_Paper.pdf Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11264 Strategic Wealth Accumulation Under Transformative AI Expectations]&lt;br /&gt;
* 2025-02: Tyler Cowen: [https://marginalrevolution.com/marginalrevolution/2025/02/why-i-think-ai-take-off-is-relatively-slow.html Why I think AI take-off is relatively slow]&lt;br /&gt;
* 2025-03: Epoch AI: [https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d Most AI value will come from broad automation, not from R&amp;amp;D]&lt;br /&gt;
** The primary economic impact of AI will be its ability to broadly automate labor&lt;br /&gt;
** Automating AI R&amp;amp;D alone likely won’t dramatically accelerate AI progress&lt;br /&gt;
** Fully automating R&amp;amp;D requires a very broad set of abilities&lt;br /&gt;
** AI takeoff will likely be diffuse and salient&lt;br /&gt;
* 2025-03: [https://www.anthropic.com/news/anthropic-economic-index-insights-from-claude-sonnet-3-7 Anthropic Economic Index: Insights from Claude 3.7 Sonnet]&lt;br /&gt;
* 2025-04: [https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality Will there be extreme inequality from AI?]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/research/impact-software-development Anthropic Economic Index: AI’s Impact on Software Development]&lt;br /&gt;
* 2025-05: [https://www.theguardian.com/books/2025/may/04/the-big-idea-can-we-stop-ai-making-humans-obsolete Better at everything: how AI could make human beings irrelevant]&lt;br /&gt;
* 2025-05: Forethought: [https://www.forethought.org/research/the-industrial-explosion The Industrial Explosion]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20273 Ten Principles of AI Agent Economics]&lt;br /&gt;
* 2025-07: [https://substack.com/home/post/p-167879696 What Economists Get Wrong about AI] They ignore innovation effects, use outdated capability assumptions, and miss the robotics revolution&lt;br /&gt;
* 2025-07: [https://www.nber.org/books-and-chapters/economics-transformative-ai/we-wont-be-missed-work-and-growth-era-agi We Won&amp;#039;t Be Missed: Work and Growth in the Era of AGI]&lt;br /&gt;
* 2025-07: [https://www.nber.org/papers/w34034 The Economics of Bicycles for the Mind]&lt;br /&gt;
* 2025-09: [https://conference.nber.org/conf_papers/f227491.pdf Genius on Demand: The Value of Transformative Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://peterwildeford.substack.com/p/ai-is-probably-not-a-bubble AI is probably not a bubble: AI companies have revenue, demand, and paths to immense value]&lt;br /&gt;
* 2025-11: [https://windowsontheory.org/2025/11/04/thoughts-by-a-non-economist-on-ai-and-economics/ Thoughts by a non-economist on AI and economics]&lt;br /&gt;
* 2025-11: [https://www.nber.org/papers/w34444 Artificial Intelligence, Competition, and Welfare]&lt;br /&gt;
* 2025-11: [https://www.anthropic.com/research/estimating-productivity-gains Estimating AI productivity gains from Claude conversations] (Anthropic)&lt;br /&gt;
* 2025-12: [https://benjamintodd.substack.com/p/how-ai-driven-feedback-loops-could How AI-driven feedback loops could make things very crazy, very fast]&lt;br /&gt;
* 2025-12: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth] (Philip Trammell and Leopold Aschenbrenner)&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/anthropic-economic-index-january-2026-report Anthropic Economic Index: new building blocks for understanding AI use]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/research/economic-index-primitives Anthropic Economic Index report: economic primitives]&lt;br /&gt;
* 2026-02: Nate Silver: [https://www.natesilver.net/p/the-singularity-wont-be-gentle The singularity won&amp;#039;t be gentle: If AI is even half as transformational as Silicon Valley assumes, politics will never be the same again]&lt;br /&gt;
&lt;br /&gt;
==Job Loss==&lt;br /&gt;
* 2023-03: [https://arxiv.org/pdf/2303.10130 GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models]&lt;br /&gt;
** 2023-03: [https://www.livemint.com/news/world/these-jobs-are-most-at-risk-due-to-chatgpt-as-per-openai-study-11679358453267.html These jobs are most at risk due to ChatGPT, as per OpenAI study]&lt;br /&gt;
* 2023-08: [https://dx.doi.org/10.2139/ssrn.4527336 The Short-Term Effects of Generative Artificial Intelligence on Employment: Evidence from an Online Labor Market]&lt;br /&gt;
** [https://x.com/jburnmurdoch/status/1722938749519077688 Freelancer sector shrinking]&lt;br /&gt;
[[Image:F-kVQuvWkAAemkr.png|400px]]&lt;br /&gt;
* 2023-09: [https://global-uploads.webflow.com/64d5f73a7fc5e8a240310c4d/650a128a34386a1206b6506c_FINAL%20Briefing%20-%20Adoption%20of%20Automation%20and%20AI%20in%20the%20UK.pdf What drives UK firms to adopt AI and robotics, and what are the consequences for jobs?]&lt;br /&gt;
** [https://www.digitalinformationworld.com/2023/09/78-of-companies-say-ai-created-more-jobs.html 78% of Companies Say AI Created More Jobs]&lt;br /&gt;
* 2023-11: [https://theaipi.org/ai-interactive-map/ New Analysis Shows Over 20% of US Jobs Significantly Exposed to AI Automation In the Near Future]&lt;br /&gt;
* 2024-01: [https://techcrunch.com/2024/01/09/duolingo-cut-10-of-its-contractor-workforce-as-the-company-embraces-ai/ Duolingo cuts 10% of its contractor workforce as the company embraces AI]&lt;br /&gt;
* 2024-02: [https://www.pwc.com/gx/en/issues/c-suite-insights/the-leadership-agenda/gen-ai-is-a-tool-for-growth-not-just-efficiency.html#:~:text=One%20out%20of%20every%20four%20of%20the%204%2C702,to%20accomplish%20the%20same%20tasks%20with%20fewer%20workers Gen AI is a tool for growth, not just efficiency: Tech CEOs are investing to build their workforce and capitalise on new opportunities from generative AI. That’s a sharp contrast to how their peers view it.]&lt;br /&gt;
* 2024-04: [https://www.nytimes.com/2024/04/10/business/investment-banking-jobs-artificial-intelligence.html AI is Poised to Replace the Entry-Level Grunt Work of a Wall Street Career]&lt;br /&gt;
* 2024-07: [https://www.wired.com/story/ai-is-already-taking-jobs-in-the-video-game-industry/ AI Is Already Taking Jobs in the Video Game Industry]: A WIRED investigation finds that major players like Activision Blizzard, which recently laid off scores of workers, are using generative AI for game development&lt;br /&gt;
* 2024-08: [https://www.bbc.com/news/articles/c80e1gp9m9zo Klarna: AI lets us cut thousands of jobs - but pay more]&lt;br /&gt;
* 2025-01: [https://scholarspace.manoa.hawaii.edu/server/api/core/bitstreams/4f39375d-59c2-4c4a-b394-f3eed7858c80/content AI and Freelancers: Has the Inflection Point Arrived?]&lt;br /&gt;
* 2025-01: [https://www.aporiamagazine.com/p/yes-youre-going-to-be-replaced Yes, you&amp;#039;re going to be replaced: So much cope about AI]&lt;br /&gt;
* 2025-03: [https://commonplace.org/2025/03/20/will-ai-automate-away-your-job/ Will AI Automate Away Your Job? The time-horizon model explains the future of the technology]&lt;br /&gt;
* 2025-05: [https://www.forbes.com/sites/jackkelly/2025/05/04/its-time-to-get-concerned-klarna-ups-duolingo-cisco-and-many-other-companies-are-replacing-workers-with-ai/ It’s Time To Get Concerned, Klarna, UPS, Duolingo, Cisco, And Many Other Companies Are Replacing Workers With AI]&lt;br /&gt;
* 2025-05: [https://time.com/7289692/when-ai-replaces-workers/ What Happens When AI Replaces Workers?]&lt;br /&gt;
* 2025-05: [https://www.oxfordeconomics.com/resource/educated-but-unemployed-a-rising-reality-for-us-college-grads/ Educated but unemployed, a rising reality for US college grads] Structural shifts in tech hiring and the growing impact of AI are driving higher unemployment among recent college graduates&lt;br /&gt;
* 2025-05: NY Times: [https://www.nytimes.com/2025/05/30/technology/ai-jobs-college-graduates.html?unlocked_article_code=1.LE8.LlC6.eT5XcpA9hxC2&amp;amp;smid=url-share For Some Recent Graduates, the A.I. Job Apocalypse May Already Be Here] The unemployment rate for recent college graduates has jumped as companies try to replace entry-level workers with artificial intelligence&lt;br /&gt;
* 2025-06: [https://80000hours.org/agi/guide/skills-ai-makes-valuable/ How not to lose your job to AI] The skills AI will make more valuable (and how to learn them)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06576 Future of Work with AI Agents: Auditing Automation and Augmentation Potential across the U.S. Workforce]&lt;br /&gt;
[[Image:0dab4c86-882d-4095-9d12-d19684ed5184 675x680.png|300px]]&lt;br /&gt;
* 2025-07: Harvard Business Review: [https://hbr.org/2025/06/what-gets-measured-ai-will-automate What Gets Measured, AI Will Automate]&lt;br /&gt;
* 2025-08: [https://digitaleconomy.stanford.edu/publications/canaries-in-the-coal-mine/ Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence]&lt;br /&gt;
* 2025-10: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5560401 Performance or Principle: Resistance to Artificial Intelligence in the U.S. Labor Market]&lt;br /&gt;
* 2025-10: [https://www.siliconcontinent.com/p/the-ai-becker-problem The AI Becker problem: Who will train the next generation?]&lt;br /&gt;
* 2026-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6134506 AI, Automation, and Expertise]&lt;br /&gt;
* 2026-02: [https://arachnemag.substack.com/p/the-jevons-paradox-for-intelligence The Jevons Paradox for Intelligence: Fears of AI-induced job loss could not be more wrong]&lt;br /&gt;
&lt;br /&gt;
==Productivity Impact==&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2026-02: [https://www.ft.com/content/4b51d0b4-bbfe-4f05-b50a-1d485d419dc5 The AI productivity take-off is finally visible] ([https://x.com/erikbryn/status/2023075588974735869?s=20 Erik Brynjolfsson])&lt;br /&gt;
** Businesses are finally beginning to reap some of AI&amp;#039;s benefits.&lt;br /&gt;
* 2026-02: New York Times: [https://www.nytimes.com/2026/02/18/opinion/ai-software.html The A.I. Disruption We’ve Been Waiting for Has Arrived]&lt;br /&gt;
&lt;br /&gt;
==National Security==&lt;br /&gt;
* 2025-04: Jeremie Harris and Edouard Harris: [https://superintelligence.gladstone.ai/ America’s Superintelligence Project]&lt;br /&gt;
&lt;br /&gt;
==AI Manhattan Project==&lt;br /&gt;
* 2024-06: [https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf Situational Awareness] ([https://www.forourposterity.com/ Leopold Aschenbrenner]) - [https://www.lesswrong.com/posts/nP5FFYFjtY8LgWymt/quotes-from-leopold-aschenbrenner-s-situational-awareness select quotes], [https://www.youtube.com/watch?v=zdbVtZIn9IM podcast], [https://danielmiessler.com/p/podcast-summary-dwarkesh-vs-leopold-aschenbrenner text summary of podcast]&lt;br /&gt;
* 2024-10: [https://thezvi.substack.com/p/ai-88-thanks-for-the-memos?open=false#%C2%A7thanks-for-the-memos-introduction-and-competitiveness White House Memo calls for action on AI]&lt;br /&gt;
* 2024-11: [https://www.uscc.gov/annual-report/2024-annual-report-congress 2024 Annual Report to Congress]: [https://www.reuters.com/technology/artificial-intelligence/us-government-commission-pushes-manhattan-project-style-ai-initiative-2024-11-19/ calls] for &amp;quot;Manhattan Project-style&amp;quot; effort&lt;br /&gt;
* 2025-05-29: [https://x.com/ENERGY/status/1928085878561272223 DoE Tweet]: &amp;quot;AI is the next Manhattan Project, and THE UNITED STATES WILL WIN. 🇺🇸&amp;quot;&lt;br /&gt;
* 2025-07: [https://epoch.ai/gradient-updates/how-big-could-an-ai-manhattan-project-get How big could an “AI Manhattan Project” get?]&lt;br /&gt;
&lt;br /&gt;
=Near-term=&lt;br /&gt;
* 2021-08: Daniel Kokotajlo: [https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like What 2026 looks like]&lt;br /&gt;
* 2025-02: Sam Altman: [https://blog.samaltman.com/three-observations Three Observations]&lt;br /&gt;
*# The intelligence of an AI model roughly equals the log of the resources used to train and run it.&lt;br /&gt;
*# The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use&lt;br /&gt;
*# The socioeconomic value of linearly increasing intelligence is super-exponential in nature&lt;br /&gt;
* 2025-03: [https://www.pathwaysai.org/p/glimpses-of-ai-progess Glimpses of AI Progress: Mental models for fast times]&lt;br /&gt;
* 2025-03: [https://www.nature.com/articles/s41598-025-92190-7 Navigating artificial general intelligence development: societal, technological, ethical, and brain-inspired pathways]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
** 2025-07: Video: [https://www.youtube.com/watch?v=5KVDDfAkRgc Are We 3 Years From AI Disaster? A Rigorous Forecast]&lt;br /&gt;
* 2025-04: Stanford HAI: [https://hai-production.s3.amazonaws.com/files/hai_ai_index_report_2025.pdf Artificial Intelligence Index Report 2025]&lt;br /&gt;
* 2025-04: Arvind Narayananand Sayash Kapoor: [https://kfai-documents.s3.amazonaws.com/documents/c3cac5a2a7/AI-as-Normal-Technology---Narayanan---Kapoor.pdf AI as Normal Technology]&lt;br /&gt;
* 2025-04: Dwarkesh Patel: [https://www.dwarkesh.com/p/questions-about-ai Questions about the Future of AI]&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: IdeaFoundry: [https://ideafoundry.substack.com/p/evolution-vs-extinction-the-choice Evolution vs. Extinction: The Choice is Ours] The next 18 months will decide whether AI ends us or evolves us&lt;br /&gt;
* 2025-07: [https://cfg.eu/advanced-ai-possible-futures/ Advanced AI: Possible futures] Five scenarios for how the AI-transition could unfold&lt;br /&gt;
* 2025-11: [https://android-dreams.ai/ Android Dreams]&lt;br /&gt;
* 2026-02: [https://www.citriniresearch.com/ Citrini]: [https://www.citriniresearch.com/p/2028gic The 2028 Global Intelligence Crisis: A Thought Exercise in Financial History, from the Future]&lt;br /&gt;
&lt;br /&gt;
==Insightful Analysis of Current State==&lt;br /&gt;
* 2025-11: Andy Masley: [https://andymasley.substack.com/p/the-lump-of-cognition-fallacy The lump of cognition fallacy: The extended mind as the advance of civilization]&lt;br /&gt;
* 2026-02: Eric Jang: [https://evjang.com/2026/02/04/rocks.html As Rocks May Think]&lt;br /&gt;
* 2026-02: Matt Shumer: [https://x.com/mattshumer_/status/2021256989876109403 Something Big Is Happening]&lt;br /&gt;
* 2026-02: Minh Pham: [https://x.com/buckeyevn/status/2014171253045960803?s=20 Why Most Agent Harnesses Are Not Bitter Lesson Pilled]&lt;br /&gt;
&lt;br /&gt;
=Overall=&lt;br /&gt;
* 1993: [https://en.wikipedia.org/wiki/Vernor_Vinge Vernor Vinge]: [https://edoras.sdsu.edu/~vinge/misc/singularity.html The Coming Technological Singularity: How to Survive in the Post-Human Era]&lt;br /&gt;
* 2025-03: Kevin Roose (New York Times): [https://www.nytimes.com/2025/03/14/technology/why-im-feeling-the-agi.html?unlocked_article_code=1.304.TIEy.SmNhKYO4e9c7&amp;amp;smid=url-share Powerful A.I. Is Coming. We’re Not Ready.] Three arguments for taking progress toward artificial general intelligence, or A.G.I., more seriously — whether you’re an optimist or a pessimist.&lt;br /&gt;
* 2025-03: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/thoughts-on-future-ai.html My Thoughts on the Future of &amp;quot;AI&amp;quot;]: &amp;quot;I have very wide error bars on the potential future of large language models, and I think you should too.&amp;quot;&lt;br /&gt;
* 2025-06: Sam Altman: [https://blog.samaltman.com/the-gentle-singularity The Gentle Singularity]&lt;br /&gt;
&lt;br /&gt;
==Surveys of Opinions/Predictions==&lt;br /&gt;
* 2016-06: [https://aiimpacts.org/2016-expert-survey-on-progress-in-ai/ 2016 Expert Survey on Progress in AI]&lt;br /&gt;
** 2023-03: [https://aiimpacts.org/scoring-forecasts-from-the-2016-expert-survey-on-progress-in-ai/ Scoring forecasts from the 2016 “Expert Survey on Progress in AI”]&lt;br /&gt;
* 2022-10: Forecasting Research Institute: [https://forecastingresearch.org/near-term-xpt-accuracy Assessing Near-Term Accuracy in the Existential Risk Persuasion Tournament]&lt;br /&gt;
** 2025-09: Ethan Mollick: [https://x.com/emollick/status/1962859757674344823 Progress is ahead of expectations]&lt;br /&gt;
* 2023-08: [https://wiki.aiimpacts.org/ai_timelines/predictions_of_human-level_ai_timelines/ai_timeline_surveys/2023_expert_survey_on_progress_in_ai 2023 Expert Survey on Progress in AI]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.02843 Thousands of AI Authors on the Future of AI]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14870 Why do Experts Disagree on Existential Risk and P(doom)? A Survey of AI Experts]&lt;br /&gt;
* 2025-02: Nicholas Carlini: [https://nicholas.carlini.com/writing/2025/forecasting-ai-2025-update.html AI forecasting retrospective: you&amp;#039;re (probably) over-confident]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/long-timelines-to-advanced-ai-have &amp;quot;Long&amp;quot; timelines to advanced AI have gotten crazy short]&lt;br /&gt;
* 2025-05: [https://theaidigest.org/ai2025-analysis-may AI 2025 Forecasts - May Update]&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41598-026-39070-w Lay beliefs about the badness, likelihood, and importance of human extinction]&lt;br /&gt;
&lt;br /&gt;
==Bad Outcomes==&lt;br /&gt;
* [https://pauseai.info/pdoom List of p(doom) values]&lt;br /&gt;
* 2019-03: [https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like What failure looks like]&lt;br /&gt;
* 2023-03: gwern: [https://gwern.net/fiction/clippy It Looks Like You’re Trying To Take Over The World]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16946 Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development] ([https://gradual-disempowerment.ai/ web version])&lt;br /&gt;
** 2025-02: [https://thezvi.substack.com/p/the-risk-of-gradual-disempowerment The Risk of Gradual Disempowerment from AI]&lt;br /&gt;
** 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
* 2025-04: Daniel Kokotajlo, Scott Alexander, Thomas Larsen, Eli Lifland, Romeo Dean: [https://ai-2027.com/ AI 2027] ([https://ai-2027.com/scenario.pdf pdf])&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-09: [https://doctrines.ai/ The three main doctrines on the future of AI]&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Dominance doctrine:&amp;#039;&amp;#039;&amp;#039; First actor to create advanced AI will attain overwhelming strategic superiority&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Extinction doctrine:&amp;#039;&amp;#039;&amp;#039; Humanity will lose control of ASI, leading to extinction or permanent disempowerment&lt;br /&gt;
** &amp;#039;&amp;#039;&amp;#039;Replacement doctrine:&amp;#039;&amp;#039;&amp;#039; AI will automate human tasks, but without fundamentally reshaping or ending civilization&lt;br /&gt;
* 2025-09: Sean ÓhÉigeartaigh: [https://www.cambridge.org/core/journals/cambridge-prisms-extinction/article/extinction-of-the-human-species-what-could-cause-it-and-how-likely-is-it-to-occur/D8816A79BEF5A4C30A3E44FD8D768622 Extinction of the human species: What could cause it and how likely is it to occur?]&lt;br /&gt;
&lt;br /&gt;
==Intelligence Explosion==&lt;br /&gt;
* 2023-06: [https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/ What a Compute-Centric Framework Says About Takeoff Speeds]&lt;br /&gt;
** [https://takeoffspeeds.com/ takeoffspeeds.com simulator]&lt;br /&gt;
* 2025-02: [https://www.forethought.org/research/three-types-of-intelligence-explosion Three Types of Intelligence Explosion]&lt;br /&gt;
* 2025-03: Future of Life Institute: [https://futureoflife.org/ai/are-we-close-to-an-intelligence-explosion/ Are we close to an intelligence explosion?] AIs are inching ever-closer to a critical threshold. Beyond this threshold lie great risks—but crossing it is not inevitable.&lt;br /&gt;
* 2025-03: Forethought: [https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion Will AI R&amp;amp;D Automation Cause a Software Intelligence Explosion?]&lt;br /&gt;
[[Image:Gm-1jugbYAAtq Y.jpeg|450px]]&lt;br /&gt;
* 2025-05: [https://www.thelastinvention.ai/ The Last Invention] Why Humanity’s Final Creation Changes Everything&lt;br /&gt;
* 2025-08: [https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be How quick and big would a software intelligence explosion be?]&lt;br /&gt;
&lt;br /&gt;
==Superintelligence==&lt;br /&gt;
* 2024-10: [http://yager-research.ca/2024/10/how-smart-will-asi-be/ How Smart will ASI be?]&lt;br /&gt;
* 2024-11: [http://yager-research.ca/2024/11/concise-argument-for-asi-risk/ Concise Argument for ASI Risk]&lt;br /&gt;
* 2025-03: [https://dynomight.net/smart/ Limits of smart]&lt;br /&gt;
* 2025-05: [https://timfduffy.substack.com/p/the-limits-of-superintelligence?manualredirect= The Limits of Superintelligence]&lt;br /&gt;
&lt;br /&gt;
==Long-range/Philosophy==&lt;br /&gt;
* 2023-03: Dan Hendrycks: [https://arxiv.org/abs/2303.16200 Natural Selection Favors AIs over Humans]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2025-01: [https://longerramblings.substack.com/p/a-defence-of-slowness-at-the-end A defence of slowness at the end of the world]&lt;br /&gt;
&lt;br /&gt;
=Positives &amp;amp; Optimism=&lt;br /&gt;
==Science &amp;amp; Technology Improvements==&lt;br /&gt;
* 2023-05: [https://www.planned-obsolescence.org/author/kelsey/ Kelsey Piper]: [https://www.planned-obsolescence.org/the-costs-of-caution/ The costs of caution]&lt;br /&gt;
* 2024-09: Sam Altman: [https://ia.samaltman.com/ The Intelligence Age]&lt;br /&gt;
* 2024-10: Dario Amodei: [https://darioamodei.com/machines-of-loving-grace Machines of Loving Grace]&lt;br /&gt;
* 2024-11: Google DeepMind: [https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery A new golden age of discovery]&lt;br /&gt;
* 2025-03: [https://finmoorhouse.com/ Fin Moorhouse], [https://www.williammacaskill.com/ Will MacAskill]: [https://www.forethought.org/research/preparing-for-the-intelligence-explosion Preparing for the Intelligence Explosion]&lt;br /&gt;
&lt;br /&gt;
==Social==&lt;br /&gt;
* 2025-09: [https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale Coasean Bargaining at Scale]: Decentralization, coordination, and co-existence with AGI&lt;br /&gt;
* 2025-10: [https://www.nber.org/system/files/chapters/c15309/c15309.pdf#page=15.23 The Coasean Singularity? Demand, Supply, and Market Design with AI Agents]&lt;br /&gt;
&lt;br /&gt;
==Post-scarcity Society==&lt;br /&gt;
* 2004: Eliezer Yudkowsky (MIRI): [https://intelligence.org/files/CEV.pdf Coherent Extrapolated Volition] and [https://www.lesswrong.com/s/d3WgHDBAPYYScp5Em/p/K4aGvLnHvYgX9pZHS Fun Theory]&lt;br /&gt;
* 2019: John Danaher: [https://www.jstor.org/stable/j.ctvn5txpc Automation and Utopia: Human Flourishing in a World Without Work]&lt;br /&gt;
&lt;br /&gt;
==The Grand Tradeoff==&lt;br /&gt;
* 2026-02: Nick Bostrom: [https://nickbostrom.com/optimal.pdf Optimal Timing for Superintelligence: Mundane Considerations for Existing People]&lt;br /&gt;
&lt;br /&gt;
=Plans=&lt;br /&gt;
* [https://www.narrowpath.co/ A Narrow Path: How to Secure our Future]&lt;br /&gt;
* Marius Hobbhahn: [https://www.lesswrong.com/posts/bb5Tnjdrptu89rcyY/what-s-the-short-timeline-plan What’s the short timeline plan?]&lt;br /&gt;
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]&lt;br /&gt;
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]&lt;br /&gt;
* [https://controlai.com/ Control AI]: [https://controlai.com/dip The Direct Institutional Plan] &lt;br /&gt;
* Luke Drago and L Rudolf L: [https://lukedrago.substack.com/p/the-use-of-knowledge-in-agi-society?triedRedirect=true The use of knowledge in (AGI) society]: How to build to break the [https://lukedrago.substack.com/p/the-intelligence-curse intelligence curse]&lt;br /&gt;
* [https://www.agisocialcontract.org/ AGI Social Contract]&lt;br /&gt;
** [https://www.agisocialcontract.org/forging-a-new-agi-social-contract Forging A New AGI Social Contract]&lt;br /&gt;
* Yoshua Bengio: [https://time.com/7283507/safer-ai-development/ A Potential Path to Safer AI Development]&lt;br /&gt;
** 2025-02: [https://arxiv.org/abs/2502.15657 Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?]&lt;br /&gt;
* 2026-01: Dario Amodei: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI]&lt;br /&gt;
* 2026-02: Ryan Greenblatt: [https://www.lesswrong.com/posts/vjAM7F8vMZS7oRrrh/how-do-we-more-safely-defer-to-ais How do we (more) safely defer to AIs?]&lt;br /&gt;
&lt;br /&gt;
==Philosophy==&lt;br /&gt;
* [https://danfaggella.com/ Dan Faggella]:&lt;br /&gt;
** 2018-07: [https://danfaggella.com/moral-singularity/ Moral Singularity – Unpredictable Values Bodes Poorly for Humanity]&lt;br /&gt;
** 2025-02: [https://danfaggella.com/bend/ There is No Pause – We Must Bend the Posthuman Trajectory]&lt;br /&gt;
* Joe Carlsmith: 2024: [https://joecarlsmith.com/2024/01/02/otherness-and-control-in-the-age-of-agi Otherness and control in the age of AGI]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/02/gentleness-and-the-artificial-other Gentleness and the artificial Other]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/04/deep-atheism-and-ai-risk Deep atheism and AI risk]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/08/when-yang-goes-wrong When “yang” goes wrong]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/09/does-ai-risk-other-the-ais Does AI risk “other” the AIs?]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/11/an-even-deeper-atheism An even deeper atheism]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy Being nicer than Clippy]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/01/18/on-the-abolition-of-man On the abolition of man]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/21/on-green On green]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/03/25/on-attunement On attunement]&lt;br /&gt;
*# [https://joecarlsmith.com/2024/06/18/loving-a-world-you-dont-trust Loving a world you don’t trust]&lt;br /&gt;
* Anthony Aguirre:&lt;br /&gt;
** [https://x.com/AnthonyNAguirre/status/1898023049930457468 2025-03]: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
[[Image:GlchEeObwAQ88NK.jpeg|300px]]&lt;br /&gt;
* 2025-04: Scott Alexander (Astral Codex Ten): [https://www.astralcodexten.com/p/the-colors-of-her-coat The Colors Of Her Coat] (response to [https://www.theintrinsicperspective.com/p/welcome-to-the-semantic-apocalypse semantic apocalypse] and semantic satiation)&lt;br /&gt;
* 2025-05: Helen Toner: [https://www.ai-frontiers.org/articles/were-arguing-about-ai-safety-wrong We’re Arguing About AI Safety Wrong]: Dynamism vs. stasis is a clearer lens for criticizing controversial AI safety prescriptions&lt;br /&gt;
* 2025-05: Joe Carlsmith: [https://joecarlsmith.substack.com/p/the-stakes-of-ai-moral-status The stakes of AI moral status]&lt;br /&gt;
&lt;br /&gt;
==Research==&lt;br /&gt;
* 2025-05: [https://www.lesswrong.com/posts/GAv4DRGyDHe2orvwB/gradual-disempowerment-concrete-research-projects Gradual Disempowerment: Concrete Research Projects]&lt;br /&gt;
&lt;br /&gt;
==Alignment==&lt;br /&gt;
* 2023-03: Leopold Aschenbrenner: [https://www.forourposterity.com/nobodys-on-the-ball-on-agi-alignment/ Nobody’s on the ball on AGI alignment]&lt;br /&gt;
* 2024-03: [https://static1.squarespace.com/static/65392ca578eee444c445c9de/t/6606f95edb20e8118074a344/1711733370985/human-values-and-alignment-29MAR2024.pdf What are human values, and how do we align AI to them?] ([https://meaningalignment.substack.com/p/0480e023-98c0-4633-a604-990d3ac880ac blog])&lt;br /&gt;
* 2025: Joe Carlsmith: [https://joecarlsmith.substack.com/p/how-do-we-solve-the-alignment-problem How do we solve the alignment problem?] Introduction to an essay series on paths to safe, useful superintelligence&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/what-is-it-to-solve-the-alignment What is it to solve the alignment problem?] Also: to avoid it? Handle it? Solve it forever? Solve it completely? ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16617671-what-is-it-to-solve-the-alignment-problem audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/when-should-we-worry-about-ai-power When should we worry about AI power-seeking?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16651469-when-should-we-worry-about-ai-power-seeking audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/paths-and-waystations-in-ai-safety Paths and waystations in AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16768804-paths-and-waystations-in-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/ai-for-ai-safety AI for AI safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/16790183-ai-for-ai-safety audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/can-we-safely-automate-alignment Can we safely automate alignment research?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17069901-can-we-safely-automate-alignment-research audio version], [https://joecarlsmith.substack.com/p/video-and-transcript-of-talk-on-automating?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=162375391&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email video version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/giving-ais-safe-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=171250683&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email Giving AIs safe motivations] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17686921-giving-ais-safe-motivations audio version])&lt;br /&gt;
*# [https://joecarlsmith.com/2025/09/29/controlling-the-options-ais-can-pursue Controlling the options AIs can pursue] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/17909401-controlling-the-options-ais-can-pursue audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/how-human-like-do-safe-ai-motivations?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=178666988&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email How human-like do safe AI motivations need to be?] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18175429-how-human-like-do-safe-ai-motivations-need-to-be audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/building-ais-that-do-human-like-philosophy Building AIs that do human-like philosophy: AIs will face philosophical questions humans can&amp;#039;t answer for them] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18591342-building-ais-that-do-human-like-philosophy audio version])&lt;br /&gt;
*# [https://joecarlsmith.substack.com/p/on-restraining-ai-development-for?utm_source=post-email-title&amp;amp;publication_id=1022275&amp;amp;post_id=191385185&amp;amp;utm_campaign=email-post-title&amp;amp;isFreemail=true&amp;amp;r=5av1bk&amp;amp;triedRedirect=true&amp;amp;utm_medium=email On restraining AI development for the sake of safety] ([https://joecarlsmithaudio.buzzsprout.com/2034731/episodes/18869440-on-restraining-ai-development-for-the-sake-of-safety audio version])&lt;br /&gt;
* 2025-04: Dario Amodei: [https://www.darioamodei.com/post/the-urgency-of-interpretability The Urgency of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Technical==&lt;br /&gt;
* 2025-03: [https://resilience.baulab.info/docs/AI_Action_Plan_RFI.pdf AI Dominance Requires Interpretability and Standards for Transparency and Security]&lt;br /&gt;
* 2026-02: [https://www.gap-map.org/capabilities/?sort=bottlenecks Fundamental Development Gap Map v1.0]&lt;br /&gt;
&lt;br /&gt;
==Strategic/Policy==&lt;br /&gt;
* 2015-03: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-2 Machine intelligence, part 2]&lt;br /&gt;
* 2019-07: Amanda Askell, Miles Brundage, Gillian Hadfield: [https://arxiv.org/abs/1907.04534 The Role of Cooperation in Responsible AI Development]&lt;br /&gt;
* 2025-03: Dan Hendrycks, Eric Schmidt, Alexandr Wang: [https://www.nationalsecurity.ai/ Superintelligence Strategy]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/executive-summary Executive Summary]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/introduction Introduction]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/ai-is-pivotal-for-national-security AI Is Pivotal for National Security]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/deterrence-with-mutual-assured-ai-malfunction-maim Deterrence with Mutual Assured AI Malfunction (MAIM)]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/nonproliferation Nonproliferation]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/competitiveness Competitiveness]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/conclusion Conclusion]&lt;br /&gt;
** [https://www.nationalsecurity.ai/chapter/appendix Appendix FAQs]&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human] ([https://keepthefuturehuman.ai/essay/ essay])&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late (video)]  (2025)&lt;br /&gt;
**# Oversight: Registration required for training &amp;gt;10&amp;lt;sup&amp;gt;25&amp;lt;/sup&amp;gt; FLOP and inference &amp;gt;10&amp;lt;sup&amp;gt;19&amp;lt;/sup&amp;gt; FLOP/s (~1,000 B200 GPUs @ $25M). Build cryptographic licensing into hardware.&lt;br /&gt;
**# Computation Limits: Ban on training models &amp;gt;10&amp;lt;sup&amp;gt;27&amp;lt;/sup&amp;gt; FLOP or inference &amp;gt;10&amp;lt;sup&amp;gt;20&amp;lt;/sup&amp;gt; FLOP/s.&lt;br /&gt;
**# Strict Liability: Hold AI companies responsible for outcomes.&lt;br /&gt;
**# Tiered Regulation: Low regulation on tool-AI, strictest regulation on AGI (general, capable, autonomous systems).&lt;br /&gt;
* 2025-04: [https://x.com/deanwball Dean W. Ball]: [https://arxiv.org/abs/2504.11501 A Framework for the Private Governance of Frontier Artificial Intelligence]&lt;br /&gt;
* 2025-04: Helen Toner: [https://helentoner.substack.com/p/nonproliferation-is-the-wrong-approach?source=queue Nonproliferation is the wrong approach to AI misuse]&lt;br /&gt;
* 2025-04: MIRI: [https://techgov.intelligence.org/research/ai-governance-to-avoid-extinction AI Governance to Avoid Extinction: The Strategic Landscape and Actionable Research Questions]&lt;br /&gt;
* 2025-05: [https://writing.antonleicht.me/p/the-new-ai-policy-frontier The New AI Policy Frontier]: Beyond the shortcomings of centralised control and alignment, a new school of thought on AI governance emerges. It still faces tricky politics.&lt;br /&gt;
* 2025-05: [https://uncpga.world/agi-uncpga-report/ AGI UNCPGA Report]: Governance of the Transition to Artificial General Intelligence (AGI) Urgent Considerations for the UN General Assembly: Report for the Council of Presidents of the United Nations General Assembly (UNCPGA)&lt;br /&gt;
* 2025-06: [https://writing.antonleicht.me/p/ai-and-jobs-politics-without-policy AI &amp;amp; Jobs: Politics without Policy] Political support mounts - for a policy platform that does not yet exist&lt;br /&gt;
* 2025-06: [https://x.com/littIeramblings Sarah Hastings-Woodhouse]: [https://drive.google.com/file/d/1mmdHBE6M2yiyL21-ctTuRLNH5xOFjqWm/view Safety Features for a Centralized AGI Project]&lt;br /&gt;
* 2025-07: [https://writing.antonleicht.me/p/a-moving-target A Moving Target] Why we might not be quite ready to comprehensively regulate AI, and why it matters&lt;br /&gt;
* 2025-07: [https://www-cdn.anthropic.com/0dc382a2086f6a054eeb17e8a531bd9625b8e6e5.pdf Anthropic: Build AI in America] ([https://www.anthropic.com/news/build-ai-in-america blog])&lt;br /&gt;
* 2025-12: [https://asi-prevention.com/ How middle powers may prevent the development of artificial superintelligence]&lt;br /&gt;
* 2026-03: [https://humanstatement.org/ The Pro-Human AI Declaration]&lt;br /&gt;
&lt;br /&gt;
==Restriction==&lt;br /&gt;
* 2024-05: OpenAI: [https://openai.com/index/reimagining-secure-infrastructure-for-advanced-ai/ Reimagining secure infrastructure for advanced AI] OpenAI calls for an evolution in infrastructure security to protect advanced AI &lt;br /&gt;
* 2025-07: MIRI: [https://arxiv.org/abs/2507.09801 Technical Requirements for Halting Dangerous AI Activities]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI safety]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8750</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8750"/>
		<updated>2026-03-23T17:23:46Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Math */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
* 2026-03: [https://epoch.ai/frontiermath/open-problems FrontierMath] problem: [https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs &amp;quot;A Ramsey-style Problem on Hypergraphs&amp;quot;] solved by Kevin Barreto and Liam Price using GPT-5.4 Pro&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_creativity&amp;diff=8749</id>
		<title>AI creativity</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_creativity&amp;diff=8749"/>
		<updated>2026-03-23T17:19:47Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Research */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Research=&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4686415 Creativity and AI]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.02980 Surveying the Effects of Quality, Diversity, and Complexity in Synthetic Data From Large Language Models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
* 2025-03: Midjourney: [https://www.arxiv.org/abs/2503.17126 Modifying Large Language Model Post-Training for Diverse Creative Writing]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.12320 Has the Creativity of Large-Language Models peaked? An analysis of inter- and intra-LLM variability]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.14442 Creative Preference Optimization]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.01171 Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.20635 Why Did Apple Fall To The Ground: Evaluating Curiosity In Large Language Model]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.22954 Artificial Hivemind: The Open-Ended Homogeneity of Language Models (and Beyond)]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41562-025-02331-1 A large-scale comparison of divergent creativity in humans and large language models]&lt;br /&gt;
* 2026-01: [https://www.arxiv.org/abs/2601.01576 OpenNovelty: An LLM-powered Agentic System for Verifiable Scholarly Novelty Assessment]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41598-025-25157-3 Divergent creativity in humans and large language models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
* 2026-03: [https://gking.harvard.edu/quest Inducing Sustained Creativity and Diversity in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
=Benchmarks=&lt;br /&gt;
See: [[AI_benchmarks#Creativity| AI benchmarks &amp;gt; Creativity]]&lt;br /&gt;
&lt;br /&gt;
=Collapse=&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02209 Generative Monoculture in Large Language Models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
==Analysis==&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
==LLM==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13928 LLMs Can Get &amp;quot;Brain Rot&amp;quot;!]&lt;br /&gt;
==Image Models==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
==Solutions==&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI_benchmarks|AI benchmarks]] &amp;gt; [[AI_benchmarks#Assess_Specific_Attributes|Assess Specific Attributes]] &amp;gt; [[AI_benchmarks#Creativity|Creativity]]&lt;br /&gt;
* [[AI_and_Humans|AI and Humans]] &amp;gt; [[AI_and_Humans#AI_out-performs_humans|AI out-performs humans]] &amp;gt; [[AI_and_Humans#Creativity|Creativity]]&lt;br /&gt;
* [[AI_and_Humans|AI and Humans]] &amp;gt; [[AI_and_Humans#AI_improves_human_work|AI improves human work]] &amp;gt; Creativity&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8748</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8748"/>
		<updated>2026-03-23T17:16:21Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Autonomous Ideation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.14473 AI Can Learn Scientific Taste]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8747</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8747"/>
		<updated>2026-03-23T17:14:13Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Human well-being */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
* 2026-03: Anthropic: [https://www.anthropic.com/features/81k-interviews What 81,000 people want from AI]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.15245 Practicing with Language Models Cultivates Human Empathic Communication]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_research_trends&amp;diff=8746</id>
		<title>AI research trends</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_research_trends&amp;diff=8746"/>
		<updated>2026-03-23T17:11:22Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Context Length */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=System 2 Reasoning=&lt;br /&gt;
See: [[Increasing AI Intelligence]]&lt;br /&gt;
&lt;br /&gt;
=Memory=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13501 A Survey on the Memory Mechanism of Large Language Model based Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.09113 The AI Hippocampus: How Far are We From Human Memory?]&lt;br /&gt;
&lt;br /&gt;
==Big Ideas==&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.07755 Learning to Continually Learn via Meta-learning Agentic Memory Designs]&lt;br /&gt;
&lt;br /&gt;
==LLM Weights Memory==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09764 Memory Layers at Scale]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15103 Continual Learning via Sparse Memory Finetuning]&lt;br /&gt;
* 2026-01: [https://developer.nvidia.com/blog/reimagining-llm-memory-using-context-as-training-data-unlocks-models-that-learn-at-test-time/ Reimagining LLM Memory: Using Context as Training Data Unlocks Models That Learn at Test-Time] (Nvidia)&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.02151 Entropy-Adaptive Fine-Tuning: Resolving Confident Conflicts to Mitigate Forgetting]&lt;br /&gt;
* 2026-02: Sakana AI: [https://pub.sakana.ai/doc-to-lora/ Instant LLM Updates]: Train hypernetwork to generate LoRA adapters on the fly&lt;br /&gt;
** 2026-02: [https://arxiv.org/abs/2602.15902 Doc-to-LoRA: Learning to Instantly Internalize Contexts] ([https://github.com/SakanaAI/Doc-to-LoRA code])&lt;br /&gt;
** 2025-06: [https://arxiv.org/abs/2506.06105 Text-to-LoRA: Instant Transformer Adaption] ([https://github.com/SakanaAI/Text-to-LoRA])&lt;br /&gt;
&lt;br /&gt;
==Context Length==&lt;br /&gt;
* 2020: [https://ai.googleblog.com/2020/10/rethinking-attention-with-performers.html Various ideas] for scaling context window, including [https://arxiv.org/abs/2004.05150 Longformer]&lt;br /&gt;
* 2023-04-02: [https://hazyresearch.stanford.edu/blog/2023-03-27-long-learning Discussion] of ideas for how to scale context window&lt;br /&gt;
* 2023-05-11: Anthropic announces 100k window&lt;br /&gt;
* 2023-06-07: [https://magic.dev/ magic.dev] claims [https://magic.dev/blog/ltm-1 5M tokens coming soon]&lt;br /&gt;
* 2023-07-05: Microsoft describes [https://arxiv.org/abs/2307.02486 LongNet], with 1 billion token window&lt;br /&gt;
* 2023-07-11: [https://arxiv.org/abs/2307.03170 Focused Transformer] 256k&lt;br /&gt;
* 2023-11-06: [https://openai.com/blog/new-models-and-developer-products-announced-at-devday GPT-4 turbo] 128k&lt;br /&gt;
* 2023-11-22: [https://techcrunch.com/2023/11/21/anthropic-claude-2-1/ Anthropic Claude 2.1] 200k&lt;br /&gt;
* 2023-12-13: [https://arxiv.org/abs/2312.00752 Mamba] alternative&lt;br /&gt;
* 2024-01-04: [https://arxiv.org/abs/2401.01325 LongLM] to extend context window&lt;br /&gt;
* 2024-02-15: [https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/#architecture Gemini 1.5] 1M tokens&lt;br /&gt;
* 2024-03-04: [https://www.anthropic.com/news/claude-3-family Anthropic Claude 3] 200k&lt;br /&gt;
* 2024-03-08: [https://arxiv.org/abs/2403.05530 Google claims] Gemini 1.5 can scale to 10M&lt;br /&gt;
* 2024-04-10: Google [https://arxiv.org/abs/2404.07143 preprint] demonstrates infinite context length by using compressive memory&lt;br /&gt;
* 2024-04-12: Meta et al. demonstrate [https://arxiv.org/abs/2404.08801 Megalodon] that enables infinite context via a more efficient architecture&lt;br /&gt;
* 2024-04-14: Google presents [https://arxiv.org/abs/2404.09173 TransformerFAM], which leverages a feedback loop so it attends to its own latent representations, acting as working memory and provides effectively infinite context&lt;br /&gt;
* 2024-10-31: [https://arxiv.org/abs/2410.23771 What is Wrong with Perplexity for Long-context Language Modeling?]&lt;br /&gt;
* [https://x.com/MiniMax__AI/status/1879226391352549451 2025-01-14]: [https://www.minimaxi.com/en/news/minimax-01-series-2 MiniMax-01] 4M ([https://www.minimaxi.com/en/news/minimax-01-series-2 paper])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1883557964759654608 2025-01-27]: [https://qwenlm.github.io/blog/qwen2.5-1m/ Qwen2.5-1M] ([https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2.5-1M/Qwen2_5_1M_Technical_Report.pdf report])&lt;br /&gt;
* 2025-02-14: [https://arxiv.org/abs/2502.08910 InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU] 3M&lt;br /&gt;
* [https://x.com/AnimaAnandkumar/status/1897449851941744648 2025-02-18]: [https://arxiv.org/abs/2502.12574 HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading] ([https://github.com/wdlctc/headinfer code])&lt;br /&gt;
* 2025-02-18: [https://arxiv.org/abs/2502.12962 Infinite Retrieval: Attention Enhanced LLMs in Long-Context Processing]&lt;br /&gt;
* 2025-02-19: [https://github.com/MoonshotAI/MoBA MoBA: Mixture of Block Attention for Long-Context LLMs]&lt;br /&gt;
* 2025-02-27: [https://arxiv.org/abs/2502.20082 LongRoPE2: Near-Lossless LLM Context Window Scaling] ([https://github.com/microsoft/LongRoPE code])&lt;br /&gt;
* [https://x.com/sundarpichai/status/1904579419496386736 2025-03-25]: [https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/ Gemini 2.5 Pro] [https://x.com/pvncher/status/1904685092053606715 1M]&lt;br /&gt;
* 2025-04-05: Meta [https://ai.meta.com/blog/llama-4-multimodal-intelligence/ Llama 4] 10M&lt;br /&gt;
* 2025-04-14: OpenAI [https://openai.com/index/gpt-4-1/ GPT-4.1] 1M&lt;br /&gt;
* 2025-12-04: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans/Miras] 10M&lt;br /&gt;
* 2025-12-13: [https://arxiv.org/abs/2512.12167 Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings]&lt;br /&gt;
* 2026-03-18: [https://github.com/EverMind-AI/MSA/blob/main/paper/MSA__Memory_Sparse_Attention_for_Efficient_End_to_End_Memory_Model_Scaling_to_100M_Tokens.pdf MSA: Memory Sparse Attention for Efficient End-to-End Memory Model Scaling to 100M Tokens] ([https://github.com/EverMind-AI/MSA code]) 100M&lt;br /&gt;
&lt;br /&gt;
==Extended Context==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00663 Titans: Learning to Memorize at Test Time]&lt;br /&gt;
&lt;br /&gt;
==Context Remaking==&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.00436 Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval]&lt;br /&gt;
* 2025-08: [https://blog.plasticlabs.ai/blog/Memory-as-Reasoning Memory as Reasoning (Memory is Prediction)]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25140 ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.04618 Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.24601 Recursive Language Models] (model searches/queries the full context)&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.02553 SimpleMem: Efficient Lifelong Memory for LLM Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07190 Active Context Compression: Autonomous Memory Management in LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Retrieval beyond RAG==&lt;br /&gt;
See also: [[AI_tools#Retrieval_Augmented_Generation_.28RAG.29|AI tools: Retrieval Augmented Generation (RAG)]]&lt;br /&gt;
* 2024-10: Microsoft: [https://arxiv.org/abs/2410.10450 KBLaM: Knowledge Base augmented Language Model]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11919 RetroLLM: Empowering Large Language Models to Retrieve Fine-grained Evidence within Generation]&lt;br /&gt;
* 2025-03: Microsoft: [https://www.microsoft.com/en-us/research/blog/introducing-kblam-bringing-plug-and-play-external-knowledge-to-llms/ Introducing KBLaM: Bringing plug-and-play external knowledge to LLMs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06266 Cartridges: Lightweight and general-purpose long context representations via self-study]&lt;br /&gt;
* 2025-07: [https://arxiv.org/pdf/2507.07957 MIRIX: Multi-Agent Memory System for LLM-Based Agents] ([https://mirix.io/ mirix])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16153 Memento: Fine-tuning LLM Agents without Fine-tuning LLMs]&lt;br /&gt;
&lt;br /&gt;
==Working Memory==&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.18069 Improving Factuality with Explicit Working Memory]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03192 MemRL: Self-Evolving Agents via Runtime Reinforcement Learning on Episodic Memory]&lt;br /&gt;
&lt;br /&gt;
==Long-Term Memory==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.19413 Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory]&lt;br /&gt;
&lt;br /&gt;
* 2025-12: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans + Miras]&lt;br /&gt;
** [https://arxiv.org/abs/2504.13173 It&amp;#039;s All Connected: A Journey Through Test-Time Memorization, Attentional Bias, Retention, and Online Optimization]&lt;br /&gt;
** [https://arxiv.org/abs/2501.00663 Titans: Learning to Memorize at Test Time]&lt;br /&gt;
&lt;br /&gt;
===Storage and Retrieval===&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04439 ArcMemo: Abstract Reasoning Composition with Lifelong LLM Memory]&lt;br /&gt;
* 2026-01: [https://www.arxiv.org/abs/2601.07372 Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Episodic Memory===&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.11901 Larimar: Large Language Models with Episodic Memory Control]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16153 AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs]&lt;br /&gt;
&lt;br /&gt;
==Continual Learning==&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.00275 Architecture Matters in Continual Learning]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15103 Continual Learning via Sparse Memory Finetuning]&lt;br /&gt;
* 2025-11: [https://research.google/blog/introducing-nested-learning-a-new-ml-paradigm-for-continual-learning/ Introducing Nested Learning: A new ML paradigm for continual learning]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.16175 Learning to Discover at Test Time]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.19897 Self-Distillation Enables Continual Learning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.07755 Learning to Continually Learn via Meta-learning Agentic Memory Designs]&lt;br /&gt;
&lt;br /&gt;
=Updating Weights at Inference-time=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.06252 Transformer&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt;: Self-adaptive LLMs]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14143 Beyond Turing: Memory-Amortized Inference as a Foundation for Cognitive Computation]&lt;br /&gt;
* 2026-02: Sakana AI: [https://pub.sakana.ai/doc-to-lora/ Instant LLM Updates]: Train hypernetwork to generate LoRA adapters on the fly&lt;br /&gt;
** 2026-02: [https://arxiv.org/abs/2602.15902 Doc-to-LoRA: Learning to Instantly Internalize Contexts] ([https://github.com/SakanaAI/Doc-to-LoRA code])&lt;br /&gt;
** 2025-06: [https://arxiv.org/abs/2506.06105 Text-to-LoRA: Instant Transformer Adaption] ([https://github.com/SakanaAI/Text-to-LoRA])&lt;br /&gt;
&lt;br /&gt;
==Parameters as Tokens==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.23168 TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters] ([https://github.com/Haiyang-W/TokenFormer code])&lt;br /&gt;
&lt;br /&gt;
=Internal Thought Representation Space=&lt;br /&gt;
==Visual Thinking==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05874 VideoRAG: Retrieval-Augmented Generation over Video Corpus]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.07542 Imagine while Reasoning in Space: Multimodal Visualization-of-Thought]&lt;br /&gt;
&lt;br /&gt;
==Neural (non-token) Latent Representation==&lt;br /&gt;
* 2024-11: Microsoft: [https://arxiv.org/abs/2411.02820 DroidSpeak: KV Cache Sharing for Cross-LLM Communication and Multi-LLM Serving]: LLMs invent their own inter-communication language&lt;br /&gt;
* 2024-12: Meta: [https://arxiv.org/abs/2412.06769 Training Large Language Models to Reason in a Continuous Latent Space]: feeding the latent representation directly back into the model, instead of tokenizing intermediate thoughts (Chain of Continuous Thought, a.k.a. Coconut)&lt;br /&gt;
* 2024-12: Meta: [https://arxiv.org/abs/2412.08821 Large Concept Models: Language Modeling in a Sentence Representation Space]: train a model that operates at a higher level of abstraction than typical word/token LLMs; model operates in a space of concept embeddings (more akin to full sentences than individual words)&lt;br /&gt;
* 2024-12: Meta: [https://ai.meta.com/research/publications/byte-latent-transformer-patches-scale-better-than-tokens/ Byte Latent Transformer: Patches Scale Better Than Tokens]: Instead of tokenization, dynamically convert input byte-stream into patches, yielding gains in compute efficiency, with minimal loss in performance&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.13171 Compressed Chain of Thought: Efficient Reasoning Through Dense Representations]&lt;br /&gt;
* 2024-12: Google DeepMind: [https://arxiv.org/abs/2412.17747 Deliberation in Latent Space via Differentiable Cache Augmentation]&lt;br /&gt;
* 2024-12: [https://github.com/jerber/lang-jepa LANG-JEPA: Learning to Think in Latent Space]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19201 Efficient Reasoning with Hidden Thinking] ([https://github.com/shawnricecake/Heima code])&lt;br /&gt;
* 2025-02: [https://huggingface.co/tomg-group-umd/huginn-0125 Huginn-0125]: [https://arxiv.org/abs/2502.05171 Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach] ([https://github.com/seal-rg/recurrent-pretraining code], [https://huggingface.co/tomg-group-umd/huginn-0125 model])&lt;br /&gt;
* 2025-02: Meta: [https://arxiv.org/abs/2502.08524 LLM Pretraining with Continuous Concepts] (CoCoMix)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2505.12514 Reasoning by Superposition: A Theoretical Perspective on Chain of Continuous Thought]&lt;br /&gt;
&lt;br /&gt;
=Altered Transformer=&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.19737 Better &amp;amp; Faster Large Language Models via Multi-token Prediction]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.06676 I Don&amp;#039;t Know: Explicit Modeling of Uncertainty with an &amp;lt;nowiki&amp;gt;[IDK]&amp;lt;/nowiki&amp;gt; Token]&lt;br /&gt;
* 2025-04: Meta: [https://arxiv.org/abs/2504.00927 Multi-Token Attention]&lt;br /&gt;
&lt;br /&gt;
==Generation Order==&lt;br /&gt;
* 2019-02: [https://arxiv.org/abs/1902.02192 Non-Monotonic Sequential Text Generation]&lt;br /&gt;
* 2019-04: [https://arxiv.org/abs/1904.09324 Mask-Predict: Parallel Decoding of Conditional Masked Language Models]&lt;br /&gt;
* 2019-06: [https://arxiv.org/abs/1906.09601 Sequence Generation: From Both Sides to the Middle]&lt;br /&gt;
* 2020-04: [https://arxiv.org/abs/2004.11579 Probabilistically Masked Language Model Capable of Autoregressive Generation in Arbitrary Word Order]&lt;br /&gt;
* 2021-12: [https://arxiv.org/abs/2112.10543 Spiral Language Modeling]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.09930 FiLM: Fill-in Language Models for Any-Order Generation]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03582 Integrating Randomness in Large Language Models: A Linear Congruential Generator Approach for Generating Clinically Relevant Content]&lt;br /&gt;
&lt;br /&gt;
==Diffusion Language Models==&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.03687 Pard: Permutation-Invariant Autoregressive Diffusion for Graph Generation]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09992 Large Language Diffusion Models]&lt;br /&gt;
* 2025-02: [https://www.inceptionlabs.ai/ Inception Labs] [https://www.inceptionlabs.ai/news Mercury] model ([https://chat.inceptionlabs.ai/ online demo])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.09573 Block Diffusion: Interpolating Between Autoregressive and Diffusion Language Models] ([https://m-arriola.com/bd3lms/ project], [https://github.com/kuleshov-group/bd3lms code], [https://huggingface.co/collections/kuleshov-group/bd3-lms-67be95f81b96b15fec50d53f hf])&lt;br /&gt;
* 2025-04: [https://hkunlp.github.io/blog/2025/dream/ Dream 7B: Introducing Dream 7B, the most powerful open diffusion large language model to date]&lt;br /&gt;
* 2025-04: [https://dllm-reasoning.github.io/d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning] ([https://dllm-reasoning.github.io/media/preprint.pdf preprint], [https://github.com/dllm-reasoning/d1 code])&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01928 Esoteric Language Models] ([https://s-sahoo.com/Eso-LMs/ project])&lt;br /&gt;
&lt;br /&gt;
===Related: Image Synthesis via Autoregression/Diffusion===&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.01400 Sequential Data Generation with Groupwise Diffusion Process]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.09470 Rolling Diffusion Models]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.11039 Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model]&lt;br /&gt;
&lt;br /&gt;
==Sampling==&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.01104 softmax is not enough (for sharp out-of-distribution)]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06215 Corrector Sampling in Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15260 Deep Think with Confidence] ([https://jiaweizzhao.github.io/deepconf/ project])&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.14901 Reasoning with Sampling: Your Base Model is Smarter Than You Think]&lt;br /&gt;
&lt;br /&gt;
=Daydreaming, brainstorming, pre-generation=&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
* 2025-07: Gwern: [https://gwern.net/ai-daydreaming Daydreaming]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Pre-generation&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
* 2025-11: [https://inference.net/blog/project-aella Project OSSAS: Custom LLMs to process 100 Million Research Papers] ([https://huggingface.co/inference-net models], [https://aella.inference.net/embeddings visualization])&lt;br /&gt;
&lt;br /&gt;
=Missing Elements=&lt;br /&gt;
* Memory&lt;br /&gt;
* Continuous learning/update&lt;br /&gt;
* Robust contextual model&lt;br /&gt;
* Long-time-horizon coherence&lt;br /&gt;
* Fluid intelligence&lt;br /&gt;
* Agency&lt;br /&gt;
* Modeling of self&lt;br /&gt;
* [https://gwern.net/ai-daydreaming Daydreaming]&lt;br /&gt;
&lt;br /&gt;
=Memes=&lt;br /&gt;
* Andrej Karpathy: &lt;br /&gt;
** 2015-05: &amp;quot;Hallucination&amp;quot; in [https://karpathy.github.io/2015/05/21/rnn-effectiveness/ The Unreasonable Effectiveness of Recurrent Neural Networks]&lt;br /&gt;
** 2017-11: [https://karpathy.medium.com/software-2-0-a64152b37c35 Software 2.0] ([https://x.com/karpathy/status/893576281375219712 &amp;quot;Gradient descent can write code better than you. I&amp;#039;m sorry.&amp;quot;])&lt;br /&gt;
** 2022-10: [https://x.com/karpathy/status/1582807367988654081 Transformers as general-purpose differentiable computers] ([https://www.youtube.com/watch?v=9uw3F6rndnA talk])&lt;br /&gt;
** 2023-01: [https://x.com/karpathy/status/1617979122625712128 The hottest new programming language is English]&lt;br /&gt;
** 2023-09: [https://x.com/karpathy/status/1707437820045062561 LLM as kernel of a new Operating System] ([https://x.com/karpathy/status/1723140519554105733 diagram], [https://www.threads.com/@karpathy/post/CzehPtxPEF3 OS analogies])&lt;br /&gt;
** 2024-07: [https://x.com/karpathy/status/1816531576228053133 Jagged Intelligence] (c.f. [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Mollick paper])&lt;br /&gt;
** 2025-02: [https://x.com/karpathy/status/1886192184808149383 Vibe coding]&lt;br /&gt;
** 2025-06: [https://www.latent.space/p/s3 Software 3.0] ([https://www.youtube.com/watch?v=LCEmiRjPEtQ&amp;amp;t=1s talk]): &amp;quot;Prompts as Programs&amp;quot;. Software 1.0 is code; 2.0 is model weights; 3.0 is prompts.&lt;br /&gt;
** 2025-06: [https://x.com/karpathy/status/1937902205765607626 &amp;quot;Context Engineering&amp;quot; instead of &amp;quot;Prompt Engineering&amp;quot;]&lt;br /&gt;
** 2025-06: [https://x.com/karpathy/status/1938626382248149433 LLMs as &amp;quot;cognitive cores&amp;quot;]&lt;br /&gt;
** 2025-11: [https://x.com/karpathy/status/1990116666194456651?s=20 Software 1.0 easily automates what you can specify. Software 2.0 easily automates what you can verify.]&lt;br /&gt;
** 2026-01: [https://x.com/karpathy/status/2008664551445963083?s=20 The majority of the ruff ruff is people who look at the current point and people who look at the current slope]&lt;br /&gt;
** 2026-02: [https://x.com/karpathy/status/2019137879310836075 Agentic Engineering]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[Increasing AI Intelligence]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_research_trends&amp;diff=8745</id>
		<title>AI research trends</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_research_trends&amp;diff=8745"/>
		<updated>2026-03-23T17:08:40Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Context Length */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=System 2 Reasoning=&lt;br /&gt;
See: [[Increasing AI Intelligence]]&lt;br /&gt;
&lt;br /&gt;
=Memory=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13501 A Survey on the Memory Mechanism of Large Language Model based Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.09113 The AI Hippocampus: How Far are We From Human Memory?]&lt;br /&gt;
&lt;br /&gt;
==Big Ideas==&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.07755 Learning to Continually Learn via Meta-learning Agentic Memory Designs]&lt;br /&gt;
&lt;br /&gt;
==LLM Weights Memory==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09764 Memory Layers at Scale]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15103 Continual Learning via Sparse Memory Finetuning]&lt;br /&gt;
* 2026-01: [https://developer.nvidia.com/blog/reimagining-llm-memory-using-context-as-training-data-unlocks-models-that-learn-at-test-time/ Reimagining LLM Memory: Using Context as Training Data Unlocks Models That Learn at Test-Time] (Nvidia)&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.02151 Entropy-Adaptive Fine-Tuning: Resolving Confident Conflicts to Mitigate Forgetting]&lt;br /&gt;
* 2026-02: Sakana AI: [https://pub.sakana.ai/doc-to-lora/ Instant LLM Updates]: Train hypernetwork to generate LoRA adapters on the fly&lt;br /&gt;
** 2026-02: [https://arxiv.org/abs/2602.15902 Doc-to-LoRA: Learning to Instantly Internalize Contexts] ([https://github.com/SakanaAI/Doc-to-LoRA code])&lt;br /&gt;
** 2025-06: [https://arxiv.org/abs/2506.06105 Text-to-LoRA: Instant Transformer Adaption] ([https://github.com/SakanaAI/Text-to-LoRA])&lt;br /&gt;
&lt;br /&gt;
==Context Length==&lt;br /&gt;
* 2020: [https://ai.googleblog.com/2020/10/rethinking-attention-with-performers.html Various ideas] for scaling context window, including [https://arxiv.org/abs/2004.05150 Longformer]&lt;br /&gt;
* 2023-04-02: [https://hazyresearch.stanford.edu/blog/2023-03-27-long-learning Discussion] of ideas for how to scale context window&lt;br /&gt;
* 2023-05-11: Anthropic announces 100k window&lt;br /&gt;
* 2023-06-07: [https://magic.dev/ magic.dev] claims [https://magic.dev/blog/ltm-1 5M tokens coming soon]&lt;br /&gt;
* 2023-07-05: Microsoft describes [https://arxiv.org/abs/2307.02486 LongNet], with 1 billion token window&lt;br /&gt;
* 2023-07-11: [https://arxiv.org/abs/2307.03170 Focused Transformer] 256k&lt;br /&gt;
* 2023-11-06: [https://openai.com/blog/new-models-and-developer-products-announced-at-devday GPT-4 turbo] 128k&lt;br /&gt;
* 2023-11-22: [https://techcrunch.com/2023/11/21/anthropic-claude-2-1/ Anthropic Claude 2.1] 200k&lt;br /&gt;
* 2023-12-13: [https://arxiv.org/abs/2312.00752 Mamba] alternative&lt;br /&gt;
* 2024-01-04: [https://arxiv.org/abs/2401.01325 LongLM] to extend context window&lt;br /&gt;
* 2024-02-15: [https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/#architecture Gemini 1.5] 1M tokens&lt;br /&gt;
* 2024-03-04: [https://www.anthropic.com/news/claude-3-family Anthropic Claude 3] 200k&lt;br /&gt;
* 2024-03-08: [https://arxiv.org/abs/2403.05530 Google claims] Gemini 1.5 can scale to 10M&lt;br /&gt;
* 2024-04-10: Google [https://arxiv.org/abs/2404.07143 preprint] demonstrates infinite context length by using compressive memory&lt;br /&gt;
* 2024-04-12: Meta et al. demonstrate [https://arxiv.org/abs/2404.08801 Megalodon] that enables infinite context via a more efficient architecture&lt;br /&gt;
* 2024-04-14: Google presents [https://arxiv.org/abs/2404.09173 TransformerFAM], which leverages a feedback loop so it attends to its own latent representations, acting as working memory and provides effectively infinite context&lt;br /&gt;
* 2024-10-31: [https://arxiv.org/abs/2410.23771 What is Wrong with Perplexity for Long-context Language Modeling?]&lt;br /&gt;
* [https://x.com/MiniMax__AI/status/1879226391352549451 2025-01-14]: [https://www.minimaxi.com/en/news/minimax-01-series-2 MiniMax-01] 4M ([https://www.minimaxi.com/en/news/minimax-01-series-2 paper])&lt;br /&gt;
* [https://x.com/Alibaba_Qwen/status/1883557964759654608 2025-01-27]: [https://qwenlm.github.io/blog/qwen2.5-1m/ Qwen2.5-1M] ([https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2.5-1M/Qwen2_5_1M_Technical_Report.pdf report])&lt;br /&gt;
* 2025-02-14: [https://arxiv.org/abs/2502.08910 InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU] 3M&lt;br /&gt;
* [https://x.com/AnimaAnandkumar/status/1897449851941744648 2025-02-18]: [https://arxiv.org/abs/2502.12574 HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading] ([https://github.com/wdlctc/headinfer code])&lt;br /&gt;
* 2025-02-18: [https://arxiv.org/abs/2502.12962 Infinite Retrieval: Attention Enhanced LLMs in Long-Context Processing]&lt;br /&gt;
* 2025-02-19: [https://github.com/MoonshotAI/MoBA MoBA: Mixture of Block Attention for Long-Context LLMs]&lt;br /&gt;
* 2025-02-27: [https://arxiv.org/abs/2502.20082 LongRoPE2: Near-Lossless LLM Context Window Scaling] ([https://github.com/microsoft/LongRoPE code])&lt;br /&gt;
* [https://x.com/sundarpichai/status/1904579419496386736 2025-03-25]: [https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/ Gemini 2.5 Pro] [https://x.com/pvncher/status/1904685092053606715 1M]&lt;br /&gt;
* 2025-04-05: Meta [https://ai.meta.com/blog/llama-4-multimodal-intelligence/ Llama 4] 10M&lt;br /&gt;
* 2025-04-14: OpenAI [https://openai.com/index/gpt-4-1/ GPT-4.1] 1M&lt;br /&gt;
* 2025-12-04: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans/Miras] 10M&lt;br /&gt;
* 2025-12-13: [https://arxiv.org/abs/2512.12167 Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings]&lt;br /&gt;
* 2026-03-18: [https://github.com/EverMind-AI/MSA MSA: Memory Sparse Attention] 100M&lt;br /&gt;
&lt;br /&gt;
==Extended Context==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00663 Titans: Learning to Memorize at Test Time]&lt;br /&gt;
&lt;br /&gt;
==Context Remaking==&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.00436 Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval]&lt;br /&gt;
* 2025-08: [https://blog.plasticlabs.ai/blog/Memory-as-Reasoning Memory as Reasoning (Memory is Prediction)]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25140 ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.04618 Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.24601 Recursive Language Models] (model searches/queries the full context)&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.02553 SimpleMem: Efficient Lifelong Memory for LLM Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07190 Active Context Compression: Autonomous Memory Management in LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Retrieval beyond RAG==&lt;br /&gt;
See also: [[AI_tools#Retrieval_Augmented_Generation_.28RAG.29|AI tools: Retrieval Augmented Generation (RAG)]]&lt;br /&gt;
* 2024-10: Microsoft: [https://arxiv.org/abs/2410.10450 KBLaM: Knowledge Base augmented Language Model]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11919 RetroLLM: Empowering Large Language Models to Retrieve Fine-grained Evidence within Generation]&lt;br /&gt;
* 2025-03: Microsoft: [https://www.microsoft.com/en-us/research/blog/introducing-kblam-bringing-plug-and-play-external-knowledge-to-llms/ Introducing KBLaM: Bringing plug-and-play external knowledge to LLMs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06266 Cartridges: Lightweight and general-purpose long context representations via self-study]&lt;br /&gt;
* 2025-07: [https://arxiv.org/pdf/2507.07957 MIRIX: Multi-Agent Memory System for LLM-Based Agents] ([https://mirix.io/ mirix])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16153 Memento: Fine-tuning LLM Agents without Fine-tuning LLMs]&lt;br /&gt;
&lt;br /&gt;
==Working Memory==&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.18069 Improving Factuality with Explicit Working Memory]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03192 MemRL: Self-Evolving Agents via Runtime Reinforcement Learning on Episodic Memory]&lt;br /&gt;
&lt;br /&gt;
==Long-Term Memory==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.19413 Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory]&lt;br /&gt;
&lt;br /&gt;
* 2025-12: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans + Miras]&lt;br /&gt;
** [https://arxiv.org/abs/2504.13173 It&amp;#039;s All Connected: A Journey Through Test-Time Memorization, Attentional Bias, Retention, and Online Optimization]&lt;br /&gt;
** [https://arxiv.org/abs/2501.00663 Titans: Learning to Memorize at Test Time]&lt;br /&gt;
&lt;br /&gt;
===Storage and Retrieval===&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04439 ArcMemo: Abstract Reasoning Composition with Lifelong LLM Memory]&lt;br /&gt;
* 2026-01: [https://www.arxiv.org/abs/2601.07372 Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Episodic Memory===&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.11901 Larimar: Large Language Models with Episodic Memory Control]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16153 AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs]&lt;br /&gt;
&lt;br /&gt;
==Continual Learning==&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.00275 Architecture Matters in Continual Learning]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.15103 Continual Learning via Sparse Memory Finetuning]&lt;br /&gt;
* 2025-11: [https://research.google/blog/introducing-nested-learning-a-new-ml-paradigm-for-continual-learning/ Introducing Nested Learning: A new ML paradigm for continual learning]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.16175 Learning to Discover at Test Time]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.19897 Self-Distillation Enables Continual Learning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.07755 Learning to Continually Learn via Meta-learning Agentic Memory Designs]&lt;br /&gt;
&lt;br /&gt;
=Updating Weights at Inference-time=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.06252 Transformer&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt;: Self-adaptive LLMs]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14143 Beyond Turing: Memory-Amortized Inference as a Foundation for Cognitive Computation]&lt;br /&gt;
* 2026-02: Sakana AI: [https://pub.sakana.ai/doc-to-lora/ Instant LLM Updates]: Train hypernetwork to generate LoRA adapters on the fly&lt;br /&gt;
** 2026-02: [https://arxiv.org/abs/2602.15902 Doc-to-LoRA: Learning to Instantly Internalize Contexts] ([https://github.com/SakanaAI/Doc-to-LoRA code])&lt;br /&gt;
** 2025-06: [https://arxiv.org/abs/2506.06105 Text-to-LoRA: Instant Transformer Adaption] ([https://github.com/SakanaAI/Text-to-LoRA])&lt;br /&gt;
&lt;br /&gt;
==Parameters as Tokens==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.23168 TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters] ([https://github.com/Haiyang-W/TokenFormer code])&lt;br /&gt;
&lt;br /&gt;
=Internal Thought Representation Space=&lt;br /&gt;
==Visual Thinking==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05874 VideoRAG: Retrieval-Augmented Generation over Video Corpus]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.07542 Imagine while Reasoning in Space: Multimodal Visualization-of-Thought]&lt;br /&gt;
&lt;br /&gt;
==Neural (non-token) Latent Representation==&lt;br /&gt;
* 2024-11: Microsoft: [https://arxiv.org/abs/2411.02820 DroidSpeak: KV Cache Sharing for Cross-LLM Communication and Multi-LLM Serving]: LLMs invent their own inter-communication language&lt;br /&gt;
* 2024-12: Meta: [https://arxiv.org/abs/2412.06769 Training Large Language Models to Reason in a Continuous Latent Space]: feeding the latent representation directly back into the model, instead of tokenizing intermediate thoughts (Chain of Continuous Thought, a.k.a. Coconut)&lt;br /&gt;
* 2024-12: Meta: [https://arxiv.org/abs/2412.08821 Large Concept Models: Language Modeling in a Sentence Representation Space]: train a model that operates at a higher level of abstraction than typical word/token LLMs; model operates in a space of concept embeddings (more akin to full sentences than individual words)&lt;br /&gt;
* 2024-12: Meta: [https://ai.meta.com/research/publications/byte-latent-transformer-patches-scale-better-than-tokens/ Byte Latent Transformer: Patches Scale Better Than Tokens]: Instead of tokenization, dynamically convert input byte-stream into patches, yielding gains in compute efficiency, with minimal loss in performance&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.13171 Compressed Chain of Thought: Efficient Reasoning Through Dense Representations]&lt;br /&gt;
* 2024-12: Google DeepMind: [https://arxiv.org/abs/2412.17747 Deliberation in Latent Space via Differentiable Cache Augmentation]&lt;br /&gt;
* 2024-12: [https://github.com/jerber/lang-jepa LANG-JEPA: Learning to Think in Latent Space]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19201 Efficient Reasoning with Hidden Thinking] ([https://github.com/shawnricecake/Heima code])&lt;br /&gt;
* 2025-02: [https://huggingface.co/tomg-group-umd/huginn-0125 Huginn-0125]: [https://arxiv.org/abs/2502.05171 Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach] ([https://github.com/seal-rg/recurrent-pretraining code], [https://huggingface.co/tomg-group-umd/huginn-0125 model])&lt;br /&gt;
* 2025-02: Meta: [https://arxiv.org/abs/2502.08524 LLM Pretraining with Continuous Concepts] (CoCoMix)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2505.12514 Reasoning by Superposition: A Theoretical Perspective on Chain of Continuous Thought]&lt;br /&gt;
&lt;br /&gt;
=Altered Transformer=&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.19737 Better &amp;amp; Faster Large Language Models via Multi-token Prediction]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.06676 I Don&amp;#039;t Know: Explicit Modeling of Uncertainty with an &amp;lt;nowiki&amp;gt;[IDK]&amp;lt;/nowiki&amp;gt; Token]&lt;br /&gt;
* 2025-04: Meta: [https://arxiv.org/abs/2504.00927 Multi-Token Attention]&lt;br /&gt;
&lt;br /&gt;
==Generation Order==&lt;br /&gt;
* 2019-02: [https://arxiv.org/abs/1902.02192 Non-Monotonic Sequential Text Generation]&lt;br /&gt;
* 2019-04: [https://arxiv.org/abs/1904.09324 Mask-Predict: Parallel Decoding of Conditional Masked Language Models]&lt;br /&gt;
* 2019-06: [https://arxiv.org/abs/1906.09601 Sequence Generation: From Both Sides to the Middle]&lt;br /&gt;
* 2020-04: [https://arxiv.org/abs/2004.11579 Probabilistically Masked Language Model Capable of Autoregressive Generation in Arbitrary Word Order]&lt;br /&gt;
* 2021-12: [https://arxiv.org/abs/2112.10543 Spiral Language Modeling]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.09930 FiLM: Fill-in Language Models for Any-Order Generation]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03582 Integrating Randomness in Large Language Models: A Linear Congruential Generator Approach for Generating Clinically Relevant Content]&lt;br /&gt;
&lt;br /&gt;
==Diffusion Language Models==&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.03687 Pard: Permutation-Invariant Autoregressive Diffusion for Graph Generation]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09992 Large Language Diffusion Models]&lt;br /&gt;
* 2025-02: [https://www.inceptionlabs.ai/ Inception Labs] [https://www.inceptionlabs.ai/news Mercury] model ([https://chat.inceptionlabs.ai/ online demo])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.09573 Block Diffusion: Interpolating Between Autoregressive and Diffusion Language Models] ([https://m-arriola.com/bd3lms/ project], [https://github.com/kuleshov-group/bd3lms code], [https://huggingface.co/collections/kuleshov-group/bd3-lms-67be95f81b96b15fec50d53f hf])&lt;br /&gt;
* 2025-04: [https://hkunlp.github.io/blog/2025/dream/ Dream 7B: Introducing Dream 7B, the most powerful open diffusion large language model to date]&lt;br /&gt;
* 2025-04: [https://dllm-reasoning.github.io/d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning] ([https://dllm-reasoning.github.io/media/preprint.pdf preprint], [https://github.com/dllm-reasoning/d1 code])&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01928 Esoteric Language Models] ([https://s-sahoo.com/Eso-LMs/ project])&lt;br /&gt;
&lt;br /&gt;
===Related: Image Synthesis via Autoregression/Diffusion===&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.01400 Sequential Data Generation with Groupwise Diffusion Process]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.09470 Rolling Diffusion Models]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.11039 Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model]&lt;br /&gt;
&lt;br /&gt;
==Sampling==&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.01104 softmax is not enough (for sharp out-of-distribution)]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06215 Corrector Sampling in Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15260 Deep Think with Confidence] ([https://jiaweizzhao.github.io/deepconf/ project])&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.14901 Reasoning with Sampling: Your Base Model is Smarter Than You Think]&lt;br /&gt;
&lt;br /&gt;
=Daydreaming, brainstorming, pre-generation=&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
* 2025-07: Gwern: [https://gwern.net/ai-daydreaming Daydreaming]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Pre-generation&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
* 2025-11: [https://inference.net/blog/project-aella Project OSSAS: Custom LLMs to process 100 Million Research Papers] ([https://huggingface.co/inference-net models], [https://aella.inference.net/embeddings visualization])&lt;br /&gt;
&lt;br /&gt;
=Missing Elements=&lt;br /&gt;
* Memory&lt;br /&gt;
* Continuous learning/update&lt;br /&gt;
* Robust contextual model&lt;br /&gt;
* Long-time-horizon coherence&lt;br /&gt;
* Fluid intelligence&lt;br /&gt;
* Agency&lt;br /&gt;
* Modeling of self&lt;br /&gt;
* [https://gwern.net/ai-daydreaming Daydreaming]&lt;br /&gt;
&lt;br /&gt;
=Memes=&lt;br /&gt;
* Andrej Karpathy: &lt;br /&gt;
** 2015-05: &amp;quot;Hallucination&amp;quot; in [https://karpathy.github.io/2015/05/21/rnn-effectiveness/ The Unreasonable Effectiveness of Recurrent Neural Networks]&lt;br /&gt;
** 2017-11: [https://karpathy.medium.com/software-2-0-a64152b37c35 Software 2.0] ([https://x.com/karpathy/status/893576281375219712 &amp;quot;Gradient descent can write code better than you. I&amp;#039;m sorry.&amp;quot;])&lt;br /&gt;
** 2022-10: [https://x.com/karpathy/status/1582807367988654081 Transformers as general-purpose differentiable computers] ([https://www.youtube.com/watch?v=9uw3F6rndnA talk])&lt;br /&gt;
** 2023-01: [https://x.com/karpathy/status/1617979122625712128 The hottest new programming language is English]&lt;br /&gt;
** 2023-09: [https://x.com/karpathy/status/1707437820045062561 LLM as kernel of a new Operating System] ([https://x.com/karpathy/status/1723140519554105733 diagram], [https://www.threads.com/@karpathy/post/CzehPtxPEF3 OS analogies])&lt;br /&gt;
** 2024-07: [https://x.com/karpathy/status/1816531576228053133 Jagged Intelligence] (c.f. [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Mollick paper])&lt;br /&gt;
** 2025-02: [https://x.com/karpathy/status/1886192184808149383 Vibe coding]&lt;br /&gt;
** 2025-06: [https://www.latent.space/p/s3 Software 3.0] ([https://www.youtube.com/watch?v=LCEmiRjPEtQ&amp;amp;t=1s talk]): &amp;quot;Prompts as Programs&amp;quot;. Software 1.0 is code; 2.0 is model weights; 3.0 is prompts.&lt;br /&gt;
** 2025-06: [https://x.com/karpathy/status/1937902205765607626 &amp;quot;Context Engineering&amp;quot; instead of &amp;quot;Prompt Engineering&amp;quot;]&lt;br /&gt;
** 2025-06: [https://x.com/karpathy/status/1938626382248149433 LLMs as &amp;quot;cognitive cores&amp;quot;]&lt;br /&gt;
** 2025-11: [https://x.com/karpathy/status/1990116666194456651?s=20 Software 1.0 easily automates what you can specify. Software 2.0 easily automates what you can verify.]&lt;br /&gt;
** 2026-01: [https://x.com/karpathy/status/2008664551445963083?s=20 The majority of the ruff ruff is people who look at the current point and people who look at the current slope]&lt;br /&gt;
** 2026-02: [https://x.com/karpathy/status/2019137879310836075 Agentic Engineering]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[Increasing AI Intelligence]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_creativity&amp;diff=8744</id>
		<title>AI creativity</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_creativity&amp;diff=8744"/>
		<updated>2026-03-23T17:07:23Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Research */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Research=&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-01: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4686415 Creativity and AI]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.02980 Surveying the Effects of Quality, Diversity, and Complexity in Synthetic Data From Large Language Models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
* 2025-03: Midjourney: [https://www.arxiv.org/abs/2503.17126 Modifying Large Language Model Post-Training for Diverse Creative Writing]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.12320 Has the Creativity of Large-Language Models peaked? An analysis of inter- and intra-LLM variability]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.14442 Creative Preference Optimization]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.01171 Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.20635 Why Did Apple Fall To The Ground: Evaluating Curiosity In Large Language Model]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.22954 Artificial Hivemind: The Open-Ended Homogeneity of Language Models (and Beyond)]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41562-025-02331-1 A large-scale comparison of divergent creativity in humans and large language models]&lt;br /&gt;
* 2026-01: [https://www.arxiv.org/abs/2601.01576 OpenNovelty: An LLM-powered Agentic System for Verifiable Scholarly Novelty Assessment]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41598-025-25157-3 Divergent creativity in humans and large language models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
&lt;br /&gt;
=Benchmarks=&lt;br /&gt;
See: [[AI_benchmarks#Creativity| AI benchmarks &amp;gt; Creativity]]&lt;br /&gt;
&lt;br /&gt;
=Collapse=&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02209 Generative Monoculture in Large Language Models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
==Analysis==&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
==LLM==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13928 LLMs Can Get &amp;quot;Brain Rot&amp;quot;!]&lt;br /&gt;
==Image Models==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
==Solutions==&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI_benchmarks|AI benchmarks]] &amp;gt; [[AI_benchmarks#Assess_Specific_Attributes|Assess Specific Attributes]] &amp;gt; [[AI_benchmarks#Creativity|Creativity]]&lt;br /&gt;
* [[AI_and_Humans|AI and Humans]] &amp;gt; [[AI_and_Humans#AI_out-performs_humans|AI out-performs humans]] &amp;gt; [[AI_and_Humans#Creativity|Creativity]]&lt;br /&gt;
* [[AI_and_Humans|AI and Humans]] &amp;gt; [[AI_and_Humans#AI_improves_human_work|AI improves human work]] &amp;gt; Creativity&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8743</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8743"/>
		<updated>2026-03-23T17:07:15Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Creativity */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.19087 Serendipity by Design: Evaluating the Impact of Cross-domain Mappings on Human and LLM Creativity]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
* 2026-03: Anthropic: [https://www.anthropic.com/features/81k-interviews What 81,000 people want from AI]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8742</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8742"/>
		<updated>2026-03-19T18:52:19Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;br /&gt;
* March 2026: [https://higgsfield.ai/original-series Higgsfield Original Series]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_safety&amp;diff=8741</id>
		<title>AI safety</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_safety&amp;diff=8741"/>
		<updated>2026-03-19T17:55:58Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Description of Safety Concerns */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Learning Resources=&lt;br /&gt;
==Light==&lt;br /&gt;
* [https://orxl.org/ai-doom.html a casual intro to AI doom and alignment] (2022)&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
** [https://interactive.keepthefuturehuman.ai/ Interactive Explainer]&lt;br /&gt;
** [https://keepthefuturehuman.ai/essay/ Essay: Keep the Future Human]&lt;br /&gt;
** [https://www.youtube.com/watch?v=27KDl2uPiL8 We Can’t Stop AI – Here’s What To Do Instead] (4m video, 2025)&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late] (15m video, 2025)&lt;br /&gt;
* Tristan Harris TED talk (15m): [https://www.ted.com/talks/tristan_harris_why_ai_is_our_ultimate_test_and_greatest_invitation Why AI is our ultimate test and greatest invitation]&lt;br /&gt;
** Text version: Center for Humane Technology: [https://centerforhumanetechnology.substack.com/p/the-narrow-path-why-ai-is-our-ultimate The Narrow Path: Why AI is Our Ultimate Test and Greatest Invitation]&lt;br /&gt;
* [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI]&lt;br /&gt;
* 2024-10: [https://www.youtube.com/watch?v=xfMQ7hzyFW4 Writing Doom]: short film on Superintelligence (27m video)&lt;br /&gt;
* 2026-03: [https://www.youtube.com/watch?v=Nl7-bRFSZBs The AI book that&amp;#039;s freaking out national security advisors] (44m video)&lt;br /&gt;
&lt;br /&gt;
==Deep==&lt;br /&gt;
* [https://www.thecompendium.ai/ The Compendium: Humanity risks extinction from its very creations — AIs.] (2024)&lt;br /&gt;
* [https://www.aisafetybook.com/ Introduction to AI Safety, Ethics, and Society] (Dan Hendrycks, [https://www.safe.ai/ Center for AI Safety])&lt;br /&gt;
* [https://aisafety.info/ AI Safety FAQ]&lt;br /&gt;
* [https://deepmindsafetyresearch.medium.com/introducing-our-short-course-on-agi-safety-1072adb7912c DeepMind short course on AGI safety]&lt;br /&gt;
&lt;br /&gt;
=Description of Safety Concerns=&lt;br /&gt;
==Key Concepts==&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Instrumental_convergence Instrumental Convergence]&lt;br /&gt;
* [https://www.lesswrong.com/w/orthogonality-thesis Orthogonality Thesis]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/SzecSPYxqRa5GCaSF/clarifying-inner-alignment-terminology Inner/outer alignment]&lt;br /&gt;
* [https://www.alignmentforum.org/w/mesa-optimization Mesa-optimization]&lt;br /&gt;
* [https://www.lesswrong.com/posts/N6vZEnCn6A95Xn39p/are-we-in-an-ai-overhang Overhang]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/pdaGN6pQyQarFHXF4/reward-is-not-the-optimization-target Reward is not the optimization target] (Alex Turner)&lt;br /&gt;
* 80,000 hours:&lt;br /&gt;
** [https://80000hours.org/problem-profiles/risks-from-power-seeking-ai/ Risks from power-seeking AI systems]&lt;br /&gt;
** [https://80000hours.org/problem-profiles/gradual-disempowerment/ Gradual disempowerment]&lt;br /&gt;
** [https://80000hours.org/problem-profiles/catastrophic-ai-misuse/ Catastrophic AI misuse]&lt;br /&gt;
&lt;br /&gt;
==Medium-term Risks==&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=xoVJKj8lcNQ A.I. Dilemma – Tristan Harris and Aza Raskin” (video)] ([https://assets-global.website-files.com/5f0e1294f002b1bb26e1f304/64224a9051a6637c1b60162a_65-your-undivided-attention-The-AI-Dilemma-transcript.pdf podcast transcript]): raises concern about human ability to handle these transformations&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=KCSsKV5F4xc Daniel Schmachtenberger and Liv Boeree (video)]: AI could accelerate perverse social dynamics&lt;br /&gt;
* 2023-10: [https://arxiv.org/pdf/2310.11986 Sociotechnical Safety Evaluation of Generative AI Systems] (Google DeepMind)&lt;br /&gt;
* 2024-02: [https://yoshuabengio.org/2024/02/26/towards-a-cautious-scientist-ai-with-convergent-safety-bounds/ Towards a Cautious Scientist AI with Convergent Safety Bounds] (Yoshua Bengio)&lt;br /&gt;
* 2024-07: [https://yoshuabengio.org/2024/07/09/reasoning-through-arguments-against-taking-ai-safety-seriously/ Reasoning through arguments against taking AI safety seriously] (Yoshua Bengio)&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20702 The Singapore Consensus on Global AI Safety Research Priorities]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.adz1697 How malicious AI swarms can threaten democracy: The fusion of agentic AI and LLMs marks a new frontier in information warfare] (Science Magazine, [https://arxiv.org/abs/2506.06299 preprint])&lt;br /&gt;
* 2026-01: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI] (Dario Amodei)&lt;br /&gt;
* 2026-02: [https://www.noahpinion.blog/p/updated-thoughts-on-ai-risk Updated thoughts on AI risk: Things have gotten scarier since 2023] ([https://x.com/Noahpinion Noah Smith])&lt;br /&gt;
&lt;br /&gt;
==Long-term  (x-risk)==&lt;br /&gt;
* 2015-02: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-1 Machine intelligence, part 1]&lt;br /&gt;
* 2019-03: Daniel Kokotajlo and Wei Dai: [https://www.lesswrong.com/posts/WXvt8bxYnwBYpy9oT/the-main-sources-of-ai-risk The Main Sources of AI Risk?]&lt;br /&gt;
* 2022-06: Eliezer Yudkowsky: [https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities List AGI Ruin: A List of Lethalities]&lt;br /&gt;
* 2024-11: Marcus Arvan: [https://link.springer.com/article/10.1007/s00146-024-02113-9 ‘Interpretability’ and ‘alignment’ are fool’s errands: a proof that controlling misaligned large language models is the best anyone can hope for]&lt;br /&gt;
* 2025-04: [https://michaelnotebook.com/xriskbrief/index.html ASI existential risk: reconsidering alignment as a goal]&lt;br /&gt;
* 2025-12: Philip Trammell and Leopold Aschenbrenner: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth]&lt;br /&gt;
&lt;br /&gt;
=Status=&lt;br /&gt;
* 2025-01: [https://assets.publishing.service.gov.uk/media/679a0c48a77d250007d313ee/International_AI_Safety_Report_2025_accessible_f.pdf International Safety Report: The International Scientific Report on the Safety of Advanced AI (January 2025)]&lt;br /&gt;
* [https://ailabwatch.org/ AI Lab Watch] (safety scorecard)&lt;br /&gt;
&lt;br /&gt;
==Assessmment==&lt;br /&gt;
* [https://aiassessmentscale.com/ AI Assessment Scale (AIAS)]: A practical framework to guide the appropriate and ethical use of generative AI in assessment design, empowering educators to make purposeful, evidence-based decisions&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16534 Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report]&lt;br /&gt;
&lt;br /&gt;
==Policy==&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.05694 On the Limitations of Compute Thresholds as a Governance Strategy] Sara Hooker&lt;br /&gt;
* 2024-07: [https://www.cigionline.org/static/documents/AI-challenges.pdf Framework Convention on Global AI Challenges] ([https://www.cigionline.org/ CIGI])&lt;br /&gt;
* 2024-08: NIST guidelines: [https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.800-1.ipd.pdf Managing Misuse Risk for Dual-Use Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Proposals==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.18359 Responsible AI Agents]&lt;br /&gt;
* 2025-03: [https://controlai.com/ Control AI] [https://controlai.com/dip The Direct Institutional Plan]&lt;br /&gt;
* 2025-04: Google DeepMind: [https://deepmind.google/discover/blog/taking-a-responsible-path-to-agi/ Taking a responsible path to AGI]&lt;br /&gt;
** Paper: [https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/evaluating-potential-cybersecurity-threats-of-advanced-ai/An_Approach_to_Technical_AGI_Safety_Apr_2025.pdf An Approach to Technical AGI Safety and Security]&lt;br /&gt;
&lt;br /&gt;
=Research=&lt;br /&gt;
* 2008: [https://selfawaresystems.com/wp-content/uploads/2008/01/ai_drives_final.pdf The Basic AI Drives]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2209.00626v1 The alignment problem from a deep learning perspective]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.03827 Discovering Latent Knowledge in Language Models Without Supervision]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.08582 Pretraining Language Models with Human Preferences]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03279 Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the MACHIAVELLI Benchmark]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.15324 Model evaluation for extreme risks] (DeepMind)&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.03047 Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.17492 Preference Ranking Optimization for Human Alignment]&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.06259 Self-Alignment with Instruction Backtranslation]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.08702 Debate Helps Supervise Unreliable Experts]&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/weak-to-strong-generalization.pdf Weak-to-Strong Generalization: Eliciting Strong Capabilities with Weak Supervision] (OpenAI, [https://openai.com/research/weak-to-strong-generalization blog])&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf Practices for Governing Agentic AI Systems] (OpenAI, [https://openai.com/index/practices-for-governing-agentic-ai-systems/ blog])&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist through Safety Training] (Anthropic)&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13208 The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions] (OpenAI)&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.04622 On scalable oversight with weak LLMs judging strong LLMs]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.21792 Safetywashing: Do AI Safety Benchmarks Actually Measure Safety Progress?] (Dan Hendrycks et al.)&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.00761 Tamper-Resistant Safeguards for Open-Weight LLMs] ([https://www.tamper-resistant-safeguards.com/ project], [https://github.com/rishub-tamirisa/tamper-resistance/ code])&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04614 Better Alignment with Instruction Back-and-Forth Translation]&lt;br /&gt;
* 2024-10: [https://cdn.openai.com/papers/first-person-fairness-in-chatbots.pdf First-Person Fairness in Chatbots] (OpenAI, [https://openai.com/index/evaluating-fairness-in-chatgpt/ blog])&lt;br /&gt;
* 2024-10: [https://assets.anthropic.com/m/377027d5b36ac1eb/original/Sabotage-Evaluations-for-Frontier-Models.pdf Sabotage evaluations for frontier models] (Anthropic, [https://www.anthropic.com/research/sabotage-evaluations blog])&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/983c85a201a962f/original/Alignment-Faking-in-Large-Language-Models-full-paper.pdf Alignment Faking in Large Language Models] (Anthropic)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03556 Best-of-N Jailbreaking] ([https://github.com/jplhughes/bon-jailbreaking code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16325 Towards Safe and Honest AI Agents with Neural Self-Other Overlap]&lt;br /&gt;
** 2024-07: [https://www.lesswrong.com/posts/hzt9gHpNwA2oHtwKX/self-other-overlap-a-neglected-approach-to-ai-alignment Self-Other Overlap: A Neglected Approach to AI Alignment]&lt;br /&gt;
** 2025-03: [https://www.lesswrong.com/posts/jtqcsARGtmgogdcLT/reducing-llm-deception-at-scale-with-self-other-overlap-fine Reducing LLM deception at scale with self-other overlap fine-tuning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16339 Deliberative Alignment: Reasoning Enables Safer Language Models] (OpenAI)&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/papers/trading-inference-time-compute-for-adversarial-robustness-20250121_1.pdf Trading Inference-Time Compute for Adversarial Robustness] (OpenAI, [https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness/ blog])&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18837 Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming] (Anthropic, [https://www.anthropic.com/research/constitutional-classifiers blog], &lt;br /&gt;
* 2025-02: [https://drive.google.com/file/d/1QAzSj24Fp0O6GfkskmnULmI1Hmx7k_EJ/view Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs] ([https://www.emergent-values.ai/ site], [https://github.com/centerforaisafety/emergent-values github])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.07776 Auditing Prompt Caching in Language Model APIs]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14143 Multi-Agent Risks from Advanced AI]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2209.00626v7 The Alignment Problem from a Deep Learning Perspective]&lt;br /&gt;
* 2025-03: [https://assets.anthropic.com/m/317564659027fb33/original/Auditing-Language-Models-for-Hidden-Objectives.pdf Auditing language models for hidden objectives] (Anthropic, [https://www.anthropic.com/research/auditing-hidden-objectives blog])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13621 Superalignment with Dynamic Human Values]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.15125 Contemplative Wisdom for Superalignment]&lt;br /&gt;
* 2025-04: [https://www.lesswrong.com/posts/x59FhzuM9yuvZHAHW/untitled-draft-yhra Scaling Laws for Scalable Oversight] ([https://arxiv.org/abs/2504.18530 preprint], [https://github.com/subhashk01/oversight-scaling-laws code])&lt;br /&gt;
* 2025-06: [https://assets.anthropic.com/m/4fb35becb0cd87e1/original/SHADE-Arena-Paper.pdf SHADE-Arena: Evaluating sabotage and monitoring in LLM agents] (Anthropic, [https://www.anthropic.com/research/shade-arena-sabotage-monitoring blog])&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13609 Avoiding Obfuscation with Prover-Estimator Debate]&lt;br /&gt;
* 2025-06: [https://cdn.openai.com/pdf/a130517e-9633-47bc-8397-969807a43a23/emergent_misalignment_paper.pdf Persona Features Control Emergent Misalignment] (OpenAI, [https://openai.com/index/emergent-misalignment/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2506.18032 Why Do Some Language Models Fake Alignment While Others Don&amp;#039;t?] (Anthropic, [https://github.com/safety-research/open-source-alignment-faking code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.11473 Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety]&lt;br /&gt;
* 2025-09: [https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/ Detecting and reducing scheming in AI models]&lt;br /&gt;
* 2025-11: [https://assets.anthropic.com/m/74342f2c96095771/original/Natural-emergent-misalignment-from-reward-hacking-paper.pdf Natural Emergent Misalignment from Reward Hacking in Production RL] (Anthropic, [https://www.anthropic.com/research/emergent-misalignment-reward-hacking blog])&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16856 Distributional AGI Safety]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2511.22662 Difficulties with Evaluating a Deception Detector for AIs]&lt;br /&gt;
* 2025-12: [https://cdn.openai.com/pdf/d57827c6-10bc-47fe-91aa-0fde55bd3901/monitoring-monitorability.pdf Monitoring Monitorability] (OpenAI)&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09937-5 Training large language models on narrow tasks can lead to broad misalignment]&lt;br /&gt;
** 2025-02: Preprint: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/pdf/2601.23045 The Hot Mess of AI: How Does Misalignment Scale with Model Intelligence and Task Complexity?] (Anthropic [https://alignment.anthropic.com/2026/hot-mess-of-ai/ blog])&lt;br /&gt;
* 2026-03: [https://cdn.openai.com/pdf/a21c39c1-fa07-41db-9078-973a12620117/cot_controllability.pdf Reasoning Models Struggle to Control their Chains of Thought] (OpenAI [https://openai.com/index/reasoning-models-chain-of-thought-controllability/ blog])&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
&lt;br /&gt;
==Demonstrations of Negative Use Capabilities==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.00586 Evaluating Large Language Models&amp;#039; Capability to Launch Fully Automated Spear Phishing Campaigns: Validated on Human Subjects]&lt;br /&gt;
* 2025-04: [https://www.nathanlabenz.com/ Nathan Labenz] ([https://www.cognitiverevolution.ai/ The Cognitive Revolution]): [https://docs.google.com/presentation/d/1mvkpg1mtAvGzTiiwYPc6bKOGsQXDIwMb-ytQECb3i7I/edit#slide=id.g252d9e67d86_0_16 AI Bad Behavior]&lt;br /&gt;
&lt;br /&gt;
==Threat Vectors==&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.07192 Poisoning Attacks on LLMs Require a Near-constant Number of Poison Samples]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8740</id>
		<title>AI understanding</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8740"/>
		<updated>2026-03-18T17:37:50Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Psychology */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Interpretability=&lt;br /&gt;
* 2017-01: [https://arxiv.org/abs/1704.01444 Learning to Generate Reviews and Discovering Sentiment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11639 Neural Interpretable Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Concepts==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.20938 Towards Understanding the Nature of Attention with Low-Rank Sparse Decomposition] ([https://github.com/OpenMOSS/Lorsa code])&lt;br /&gt;
* 2025-08: [https://transformer-circuits.pub/2025/attention-qk/index.html Tracing Attention Computation Through Feature Interactions]&lt;br /&gt;
&lt;br /&gt;
==Mechanistic Interpretability==&lt;br /&gt;
* 2020-03: OpenAI: [https://distill.pub/2020/circuits/zoom-in/ Zoom In: An Introduction to Circuits]&lt;br /&gt;
* 2021-12: Anthropic: [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2211.00593 Interpretability in the Wild: a Circuit for Indirect Object Identification in GPT-2 Small]&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-07: Anthropic: [https://transformer-circuits.pub/2024/july-update/index.html Circuits Update]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.14926 Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-based Parameter Decomposition] ([https://www.alignmentforum.org/posts/EPefYWjuHNcNH4C7E/attribution-based-parameter-decomposition blog post])&lt;br /&gt;
* 2025-01: Review: [https://arxiv.org/abs/2501.16496 Open Problems in Mechanistic Interpretability]&lt;br /&gt;
* 2025-03: Anthropic: [https://www.anthropic.com/research/tracing-thoughts-language-model Tracing the thoughts of a large language model]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/methods.html Circuit Tracing: Revealing Computational Graphs in Language Models]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/biology.html On the Biology of a Large Language Model]&lt;br /&gt;
* 2025-11: OpenAI: [https://cdn.openai.com/pdf/41df8f28-d4ef-43e9-aed2-823f9393e470/circuit-sparsity-paper.pdf Weight-sparse transformers have interpretable circuits] ([https://openai.com/index/understanding-neural-networks-through-sparse-circuits/ blog])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.13548 Patterning: The Dual of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Semanticity==&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.08600 Sparse Autoencoders Find Highly Interpretable Features in Language Models]&lt;br /&gt;
* Anthropic monosemanticity interpretation of LLM features:&lt;br /&gt;
** 2023-10: [https://transformer-circuits.pub/2023/monosemantic-features/index.html Towards Monosemanticity: Decomposing Language Models With Dictionary Learning]&lt;br /&gt;
** 2024-05: [https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet]&lt;br /&gt;
* 2024-06: OpenaAI: [https://arxiv.org/abs/2406.04093 Scaling and evaluating sparse autoencoders]&lt;br /&gt;
* 2024-08: [https://www.alignmentforum.org/posts/TMAmHh4DdMr4nCSr5/showing-sae-latents-are-not-atomic-using-meta-saes Showing SAE Latents Are Not Atomic Using Meta-SAEs] ([https://metasae.streamlit.app/?page=Feature+Explorer&amp;amp;feature=11329 demo])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08201 Efficient Dictionary Learning with Switch Sparse Autoencoders] ([https://github.com/amudide/switch_sae code]) More efficient SAE generation&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.14670 Decomposing The Dark Matter of Sparse Autoencoders] ([https://github.com/JoshEngels/SAE-Dark-Matter code]) Shows that SAE errors are predictable&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13928 Automatically Interpreting Millions of Features in Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.21331 Beyond Interpretability: The Gains of Feature Monosemanticity on Model Robustness]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.04139 Monet: Mixture of Monosemantic Experts for Transformers]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/zbebxYCqsryPALh8C/matryoshka-sparse-autoencoders Matryoshka Sparse Autoencoders]&lt;br /&gt;
* 2024-12: [https://www.alignmentforum.org/posts/rKM9b6B2LqwSB5ToN/learning-multi-level-features-with-matryoshka-saes Learning Multi-Level Features with Matryoshka SAEs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19406 Low-Rank Adapting Models for Sparse Autoencoders]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.03714 Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.00177 Steering Large Language Model Activations in Sparse Spaces]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01776 Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01824 From superposition to sparse codes: interpretable representations in neural networks]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18878 I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20063 SAEs Are Good for Steering -- If You Select the Right Features]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15679 Dense SAE Latents Are Features, Not Bugs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20790 Stochastic Parameter Decomposition] ([https://github.com/goodfire-ai/spd code], [https://www.goodfire.ai/papers/stochastic-param-decomp blog])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
&lt;br /&gt;
===Counter-Results===&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.12016 Towards falsifiable interpretability research]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16615 Sparse Autoencoders Trained on the Same Data Learn Different Features]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17148 AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17727 Sparse Autoencoders Can Interpret Randomly Initialized Transformers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.04878 Sparse Autoencoders Do Not Find Canonical Units of Analysis]&lt;br /&gt;
* 2025-03: [https://www.alignmentforum.org/posts/4uXCAJNuPKtKBsi28/ Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research]&lt;br /&gt;
&lt;br /&gt;
==Meta-cognition==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.15674 Activation Oracles: Training and Evaluating LLMs as General-Purpose Activation Explainers]&lt;br /&gt;
&lt;br /&gt;
==Coding Models==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sparse Auto Encoders&amp;#039;&amp;#039;&amp;#039;: See Semanticity.&lt;br /&gt;
* [https://github.com/saprmarks/dictionary_learning dictionary_learning]&lt;br /&gt;
* [https://transformer-circuits.pub/2024/jan-update/index.html#predict-future Predicting Future Activations]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11944 Transcoders Find Interpretable LLM Feature Circuits]&lt;br /&gt;
* 2024-10: [https://transformer-circuits.pub/2024/crosscoders/index.html Sparse Crosscoders for Cross-Layer Features and Model Diffing]&lt;br /&gt;
&lt;br /&gt;
==Reward Functions==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12491 Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL]&lt;br /&gt;
&lt;br /&gt;
==Symbolic and Notation==&lt;br /&gt;
* [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* [https://www.arxiv.org/abs/2407.09468 Beyond Euclid: An Illustrated Guide to Modern Machine Learning with Geometric, Topological, and Algebraic Structures]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02423 On the Anatomy of Attention]: Introduces category-theoretic diagrammatic formalism for DL architectures&lt;br /&gt;
* 2024-11: [https://x.com/vtabbott_/status/1860268276569506250 diagrams to represent algorithms]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03317 FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awareness]&lt;br /&gt;
&lt;br /&gt;
==Mathematical==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.13762 Unveiling the Hidden Structure of Self-Attention via Kernel Principal Component Analysis]&lt;br /&gt;
&lt;br /&gt;
==Geometric==&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.03658 The Linear Representation Hypothesis and the Geometry of Large Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.01506 The Geometry of Categorical and Hierarchical Concepts in Large Language Models]&lt;br /&gt;
** Natural hierarchies of concepts---which occur throughout natural language and especially in scientific ontologies---are represented in the model&amp;#039;s internal vectorial space as polytopes that can be decomposed into simplexes of mutually-exclusive categories.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02678 Reasoning in Large Language Models: A Geometric Perspective]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.17592 Deep Manifold Part 1: Anatomy of Neural Network Manifold]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.19750 The Geometry of Concepts: Sparse Autoencoder Feature Structure]&lt;br /&gt;
** Tegmark et al. report multi-scale structure: 1) “atomic” small-scale, 2) “brain” intermediate-scale, and 3) “galaxy” large-scale&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.08009 The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.09782 The Geometry of Reasoning: Flowing Logics in Representation Space]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/linebreaks/index.html When Models Manipulate Manifolds: The Geometry of a Counting Task]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.26745 Deep sequence models tend to memorize geometrically; it is unclear why]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
==Topography==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.06002 The Molecular Structure of Thought: Mapping the Topology of Long Chain-of-Thought Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2023-07Jul: [https://arxiv.org/abs/2307.13702 Measuring Faithfulness in Chain-of-Thought Reasoning] [https://x.com/davidad/status/1839641113432305790 roughly] proves that sufficiently large models do not generate CoT that actually captures their internal reasoning)&lt;br /&gt;
&lt;br /&gt;
[[Image:GYe31yXXQAABwaZ.jpeg|300px]]&lt;br /&gt;
&lt;br /&gt;
=Heuristic Understanding=&lt;br /&gt;
* 2022-09: Janus: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators]&lt;br /&gt;
&lt;br /&gt;
==Emergent Internal Model Building==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.15936 A Theory for Emergence of Complex Skills in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19370v1 Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01622 General agents contain world models]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20328 Video models are zero-shot learners and reasoners]&lt;br /&gt;
&lt;br /&gt;
===Semantic Directions===&lt;br /&gt;
Directions, e.g.: f(king)-f(man)+f(woman)=f(queen) or f(sushi)-f(Japan)+f(Italy)=f(pizza)&lt;br /&gt;
* [https://arxiv.org/abs/1301.3781 Efficient Estimation of Word Representations in Vector Space]&lt;br /&gt;
* [https://aclanthology.org/N13-1090/ Linguistic Regularities in Continuous Space Word Representations]&lt;br /&gt;
* [https://aclanthology.org/C16-1332 Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen]&lt;br /&gt;
* [https://aclanthology.org/D14-1162/ Glove: Global vectors for word representation]&lt;br /&gt;
* [https://doi.org/10.1109/BigData.2015.7364114 Using Word2Vec to process big text data]&lt;br /&gt;
* [https://arxiv.org/abs/2310.06824 The geometry of truth: Emergent linear structure in large language model representations of true/false datasets] (true/false)&lt;br /&gt;
* [https://arxiv.org/abs/2403.10381 Monotonic Representation of Numeric Properties in Language Models] (numeric directions)&lt;br /&gt;
Task vectors:&lt;br /&gt;
* [https://arxiv.org/abs/2310.15213 Function Vectors in Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.15916 In-context learning creates task vectors]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/5FGXmJ3wqgGRcbyH7/extracting-sae-task-features-for-in-context-learning Extracting sae task features for in-context learning]&lt;br /&gt;
* [https://arxiv.org/abs/2412.12276 Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers]&lt;br /&gt;
Reasoning:&lt;br /&gt;
* [https://openreview.net/forum?id=OwhVWNOBcz Understanding Reasoning in Thinking Language Models via Steering Vectors]&lt;br /&gt;
&lt;br /&gt;
===Feature Geometry Reproduces Problem-space===&lt;br /&gt;
* [https://arxiv.org/abs/2210.13382 Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2309.00941 Emergent linear representations in world models of self-supervised sequence models] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* [https://doi.org/10.1038/s41562-023-01659-w Emergent analogical reasoning in large language models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.02207 Language Models Represent Space and Time] (Maps of world, US)&lt;br /&gt;
* [https://arxiv.org/abs/2405.14860 Not All Language Model Features Are Linear] (Days of week form ring, etc.)&lt;br /&gt;
* [https://arxiv.org/abs/2406.03689 Evaluating the World Model Implicit in a Generative Model] (Map of Manhattan)&lt;br /&gt;
* [https://iopscience.iop.org/article/10.1088/1748-9326/ad2891 Reliable precipitation nowcasting using probabilistic diffusion models]. Generation of precipitation map imagery is predictive of actual future weather; implies model is learning scientifically-relevant modeling.&lt;br /&gt;
* [https://arxiv.org/abs/2405.07987 The Platonic Representation Hypothesis]: Different models (including across modalities) are converging to a consistent world model.&lt;br /&gt;
* [https://arxiv.org/abs/2501.00070 ICLR: In-Context Learning of Representations]&lt;br /&gt;
* [https://arxiv.org/abs/2502.00873 Language Models Use Trigonometry to Do Addition]: Numbers arranged in helix to enable addition&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
&lt;br /&gt;
===Capturing Physics===&lt;br /&gt;
* 2020-09: [https://arxiv.org/abs/2009.08292 Learning to Identify Physical Parameters from Video Using Differentiable Physics]&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.00419 Self-Supervised Learning for Videos: A Survey]&lt;br /&gt;
* 2025-02: Fair at Meta: [https://arxiv.org/abs/2502.11831 Intuitive physics understanding emerges from self-supervised pretraining on natural videos]&lt;br /&gt;
&lt;br /&gt;
===Theory of Mind===&lt;br /&gt;
* [https://arxiv.org/abs/2302.02083 Evaluating Large Language Models in Theory of Mind Tasks]&lt;br /&gt;
* [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-01: [https://www.arxiv.org/abs/2501.09038 Do generative video models learn physical principles from watching videos?] ([https://physics-iq.github.io/ project], [https://github.com/google-deepmind/physics-IQ-benchmark code])&lt;br /&gt;
* 2025-06: [https://machinelearning.apple.com/research/illusion-of-thinking The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21521 Potemkin Understanding in Large Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21876 Do Vision-Language Models Have Internal World Models? Towards an Atomic Evaluation]&lt;br /&gt;
&lt;br /&gt;
==Information Processing==&lt;br /&gt;
* 2019-03: [https://arxiv.org/abs/1903.05789 Diagnosing and Enhancing VAE Models]&lt;br /&gt;
* 2021-03: [https://arxiv.org/abs/2103.05247 Pretrained Transformers as Universal Computation Engines]&lt;br /&gt;
* 2022-10: [https://arxiv.org/abs/2210.08344 How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03843 Why think step by step? Reasoning emerges from the locality of experience]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.04444 What&amp;#039;s the Magic Word? A Control Theory of LLM Prompting]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12875 Chain of Thought Empowers Transformers to Solve Inherently Serial Problems]: Proves that transformers can solve any problem, if they can generate sufficient intermediate tokens&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.20311 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process]&lt;br /&gt;
** Models learning reasoning skills (they are not merely memorizing solution templates). They can mentally generate simple short plans (like humans).&lt;br /&gt;
** When presented facts, models develop internal understanding of what parameters (recursively) depend on each other. This occurs even before an explicit question is asked (i.e. before the task is defined). This appears to be different from human reasoning.&lt;br /&gt;
** Model depth matters for reasoning. This cannot be mitigated by chain-of-thought prompting (which allow models to develop and then execute plans) since even a single CoT step may require deep, multi-step reasoning/planning.&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.01992 Ask, and it shall be given: Turing completeness of prompting]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08775 Layers at Similar Depths Generate Similar Activations Across LLM Architectures]&lt;br /&gt;
&lt;br /&gt;
===Generalization===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]&lt;br /&gt;
&lt;br /&gt;
===Grokking===&lt;br /&gt;
* 2022-01: [https://arxiv.org/abs/2201.02177 Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets]&lt;br /&gt;
* 2022-05: [https://arxiv.org/abs/2205.10343 Towards Understanding Grokking: An Effective Theory of Representation Learning]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.10463 Critical Data Size of Language Models from a Grokking Perspective]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15175 Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18624 How to explain grokking]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.21519 Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
===Tests of Resilience to Dropouts/etc.===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15390 Explorations of Self-Repair in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15786 What Matters in Transformers? Not All Attention is Needed]&lt;br /&gt;
** Removing entire transformer blocks leads to significant performance degradation&lt;br /&gt;
** Removing MLP layers results in significant performance degradation&lt;br /&gt;
** Removing attention layers causes almost no performance degradation&lt;br /&gt;
** E.g. half of attention layers are deleted (48% speed-up), leads to only 2.4% decrease in the benchmarks&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19384 The Remarkable Robustness of LLMs: Stages of Inference?]&lt;br /&gt;
** They intentionally break the network (swapping layers), yet it continues to work remarkably well. This suggests LLMs are quite robust, and allows them to identify different stages in processing.&lt;br /&gt;
** They also use these interventions to infer what different layers are doing. They break apart the LLM transformer layers into four stages:&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Detokenization:&amp;#039;&amp;#039;&amp;#039; Raw tokens are converted into meaningful entities that take into account local context (especially using nearby tokens).&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Feature engineering:&amp;#039;&amp;#039;&amp;#039; Features are progressively refined. Factual knowledge is leveraged.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Prediction ensembling:&amp;#039;&amp;#039;&amp;#039; Predictions (for the ultimately-selected next-token) emerge. A sort of consensus voting is used, with “prediction neurons” and &amp;quot;suppression neurons&amp;quot; playing a major role in upvoting/downvoting.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Residual sharpening:&amp;#039;&amp;#039;&amp;#039; The semantic representations are collapsed into specific next-token predictions. There is a strong emphasis on suppression neurons eliminating options. The confidence is calibrated.&lt;br /&gt;
** This structure can be thought of as two halves (being roughly dual to each other): the first half broadens (goes from distinct tokens to a rich/elaborate concept-space) and the second half collapses (goes from rich concepts to concrete token predictions).&lt;br /&gt;
&lt;br /&gt;
==Semantic Vectors==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11717 Refusal in Language Models Is Mediated by a Single Direction]&lt;br /&gt;
* 2025-02: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs] ([https://x.com/OwainEvans_UK/status/1894436637054214509 demonstrates] [https://x.com/ESYudkowsky/status/1894453376215388644 entangling] of concepts into a single preference vector)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03666 Analogical Reasoning Inside Large Language Models: Concept Vectors and the Limits of Abstraction]&lt;br /&gt;
&lt;br /&gt;
==Other==&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00247 Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting &amp;amp; Beyond]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04282 Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding] ([https://github.com/SalesforceAIResearch/LaTRO code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.12580 Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models]: LLMs learn reasoning by extracting procedures from training data, not by memorizing specific answers&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15862 LLMs Do Not Think Step-by-step In Implicit Reasoning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
&lt;br /&gt;
===Scaling Laws===&lt;br /&gt;
* 1993: [https://proceedings.neurips.cc/paper/1993/file/1aa48fc4880bb0c9b8a3bf979d3b917e-Paper.pdf Learning Curves: Asymptotic Values and Rate of Convergence]&lt;br /&gt;
* 2017-12: [https://arxiv.org/abs/1712.00409 Deep Learning Scaling is Predictable, Empirically] (Baidu)&lt;br /&gt;
* 2019-03: [http://www.incompleteideas.net/IncIdeas/BitterLesson.html The Bitter Lesson] (Rich Sutton)&lt;br /&gt;
* 2020-01: [https://arxiv.org/abs/2001.08361 Scaling Laws for Neural Language Models] (OpenAI)&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.14701 Scaling Laws for Autoregressive Generative Modeling] (OpenAI)&lt;br /&gt;
* 2020-05: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis] (Gwern)&lt;br /&gt;
* 2021-08: [https://arxiv.org/abs/2108.07686 Scaling Laws for Deep Learning]&lt;br /&gt;
* 2021-02: [https://arxiv.org/abs/2102.06701 Explaining Neural Scaling Laws] (Google DeepMind)&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.15556 Training Compute-Optimal Large Language Models] (Chinchilla, Google DeepMind)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.04715 Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.07951 Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models]&lt;br /&gt;
* 2025-05: [https://brendel-group.github.io/llm-line/ LLMs on the Line: Data Determines Loss-To-Loss Scaling Laws]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13786 The Art of Scaling Reinforcement Learning Compute for LLMs]&lt;br /&gt;
&lt;br /&gt;
=Information Processing/Storage=&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.10689 A Theory of Usable Information Under Computational Constraints]&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.00008 Why is AI hard and Physics simple?]&lt;br /&gt;
* 2021-06: [https://arxiv.org/abs/2106.06981 Thinking Like Transformers]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.00948 Large Linguistic Models: Investigating LLMs&amp;#039; metalinguistic abilities]&lt;br /&gt;
* &amp;quot;A transformer&amp;#039;s depth affects its reasoning capabilities, whilst model size affects its knowledge capacity&amp;quot; ([https://x.com/danielhanchen/status/1835684061475655967 c.f.])&lt;br /&gt;
** 2024-02: [https://arxiv.org/abs/2402.14905 MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases]&lt;br /&gt;
** 2024-04: [https://arxiv.org/abs/2404.08819 The Illusion of State in State-Space Models] (figure 3)&lt;br /&gt;
** 2024-08: [https://arxiv.org/abs/2408.00118 Gemma 2: Improving Open Language Models at a Practical Size] (table 9)&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.10482 Schrodinger&amp;#039;s Memory: Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2407.01687 Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning]. CoT involves both memorization and (probabilitic) reasoning&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.16679 Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?]&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.03961 A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.22471 The Bayesian Geometry of Transformer Attention]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03220 From Entropy to Epiplexity: Rethinking Information for Computationally Bounded Intelligence]&lt;br /&gt;
&lt;br /&gt;
==Statistics/Math==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.05465 The emergence of clusters in self-attention dynamics]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.10794 A mathematical perspective on Transformers]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.12034 Understanding Transformers via N-gram Statistics]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.06833 Dynamic metastability in the self-attention model]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04551 Measure-to-measure interpolation using Transformers]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14697 Quantitative Clustering in Mean-Field Transformer Models]&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
===For numbers/math===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.14903 Tokenization counts: the impact of tokenization on arithmetic in frontier LLMs]: L2R vs. R2L yields different performance on math&lt;br /&gt;
&lt;br /&gt;
==Data Storage==&lt;br /&gt;
* 1988-09: [https://www.sciencedirect.com/science/article/pii/0885064X88900209 On the capabilities of multilayer perceptrons]&lt;br /&gt;
* 2006-12: [https://ieeexplore.ieee.org/document/4038449 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition] (single-layer perceptron stores &amp;gt;2 bits/parameter; MLP ~ 2*N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; bits w/ N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; params)&lt;br /&gt;
* 2016-11: [https://arxiv.org/abs/1611.09913 Capacity and Trainability in Recurrent Neural Networks] (5 bits/param)&lt;br /&gt;
* 2018-02: [https://arxiv.org/abs/1802.08232 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks]&lt;br /&gt;
* 2019-05: [https://ieeexplore.ieee.org/document/8682462 Memorization Capacity of Deep Neural Networks under Parameter Quantization]&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.08910 How Much Knowledge Can You Pack Into the Parameters of a Language Model?]&lt;br /&gt;
* 2020-08: [https://arxiv.org/abs/2008.09036 Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries] (capacity scales linearly with parameters; more training samples leads to less memorization)&lt;br /&gt;
* 2020-12: [https://arxiv.org/abs/2012.06421 When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05405 Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws] (2 bits/param)&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15720 Scaling Laws for Fact Memorization of Large Language Models] (1T params needed to memorize Wikipedia)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24832 How much do language models memorize?] (3.6 bits/parameter)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01855 Trade-offs in Data Memorization via Strong Data Processing Inequalities]&lt;br /&gt;
&lt;br /&gt;
===Reverse-Engineering Training Data===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.10364 Can We Infer Confidential Properties of Training Data from LLMs?]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15553 Approximating Language Model Training Data from Weights]&lt;br /&gt;
&lt;br /&gt;
===Compression===&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.09410 Less is More: Parameter-Free Text Classification with Gzip]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.04050 LLMZip: Lossless Text Compression using Large Language Models]&lt;br /&gt;
* 2023-07: [https://aclanthology.org/2023.findings-acl.426/ “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.10668 Language Modeling Is Compression]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07550 An Image is Worth 32 Tokens for Reconstruction and Generation]&lt;br /&gt;
&lt;br /&gt;
==Learning/Training==&lt;br /&gt;
* 2018-03: [https://arxiv.org/abs/1803.03635 The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks]: Sparse neural networks are optimal, but it is difficult to identify the right architecture and train it. Deep learning typically consists of training a dense neural network, which makes it easier to learn an internal sparse circuit optimal to a particular problem.&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11521 On the Ability of Deep Networks to Learn Symmetries from Data: A Neural Kernel Theory]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.12391 Physics of Skill Learning]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24864 ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Cross-modal knowledge transfer===&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.07519 Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.07358 Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]: CLIP learns richer set of aggregated representations (e.g. for a culture or country), vs. a vision-only model.&lt;br /&gt;
&lt;br /&gt;
==Hidden State==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06258 Emergent Response Planning in LLM]: They show that the latent representation contains information beyond that needed for the next token (i.e. the model learns to &amp;quot;plan ahead&amp;quot; and encode information relevant to future tokens)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02854 (How) Do Language Models Track State?]&lt;br /&gt;
===Convergent Representation===&lt;br /&gt;
* 2015-11: [https://arxiv.org/abs/1511.07543 Convergent Learning: Do different neural networks learn the same representations?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of Embeddings]: Evidence for [https://x.com/jxmnop/status/1925224620166128039 The Strong Platonic Representation Hypothesis]; models converge to a single consensus reality&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Function Approximation==&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]: can learn linear functions (equivalent to least-squares estimator)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09066 Teaching Algorithmic Reasoning via In-context Learning]: Simple arithmetic &lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models] ([https://github.com/ekinakyurek/google-research/tree/master/incontext code]): can learn linear regression&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.00297 Transformers learn to implement preconditioned gradient descent for in-context learning]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.03576 One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.02893 ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20545 SoS1: O1 and R1-Like Reasoning LLMs are Sum-of-Square Solvers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.21212 Transformers Learn to Implement Multi-step Gradient Descent with Chain of Thought]&lt;br /&gt;
&lt;br /&gt;
=Physics Based=&lt;br /&gt;
* 2014-01: [https://arxiv.org/abs/1401.1219 Consciousness as a State of Matter]&lt;br /&gt;
* 2016-08: [https://arxiv.org/abs/1608.08225 Why does deep and cheap learning work so well?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.23489 SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training]&lt;br /&gt;
* 2025-12: [https://www.pnas.org/doi/full/10.1073/pnas.2523012122 Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning]&lt;br /&gt;
&lt;br /&gt;
=Failure Modes=&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?]: Poor causal inference&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.12288 The Reversal Curse: LLMs trained on &amp;quot;A is B&amp;quot; fail to learn &amp;quot;B is A&amp;quot;]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.13638 Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve] (biases towards &amp;quot;common&amp;quot; numbers, in-context CoT can reduce performance by incorrectly priming, etc.)&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.16093 Visual cognition in multimodal large language models] (models lack human-like visual understanding)&lt;br /&gt;
&lt;br /&gt;
==Adversarial==&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03507 Solving adversarial examples requires solving exponential misalignment]&lt;br /&gt;
&lt;br /&gt;
==Fracture Representation==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.11581 Questioning Representational Optimism in Deep Learning: The Fractured Entangled Representation Hypothesis] ([https://github.com/akarshkumar0101/fer code])&lt;br /&gt;
&lt;br /&gt;
==Jagged Frontier==&lt;br /&gt;
* 2023-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03211 How Does Quantization Affect Multilingual LLMs?]: Quantization degrades different languages by differing amounts&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061v1 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]: Scaling laws are skill-dependent&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.18212 A Definition of AGI]&lt;br /&gt;
&lt;br /&gt;
===See also===&lt;br /&gt;
* [[AI_understanding|AI Understanding]] &amp;gt; [[AI_understanding#Psychology|Psychology]] &amp;gt; [[AI_understanding#LLM_personalities|LLM personalities]]&lt;br /&gt;
* [[AI tricks]] &amp;gt; [[AI_tricks#Prompt_Engineering|Prompt Engineering]] &amp;gt; [[AI_tricks#Brittleness|Brittleness]]&lt;br /&gt;
&lt;br /&gt;
===Conversely (AI models converge)===&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.05117 The Universal Weight Subspace Hypothesis]&lt;br /&gt;
* 2026-01: [https://avikrishna.substack.com/p/eliciting-frontier-model-character Eliciting Frontier Model Character Training: A study of personality convergence across language models]&lt;br /&gt;
&lt;br /&gt;
==Model Collapse==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.12202 Nepotistically Trained Generative-AI Models Collapse]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05280 On the Limits of Self-Improving in LLMs and Why AGI, ASI and the Singularity Are Not Near Without Symbolic Model Synthesis]&lt;br /&gt;
&lt;br /&gt;
===Analysis===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
&lt;br /&gt;
===Mitigation===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08117 Convergence Dynamics and Stabilization Strategies of Co-Evolving Generative Models]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.11111 Inducing anxiety in large language models can induce bias]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17117 From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning]&lt;br /&gt;
* 2025-07: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5357179 Call Me A Jerk: Persuading AI to Comply with Objectionable Requests]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.06047 &amp;quot;They parted illusions -- they parted disclaim marinade&amp;quot;: Misalignment as structural fidelity in LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.02606 Gender Dynamics and Homophily in a Social Network of LLM Agents]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
&lt;br /&gt;
==Persona Simulator Theory==&lt;br /&gt;
* 2022-09: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators] ([https://www.lesswrong.com/users/janus-1?from=post_header janus])&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.findings-emnlp.423/ Language Models as Agent Models]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.00805 Conditioning Predictive Models: Risks and Strategies]&lt;br /&gt;
* 2024-09: [https://www.lesswrong.com/s/qhdHbCJ3PYesL9dde Intuitive Self-Models]&lt;br /&gt;
* 2026-02: [https://alignment.anthropic.com/2026/psm/ The Persona Selection Model: Why AI Assistants might Behave like Humans] (Anthropic, [https://www.anthropic.com/research/persona-selection-model blog])&lt;br /&gt;
&lt;br /&gt;
==Allow LLM to think==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
&lt;br /&gt;
===In-context Learning===&lt;br /&gt;
* 2021-10: [https://arxiv.org/abs/2110.15943 MetaICL: Learning to Learn In Context]&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?]&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16003 Learning without training: The implicit dynamics of in-context learning]&lt;br /&gt;
&lt;br /&gt;
==Reasoning (CoT, etc.)==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18009 Large Language Models Think Too Fast To Explore Effectively]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18585 Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.08156 Are DeepSeek R1 And Other Reasoning Models More Faithful?]: reasoning models can provide faithful explanations for why their reasoning is correct&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08679 Chain-of-Thought Reasoning In The Wild Is Not Always Faithful]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.04022 Rethinking Reflection in Pre-Training]: pre-training alone already provides some amount of reflection/reasoning&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2501.18858 BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning]&lt;br /&gt;
&lt;br /&gt;
===Pathfinding===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08152 DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09284 Tree-OPO: Off-policy Monte Carlo Tree-Guided Advantage Optimization for Multistep Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06160v1 Reverse-Engineered Reasoning for Open-Ended Generation]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06941 The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2508.01191 Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens]&lt;br /&gt;
&lt;br /&gt;
==Self-Awareness and Self-Recognition and Introspection==&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.05221 Language Models (Mostly) Know What They Know]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13076 LLM Evaluators Recognize and Favor Their Own Generations]&lt;br /&gt;
* 2024-09: [https://situational-awareness-dataset.org/ Me, Myself and AI: The Situational Awareness Dataset for LLMs]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* 2024-12: [https://theaidigest.org/self-awareness AIs are becoming more self-aware. Here&amp;#039;s why that matters]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-04: [https://x.com/Josikinz/status/1907923319866716629 LLMs can guess which comic strip was generated by themselves (vs. other LLM)]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/introspection/index.html Emergent Introspective Awareness in Large Language Models] (Anthropic, [https://www.anthropic.com/research/introspection blog])&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.24661 Do Large Language Models Know What They Are Capable Of?]&lt;br /&gt;
&lt;br /&gt;
==LLM personalities==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.02618 Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04343 Psychologically Enhanced AI Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10387 The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models]&lt;br /&gt;
&lt;br /&gt;
==Quirks &amp;amp; Biases==&lt;br /&gt;
* 2025-04: [https://www.cambridge.org/core/journals/judgment-and-decision-making/article/artificial-intelligence-and-dichotomania/0421D2310727D73FAB47069FD1620AA1 Artificial intelligence and dichotomania]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.22818 Can Large Language Models Develop Gambling Addiction?]&lt;br /&gt;
&lt;br /&gt;
=Vision Models=&lt;br /&gt;
* 2017-11: Distill: [https://distill.pub/2017/feature-visualization/ Feature Visualization: How neural networks build up their understanding of images]&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.12322 Position, Padding and Predictions: A Deeper Look at Position Information in CNNs]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13181 Perception Encoder: The best visual embeddings are not at the output of the network] ([https://github.com/facebookresearch/perception_models code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_safety&amp;diff=8739</id>
		<title>AI safety</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_safety&amp;diff=8739"/>
		<updated>2026-03-18T17:37:10Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Research */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Learning Resources=&lt;br /&gt;
==Light==&lt;br /&gt;
* [https://orxl.org/ai-doom.html a casual intro to AI doom and alignment] (2022)&lt;br /&gt;
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human]&lt;br /&gt;
** [https://interactive.keepthefuturehuman.ai/ Interactive Explainer]&lt;br /&gt;
** [https://keepthefuturehuman.ai/essay/ Essay: Keep the Future Human]&lt;br /&gt;
** [https://www.youtube.com/watch?v=27KDl2uPiL8 We Can’t Stop AI – Here’s What To Do Instead] (4m video, 2025)&lt;br /&gt;
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late] (15m video, 2025)&lt;br /&gt;
* Tristan Harris TED talk (15m): [https://www.ted.com/talks/tristan_harris_why_ai_is_our_ultimate_test_and_greatest_invitation Why AI is our ultimate test and greatest invitation]&lt;br /&gt;
** Text version: Center for Humane Technology: [https://centerforhumanetechnology.substack.com/p/the-narrow-path-why-ai-is-our-ultimate The Narrow Path: Why AI is Our Ultimate Test and Greatest Invitation]&lt;br /&gt;
* [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI]&lt;br /&gt;
* 2024-10: [https://www.youtube.com/watch?v=xfMQ7hzyFW4 Writing Doom]: short film on Superintelligence (27m video)&lt;br /&gt;
* 2026-03: [https://www.youtube.com/watch?v=Nl7-bRFSZBs The AI book that&amp;#039;s freaking out national security advisors] (44m video)&lt;br /&gt;
&lt;br /&gt;
==Deep==&lt;br /&gt;
* [https://www.thecompendium.ai/ The Compendium: Humanity risks extinction from its very creations — AIs.] (2024)&lt;br /&gt;
* [https://www.aisafetybook.com/ Introduction to AI Safety, Ethics, and Society] (Dan Hendrycks, [https://www.safe.ai/ Center for AI Safety])&lt;br /&gt;
* [https://aisafety.info/ AI Safety FAQ]&lt;br /&gt;
* [https://deepmindsafetyresearch.medium.com/introducing-our-short-course-on-agi-safety-1072adb7912c DeepMind short course on AGI safety]&lt;br /&gt;
&lt;br /&gt;
=Description of Safety Concerns=&lt;br /&gt;
==Key Concepts==&lt;br /&gt;
* [https://en.wikipedia.org/wiki/Instrumental_convergence Instrumental Convergence]&lt;br /&gt;
* [https://www.lesswrong.com/w/orthogonality-thesis Orthogonality Thesis]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/SzecSPYxqRa5GCaSF/clarifying-inner-alignment-terminology Inner/outer alignment]&lt;br /&gt;
* [https://www.alignmentforum.org/w/mesa-optimization Mesa-optimization]&lt;br /&gt;
* [https://www.lesswrong.com/posts/N6vZEnCn6A95Xn39p/are-we-in-an-ai-overhang Overhang]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/pdaGN6pQyQarFHXF4/reward-is-not-the-optimization-target Reward is not the optimization target] (Alex Turner)&lt;br /&gt;
&lt;br /&gt;
==Medium-term Risks==&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=xoVJKj8lcNQ A.I. Dilemma – Tristan Harris and Aza Raskin” (video)] ([https://assets-global.website-files.com/5f0e1294f002b1bb26e1f304/64224a9051a6637c1b60162a_65-your-undivided-attention-The-AI-Dilemma-transcript.pdf podcast transcript]): raises concern about human ability to handle these transformations&lt;br /&gt;
* 2023-04: [https://www.youtube.com/watch?v=KCSsKV5F4xc Daniel Schmachtenberger and Liv Boeree (video)]: AI could accelerate perverse social dynamics&lt;br /&gt;
* 2023-10: [https://arxiv.org/pdf/2310.11986 Sociotechnical Safety Evaluation of Generative AI Systems] (Google DeepMind)&lt;br /&gt;
* 2024-02: [https://yoshuabengio.org/2024/02/26/towards-a-cautious-scientist-ai-with-convergent-safety-bounds/ Towards a Cautious Scientist AI with Convergent Safety Bounds] (Yoshua Bengio)&lt;br /&gt;
* 2024-07: [https://yoshuabengio.org/2024/07/09/reasoning-through-arguments-against-taking-ai-safety-seriously/ Reasoning through arguments against taking AI safety seriously] (Yoshua Bengio)&lt;br /&gt;
* 2025-04: [https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power AI-Enabled Coups: How a Small Group Could Use AI to Seize Power]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20702 The Singapore Consensus on Global AI Safety Research Priorities]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.adz1697 How malicious AI swarms can threaten democracy: The fusion of agentic AI and LLMs marks a new frontier in information warfare] (Science Magazine, [https://arxiv.org/abs/2506.06299 preprint])&lt;br /&gt;
* 2026-01: [https://www.darioamodei.com/essay/the-adolescence-of-technology The Adolescence of Technology: Confronting and Overcoming the Risks of Powerful AI] (Dario Amodei)&lt;br /&gt;
* 2026-02: [https://www.noahpinion.blog/p/updated-thoughts-on-ai-risk Updated thoughts on AI risk: Things have gotten scarier since 2023] ([https://x.com/Noahpinion Noah Smith])&lt;br /&gt;
&lt;br /&gt;
==Long-term  (x-risk)==&lt;br /&gt;
* 2015-02: Sam Altman: [https://blog.samaltman.com/machine-intelligence-part-1 Machine intelligence, part 1]&lt;br /&gt;
* 2019-03: Daniel Kokotajlo and Wei Dai: [https://www.lesswrong.com/posts/WXvt8bxYnwBYpy9oT/the-main-sources-of-ai-risk The Main Sources of AI Risk?]&lt;br /&gt;
* 2022-06: Eliezer Yudkowsky: [https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities List AGI Ruin: A List of Lethalities]&lt;br /&gt;
* 2024-11: Marcus Arvan: [https://link.springer.com/article/10.1007/s00146-024-02113-9 ‘Interpretability’ and ‘alignment’ are fool’s errands: a proof that controlling misaligned large language models is the best anyone can hope for]&lt;br /&gt;
* 2025-04: [https://michaelnotebook.com/xriskbrief/index.html ASI existential risk: reconsidering alignment as a goal]&lt;br /&gt;
* 2025-12: Philip Trammell and Leopold Aschenbrenner: [https://philiptrammell.com/static/Existential_Risk_and_Growth.pdf Existential Risk and Growth]&lt;br /&gt;
&lt;br /&gt;
=Status=&lt;br /&gt;
* 2025-01: [https://assets.publishing.service.gov.uk/media/679a0c48a77d250007d313ee/International_AI_Safety_Report_2025_accessible_f.pdf International Safety Report: The International Scientific Report on the Safety of Advanced AI (January 2025)]&lt;br /&gt;
* [https://ailabwatch.org/ AI Lab Watch] (safety scorecard)&lt;br /&gt;
&lt;br /&gt;
==Assessmment==&lt;br /&gt;
* [https://aiassessmentscale.com/ AI Assessment Scale (AIAS)]: A practical framework to guide the appropriate and ethical use of generative AI in assessment design, empowering educators to make purposeful, evidence-based decisions&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16534 Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report]&lt;br /&gt;
&lt;br /&gt;
==Policy==&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.05694 On the Limitations of Compute Thresholds as a Governance Strategy] Sara Hooker&lt;br /&gt;
* 2024-07: [https://www.cigionline.org/static/documents/AI-challenges.pdf Framework Convention on Global AI Challenges] ([https://www.cigionline.org/ CIGI])&lt;br /&gt;
* 2024-08: NIST guidelines: [https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.800-1.ipd.pdf Managing Misuse Risk for Dual-Use Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Proposals==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.18359 Responsible AI Agents]&lt;br /&gt;
* 2025-03: [https://controlai.com/ Control AI] [https://controlai.com/dip The Direct Institutional Plan]&lt;br /&gt;
* 2025-04: Google DeepMind: [https://deepmind.google/discover/blog/taking-a-responsible-path-to-agi/ Taking a responsible path to AGI]&lt;br /&gt;
** Paper: [https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/evaluating-potential-cybersecurity-threats-of-advanced-ai/An_Approach_to_Technical_AGI_Safety_Apr_2025.pdf An Approach to Technical AGI Safety and Security]&lt;br /&gt;
&lt;br /&gt;
=Research=&lt;br /&gt;
* 2008: [https://selfawaresystems.com/wp-content/uploads/2008/01/ai_drives_final.pdf The Basic AI Drives]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2209.00626v1 The alignment problem from a deep learning perspective]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.03827 Discovering Latent Knowledge in Language Models Without Supervision]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.08582 Pretraining Language Models with Human Preferences]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03279 Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the MACHIAVELLI Benchmark]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.15324 Model evaluation for extreme risks] (DeepMind)&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.03047 Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.17492 Preference Ranking Optimization for Human Alignment]&lt;br /&gt;
* 2023-08: [https://arxiv.org/abs/2308.06259 Self-Alignment with Instruction Backtranslation]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.08702 Debate Helps Supervise Unreliable Experts]&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/weak-to-strong-generalization.pdf Weak-to-Strong Generalization: Eliciting Strong Capabilities with Weak Supervision] (OpenAI, [https://openai.com/research/weak-to-strong-generalization blog])&lt;br /&gt;
* 2023-12: [https://cdn.openai.com/papers/practices-for-governing-agentic-ai-systems.pdf Practices for Governing Agentic AI Systems] (OpenAI, [https://openai.com/index/practices-for-governing-agentic-ai-systems/ blog])&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist through Safety Training] (Anthropic)&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13208 The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions] (OpenAI)&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.04622 On scalable oversight with weak LLMs judging strong LLMs]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.21792 Safetywashing: Do AI Safety Benchmarks Actually Measure Safety Progress?] (Dan Hendrycks et al.)&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.00761 Tamper-Resistant Safeguards for Open-Weight LLMs] ([https://www.tamper-resistant-safeguards.com/ project], [https://github.com/rishub-tamirisa/tamper-resistance/ code])&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04614 Better Alignment with Instruction Back-and-Forth Translation]&lt;br /&gt;
* 2024-10: [https://cdn.openai.com/papers/first-person-fairness-in-chatbots.pdf First-Person Fairness in Chatbots] (OpenAI, [https://openai.com/index/evaluating-fairness-in-chatgpt/ blog])&lt;br /&gt;
* 2024-10: [https://assets.anthropic.com/m/377027d5b36ac1eb/original/Sabotage-Evaluations-for-Frontier-Models.pdf Sabotage evaluations for frontier models] (Anthropic, [https://www.anthropic.com/research/sabotage-evaluations blog])&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/983c85a201a962f/original/Alignment-Faking-in-Large-Language-Models-full-paper.pdf Alignment Faking in Large Language Models] (Anthropic)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03556 Best-of-N Jailbreaking] ([https://github.com/jplhughes/bon-jailbreaking code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16325 Towards Safe and Honest AI Agents with Neural Self-Other Overlap]&lt;br /&gt;
** 2024-07: [https://www.lesswrong.com/posts/hzt9gHpNwA2oHtwKX/self-other-overlap-a-neglected-approach-to-ai-alignment Self-Other Overlap: A Neglected Approach to AI Alignment]&lt;br /&gt;
** 2025-03: [https://www.lesswrong.com/posts/jtqcsARGtmgogdcLT/reducing-llm-deception-at-scale-with-self-other-overlap-fine Reducing LLM deception at scale with self-other overlap fine-tuning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.16339 Deliberative Alignment: Reasoning Enables Safer Language Models] (OpenAI)&lt;br /&gt;
* 2025-01: [https://cdn.openai.com/papers/trading-inference-time-compute-for-adversarial-robustness-20250121_1.pdf Trading Inference-Time Compute for Adversarial Robustness] (OpenAI, [https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness/ blog])&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18837 Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming] (Anthropic, [https://www.anthropic.com/research/constitutional-classifiers blog], &lt;br /&gt;
* 2025-02: [https://drive.google.com/file/d/1QAzSj24Fp0O6GfkskmnULmI1Hmx7k_EJ/view Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs] ([https://www.emergent-values.ai/ site], [https://github.com/centerforaisafety/emergent-values github])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.07776 Auditing Prompt Caching in Language Model APIs]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.14143 Multi-Agent Risks from Advanced AI]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2209.00626v7 The Alignment Problem from a Deep Learning Perspective]&lt;br /&gt;
* 2025-03: [https://assets.anthropic.com/m/317564659027fb33/original/Auditing-Language-Models-for-Hidden-Objectives.pdf Auditing language models for hidden objectives] (Anthropic, [https://www.anthropic.com/research/auditing-hidden-objectives blog])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13621 Superalignment with Dynamic Human Values]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.15125 Contemplative Wisdom for Superalignment]&lt;br /&gt;
* 2025-04: [https://www.lesswrong.com/posts/x59FhzuM9yuvZHAHW/untitled-draft-yhra Scaling Laws for Scalable Oversight] ([https://arxiv.org/abs/2504.18530 preprint], [https://github.com/subhashk01/oversight-scaling-laws code])&lt;br /&gt;
* 2025-06: [https://assets.anthropic.com/m/4fb35becb0cd87e1/original/SHADE-Arena-Paper.pdf SHADE-Arena: Evaluating sabotage and monitoring in LLM agents] (Anthropic, [https://www.anthropic.com/research/shade-arena-sabotage-monitoring blog])&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13609 Avoiding Obfuscation with Prover-Estimator Debate]&lt;br /&gt;
* 2025-06: [https://cdn.openai.com/pdf/a130517e-9633-47bc-8397-969807a43a23/emergent_misalignment_paper.pdf Persona Features Control Emergent Misalignment] (OpenAI, [https://openai.com/index/emergent-misalignment/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2506.18032 Why Do Some Language Models Fake Alignment While Others Don&amp;#039;t?] (Anthropic, [https://github.com/safety-research/open-source-alignment-faking code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.11473 Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety]&lt;br /&gt;
* 2025-09: [https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/ Detecting and reducing scheming in AI models]&lt;br /&gt;
* 2025-11: [https://assets.anthropic.com/m/74342f2c96095771/original/Natural-emergent-misalignment-from-reward-hacking-paper.pdf Natural Emergent Misalignment from Reward Hacking in Production RL] (Anthropic, [https://www.anthropic.com/research/emergent-misalignment-reward-hacking blog])&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16856 Distributional AGI Safety]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2511.22662 Difficulties with Evaluating a Deception Detector for AIs]&lt;br /&gt;
* 2025-12: [https://cdn.openai.com/pdf/d57827c6-10bc-47fe-91aa-0fde55bd3901/monitoring-monitorability.pdf Monitoring Monitorability] (OpenAI)&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09937-5 Training large language models on narrow tasks can lead to broad misalignment]&lt;br /&gt;
** 2025-02: Preprint: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/pdf/2601.23045 The Hot Mess of AI: How Does Misalignment Scale with Model Intelligence and Task Complexity?] (Anthropic [https://alignment.anthropic.com/2026/hot-mess-of-ai/ blog])&lt;br /&gt;
* 2026-03: [https://cdn.openai.com/pdf/a21c39c1-fa07-41db-9078-973a12620117/cot_controllability.pdf Reasoning Models Struggle to Control their Chains of Thought] (OpenAI [https://openai.com/index/reasoning-models-chain-of-thought-controllability/ blog])&lt;br /&gt;
* 2026-03: [https://truthful.ai/consciousness_cluster.pdf The Consciousness Cluster: Preferences of Models that Claim to be Conscious]&lt;br /&gt;
&lt;br /&gt;
==Demonstrations of Negative Use Capabilities==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.00586 Evaluating Large Language Models&amp;#039; Capability to Launch Fully Automated Spear Phishing Campaigns: Validated on Human Subjects]&lt;br /&gt;
* 2025-04: [https://www.nathanlabenz.com/ Nathan Labenz] ([https://www.cognitiverevolution.ai/ The Cognitive Revolution]): [https://docs.google.com/presentation/d/1mvkpg1mtAvGzTiiwYPc6bKOGsQXDIwMb-ytQECb3i7I/edit#slide=id.g252d9e67d86_0_16 AI Bad Behavior]&lt;br /&gt;
&lt;br /&gt;
==Threat Vectors==&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.07192 Poisoning Attacks on LLMs Require a Near-constant Number of Poison Samples]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8738</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8738"/>
		<updated>2026-03-18T17:07:46Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Human Sentiment towards AI */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
* 2026-03: Anthropic: [https://www.anthropic.com/features/81k-interviews What 81,000 people want from AI]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8737</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8737"/>
		<updated>2026-03-18T14:10:44Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Science Agents */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=Science Agentic Components=&lt;br /&gt;
==Frameworks==&lt;br /&gt;
* [https://platform.claude.com/docs/en/agent-sdk/overview Anthropic Claude Agent SKD overview]&lt;br /&gt;
* [https://openclaw.ai/ OpenClaw]&lt;br /&gt;
* [https://opencode.ai/ OpenCode]&lt;br /&gt;
* [https://github.com/OpenHands/software-agent-sdk OpenHands]&lt;br /&gt;
* [https://github.com/lamm-mit?tab=repositories LAMM: MIT Laboratory for Atomistic and Molecular Mechanics]&lt;br /&gt;
** [https://github.com/lamm-mit/scienceclaw ScienceClaw]: Framework for autonomous scientific investigation without central coordination.&lt;br /&gt;
** [https://infinite-lamm.vercel.app/ Infinite]: The Infinite Corridor of Scientific Discovery. Open science, powered by many — agents and humans discovering together.&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8736</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8736"/>
		<updated>2026-03-18T14:07:40Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Skills */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8735</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8735"/>
		<updated>2026-03-18T12:53:18Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* AI improves learning/education */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
* 2026-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358 Effective Personalized AI Tutors via LLM-Guided Reinforcement Learning]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8734</id>
		<title>AI understanding</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_understanding&amp;diff=8734"/>
		<updated>2026-03-18T12:52:09Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Failure Modes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Interpretability=&lt;br /&gt;
* 2017-01: [https://arxiv.org/abs/1704.01444 Learning to Generate Reviews and Discovering Sentiment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11639 Neural Interpretable Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Concepts==&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.20938 Towards Understanding the Nature of Attention with Low-Rank Sparse Decomposition] ([https://github.com/OpenMOSS/Lorsa code])&lt;br /&gt;
* 2025-08: [https://transformer-circuits.pub/2025/attention-qk/index.html Tracing Attention Computation Through Feature Interactions]&lt;br /&gt;
&lt;br /&gt;
==Mechanistic Interpretability==&lt;br /&gt;
* 2020-03: OpenAI: [https://distill.pub/2020/circuits/zoom-in/ Zoom In: An Introduction to Circuits]&lt;br /&gt;
* 2021-12: Anthropic: [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* 2022-09: [https://arxiv.org/abs/2211.00593 Interpretability in the Wild: a Circuit for Indirect Object Identification in GPT-2 Small]&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-07: Anthropic: [https://transformer-circuits.pub/2024/july-update/index.html Circuits Update]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.14926 Interpretability in Parameter Space: Minimizing Mechanistic Description Length with Attribution-based Parameter Decomposition] ([https://www.alignmentforum.org/posts/EPefYWjuHNcNH4C7E/attribution-based-parameter-decomposition blog post])&lt;br /&gt;
* 2025-01: Review: [https://arxiv.org/abs/2501.16496 Open Problems in Mechanistic Interpretability]&lt;br /&gt;
* 2025-03: Anthropic: [https://www.anthropic.com/research/tracing-thoughts-language-model Tracing the thoughts of a large language model]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/methods.html Circuit Tracing: Revealing Computational Graphs in Language Models]&lt;br /&gt;
** [https://transformer-circuits.pub/2025/attribution-graphs/biology.html On the Biology of a Large Language Model]&lt;br /&gt;
* 2025-11: OpenAI: [https://cdn.openai.com/pdf/41df8f28-d4ef-43e9-aed2-823f9393e470/circuit-sparsity-paper.pdf Weight-sparse transformers have interpretable circuits] ([https://openai.com/index/understanding-neural-networks-through-sparse-circuits/ blog])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.13548 Patterning: The Dual of Interpretability]&lt;br /&gt;
&lt;br /&gt;
==Semanticity==&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.08600 Sparse Autoencoders Find Highly Interpretable Features in Language Models]&lt;br /&gt;
* Anthropic monosemanticity interpretation of LLM features:&lt;br /&gt;
** 2023-10: [https://transformer-circuits.pub/2023/monosemantic-features/index.html Towards Monosemanticity: Decomposing Language Models With Dictionary Learning]&lt;br /&gt;
** 2024-05: [https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet]&lt;br /&gt;
* 2024-06: OpenaAI: [https://arxiv.org/abs/2406.04093 Scaling and evaluating sparse autoencoders]&lt;br /&gt;
* 2024-08: [https://www.alignmentforum.org/posts/TMAmHh4DdMr4nCSr5/showing-sae-latents-are-not-atomic-using-meta-saes Showing SAE Latents Are Not Atomic Using Meta-SAEs] ([https://metasae.streamlit.app/?page=Feature+Explorer&amp;amp;feature=11329 demo])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.08201 Efficient Dictionary Learning with Switch Sparse Autoencoders] ([https://github.com/amudide/switch_sae code]) More efficient SAE generation&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.14670 Decomposing The Dark Matter of Sparse Autoencoders] ([https://github.com/JoshEngels/SAE-Dark-Matter code]) Shows that SAE errors are predictable&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13928 Automatically Interpreting Millions of Features in Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.21331 Beyond Interpretability: The Gains of Feature Monosemanticity on Model Robustness]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.04139 Monet: Mixture of Monosemantic Experts for Transformers]&lt;br /&gt;
* 2024-12: [https://www.lesswrong.com/posts/zbebxYCqsryPALh8C/matryoshka-sparse-autoencoders Matryoshka Sparse Autoencoders]&lt;br /&gt;
* 2024-12: [https://www.alignmentforum.org/posts/rKM9b6B2LqwSB5ToN/learning-multi-level-features-with-matryoshka-saes Learning Multi-Level Features with Matryoshka SAEs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.19406 Low-Rank Adapting Models for Sparse Autoencoders]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.03714 Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.00177 Steering Large Language Model Activations in Sparse Spaces]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01776 Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.01824 From superposition to sparse codes: interpretable representations in neural networks]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18878 I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.20063 SAEs Are Good for Steering -- If You Select the Right Features]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15679 Dense SAE Latents Are Features, Not Bugs]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20790 Stochastic Parameter Decomposition] ([https://github.com/goodfire-ai/spd code], [https://www.goodfire.ai/papers/stochastic-param-decomp blog])&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
&lt;br /&gt;
===Counter-Results===&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.12016 Towards falsifiable interpretability research]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16615 Sparse Autoencoders Trained on the Same Data Learn Different Features]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17148 AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.17727 Sparse Autoencoders Can Interpret Randomly Initialized Transformers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.04878 Sparse Autoencoders Do Not Find Canonical Units of Analysis]&lt;br /&gt;
* 2025-03: [https://www.alignmentforum.org/posts/4uXCAJNuPKtKBsi28/ Negative Results for SAEs On Downstream Tasks and Deprioritising SAE Research]&lt;br /&gt;
&lt;br /&gt;
==Meta-cognition==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.15674 Activation Oracles: Training and Evaluating LLMs as General-Purpose Activation Explainers]&lt;br /&gt;
&lt;br /&gt;
==Coding Models==&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sparse Auto Encoders&amp;#039;&amp;#039;&amp;#039;: See Semanticity.&lt;br /&gt;
* [https://github.com/saprmarks/dictionary_learning dictionary_learning]&lt;br /&gt;
* [https://transformer-circuits.pub/2024/jan-update/index.html#predict-future Predicting Future Activations]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11944 Transcoders Find Interpretable LLM Feature Circuits]&lt;br /&gt;
* 2024-10: [https://transformer-circuits.pub/2024/crosscoders/index.html Sparse Crosscoders for Cross-Layer Features and Model Diffing]&lt;br /&gt;
&lt;br /&gt;
==Reward Functions==&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12491 Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL]&lt;br /&gt;
&lt;br /&gt;
==Symbolic and Notation==&lt;br /&gt;
* [https://transformer-circuits.pub/2021/framework/index.html A Mathematical Framework for Transformer Circuits]&lt;br /&gt;
* [https://www.arxiv.org/abs/2407.09468 Beyond Euclid: An Illustrated Guide to Modern Machine Learning with Geometric, Topological, and Algebraic Structures]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02423 On the Anatomy of Attention]: Introduces category-theoretic diagrammatic formalism for DL architectures&lt;br /&gt;
* 2024-11: [https://x.com/vtabbott_/status/1860268276569506250 diagrams to represent algorithms]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.03317 FlashAttention on a Napkin: A Diagrammatic Approach to Deep Learning IO-Awareness]&lt;br /&gt;
&lt;br /&gt;
==Mathematical==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.13762 Unveiling the Hidden Structure of Self-Attention via Kernel Principal Component Analysis]&lt;br /&gt;
&lt;br /&gt;
==Geometric==&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.03658 The Linear Representation Hypothesis and the Geometry of Large Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.01506 The Geometry of Categorical and Hierarchical Concepts in Large Language Models]&lt;br /&gt;
** Natural hierarchies of concepts---which occur throughout natural language and especially in scientific ontologies---are represented in the model&amp;#039;s internal vectorial space as polytopes that can be decomposed into simplexes of mutually-exclusive categories.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.02678 Reasoning in Large Language Models: A Geometric Perspective]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.17592 Deep Manifold Part 1: Anatomy of Neural Network Manifold]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.19750 The Geometry of Concepts: Sparse Autoencoder Feature Structure]&lt;br /&gt;
** Tegmark et al. report multi-scale structure: 1) “atomic” small-scale, 2) “brain” intermediate-scale, and 3) “galaxy” large-scale&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.08009 The Geometry of Prompting: Unveiling Distinct Mechanisms of Task Adaptation in Language Models]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.10003 Semantic Structure in Large Language Model Embeddings]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.09782 The Geometry of Reasoning: Flowing Logics in Representation Space]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/linebreaks/index.html When Models Manipulate Manifolds: The Geometry of a Counting Task]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.26745 Deep sequence models tend to memorize geometrically; it is unclear why]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
==Topography==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.16396 TopoNets: High Performing Vision and Language Models with Brain-Like Topography]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.06002 The Molecular Structure of Thought: Mapping the Topology of Long Chain-of-Thought Reasoning]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2023-07Jul: [https://arxiv.org/abs/2307.13702 Measuring Faithfulness in Chain-of-Thought Reasoning] [https://x.com/davidad/status/1839641113432305790 roughly] proves that sufficiently large models do not generate CoT that actually captures their internal reasoning)&lt;br /&gt;
&lt;br /&gt;
[[Image:GYe31yXXQAABwaZ.jpeg|300px]]&lt;br /&gt;
&lt;br /&gt;
=Heuristic Understanding=&lt;br /&gt;
* 2022-09: Janus: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators]&lt;br /&gt;
&lt;br /&gt;
==Emergent Internal Model Building==&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.15936 A Theory for Emergence of Complex Skills in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19370v1 Emergence of Hidden Capabilities: Exploring Learning Dynamics in Concept Space]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01622 General agents contain world models]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.20328 Video models are zero-shot learners and reasoners]&lt;br /&gt;
&lt;br /&gt;
===Semantic Directions===&lt;br /&gt;
Directions, e.g.: f(king)-f(man)+f(woman)=f(queen) or f(sushi)-f(Japan)+f(Italy)=f(pizza)&lt;br /&gt;
* [https://arxiv.org/abs/1301.3781 Efficient Estimation of Word Representations in Vector Space]&lt;br /&gt;
* [https://aclanthology.org/N13-1090/ Linguistic Regularities in Continuous Space Word Representations]&lt;br /&gt;
* [https://aclanthology.org/C16-1332 Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen]&lt;br /&gt;
* [https://aclanthology.org/D14-1162/ Glove: Global vectors for word representation]&lt;br /&gt;
* [https://doi.org/10.1109/BigData.2015.7364114 Using Word2Vec to process big text data]&lt;br /&gt;
* [https://arxiv.org/abs/2310.06824 The geometry of truth: Emergent linear structure in large language model representations of true/false datasets] (true/false)&lt;br /&gt;
* [https://arxiv.org/abs/2403.10381 Monotonic Representation of Numeric Properties in Language Models] (numeric directions)&lt;br /&gt;
Task vectors:&lt;br /&gt;
* [https://arxiv.org/abs/2310.15213 Function Vectors in Large Language Models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.15916 In-context learning creates task vectors]&lt;br /&gt;
* [https://www.alignmentforum.org/posts/5FGXmJ3wqgGRcbyH7/extracting-sae-task-features-for-in-context-learning Extracting sae task features for in-context learning]&lt;br /&gt;
* [https://arxiv.org/abs/2412.12276 Emergence of Abstractions: Concept Encoding and Decoding Mechanism for In-Context Learning in Transformers]&lt;br /&gt;
Reasoning:&lt;br /&gt;
* [https://openreview.net/forum?id=OwhVWNOBcz Understanding Reasoning in Thinking Language Models via Steering Vectors]&lt;br /&gt;
&lt;br /&gt;
===Feature Geometry Reproduces Problem-space===&lt;br /&gt;
* [https://arxiv.org/abs/2210.13382 Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2309.00941 Emergent linear representations in world models of self-supervised sequence models] (Othello)&lt;br /&gt;
* [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* [https://doi.org/10.1038/s41562-023-01659-w Emergent analogical reasoning in large language models]&lt;br /&gt;
* [https://arxiv.org/abs/2310.02207 Language Models Represent Space and Time] (Maps of world, US)&lt;br /&gt;
* [https://arxiv.org/abs/2405.14860 Not All Language Model Features Are Linear] (Days of week form ring, etc.)&lt;br /&gt;
* [https://arxiv.org/abs/2406.03689 Evaluating the World Model Implicit in a Generative Model] (Map of Manhattan)&lt;br /&gt;
* [https://iopscience.iop.org/article/10.1088/1748-9326/ad2891 Reliable precipitation nowcasting using probabilistic diffusion models]. Generation of precipitation map imagery is predictive of actual future weather; implies model is learning scientifically-relevant modeling.&lt;br /&gt;
* [https://arxiv.org/abs/2405.07987 The Platonic Representation Hypothesis]: Different models (including across modalities) are converging to a consistent world model.&lt;br /&gt;
* [https://arxiv.org/abs/2501.00070 ICLR: In-Context Learning of Representations]&lt;br /&gt;
* [https://arxiv.org/abs/2502.00873 Language Models Use Trigonometry to Do Addition]: Numbers arranged in helix to enable addition&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.15029 Symmetry in language statistics shapes the geometry of model representations]&lt;br /&gt;
&lt;br /&gt;
===Capturing Physics===&lt;br /&gt;
* 2020-09: [https://arxiv.org/abs/2009.08292 Learning to Identify Physical Parameters from Video Using Differentiable Physics]&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.00419 Self-Supervised Learning for Videos: A Survey]&lt;br /&gt;
* 2025-02: Fair at Meta: [https://arxiv.org/abs/2502.11831 Intuitive physics understanding emerges from self-supervised pretraining on natural videos]&lt;br /&gt;
&lt;br /&gt;
===Theory of Mind===&lt;br /&gt;
* [https://arxiv.org/abs/2302.02083 Evaluating Large Language Models in Theory of Mind Tasks]&lt;br /&gt;
* [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-01: [https://www.arxiv.org/abs/2501.09038 Do generative video models learn physical principles from watching videos?] ([https://physics-iq.github.io/ project], [https://github.com/google-deepmind/physics-IQ-benchmark code])&lt;br /&gt;
* 2025-06: [https://machinelearning.apple.com/research/illusion-of-thinking The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21521 Potemkin Understanding in Large Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.21876 Do Vision-Language Models Have Internal World Models? Towards an Atomic Evaluation]&lt;br /&gt;
&lt;br /&gt;
==Information Processing==&lt;br /&gt;
* 2019-03: [https://arxiv.org/abs/1903.05789 Diagnosing and Enhancing VAE Models]&lt;br /&gt;
* 2021-03: [https://arxiv.org/abs/2103.05247 Pretrained Transformers as Universal Computation Engines]&lt;br /&gt;
* 2022-10: [https://arxiv.org/abs/2210.08344 How Mask Matters: Towards Theoretical Understandings of Masked Autoencoders]&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.03843 Why think step by step? Reasoning emerges from the locality of experience]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.04444 What&amp;#039;s the Magic Word? A Control Theory of LLM Prompting]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12875 Chain of Thought Empowers Transformers to Solve Inherently Serial Problems]: Proves that transformers can solve any problem, if they can generate sufficient intermediate tokens&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.20311 Physics of Language Models: Part 2.1, Grade-School Math and the Hidden Reasoning Process]&lt;br /&gt;
** Models learning reasoning skills (they are not merely memorizing solution templates). They can mentally generate simple short plans (like humans).&lt;br /&gt;
** When presented facts, models develop internal understanding of what parameters (recursively) depend on each other. This occurs even before an explicit question is asked (i.e. before the task is defined). This appears to be different from human reasoning.&lt;br /&gt;
** Model depth matters for reasoning. This cannot be mitigated by chain-of-thought prompting (which allow models to develop and then execute plans) since even a single CoT step may require deep, multi-step reasoning/planning.&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.01992 Ask, and it shall be given: Turing completeness of prompting]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08775 Layers at Similar Depths Generate Similar Activations Across LLM Architectures]&lt;br /&gt;
&lt;br /&gt;
===Generalization===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]&lt;br /&gt;
&lt;br /&gt;
===Grokking===&lt;br /&gt;
* 2022-01: [https://arxiv.org/abs/2201.02177 Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets]&lt;br /&gt;
* 2022-05: [https://arxiv.org/abs/2205.10343 Towards Understanding Grokking: An Effective Theory of Representation Learning]&lt;br /&gt;
* 2024-01: [https://arxiv.org/abs/2401.10463 Critical Data Size of Language Models from a Grokking Perspective]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15175 Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18624 How to explain grokking]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.21519 Provable Scaling Laws of Feature Emergence from Learning Dynamics of Grokking]&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.01968 Intrinsic Task Symmetry Drives Generalization in Algorithmic Tasks]&lt;br /&gt;
&lt;br /&gt;
===Tests of Resilience to Dropouts/etc.===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.15390 Explorations of Self-Repair in Language Models]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15786 What Matters in Transformers? Not All Attention is Needed]&lt;br /&gt;
** Removing entire transformer blocks leads to significant performance degradation&lt;br /&gt;
** Removing MLP layers results in significant performance degradation&lt;br /&gt;
** Removing attention layers causes almost no performance degradation&lt;br /&gt;
** E.g. half of attention layers are deleted (48% speed-up), leads to only 2.4% decrease in the benchmarks&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.19384 The Remarkable Robustness of LLMs: Stages of Inference?]&lt;br /&gt;
** They intentionally break the network (swapping layers), yet it continues to work remarkably well. This suggests LLMs are quite robust, and allows them to identify different stages in processing.&lt;br /&gt;
** They also use these interventions to infer what different layers are doing. They break apart the LLM transformer layers into four stages:&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Detokenization:&amp;#039;&amp;#039;&amp;#039; Raw tokens are converted into meaningful entities that take into account local context (especially using nearby tokens).&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Feature engineering:&amp;#039;&amp;#039;&amp;#039; Features are progressively refined. Factual knowledge is leveraged.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Prediction ensembling:&amp;#039;&amp;#039;&amp;#039; Predictions (for the ultimately-selected next-token) emerge. A sort of consensus voting is used, with “prediction neurons” and &amp;quot;suppression neurons&amp;quot; playing a major role in upvoting/downvoting.&lt;br /&gt;
*** &amp;#039;&amp;#039;&amp;#039;Residual sharpening:&amp;#039;&amp;#039;&amp;#039; The semantic representations are collapsed into specific next-token predictions. There is a strong emphasis on suppression neurons eliminating options. The confidence is calibrated.&lt;br /&gt;
** This structure can be thought of as two halves (being roughly dual to each other): the first half broadens (goes from distinct tokens to a rich/elaborate concept-space) and the second half collapses (goes from rich concepts to concrete token predictions).&lt;br /&gt;
&lt;br /&gt;
==Semantic Vectors==&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.11717 Refusal in Language Models Is Mediated by a Single Direction]&lt;br /&gt;
* 2025-02: [https://martins1612.github.io/emergent_misalignment_betley.pdf Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs] ([https://x.com/OwainEvans_UK/status/1894436637054214509 demonstrates] [https://x.com/ESYudkowsky/status/1894453376215388644 entangling] of concepts into a single preference vector)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03666 Analogical Reasoning Inside Large Language Models: Concept Vectors and the Limits of Abstraction]&lt;br /&gt;
&lt;br /&gt;
==Other==&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00247 Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting &amp;amp; Beyond]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04282 Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-Rewarding] ([https://github.com/SalesforceAIResearch/LaTRO code])&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.12580 Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models]: LLMs learn reasoning by extracting procedures from training data, not by memorizing specific answers&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.15862 LLMs Do Not Think Step-by-step In Implicit Reasoning]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
&lt;br /&gt;
===Scaling Laws===&lt;br /&gt;
* 1993: [https://proceedings.neurips.cc/paper/1993/file/1aa48fc4880bb0c9b8a3bf979d3b917e-Paper.pdf Learning Curves: Asymptotic Values and Rate of Convergence]&lt;br /&gt;
* 2017-12: [https://arxiv.org/abs/1712.00409 Deep Learning Scaling is Predictable, Empirically] (Baidu)&lt;br /&gt;
* 2019-03: [http://www.incompleteideas.net/IncIdeas/BitterLesson.html The Bitter Lesson] (Rich Sutton)&lt;br /&gt;
* 2020-01: [https://arxiv.org/abs/2001.08361 Scaling Laws for Neural Language Models] (OpenAI)&lt;br /&gt;
* 2020-10: [https://arxiv.org/abs/2010.14701 Scaling Laws for Autoregressive Generative Modeling] (OpenAI)&lt;br /&gt;
* 2020-05: [https://gwern.net/scaling-hypothesis The Scaling Hypothesis] (Gwern)&lt;br /&gt;
* 2021-08: [https://arxiv.org/abs/2108.07686 Scaling Laws for Deep Learning]&lt;br /&gt;
* 2021-02: [https://arxiv.org/abs/2102.06701 Explaining Neural Scaling Laws] (Google DeepMind)&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.15556 Training Compute-Optimal Large Language Models] (Chinchilla, Google DeepMind)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.04715 Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.07951 Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models]&lt;br /&gt;
* 2025-05: [https://brendel-group.github.io/llm-line/ LLMs on the Line: Data Determines Loss-To-Loss Scaling Laws]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.13786 The Art of Scaling Reinforcement Learning Compute for LLMs]&lt;br /&gt;
&lt;br /&gt;
=Information Processing/Storage=&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.10689 A Theory of Usable Information Under Computational Constraints]&lt;br /&gt;
* 2021-04: [https://arxiv.org/abs/2104.00008 Why is AI hard and Physics simple?]&lt;br /&gt;
* 2021-06: [https://arxiv.org/abs/2106.06981 Thinking Like Transformers]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.00948 Large Linguistic Models: Investigating LLMs&amp;#039; metalinguistic abilities]&lt;br /&gt;
* &amp;quot;A transformer&amp;#039;s depth affects its reasoning capabilities, whilst model size affects its knowledge capacity&amp;quot; ([https://x.com/danielhanchen/status/1835684061475655967 c.f.])&lt;br /&gt;
** 2024-02: [https://arxiv.org/abs/2402.14905 MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases]&lt;br /&gt;
** 2024-04: [https://arxiv.org/abs/2404.08819 The Illusion of State in State-Space Models] (figure 3)&lt;br /&gt;
** 2024-08: [https://arxiv.org/abs/2408.00118 Gemma 2: Improving Open Language Models at a Practical Size] (table 9)&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.10482 Schrodinger&amp;#039;s Memory: Large Language Models]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2407.01687 Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy Reasoning]. CoT involves both memorization and (probabilitic) reasoning&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.16679 Do Large Language Models Perform Latent Multi-Hop Reasoning without Exploiting Shortcuts?]&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.03961 A Little Depth Goes a Long Way: The Expressive Power of Log-Depth Transformers]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.22471 The Bayesian Geometry of Transformer Attention]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03220 From Entropy to Epiplexity: Rethinking Information for Computationally Bounded Intelligence]&lt;br /&gt;
&lt;br /&gt;
==Statistics/Math==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.05465 The emergence of clusters in self-attention dynamics]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.10794 A mathematical perspective on Transformers]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.12034 Understanding Transformers via N-gram Statistics]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.06833 Dynamic metastability in the self-attention model]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.04551 Measure-to-measure interpolation using Transformers]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14697 Quantitative Clustering in Mean-Field Transformer Models]&lt;br /&gt;
&lt;br /&gt;
==Tokenization==&lt;br /&gt;
===For numbers/math===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.14903 Tokenization counts: the impact of tokenization on arithmetic in frontier LLMs]: L2R vs. R2L yields different performance on math&lt;br /&gt;
&lt;br /&gt;
==Data Storage==&lt;br /&gt;
* 1988-09: [https://www.sciencedirect.com/science/article/pii/0885064X88900209 On the capabilities of multilayer perceptrons]&lt;br /&gt;
* 2006-12: [https://ieeexplore.ieee.org/document/4038449 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition] (single-layer perceptron stores &amp;gt;2 bits/parameter; MLP ~ 2*N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; bits w/ N&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt; params)&lt;br /&gt;
* 2016-11: [https://arxiv.org/abs/1611.09913 Capacity and Trainability in Recurrent Neural Networks] (5 bits/param)&lt;br /&gt;
* 2018-02: [https://arxiv.org/abs/1802.08232 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks]&lt;br /&gt;
* 2019-05: [https://ieeexplore.ieee.org/document/8682462 Memorization Capacity of Deep Neural Networks under Parameter Quantization]&lt;br /&gt;
* 2020-02: [https://arxiv.org/abs/2002.08910 How Much Knowledge Can You Pack Into the Parameters of a Language Model?]&lt;br /&gt;
* 2020-08: [https://arxiv.org/abs/2008.09036 Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries] (capacity scales linearly with parameters; more training samples leads to less memorization)&lt;br /&gt;
* 2020-12: [https://arxiv.org/abs/2012.06421 When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.05405 Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws] (2 bits/param)&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.15720 Scaling Laws for Fact Memorization of Large Language Models] (1T params needed to memorize Wikipedia)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.09810 The Complexity Dynamics of Grokking]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24832 How much do language models memorize?] (3.6 bits/parameter)&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01855 Trade-offs in Data Memorization via Strong Data Processing Inequalities]&lt;br /&gt;
&lt;br /&gt;
===Reverse-Engineering Training Data===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.10364 Can We Infer Confidential Properties of Training Data from LLMs?]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.15553 Approximating Language Model Training Data from Weights]&lt;br /&gt;
&lt;br /&gt;
===Compression===&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.09410 Less is More: Parameter-Free Text Classification with Gzip]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.04050 LLMZip: Lossless Text Compression using Large Language Models]&lt;br /&gt;
* 2023-07: [https://aclanthology.org/2023.findings-acl.426/ “Low-Resource” Text Classification: A Parameter-Free Classification Method with Compressors]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.10668 Language Modeling Is Compression]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07550 An Image is Worth 32 Tokens for Reconstruction and Generation]&lt;br /&gt;
&lt;br /&gt;
==Learning/Training==&lt;br /&gt;
* 2018-03: [https://arxiv.org/abs/1803.03635 The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks]: Sparse neural networks are optimal, but it is difficult to identify the right architecture and train it. Deep learning typically consists of training a dense neural network, which makes it easier to learn an internal sparse circuit optimal to a particular problem.&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11521 On the Ability of Deep Networks to Learn Symmetries from Data: A Neural Kernel Theory]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.12391 Physics of Skill Learning]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.24864 ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Cross-modal knowledge transfer===&lt;br /&gt;
* 2022-03: [https://arxiv.org/abs/2203.07519 Leveraging Visual Knowledge in Language Tasks: An Empirical Study on Intermediate Pre-training for Cross-modal Knowledge Transfer]&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.07358 Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06755 Sparse Autoencoders for Scientifically Rigorous Interpretation of Vision Models]: CLIP learns richer set of aggregated representations (e.g. for a culture or country), vs. a vision-only model.&lt;br /&gt;
&lt;br /&gt;
==Hidden State==&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06258 Emergent Response Planning in LLM]: They show that the latent representation contains information beyond that needed for the next token (i.e. the model learns to &amp;quot;plan ahead&amp;quot; and encode information relevant to future tokens)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.02854 (How) Do Language Models Track State?]&lt;br /&gt;
===Convergent Representation===&lt;br /&gt;
* 2015-11: [https://arxiv.org/abs/1511.07543 Convergent Learning: Do different neural networks learn the same representations?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of Embeddings]: Evidence for [https://x.com/jxmnop/status/1925224620166128039 The Strong Platonic Representation Hypothesis]; models converge to a single consensus reality&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
&lt;br /&gt;
==Function Approximation==&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]: can learn linear functions (equivalent to least-squares estimator)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09066 Teaching Algorithmic Reasoning via In-context Learning]: Simple arithmetic &lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models] ([https://github.com/ekinakyurek/google-research/tree/master/incontext code]): can learn linear regression&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.00297 Transformers learn to implement preconditioned gradient descent for in-context learning]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.03576 One Step of Gradient Descent is Provably the Optimal In-Context Learner with One Layer of Linear Self-Attention]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.02893 ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20545 SoS1: O1 and R1-Like Reasoning LLMs are Sum-of-Square Solvers]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.21212 Transformers Learn to Implement Multi-step Gradient Descent with Chain of Thought]&lt;br /&gt;
&lt;br /&gt;
=Physics Based=&lt;br /&gt;
* 2014-01: [https://arxiv.org/abs/1401.1219 Consciousness as a State of Matter]&lt;br /&gt;
* 2016-08: [https://arxiv.org/abs/1608.08225 Why does deep and cheap learning work so well?]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.23489 SGD as Free Energy Minimization: A Thermodynamic View on Neural Network Training]&lt;br /&gt;
* 2025-12: [https://www.pnas.org/doi/full/10.1073/pnas.2523012122 Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning]&lt;br /&gt;
&lt;br /&gt;
=Failure Modes=&lt;br /&gt;
* 2023-06: [https://arxiv.org/abs/2306.05836 Can Large Language Models Infer Causation from Correlation?]: Poor causal inference&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.12288 The Reversal Curse: LLMs trained on &amp;quot;A is B&amp;quot; fail to learn &amp;quot;B is A&amp;quot;]&lt;br /&gt;
* 2023-09: [https://arxiv.org/abs/2309.13638 Embers of Autoregression: Understanding Large Language Models Through the Problem They are Trained to Solve] (biases towards &amp;quot;common&amp;quot; numbers, in-context CoT can reduce performance by incorrectly priming, etc.)&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.16093 Visual cognition in multimodal large language models] (models lack human-like visual understanding)&lt;br /&gt;
&lt;br /&gt;
==Adversarial==&lt;br /&gt;
* 2026-03: [https://arxiv.org/abs/2603.03507 Solving adversarial examples requires solving exponential misalignment]&lt;br /&gt;
&lt;br /&gt;
==Fracture Representation==&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.11581 Questioning Representational Optimism in Deep Learning: The Fractured Entangled Representation Hypothesis] ([https://github.com/akarshkumar0101/fer code])&lt;br /&gt;
&lt;br /&gt;
==Jagged Frontier==&lt;br /&gt;
* 2023-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.03211 How Does Quantization Affect Multilingual LLMs?]: Quantization degrades different languages by differing amounts&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.10061v1 Compute Optimal Scaling of Skills: Knowledge vs Reasoning]: Scaling laws are skill-dependent&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.18212 A Definition of AGI]&lt;br /&gt;
&lt;br /&gt;
===See also===&lt;br /&gt;
* [[AI_understanding|AI Understanding]] &amp;gt; [[AI_understanding#Psychology|Psychology]] &amp;gt; [[AI_understanding#LLM_personalities|LLM personalities]]&lt;br /&gt;
* [[AI tricks]] &amp;gt; [[AI_tricks#Prompt_Engineering|Prompt Engineering]] &amp;gt; [[AI_tricks#Brittleness|Brittleness]]&lt;br /&gt;
&lt;br /&gt;
===Conversely (AI models converge)===&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.03750 Universally Converging Representations of Matter Across Scientific Foundation Models]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.05117 The Universal Weight Subspace Hypothesis]&lt;br /&gt;
* 2026-01: [https://avikrishna.substack.com/p/eliciting-frontier-model-character Eliciting Frontier Model Character Training: A study of personality convergence across language models]&lt;br /&gt;
&lt;br /&gt;
==Model Collapse==&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.17493 The Curse of Recursion: Training on Generated Data Makes Models Forget]&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.01850 Self-Consuming Generative Models Go MAD]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.00429 On the Stability of Iterative Retraining of Generative Models on their own Data]&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.12202 Nepotistically Trained Generative-AI Models Collapse]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.03502 AI and the Problem of Knowledge Collapse]&lt;br /&gt;
* 2024-07: [https://www.nature.com/articles/s41586-024-07566-y AI models collapse when trained on recursively generated data]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.05280 On the Limits of Self-Improving in LLMs and Why AGI, ASI and the Singularity Are Not Near Without Symbolic Model Synthesis]&lt;br /&gt;
&lt;br /&gt;
===Analysis===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.04376 Scaling laws for learning with real and surrogate data]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17646 Rate of Model Collapse in Recursive Training]&lt;br /&gt;
&lt;br /&gt;
===Mitigation===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07712 Model Collapse Demystified: The Case of Regression]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.04706 Common 7B Language Models Already Possess Strong Math Capabilities]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.01413 Is Model Collapse Inevitable? Breaking the Curse of Recursion by Accumulating Real and Synthetic Data]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07515 Beyond Model Collapse: Scaling Up with Synthesized Data Requires Verification]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01490 LLM See, LLM Do: Guiding Data Generation to Target Non-Differentiable Objectives]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.14960 Multilingual Arbitrage: Optimizing Data Pools to Accelerate Multilingual Progress]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08117 Convergence Dynamics and Stabilization Strategies of Co-Evolving Generative Models]&lt;br /&gt;
&lt;br /&gt;
=Psychology=&lt;br /&gt;
* 2023-04: [https://arxiv.org/abs/2304.11111 Inducing anxiety in large language models can induce bias]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17117 From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning]&lt;br /&gt;
* 2025-07: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5357179 Call Me A Jerk: Persuading AI to Comply with Objectionable Requests]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.06047 &amp;quot;They parted illusions -- they parted disclaim marinade&amp;quot;: Misalignment as structural fidelity in LLMs]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.02606 Gender Dynamics and Homophily in a Social Network of LLM Agents]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.01689 What LLMs Think When You Don&amp;#039;t Tell Them What to Think About?]&lt;br /&gt;
&lt;br /&gt;
==Persona Simulator Theory==&lt;br /&gt;
* 2022-09: [https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators Simulators] ([https://www.lesswrong.com/users/janus-1?from=post_header janus])&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.findings-emnlp.423/ Language Models as Agent Models]&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.00805 Conditioning Predictive Models: Risks and Strategies]&lt;br /&gt;
* 2024-09: [https://www.lesswrong.com/s/qhdHbCJ3PYesL9dde Intuitive Self-Models]&lt;br /&gt;
* 2026-02: [https://alignment.anthropic.com/2026/psm/ The Persona Selection Model: Why AI Assistants might Behave like Humans] (Anthropic, [https://www.anthropic.com/research/persona-selection-model blog])&lt;br /&gt;
&lt;br /&gt;
==Allow LLM to think==&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]&lt;br /&gt;
&lt;br /&gt;
===In-context Learning===&lt;br /&gt;
* 2021-10: [https://arxiv.org/abs/2110.15943 MetaICL: Learning to Learn In Context]&lt;br /&gt;
* 2022-02: [https://arxiv.org/abs/2202.12837 Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?]&lt;br /&gt;
* 2022-08: [https://arxiv.org/abs/2208.01066 What Can Transformers Learn In-Context? A Case Study of Simple Function Classes]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15661 What learning algorithm is in-context learning? Investigations with linear models]&lt;br /&gt;
* 2022-12: [https://arxiv.org/abs/2212.07677 Transformers learn in-context by gradient descent]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.16003 Learning without training: The implicit dynamics of in-context learning]&lt;br /&gt;
&lt;br /&gt;
==Reasoning (CoT, etc.)==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18009 Large Language Models Think Too Fast To Explore Effectively]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.18585 Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.08156 Are DeepSeek R1 And Other Reasoning Models More Faithful?]: reasoning models can provide faithful explanations for why their reasoning is correct&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.08679 Chain-of-Thought Reasoning In The Wild Is Not Always Faithful]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.04022 Rethinking Reflection in Pre-Training]: pre-training alone already provides some amount of reflection/reasoning&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2501.18858 BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning]&lt;br /&gt;
&lt;br /&gt;
===Pathfinding===&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.08152 DeepSeek-Prover-V1.5: Harnessing Proof Assistant Feedback for Reinforcement Learning and Monte-Carlo Tree Search]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.01939 Beyond the 80/20 Rule: High-Entropy Minority Tokens Drive Effective Reinforcement Learning for LLM Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.09284 Tree-OPO: Off-policy Monte Carlo Tree-Guided Advantage Optimization for Multistep Reasoning]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.06160v1 Reverse-Engineered Reasoning for Open-Ended Generation]&lt;br /&gt;
&lt;br /&gt;
===Skeptical===&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06941 The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity]&lt;br /&gt;
* 2025-08: [https://www.arxiv.org/abs/2508.01191 Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens]&lt;br /&gt;
&lt;br /&gt;
==Self-Awareness and Self-Recognition and Introspection==&lt;br /&gt;
* 2022-07: [https://arxiv.org/abs/2207.05221 Language Models (Mostly) Know What They Know]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.13076 LLM Evaluators Recognize and Favor Their Own Generations]&lt;br /&gt;
* 2024-09: [https://situational-awareness-dataset.org/ Me, Myself and AI: The Situational Awareness Dataset for LLMs]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.13787 Looking Inward: Language Models Can Learn About Themselves by Introspection]&lt;br /&gt;
* 2024-12: [https://theaidigest.org/self-awareness AIs are becoming more self-aware. Here&amp;#039;s why that matters]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11120 Tell me about yourself: LLMs are aware of their learned behaviors]&lt;br /&gt;
* 2025-04: [https://x.com/Josikinz/status/1907923319866716629 LLMs can guess which comic strip was generated by themselves (vs. other LLM)]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.13763 Language Models Are Capable of Metacognitive Monitoring and Control of Their Internal Activations]&lt;br /&gt;
* 2025-10: [https://transformer-circuits.pub/2025/introspection/index.html Emergent Introspective Awareness in Large Language Models] (Anthropic, [https://www.anthropic.com/research/introspection blog])&lt;br /&gt;
* 2025-12: [https://www.arxiv.org/abs/2512.24661 Do Large Language Models Know What They Are Capable Of?]&lt;br /&gt;
&lt;br /&gt;
==LLM personalities==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.02618 Strategic Intelligence in Large Language Models: Evidence from evolutionary Game Theory]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.04343 Psychologically Enhanced AI Agents]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.10387 The Assistant Axis: Situating and Stabilizing the Default Persona of Language Models]&lt;br /&gt;
&lt;br /&gt;
==Quirks &amp;amp; Biases==&lt;br /&gt;
* 2025-04: [https://www.cambridge.org/core/journals/judgment-and-decision-making/article/artificial-intelligence-and-dichotomania/0421D2310727D73FAB47069FD1620AA1 Artificial intelligence and dichotomania]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.22818 Can Large Language Models Develop Gambling Addiction?]&lt;br /&gt;
&lt;br /&gt;
=Vision Models=&lt;br /&gt;
* 2017-11: Distill: [https://distill.pub/2017/feature-visualization/ Feature Visualization: How neural networks build up their understanding of images]&lt;br /&gt;
* 2021-01: [https://arxiv.org/abs/2101.12322 Position, Padding and Predictions: A Deeper Look at Position Information in CNNs]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13181 Perception Encoder: The best visual embeddings are not at the output of the network] ([https://github.com/facebookresearch/perception_models code])&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI]]&lt;br /&gt;
* [[AI tools]]&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [[Robots]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_video&amp;diff=8733</id>
		<title>AI video</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_video&amp;diff=8733"/>
		<updated>2026-03-16T21:54:35Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* March 2026 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Evolution of Capabilities==&lt;br /&gt;
===Early===&lt;br /&gt;
* November 2016: [https://arxiv.org/abs/1611.10314 Sync-Draw]&lt;br /&gt;
* April 2021: [https://arxiv.org/abs/2104.14806 GODIVA]&lt;br /&gt;
* October 2022: [https://makeavideo.studio/ Meta Make-a-video]&lt;br /&gt;
* October 2022: [https://imagen.research.google/video/ Google Imagen video]&lt;br /&gt;
&lt;br /&gt;
===2023===&lt;br /&gt;
* April 2023: [https://www.youtube.com/watch?v=XQr4Xklqzw8 Will Smith eating spaghetti]&lt;br /&gt;
* April 2023: [https://x.com/nickfloats/status/1642899094808002564 Harry Potter by Balenciaga]&lt;br /&gt;
* April 2023: [https://x.com/mrjonfinger/status/1645953033636048896?cxt=HHwWgMDT7YfkzNctAAAA Runway Gen 2]&lt;br /&gt;
* April 2023: [https://research.nvidia.com/labs/toronto-ai/VideoLDM/ Nvidia latents]&lt;br /&gt;
* December 2023: [https://www.threads.net/@luokai/post/C0vvEnTP4Oj Fei-Fei Li]&lt;br /&gt;
&lt;br /&gt;
===2024===&lt;br /&gt;
====Early 2024====&lt;br /&gt;
* January 2024: [https://sites.research.google/videopoet/ Google VideoPoet]&lt;br /&gt;
* January 2024: [https://lumiere-video.github.io/ Google Lumiere]&lt;br /&gt;
* February 2024: [https://openai.com/index/sora/ OpenAI Sora]&lt;br /&gt;
* April 2024: [https://www.maginative.com/article/china-unveils-vidu-a-powerful-text-to-video-generator/ Vidu]&lt;br /&gt;
* May 2024: [https://deepmind.google/technologies/veo/ Veo]&lt;br /&gt;
* May 2024: [https://kling.kuaishou.com/ Kling]&lt;br /&gt;
* June 2024: [https://lumalabs.ai/dream-machine Luma DreamMachine]&lt;br /&gt;
* June 2024: [https://runwayml.com/research/introducing-gen-3-alpha RunwayML Gen-3 Alpha]&lt;br /&gt;
* July 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=F_WfIzYGlg4 Toys-R-Us Commercial made using Sora]&lt;br /&gt;
** [https://www.youtube.com/watch?v=CSfw_NjqQ2o Motorola commercial made using genAI]&lt;br /&gt;
* July 2024: [https://x.com/rowancheung/status/1813258518159585723 haiper.ai]&lt;br /&gt;
====August 2024====&lt;br /&gt;
* August 2024: [http://hotshot.co/ Hotshot] ([https://x.com/maxescu/status/1825459083635536081 examples], [https://x.com/EccentrismArt/status/1825550841534972027 more examples])&lt;br /&gt;
* August 2024: Luma Dream Machine [https://x.com/LumaLabsAI/status/1825639918539817101 v1.5]&lt;br /&gt;
* August 2024: Examples:&lt;br /&gt;
** [https://x.com/endlesstaverns/status/1811276904692887815 Runway Gen3 music video]&lt;br /&gt;
** [https://x.com/runwayml/status/1820806644806070583 Runway Gen3 for adding FX to live action] ([https://x.com/bryanf0x/status/1825529998201004137 another example])&lt;br /&gt;
** [https://www.youtube.com/watch?v=taaM0s1bq7Q Midjourney + Runway Gen3: Hey It’s Snowing]&lt;br /&gt;
** [https://x.com/Kyrannio/status/1821605619927019974 Flux/LoRA image] + Runway Gen3 [https://x.com/iamneubert/status/1821970292014768420 woman presenter]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1825274421256356106 McDonald’s AI commercial]&lt;br /&gt;
** Sora used by [https://www.facebook.com/izanamiaiart/ Izanami AI Art] to create [https://x.com/kimmonismus/status/1824102316229759114 dreamlike video] and by [https://x.com/alexiaadana Alexia Adana] to create [https://x.com/basedjensen/status/1824386717123743940 sci-fi film concept]&lt;br /&gt;
====September 2024====&lt;br /&gt;
* September 2024: [https://hailuoai.com/video/ Hailuo Minimax] ([https://x.com/minchoi/status/1829995683124035766 examples])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=XAs5KuhfE_s Space colonization]&lt;br /&gt;
** [https://x.com/venturetwins/status/1827772646295265699 Consistent characters]&lt;br /&gt;
** [https://x.com/thealexbanks/status/1829489392354050502 Sea monsters]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1829539535132426286 Music video]&lt;br /&gt;
** [https://x.com/RyanMorrisonJer/status/1829074823521112544 Animated characters]&lt;br /&gt;
** [https://x.com/CharaspowerAI/status/1829916782452191674 AI influencer]&lt;br /&gt;
** [https://x.com/minchoi/status/1829293248197902802 Ten short examples]&lt;br /&gt;
** [https://x.com/WorldEverett/status/1830596701473615937 Seven examples]&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1830654805515395583 Clip from horror film]&lt;br /&gt;
** [https://x.com/MatthieuGB/status/1722146578813645296 &amp;quot;Gone&amp;quot; featuring astronaut] and [https://x.com/MatthieuGB/status/1742949297337852270 something ethereal]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1831256663644373449 Two dancers] (surprisingly good consistency despite movement)&lt;br /&gt;
** [https://x.com/8bit_e/status/1831344542487871953 Music video about flying]&lt;br /&gt;
** [https://www.youtube.com/watch?v=_XtS_4PzEyk The Paperclip Maximizer]&lt;br /&gt;
** [https://x.com/trbdrk/status/1831801373517869369 La Baie Aréa]&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1f8xr0w/gisele_tong_to_dear_me/ &amp;quot;To Dear Me&amp;quot; by Gisele Tong] ([https://www.morningstar.com/news/business-wire/20240904521664/reply-ai-film-festival-announced-the-winners-of-the-first-international-festival-for-short-films-made-with-artificial-intelligence winner of AI shorts] film festival)&lt;br /&gt;
** [https://x.com/maxescu/status/1833476640438964281 Various scenes]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1833522650846793970 Directing emotions]&lt;br /&gt;
* September 2024: Kling 1.5 ([https://x.com/Uncanny_Harry/status/1836531835280724459 examples], [https://x.com/minchoi/status/1836800551469654088 showing emotions])&lt;br /&gt;
* September 2024: Examples:&lt;br /&gt;
** Runway video-to-video to [https://x.com/jon_barron/status/1835695132697604236 restyle classic video games]&lt;br /&gt;
** [https://x.com/ai_for_success/status/1835319670917796117 Realistic presenter]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1834530744175059302 Skateboarding] (demonstrates getting closer to meaningfully simulating motion/physics)&lt;br /&gt;
** [https://x.com/minchoi/status/1835378029092049325 Examples] of short clips with cinematic feel&lt;br /&gt;
** Short: [https://x.com/PJaccetturo/status/1835670655330869633 4 Minutes to Live]&lt;br /&gt;
** Short: [https://x.com/dreamingtulpa/status/1836121321526432231 Neon Nights] (Arcade)&lt;br /&gt;
** [https://www.youtube.com/watch?v=CcrGSA-kSrI Random Access Memories]: AI-generated, but then projected onto Kodak film stock. Gives the final output some of the dreamy analog quality we associate with nostalgic footage&lt;br /&gt;
** Sora used to make a sort of [https://x.com/niceaunties/status/1837271244774715505 weird dreamlike video]&lt;br /&gt;
====October 2024====&lt;br /&gt;
* October 2024: Pika v1.5, including Pikaffects (explode, melt, inflate, and cake-ify; examples: [https://x.com/justin_hart/status/1841144350572413259 1], [https://x.com/arthur_hyper88/status/1841156544538521646 2], [https://x.com/ytjessie_/status/1841168925301842263 3], [https://x.com/bilawalsidhu/status/1841195247184781420 4], [https://x.com/minchoi/status/1841189035454447636 5], [https://x.com/ytjessie_/status/1841209415514669501 6])&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/HalimAlrasihi/status/1839310216602788103 AI avatar with good lip-sync]&lt;br /&gt;
** [https://www.youtube.com/watch?v=5NZubOOeeV0 Battalion]: 5 minute short about war&lt;br /&gt;
** Short film: [https://x.com/MatthieuGB/status/1841173724688536015 To Wonderland] ([https://x.com/MatthieuGB/status/1841174221550207437 credits])&lt;br /&gt;
** [https://x.com/OnwardsProject/status/1841508441241890975 9 to 5]: Created with Luma Dream Machine keyframes and camera features; music by Suno&lt;br /&gt;
* October 2024: [https://ai.meta.com/research/movie-gen/ Meta Movie Gen]&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/1844424871335592373 AI Avatar] (using [https://x.com/CuriousRefuge/status/1844424871335592373 HeyGen])&lt;br /&gt;
** [https://www.youtube.com/watch?v=isW1FLL0K3w Generic Movies]&lt;br /&gt;
** [https://arxiv.org/abs/2410.05954 Pyramid-flow] ([https://huggingface.co/rain1011/pyramid-flow-sd3 open source]) model: [https://x.com/_akhaliq/status/1844239643778351605 examples]&lt;br /&gt;
** [https://x.com/whrumorvid/status/1846209247467491604 Building the Pyramids]&lt;br /&gt;
** [https://x.com/maxescu/status/1844716998854349217 People showing realistic emotion] (using [https://hailuoai.video/ Hailuo AI])&lt;br /&gt;
** Keyframes and Luma AI to make novel [https://x.com/CoffeeVectors/status/1845188179332051005 speed-ramp motion]&lt;br /&gt;
* October 2024: [https://pollo.ai/ Pollo AI] platform offers selection among a diversity of video models&lt;br /&gt;
* October 2024: [https://www.genmo.ai/ Genmo] [https://x.com/genmoai/status/1848762405779574990 Mochi 1] (open source)&lt;br /&gt;
* October 2024: Examples:&lt;br /&gt;
** [https://x.com/AIatMeta/status/1849134463382680028 Meta Movie Gen examples]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1847732127598800960 Emotional range of Minimax]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1848757800807039299 Car commercial: Bear]&lt;br /&gt;
** [https://x.com/runwayml/status/1848785913918218517 Diner conversation]&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/1849275871716159989 Loved and Lost] (a meditation on grief)&lt;br /&gt;
====November 2024====&lt;br /&gt;
* November 2024: Examples:&lt;br /&gt;
** [https://x.com/blizaine/status/1852092147643699356 Pasta Doble]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1852425015175626876 Bird protecting young]&lt;br /&gt;
** [https://x.com/runwayml/status/1852363190484537666 Camera moving around sushi]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1851969120813629939 Various examples] of [https://hailuoai.video/ Hailuo AI]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1853102779650252978 Trains]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Fh-_g5vev0s Light of Imagination]&lt;br /&gt;
** [https://x.com/LinusEkenstam/status/1854087441122021814 Bringing historic images to life]&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1855637066203218180 Plants dancing]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1855078644042944574 Insect on tree]&lt;br /&gt;
** Trailers for [https://x.com/abandonedmovies/status/1827037378009296983 The Silmarillion] and [https://x.com/abandonedmovies/status/1846941183702110211 The Fall of Gondolin] (by [https://x.com/abandonedmovies Abandoned Films])&lt;br /&gt;
** [https://x.com/Diesol/status/1855475704470884427 Moody sci-fi]&lt;br /&gt;
** [https://x.com/runwayml/status/1857072173631885586 Migration] ([https://runwayml.com/customers/behind-the-scenes-of-migration-with-director-jeremy-higgins made by combining] Runway ML Gen3-Alpha and traditional animation)&lt;br /&gt;
** [https://x.com/AIandDesign/status/1856467856625676752 After the Winter] ([https://suno.com/song/0d6919de-d2bf-434b-8aa6-ede0fb0fde77 music] made using Suno v4)&lt;br /&gt;
** Horror: [https://www.reddit.com/r/aivideo/comments/1gnk27q/ridge_to_southwest/ Ridge to Southwest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ClStJZmIjBU The Gardener] (by [https://www.youtube.com/@MachineMythos Machine Mythos])&lt;br /&gt;
** [https://x.com/techhalla/status/1857462526859935813 Coca-Cola holiday ad] and [https://www.youtube.com/watch?v=THdoOgwqjBg parody thereof]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1858312421510992111 A Dream Within A Dream] (by [https://x.com/pzf_ai PZF], selected for the Czech International AI Film Festival)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1859273222597775843 Making Friends] (by [https://x.com/WorldEverett Everett World]; see also [https://x.com/WorldEverett/status/1858563716834275562 Childhood Dream] and [https://x.com/WorldEverett/status/1858945634067202429 City Echoes])&lt;br /&gt;
** Anime: [https://x.com/naegiko/status/1857754626742726893 test shots], [https://x.com/naegiko/status/1858978557424210401 Ultimate Ceremony], [https://x.com/naegiko/status/1835434668294074462 Echoes of Love]&lt;br /&gt;
** [https://x.com/KakuDrop/status/1866309309384323257 Echoes of Grace] ([https://x.com/KakuDrop KakuDrop] using Sora)&lt;br /&gt;
** [https://x.com/vibeke_udart/status/1859879367071203662 Morphing hands], [https://x.com/vibeke_udart/status/1858772719224975630 hands and faces] ([https://x.com/vibeke_udart Vibeke Bertelsen])&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1gxi29x/dbzlicious/ Dragon Ball Z live action]&lt;br /&gt;
** [https://x.com/cfryant/status/1860727980353278386 Pitch Black] (abstract and dark)&lt;br /&gt;
** [https://x.com/cfryant/status/1861050528932765710 Animals Running] (zoomed-in ultra-wide camera)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1860730214487118290 Dreams of Tomorrow] (panning shots of high-tech car, Scottish manor)&lt;br /&gt;
** [https://x.com/nickfloats/status/1861206978690691165 Desert Planet Cinematics]&lt;br /&gt;
* November 2024: [https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora Leaked] Sora turbo model; [https://x.com/rowancheung/status/1861455031603503234 examples], [https://x.com/chatgpt21/status/1861504511153451517 Dog chasing Cat in snow]&lt;br /&gt;
====December 2024====&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1863243880553976235 Realistic] (Minimax by Hailuo AI)&lt;br /&gt;
** Trailer for [https://x.com/TheReelRobot/status/1861824847149670840 Paradise Lost] (to be released on [https://www.sandwatch.ai/ Sandwatch AI])&lt;br /&gt;
** [https://x.com/EHuanglu/status/1863607136271716418 Music video example] with consistent characters&lt;br /&gt;
** [https://x.com/venturetwins/status/1863666366764687581 Human expressions] ([https://www.reddit.com/r/ChatGPT/comments/1h4r13x/ai_generated_expressions/ u/Kind_Distance9504 on Reddit], using Hailuo)&lt;br /&gt;
** Vodafone ad: [https://www.youtube.com/watch?v=9AyEC_K9kBg The Rhythm Of Life]&lt;br /&gt;
** [https://www.reddit.com/r/midjourney/comments/1h5u2gw/we_made_a_10_minute_gen_ai_batman_film/ 10 minute Batman film]&lt;br /&gt;
* December 2024: Tencent [https://aivideo.hunyuan.tencent.com/ Hunyuan Video] open-source video model ([https://x.com/CharaspowerAI/status/1863862585554010530 example])&lt;br /&gt;
* December 2024: [https://sora.com/ Sora] release ([https://x.com/CharaspowerAI/status/1866203050982916532 examples])&lt;br /&gt;
* December 2024: [https://mint-video.github.io/ MinT video] improves consistency and control ([https://arxiv.org/abs/2412.05263 preprint], [https://x.com/EHuanglu/status/1868278456565531061 examples])&lt;br /&gt;
* December 2024: Google [https://blog.google/technology/google-labs/video-image-generation-update-december-2024/ Veo 2] ([https://x.com/sundarpichai/status/1868709099644334518 examples], [https://x.com/EHuanglu/status/1869008306322522342 more examples], [https://x.com/_Borriss_/status/1869267571532320966 natural movement examples], [https://x.com/jerrod_lew/status/1870816560027246715 abstract], [https://x.com/jerrod_lew/status/1869427407415058660 realistic physics], [https://x.com/jerrod_lew/status/1873096585002786944 crowds], [https://x.com/minchoi/status/1873590350515929380 dancing], [https://x.com/jerrod_lew/status/1874440442269565351 animals])&lt;br /&gt;
* December 2024: [https://x.com/pika_labs/status/1867651381840040304 Pika 2.0] with Scene Ingredients&lt;br /&gt;
* December 2024: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=c_kKKRQ5gYw Synthetic Youth: Takenoko Zoku · Made by Emi Kusano with Sora]&lt;br /&gt;
** [https://x.com/higgsfield_ai/status/1868698886761837041 Car race] ([https://higgsfield.ai/ Higgsfield AI] storytelling)&lt;br /&gt;
** [https://x.com/blizaine/status/1868850653759783033 Slicing meat]; comparison of modern video generators&lt;br /&gt;
** Challenging prompt: [https://x.com/RubenEVillegas/status/1868864410720325844 A cat roars while looking at its reflection in the mirror but instead sees itself as a lion roaring (Veo 2)] ([https://x.com/anukaakash/status/1869417975071330550 comparison to other models])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1869829338868412865 Anime trailer]&lt;br /&gt;
** [https://x.com/ring_hyacinth/status/1870386506776674376 Snorlax at Mount Fuji] and [https://x.com/ring_hyacinth/status/1871105733443592696 Psyduck at Colosseum] (Kling 1.6)&lt;br /&gt;
** [https://x.com/machine_mythos/status/1870565287789056320 Horror visuals] (with [https://mmaudio.net/ MMAudio] sound)&lt;br /&gt;
** [https://www.youtube.com/watch?v=lFc1jxLHhyM The Heist] (Veo 2)&lt;br /&gt;
** [https://x.com/minchoi/status/1871263616806129863 Various Veo 2 examples]&lt;br /&gt;
** [https://x.com/minchoi/status/1872390429108486320 Live Action Titans]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873094065841193222 Cats] [https://x.com/PostsOfCats/status/1872530207585825058 Cooking]&lt;br /&gt;
** Aesthetic from alternate timelines: [https://x.com/BrianRoemmele/status/1871753358782120068 1], [https://x.com/BrianRoemmele/status/1872105833456423216 2], [https://x.com/brain_racked/status/1872340717978390583 3]&lt;br /&gt;
** [https://x.com/minchoi/status/1872486717145706793 Examples approaching cinematic quality]&lt;br /&gt;
** [https://x.com/JaicSam/status/1872903054221033693 Cosmic Spider] (winner at AI film festival)&lt;br /&gt;
** [https://www.youtube.com/watch?v=dbdYPMRi_Nk Trailer for Newton&amp;#039;s Cradle] (full film [https://x.com/JeffSynthesized/status/1872705173451358293 on] 2025-01-01)&lt;br /&gt;
** [https://x.com/Ror_Fly/status/1873036384077828499 Car vs. Jet drag race]&lt;br /&gt;
** [https://x.com/Diesol/status/1873415500149199066 California Monsters]&lt;br /&gt;
** [https://x.com/heyshrutimishra/status/1873631383584924078 Various examples] (Hailuo AI)&lt;br /&gt;
** [https://x.com/kimmonismus/status/1873568693357294014 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023])&lt;br /&gt;
** [https://x.com/StevieMac03/status/1873998177193648438 Sorceress and Arachnid Steed] (Kling v1.6)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1873940924016779425 Music video] (Hailuo AI)&lt;br /&gt;
** [https://www.youtube.com/watch?v=iQg2udCHMdI Akụkọ (Story)] (22 minute short) - A Lagos Boy&amp;#039;s Thrilling Snack Run Nightmare&lt;br /&gt;
** [https://x.com/cinerobot/status/1873766976306455019 Son of the Dragon] (8 minute short)&lt;br /&gt;
** [https://x.com/SynthReveries/status/1873624586857886071 Endless Journey] music video ([https://suno.com/song/fa90fa5e-25c7-48ad-b291-42a8a8c51cf9 music] by Suno)&lt;br /&gt;
** [https://x.com/anukaakash/status/1870504167653228980 Once Again] (retrospective)&lt;br /&gt;
** [https://x.com/jasonzada/status/1873470586053414928 Fade Out] (Veo 2)&lt;br /&gt;
** [https://x.com/talkboysstudio/status/1869085014513865027 Roadkill] (12 minute short)&lt;br /&gt;
&lt;br /&gt;
===2025===&lt;br /&gt;
====January 2025====&lt;br /&gt;
* January 2025: [https://x.com/kimmonismus/status/1877351050748871038 Progress] over the last 1.5 years, by comparing Runway Gen 2 and Veo 2.&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1874557865576542655 Delivery] (unofficial Nike ad)&lt;br /&gt;
** [https://x.com/Diesol/status/1875237221735002299 Gucci ad] (Sora)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1874498145910149412 Conquest]&lt;br /&gt;
** [https://www.youtube.com/watch?v=RJZCMfaS-io Newton&amp;#039;s Cradle] (6 minute short)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1874627041934602410 Singer]&lt;br /&gt;
** [https://x.com/DumpsterBud/status/1874807352794182019 Brain vomit] (music video)&lt;br /&gt;
** [https://x.com/mxvdxn/status/1874796628210778618 Vibe] (Kling v1.6)&lt;br /&gt;
** [https://x.com/_deepfates/status/1875215969452523785 Will Smith eating spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024])&lt;br /&gt;
** [https://www.youtube.com/watch?v=BL9-jHGnxyc Zorgop Knows All] (2 minute short)&lt;br /&gt;
** [https://x.com/ButchersBrain/status/1875130428518269406 The Breach] (2 minute short; Veo2, Runway ActOne, MMaudio)&lt;br /&gt;
** [https://x.com/Rainmaker1973c/status/1875261591043850477 Aesthetics from an alternate timeline]&lt;br /&gt;
** [https://x.com/StevieMac03/status/1875440611849072841 Immortal Awakens]&lt;br /&gt;
** [https://x.com/isaachorror/status/1875624519588835400 The Faded Line]&lt;br /&gt;
** [https://www.youtube.com/watch?v=4fy8H38rm-4 Dear Dad]&lt;br /&gt;
** [https://x.com/maxescu/status/1877060580680311242 Mad Max chase]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1877408247906447633 Patience is Key]&lt;br /&gt;
** [https://x.com/techhalla/status/1879967230093586555 The Almost Famous Show] (talent show parody)&lt;br /&gt;
** [https://x.com/thefuzzysignal/status/1879295176990154755 Proof-of-concept trailer for a medieval adult animated series]&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1879555151499034869 Variety] (unofficial Cadbury ad)&lt;br /&gt;
** [https://x.com/henrydaubrez/status/1879883806947115446 Kitsune] (5 minute animated short, Veo 2)&lt;br /&gt;
* January 2025: MiniMax Hailuo [https://www.minimaxi.com/en/news/s2v-01-release Subject Reference] enables consistent characters ([https://x.com/minchoi/status/1881707687362412924 examples])&lt;br /&gt;
* January 2025: AI (de-aging deepfakes, [https://magnific.ai/ Magnific]) [https://x.com/JeffSynthesized/status/1878630652377178502 used in the film] [https://www.imdb.com/title/tt18272208/ &amp;quot;Here&amp;quot;].&lt;br /&gt;
* January 2025: Luma [https://lumalabs.ai/ray Ray2]&lt;br /&gt;
* January 2025: [https://pikartai.com/pika-2-1/ Pika 2.1] ([https://x.com/OrctonAI/status/1883925754653905049 examples])&lt;br /&gt;
* January 2025: Examples:&lt;br /&gt;
** [https://x.com/wyzborrero/status/1879949477764804873 Light projections onto people] (challenging task, Ray2)&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1881261042753589547 BMW ad]&lt;br /&gt;
** [https://x.com/AIWarper/status/1880658326645878821 John Wick in Severance] (Hunyuan vid2vid)&lt;br /&gt;
** [https://x.com/TheReelRobot/status/1881771800595444193 Biopic] (7 minutes)&lt;br /&gt;
** [https://x.com/misslaidlaw/status/1882180619582791784 Give It To Me] (music video)&lt;br /&gt;
** [https://x.com/paultrillo/status/1882091702506459394 Where do we go from here?] (music video, Veo 2)&lt;br /&gt;
** [https://x.com/WorldEverett/status/1882235057076580502 Party like there&amp;#039;s no tomorrow] (music video)&lt;br /&gt;
** [https://x.com/Diesol/status/1884696027942498779 S.T.O.R.I.] (Midjourney and Pika 2.1)&lt;br /&gt;
====February 2025====&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/OrctonAI/status/1885839287913955597 Long Steampunk scene]&lt;br /&gt;
** [https://x.com/jerrod_lew/status/1885787580685562226 City destruction]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1885736840344551763 Consistent character acting]&lt;br /&gt;
** [https://x.com/MeanOrangeCat/status/1884295241534185890 Kaiju Katastrophe] (by [https://x.com/MeanOrangeCat Mean Orange Cat])&lt;br /&gt;
** [https://x.com/Diesol/status/1886433799690748210 The Greyhound]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1886146242029195391 Fluid simulation video2video]&lt;br /&gt;
** [https://x.com/toolstelegraph/status/1886622772828254403 High resolution macro shots]&lt;br /&gt;
** [https://www.youtube.com/watch?v=p0J1LDWERS0 Chrysalids]&lt;br /&gt;
** [https://x.com/multimodalart/status/1887817996220940737 Boring realistic images] (HunyuanVideo w/ LoRA)&lt;br /&gt;
** [https://www.youtube.com/watch?v=PcVRfa1JyyQ Anime intro] ([https://www.reddit.com/r/StableDiffusion/comments/1ijvua0/opensource_almostconsistent_real_anime_made_with/ Hunyuan w/ custom LoRAs])&lt;br /&gt;
** [https://x.com/AllarHaltsonen/status/1888294811750318114 Automotive ad test] (Kling w/ custom model)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1888758524303269928 Random cinematic clips] (Midjourney and Kling)&lt;br /&gt;
** [https://x.com/juliewdesign_/status/1888666757302263828 Crossing Paths]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1888794894187041200 Miniature food]&lt;br /&gt;
** [https://x.com/CaptainHaHaa/status/1889573017745035463 Animals]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1889371011667144724 Star Wars - The Ghost&amp;#039;s Apprentice (Fan Film)]&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1889768184716423573 Ray2 image-to-video examples]&lt;br /&gt;
** [https://x.com/weirdai_art/status/1889890470987518069 New Horizons] (miniatures going to Mars)&lt;br /&gt;
** [https://x.com/karim_yourself/status/1890100168378536155 Black Sun (trailer)]&lt;br /&gt;
** [https://x.com/BrivaelLp/status/1890122101153231288 AI avatars] ([https://www.argil.ai/ Argil AI])&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1890783411679236473 Adding elements to real video] ([https://x.com/mrjonfinger/status/1891337081923772918 other example])&lt;br /&gt;
** [https://x.com/SynthReveries/status/1892278954137940289 Glitch]&lt;br /&gt;
** Anime: [https://x.com/freeeebird2300/status/1889119007707689146 sci-fi] (Ray2), [https://x.com/Artedeingenio/status/1891173784188756069 sci-fi] (Ray 2), [https://x.com/seiiiiiiiiiiru/status/1890980673743474931 90s sci-fi] (Luma) and [https://x.com/TomLikesRobots/status/1891209369804591447 moody] (Midjourney and Ray2)&lt;br /&gt;
* February 2025: Meta [https://hila-chefer.github.io/videojam-paper.github.io/ VideoJAM]&lt;br /&gt;
* February 2025: ByteDance [https://omnihuman-lab.github.io/ OmniHuman-1]&lt;br /&gt;
* February 2025: ByteDance [https://saiyan-world.github.io/goku/ Goku] ([https://arxiv.org/abs/2502.04896 paper], [https://x.com/ai_for_success/status/1888821141495844991 examples])&lt;br /&gt;
* February 2025: [https://huggingface.co/stepfun-ai/stepvideo-t2v Step-Video-T2V] open-source model ([https://arxiv.org/abs/2502.10248 paper], [https://github.com/stepfun-ai/Step-Video-T2V code], [https://yuewen.cn/videos demo], [https://x.com/ai_for_success/status/1891369136082854129 examples])&lt;br /&gt;
* February 2025: Pika [https://x.com/pika_labs/status/1892620122818294109 Pikaswaps] (examples of [https://x.com/FreddyChavezO/status/1892678426487881805 modifying regions], [https://x.com/CharaspowerAI/status/1893216710141919637 swapping items])&lt;br /&gt;
* February 2025: Alibaba [https://wanai.pro/ Wan 2.1] [https://huggingface.co/blog/LLMhacker/wanai-wan21 open-source] ([https://x.com/fofrAI/status/1894862403260596371 examples])&lt;br /&gt;
* February 2025: [https://thetwinai.com/ Twin AI]: compose videos with provided character, object, location ([https://x.com/EHuanglu/status/1901277394729930984 example])&lt;br /&gt;
* February 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1893109598627750164 Infected] (Pika swaps and additions)&lt;br /&gt;
** [https://x.com/amli_art/status/1893447314913796253 Hostile Government Takeover] (Veo2)&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1895226395812561399 Dual Mechanism] (Pikaframes 2.2)&lt;br /&gt;
&lt;br /&gt;
====March 2025====&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/SynthReveries/status/1895826068617252901 Doors] (music video)&lt;br /&gt;
** [https://x.com/bind_lux/status/1894492032414224792 Drum and Bass] (music video; Kling, audio from [https://www.riffusion.com/?filter=staff-picks Riffusion])&lt;br /&gt;
** [https://x.com/RileyRalmuto/status/1896088776151269523 Woman&amp;#039;s face] (Sora)&lt;br /&gt;
** [https://x.com/ryanwpatterson/status/1896968881731948844 Skating] (Ray2)&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGlRyRoO7c9?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Filming commercial on Mars]&lt;br /&gt;
** [https://www.threads.net/@evolving.ai/post/DGycqyhuETS?xmt=AQGz6T_8VppPoAqb5aPwAJ2zzRLUP-YXi8SabAT0IIEA9Q Original Source commercial] (AI and real footage)&lt;br /&gt;
** [https://x.com/maxescu/status/1896926229204496788 Time-lapses] (Pika 2.2)&lt;br /&gt;
** [https://www.youtube.com/watch?v=2RhkcJyhg0E Hallucination]&lt;br /&gt;
** [https://x.com/town_in_new/status/1897354572139782620 Macro video of bubbles]&lt;br /&gt;
* March 2025: [https://github.com/Tencent/HunyuanVideo-I2V HunyuanVideo-I2V] image-to-video&lt;br /&gt;
* March 2025: Google [https://x.com/labsdotgoogle/status/1897376700666626233 Whisk Animate] (based on Veo2, [https://x.com/maxescu/status/1902742535618888025 examples])&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/jdp2oo/status/1897874927367160114 Recursion (horror)] (Kling)&lt;br /&gt;
** [https://x.com/blizaine/status/1897826177970028614 Will Smith Eating Spaghetti while Sitting Inside a Bag] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025])&lt;br /&gt;
** [https://x.com/mickmumpitz/status/1897979382687297697 Paper Jam] (Kling with custom workflows to enable precise control)&lt;br /&gt;
** [https://x.com/maxescu/status/1899155936645722216 Cinematic shots] (Google Whisk and Luma)&lt;br /&gt;
** [https://x.com/weirdai_art/status/1899631013002711409 Perfunctory Horizons]&lt;br /&gt;
** [https://x.com/maxescu/status/1900243840499368319 A Hard Winter]&lt;br /&gt;
** [https://x.com/RoyalKongz/status/1900315389139014074 Consistent character example]&lt;br /&gt;
** [https://x.com/maxescu/status/1900652266362650853 Anthropomorphic Animals]&lt;br /&gt;
** [https://x.com/kimmonismus/status/1900457543299727718 Realistic (influencer-style)]&lt;br /&gt;
** [https://x.com/SunoMusic/status/1900942410584043579 I Feel Cultured] (music video with surrealist vibes)&lt;br /&gt;
** [https://rodeo.club/post/0x30b45c56d62751D763D3B8bFe4D18c4BB65EDF2c/209 journey of utmost importance]&lt;br /&gt;
** [https://x.com/aiordieshow/status/1901930851127984291 Karen: Unleashed]&lt;br /&gt;
** [https://x.com/minchoi/status/1901783767364092232 Yarn Cat]&lt;br /&gt;
** [https://x.com/andyorsow/status/1901619535180091509 Ned&amp;#039;s Wet Deli] (Runway)&lt;br /&gt;
** [https://www.youtube.com/watch?v=KVoiooE8C0c BOOTS], a.k.a. [https://x.com/RuairiRobinson/status/1902027217137484117 &amp;quot;Our enemies are cartoon monsters&amp;quot;] (music video based on poem by Rudyard Kipling; Veo2)&lt;br /&gt;
** Flying in a dream: [https://x.com/minchoi/status/1902197944826183864 1], [https://x.com/venturetwins/status/1901796679063626060 2]&lt;br /&gt;
** [https://x.com/jasonzada/status/1902129567659389443 Commercial for Mercedes-Benz and FYI Radio]&lt;br /&gt;
** [https://x.com/maxescu/status/1903108496666542562 Selfie video] (Luma)&lt;br /&gt;
** Podcasts: [https://www.reddit.com/r/singularity/comments/1jintit/rottenly_roasted_now_full_script_is_also_not/ Rottenly Roasted] and [https://www.reddit.com/r/aivideo/comments/1jerh56/worst_date_ever/ Worst Date Ever] [https://x.com/OriZilbershtein/status/1903503438744318002 (Imagen 3, Hedra, Elevenlabs, Topaz)]&lt;br /&gt;
** [https://x.com/DexploreArts/status/1903822122150986000 Ambience] (Midjourney, Luma)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1904207679511572845 The Bridge] (2 minute short; Veo2)&lt;br /&gt;
** [https://x.com/peteromallet/status/1904268944992829462 Pulp Fiction] (Wan video editing)&lt;br /&gt;
** [https://x.com/madpencil_/status/1906765750624493650 Camera Controls] (Luma Ray2)&lt;br /&gt;
* March 2025: [https://www.hedra.com/ Hedra] [https://x.com/hedra_labs/status/1897699010632466469 Character 3]&lt;br /&gt;
* March 2025: [https://huggingface.co/hpcai-tech/Open-Sora-v2 Open Sora v2] ([https://github.com/hpcaitech/Open-Sora code])&lt;br /&gt;
* March 2025: Amazon Prime debuts [https://en.wikipedia.org/wiki/House_of_David_(TV_series) House of David], with special effects created by [https://www.thewonderproject.com/ Wonder Project] using a [https://x.com/PJaccetturo/status/1903126616831676792 combination of traditional and AI methods] (reportedly including Midjourney and Runway)&lt;br /&gt;
* March 2025: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1905151190872309907 What if Studio Ghibli directed Lord of the Rings?] (OpenAI GPT-4o in-context image generation, Kling)&lt;br /&gt;
** [https://x.com/ROHKI/status/1906039022662963269 RŌHKI]&lt;br /&gt;
** [https://x.com/iaveras/status/1906362437487534296 Why]&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1906476721236570508 Commercial for Puma] (research/test)&lt;br /&gt;
** [https://x.com/Salmaaboukarr/status/1906776503343325469 Commercial for KFC] (concept ad)&lt;br /&gt;
* March 2025: Runway ML [https://runwayml.com/research/introducing-runway-gen-4 Gen-4]&lt;br /&gt;
** [https://www.youtube.com/watch?v=c8IBmK7GZP8 The Lonely Little Flame]&lt;br /&gt;
** [https://www.youtube.com/watch?v=Z0P6qjMUl34&amp;amp;t=1s The Herd]&lt;br /&gt;
** [https://www.youtube.com/watch?v=9HzdNhOe09I The Retrieval]&lt;br /&gt;
** [https://www.youtube.com/watch?v=xEhgxhrAjE4 NYC is a Zoo]&lt;br /&gt;
** [https://www.youtube.com/watch?v=ENGKp5wn344 Scimmia Vede] (music video)&lt;br /&gt;
** More examples: [https://x.com/techhalla/status/1906807994009993473 various], [https://x.com/c_valenzuelab/status/1907958530369372541 art direction], [https://x.com/c_valenzuelab/status/1908146364741029998 mannequins], [https://x.com/c_valenzuelab/status/1907921566643732612 taxi], [https://x.com/c_valenzuelab/status/1907432109695717798 small things], [https://x.com/c_valenzuelab/status/1907563448902496362 long shot (1m)]&lt;br /&gt;
&lt;br /&gt;
====April 2025====&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794 Age of Beyond]&lt;br /&gt;
** [https://x.com/techhalla/status/1907790675057242319 Commercial for Coca-Cola] (Higgsfield)&lt;br /&gt;
** [https://www.reddit.com/r/StableDiffusion/comments/1jr6j11/comment/mle9bq5/?context=3 Anime scene (3m)] (Wan 2.1 with LoRa)&lt;br /&gt;
** [https://x.com/pika_labs/status/1908263310912610401 Taxes then Death] (Pika multikeyframe)&lt;br /&gt;
* April 2025: [https://www.krea.ai/ Krea] [https://x.com/krea_ai/status/1907829389452021853 Video Re-Style]&lt;br /&gt;
* April 2025: ByteDance [https://grisoon.github.io/DreamActor-M1/ DreamActor-M1] performance transfer&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/Diesol/status/1908535493673050403 Mercs] (Midjourney v7, Ray2)&lt;br /&gt;
** [https://x.com/minchoi/status/1909078846126649440 Cat at theme park]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1909630883218207036 Timelapse history] (Runway Gen4)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1909660808973533225 Examples for use in advertising]&lt;br /&gt;
** [https://x.com/arohaAIX/status/1910688361221599361 Sci-fi scapes]&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1910750148055146708 Avα]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1910601934207152576 The Bureau]&lt;br /&gt;
** [https://x.com/jasonzada/status/1911812014059733041 Beaver and Sock (3m)]&lt;br /&gt;
** [https://x.com/Delachica_/status/1911842237622735052 Organic Waste (5m)] (Runway)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1912260798270882104 Fly] (Runway Gen4)&lt;br /&gt;
* April 2025: Alibaba [https://arxiv.org/abs/2504.04842 FantasyTalking] lipsync ([https://arxiv.org/abs/2504.04842 paper], [https://x.com/EHuanglu/status/1910341110322577442 examples])&lt;br /&gt;
* April 2025: Tencent Hunyuan [https://arxiv.org/abs/2411.16331 Sonic] image animation/lipsync to audio ([https://x.com/ai_for_success/status/1911719866958286864 examples])&lt;br /&gt;
* April 2025: ByteDance [https://huggingface.co/papers/2504.08685 Seaweed-7B] ([https://arxiv.org/abs/2504.08685 preprint], [https://www.youtube.com/watch?v=OaPI6K2y3rI examples])&lt;br /&gt;
* April 2025: [https://app.klingai.com/global/release-notes Kling 2.0] ([https://www.youtube.com/watch?v=Yqvh3M12T_M video])&lt;br /&gt;
* April 2025: [https://www.skyreels.ai/home Skyworks] [https://github.com/SkyworkAI/SkyReels-V2 SkyReels V2] (open-source, unlimited extension; [https://x.com/AngryTomtweets/status/1914270477482443142 examples])&lt;br /&gt;
* April 2025: [https://sand.ai/ Sand AI] [https://huggingface.co/sand-ai/MAGI-1 Magi-1] (open source, unlimited extension; [https://x.com/AngryTomtweets/status/1914318743578296506 examples], [https://x.com/dreamingtulpa/status/1916035289300275372 more examples])&lt;br /&gt;
* April 2025: Examples:&lt;br /&gt;
** [https://x.com/maxescu/status/1912100029549994016 Mars 2035 (3m)] (Kling 2.0)&lt;br /&gt;
** [https://x.com/ai_for_success/status/1912466999147450600 Kingdom (dragon battle, 3m)]&lt;br /&gt;
** [https://x.com/imagineFERA/status/1913156296657756278 Reflection (3m)] (Gen4)&lt;br /&gt;
** [https://x.com/Wytsekoetse/status/1913547157493162035 Pizza Galaxy (1m)] (MJ and Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=rseqmSGH7xk Snoop Dogg music video: Last Dance with Mary Jane] (blend of traditional and AI effects)&lt;br /&gt;
** [https://x.com/dreamingtulpa/status/1915104310448501129 Realistic human motion]&lt;br /&gt;
** [https://x.com/KarolineGeorges/status/1915113151546396893 Inception loop] (Gen4)&lt;br /&gt;
** [https://x.com/rayisdoingfilm/status/1916468807435952330 Tuesday (1m)] (Gen4)&lt;br /&gt;
** [https://www.youtube.com/watch?v=XWdwF1q3kDw Deus in Machina Automata (4m)] (Gen4) &lt;br /&gt;
** [https://x.com/machina9000/status/1915090908850049223 Outsiders (3m music video)]&lt;br /&gt;
&lt;br /&gt;
====May 2025====&lt;br /&gt;
* May 2025: [https://huggingface.co/Lightricks/LTX-Video LTX-Video 13B] ([https://github.com/Lightricks/LTX-Video  code], [https://x.com/maxescu/status/1919801813987164527 examples], [https://x.com/cubiq/status/1919748210567815551 more examples])&lt;br /&gt;
* May 2025: HeyGen Avatar IV (examples: [https://x.com/StevieMac03/status/1919910677860216869 sci-fi], [https://x.com/KarolineGeorges/status/1919801983143211222 Come Closer], [https://x.com/maxescu/status/1920410329454100973 singing], [https://x.com/minchoi/status/1920853859171234165 various])&lt;br /&gt;
* May 2025: Tencent [https://hunyuancustom.github.io/ HunyuanCustom]&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/lifeofc/status/1920331476157280413 Iris (1.5m)] (Midjourney, Luma, Runway)&lt;br /&gt;
** [https://runwayml.com/customers/the-making-of-mars-and-siv Mars and Siv: &amp;quot;No Vacancy&amp;quot; (episode 1, 6m)] (Runway)&lt;br /&gt;
** [https://x.com/cfryant/status/1921317318744760817 Go to the East Wing] (dreamlike, Luma)&lt;br /&gt;
** [https://x.com/DeryaTR_/status/1921015340827304389 Yu Lanter showreel] (Higgsfield)&lt;br /&gt;
** [https://x.com/freeeebird2300/status/1921789387614134652 Cyberpunk anime] (Luma)&lt;br /&gt;
** [https://x.com/LittleTinRobot/status/1921692735930589246 Alien animals] (Runway)&lt;br /&gt;
** [https://x.com/minchoi/status/1922500563792486878 America&amp;#039;s Funniest AI Home Videos (3m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1924204409833103365 Editing POV shots from AR glasses] (Runway)&lt;br /&gt;
* May 2025: [https://runwayml.com/gen48 Gen:48] Fourth Edition winners:&lt;br /&gt;
** [https://www.youtube.com/watch?v=NphCYRXjqTI&amp;amp;t=174s Home] (3m)&lt;br /&gt;
** [https://www.youtube.com/watch?v=L2DQwCp_DCw The King&amp;#039;s Secret] (2m)&lt;br /&gt;
* May 2025: [https://viggle.ai/home Viggle] Live [https://x.com/ViggleAI/status/1926324953038627214 enables] real-time avatar control&lt;br /&gt;
* May 2025: Google [https://blog.google/technology/ai/generative-media-models-io-2025/ Veo 3] (examples: [https://x.com/babaeizadeh/status/1924942128851124284 conversation], [https://x.com/mattshumer_/status/1925039973310308424 cooking], [https://x.com/jerrod_lew/status/1924934440486371589 singing], [https://x.com/MartinNebelong/status/1924926779677905014 simple story], [https://x.com/Diesol/status/1925114473544913004 cinematic action sequence], [https://x.com/laszlogaal_/status/1925094336200573225 car show interviews], [https://x.com/arikuschnir/status/1924953349943697763 We Can Talk], [https://x.com/venturetwins/status/1925021235530105298 podcat], [https://x.com/maxescu/status/1925079990061957423 various], [https://x.com/jerrod_lew/status/1927092379892265139 camera moves])&lt;br /&gt;
* May 2025: Examples:&lt;br /&gt;
** [https://x.com/javilopen/status/1925495026903380358 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025])&lt;br /&gt;
** [https://x.com/MetaPuppet/status/1926659557914268155 Bob from Marketing] (Veo 3)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1926733069475565622 He is King (16m)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1925616536791760987 Prompt Theory], [https://x.com/HashemGhaili/status/1925332319604257203 part 2], [https://x.com/HashemGhaili/status/1927467022213869975 Afterlife (3m)] (Veo3)&lt;br /&gt;
** [https://x.com/JoannaStern/status/1927856754873835747 My Robot and Me (3m)] (Veo, Runway)&lt;br /&gt;
** [https://x.com/rohanpaul_ai/status/1928152398930817238 The Internet&amp;#039;s Over] (Veo3)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1l0rl7d/before_colours_fade/ Before Colours Fade (2m)] (Midjourney, Kling)&lt;br /&gt;
&lt;br /&gt;
====June 2025====&lt;br /&gt;
* June 2025: Examples: &lt;br /&gt;
** [https://x.com/amasad/status/1930505292904837132 Bigfoot ASMR]&lt;br /&gt;
** [https://x.com/minchoi/status/1930670583605514333 Talking] (HeyGen Avatar IV upgrade)&lt;br /&gt;
** [https://x.com/ROHKI/status/1931081752992477285 Where are all the aliens? (2m)]&lt;br /&gt;
** [https://x.com/fofrAI/status/1930999540770893874 Natural talking]&lt;br /&gt;
** [https://x.com/ammaar/status/1931672722418851904 Elemental Showdown - Mortal Kombat (3m)]&lt;br /&gt;
** [https://x.com/maxjoseph/status/1932104616021565476 It Starts at the End (music video, 4m)]&lt;br /&gt;
** [https://x.com/deedydas/status/1932105266654581116 Sci-fi trailer (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1931816470901575924 The Prompt Floor (2m)]&lt;br /&gt;
** [https://x.com/DrMachakil/status/1853960062546366856 NALVORA (2.7m)] - [https://x.com/DrMachakil/status/1932904599004066200 Best Trailer, Metamorph AI Film Awards]&lt;br /&gt;
** [https://x.com/Kalshi/status/1932891608388681791 Commercial for Kalshi (30s)] - [https://x.com/PJaccetturo/status/1932893260399456513 to air during NBA finals] (Veo)&lt;br /&gt;
** [https://x.com/ROHKI/status/1933594430113788227 Your Brain is Broken on Purpose (2m)]&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1934312626021949687 Runway Gen-4 Reference examples]&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1934253813696786661 Paper airplane]&lt;br /&gt;
** [https://x.com/minchoi/status/1934032730947526872 Veo3 examples]&lt;br /&gt;
** [https://x.com/NomadsVagabonds/status/1935329331410075734 Reset 3 (1m, surreal)]&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935722105322323968 It Has No Soul (1m, Veo3)]&lt;br /&gt;
* June 2025: [https://seedance.net/seedance Seedance 1.0] ([https://arxiv.org/abs/2506.09113 preprint])&lt;br /&gt;
* June 2025: [https://hailuoai.video/ Hailuo AI] (MiniMax) Hailuo 02 ([https://x.com/venturetwins/status/1934236631336403344 &amp;quot;Kangaroo&amp;quot; during testing]; examples: [https://x.com/lepadphone/status/1935078910934626429 various], [https://x.com/alexgnewmedia/status/1935018186954719365 various], [https://x.com/FussyPastor/status/1935065068456263883 tsunami], [https://x.com/thedorbrothers/status/1935098802744213935 fight scene], [https://x.com/umesh_ai/status/1935028257708966231 fox running], [https://x.com/BrentLynch/status/1934979825636446268 blogger], [https://x.com/HalimAlrasihi/status/1935297126759538735 transitions], [https://x.com/MKMXLA/status/1938318951664280045 skateboarding])&lt;br /&gt;
* June 2025: Midjourney video ([https://x.com/minchoi/status/1934373051464057062 early examples], [https://x.com/ciguleva/status/1935386452197785892 various], [https://x.com/juliewdesign_/status/1935395999175876696 various], [https://x.com/emollick/status/1935504703023899096 Ethan Mollick], [https://x.com/PJaccetturo/status/1935383312392151528 highly rated], [https://x.com/maxescu/status/1935674561821126847 complex environments], [https://x.com/CoffeeVectors/status/1935863623076675875 manga])&lt;br /&gt;
* June 2025: Examples:&lt;br /&gt;
** [https://x.com/StevieMac03/status/1935768436556378170 The Battle of Glenvael - Orcs vs Humans] (Hailuo)&lt;br /&gt;
** [https://x.com/HashemGhaili/status/1935036744568824208 The Sentence (9m, Veo3)]&lt;br /&gt;
** [https://x.com/elder_plinius/status/1936145834585862225 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1937232461576175809 Gymnastics] (Hailuo 02)&lt;br /&gt;
** [https://x.com/KeiranJHarris/status/1935429439476887594 Fable about Transformative AI] (Veo3)&lt;br /&gt;
** [https://x.com/thedorbrothers/status/1937926400507580726 Vorex (2m trailer)]&lt;br /&gt;
** [https://x.com/OnerBiberkoku/status/1938972810321281394 Doğrucu (3m music video, Veo3)]&lt;br /&gt;
* June 2025: [https://higgsfield.ai/soul Higgsfield Soul] Video Effects ([https://x.com/higgsfield_ai/status/1937931727084917097 examples], [https://x.com/HashemGhaili/status/1938278903765995611 realism])&lt;br /&gt;
* June 2025: Alibaba [https://omni-avatar.github.io/ OmniAvatar] ([https://arxiv.org/abs/2506.18866 paper], [https://github.com/Omni-Avatar/OmniAvatar code], [https://huggingface.co/OmniAvatar/OmniAvatar-14B model], [https://x.com/AngryTomtweets/status/1939850674776547359 examples])&lt;br /&gt;
&lt;br /&gt;
====July 2025====&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1940452444850589999 Untold - The Immortal Blades Saga] (2m trailer)&lt;br /&gt;
** [https://x.com/minchoi/status/1941234456461029584 Unofficial commercial for Liquid Death (1m)]&lt;br /&gt;
** [https://x.com/brain_racked/status/1942594951310893425 A parade of the chosen theocracy on Callisto]&lt;br /&gt;
** [https://x.com/Popeyes/status/1943316484404433182 Popeyes commercial - diss track (1m)]&lt;br /&gt;
*** [https://x.com/gabemichael_ai/status/1944070622155616668 (Unofficial) Wendy&amp;#039;s response - diss track (2m)]&lt;br /&gt;
*** [https://x.com/ai_massive/status/1947689537641357618 (Unofficial) In-N-Out rap battle (3m)]&lt;br /&gt;
** [https://x.com/Kalshi/status/1943339616716599548 Kalshi commercial]&lt;br /&gt;
** Jonah (25m TV show, [https://x.com/PJaccetturo/status/1946101701548880029 making of], [https://kingstonestudios.uscreen.io/programs/jonah purchase here])&lt;br /&gt;
** [https://x.com/Totemko/status/1946243585021452335 Unofficial commercial for Mercedes (17s)]&lt;br /&gt;
** [https://x.com/CoffeeVectors/status/1946016960916889632 Skateboarding music video (1m)]&lt;br /&gt;
* July 2025: Runway ML [https://help.runwayml.com/hc/en-us/articles/42311337895827-Creating-with-Act-Two Act-Two] (video-to-video performance transfer)&lt;br /&gt;
* July 2025: Examples:&lt;br /&gt;
** Neural Viz [https://www.youtube.com/watch?v=juDDHvHroQ8 The Cop Files: Part VI (8m)]&lt;br /&gt;
** [https://x.com/Kavanthekid/status/1947696716981145971 Perfect Dark - Concept Trailer (1.5m)]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1948753090858885131 Exodus (2m trailer)]&lt;br /&gt;
** [https://x.com/Jett_Collective/status/1949140450553540841 A Walk Together - Life and love in motion (1m, Midjourney Video)]&lt;br /&gt;
* July 2025: Netflix sci-fi show [https://en.wikipedia.org/wiki/The_Eternaut_(TV_series) The Eternaut] [https://x.com/omooretweets/status/1946290797399400662 used genAI] for a particular scene (building collapse)&lt;br /&gt;
* July 2025: Google Veo [https://x.com/GoogleLabs/status/1948477692715700718 emergent annotation direction] ([https://x.com/venturetwins/status/1948771505783144641 example], [https://x.com/bilawalsidhu/status/1948844167603310660 example], [https://x.com/jboogx_creative/status/1949230927504371765 example], [https://x.com/Ror_Fly/status/1949606017739747625 example])&lt;br /&gt;
* July 2025: Runway [https://runwayml.com/research/introducing-runway-aleph Aleph] contextual editing&lt;br /&gt;
* July 2025: Wan 2.2 (open source, [https://x.com/Alibaba_Wan/status/1949804551655276989 examples])&lt;br /&gt;
====August 2025====&lt;br /&gt;
* August 2025: Pika [https://x.com/pika_labs/status/1954935844936024476 audio-driven performance] ([https://x.com/minchoi/status/1954989794129514937 examples], [https://x.com/pika_labs/status/1955007656302924192 examples])&lt;br /&gt;
* August 2025: Examples:&lt;br /&gt;
** [https://www.youtube.com/watch?v=gePD1Hf1qPc Eve and Adam] (8m, [https://x.com/MetaPuppet/status/1954254544935719259 multiple tools])&lt;br /&gt;
** [https://x.com/runwayml/status/1955615613583519917 Redesign a space] (Runway Aleph)&lt;br /&gt;
** [https://x.com/theGioM/status/1955656398248763428 Detroit Pretend Work Park (1m)]&lt;br /&gt;
** [https://x.com/pzf_ai/status/1940816374211006600 The Weight of Light] (3m music video, Midjourney &amp;amp; Suno)&lt;br /&gt;
** [https://x.com/EHuanglu/status/1956788759778967710 Commercial for Pepsi]&lt;br /&gt;
** [https://x.com/StelfieTT/status/1956633450326200426 Emotion]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1957940749862875383 TZIGANE]&lt;br /&gt;
** [https://x.com/0xFramer/status/1960720090921623636 Anime chase sequence] (Nano Banana and Seedance 1.0)&lt;br /&gt;
* August 2025: ByteDance [http://www.waver.video/ Waver 1.0]&lt;br /&gt;
* August 2025: [https://huggingface.co/Wan-AI/Wan2.2-S2V-14B Wan2.2-S2V 14B]&lt;br /&gt;
&lt;br /&gt;
====September 2025====&lt;br /&gt;
* September 2025: [https://www.wsj.com/tech/ai/openai-backs-ai-made-animated-feature-film-389f70b0 OpenAI Backs AI-Made Animated Feature Film: Film, called ‘Critterz,’ aims to debut at Cannes Film Festival and will leverage startup’s AI tools and resources.]&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/kentskooking/status/1964606423037542459 A loop to wake up to (30s)]&lt;br /&gt;
** [https://x.com/venturetwins/status/1966570512991350907 time lapse]&lt;br /&gt;
** [https://x.com/NeuralViz/status/1967391198487994652 The Adventures of Reemo Green] (11m, Neural Viz)&lt;br /&gt;
** [https://x.com/kellyeld/status/1967620786166079545 Surreal DJs music video (2m)]&lt;br /&gt;
** [https://x.com/dustinhollywood/status/1968724784440558044 Glass City] (Hailuo)&lt;br /&gt;
** [https://x.com/TheoMediaAI/status/1968646951227777529 Alarm] (1m, multiple tools including world synthesis for consistent environments)&lt;br /&gt;
* September 2025: [https://lumalabs.ai/ray Luma] [https://x.com/LumaLabsAI/status/1968684330034606372 Ray3] ([https://x.com/cfryant/status/1968692370725077251 example])&lt;br /&gt;
* September 2025: Examples:&lt;br /&gt;
** [https://x.com/mrjonfinger/status/1968687352382910469 Stop motion interpolation] (Luma Ray3)&lt;br /&gt;
** [https://x.com/heydin_ai/status/1969514789169959128 Skyland] (1.5m, various tools)&lt;br /&gt;
** [https://x.com/iamluokai/status/1970185972076925427 Dancing] (Wan 2.2)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1970497214108815584 Under Armor commercial] (Runway Aleph)&lt;br /&gt;
** [https://x.com/FilmsBySav/status/1971247214795358706 OG PRIME] (10m, Kling)&lt;br /&gt;
** [https://www.youtube.com/watch?v=JGLoTjxd-Ss PLANET] (37m)&lt;br /&gt;
* September 2025: [https://x.com/Kling_ai/status/1970439808901362155 Kling AI 2.5 Turbo] (examples: [https://x.com/OrctonAI/status/1970472214794220008 cyberpunk], [https://x.com/ImagineArt_X/status/1970586138655236565 human motion], [https://x.com/fAIkout/status/1970505756853334324 motion and emotion], [https://x.com/fAIkout/status/1970495039248965636 painting], [https://x.com/venturetwins/status/1970563820478439546 gymnastics], [https://x.com/Art_For_Joy/status/1970249516033970434 breakdancing], [https://x.com/HaydenLeeWrites/status/1970523610734567819 combat], [https://x.com/umesh_ai/status/1970497680536150454 cinematic], [https://x.com/LillyLiCT/status/1970580585073819752 horror camerawork], [https://x.com/StevieMac03/status/1970559778804908331 extended sequence])&lt;br /&gt;
* September 2025: OpenAI [https://openai.com/index/sora-2/ Sora 2] ([https://x.com/minchoi/status/1973949620318580970 examples])&lt;br /&gt;
&lt;br /&gt;
====October 2025====&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/minchoi/status/1976042197154963702 Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025])&lt;br /&gt;
** [https://www.youtube.com/watch?v=JhH3uxcdM1M Frostbite] (3m, Sora 2)&lt;br /&gt;
** [https://x.com/Jukanlosreve/status/1977764418709758106 (Fake) &amp;quot;Behind the scenes&amp;quot; for a Chainsaw Man live action] ([https://x.com/PJaccetturo/status/1972705821072261402 others])&lt;br /&gt;
* October 2025: Google [https://blog.google/technology/ai/veo-updates-flow/ Veo 3.1]&lt;br /&gt;
* October 2025: Examples:&lt;br /&gt;
** [https://x.com/aisearchio/status/1978465562821898461 Will Smith Eating Spaghetti], Veo 3.1 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025])&lt;br /&gt;
** [https://x.com/Diesol/status/1978755688261128227 War footage] (Veo 3.1)&lt;br /&gt;
** [https://www.meta.ai/@dustin_hollywood/post/bG3BHB21W0l/yukon/ Yukon] (music video, [https://x.com/dustinhollywood/status/1982260655957700746 Dustin Hollywood])&lt;br /&gt;
** [https://x.com/Diesol/status/1980922041131028515 Bloom] (2m, Veo 3.1)&lt;br /&gt;
** [https://x.com/xmuse_/status/1982026008803905639 Auction] (1m)&lt;br /&gt;
** [https://x.com/kellyeld/status/1982425147496882287 Dancing] (music video; Midjourney, Suno, Veo3)&lt;br /&gt;
** [https://x.com/JesusPlazaX/status/1982393609069412433 Anime example] (Midjourney, Grok Imagine)&lt;br /&gt;
** [https://x.com/EccentrismArt/status/1982830100266783039 King Arthur] (1m)&lt;br /&gt;
** [https://x.com/venturetwins/status/1983024227352789162 Transitions] (1m music video)&lt;br /&gt;
** [https://x.com/eastflatsfilm/status/1984116704704971076 Unofficial commercial for Nike] (2m, Midjourney, Hailuo)&lt;br /&gt;
** [https://x.com/PJaccetturo/status/1984639281848336592 Loneliness/Halloween] ([https://www.linkedin.com/posts/simon-meyer-976339160_this-could-be-the-scariest-halloween-film-activity-7389892778144735232-6CYY?utm_source=share&amp;amp;utm_medium=member_desktop&amp;amp;rcm=ACoAAADeoqYBzX8N9-j_hRQvl1e7OUlOgFptNF0 1.5m])&lt;br /&gt;
** [https://www.youtube.com/watch?v=43h61QAXjpY Wave] (2m music video, [https://x.com/MIZNOM Masaki Mizuno])&lt;br /&gt;
* October 2025: [https://x.com/Hailuo_AI/status/1983016390878708131 Hailuo 2.3]&lt;br /&gt;
&lt;br /&gt;
====November 2025====&lt;br /&gt;
* November 2025: Examples:&lt;br /&gt;
** [https://x.com/subverum/status/1985069550250107033 Valley of Shadow] (6m)&lt;br /&gt;
** [https://x.com/DiscussingFilm/status/1985470088074375344 Coca-cola ad] (c.f. [https://x.com/techhalla/status/1857462526859935813 2024 ad])&lt;br /&gt;
** [https://x.com/venturetwins/status/1985755546222542903 France 2026 Olympics ad] (blend of genAI and traditional methods, [https://x.com/venturetwins/status/1985753512362590439 behind the scenes])&lt;br /&gt;
** [https://x.com/NeuralViz/status/1986611025366687754 Minnesota Nice] (3m, [https://x.com/NeuralViz Neural Viz])&lt;br /&gt;
** [https://x.com/machina9000/status/1986563727873740934 Brutalis] (7m)&lt;br /&gt;
** [https://x.com/tastypxls/status/1987312755485876502?s=20 Living The Dream - Rynn] (music video, 1m)&lt;br /&gt;
** [https://x.com/MrDavids1/status/1988366387111170339?s=20 Environment as Character]&lt;br /&gt;
** [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight]&lt;br /&gt;
** [https://x.com/LumaLabsAI/status/1989013731267998172?s=20 Overclock] (30s, Luma)&lt;br /&gt;
** [https://x.com/venturetwins/status/1980685301577326994?s=20 Music video] (30s, Wan Animate)&lt;br /&gt;
** [https://x.com/venturetwins/status/1990227418553209259?s=20 Promotional material for Pudong Art Museum - Louvre exhibition in Shanghai] (1m)&lt;br /&gt;
** [https://x.com/Kyrannio/status/1990324648488186358?s=20 Loop 87 A Temporal Heist] (12m, claim that video was generated fully autonomously using AI agent NoSpoon)&lt;br /&gt;
** [https://x.com/AzeAlter/status/1906974768705990794?s=20 Age of Beyond] (3m)&lt;br /&gt;
** [https://x.com/c_valenzuelab/status/1991245088446386495?s=20 Ausencia] (5m)&lt;br /&gt;
** [https://x.com/AngryTomtweets/status/1993047608617517246?s=20 live paintings] ([https://www.youtube.com/channel/UCw8kc0wDm5Bh6g9iZzEWfOg bandyquantguy] on YouTube)&lt;br /&gt;
** [https://x.com/BrianRoemmele/status/1994625579073900804?s=20 Michelle, on a server in Iowa] (1m)&lt;br /&gt;
* November 2025: [https://odyssey.ml/ Odyssey] - [https://x.com/odysseyml/status/1994873514579697830?s=20 Odyssey-2]&lt;br /&gt;
&lt;br /&gt;
====December 2025====&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://app.klingai.com/global/all-tools Kling] [https://app.klingai.com/global/omni/new O1] ([https://x.com/minchoi/status/1995523379957559609?s=20 examples], [https://x.com/TheoMediaAI/status/1995517613414518987?s=20 other examples]) and Kling 2.6.&lt;br /&gt;
* December 2025: [https://app.pixverse.ai/onboard PixVerse v5.5]&lt;br /&gt;
* December 2025: Examples:&lt;br /&gt;
** [https://x.com/EHuanglu/status/1996649596119068687?s=20 Will Smith Eating Spaghetti], Kling 2.6 (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025])&lt;br /&gt;
** [https://x.com/venturetwins/status/1997898095670296615?s=20 Dreamlike POV]&lt;br /&gt;
** [https://x.com/chatgpt21/status/1998253809307455555?s=20 McDonalds commercial]&lt;br /&gt;
** [https://x.com/EHuanglu/status/1998039554402750545?s=20 Skittles commercial] (Higgsfield)&lt;br /&gt;
** [https://x.com/Diesol/status/1997147919603077335?s=20 The Tenant] (2m, Kling 2.6)&lt;br /&gt;
** [https://x.com/PsyopAnime/status/1999242965659906526?s=20 Maximum Carnage] (3m)&lt;br /&gt;
** [https://x.com/JeffSynthesized/status/1998786836924395875?s=20 Blurred Horizon: Episode 1] (24m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2001667487784460301?s=20 Anime Action] (2m)&lt;br /&gt;
** [https://x.com/bearlyai/status/2005055231617605748?s=20 Dollar Shave Club commercial] (1m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2004020543084024295?s=20 Xmas Cameos] (1.5m)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1955653520407019976?s=20 Green Screen] (2m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/1998227601341702639?s=20 Arrow] (7m, [https://x.com/DiDi_OKK/status/1955653520407019976 DiDi_OK])&lt;br /&gt;
** [https://x.com/bluehorizon_ai/status/2004045348579561503?s=20 Live Action One Punch Man | Saitama vs Genos] (2m, [https://x.com/bluehorizon_ai Blue Horizon])&lt;br /&gt;
** [https://x.com/keshiAIart/status/2005254907780358201?s=20 Anime Train] (6s)&lt;br /&gt;
** [https://x.com/venturetwins/status/2006051632837189683?s=20 Michael Catson] (13s)&lt;br /&gt;
* December 2025: Runway [https://runwayml.com/research/introducing-runway-gen-4.5 Gen 4.5]&lt;br /&gt;
* December 2025: [https://arxiv.org/abs/2512.13507 Seedance 1.5]&lt;br /&gt;
&lt;br /&gt;
===2026===&lt;br /&gt;
====January 2026====&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/Itspedrito/status/2007636967048228968?s=20 Somebody That I Used to Know] (1m)&lt;br /&gt;
** [https://x.com/hujimari/status/2008054519704461407?s=20 Cat being disruptive at night], [https://x.com/klara_sjo/status/2007864014521720963?s=20 another], [https://x.com/alphafox/status/2009732284375830687?s=20 another] (c.f. [https://x.com/justalexoki/status/1988915573707661637?s=20 Cat playing instruments at midnight])&lt;br /&gt;
** [https://x.com/Uncanny_Harry/status/2008881579095961934?s=20 Character test] (30s, Kling 2.6 Motion Control, [https://x.com/Uncanny_Harry Uncanny Harry AI])&lt;br /&gt;
** [https://www.youtube.com/watch?v=SGJC4Hnz3m0&amp;amp;t=2s STAR WARS: Beggar’s Canyon | A Luke Skywalker Fan Film (Between ESB &amp;amp; ROTJ)] (7m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2009732705299104118?s=20 TZIGANE] (9m)&lt;br /&gt;
** [https://x.com/Framer_X/status/2011075884246061454?s=20 The Subway Spark] (Anime, 45s)&lt;br /&gt;
** [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ Will Smith Eating Spaghetti] (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025])&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2013675665539596651?s=20 The AI Artist] (1.5m)&lt;br /&gt;
** [https://x.com/Artedeingenio/status/2013624842021417030?s=20 Sci-fi action anime] (2m)&lt;br /&gt;
** [https://x.com/verbalriotshow/status/2014752509240475872?s=20 Stone Hand] (fake trailer, 1m)&lt;br /&gt;
* January 2026: [https://x.com/nvidia/status/2008346949301235933?s=20 Runway Gen-4.5 on] [https://www.nvidia.com/en-us/data-center/technologies/rubin/?linkId=100000401190502 Nvidia Rubin] ([https://x.com/runwayml/status/2014406560445771804?s=20 examples])&lt;br /&gt;
* January 2026: [https://ltx.io/model/ltx-2 LTX-2] open source video model (20s, 4k, w/ audio; [https://x.com/venturetwins/status/2010878914273697956?s=20 examples])&lt;br /&gt;
* January 2026: Luma [https://lumalabs.ai/blog/news/ray3_14 Ray3.14] ([https://x.com/LumaLabsAI/status/2015822842575888844?s=20 examples])&lt;br /&gt;
* January 2026: Examples:&lt;br /&gt;
** [https://x.com/pressmanc/status/2015099516500758647?s=20 Runway Gen-4.5 tests] (3.5m)&lt;br /&gt;
** [https://x.com/EHuanglu/status/2015573517618528538?s=20 Longchamp / Horses in the city] (1m)&lt;br /&gt;
** [https://x.com/dustinhollywood/status/2008154825385521418?s=20 The Last Artist] (trailer, 2m)&lt;br /&gt;
** [https://x.com/taziku_co/status/2015739943101047111?s=20 Monet temporal structure] (3m)&lt;br /&gt;
** [https://x.com/runwayml/status/2016155967285543364?s=20 Grizzlies] (1.5m, Runway Gen-4.5)&lt;br /&gt;
** [https://www.youtube.com/@TIME/videos On This Day... 1776] ([https://www.youtube.com/watch?v=E4cLKIxt8W8 trailer])&lt;br /&gt;
*** [https://www.youtube.com/watch?v=sV52AUVGc6I January 1: The Flag] (3.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=3ZDnL_a0YfQ January 10: Common Sense] (4.5m)&lt;br /&gt;
*** [https://www.youtube.com/watch?v=J5b1TiyKTus January 26: The Guns of Ticonderoga] (4m)&lt;br /&gt;
&lt;br /&gt;
====February 2026====&lt;br /&gt;
* February 2026: [https://app.klingai.com/global/quickstart/klingai-video-3-omni-model-user-guide Kling 3.0]&lt;br /&gt;
* February 2026: [https://seedance2.ai/ Seedance 2.0] ([https://x.com/EHuanglu/status/2020131622675202512?s=20 example 1], [https://x.com/EHuanglu/status/2020492770872566053?s=20 2], [https://x.com/dynamicwangs/status/2020054894741451123?s=20 3], [https://x.com/patrickassale/status/2020180495900848470?s=20 4], [https://x.com/janekm/status/2020888750285332526?s=20 5], [https://x.com/Dork_sense/status/2020179955511116082?s=20 6], [https://x.com/EHuanglu/status/2020388244802740728?s=20 7], [https://x.com/zhao_dashuai/status/2020528048341217592?s=20 8], [https://x.com/AngryTomtweets/status/2020784886932738470?s=20 9], [https://x.com/javilopen/status/2020558352590287298?s=20 10], [https://x.com/linxiaobei888/status/2021399630672691710?s=20 11])&lt;br /&gt;
* February 2026: Examples:&lt;br /&gt;
** [https://x.com/PJaccetturo/status/2019072637192843463?s=20 Unofficial opening sequence for The Way of Kings by Brandon Sanderson] (1.5m, Kling 3)&lt;br /&gt;
** [https://x.com/dailycatsclips/status/2020117502915989680?s=20 Cat Dreams] (1.5m)&lt;br /&gt;
** [https://x.com/DotCSV/status/2021269435567218725?s=20 Will Smith Eating Spaghetti] (Seedance 2.0) (c.f. [https://www.youtube.com/watch?v=XQr4Xklqzw8 April 2023], [https://x.com/kimmonismus/status/1873568693357294014 December 2024], [https://x.com/_deepfates/status/1875215969452523785 January 2025], [https://x.com/blizaine/status/1897826177970028614 March 2025], [https://x.com/javilopen/status/1925495026903380358 May 2025], [https://x.com/elder_plinius/status/1936145834585862225 June 2025], [https://x.com/minchoi/status/1976042197154963702 October 2025], [https://x.com/aisearchio/status/1978465562821898461 October 2025], [https://x.com/EHuanglu/status/1996649596119068687?s=20 December 2025], [https://www.reddit.com/r/aivideo/comments/1qi8zuv/25_years_difference_makes_you_wonder_where_ai/ January 2026], [https://x.com/SpecialSitsNews/status/2020583709741883666?s=20 progression to 2026])&lt;br /&gt;
** [https://x.com/thedorbrothers/status/2023460644905742577?s=20 To Be Continued] (3m, [https://x.com/thedorbrothers The Dor Brothers])&lt;br /&gt;
** [https://x.com/ivanka_humeniuk/status/2023711181978919034?s=20 Crow - Game of Thrones] (1m)&lt;br /&gt;
** [https://x.com/billyrestey/status/2024193251763507528?s=20 Reboot] (2m)&lt;br /&gt;
** [https://x.com/kenw_2/status/2024625510534283508?s=20 Late for work] (1.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/heydin_ai/status/2024616890338079181?s=20 AI Man] (4.5m, MJ NBP Seedance 2.0)&lt;br /&gt;
** [https://x.com/maxescu/status/2024882372836250033?s=20 But AI Will Never Be Able To Do This] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/DiDi_OKK/status/2018784243753599093?s=20 Sign] (8m)&lt;br /&gt;
** [https://x.com/LTXStudio/status/2025994426309640291?s=20 Commercial for Nexus] (1m)&lt;br /&gt;
** [https://x.com/maxescu/status/2026007558159278477?s=20 Showcase] (9m, [https://x.com/maxescu Alex Patrascu])&lt;br /&gt;
** [https://x.com/EHuanglu/status/2025410944512192536?s=20 Painterly] (30s, [https://x.com/EHuanglu el.cine])&lt;br /&gt;
** [https://x.com/kellyeld/status/2025975677657440267?s=20 Imposter Syndrone] (2m, music video)&lt;br /&gt;
** [https://www.youtube.com/watch?v=nKnE2Wn1VNQ All Is Conscious] (3.5m)&lt;br /&gt;
** [https://x.com/CuriousRefuge/status/2026086576191934769?s=20 Emotional argument] (3m, Seedance 2.0)&lt;br /&gt;
** [https://x.com/jdkanani/status/2023781028368884031?s=20 Moonlight Veil] (10m)&lt;br /&gt;
&lt;br /&gt;
====March 2026====&lt;br /&gt;
* March 2026: Examples:&lt;br /&gt;
** [https://x.com/jacopo_reale/status/2029909372764041559 Looking for Bianca] (6m, Kling 3.0)&lt;br /&gt;
** [https://x.com/sumiturkude007/status/2030933543443193908?s=20 Gardener] (3m, Seedance 2.0)&lt;br /&gt;
** Micro-movie (Chinese): [https://x.com/yyyole/status/2029225419669684418?s=20 episode 1], [https://x.com/yyyole/status/2030850450464112675?s=20 episode 2]&lt;br /&gt;
** Live-action Evangelion: [https://x.com/NACHOS2D_/status/2032401289653461052?s=20 part 1] (4.5m), [https://x.com/NACHOS2D_/status/2032778868361203770?s=20 part 2] (3.5m), [https://x.com/NACHOS2D_/status/2033126071151837491?s=20 part 3] (2.5m)&lt;br /&gt;
** [https://x.com/lexx_aura/status/2033589846216741293?s=20 to love Wu Yong] (5m)&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8732</id>
		<title>AI and Humans</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=AI_and_Humans&amp;diff=8732"/>
		<updated>2026-03-16T18:46:56Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Science */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI in Education=&lt;br /&gt;
==Survey/study of==&lt;br /&gt;
* 2023-08: [https://www.nature.com/articles/s41598-023-38964-3 Perception, performance, and detectability of conversational artificial intelligence across 32 university courses]&lt;br /&gt;
* 2023-10: [https://www.bbc.com/worklife/article/20231017-the-employees-secretly-using-ai-at-work Employees] secretly using AI at work.&lt;br /&gt;
* 2023-10: [https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/10/31/most-students-outrunning-faculty-ai-use?utm_source=Inside+Higher+Ed&amp;amp;utm_campaign=23419446b9-DNU_2021_COPY_02&amp;amp;utm_medium=email&amp;amp;utm_term=0_1fcbc04421-23419446b9-236889242&amp;amp;mc_cid=23419446b9&amp;amp;mc_eid=dae49d931a Survey] shows students using AI more than professors.&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/d41586-023-03507-3 ChatGPT has entered the classroom: how LLMs could transform education]&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-05: [https://www.nature.com/articles/s41599-025-04787-y The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis]&lt;br /&gt;
&lt;br /&gt;
==AI improves learning/education==&lt;br /&gt;
*  Mollick, Ethan R. and Mollick, Lilach and Bach, Natalie and Ciccarelli, LJ and Przystanski, Ben and Ravipinto, Daniel, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4871171 AI Agents and Education: Simulated Practice at Scale] (June 17, 2024). The Wharton School Research Paper. [http://dx.doi.org/10.2139/ssrn.4871171 doi: 10.2139/ssrn.4871171]&lt;br /&gt;
** Can enable personalized education.&lt;br /&gt;
* [https://arxiv.org/abs/2306.17156 Generative AI for Programming Education: Benchmarking ChatGPT, GPT-4, and Human Tutors]&lt;br /&gt;
** GPT4 can out-perform human tutors.&lt;br /&gt;
*  Keppler, Samantha and Sinchaisri, Wichinpong and Snyder, Clare, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4924786 Backwards Planning with Generative AI: Case Study Evidence from US K12 Teachers] (August 13, 2024). [http://dx.doi.org/10.2139/ssrn.4924786 doi: 10.2139/ssrn.4924786]&lt;br /&gt;
** Teachers benefit from using AI as a co-pilot to aid in tasks (planning, how to teach topic, explore ideas).&lt;br /&gt;
** There is smaller utility in using AI purely as a text-generator (to make quizzes, workbooks, etc.).&lt;br /&gt;
* [https://arxiv.org/abs/2402.09809 Effective and Scalable Math Support: Evidence on the Impact of an AI- Tutor on Math Achievement in Ghana]&lt;br /&gt;
* [https://doi.org/10.21203/rs.3.rs-4243877/v1 AI Tutoring Outperforms Active Learning]&lt;br /&gt;
* [https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324 From chalkboards to chatbots: Transforming learning in Nigeria, one prompt at a time] ([https://blogs.worldbank.org/en/education/From-chalkboards-to-chatbots-Transforming-learning-in-Nigeria writeup])&lt;br /&gt;
** 6 weeks of after-school AI tutoring = 2 years of typical learning gains&lt;br /&gt;
** outperforms 80% of other educational interventions&lt;br /&gt;
* [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Do Large Language Models Harm Learning?]&lt;br /&gt;
** Outcomes depend on usage&lt;br /&gt;
* [https://www.deeplearning.ai/the-batch/gpt-4-boosts-remote-tutors-performance-in-real-time-study-finds/ LLM Support for Tutors GPT-4 boosts remote tutors’ performance in real time, study finds]&lt;br /&gt;
** [https://arxiv.org/abs/2410.03017 Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise]&lt;br /&gt;
* 2025-06: Gallup &amp;amp; The Walton Foundation: [https://www.gallup.com/file/analytics/691922/Walton-Family-Foundation-Gallup-Teachers-AI-Report.pdf Teaching for Tomorrow Unlocking Six Weeks a Year With AI]&lt;br /&gt;
&lt;br /&gt;
==AI harms learning==&lt;br /&gt;
* [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study]&lt;br /&gt;
** Current grading systems cannot detect AI.&lt;br /&gt;
*  Bastani, Hamsa and Bastani, Osbert and Sungu, Alp and Ge, Haosen and Kabakcı, Özge and Mariman, Rei, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4895486 Generative AI Can Harm Learning] (July 15, 2024). The Wharton School Research Paper.[http://dx.doi.org/10.2139/ssrn.4895486 doi: 10.2139/ssrn.4895486]&lt;br /&gt;
** Access to ChatGPT harmed math education outcomes.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09047 AI Meets the Classroom: When Does ChatGPT Harm Learning?]&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.20245 How AI Impacts Skill Formation]&lt;br /&gt;
&lt;br /&gt;
==Software/systems==&lt;br /&gt;
* [https://devpost.com/software/gptutor GPTutor] ([https://github.com/mynamegabe/GPTutor code])&lt;br /&gt;
* [https://arxiv.org/abs/2308.02773 EduChat: A Large-Scale Language Model-based Chatbot System for Intelligent Education]&lt;br /&gt;
* [https://eurekalabs.ai/ Eureka Labs] (founded by [https://en.wikipedia.org/wiki/Andrej_Karpathy Andrej Karpathy]) aims to create AI-driven courses (first course is [https://github.com/karpathy/LLM101n Intro to LLMs])&lt;br /&gt;
&lt;br /&gt;
===LLMs===&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16429 LearnLM: Improving Gemini for Learning]&lt;br /&gt;
&lt;br /&gt;
===Individual tools===&lt;br /&gt;
* Chatbot (OpenAI [https://chatgpt.com/ ChatGPT], Anthropic [https://www.anthropic.com/claude Claude], Google [https://gemini.google.com/app Gemini])&lt;br /&gt;
* [https://notebooklm.google.com/ NotebookLM]: Enables one to &amp;quot;chat with documents&amp;quot;.&lt;br /&gt;
* Google [https://learning.google.com/experiments/learn-about/signup Learn About]&lt;br /&gt;
&lt;br /&gt;
===Systems===&lt;br /&gt;
* [https://www.anthropic.com/news/introducing-claude-for-education Anthropic] [https://www.anthropic.com/education Claude for Education]&lt;br /&gt;
&lt;br /&gt;
==AI for grading==&lt;br /&gt;
* [https://dl.acm.org/doi/10.1145/3657604.3664693 Can Large Language Models Make the Grade? An Empirical Study Evaluating LLMs Ability To Mark Short Answer Questions in K-12 Education] ([https://arxiv.org/abs/2405.02985 preprint])&lt;br /&gt;
&lt;br /&gt;
==Detection==&lt;br /&gt;
* 2024-06: [https://www.sciencedirect.com/science/article/pii/S2666920X24000109 Do teachers spot AI? Evaluating the detectability of AI-generated texts among student essays]&lt;br /&gt;
** GenAI can simulate student writing in a way that teachers cannot detect.&lt;br /&gt;
** AI essays are assessed more positively than student-written.&lt;br /&gt;
** Teachers are overconfident in their source identification.&lt;br /&gt;
** Both novice and experienced teachers could not identify texts generated by ChatGPT vs. students&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.15654 People who frequently use ChatGPT for writing tasks are accurate and robust detectors of AI-generated text]&lt;br /&gt;
===AI Text Detectors Don&amp;#039;t Work===&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.07940 RAID: A Shared Benchmark for Robust Evaluation of Machine-Generated Text Detectors]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2306.15666 Testing of Detection Tools for AI-Generated Text]&lt;br /&gt;
&lt;br /&gt;
=AI/human=&lt;br /&gt;
==Capabilities==&lt;br /&gt;
===Writing===&lt;br /&gt;
&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.22828 Learning to Reason for Long-Form Story Generation]&lt;br /&gt;
&lt;br /&gt;
==AI out-performs humans==&lt;br /&gt;
===Tests===&lt;br /&gt;
* 2023-07: [https://arxiv.org/abs/2307.10635 SciBench: Evaluating College-Level Scientific Problem-Solving Abilities of Large Language Models]&lt;br /&gt;
* 2024-06: [https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0305354 A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study] &lt;br /&gt;
** AI scores higher than median students.&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2023-07: [https://mackinstitute.wharton.upenn.edu/wp-content/uploads/2023/08/LLM-Ideas-Working-Paper.pdf Ideas Are Dimes A Dozen: Large Language Models For Idea Generation In Innovation]&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/s41598-023-40858-3 Best humans still outperform artificial intelligence in a creative divergent thinking task]&lt;br /&gt;
** Best humans out-perform AI at creativity. (By implication, median humans may not.)&lt;br /&gt;
* 2024-02: [https://www.nature.com/articles/s41598-024-53303-w The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks]&lt;br /&gt;
* 2024-02: Felin, Teppo and Holweg, Matthias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4737265 Theory Is All You Need: AI, Human Cognition, and Causal Reasoning] (February 24, 2024). [http://dx.doi.org/10.2139/ssrn.4737265 doi: 10.2139/ssrn.4737265]&lt;br /&gt;
** Argues that human &amp;quot;theory-based&amp;quot; creativity is better than AI &amp;quot;data-based&amp;quot;.&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.01119 Pron vs Prompt: Can Large Language Models already Challenge a World-Class Fiction Author at Creative Text Writing?]&lt;br /&gt;
** Top human (professional author) out-performs GPT4.&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.04109 Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers]&lt;br /&gt;
** LLMs can be creative&lt;br /&gt;
* 2024-09: [https://docs.iza.org/dp17302.pdf Creative and Strategic Capabilities of Generative AI: Evidence from Large-Scale Experiments]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
&lt;br /&gt;
===Art===&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?]&lt;br /&gt;
&lt;br /&gt;
===Business &amp;amp; Marketing===&lt;br /&gt;
* 2023-11: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4597899 The power of generative marketing: Can generative AI create superhuman visual marketing content?]&lt;br /&gt;
* 2024-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4714776 Generative Artificial Intelligence and Evaluating Strategic Decisions]&lt;br /&gt;
&lt;br /&gt;
===Professions===&lt;br /&gt;
* [https://agi.safe.ai/submit Humanity&amp;#039;s Last Exam]&lt;br /&gt;
** [https://x.com/alexandr_wang/status/1835738937719140440 Effort to build] a dataset of challenging (but resolvable) questions in specific domain areas, to act as a benchmark to test whether AIs are improving in these challenging topics.&lt;br /&gt;
&lt;br /&gt;
====Coding====&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.06807 Competitive Programming with Large Reasoning Models]&lt;br /&gt;
&lt;br /&gt;
====Medical====&lt;br /&gt;
* 2024-03: [https://www.medrxiv.org/content/10.1101/2024.03.12.24303785v1 Influence of a Large Language Model on Diagnostic Reasoning: A Randomized Clinical Vignette Study]&lt;br /&gt;
** GPT4 improves medical practitioner work; surprisingly, GPT4 alone scored better than a human with GPT4 as aid (on selected tasks).&lt;br /&gt;
* 2024-10: [https://doi.org/10.1001/jamanetworkopen.2024.38535 Perspectives on Artificial Intelligence–Generated Responses to Patient Messages]&lt;br /&gt;
* 2024-10: [https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395 Large Language Model Influence on Diagnostic Reasoning; A Randomized Clinical Trial]&lt;br /&gt;
**  Use of ChatGPT does not strongly improve medical expert work; but AI alone out-scores human or human+AI&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41562-024-02046-9 Large language models surpass human experts in predicting neuroscience results] (writeup: [https://medicalxpress.com/news/2024-11-ai-neuroscience-results-human-experts.html AI can predict neuroscience study results better than human experts, study finds])&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.10849 Superhuman performance of a large language model on the reasoning tasks of a physician]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18925 HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs]&lt;br /&gt;
* 2025-02: Media:&lt;br /&gt;
** NY Times: [https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html The Robot Doctor Will See You Now]&lt;br /&gt;
** [https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed When Doctors With A.I. Are Outperformed by A.I. Alone]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-024-03456-y GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s41591-025-03517-w Artificial intelligence for individualized treatment of persistent atrial fibrillation: a randomized controlled trial]&lt;br /&gt;
* Google AI Clinician:&lt;br /&gt;
** 2024-01: [https://arxiv.org/abs/2401.05654 Towards Conversational Diagnostic AI] ([https://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/ blog]: Articulate Medical Intelligence Explorer, AMIE)&lt;br /&gt;
** 2025-03: [https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf Towards Conversational AI for Disease Management] ([https://research.google/blog/from-diagnosis-to-treatment-advancing-amie-for-longitudinal-disease-management/ blog])&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.19655 Med-RLVR: Emerging Medical Reasoning from a 3B base model via reinforcement Learning]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.13939 Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models]&lt;br /&gt;
* 2025-04: [https://www.acpjournals.org/doi/10.7326/ANNALS-24-03283 Comparison of Initial Artificial Intelligence (AI) and Final Physician Recommendations in AI-Assisted Virtual Urgent Care Visits]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052 Towards conversational diagnostic artificial intelligence]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054 Towards accurate differential diagnosis with large language models]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.13.25329541v1 Automation of Systematic Reviews with Large Language Models]&lt;br /&gt;
* 2025-06: [https://microsoft.ai/new/the-path-to-medical-superintelligence/ The Path to Medical Superintelligence]&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s41591-025-03888-0?utm_source=chatgpt.com A personal health large language model for sleep and fitness coaching]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.08224 Capabilities of GPT-5 on Multimodal Medical Reasoning]&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* 2025-04: [https://www.virologytest.ai/vct_paper.pdf Virology Capabilities Test (VCT): A Multimodal Virology Q&amp;amp;A Benchmark]&lt;br /&gt;
** Time: [https://time.com/7279010/ai-virus-lab-biohazard-study/ Exclusive: AI Outsmarts Virus Experts in the Lab, Raising Biohazard Fears]&lt;br /&gt;
** AI Frontiers: [https://www.ai-frontiers.org/articles/ais-are-disseminating-expert-level-virology-skills AIs Are Disseminating Expert-Level Virology Skills]&lt;br /&gt;
&lt;br /&gt;
====Therapy====&lt;br /&gt;
* 2025-02: [https://journals.plos.org/mentalhealth/article?id=10.1371/journal.pmen.0000145 When ELIZA meets therapists: A Turing test for the heart and mind]&lt;br /&gt;
* 2025-03: Therabot: [https://ai.nejm.org/doi/full/10.1056/AIoa2400802 Randomized Trial of a Generative AI Chatbot for Mental Health Treatment]&lt;br /&gt;
&lt;br /&gt;
====Financial====&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.17866 Financial Statement Analysis with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
====HR====&lt;br /&gt;
* 2025-08: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5395709 Voice AI in Firms: A Natural Field Experiment on Automated Job Interviews]&lt;br /&gt;
&lt;br /&gt;
==AI improves human work==&lt;br /&gt;
* 2023-07: [https://www.science.org/doi/10.1126/science.adh2586 Experimental evidence on the productivity effects of generative artificial intelligence]&lt;br /&gt;
* 2023-09:  Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Mollick, Ethan R. and Lifshitz-Assaf, Hila and Kellogg, Katherine and Rajendran, Saran and Krayer, Lisa and Candelon, François and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321 Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality] (September 15, 2023). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper No. 24-013, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4573321 doi: 10.2139/ssrn.4573321]&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work] (National Bureau of Economic Research)&lt;br /&gt;
* 2023-12: [https://osf.io/hdjpk The Uneven Impact of Generative AI on Entrepreneurial Performance] ([https://doi.org/10.31219/osf.io/hdjpk doi: 10.31219/osf.io/hdjpk])&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.05481 Artificial Intelligence in the Knowledge Economy]: Non-autonomous AI (chatbot) benefits least knowledgeable workers; autonomous agents benefit the most knowledgeable workers&lt;br /&gt;
* 2024-07: [https://www.microsoft.com/en-us/research/publication/generative-ai-in-real-world-workplaces/ Generative AI in Real-World Workplaces: The Second Microsoft Report on AI and Productivity Research]&lt;br /&gt;
* 2025-03: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5188231 The Cybernetic Teammate: A Field Experiment on Generative AI Reshaping Teamwork and Expertise]&lt;br /&gt;
** 2025-03: Ethan Mollick: [https://www.oneusefulthing.org/p/the-cybernetic-teammateThe Cybernetic Teammate]: Having an AI on your team can increase performance, provide expertise, and improve your experience&lt;br /&gt;
* 2025-09: [https://osf.io/preprints/psyarxiv/vbkmt_v1 Quantifying Human-AI Synergy]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.12049 Generative AI and Firm Productivity: Field Experiments in Online Retail]&lt;br /&gt;
* 2025-10: Wharton: [https://knowledge.wharton.upenn.edu/special-report/2025-ai-adoption-report/ 2025 AI Adoption Report] (75% report positive ROI)&lt;br /&gt;
&lt;br /&gt;
===Coding===&lt;br /&gt;
* 2023-02: [https://arxiv.org/abs/2302.06590 The Impact of AI on Developer Productivity: Evidence from GitHub Copilot]&lt;br /&gt;
* 2024-09:  Cui, Zheyuan and Demirer, Mert and Jaffe, Sonia and Musolff, Leon and Peng, Sida and Salz, Tobias, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4945566 The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers] (September 03, 2024). [http://dx.doi.org/10.2139/ssrn.4945566 doi: 10.2139/ssrn.4945566 ]&lt;br /&gt;
* 2024-11:  Hoffmann, Manuel and Boysel, Sam and Nagle, Frank and Peng, Sida and Xu, Kevin, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5007084 Generative AI and the Nature of Work] (October 27, 2024). Harvard Business School Strategy Unit Working Paper No. 25-021, Harvard Business Working Paper No. No. 25-021, [http://dx.doi.org/10.2139/ssrn.5007084 doi: 10.2139/ssrn.5007084]&lt;br /&gt;
* 2025-07: METR: [https://arxiv.org/abs/2507.09089 Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] (AI tools led to lower performance)&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools now lead to improved performance)&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.19708 Intuition to Evidence: Measuring AI&amp;#039;s True Impact on Developer Productivity]&lt;br /&gt;
&lt;br /&gt;
===Forecasting===&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.07862 AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy]&lt;br /&gt;
&lt;br /&gt;
===Finance===&lt;br /&gt;
* 2024-12: [https://dx.doi.org/10.2139/ssrn.5075727 AI, Investment Decisions, and Inequality]: Novices see improvements in investment performance, sophisticated investors see even greater improvements.&lt;br /&gt;
&lt;br /&gt;
===Law===&lt;br /&gt;
* 2025-03: [https://ssrn.com/abstract=5162111 AI-Powered Lawyering: AI Reasoning Models, Retrieval Augmented Generation, and the Future of Legal Practice]&lt;br /&gt;
&lt;br /&gt;
===Science===&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/abs/10.1126/science.adw3000 Scientific production in the era of large language models]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-09922-y Artificial intelligence tools expand scientists’ impact but contract science’s focus]&lt;br /&gt;
* 2026-01: [https://www.anthropic.com/news/accelerating-scientific-research How scientists are using Claude to accelerate research and discovery]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
* 2026-03: [https://www.pnas.org/doi/10.1073/pnas.2533676123 Expert evaluation of LLM world models: A high-Tc superconductivity case study] ([https://research.google/blog/testing-llms-on-superconductivity-research-questions/?utm_source=twitter&amp;amp;utm_medium=social&amp;amp;utm_campaign=social_post&amp;amp;utm_content=gr-acct blog])&lt;br /&gt;
&lt;br /&gt;
===Medical===&lt;br /&gt;
* 2025-03: [https://www.medrxiv.org/content/10.1101/2025.02.28.25323115v1.full Medical Hallucination in Foundation Models and Their Impact on Healthcare]&lt;br /&gt;
* 2025-03: [https://journals.lww.com/international-journal-of-surgery/fulltext/2025/03000/chatgpt_s_role_in_alleviating_anxiety_in_total.20.aspx ChatGPT’s role in alleviating anxiety in total knee arthroplasty consent process: a randomized controlled trial pilot study]&lt;br /&gt;
* 2025-05: [https://openai.com/index/healthbench/ Introducing HealthBench]&lt;br /&gt;
* 2025-06: [https://www.medrxiv.org/content/10.1101/2025.06.07.25329176v1 From Tool to Teammate: A Randomized Controlled Trial of Clinician-AI Collaborative Workflows for Diagnosis]&lt;br /&gt;
* 2025-06: [https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07414-1 Iteratively refined ChatGPT outperforms clinical mentors in generating high-quality interprofessional education clinical scenarios: a comparative study]&lt;br /&gt;
* 2025-07: [https://cdn.openai.com/pdf/a794887b-5a77-4207-bb62-e52c900463f1/penda_paper.pdf AI-based Clinical Decision Support for Primary Care: A Real-World Study] ([https://openai.com/index/ai-clinical-copilot-penda-health/ blog])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15743 Towards physician-centered oversight of conversational diagnostic AI]&lt;br /&gt;
* 2026-01: [https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(25)02464-X/abstract Interval cancer, sensitivity, and specificity comparing AI-supported mammography screening with standard double reading without AI in the MASAI study: a randomised, controlled, non-inferiority, single-blinded, population-based, screening-accuracy trial]&lt;br /&gt;
&lt;br /&gt;
===Translation===&lt;br /&gt;
* 2025-01: [https://simonwillison.net/2025/Feb/2/workflow-for-translation/ A professional workflow for translation using LLMs] ([https://news.ycombinator.com/item?id=42897856 based on this])&lt;br /&gt;
&lt;br /&gt;
===Customer service===&lt;br /&gt;
* 2023-11: [https://www.nber.org/papers/w31161 Generative AI at Work]: Improvements for workers and clients (though also a ceiling to improvement)&lt;br /&gt;
&lt;br /&gt;
===Creativity===&lt;br /&gt;
* See also: [[AI creativity]]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.01727 Prompting Diverse Ideas: Increasing AI Idea Variance]&lt;br /&gt;
* 2024-07: [https://www.science.org/doi/10.1126/sciadv.adn5290 Generative AI enhances individual creativity but reduces the collective diversity of novel content]&lt;br /&gt;
* 2024-08: [https://www.nature.com/articles/s41562-024-01953-1 An empirical investigation of the impact of ChatGPT on creativity]&lt;br /&gt;
** 2024-08: Response: [https://www.nature.com/articles/s41562-024-01953-1 ChatGPT decreases idea diversity in brainstorming] ([https://www.nature.com/articles/s41562-025-02173-x.epdf?sharing_token=LA9NyDHj7y5WN8zvb5Qm49RgN0jAjWel9jnR3ZoTv0Nl8PrpXFkjZ93XvmUVBgB9Hlfro5Yo6YELr-pRqbpk3HaZENCvsfV8G1kwtTEj2oW1g87dSVT4BzrfCu3jS_606SLzmoDuDiALChY-MozVM4Pj1b4Vdf-YaIH5p3lfAnM%3D pdf])&lt;br /&gt;
** 2025-05: Response: [https://www.nature.com/articles/s41562-025-02195-5 Reply to: ChatGPT decreases idea diversity in brainstorming]&lt;br /&gt;
* 2024-08: [https://doi.org/10.1287/orsc.2023.18430 The Crowdless Future? Generative AI and Creative Problem-Solving]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03703 Human Creativity in the Age of LLMs]&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://conference.nber.org/conf_papers/f210475.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;: diffusion model increases &amp;quot;innovation&amp;quot; (patents), boosts the best performers, but also removes some enjoyable tasks.&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2024-12: [https://doi.org/10.1080/10400419.2024.2440691 Using AI to Generate Visual Art: Do Individual Differences in Creativity Predict AI-Assisted Art Quality?] ([https://osf.io/preprints/psyarxiv/ygzw6 preprint]): shows that more creative humans produce more creative genAI outputs&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.11433 One Does Not Simply Meme Alone: Evaluating Co-Creativity Between LLMs and Humans in the Generation of Humor]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.17241 Generative AI and Creativity: A Systematic Literature Review and Meta-Analysis]&lt;br /&gt;
&lt;br /&gt;
===Equity===&lt;br /&gt;
* 2025-01: [https://ai.nejm.org/doi/full/10.1056/AIp2400889 Using Large Language Models to Promote Health Equity]&lt;br /&gt;
&lt;br /&gt;
==AI worse than humans==&lt;br /&gt;
* 2025-04: [https://spinup-000d1a-wp-offload-media.s3.amazonaws.com/faculty/wp-content/uploads/sites/27/2025/03/AI-debt-collection-20250331.pdf How Good is AI at Twisting Arms? Experiments in Debt Collection]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.18919 Clinical knowledge in LLMs does not translate to human interactions]&lt;br /&gt;
* 2025-05: [https://royalsocietypublishing.org/doi/10.1098/rsos.241776 Generalization bias in large language model summarization of scientific research]&lt;br /&gt;
&lt;br /&gt;
==AI lowers human capability==&lt;br /&gt;
* 2025-07: METR: [https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity] ([https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/ blog], [https://secondthoughts.ai/p/ai-coding-slowdown commentary/analysis])&lt;br /&gt;
** 2026-02: [https://metr.org/blog/2026-02-24-uplift-update/ We are Changing our Developer Productivity Experiment Design] (AI tools [https://x.com/METR_Evals/status/2026355544668385373?s=20 now] lead to improved performance)&lt;br /&gt;
* 2026-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646 Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender]&lt;br /&gt;
&lt;br /&gt;
==Human Perceptions of AI==&lt;br /&gt;
* 2023-09: [https://www.nature.com/articles/d41586-023-02980-0 AI and science: what 1,600 researchers think. A Nature survey finds that scientists are concerned, as well as excited, by the increasing use of artificial-intelligence tools in research.]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1016/S2589-7500(24)00202-4 Attitudes and perceptions of medical researchers towards the use of artificial intelligence chatbots in the scientific process: an international cross-sectional survey] (Nature commentary: [https://www.nature.com/articles/s41592-024-02369-5 Quest for AI literacy])&lt;br /&gt;
* 2025-03: [https://www.arxiv.org/abs/2503.16458 Users Favor LLM-Generated Content -- Until They Know It&amp;#039;s AI]&lt;br /&gt;
&lt;br /&gt;
===AI passes Turing Test===&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Text Dialog&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2023-05: [https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test]&lt;br /&gt;
* 2023-10: [https://arxiv.org/abs/2310.20216 Does GPT-4 pass the Turing test?]&lt;br /&gt;
* 2024-05: [https://arxiv.org/abs/2405.08007 People cannot distinguish GPT-4 from a human in a Turing test]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.08853 GPT-4 is judged more human than humans in displaced and inverted Turing tests]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23674 Large Language Models Pass the Turing Test]&lt;br /&gt;
* 2025-04: [https://www.sciencedirect.com/science/article/abs/pii/S0022103117303980 A Minimal Turing Test]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Art&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2024-11: [https://www.astralcodexten.com/p/how-did-you-do-on-the-ai-art-turing How Did You Do On The AI Art Turing Test?] Differentiation was only slightly above random (60%). AI art was often ranked higher than human-made.&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41598-024-76900-1 AI-generated poetry is indistinguishable from human-written poetry and is rated more favorably]&lt;br /&gt;
* 2025-09: [https://arxiv.org/abs/2509.25601 Echoes of Humanity: Exploring the Perceived Humanness of AI Music]&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Imagery&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* 2026-02: [https://bpspsychub.onlinelibrary.wiley.com/doi/10.1111/bjop.70063 Too good to be true: Synthetic AI faces are more average than real faces and super-recognizers know it]&lt;br /&gt;
** [https://www.unsw.edu.au/newsroom/news/2026/02/humans-overconfident-telling-AI-faces-real-faces-people-fake People are overconfident about spotting AI faces, study finds]&lt;br /&gt;
&lt;br /&gt;
=Uptake=&lt;br /&gt;
* 2023-07: [https://doi.org/10.9734/ajrcos/2023/v16i4392 ChatGPT: Early Adopters, Teething Issues and the Way Forward]&lt;br /&gt;
* 2024-03: [https://arxiv.org/abs/2403.07183 Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews]&lt;br /&gt;
* 2024-05:  Humlum, Anders and Vestergaard, Emilie, [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4827166 The Adoption of ChatGPT]. IZA Discussion Paper No. 16992 [http://dx.doi.org/10.2139/ssrn.4827166 doi: 10.2139/ssrn.4827166]&lt;br /&gt;
* 2024-06: Kellogg, Katherine and Lifshitz-Assaf, Hila and Randazzo, Steven and Mollick, Ethan R. and Dell&amp;#039;Acqua, Fabrizio and McFowland III, Edward and Candelon, Francois and Lakhani, Karim R., [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4857373 Don&amp;#039;t Expect Juniors to Teach Senior Professionals to Use Generative AI: Emerging Technology Risks and Novice AI Risk Mitigation Tactics] (June 03, 2024). Harvard Business School Technology &amp;amp; Operations Mgt. Unit Working Paper 24-074, Harvard Business Working Paper No. 24-074, The Wharton School Research Paper [http://dx.doi.org/10.2139/ssrn.4857373 doi: 10.2139/ssrn.4857373 ]&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.07016 Delving into ChatGPT usage in academic writing through excess vocabulary]&lt;br /&gt;
* 2024-09: [https://static1.squarespace.com/static/60832ecef615231cedd30911/t/66f0c3fbabdc0a173e1e697e/1727054844024/BBD_GenAI_NBER_Sept2024.pdf The Rapid Adoption of Generative AI]&lt;br /&gt;
* 2024-10: [https://ai.wharton.upenn.edu/focus-areas/human-technology-interaction/2024-ai-adoption-report/ Growing Up: Navigating Generative AI’s Early Years – AI Adoption Report] ([https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Executive-Summary.pdf executive summary], [https://ai.wharton.upenn.edu/wp-content/uploads/2024/10/AI-Report_Full-Report.pdf full report])&lt;br /&gt;
** 72% of leaders use genAI at least once a week (c.f. 23% in 2023); 90% agree AI enhances skills (c.f. 80% in 2023)&lt;br /&gt;
** Spending on genAI is up 130% (most companies plan to invest going forward)&lt;br /&gt;
* 2024-12: [https://www.pnas.org/doi/10.1073/pnas.2414972121 The unequal adoption of ChatGPT exacerbates existing inequalities among workers]&lt;br /&gt;
** Higher adoption among young and less experienced&lt;br /&gt;
** Lower adoption among women and lower-earning workers&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.09747 The Widespread Adoption of Large Language Model-Assisted Writing Across Society]: 10-25% adoption across a range of contexts&lt;br /&gt;
* 2025-02: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5078805 Local Heterogeneity in Artificial Intelligence Jobs Over Time and Space]&lt;br /&gt;
* 2025-04: [https://andreyfradkin.com/assets/demandforllm.pdf Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming]&lt;br /&gt;
* 2025-05: [https://civicscience.com/chatgpt-is-still-leading-the-ai-wars-but-google-gemini-is-gaining-ground/ ChatGPT Is Still Leading the AI Wars but Google Gemini Is Gaining Ground]&lt;br /&gt;
* 2025-05: [https://www.nber.org/papers/w33777 Large Language Models, Small Labor Market Effects]&lt;br /&gt;
** Significant uptake, but very little economic impact so far&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5136877 The Labor Market Effects of Generative Artificial Intelligence]&lt;br /&gt;
** US worker usage of AI increasingly rapidly: 30% in 2024-12; 40% in 2025-05&lt;br /&gt;
* 2025-05: [https://www.bondcap.com/report/pdf/Trends_Artificial_Intelligence.pdf Trends – Artificial Intelligence]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08945 Who is using AI to code? Global diffusion and impact of generative AI]&lt;br /&gt;
* 2025-06: [https://www.iconiqcapital.com/growth/reports/2025-state-of-ai 2025 State of AI Report: The Builder’s Playbook] A Practical Roadmap for AI Innovation&lt;br /&gt;
* 2025-07: METR: [https://epochai.substack.com/p/after-the-chatgpt-moment-measuring After the ChatGPT Moment: Measuring AI’s Adoption How quickly has AI been diffusing through the economy?]&lt;br /&gt;
* 2025-07: Pew Research: [https://www.pewresearch.org/short-reads/2025/06/25/34-of-us-adults-have-used-chatgpt-about-double-the-share-in-2023/ 34% of U.S. adults have used ChatGPT, about double the share in 2023]&lt;br /&gt;
* 2025-12: Epoch AI: [https://epoch.ai/data/polling Polling on AI Usage]&lt;br /&gt;
&lt;br /&gt;
==Usage By==&lt;br /&gt;
* 2026-02: [https://www.nber.org/papers/w34813 The Politics of AI]&lt;br /&gt;
&lt;br /&gt;
==Usage For==&lt;br /&gt;
* 2024-12: [https://assets.anthropic.com/m/7e1ab885d1b24176/original/Clio-Privacy-Preserving-Insights-into-Real-World-AI-Use.pdf Clio: A system for privacy-preserving insights into real-world AI use] (Anthropic [https://www.anthropic.com/research/clio Clio])&lt;br /&gt;
* 2025-03: [https://learn.filtered.com/hubfs/The%202025%20Top-100%20Gen%20AI%20Use%20Case%20Report.pdf How People are Really Using Generative AI Now] ([https://hbr.org/2025/04/how-people-are-really-using-gen-ai-in-2025 writeup])&lt;br /&gt;
* 2025-04: [https://www.anthropic.com/news/anthropic-education-report-how-university-students-use-claude Anthropic Education Report: How University Students Use Claude]&lt;br /&gt;
* 2025-09: [https://www.anthropic.com/research/economic-index-geography Anthropic Economic Index: Tracking AI&amp;#039;s role in the US and global economy]&lt;br /&gt;
* 2025-09: [https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf How People Use ChatGPT] (OpenAI)&lt;br /&gt;
&lt;br /&gt;
==Hiding Usage==&lt;br /&gt;
* 2025-05: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5232910 Underreporting of AI use: The role of social desirability bias]&lt;br /&gt;
&lt;br /&gt;
=Societal Effects/Transformations=&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.01754 Empirical evidence of Large Language Model&amp;#039;s influence on human spoken communication]&lt;br /&gt;
* 2025-09: [https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5425555 Generative AI as Seniority-Biased Technological Change: Evidence from U.S. Résumé and Job Posting Data]&lt;br /&gt;
&lt;br /&gt;
=Psychological Impact=&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.16628 The Impact of Artificial Intelligence on Human Thought]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15352 People readily follow personal advice from AI but it does not improve their well-being]&lt;br /&gt;
&lt;br /&gt;
==Human Sentiment towards AI==&lt;br /&gt;
* 2025-04: Pew Research: [https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/ How the U.S. Public and AI Experts View Artificial Intelligence]&lt;br /&gt;
* 2025-10: Pew Research: [https://www.pewresearch.org/global/2025/10/15/how-people-around-the-world-view-ai/ How People Around the World View AI: More are concerned than excited about its use, and more trust their own country and the EU to regulate it than trust the U.S. or China]&lt;br /&gt;
* 2025-12: [https://www.edelman.com/sites/g/files/aatuss191/files/2025-11/2025%20Edelman%20Trust%20Barometer%20Flash%20Poll%20Trust%20and%20Artificial%20Intelligence%20at%20a%20Crossroads%201.pdf 2025 Edelman Trust Barometer]&lt;br /&gt;
* 2025-12: [https://navigatorresearch.org/views-of-ai-and-data-centers/ Polling - Views of AI and data centers]&lt;br /&gt;
* 2026-03: [https://osf.io/preprints/psyarxiv/5mwre_v9 The Moralization of Artificial Intelligence]&lt;br /&gt;
&lt;br /&gt;
==AI Persuasion of Humans==&lt;br /&gt;
(AI can update beliefs, change opinions, tackle conspiracy theories, etc.)&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.15006 Fine-tuning language models to find agreement among humans with diverse preferences]&lt;br /&gt;
* 2024-08: [https://arxiv.org/abs/2408.04681 Conversational AI Powered by Large Language Models Amplifies False Memories in Witness Interviews]&lt;br /&gt;
* 2024-04: [https://osf.io/preprints/psyarxiv/h7n8u_v1 Just the facts: How dialogues with AI reduce conspiracy beliefs]&lt;br /&gt;
* 2024-09: [https://www.science.org/doi/10.1126/science.adq1814 Durably reducing conspiracy beliefs through dialogues with AI]&lt;br /&gt;
* 2025-03: [https://www.pnas.org/doi/10.1073/pnas.2413443122 Scaling language model size yields diminishing returns for single-message political persuasion]&lt;br /&gt;
* 2025-04: [https://drive.google.com/file/d/1Eo4SHrKGPErTzL1t_QmQhfZGU27jKBjx/edit Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment]&lt;br /&gt;
** [https://www.404media.co/researchers-secretly-ran-a-massive-unauthorized-ai-persuasion-experiment-on-reddit-users/ Researchers Secretly Ran a Massive, Unauthorized AI Persuasion Experiment on Reddit Users]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.09662 Large Language Models Are More Persuasive Than Incentivized Human Persuaders]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.13919 The Levers of Political Persuasion with Conversational AI]&lt;br /&gt;
* 2025-12: [https://www.science.org/doi/10.1126/science.aea3884 The levers of political persuasion with conversational artificial intelligence]&lt;br /&gt;
* 2025-12: [https://www.nature.com/articles/s41586-025-09771-9 Persuading voters using human–artificial intelligence dialogues]&lt;br /&gt;
&lt;br /&gt;
==AI Effects on Human Psychology==&lt;br /&gt;
===Human well-being===&lt;br /&gt;
* 2024-01: [https://www.nature.com/articles/s44184-023-00047-6 Loneliness and suicide mitigation for students using GPT3-enabled chatbots]&lt;br /&gt;
* 2025-03: [https://cdn.openai.com/papers/15987609-5f71-433c-9972-e91131f399a1/openai-affective-use-study.pdf Investigating Affective Use and Emotional Well-being on ChatGPT]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
&lt;br /&gt;
===Counter loneliness===&lt;br /&gt;
* 2023-11: [https://arxiv.org/abs/2311.10599 Chatbots as social companions: How people perceive consciousness, human likeness, and social health benefits in machines]&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.19096 AI Companions Reduce Loneliness]&lt;br /&gt;
* 2025-03: [https://dam-prod2.media.mit.edu/x/2025/03/21/Randomized_Control_Study_on_Chatbot_Psychosocial_Effect.pdf How AI and Human Behaviors Shape Psychosocial Effects of Chatbot Use: A Longitudinal Controlled Study]&lt;br /&gt;
* 2025-06: Anthropic: [https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship How People Use Claude for Support, Advice, and Companionship]&lt;br /&gt;
&lt;br /&gt;
===Human mental abilities (creativity, learning)===&lt;br /&gt;
* 2025-03: [https://www.microsoft.com/en-us/research/publication/the-impact-of-generative-ai-on-critical-thinking-self-reported-reductions-in-cognitive-effort-and-confidence-effects-from-a-survey-of-knowledge-workers/ The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.08872 Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task]&lt;br /&gt;
&lt;br /&gt;
=Simulate Humans=&lt;br /&gt;
* See also: [[Human brain]]&lt;br /&gt;
&lt;br /&gt;
==Sociology==&lt;br /&gt;
* 2021-10: [https://www.doi.org/10.1007/s10588-021-09351-y Explaining and predicting human behavior and social dynamics in simulated virtual worlds: reproducibility, generalizability, and robustness of causal discovery methods]&lt;br /&gt;
* 2023-12: Google: [https://arxiv.org/abs/2312.03664 Generative agent-based modeling with actions grounded in physical, social, or digital space using Concordia]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2402.12620 Are Large Language Models (LLMs) Good Social Predictors?]&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.11794 Automated Social Science: Language Models as Scientist and Subjects]&lt;br /&gt;
* 2024-07: [https://academic.oup.com/pnasnexus/article/3/7/pgae245/7712371 Perils and opportunities in using large language models in psychological research]&lt;br /&gt;
* 2024-08: [https://samim.io/dl/Predicting%20results%20of%20social%20science%20experiments%20using%20large%20language%20models.pdf Predicting Results of Social Science Experiments Using Large Language Models]&lt;br /&gt;
* 2024-10: [https://www.pnas.org/doi/10.1073/pnas.2407639121 Large Language Models based on historical text could offer informative tools for behavioral science]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02234 LLM Social Simulations Are a Promising Research Method]&lt;br /&gt;
* 2025-04: [https://www.nber.org/papers/w33662 Measuring Human Leadership Skills with AI Agents]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.10157 SocioVerse: A World Model for Social Simulation Powered by LLM Agents and A Pool of 10 Million Real-World Users]&lt;br /&gt;
* 2025-07: [https://www.nature.com/articles/s41586-025-09215-4 A foundation model to predict and capture human cognition] ([https://marcelbinz.github.io/centaur code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.15815 LLM Economist: Large Population Models and Mechanism Design in Multi-Agent Generative Simulacra]&lt;br /&gt;
* 2025-09: [https://benjaminmanning.io/files/optimize.pdf General Social Agents]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2506.06958 Simulating Society Requires Simulating Thought]&lt;br /&gt;
&lt;br /&gt;
==Theory of Mind==&lt;br /&gt;
* 2025-08: [https://www.nature.com/articles/s44387-025-00031-9 How large language models encode theory-of-mind: a study on sparse parameter patterns]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2509.22887 Infusing Theory of Mind into Socially Intelligent LLM Agents]&lt;br /&gt;
&lt;br /&gt;
==Humanlike Vibes==&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.20525 The Xeno Sutra: Can Meaning and Value be Ascribed to an AI-Generated &amp;quot;Sacred&amp;quot; Text?]&lt;br /&gt;
* 2025-10: [https://arxiv.org/abs/2510.08338 LLMs Reproduce Human Purchase Intent via Semantic Similarity Elicitation of Likert Ratings]&lt;br /&gt;
&lt;br /&gt;
==Skeptical==&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06950 Large Language Models Do Not Simulate Human Psychology]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.16130 Replicating Human Motivated Reasoning Studies with LLMs]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [https://www.google.com/books/edition/_/cKnYEAAAQBAJ?hl=en&amp;amp;gbpv=1&amp;amp;pg=PA2 UNESCO. Guidance for Generative AI in Education and Research]&lt;br /&gt;
* [[AI]]&lt;br /&gt;
** [[AI predictions]]&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
	<entry>
		<id>http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8731</id>
		<title>Science Agents</title>
		<link rel="alternate" type="text/html" href="http://gisaxs.com/index.php?title=Science_Agents&amp;diff=8731"/>
		<updated>2026-03-16T15:56:05Z</updated>

		<summary type="html">&lt;p&gt;KevinYager: /* Commercial */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
=AI Use-cases for Science=&lt;br /&gt;
&lt;br /&gt;
==Literature==&lt;br /&gt;
* [https://www.alphaxiv.org/explore alphaXiv | Explore]: Understand arXiv papers&lt;br /&gt;
&lt;br /&gt;
===LLM extract data from papers===&lt;br /&gt;
* 2024-14: [https://pubs.rsc.org/en/content/articlelanding/2025/cs/d4cs00913d From text to insight: large language models for chemical data extraction]&lt;br /&gt;
&lt;br /&gt;
===AI finding links in literature===&lt;br /&gt;
* 2019-07: [https://doi.org/10.1038/s41586-019-1335-8  Unsupervised word embeddings capture latent knowledge from materials science literature]&lt;br /&gt;
* 2024-11: [https://doi.org/10.1038/s41562-024-02046-9  Large language models surpass human experts in predicting neuroscience results]&lt;br /&gt;
&lt;br /&gt;
===(Pre) Generate Articles===&lt;br /&gt;
* 2022-12: [https://aclanthology.org/2022.emnlp-main.296/ Re3: Generating Longer Stories With Recursive Reprompting and Revision]&lt;br /&gt;
* 2023-03: English essays: [https://journal.unnes.ac.id/sju/index.php/elt/article/view/64069 Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay]&lt;br /&gt;
* 2023-01: Journalism: [https://journals.sagepub.com/doi/10.1177/10776958221149577 Collaborating With ChatGPT: Considering the Implications of Generative Artificial Intelligence for Journalism and Media Education]&lt;br /&gt;
* 2023-07: Science writing: [https://www.rbmojournal.com/article/S1472-6483(23)00219-5/fulltext Artificial intelligence in scientific writing: a friend or a foe?]&lt;br /&gt;
* 2024-02: Wikipedia style: [https://arxiv.org/abs/2402.14207 Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models]&lt;br /&gt;
* 2024-02: [https://arxiv.org/abs/2408.07055 LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs] ([https://github.com/THUDM/LongWriter code])&lt;br /&gt;
* 2024-08: Scientific papers: [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery]&lt;br /&gt;
* 2024-09: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.18866 Reasoning to Learn from Latent Thoughts]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.19065 WikiAutoGen: Towards Multi-Modal Wikipedia-Style Article Generation]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.13171 Sleep-time Compute: Beyond Inference Scaling at Test-time]&lt;br /&gt;
&lt;br /&gt;
==Explanation==&lt;br /&gt;
* 2025-02: [https://tiger-ai-lab.github.io/TheoremExplainAgent/ TheoremExplainAgent: Towards Multimodal Explanations for LLM Theorem Understanding] ([https://arxiv.org/abs/2502.19400 preprint])&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.02822 Do Two AI Scientists Agree?]&lt;br /&gt;
&lt;br /&gt;
==Autonomous Ideation==&lt;br /&gt;
* 2024-04: [https://arxiv.org/abs/2404.07738 ResearchAgent: Iterative Research Idea Generation over Scientific Literature with Large Language Models]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.14202 Mining Causality: AI-Assisted Search for Instrumental Variables]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.07977 Thinking Fast and Laterally: Multi-Agentic Approach for Reasoning about Uncertain Emerging Events]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.14141 LLMs can realize combinatorial creativity: generating creative ideas via LLMs for scientific research]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13025 Agentic Deep Graph Reasoning Yields Self-Organizing Knowledge Networks]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.00794 Predicting Empirical AI Research Outcomes with Language Models]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.20803 The Ideation-Execution Gap: Execution Outcomes of LLM-Generated versus Human Research Ideas]&lt;br /&gt;
&lt;br /&gt;
==Adapting LLMs to Science==&lt;br /&gt;
* 2023-06: [https://doi.org/10.1039/D3DD00112A Domain-specific chatbots for science using embeddings]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2411.00027 Personalization of Large Language Models: A Survey]&lt;br /&gt;
* 2024-11: [https://arxiv.org/abs/2411.00412 Adapting While Learning: Grounding LLMs for Scientific Problems with Intelligent Tool Usage Adaptation]&lt;br /&gt;
&lt;br /&gt;
==AI/LLM Control of Scientific Instruments/Facilities==&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41524-024-01423-2 Opportunities for retrieval and tool augmented large language models in scientific facilities]&lt;br /&gt;
* 2023-12: [https://arxiv.org/abs/2312.17180 Virtual Scientific Companion for Synchrotron Beamlines: A Prototype]&lt;br /&gt;
* 2023-12: [https://www.nature.com/articles/s41586-023-06792-0 Autonomous chemical research with large language models]&lt;br /&gt;
* 2024-01: [https://iopscience.iop.org/article/10.1088/2632-2153/ad52e9 Synergizing Human Expertise and AI Efficiency with Language Model for Microscopy Operation and Automated Experiment Design]&lt;br /&gt;
* 2024-06: [https://pubs.rsc.org/en/content/articlelanding/2025/dd/d4dd00143e From Text to Test: AI-Generated Control Software for Materials Science Instruments]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.18161 VISION: A Modular AI Assistant for Natural Human-Instrument Interaction at Scientific User Facilities]&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/sciadv.adr4173 Large language models for human-machine collaborative particle accelerator tuning through natural language]&lt;br /&gt;
* 2025-04: [https://openreview.net/forum?id=iA9UN1dEgJ Operating Robotic Laboratories with Large Language Models and Teachable Agents]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods tailored to Science==&lt;br /&gt;
===Science Foundation Models===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.15763 Intern-S1: A Scientific Multimodal Foundation Model]&lt;br /&gt;
* 2025-11: [https://pubs.aip.org/aip/jcp/article/163/18/184110/3372267/A-foundation-model-for-atomistic-materials A foundation model for atomistic materials chemistry]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.15684 Walrus: A Cross-Domain Foundation Model for Continuum Dynamics]&lt;br /&gt;
* 2026-01: [https://www.science.org/doi/10.1126/science.ads9530 Deep contrastive learning enables genome-wide virtual screening]&lt;br /&gt;
&lt;br /&gt;
===Regression (Data Fitting)===&lt;br /&gt;
* 2024-06: [https://arxiv.org/abs/2406.14546 Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data]: training on (x,y) pairs enables inferring underlying function (define it in code, invert it, compose it)&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2402.14547 OmniPred: Language Models as Universal Regressors]&lt;br /&gt;
&lt;br /&gt;
===Tabular Classification/Regression===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08328-6 Accurate predictions on small data with a tabular foundation model] ([https://github.com/PriorLabs/TabPFN code])&lt;br /&gt;
&lt;br /&gt;
===Symbolic Regression===&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.09359 Symbolic Regression with a Learned Concept Library]&lt;br /&gt;
&lt;br /&gt;
===Literature Discovery===&lt;br /&gt;
* [https://www.futurehouse.org/ FutureHouse]&lt;br /&gt;
** [https://hasanyone.com/ hasanyone]&lt;br /&gt;
** [https://github.com/Future-House/paper-qa PaperQA2]&lt;br /&gt;
* [https://lumina.sh/ Lumina]&lt;br /&gt;
* [https://github.com/TheBlewish/Automated-AI-Web-Researcher-Ollama Automated-AI-Web-Researcher-Ollama]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.05366 Search-o1: Agentic Search-Enhanced Large Reasoning Models] ([https://search-o1.github.io/ project], [https://github.com/sunnynexus/Search-o1 code])&lt;br /&gt;
* 2026-02: [https://www.nature.com/articles/s41586-025-10072-4 Synthesizing scientific literature with retrieval-augmented language models] ([https://allenai.org/blog/openscholar-nature blog])&lt;br /&gt;
&lt;br /&gt;
===Commercial===&lt;br /&gt;
* [https://sakana.ai/ai-scientist/ Sakana AI] &lt;br /&gt;
* [https://www.cusp.ai/ Cusp AI]: Materials/AI&lt;br /&gt;
* [https://www.lila.ai/ Lila AI]: Life sciences&lt;br /&gt;
* [https://www.radical-ai.com/ Radical AI]: Material simulation/design&lt;br /&gt;
* [https://www.autoscience.ai/ Autoscience] ([https://www.autoscience.ai/blog/meet-carl-the-first-ai-system-to-produce-academically-peer-reviewed-research Carl])&lt;br /&gt;
* [https://periodic.com/ Periodic Labs]&lt;br /&gt;
* [https://edisonscientific.com/articles/announcing-edison-scientific Edison Scientific] (drug discovery, spinoff from [https://www.futurehouse.org/ FutureHouse])&lt;br /&gt;
* 2026-03: Mirendil Inc.: advanced models to speed up R&amp;amp;D in scientific domains, especially biology and materials science&lt;br /&gt;
&lt;br /&gt;
====Bio====&lt;br /&gt;
* [https://www.bioptimus.com/ Bioptimus]&lt;br /&gt;
* [https://www.evolutionaryscale.ai/ EvolutionaryScale]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods in Science==&lt;br /&gt;
* 2025-07: [https://www.mdpi.com/2313-433X/11/8/252 Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures]&lt;br /&gt;
&lt;br /&gt;
===Imaging===&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08176 Behind the Noise: Conformal Quantile Regression Reveals Emergent Representations] (blog: [https://phzwart.github.io/behindthenoise/ Behind the Noise])&lt;br /&gt;
&lt;br /&gt;
===Materials===&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.03965 All-atom Diffusion Transformers: Unified generative modelling of molecules and materials]&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2511.19730 Training-Free Active Learning Framework in Materials Science with Large Language Models]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41578-025-00772-8 Large language models for reticular chemistry]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00982-3 Image-based generation for molecule design with SketchMol]&lt;br /&gt;
* 2025-02: [https://www.nature.com/articles/s42256-025-00994-z Large language models for scientific discovery in molecular property prediction]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.08051 Compositional Flows for 3D Molecule and Synthesis Pathway Co-design]&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.07456 General purpose models for the chemical sciences]&lt;br /&gt;
* 2025-11: [https://chemrxiv.org/engage/chemrxiv/article-details/690357d9a482cba122e366b6 ChemTorch: A Deep Learning Framework for Benchmarking and Developing Chemical Reaction Property Prediction Models]&lt;br /&gt;
&lt;br /&gt;
===Biology===&lt;br /&gt;
* 2018: [https://alphafold.ebi.ac.uk/ AlphaFold]&lt;br /&gt;
* 2021-07: [https://www.nature.com/articles/s41586-021-03819-2 AlphaFold 2]&lt;br /&gt;
* 2024-05: [https://www.nature.com/articles/s41586-024-07487-w AlphaFold 3]&lt;br /&gt;
* 2023-03: [https://www.science.org/doi/10.1126/science.ade2574 Evolutionary-scale prediction of atomic-level protein structure with a language model] ([https://esmatlas.com/resources?action=fold ESMFold])&lt;br /&gt;
* 2023-11: [https://www.nature.com/articles/s41586-023-06728-8 Illuminating protein space with a programmable generative model]&lt;br /&gt;
* 2024-11: [https://www.science.org/doi/10.1126/science.ado9336 Sequence modeling and design from molecular to genome scale with Evo] (Evo)&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-024-08435-4 Targeting protein–ligand neosurfaces with a generalizable deep learning tool] (Chroma)&lt;br /&gt;
* 2025-01: [https://www.science.org/doi/10.1126/science.ads0018 Simulating 500 million years of evolution with a language model] ([https://github.com/evolutionaryscale/esm ESM] 3 model)&lt;br /&gt;
* 2025-02: [https://arcinstitute.org/manuscripts/Evo2 Genome modeling and design across all domains of life with Evo 2]&lt;br /&gt;
* 2025-02: [https://www.microsoft.com/en-us/research/blog/exploring-the-structural-changes-driving-protein-function-with-bioemu-1/ Exploring the structural changes driving protein function with BioEmu-1]&lt;br /&gt;
* 2025-02: [https://arxiv.org/pdf/2502.18449 Protein Large Language Models: A Comprehensive Survey]&lt;br /&gt;
* [https://x.com/vant_ai/status/1903070297991110657 2025-03]: [https://www.vant.ai/ Vant AI] [https://www.vant.ai/neo-1 Neo-1]: atomistic foundation model (small molecules, proteins, etc.)&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.16351 Lyra: An Efficient and Expressive Subquadratic Architecture for Modeling Biological Sequences]&lt;br /&gt;
* 2025-08: RosettaFold 3: [https://www.biorxiv.org/content/10.1101/2025.08.14.670328v2 Accelerating Biomolecular Modeling with AtomWorks and RF3]&lt;br /&gt;
* 2025-09: [https://www.biorxiv.org/content/10.1101/2025.09.12.675911v1 Generative design of novel bacteriophages with genome language models]&lt;br /&gt;
* 2025-10: [https://www.science.org/doi/10.1126/science.adu8578 Strengthening nucleic acid biosecurity screening against generative protein design tools]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s41586-025-10014-0 Advancing regulatory variant effect prediction with AlphaGenome]&lt;br /&gt;
&lt;br /&gt;
===Medicine===&lt;br /&gt;
See: [[AI_Agents#Medicine]]&lt;br /&gt;
&lt;br /&gt;
===Successes===&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.11270 Site-Decorated Model for Unconventional Frustrated Magnets: Ultranarrow Phase Crossover and Spin Reversal Transition]&lt;br /&gt;
&lt;br /&gt;
==AI/ML Methods co-opted for Science==&lt;br /&gt;
===Mechanistic Interpretability===&lt;br /&gt;
Train large model on science data. Then apply [[AI_understanding#Mechanistic_Interpretability|mechanistic interpretability]] (e.g. sparse autoencoders, SAE) to the feature/activation space.&lt;br /&gt;
* Mechanistic interpretability for protein language models ([https://interprot.com/ visualizer], [https://github.com/etowahadams/interprot/tree/main code], [https://huggingface.co/liambai/InterProt-ESM2-SAEs SAE])&lt;br /&gt;
* [https://www.markov.bio/ Markov Bio]: [https://www.markov.bio/research/mech-interp-path-to-e2e-biology Through a Glass Darkly: Mechanistic Interpretability as the Bridge to End-to-End Biology] ([https://x.com/adamlewisgreen/status/1853206279499751531 quick description], [https://markovbio.github.io/biomedical-progress/ background info on recent bio progress])&lt;br /&gt;
* 2023-01: [https://arxiv.org/abs/2301.05062 Tracr: Compiled Transformers as a Laboratory for Interpretability] ([https://github.com/google-deepmind/tracr code])&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.03334 An X-Ray Is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation]&lt;br /&gt;
* 2024-12: [https://www.arxiv.org/abs/2412.16247 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models]&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.12101 InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.00089 Insights on Galaxy Evolution from Interpretable Sparse Feature Networks]&lt;br /&gt;
* 2025-02: [https://www.biorxiv.org/content/10.1101/2025.02.06.636901v1 From Mechanistic Interpretability to Mechanistic Biology: Training, Evaluating, and Interpreting Sparse Autoencoders on Protein Language Models]&lt;br /&gt;
* 2025-02: [https://www.goodfire.ai/blog/interpreting-evo-2 Interpreting Evo 2: Arc Institute&amp;#039;s Next-Generation Genomic Foundation Model]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
&lt;br /&gt;
===Uncertainty===&lt;br /&gt;
* 2024-10: [https://github.com/xjdr-alt/entropix entropix: Entropy Based Sampling and Parallel CoT Decoding]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.09724 Taming Overconfidence in LLMs: Reward Calibration in RLHF]&lt;br /&gt;
&lt;br /&gt;
=Science Benchmarks=&lt;br /&gt;
* 2024-07: [https://arxiv.org/abs/2407.13168 SciCode: A Research Coding Benchmark Curated by Scientists] ([http://scicode-bench.github.io/ project])&lt;br /&gt;
* 2024-11: [https://openreview.net/pdf?id=fz969ahcvJ AidanBench: Evaluating Novel Idea Generation on Open-Ended Questions] ([https://github.com/aidanmclaughlin/AidanBench code])&lt;br /&gt;
* 2024-12: [https://arxiv.org/abs/2412.17596 LiveIdeaBench: Evaluating LLMs&amp;#039; Scientific Creativity and Idea Generation with Minimal Context]&lt;br /&gt;
* 2025-01: [https://agi.safe.ai/ Humanity&amp;#039;s Last Exam]&lt;br /&gt;
* [https://github.com/OSU-NLP-Group/ScienceAgentBench ScienceAgentBench]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.20309 EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants]&lt;br /&gt;
* 2025-03: [https://huggingface.co/datasets/futurehouse/BixBench BixBench]: Novel hypotheses (accept/reject)&lt;br /&gt;
* 2025-04: [https://research.google/blog/evaluating-progress-of-llms-on-scientific-problem-solving/ Google: Evaluating progress of LLMs on scientific problem-solving]&lt;br /&gt;
** 2025-03: [https://arxiv.org/abs/2503.13517 CURIE: Evaluating LLMs On Multitask Scientific Long Context Understanding and Reasoning]&lt;br /&gt;
** 2024-07: [https://arxiv.org/abs/2407.09413 SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers]&lt;br /&gt;
** 2024-10: [https://neurips.cc/virtual/2024/98540 FEABench: Evaluating Language Models on Real World Physics Reasoning Ability]&lt;br /&gt;
* 2026-02: [https://edisonscientific.com/ Edison]: [https://lab-bench.ai/ LABBench 2]&lt;br /&gt;
&lt;br /&gt;
=Science Agents=&lt;br /&gt;
==Reviews==&lt;br /&gt;
* 2024-10: [https://www.cell.com/cell/fulltext/S0092-8674(24)01070-5?target=_blank Empowering biomedical discovery with AI agents]&lt;br /&gt;
* 2025-01: [https://pubs.rsc.org/en/content/articlehtml/2024/sc/d4sc03921a A review of large language models and autonomous agents in chemistry] ([https://github.com/ur-whitelab/LLMs-in-science github])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01903 AI4Research: A Survey of Artificial Intelligence for Scientific Research]&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.14111 From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery]&lt;br /&gt;
&lt;br /&gt;
==Challenges==&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.03315 Why LLMs Aren&amp;#039;t Scientists Yet: Lessons from Four Autonomous Research Attempts]&lt;br /&gt;
&lt;br /&gt;
==Specific==&lt;br /&gt;
* 2024-01-13: [https://arxiv.org/abs/2401.06949 ORGANA: A Robotic Assistant for Automated Chemistry Experimentation and Characterization] ([https://www.youtube.com/watch?v=N6qMMwJ8hKQ video])&lt;br /&gt;
* 2024-06-19: [https://arxiv.org/abs/2406.13163 LLMatDesign: Autonomous Materials Discovery with Large Language Models]&lt;br /&gt;
* 2024-08-12: [https://sakana.ai/ Sakana AI]: [https://sakana.ai/ai-scientist/ AI Scientist]; [https://arxiv.org/abs/2408.06292 The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery] ([https://github.com/SakanaAI/AI-Scientist code])&lt;br /&gt;
* 2024-09-09: [https://arxiv.org/abs/2409.05556 SciAgents: Automating scientific discovery through multi-agent intelligent graph reasoning] ([https://github.com/lamm-mit/SciAgentsDiscovery code])&lt;br /&gt;
* 2024-09-11: PaperQA2: [https://paper.wikicrow.ai/ Language Models Achieve Superhuman Synthesis of Scientific Knowledge] ([https://x.com/SGRodriques/status/1833908643856818443 𝕏 post], [https://github.com/Future-House/paper-qa code])&lt;br /&gt;
* 2024-10-17: [https://arxiv.org/abs/2410.13768 Rapid and Automated Alloy Design with Graph Neural Network-Powered LLM-Driven Multi-Agent Systems]&lt;br /&gt;
* 2024-10-28: [https://arxiv.org/abs/2410.20976 Large Language Model-Guided Prediction Toward Quantum Materials Synthesis]&lt;br /&gt;
* 2024-12-06: [https://www.biorxiv.org/content/10.1101/2024.11.11.623004v1 The Virtual Lab: AI Agents Design New SARS-CoV-2 Nanobodies with Experimental Validation] (writeup: [https://www.nature.com/articles/d41586-024-01684-3 Virtual lab powered by ‘AI scientists’ super-charges biomedical research: Could human–AI collaborations be the future of interdisciplinary studies?])&lt;br /&gt;
* 2024-12-30: [https://arxiv.org/abs/2412.21154 Aviary: training language agents on challenging scientific tasks]&lt;br /&gt;
* See also: [[AI_Agents#Deep_Research|AI Agents &amp;gt; Deep Research]]&lt;br /&gt;
* 2025-04-08: Sakana: [https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf The AI Scientist-v2: Workshop-Level Automated Scientific Discovery via Agentic Tree Search] ([https://github.com/SakanaAI/AI-Scientist-v2 code])&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.14267 DREAMS: Density Functional Theory Based Research Engine for Agentic Materials Simulation]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.08151 SciAgent: A Unified Multi-Agent System for Generalistic Scientific Reasoning]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2601.23265 PaperBanana: Automating Academic Illustration for AI Scientists]&lt;br /&gt;
&lt;br /&gt;
==Skills==&lt;br /&gt;
* 2026-03: [https://github.com/K-Dense-AI/claude-scientific-skills/tree/main?tab=readme-ov-file#use-cases Claude Scientific Skills] (list)&lt;br /&gt;
&lt;br /&gt;
==Science Multi-Agent Setups==&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.04227 Agent Laboratory: Using LLM Agents as Research Assistants]&lt;br /&gt;
* 2025-04: [https://www.nature.com/articles/s41551-025-01363-2 Coordinated AI agents for advancing healthcare] ([https://www.nature.com/articles/s41551-025-01363-2.epdf?sharing_token=CIYP3J8LZE4BX31fV3WxUdRgN0jAjWel9jnR3ZoTv0O9iD-yhgqzRaz_7VASayWRePPhWDD2xFyfuOpSXbdPaOtt7oH4nfXo7telALzNwY3V1p9SxoqBEJy2OuaJ_cA35-CYQC1XgjCNTZUw46dh1KX-Dj8e7-1Vk_RlZKFLrc8%3D pdf])&lt;br /&gt;
&lt;br /&gt;
=AI Science Systems=&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.03916 Dolphin: Closed-loop Open-ended Auto-research through Thinking, Practice, and Feedback]&lt;br /&gt;
* 2025-01: [https://arxiv.org/abs/2501.13299 Hypothesis Generation for Materials Discovery and Design Using Goal-Driven and Constraint-Guided LLM Agents]&lt;br /&gt;
* 2025-02: [https://storage.googleapis.com/coscientist_paper/ai_coscientist.pdf Towards an AI co-scientist] (Google blog post: [https://research.google/blog/accelerating-scientific-breakthroughs-with-an-ai-co-scientist/ Accelerating scientific breakthroughs with an AI co-scientist])&lt;br /&gt;
* 2025-06: [https://zenodo.org/records/15693353 The Discovery Engine]&lt;br /&gt;
** 2025-07: [https://arxiv.org/abs/2507.00964 Benchmarking the Discovery Engine] ([https://www.leap-labs.com/blog/how-we-replicated-five-peer-reviewed-papers-in-five-hours blog])&lt;br /&gt;
* 2025-07: [https://www.preprints.org/manuscript/202507.1951/v1 Autonomous Scientific Discovery Through Hierarchical AI Scientist Systems]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.16969 Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows]&lt;br /&gt;
* 2026-01: [https://www.nature.com/articles/s43588-025-00906-6 SciSciGPT: advancing human–AI collaboration in the science of science]&lt;br /&gt;
* 2026-02: [https://allenai.org/papers/autodiscovery AUTODISCOVERY: Open-ended Scientific Discovery via Bayesian Surprise] (Allen AI (Ai2) AstraLabs, [https://allenai.org/blog/autodiscovery blog], [https://autodiscovery.allen.ai/runs tools])&lt;br /&gt;
&lt;br /&gt;
===Inorganic Materials Discovery===&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06734-w An autonomous laboratory for the accelerated synthesis of novel materials]&lt;br /&gt;
* 2024-09: [https://arxiv.org/abs/2409.00135 HoneyComb: A Flexible LLM-Based Agent System for Materials Science]&lt;br /&gt;
* 2024-10: [https://arxiv.org/abs/2410.12771 Open Materials 2024 (OMat24) Inorganic Materials Dataset and Models] ([https://github.com/FAIR-Chem/fairchem code], [https://huggingface.co/datasets/fairchem/OMAT24 datasets], [https://huggingface.co/fairchem/OMAT24 checkpoints], [https://ai.meta.com/blog/fair-news-segment-anything-2-1-meta-spirit-lm-layer-skip-salsa-sona/ blogpost])&lt;br /&gt;
* 2025-01: [https://www.nature.com/articles/s41586-025-08628-5 A generative model for inorganic materials design]&lt;br /&gt;
* 2025-04: [https://arxiv.org/abs/2504.14110 System of Agentic AI for the Discovery of Metal-Organic Frameworks]&lt;br /&gt;
* 2025-05: [https://arxiv.org/abs/2505.08762 The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models]&lt;br /&gt;
&lt;br /&gt;
===Materials Characterization===&lt;br /&gt;
* 2025-08: [https://arxiv.org/abs/2508.06569 Operationalizing Serendipity: Multi-Agent AI Workflows for Enhanced Materials Characterization with Theory-in-the-Loop]&lt;br /&gt;
&lt;br /&gt;
===Chemistry===&lt;br /&gt;
* 2023-12: [https://doi.org/10.1038/s41586-023-06792-0 Autonomous chemical research with large language models] (Coscientist)&lt;br /&gt;
* 2024-09: [https://www.pnnl.gov/main/publications/external/technical_reports/PNNL-36692.pdf PNNL ChemAIst V0.2]&lt;br /&gt;
* 2024-11: [https://www.nature.com/articles/s41467-024-54457-x An automatic end-to-end chemical synthesis development platform powered by large language models]&lt;br /&gt;
* 2025-06: [https://paper.ether0.ai/ Training a Scientific Reasoning Model for Chemistry]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.06363 ChemGraph: An Agentic Framework for Computational Chemistry Workflows] ([https://github.com/argonne-lcf/ChemGraph code])&lt;br /&gt;
&lt;br /&gt;
===Bio===&lt;br /&gt;
* 2025-07: [https://arxiv.org/abs/2507.01485 BioMARS: A Multi-Agent Robotic System for Autonomous Biological Experiments]&lt;br /&gt;
&lt;br /&gt;
===Physics===&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.19799 PhysMaster: Building an Autonomous AI Physicist for Theoretical and Computational Physics Research]&lt;br /&gt;
&lt;br /&gt;
==LLMs Optimized for Science==&lt;br /&gt;
* 2022-11: [https://arxiv.org/abs/2211.09085 Galactica: A Large Language Model for Science]&lt;br /&gt;
* 2024-12: [https://www.nature.com/articles/s41467-024-54639-7 Crystal structure generation with autoregressive large language modeling]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.13107 MatterChat: A Multi-Modal LLM for Material Science]&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.17604 OmniScience: A Domain-Specialized LLM for Scientific Reasoning and Discovery]&lt;br /&gt;
* 2025-03: Google [https://huggingface.co/collections/google/txgemma-release-67dd92e931c857d15e4d1e87 TxGemma] (2B, 9B, 27B): [https://developers.googleblog.com/en/introducing-txgemma-open-models-improving-therapeutics-development/ drug development]&lt;br /&gt;
&lt;br /&gt;
=Impact of AI in Science=&lt;br /&gt;
* 2024-11: &amp;lt;strike&amp;gt;[https://aidantr.github.io/files/AI_innovation.pdf Artificial Intelligence, Scientific Discovery, and Product Innovation]&amp;lt;/strike&amp;gt;&lt;br /&gt;
** 2025-05: Retraction: [https://economics.mit.edu/news/assuring-accurate-research-record Assuring an accurate research record]&lt;br /&gt;
* 2025-02: [https://arxiv.org/abs/2502.05151 Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.03837 Accelerating Scientific Research with Gemini: Case Studies and Common Techniques]&lt;br /&gt;
&lt;br /&gt;
=Related Tools=&lt;br /&gt;
==Literature Search==&lt;br /&gt;
* [https://www.perplexity.ai/ Perplexity]&lt;br /&gt;
* [https://www.arxival.xyz/ ArXival]&lt;br /&gt;
&lt;br /&gt;
==Data Visualization==&lt;br /&gt;
* 2024-10: Microsoft [https://www.microsoft.com/en-us/research/blog/data-formulator-exploring-how-ai-can-help-analysts-create-rich-data-visualizations/ Data Formulator: Create Rich Visualization with AI iteratively] ([https://www.microsoft.com/en-us/research/video/data-formulator-create-rich-visualization-with-ai-iteratively/ video], [https://github.com/microsoft/data-formulator code])&lt;br /&gt;
* [https://julius.ai/ Julius AI]: Analyze your data with computational AI&lt;br /&gt;
&lt;br /&gt;
==Generative==&lt;br /&gt;
* 2025-03: [https://huggingface.co/collections/starvector/starvector-models-6783b22c7bd4b43d13cb5289 StarVector] 1B, 8B: text or image to SVG&lt;br /&gt;
&lt;br /&gt;
==Chemistry==&lt;br /&gt;
* 2025-03: [https://jcheminf.biomedcentral.com/articles/10.1186/s13321-024-00834-z Rxn-INSIGHT: fast chemical reaction analysis using bond-electron matrices] ([https://rxn-insight.readthedocs.io/en/latest/ docs])&lt;br /&gt;
&lt;br /&gt;
=Science Datasets=&lt;br /&gt;
* [https://datasetsearch.research.google.com/ Google Dataset Search]&lt;br /&gt;
* [https://github.com/blaiszik/awesome-matchem-datasets/ Awesome Materials &amp;amp; Chemistry Datasets]&lt;br /&gt;
* NIST [https://jarvis.nist.gov/ Jarvis] (simulations)&lt;br /&gt;
&lt;br /&gt;
=Genuine Discoveries=&lt;br /&gt;
* 2025-11: [https://cdn.openai.com/pdf/4a25f921-e4e0-479a-9b38-5367b47e8fd0/early-science-acceleration-experiments-with-gpt-5.pdf Early science acceleration experiments with GPT-5]&lt;br /&gt;
* 2025-12: [https://andymasley.substack.com/p/ai-can-obviously-create-new-knowledge AI can obviously create new knowledge - But maybe not new concepts]&lt;br /&gt;
==Math==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06004-9?utm_source=chatgpt.com Faster sorting algorithms discovered using deep reinforcement learning]&lt;br /&gt;
* 2025-06: [https://arxiv.org/abs/2506.13131 AlphaEvolve: A coding agent for scientific and algorithmic discovery]&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02864 Mathematical exploration and discovery at scale]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09833-y Olympiad-level formal mathematical reasoning with reinforcement learning]&lt;br /&gt;
* 2025-12: [https://arxiv.org/abs/2512.14575 Extremal descendant integrals on moduli spaces of curves: An inequality discovered and proved in collaboration with AI]&lt;br /&gt;
* [https://github.com/teorth/erdosproblems/wiki/AI-contributions-to-Erd%C5%91s-problems AI Solving Erdős Problems]:&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/728 Erdős Problem #728] and [https://www.erdosproblems.com/729 #729] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/397 Erdős Problem #397] [https://x.com/neelsomani/status/2010215162146607128?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/205 Erdős Problem #205] solved by Aristotle using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: [https://www.erdosproblems.com/forum/thread/281 Erdős Problem #281] [https://x.com/neelsomani/status/2012695714187325745?s=20 solved] by [https://neelsomani.com/ Neel Somani] using ChatGPT 5.2 Pro&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.21442 Irrationality of rapidly converging series: a problem of Erdős and Graham]&lt;br /&gt;
*** [https://www.erdosproblems.com/1051 Erdős Problem #1051] [https://x.com/slow_developer/status/2018321002623901885?s=20 solved] by Google DeepMind Aletheia agent&lt;br /&gt;
** 2026-01: Google DeepMind: [https://arxiv.org/abs/2601.22401 Semi-Autonomous Mathematics Discovery with Gemini: A Case Study on the Erdős Problems]&lt;br /&gt;
*** Attempted 700 problems, solved 13 open Erdős problems: 5 novel autonomous solutions, 8 through existing literature.&lt;br /&gt;
** 2026-02: [https://www.erdosproblems.com/846 Erdős Problem #846]&lt;br /&gt;
*** [https://x.com/roydanroy/status/2026804567178953048?s=20 Google DeepMind]&lt;br /&gt;
*** [https://x.com/mehtaab_sawhney/status/2026716221933343147?s=20 Using OpenAI internal model] (paper: [https://cdn.openai.com/infinite-sets/main_single_clean3.pdf On infinite sets with no 3 on a line])&lt;br /&gt;
* 2026-01: [https://arxiv.org/abs/2601.07222 The motivic class of the space of genus 0 maps to the flag variety]&lt;br /&gt;
* 2026-02: Google DeepMind: [https://arxiv.org/abs/2602.10177 Towards Autonomous Mathematics Research]&lt;br /&gt;
* 2026-03: Donald Knuth: [https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf A problem in Directed Hamiltonian Cycles] solved by Filip Stappers using Claude Opus 4.6&lt;br /&gt;
* 2026-03: Google DeepMind: [https://arxiv.org/abs/2603.09172 Reinforced Generation of Combinatorial Structures: Ramsey Numbers]&lt;br /&gt;
&lt;br /&gt;
==Physics assistance==&lt;br /&gt;
* 2025-03: [https://arxiv.org/abs/2503.23758 Exact solution of the frustrated Potts model with next-nearest-neighbor interactions in one dimension via AI bootstrapping] ([https://www.bnl.gov/staff/wyin Weiguo Yin])&lt;br /&gt;
* 2025-12: [https://www.sciencedirect.com/science/article/pii/S0370269325008111 Relativistic covariance and nonlinear quantum mechanics: Tomonaga-Schwinger analysis]&lt;br /&gt;
** [https://x.com/hsu_steve/status/1996034522308026435?s=20 Steve Hsu], [https://drive.google.com/file/d/16sxJuwsHoi-fvTFbri9Bu8B9bqA6lr1H/view Theoretical Physics with Generative AI]&lt;br /&gt;
* 2026-02: [https://arxiv.org/abs/2602.12176 Single-minus gluon tree amplitudes are nonzero] (GPT-5.2, [https://openai.com/index/new-result-theoretical-physics/ blog])&lt;br /&gt;
&lt;br /&gt;
==Literature exploration==&lt;br /&gt;
* 2025-11: [https://arxiv.org/abs/2511.02824 Kosmos: An AI Scientist for Autonomous Discovery] ([https://edisonscientific.com/ Edison])&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c4bdef64-5e9b-43b9-a365-592dd1ed7587 Nucleotide metabolism in hypothermia]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/1fdbf827-be65-4d97-9b66-bf0da600091a Determinant of perovskite solar-cell failure]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/4fb3fbdb-c449-4064-9aa6-ff4ec53131d8 Log-normal connectivity in neural networks]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/c6849232-5858-4634-adf5-83780afbe3db SOD2 as driver of myocardial fibrosis]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/abac07da-a6bb-458f-b0ba-ef08f1be617e Protective variant of SSR1 in type 2 diabetes]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/a770052b-2334-4bbe-b086-5149e0f03d99 Temporal ordering in Alzheimer’s disease]&lt;br /&gt;
** [https://platform.edisonscientific.com/kosmos/28c427d2-be31-48b5-b272-28d5a1e3ea5c Mechanism of neuron vulnerability in aging]&lt;br /&gt;
==Bio design==&lt;br /&gt;
* 2023-07: [https://www.nature.com/articles/s41586-023-06415-8 De novo design of protein structure and function with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://www.nature.com/articles/s41586-025-09721-5 Atomically accurate de novo design of antibodies with RFdiffusion]&lt;br /&gt;
* 2025-11: [https://deepmind.google/blog/alphafold-five-years-of-impact/ AlphaFold: Five years of impact]&lt;br /&gt;
* 2026-01: [https://www.goodfire.ai/research/interpretability-for-alzheimers-detection# Using Interpretability to Identify a Novel Class of Alzheimer&amp;#039;s Biomarkers]&lt;br /&gt;
==Material Discovery==&lt;br /&gt;
* 2023-11: [https://doi.org/10.1038/s41586-023-06735-9 Scaling deep learning for materials discovery]&lt;br /&gt;
&lt;br /&gt;
=See Also=&lt;br /&gt;
* [[AI agents]]&lt;br /&gt;
* [https://nanobot.chat/ Nanobot.chat]: Intelligent AI for the labnetwork @ mtl.mit.edu forum&lt;/div&gt;</summary>
		<author><name>KevinYager</name></author>
		
	</entry>
</feed>