Difference between revisions of "AI research trends"

From GISAXS
Jump to: navigation, search
(Context Length)
(Context Length)
 
Line 49: Line 49:
 
* 2025-12-04: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans/Miras] 10M
 
* 2025-12-04: Google [https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/ Titans/Miras] 10M
 
* 2025-12-13: [https://arxiv.org/abs/2512.12167 Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings]
 
* 2025-12-13: [https://arxiv.org/abs/2512.12167 Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings]
* 2026-03-18: [https://github.com/EverMind-AI/MSA MSA: Memory Sparse Attention] 100M
+
* 2026-03-18: [https://github.com/EverMind-AI/MSA/blob/main/paper/MSA__Memory_Sparse_Attention_for_Efficient_End_to_End_Memory_Model_Scaling_to_100M_Tokens.pdf MSA: Memory Sparse Attention for Efficient End-to-End Memory Model Scaling to 100M Tokens] ([https://github.com/EverMind-AI/MSA code]) 100M
  
 
==Extended Context==
 
==Extended Context==

Latest revision as of 13:11, 23 March 2026

System 2 Reasoning

See: Increasing AI Intelligence

Memory

Reviews

Big Ideas

LLM Weights Memory

Context Length

Extended Context

Context Remaking

Retrieval beyond RAG

See also: AI tools: Retrieval Augmented Generation (RAG)

Working Memory

Long-Term Memory

Storage and Retrieval

Episodic Memory

Continual Learning

Updating Weights at Inference-time

Parameters as Tokens

Internal Thought Representation Space

Visual Thinking

Neural (non-token) Latent Representation

Altered Transformer

Tokenization

Generation Order

Diffusion Language Models

Related: Image Synthesis via Autoregression/Diffusion

Sampling

Daydreaming, brainstorming, pre-generation

Pre-generation

Missing Elements

  • Memory
  • Continuous learning/update
  • Robust contextual model
  • Long-time-horizon coherence
  • Fluid intelligence
  • Agency
  • Modeling of self
  • Daydreaming

Memes

See Also