Difference between revisions of "AI research trends"

From GISAXS
Jump to: navigation, search
(Context Length)
(Memory)
Line 24: Line 24:
 
* 2024-Apr-12: Meta et al. demonstrate [https://arxiv.org/abs/2404.08801 Megalodon] that enables infinite context via a more efficient architecture
 
* 2024-Apr-12: Meta et al. demonstrate [https://arxiv.org/abs/2404.08801 Megalodon] that enables infinite context via a more efficient architecture
 
* 2024-Apr-14: Google presents [https://arxiv.org/abs/2404.09173 TransformerFAM], which leverages a feedback loop so it attends to its own latent representations, acting as working memory and provides effectively infinite context
 
* 2024-Apr-14: Google presents [https://arxiv.org/abs/2404.09173 TransformerFAM], which leverages a feedback loop so it attends to its own latent representations, acting as working memory and provides effectively infinite context
 +
 +
==Retrieval beyond RAG==
 +
* 2024-12: [https://arxiv.org/abs/2412.11536 Let your LLM generate a few tokens and you will reduce the need for retrieval]
 +
* 2024-12: [https://arxiv.org/abs/2412.11919 RetroLLM: Empowering Large Language Models to Retrieve Fine-grained Evidence within Generation]
  
 
==Working Memory==
 
==Working Memory==

Revision as of 09:17, 28 December 2024

Novel Tokenization and/or Sampling

System 2 Reasoning

See: Increasing AI Agent Intelligence

Memory

Context Length

Retrieval beyond RAG

Working Memory

Episodic Memory

Neural (non-token) Latent Representation