Difference between revisions of "AI research trends"

From GISAXS
Jump to: navigation, search
(Neural (non-token) Latent Representation)
 
(One intermediate revision by the same user not shown)
Line 46: Line 46:
 
* 2024-03: [https://arxiv.org/abs/2403.11901 Larimar: Large Language Models with Episodic Memory Control]
 
* 2024-03: [https://arxiv.org/abs/2403.11901 Larimar: Large Language Models with Episodic Memory Control]
  
=Neural (non-token) Latent Representation=
+
=Updating Weights at Inference-time=
 +
* 2025-01: [https://arxiv.org/abs/2501.06252 Transformer<sup>2</sup>: Self-adaptive LLMs]
 +
 
 +
=Internal Though Representation Space=
 +
==Visual Thinking==
 +
* 2025-01: [https://arxiv.org/abs/2501.05874 VideoRAG: Retrieval-Augmented Generation over Video Corpus]
 +
* 2025-01: [https://arxiv.org/abs/2501.07542 Imagine while Reasoning in Space: Multimodal Visualization-of-Thought]
 +
 
 +
==Neural (non-token) Latent Representation==
 
* 2024-11: Microsoft: [https://arxiv.org/abs/2411.02820 DroidSpeak: KV Cache Sharing for Cross-LLM Communication and Multi-LLM Serving]: LLMs invent their own inter-communication language
 
* 2024-11: Microsoft: [https://arxiv.org/abs/2411.02820 DroidSpeak: KV Cache Sharing for Cross-LLM Communication and Multi-LLM Serving]: LLMs invent their own inter-communication language
 
* 2024-12: Meta: [https://arxiv.org/abs/2412.06769 Training Large Language Models to Reason in a Continuous Latent Space]: feeding the latent representation directly back into the model, instead of tokenizing intermediate thoughts (Chain of Continuous Thought, a.k.a. Coconut)
 
* 2024-12: Meta: [https://arxiv.org/abs/2412.06769 Training Large Language Models to Reason in a Continuous Latent Space]: feeding the latent representation directly back into the model, instead of tokenizing intermediate thoughts (Chain of Continuous Thought, a.k.a. Coconut)

Latest revision as of 09:14, 15 January 2025

Novel Tokenization and/or Sampling

System 2 Reasoning

See: Increasing AI Intelligence

Memory

LLM Weights Memory

Context Length

Extended Context

Retrieval beyond RAG

See also: AI tools: Retrieval Augmented Generation (RAG)

Working Memory

Episodic Memory

Updating Weights at Inference-time

Internal Though Representation Space

Visual Thinking

Neural (non-token) Latent Representation

See Also