Difference between revisions of "AI tricks"

From GISAXS
Jump to: navigation, search
(Prompt Engineering)
(Position Bias)
 
(3 intermediate revisions by the same user not shown)
Line 4: Line 4:
  
 
==In-Context Learning==
 
==In-Context Learning==
* [https://arxiv.org/abs/2005.14165 Language Models are Few-Shot Learners]
+
* 2020-05: [https://arxiv.org/abs/2005.14165 Language Models are Few-Shot Learners]
 +
* 2025-03: [https://arxiv.org/abs/2503.08030 Learning to Search Effective Example Sequences for In-Context Learning]
  
 
==Chain of Thought (CoT)==
 
==Chain of Thought (CoT)==
Line 16: Line 17:
 
==Retrieval-Augmented Generation (RAG)==
 
==Retrieval-Augmented Generation (RAG)==
 
* [https://arxiv.org/abs/2005.11401 Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks]
 
* [https://arxiv.org/abs/2005.11401 Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks]
 +
 +
==Input/Output Formats==
 +
* 2024-08: [https://arxiv.org/abs/2408.08656 LLMs Are Biased Towards Output Formats! Systematically Evaluating and Mitigating Output Format Bias of LLMs]
 +
* 2024-11: [https://arxiv.org/abs/2411.10541 Does Prompt Formatting Have Any Impact on LLM Performance?]
 +
 +
==Position Bias==
 +
* 2023-07: [https://arxiv.org/abs/2307.03172 Lost in the Middle: How Language Models Use Long Contexts]
 +
* 2024-11: [https://arxiv.org/abs/2411.01101 Self-Consistency Falls Short! The Adverse Effects of Positional Bias on Long-Context Problems]
 +
* 2025-02: [https://arxiv.org/abs/2502.01951 On the Emergence of Position Bias in Transformers]
 +
* '''Testing models:'''
 +
** [https://github.com/gkamradt/LLMTest_NeedleInAHaystack?utm_source=chatgpt.com Needle-in-a-Haystack tests]
 +
** 2023-08: [https://arxiv.org/abs/2308.14508 LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding]
 +
** 2024-02: [https://arxiv.org/abs/2402.13718 ∞Bench: Extending Long Context Evaluation Beyond 100K Tokens]
 +
** 2024-06: [Multimodal Needle in a Haystack: Benchmarking Long-Context Capability of Multimodal Large Language Models https://arxiv.org/abs/2406.11230]
 +
** 2024-07: [https://arxiv.org/abs/2407.16695 Stress-Testing Long-Context Language Models with Lifelong ICL and Task Haystack]
 +
** 2025-04: [https://arxiv.org/abs/2504.04150 Reasoning on Multiple Needles In A Haystack]
 +
* '''Mitigation:'''
 +
** 2023-10: [https://arxiv.org/abs/2310.01427 Attention Sorting Combats Recency Bias In Long Context Language Models]
 +
** 2024-07: [https://arxiv.org/abs/2407.01100 Eliminating Position Bias of Language Models: A Mechanistic Approach]
  
 
=Generation=
 
=Generation=
 
* [https://github.com/Zhen-Tan-dmml/LLM4Annotation Large Language Models for Data Annotation and Synthesis: A Survey]
 
* [https://github.com/Zhen-Tan-dmml/LLM4Annotation Large Language Models for Data Annotation and Synthesis: A Survey]

Latest revision as of 08:16, 8 May 2025

Prompt Engineering

In-Context Learning

Chain of Thought (CoT)

"Let's think step-by-step"

Multi-step

Tool-use, feedback, agentic

Retrieval-Augmented Generation (RAG)

Input/Output Formats

Position Bias

Generation