Difference between revisions of "AI tutorials"

From GISAXS
Jump to: navigation, search
(LLM)
(Transformer)
 
(8 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
==General==
 
==General==
 
* [https://mlu-explain.github.io/ MLU-EXPLAIN]
 
* [https://mlu-explain.github.io/ MLU-EXPLAIN]
 +
* [https://arxiv.org/abs/2503.02113 Deep Learning is Not So Mysterious or Different]
 +
* [https://theaidigest.org/ AI Digest]: Interactive AI explainers
 +
* [https://academy.openai.com/ OpenAI Academy]
 +
* [https://goyalpramod.github.io/blogs/evolution_of_LLMs/ Evolution of LLMs]
  
 
==Loss Functions==
 
==Loss Functions==
Line 9: Line 13:
 
* [https://jalammar.github.io/illustrated-transformer/ The Illustrated Transformer]
 
* [https://jalammar.github.io/illustrated-transformer/ The Illustrated Transformer]
 
* [https://towardsdatascience.com/transformers-explained-visually-not-just-how-but-why-they-work-so-well-d840bd61a9d3 Transformers Explained Visually — Not Just How, but Why They Work So Well]
 
* [https://towardsdatascience.com/transformers-explained-visually-not-just-how-but-why-they-work-so-well-d840bd61a9d3 Transformers Explained Visually — Not Just How, but Why They Work So Well]
 +
* [https://erdem.pl/2021/05/understanding-positional-encoding-in-transformers#positional-encoding-visualization Understanding Positional Encoding in Transformers]
 +
 
===Visualizations===
 
===Visualizations===
 
* [https://bbycroft.net/llm LLM Visualization]
 
* [https://bbycroft.net/llm LLM Visualization]
Line 16: Line 22:
  
 
==LLM==
 
==LLM==
 +
* [https://transformer-circuits.pub/2022/toy_model/index.html Toy Models of Superposition]
 
* [https://www.techrxiv.org/doi/full/10.36227/techrxiv.23589741.v1 A Survey on Large Language Models: Applications, Challenges, Limitations, and Practical Usage]
 
* [https://www.techrxiv.org/doi/full/10.36227/techrxiv.23589741.v1 A Survey on Large Language Models: Applications, Challenges, Limitations, and Practical Usage]
 
* [https://aman.ai/ Aman AI]: [https://aman.ai/primers/ai/LLM/ Overview of Large Language Models]
 
* [https://aman.ai/ Aman AI]: [https://aman.ai/primers/ai/LLM/ Overview of Large Language Models]
Line 23: Line 30:
  
 
===Video===
 
===Video===
* [https://www.youtube.com/watch?v=EWvNQjAaOHw How I use LLMs]
+
* Andrej Karpathy:
* [https://www.youtube.com/watch?v=7xTGNNLPyMI Deep Dive into LLMs like ChatGPT]
+
** [https://www.youtube.com/watch?v=EWvNQjAaOHw How I use LLMs]
 +
** [https://www.youtube.com/watch?v=7xTGNNLPyMI Deep Dive into LLMs like ChatGPT]
 +
 
 +
===Prompt Engineering===
 +
* 2025-04: Lee Boonstra (Google): [https://www.kaggle.com/whitepaper-prompt-engineering Prompt Engineering]
  
 
==Other Visualizations==
 
==Other Visualizations==
 +
* [https://distill.pub/2019/visual-exploration-gaussian-processes/ A Visual Exploration of Gaussian Processes]
 
* [https://pytorch.org/blog/inside-the-matrix/ Inside the Matrix: Visualizing Matrix Multiplication, Attention and Beyond]
 
* [https://pytorch.org/blog/inside-the-matrix/ Inside the Matrix: Visualizing Matrix Multiplication, Attention and Beyond]
 
* [https://sohl-dickstein.github.io/2024/02/12/fractal.html Neural network training makes beautiful fractals]
 
* [https://sohl-dickstein.github.io/2024/02/12/fractal.html Neural network training makes beautiful fractals]

Latest revision as of 15:12, 29 June 2025