Difference between revisions of "AI compute"

From GISAXS
Jump to: navigation, search
(Energy Use)
 
(6 intermediate revisions by the same user not shown)
Line 20: Line 20:
 
* [https://hyperbolic.xyz/ Hyperbolic AI] (open models, Enterprise tier)
 
* [https://hyperbolic.xyz/ Hyperbolic AI] (open models, Enterprise tier)
 
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub]
 
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub]
 +
* [https://www.asksage.ai/ AskSage]
  
 
==Multi-model with Model Selection==
 
==Multi-model with Model Selection==
Line 60: Line 61:
 
* 2024-02: [https://www.nature.com/articles/s41598-024-54271-x The carbon emissions of writing and illustrating are lower for AI than for humans]
 
* 2024-02: [https://www.nature.com/articles/s41598-024-54271-x The carbon emissions of writing and illustrating are lower for AI than for humans]
 
* 2025-04: [https://andymasley.substack.com/p/a-cheat-sheet-for-conversations-about Why using ChatGPT is not bad for the environment - a cheat sheet]
 
* 2025-04: [https://andymasley.substack.com/p/a-cheat-sheet-for-conversations-about Why using ChatGPT is not bad for the environment - a cheat sheet]
 +
** A single LLM response uses only ~3 Wh = 11 kJ (~10 Google searches; [https://docs.google.com/document/d/1pDdpPq3MyPdEAoTkho9YABZ0NBEhBH2v4EA98fm3pXQ/edit?usp=sharing examples of 3 Wh energy usage])
 +
** Reading an LLM-generated response (computer running for a few minutes) typically uses more energy than the LLM generation of the text.
 +
* 2025-07: Mistral: [https://mistral.ai/news/our-contribution-to-a-global-environmental-standard-for-ai Our contribution to a global environmental standard for AI]
 +
* 2025-08: [https://services.google.com/fh/files/misc/measuring_the_environmental_impact_of_delivering_ai_at_google_scale.pdf Measuring the environmental impact of delivering AI at Google Scale] ([https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference blog])

Latest revision as of 14:34, 21 August 2025

Cloud GPU

Cloud Training Compute

Cloud LLM Routers & Inference Providers

Multi-model with Model Selection

Multi-model Web Chat Interfaces

Multi-model Web Playground Interfaces

Local Router

Acceleration Hardware

Energy Use