Difference between revisions of "AI compute"

From GISAXS
Jump to: navigation, search
(Cloud LLM Routers & Inference Providers)
(Energy Use)
 
Line 64: Line 64:
 
** Reading an LLM-generated response (computer running for a few minutes) typically uses more energy than the LLM generation of the text.
 
** Reading an LLM-generated response (computer running for a few minutes) typically uses more energy than the LLM generation of the text.
 
* 2025-07: Mistral: [https://mistral.ai/news/our-contribution-to-a-global-environmental-standard-for-ai Our contribution to a global environmental standard for AI]
 
* 2025-07: Mistral: [https://mistral.ai/news/our-contribution-to-a-global-environmental-standard-for-ai Our contribution to a global environmental standard for AI]
 +
* 2025-08: [https://services.google.com/fh/files/misc/measuring_the_environmental_impact_of_delivering_ai_at_google_scale.pdf Measuring the environmental impact of delivering AI at Google Scale] ([https://cloud.google.com/blog/products/infrastructure/measuring-the-environmental-impact-of-ai-inference blog])

Latest revision as of 14:34, 21 August 2025

Cloud GPU

Cloud Training Compute

Cloud LLM Routers & Inference Providers

Multi-model with Model Selection

Multi-model Web Chat Interfaces

Multi-model Web Playground Interfaces

Local Router

Acceleration Hardware

Energy Use