Difference between revisions of "AI compute"

From GISAXS
Jump to: navigation, search
(Energy Use)
(Cloud LLM Routers & Inference Providers)
 
Line 24: Line 24:
 
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub]
 
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub]
 
* [https://www.asksage.ai/ AskSage]
 
* [https://www.asksage.ai/ AskSage]
 +
* [https://opencode.ai/docs/zen/ Opencode Zen] (for coding agents)
  
 
==Multi-model with Model Selection==
 
==Multi-model with Model Selection==

Latest revision as of 07:54, 16 September 2025

Analysis

Cloud GPU

Cloud Training Compute

Cloud LLM Routers & Inference Providers

Multi-model with Model Selection

Multi-model Web Chat Interfaces

Multi-model Web Playground Interfaces

Local Router

Acceleration Hardware

Energy Use

Examples

  • LLM query
    • 3 kW * 4s = 3 Wh = 11 kJ
  • Human brain
    • 20 W * 8h = 106 Wh
    • 20 W * 1h = 20 Wh
    • 20 W * 10m = 3 Wh = 10 kJ
  • Human brain excess thinking
    • 2 W * 8h = 11 Wh
    • 2 W * 1.7h = 3 Wh
  • Regular computer
    • 200 W * 8h = 1,600 Wh = 5,700 kJ
    • 200 W * 1m = 3 Wh = 10kJ