Difference between revisions of "AI compute"
KevinYager (talk | contribs) (→Multi-model Web Chat Interfaces) |
KevinYager (talk | contribs) (→Multi-model Web Chat Interfaces) |
||
Line 26: | Line 26: | ||
* [https://simtheory.ai/ SimTheory] | * [https://simtheory.ai/ SimTheory] | ||
* [https://abacus.ai/ Abacus AI] [https://chatllm.abacus.ai/ ChatLLM] | * [https://abacus.ai/ Abacus AI] [https://chatllm.abacus.ai/ ChatLLM] | ||
+ | * [https://poe.com/about Poe] | ||
==Multi-model Web Playground Interfaces== | ==Multi-model Web Playground Interfaces== |
Revision as of 11:00, 26 February 2025
Contents
Cloud GPU
Cloud Training Compute
Cloud LLM Routers & Inference Providers
- OpenRouter (open and closed models, no Enterprise tier)
- LiteLLM (closed models, Enterprise tier)
- Cent ML (open models, Enterprise tier)
- Fireworks AI (open models, Enterprise tier)
- Abacus AI (open and closed models, Enterprise tier)
- Portkey (open? and closed models, Enterprise tier)
- Huggingface Inference Providers Hub
Multi-model with Model Selection
Multi-model Web Chat Interfaces
Multi-model Web Playground Interfaces
Local Router
Acceleration Hardware
- Nvidia GPUs
- Google TPU
- Etched: Transformer ASICs
- Cerebras
- Untether AI
- Graphcore
- SambaNova Systems
- Groq
- Tesla Dojo
- Deep Silicon: Combined hardware/software solution for accelerated AI (e.g. ternary math)