Difference between revisions of "AI compute"
KevinYager (talk | contribs) (→Cloud LLM Routers & Inference Providers) |
KevinYager (talk | contribs) (→Cloud LLM Routers & Inference Providers) |
||
Line 12: | Line 12: | ||
==Cloud LLM Routers & Inference Providers== | ==Cloud LLM Routers & Inference Providers== | ||
− | * [https://openrouter.ai/ OpenRouter] | + | * [https://openrouter.ai/ OpenRouter] (open and closed models, no Enterprise tier) |
− | * [https://www.litellm.ai/ LiteLLM] | + | * [https://www.litellm.ai/ LiteLLM] (closed models, Enterprise tier) |
− | * [https://centml.ai/ Cent ML] | + | * [https://centml.ai/ Cent ML] (open models, Enterprise tier) |
− | * [https://fireworks.ai/ Fireworks AI] | + | * [https://fireworks.ai/ Fireworks AI] (open models, Enterprise tier) |
+ | * [https://abacus.ai/ Abacus AI] (open and closed models, Enterprise tier) | ||
+ | * [https://portkey.ai/ Portkey] (open? and closed models, Enterprise tier) | ||
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub] | * Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub] | ||
− | |||
− | |||
===Multi-model Web Chat Interfaces=== | ===Multi-model Web Chat Interfaces=== |
Revision as of 15:29, 18 February 2025
Contents
Cloud GPU
Cloud Training Compute
Cloud LLM Routers & Inference Providers
- OpenRouter (open and closed models, no Enterprise tier)
- LiteLLM (closed models, Enterprise tier)
- Cent ML (open models, Enterprise tier)
- Fireworks AI (open models, Enterprise tier)
- Abacus AI (open and closed models, Enterprise tier)
- Portkey (open? and closed models, Enterprise tier)
- Huggingface Inference Providers Hub
Multi-model Web Chat Interfaces
Multi-model Web Playground Interfaces
Acceleration Hardware
- Nvidia GPUs
- Google TPU
- Etched: Transformer ASICs
- Cerebras
- Untether AI
- Graphcore
- SambaNova Systems
- Groq
- Tesla Dojo
- Deep Silicon: Combined hardware/software solution for accelerated AI (e.g. ternary math)