Difference between revisions of "AI compute"
KevinYager (talk | contribs) (→Cloud LLM Routers & Inference Providers) |
KevinYager (talk | contribs) (→Cloud LLM Routers & Inference Providers) |
||
(One intermediate revision by the same user not shown) | |||
Line 1: | Line 1: | ||
− | + | =Cloud GPU= | |
− | |||
* [https://lambdalabs.com/ Lambda] | * [https://lambdalabs.com/ Lambda] | ||
* [https://vast.ai/ Vast AI] | * [https://vast.ai/ Vast AI] | ||
Line 7: | Line 6: | ||
* [https://hpc-ai.com/ HPC-AI] | * [https://hpc-ai.com/ HPC-AI] | ||
− | + | =Cloud Training Compute= | |
* [https://nebius.ai/ Nebius AI] | * [https://nebius.ai/ Nebius AI] | ||
* [https://glaive.ai/ Glaive AI] | * [https://glaive.ai/ Glaive AI] | ||
− | + | =Cloud LLM Routers & Inference Providers= | |
* [https://openrouter.ai/ OpenRouter] (open and closed models, no Enterprise tier) | * [https://openrouter.ai/ OpenRouter] (open and closed models, no Enterprise tier) | ||
* [https://www.litellm.ai/ LiteLLM] (closed models, Enterprise tier) | * [https://www.litellm.ai/ LiteLLM] (closed models, Enterprise tier) | ||
Line 20: | Line 19: | ||
* Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub] | * Huggingface [https://huggingface.co/blog/inference-providers Inference Providers Hub] | ||
− | ===Multi-model Web Chat Interfaces | + | ==Multi-model with Model Selection== |
+ | * [https://www.notdiamond.ai/ Not Diamond ¬⋄] | ||
+ | * [https://withmartian.com/ Martian] | ||
+ | |||
+ | ==Multi-model Web Chat Interfaces== | ||
* [https://simtheory.ai/ SimTheory] | * [https://simtheory.ai/ SimTheory] | ||
* [https://abacus.ai/ Abacus AI] | * [https://abacus.ai/ Abacus AI] | ||
− | + | ==Multi-model Web Playground Interfaces== | |
* [https://www.together.ai/ Together AI] | * [https://www.together.ai/ Together AI] | ||
* [https://hyperbolic.xyz/ Hyperbolic AI] | * [https://hyperbolic.xyz/ Hyperbolic AI] | ||
− | + | =Local Router= | |
* [https://ollama.com/ Ollama] | * [https://ollama.com/ Ollama] | ||
* [https://github.com/mudler/LocalAI LocalAI] | * [https://github.com/mudler/LocalAI LocalAI] | ||
− | + | =Acceleration Hardware= | |
* [https://www.nvidia.com/ Nvidia] GPUs | * [https://www.nvidia.com/ Nvidia] GPUs | ||
* Google [https://en.wikipedia.org/wiki/Tensor_Processing_Unit TPU] | * Google [https://en.wikipedia.org/wiki/Tensor_Processing_Unit TPU] |
Latest revision as of 12:53, 21 February 2025
Contents
Cloud GPU
Cloud Training Compute
Cloud LLM Routers & Inference Providers
- OpenRouter (open and closed models, no Enterprise tier)
- LiteLLM (closed models, Enterprise tier)
- Cent ML (open models, Enterprise tier)
- Fireworks AI (open models, Enterprise tier)
- Abacus AI (open and closed models, Enterprise tier)
- Portkey (open? and closed models, Enterprise tier)
- Huggingface Inference Providers Hub
Multi-model with Model Selection
Multi-model Web Chat Interfaces
Multi-model Web Playground Interfaces
Local Router
Acceleration Hardware
- Nvidia GPUs
- Google TPU
- Etched: Transformer ASICs
- Cerebras
- Untether AI
- Graphcore
- SambaNova Systems
- Groq
- Tesla Dojo
- Deep Silicon: Combined hardware/software solution for accelerated AI (e.g. ternary math)