⚙️
LLM Inference & Serving
70 tools
Inference runtimes, model serving platforms, fine-tuning infra, and GPU/accelerator providers for LLMs.
BytePlus
BytePlus Pte Ltd.
AI-Native Cloud for Enterprise Growth
1 story
Hugging Face Hub
Hugging Face
The central place to collaborate on models, datasets, and Spaces.
1 story
OpenRouter
OpenRouter, Inc.
The Unified Interface For LLMs
1 story
AFM Playground
Arcee AI
Playground for AFM-4.5B
0 stories
AI Studio
Google
Fastest way to start building with Gemini
0 stories
AI/ML API
AI/ML API
One API +400 AI models
0 stories
Amazon Bedrock
Amazon Web Services
The platform for building generative AI applications and agents at production scale
0 stories
Anthropic
Anthropic
AI for problem solvers
0 stories
Baidu AI Studio LLM API
Baidu
OpenAI-compatible LLM API for Baidu AI Studio
0 stories
Baidu Qianfan
Baidu
以Agent为核心的一站式企业级大模型服务平台
0 stories
Baseten
Baseten
Inference Platform: Deploy AI models in production
0 stories
Cerebras
Cerebras Systems
AI training and inference platform
0 stories
Coding Agents
Baseten
The best coding agents run on Baseten
0 stories
Conway
Conway Research
Infrastructure for self-improving, self-replicating, autonomous AI
0 stories
Decoupled DiLoCo
Google DeepMind
Resilient, distributed AI training at scale
0 stories
D
DeepClaude
Asterisk
1+1 > 2 - Combine Advanced Reasoning and Coding
0 stories
DeepEP
DeepSeek
An efficient expert-parallel communication library
0 stories
DeepGEMM
DeepSeek
clean and efficient FP8 GEMM kernels with fine-grained scaling
0 stories
DGX Spark
NVIDIA
AI supercomputer on your desk
0 stories
Exo
Exo Labs
Run frontier AI locally.
0 stories
fal
fal
Generative media platform for developers.
0 stories
Fireworks AI
Fireworks AI, Inc.
Fastest inference for generative AI
0 stories
FlashMLA
DeepSeek
Efficient Multi-head Latent Attention Kernels
0 stories
FlashQLA
Alibaba Cloud
Alibaba Cloud software product
0 stories
Gemini Live API
Google
Real-time, bidirectional multimodal API for Gemini.
0 stories
Google AI Edge Gallery
Google LLC
Explore, Experience, and Evaluate the Future of On-Device Generative AI with Google AI Edge.
0 stories
Google Cloud
Google
Cloud computing services from Google
0 stories
GuideLLM
Red Hat
SLO-aware Benchmarking and Evaluation Platform for Optimizing Real-World LLM Inference
0 stories
Hugging Face
Hugging Face
The AI community building the future.
0 stories
Interfaze
JigsawStack, Inc.
AI interface platform
0 stories
Keras Kinetic
Keras
Run ML workloads remotely on cloud TPUs and GPUs.
0 stories
Lightning AI
Lightning AI
Idea to AI product, ⚡️ fast.
0 stories
LiteLLM
BerriAI
AI Gateway to provide model access, fallbacks and spend tracking across 100+ LLMs. All in the OpenAI format.
0 stories
LM Studio
Element Labs, Inc.
Run AI models, locally and privately.
0 stories
Miles RL Training
RadixArk
Enterprise-Grade Reinforcement Learning for Large-Scale Model Training
0 stories
ModelScope
Alibaba Cloud
Open-source AI model community and MaaS platform
0 stories
Modular
Modular
Inference from Kernel to Cloud.
0 stories
Mooncake
KVCache.ai
A KVCache-centric Disaggregated Architecture for LLM Serving
0 stories
Multimodal Max
Modular
GenAI-native serving and modeling, built for performance.
0 stories
NeMo-RL
NVIDIA
Scalable RL post-training for language models.
0 stories
N
Nous Portal
Nous Research
Portal for Nous models and services.
0 stories
NVIDIA DGX 8xB200
NVIDIA
Unified AI platform for develop-to-deploy AI pipelines
0 stories
NVIDIA NIM
NVIDIA
Designed for rapid, reliable deployment of accelerated generative AI inference anywhere.
0 stories
Ollama
Ollama Inc.
Ollama is the easiest way to run open AI models locally or in the cloud, with a simple API and 40,000+ integrations.
0 stories
Open Generative AI
Muapi
Free Higgsfield AI, Freepik & Krea AI Alternative
0 stories
O
Open Responses
Open Responses
Open-source specification and ecosystem for interoperable LLM interfaces.
0 stories
O
OpenAI
OpenAI
AI platform and product suite
0 stories
PaddlePaddle AI Studio
Baidu, Inc.
人工智能学习与实训社区
0 stories
Pocket TTS
Kyutai
A lightweight text-to-speech application designed to run efficiently on CPUs.
0 stories
Prime Intellect
Prime Intellect
Distributed training and inference infrastructure
0 stories
Prime Intellect Lab
Prime Intellect
AI lab for experimenting with language models
0 stories
RunPod
Runpod, Inc.
Everything you need to train, deploy, and scale AI all in one place.
0 stories
SGLang
LMSYS Corp.
High-Performance Serving Framework for LLMs and VLMs
0 stories
Tile Kernels
Hangzhou DeepSeek Artificial Intelligence Co., Ltd.
Optimized GPU kernels for LLM operations, built with TileLang.
0 stories
TileLang-Ascend
Tile AI
Ascend TileLang adapter
0 stories
Together Fine-Tuning
Together AI
Fine-tune open-source models for real production use
0 stories
Train Models
TrainEngine.ai
Train Models privately
0 stories
Unsloth
Unsloth AI
Easily run & train models locally.
0 stories
Unsloth AI
Unsloth
Train and Run Models Locally
0 stories
U
Unsloth Studio
Unsloth
open-source, no-code web UI for training, running and exporting open models in one unified local interface
0 stories
Vast.ai
Vast.ai
Launch Fast, Pay Less
0 stories
Venice API
Venice.ai
Developer API for AI models
0 stories
Vertex AI
Google Cloud
Build, deploy, and scale machine learning models.
0 stories
vLLM
vLLM Project
The High-Throughput and Memory-Efficient inference and serving engine for LLMs
0 stories
vLLM Omni
vLLM Project
Easy, fast, and cheap omni-modality model serving for everyone
0 stories
xAI
xAI
AI for all humanity
0 stories
Xiaomi MiMo Orbit
Xiaomi
100T-token creator incentive program for MiMo builders
0 stories
ZenMux
AI Force Singapore Pte. Ltd.
AI software platform
0 stories
Zyphra Cloud
Zyphra Technologies Inc.
A full-stack AI platform on AMD powered by TensorWave
0 stories
Zyphra Inference
Zyphra
Serverless inference for frontier open-weight models
0 stories