TOOL2 stories
SGLang
High-performance serving framework for LLM and VLM workloads.
Stories
Filter storiesRELEASE11th May
OpenBMB releases MiniCPM-V 4.6 1.3B with 75.7 ms TTFT and 19x token efficiency
OpenBMB released MiniCPM-V 4.6 1.3B, claiming 55.8 percent lower vision-encoding FLOPs, 75.7 ms TTFT on a 4090, and about 1.5x token throughput over Qwen3.5 0.8B. It targets edge deployment across mobile platforms and common inference stacks.
RELEASE2mo ago
OpenClaw adds live Chrome control via Chrome DevTools MCP
OpenClaw beta added live control of a real Chrome session through Chrome DevTools MCP; the project also added native SGLang provider support and parallel tool calling work. Try it if you need self-hosted agents to handle authenticated browser flows with local inference backends.