use-local-llm
React hooks for streaming local LLM responses
React hooks for streaming responses from local LLMs — Ollama, LM Studio, llama.cpp, and any OpenAI-compatible endpoint. Direct browser-to-localhost streaming with no server, no API routes, and zero runtime dependencies beyond React.
- Direct browser → localhost streaming — no server, no API routes
- Multi-backend: Ollama, LM Studio, llama.cpp, OpenAI-compatible
- Full chat state — message history, abort, clear, error handling
- Token-by-token streaming with onToken callbacks
$ pnpm add use-local-llm






















