N0x
LLM inference, agents, RAG, Python exec in browser, no back end
LLM inference, agents, RAG, Python exec in browser, no back end
LLM inference
Run large language models directly in the browser with WebGPU acceleration
Web search integration
Access real-time information from the internet within AI workflows
RAG (Retrieval-Augmented Generation)
Implement document-based AI systems for context-aware responses
In-browser Python execution
Write and run Python code without leaving the interface
Image generation
Create images using generative models entirely client-side
Memory and persistent context
Maintain conversation history and system state across sessions
Text-to-speech
Convert AI-generated or user text into natural-sounding audio
Rapid prototyping of AI agents and chatbots without backend infrastructure
Building RAG applications that reference local documents or web content
Educational exploration of LLMs and generative AI concepts
Privacy-sensitive applications where processing data locally is essential
Creating standalone AI-powered web applications that don't require server costs