Skip to content

Chipper in Action 🎬 ​

Experience Chipper live right here! Please note that the demo's functionality is subject to limitations. It utilizes the Hugging Face serverless inference API, which operates under strict quotas and rate limits. Each user is allowed up to 8 prompts per day. The embedded documents are the chipper test data.

Live Demo ​

Maybe you ask Chipper now:

plain
Tell me a story about Chipper, the brilliant golden retriever.

or visit the demo at: https://demo.chipper.tilmangriesel.com/

This demo is limited by the free resources available from Hugging Face and myself. The quality of the results depends on the specific model and its size, with larger models typically producing more accurate outcomes. Please note that model and index selection options are disabled in this demo.

Additionally, the serverless inference API may sometimes queue your request, leading to delays. During this period, you’ll see the "Chipper is thinking..." message. Keep in mind that the performance of Hugging Face inference in this demo may vary and does not represent the experience of a self-hosted setup.

Demos ​

Web Interface ​

Use Chipper's built-in web interface to set up and customize RAG pipelines with ease. Built with vanilla JavaScript and TailwindCSS, it works offline and doesn't require any framework-specific knowledge. Run the /help command to learn how to switch models, update the embeddings index, and more.

chipper_demo_chat

Code Output ​

Automatic syntax highlighting for popular programming languages in the web interface.

chipper_demo_code_gen

Reasoning ​

For models like DeepSeek-R1, Chipper suppresses the "think" output in the UI while preserving the reasoning steps in the console output.

chipper_demo_deepseek

CLI Interface ​

Full support for the Ollama CLI and API, including reflection and proxy capabilities, with API key route decorations.

chipper_demo_ollama_cli

Third-Party Client ​

Enhance every third-party Ollama client with server-side knowledge base embeddings, allowing server side model selection, query parameters, and system prompt overrides. Enable RAG for any Ollama client or use Chipper as a centralized knowledge base.

chipper_demo_ollamac

Experimentation πŸ§ͺ ​

Since Chipper uses embeddings, you can ask him about his adventures based on the embedded stories.

Live-Demo Setup ​

Server Specs ​

The demo operates on a Scaleway Stardust 1 instance, utilizing the latest Chipper stack along with Elasticsearch from Docker Hub, powered by the Hugging Face provider.

SpecificationDetails
CPUAMD EPYC 7282 (2.8 GHz)
CPU Architectureamd64
Sizing1 vCPU, 1 GiB RAM, 8 GiB Swap
Storage25 GB

More about the stardust instance


INFO

This section is still under construction.


Have a great day! And don’t forget to check out the Chipper Quickstart Guide! πŸš€