Chipper in Action π¬ β
Experience Chipper live right here! Please note that the demo's functionality is subject to limitations. It utilizes the Hugging Face serverless inference API, which operates under strict quotas and rate limits. Each user is allowed up to 8 prompts per day. The embedded documents are the chipper test data.
Live Demo β
Maybe you ask Chipper now:
Tell me a story about Chipper, the brilliant golden retriever.
or visit the demo at: https://demo.chipper.tilmangriesel.com/
This demo is limited by the free resources available from Hugging Face and myself. The quality of the results depends on the specific model and its size, with larger models typically producing more accurate outcomes. Please note that model and index selection options are disabled in this demo.
Additionally, the serverless inference API may sometimes queue your request, leading to delays. During this period, youβll see the "Chipper is thinking..." message. Keep in mind that the performance of Hugging Face inference in this demo may vary and does not represent the experience of a self-hosted setup.
Demos β
Web Interface β
Use Chipper's built-in web interface to set up and customize RAG pipelines with ease. Built with vanilla JavaScript and TailwindCSS, it works offline and doesn't require any framework-specific knowledge. Run the /help
command to learn how to switch models, update the embeddings index, and more.
Code Output β
Automatic syntax highlighting for popular programming languages in the web interface.
Reasoning β
For models like DeepSeek-R1, Chipper suppresses the "think" output in the UI while preserving the reasoning steps in the console output.
CLI Interface β
Full support for the Ollama CLI and API, including reflection and proxy capabilities, with API key route decorations.
Third-Party Client β
Enhance every third-party Ollama client with server-side knowledge base embeddings, allowing server side model selection, query parameters, and system prompt overrides. Enable RAG for any Ollama client or use Chipper as a centralized knowledge base.
Experimentation π§ͺ β
Since Chipper uses embeddings, you can ask him about his adventures based on the embedded stories.
Live-Demo Setup β
- Inference: meta-llama/Meta-Llama-3-8B-Instruct
- Embedding: sentence-transformers/all-mpnet-base-v2
Server Specs β
The demo operates on a Scaleway Stardust 1 instance, utilizing the latest Chipper stack along with Elasticsearch from Docker Hub, powered by the Hugging Face provider.
Specification | Details |
---|---|
CPU | AMD EPYC 7282 (2.8 GHz) |
CPU Architecture | amd64 |
Sizing | 1 vCPU, 1 GiB RAM, 8 GiB Swap |
Storage | 25 GB |
More about the stardust instance
INFO
This section is still under construction.
Have a great day! And donβt forget to check out the Chipper Quickstart Guide! π