Web Search RAG Added to Ollama Colab Private Chat
Ollama Colab Private Chat now supports Web Search RAG. Just flip the 🔍 Web Search toggle in the chat UI to include DuckDuckGo search results as context when sending messages to the LLM.
大規模言語モデル(LLM)の活用や評価に関する記事。
View all tagsOllama Colab Private Chat now supports Web Search RAG. Just flip the 🔍 Web Search toggle in the chat UI to include DuckDuckGo search results as context when sending messages to the LLM.
ChatGPT and Claude are great tools, but do you ever hesitate before typing something work-related or personal into them? On the other hand, setting up a local GPU environment to run Ollama yourself is a steep barrier in terms of both effort and cost.
So I built a private LLM chat environment using Google Colab's free GPU — one that never sends conversation logs outside the instance — and packaged it into a single notebook. Run the cells from top to bottom, and it's ready to go.
Today I'm releasing Ollama Colab Free Server, an open-source notebook that runs Ollama on Google Colab's free GPU and makes it instantly available as a backend for Claude Code and Continue. Execute the cells top to bottom, and a publicly accessible LLM server is up within minutes.
Today we are releasing Ollama Multi-Model Benchmarker, an open-source tool for evaluating local LLMs before committing to a setup. It runs multiple Ollama models sequentially on Google Colab's free T4 GPU and produces a side-by-side comparison of generation speed, responsiveness, model size, and more.