Olama Deep Researcher is an open source project developed by the LangChain team that aims to provide a fully localized web research and report writing assistant. The tool combines the local large language model (LLM) and a web search engine to automatically perform information retrieval, summary, and iterative optimization, and ultimately generate a Markdown format research report containing cited sources.
Project core functions
- Fully local operation: Use local LLM hosted by Ollama or LMStudio to ensure data privacy without having to send data to the cloud.
- Intelligent iteration research: Automatically optimize search strategies through multiple rounds of iteration (default 3 rounds, configurable) to fill knowledge gaps.
- Multiple search engine support: DuckDuckGo is used by default, and Tavily or Perplexity can also be configured as a search engine.
- Visualize the research process: Monitor the research progress in real time through LangGraph Studio, including the generation of search results, summaries and final reports.
- Academic output: Automatically generate research reports in Markdown format with full citations for further editing and sharing.
ˇWork process overview
- generate a search query: After the user enters a research topic, LLM generates an initial web search query.
- Get and summarize search results: Obtain relevant web content through the configured search engine and summarize it by LLM.
- Identify knowledge gaps and iteratively optimize: LLM analyzes the summary, identifies information gaps, generates new search queries, and repeats the above steps for optimization.
- Generate final report: After multiple iterations, a Markdown format research report containing all cited sources is generated.
Quick Getting Started Guide (Take Mac as an example)
- Install the Ollama app: From Ollama official website Download and install apps for Mac.
- Pull local LLM models: For example, use the following command to pull the DeepSeek R1 model:
ollama pull deepseek-r1:8b(GitHub – langchain-ai/local-deep-researcher: Fully local web research and report writing assistant)
- Configure search engine API (optional): Register Tavily, obtain the API key, and set environment variables
TAVILY_API_KEY。 - Clone project and start service:
git clone https://github.com/langchain-ai/ollama-deep-researcher.git cd ollama-deep-researcher python -m venv .venv source .venv/bin/activate pip install -e . langgraph dev(GitHub – langchain-ai/ollama-deep-researcher: Fully local web research and report writing assistant)
- Visit LangGraph Studio: Open in browser
http://127.0.0.1:2024, enter LangGraph Studio, configure research parameters and start using them. (올라마 딥 리서치, 오픈AI 딥 리서처의 오픈 소스 대안)
🧠Project inspiration and technical background
The design inspiration of Olama Deep Researcher comes from the IterDRAG method, which decomposes queries into sub-queries, gradually obtains and summarizes information, and builds a complete knowledge graph. The project uses LangGraph to build a research flow chart and combines local LLM to achieve automated information retrieval and summary. (ollama-deep-researcher/README.md at main · langchain-ai/ollama-deep-researcher · GitHub, [AI 오픈소스] Ollama Deep Researcher – 최강의 로컬 웹 리서치 및 보고서 작성 도우미)
📁Project resources and links
- GitHub repository: (GitHub – langchain-ai/ollama-deep-researcher: Fully local web research and report writing assistant)
- Video Tutorial: The project provides detailed video tutorials built and used from scratch to help users get started quickly.
- LangGraph Studio: Through a visual interface, users can monitor the research process in real time, adjust parameters, and view intermediate results.
Github:https://github.com/langchain-ai/ollama-deep-researcher
Oil tubing: