Skip to content

A plugin for chatting with you obsidian notes trough local Ollama LLM instead of Chat GTP.

License

Notifications You must be signed in to change notification settings

brumik/obsidian-ollama-chat

Repository files navigation

🦙 Obsidian Ollama Chat

This plugin allows you to ask your local LLM about your own notes.

Ollama plugin link: https://obsidian.md/plugins?id=ollama-chat

Requirements:

Indexing is slow and hard to do in JS. Therefore you will need to run a lightweight python server to do the indexing for you next to your ollama.

For more information about progress and install see: https://github.com/brumik/ollama-obsidian-indexer

The https://github.com/brumik/obsidian-ollama-chat/tree/move-llama-inhouse branch does not need the above server to run but it is slower and less developed in general. If you cannot run the python server though you might find it useful.

To install the branch you need to build the plugin with npm run build and copy it to your obsidian install manually.

Features:

  • Run your own model locally. Set the URL to this model and you roll
  • Index your files on startup and on file modification
  • Open a modal by shortcut or command to ask your question

Future plans:

  • Text streaming when queriing to the LLM
  • Chat window for chat style communication insdead of query.
  • Add commands for useful queries to quick ask them like:
    • Summarize note
    • Summarize topic

Any feature recommendation is welcome.

About

A plugin for chatting with you obsidian notes trough local Ollama LLM instead of Chat GTP.

Resources

License

Stars

Watchers

Forks

Packages