/llm_text_analyzer

Summarize/analyze large amounts of text using local LLM models, langchain, ollama, and flask. No data leaves your computer.

Primary LanguageHTMLCreative Commons Zero v1.0 UniversalCC0-1.0

text_summarizer

summarize large amounts of text using langchain, ollama and flask. Spins out a local server using flask, works offline, no data leaves the computer.

How to set up

  • Fork or clone this repository

  • Install Ollama for easy interaction with LLMs (brew install ollama), follow instructions to spin up the ollama service (brew services start ollama)

  • Download llm models, e.g.:

    • ollama pull dolphin-llama3
    • ollama pull llama3

    Feel free to add and try out other models. For more info on models, visit ollama library

  • recommended: create new environment with venv or VScode

  • navigate to the local folder of the repo

  • install project dependencies (pip install -r requirements.txt)

How to use

  • navigate to repo local folder, enter your venv environment if set up
  • run python app.py
    • Open the link in your browser
  • provide context to work from - either copy/paste or open a file (which will populate the input window)
  • Select model to use - dropdown will list those that are available in ollama.
  • you can pick from predefined system prompts and questions, or you can edit it. It is crucial to include {context} and {input} in the system prompt, as otherwise the model won't know what data to base the answer on, and what the question is
  • Submit and... patience! Time depends hugely on your system performance and on size of the text
  • You'll get the generated response and the request metadata in a page