This is a POC for an information extraction tool for text documents using LLMs. The tool frameworks the information extraction process, with a codebook interface that makes the information extraction process as easy as possible for researchers. It builds on the llama.cpp project which lets you run LLM models locally.
Compile and run ./server
from llama.cpp and run the tool with rye run coderclient
.
To code the example document, run:
cat ./examples/doc | rye run coderclient --codebook ./examples/codebook.yaml code