promptic
is a lightweight, decorator-based Python library that simplifies the process of interacting with large language models (LLMs) using litellm. With promptic
, you can effortlessly create prompts, handle input arguments, and receive structured outputs from LLMs with just a few lines of code.
pip install promptic
from promptic import llm
@llm
def president(year):
"""Who was the President of the United States in {year}?"""
print(president(2000))
# The President of the United States in 2000 was Bill Clinton until January 20th, when George W. Bush was inaugurated as the 43rd President.
from pydantic import BaseModel
from promptic import llm
class Capital(BaseModel):
country: str
capital: str
@llm
def capital(country) -> Capital:
"""What's the capital of {country}?"""
print(capital("France"))
# country='France' capital='Paris'
Streaming Response (and litellm integration)
from promptic import llm
# most arguments are passed directly to litellm.completion
# see https://docs.litellm.ai/docs/completion
@llm(stream=True, model="claude-3-haiku-20240307")
def haiku(subject, adjective, verb="delights"):
"""Write a haiku about {subject} that is {adjective} and {verb}."""
print("".join(haiku("programming", adjective="witty")))
# Bits and bytes abound,
# Bugs and features intertwine,
# Code, the poet's rhyme.
from promptic import llm
@llm(system="you are a snarky chatbot")
def answer(question):
"""{question}"""
print(answer("What's the best programming language?"))
# Well, that's like asking what's the best flavor of ice cream.
# It really depends on what you're trying to accomplish and your personal preferences.
# But if you want to start a flame war, just bring up Python vs JavaScript.
- Decorator-based API: Easily define prompts using function docstrings and decorate them with
@promptic.llm
. - Argument interpolation: Automatically interpolate function arguments into the prompt using
{argument_name}
placeholders within docstrings. - Pydantic model support: Specify the expected output structure using Pydantic models, and
promptic
will ensure the LLM's response conforms to the defined schema. - Streaming support: Receive LLM responses in real-time by setting
stream=True
when calling the decorated function. - Simplified LLM interaction: No need to remember the exact shape of the OpenAPI response object or other LLM-specific details.
promptic
abstracts away the complexities, allowing you to focus on defining prompts and receiving structured outputs.
promptic
is designed to be simple, functional, and robust, providing exactly what you need 90% of the time when working with LLMs. It eliminates the need to remember the specific shapes of OpenAPI response objects or other LLM-specific details, allowing you to focus on creating prompts and receiving structured outputs.
With its legible and concise codebase, promptic
is reliable easy to understand. It leverages the power of litellm under the hood, ensuring compatibility with a wide range of LLMs.
promptic
is open-source software licensed under the Apache License 2.0.