/AdalFlow

AdalFlow: The “PyTorch” library to auto-optimize any LLM tasks.

Primary LanguagePythonMIT LicenseMIT

AdalFlow logo

⚡ The Library to Build and Auto-optimize LLM Applications ⚡

Try Quickstart in Colab

PyPI Version GitHub stars Open Issues License discord-invite

For AI researchers, product teams, and software engineers who want to learn the AI way.

Why AdalFlow

  1. Embracing a design pattern similar to PyTorch, AdalFlow is powerful, light, modular, and robust. AdalFlow provides Model-agnostic building blocks to build LLM task pipeline, ranging from RAG, Agents to classical NLP tasks like text classification and named entity recognition. It is easy to get high performance even with just basic manual promting.
  2. AdalFlow provides a unified auto-differentiative framework for both zero-shot prompt optimization and few-shot optimization. It advances existing auto-optimization research, including Text-Grad and DsPy. Through our research, Text-Grad 2.0 and Learn-to-Reason Few-shot In Context Learning, AdalFlow Trainer achieves the highest accuracy while being the most token-efficient.

Here is our optimization demonstration on a text classification task:

AdalFlow Auto-optimization

AdalFlow Optimized Prompt

Among all libraries, we achieved the highest accuracy with manual prompting (starting at 82%) and the highest accuracy after optimization.

Further reading: Optimize Classification

Light, Modular, and Model-agnositc Task Pipeline

LLMs are like water; AdalFlow help developers quickly shape them into any applications, from GenAI applications such as chatbots, translation, summarization, code generation, RAG, and autonomous agents to classical NLP tasks like text classification and named entity recognition.

Only two fundamental but powerful base classes: Component for the pipeline and DataClass for data interaction with LLMs. The result is a library with bare minimum abstraction, providing developers with maximum customizability.

You have full control over the prompt template, the model you use, and the output parsing for your task pipeline.

AdalFlow Task Pipeline

Further reading: How We Started, Design Philosophy and Class hierarchy.

Unified Framework for Auto-Optimization

AdalFlow provides token-efficient and high-performing prompt optimization within a unified framework. To optimize your pipeline, simply define a Parameter and pass it to our Generator. Whether you need to optimize task instructions or few-shot demonstrations, our unified framework offers an easy way to diagnose, visualize, debug, and train your pipeline.

This Trace Graph demonstrates how our auto-differentiation works.

Trainable Task Pipeline

Just define it as a Parameter and pass it to our Generator.

AdalFlow Trainable Task Pipeline

AdalComponent & Trainer

AdalComponent acts as the 'interpreter' between task pipeline and the trainer, defining training and validation steps, optimizers, evaluators, loss functions, backward engine for textual gradients or tracing the demonstrations, the teacher generator.

AdalFlow AdalComponent & Trainer

Quick Install

Install AdalFlow with pip:

pip install adalflow

Please refer to the full installation guide for more details.

Documentation

AdalFlow full documentation available at adalflow.sylph.ai:

AdalFlow: A Tribute to Ada Lovelace

AdalFlow is named in honor of Ada Lovelace, the pioneering female mathematician who first recognized that machines could do more than just calculations. As a female-led team, we aim to inspire more women to enter the AI field.

Contributors

contributors

Acknowledgements

Many existing works greatly inspired AdalFlow library! Here is a non-exhaustive list:

  • 📚 PyTorch for design philosophy and design pattern of Component, Parameter, Sequential.
  • 📚 Micrograd: A tiny autograd engine for our auto-differentiative architecture.
  • 📚 Text-Grad for the Textual Gradient Descent text optimizer.
  • 📚 DSPy for inspiring the __{input/output}__fields in our DataClass and the bootstrap few-shot optimizer.
  • 📚 OPRO for adding past text instruction along with its accuracy in the text optimizer.
  • 📚 PyTorch Lightning for the AdalComponent and Trainer.

Citation

@software{Yin2024AdalFlow,
  author = {Li Yin},
  title = {{AdalFlow: The Library for Large Language Model (LLM) Applications}},
  month = {7},
  year = {2024},
  doi = {10.5281/zenodo.12639531},
  url = {https://github.com/SylphAI-Inc/LightRAG}
}