This sample demonstrates how to create and work with AI agents driven by Azure OpenAI. It includes a Flask app that takes a topic and instruction from a user then calls a research agent that uses the Bing Search API to research the topic, a product agent that uses Azure AI Search to do a semantic similarity search for related products from a vectore store, a writer agent to combine the research and product information into a helpful article, and an editor agent to refine the article that's finally presented to the user.
- Features
- Azure account requirements
- Opening the project
- Deployment
- Testing the sample
- Costs
- Security Guidelines
- Resources
- Code of Conduct
This project template provides the following features:
- Azure OpenAI to drive the various agents
- Prompty and Prompt Flow to create, manage and evaluate the prompt into our code.
- Bing Search API to research the topic provided
- Azure AI Search for performing semantic similarity search
IMPORTANT: In order to deploy and run this example, you'll need:
- Azure account. If you're new to Azure, get an Azure account for free and you'll get some free Azure credits to get started. See guide to deploying with the free trial.
- Azure subscription with access enabled for the Azure OpenAI service. You can request access with this form. If your access request to Azure OpenAI service doesn't match the acceptance criteria, you can use OpenAI public API instead.
- Ability to deploy
gpt-35-turbo-0613
andgpt-4-1106-Preview
. - We recommend using East US 2, as this region has access to all models and services required.
- Ability to deploy
- Azure subscription with access enabled for Bing Search API
- Azure subscription with access enabled for Azure AI Search
You have a few options for setting up this project. The easiest way to get started is GitHub Codespaces, since it will setup all the tools for you, but you can also set it up locally.
-
You can run this template virtually by using GitHub Codespaces. The button will open a web-based VS Code instance in your browser:
-
Open a terminal window.
-
Sign in to your Azure account:
azd auth login
-
Provision the resources and deploy the code:
azd up
This project uses
gpt-35-turbo-0613
andgpt-4-1106-Preview
which may not be available in all Azure regions. Check for up-to-date region availability and select a region during deployment accordingly. For this project we recommend East US 2.
A related option is VS Code Dev Containers, which will open the project in your local VS Code using the Dev Containers extension:
-
Start Docker Desktop (install it if not already installed)
-
Open the project:
-
In the VS Code window that opens, once the project files show up (this may take several minutes), open a terminal window.
-
Create a new folder and switch to it in the terminal, then run this command to download the project code:
azd init -t agent-openai-python-prompty
Note that this command will initialize a git repository, so you do not need to clone this repository.
-
Install required packages:
cd src/api pip install -r requirements.txt
Once you've opened the project in Codespaces, Dev Containers, or locally, you can deploy it to Azure.
-
Sign in to your Azure account:
azd auth login
If you have any issues with that command, you may also want to try
azd auth login --use-device-code
. -
Create a new azd environment:
azd env new
This will create a folder under
.azure/
in your project to store the configuration for this deployment. You may have multiple azd environments if desired. -
Provision the resources and deploy the code:
azd up
This project uses
gpt-35-turbo-0613
andgpt-4-1106-Preview
which may not be available in all Azure regions. Check for up-to-date region availability and select a region during deployment accordingly. We recommend using East US 2 for this project. -
At the end of this process a
.env
file will be created for you. Copy this file to thesrc/api
folder.
This sample repository contains an agents folder that includes subfolders for each agent. Each agent forlder contains a prompty file where the agents prompty is defined and a python file with the code used to run it. Exploring these files will help you understand what each agent is doing. The agents folder also contains an orchestrator.py
file that can be used to run the entire flow and to create an article.
To test the sample:
- Populate the Azure AI Search vectore store index with product data.
- Change into the api/data folder:
cd src/api/data
- Install the Jupyter extension.
- Once the extension has been installed, open the
create-azure-search.ipynb
notebook. We will use this notebook to upload a catalogue of products to the Azure AI Search vector store. Clickselect kernel
in the top right hand corner of the notebook, choose Python environment and then select the recommended Python version. - Run all of the cells in the notebook. If this process was successful you should see "uploading 20 documents to index contoso-products". You're now ready to run the full promptflow.
-
Test with sample data
2.1 To run the sample using just the orchestrator logic use the following command:
``` cd .. python -m api.agents.orchestrator ```
2.2 You also have the option of testing this code locally using a Flask app. (You should run this from the src/api folder)
To run the flask webserver:
flask --debug --app api.app:app run --port 5000
http://127.0.0.1:5000/get_article?context=Write an article about camping in alaska&instruction=find specifics about what type of gear they would need and explain in detail
To understand how well our prompt flow performs using defined metrics like groundedness, coherence etc we can evaluate the results. To evaluate the prompt flow, we need to be able to compare it to what we see as "good results" in order to understand how well it aligns with our expectations.
We may be able to evaluate the flow manually (e.g., using Azure AI Studio) but for now, we'll evaluate this by running the prompt flow using gpt-4 and comparing our performance to the results obtained there. To do this, follow the instructions and steps in the notebook evaluate-chat-prompt-flow.ipynb
under the eval
folder.
You can also view the evaluation metrics by running the following commands.
In a new terminal
cd src/web
First install node packages:
npm install
Then run the web app with a local dev web server:
npm run dev
Then run evaluation
cd evaluate
python evaluate.py
Pricing may vary per region and usage. Exact costs cannot be estimated. You may try the Azure pricing calculator for the resources below:
- Azure Container Apps: Pay-as-you-go tier. Costs based on vCPU and memory used. Pricing
- Azure OpenAI: Standard tier, GPT and Ada models. Pricing per 1K tokens used, and at least 1K tokens are used per question. Pricing
- Azure Monitor: Pay-as-you-go tier. Costs based on data ingested. Pricing
This template use Managed Identity built in to eliminate the need for developers to manage these credentials. Applications can use managed identities to obtain Microsoft Entra tokens without having to manage any credentials. We also use Key Vault, specifically for Bing Search, since Managed Identity is currently not implemented for it. Additionally, we have added a GitHub Action tool that scans the infrastructure-as-code files and generates a report containing any detected issues. To ensure best practices in your repo we recommend anyone creating solutions based on our templates ensure that the Github secret scanning setting is enabled in your repos.
This project has adopted the Microsoft Open Source Code of Conduct.
Resources:
- Microsoft Open Source Code of Conduct
- Microsoft Code of Conduct FAQ
- Contact opencode@microsoft.com with questions or concerns
For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.