Azure-Samples / contoso-chat
- четверг, 21 марта 2024 г. в 00:00:07
This sample has the full End2End process of creating RAG application with Prompt Flow and AI Studio. It includes GPT 3.5 Turbo LLM application code, evaluations, deployment automation with AZD CLI, GitHub actions for evaluation and deployment and intent mapping for multiple LLM task mapping.
Warning
This sample is under active development to showcase new features and evolve with the Azure AI Studio (preview) platform. Keep in mind that the latest build may not be rigorously tested for all environments (local development, GitHub Codespaces, Skillable VM).
Instead refer to the table, identify the right commit version in context, then launch in GitHub Codespaces
Build Version | Description |
---|---|
Stable : #cc2e808 | Version tested & used in Microsoft AI Tour (works on Skillable) |
Active : main | Version under active development (breaking changes possible) |
Table Of Contents
If you find this sample useful, consider giving us a star on GitHub! If you have any questions or comments, consider filing an Issue on the source repo.
Learn to build an Large Language Model (LLM) Application with a RAG (Retrieval Augmented Generation) architecture using Azure AI Studio and Prompt Flow. By the end of this workshop you should be able to:
The repository is instrumented with a devcontainer.json
configuration that can provide you with a pre-built environment that can be launched locally, or in the cloud. You can also elect to do a manual environment setup locally, if desired. Here are the three options in increasing order of complexity and effort on your part. Pick one!
The first approach is recommended for minimal user effort in startup and maintenance. The third approach will require you to manually update or maintain your local environment, to reflect any future updates to the repo.
To setup the development environment you can leverage either GitHub Codespaces, a local Python environment (using Anaconda or venv), or a VS Code Dev Container environment (using Docker).
This is the recommended option.
Code
button on the repositoryCodespaces
tab and click Create codespace...
This should open a new browser tab with a Codespaces container setup process running. On completion, this will launch a Visual Studio Code editor in the browser, with all relevant dependencies already installed in the running development container beneath. Congratulations! Your cloud dev environment is ready!
This option uses the same devcontainer.json
configuration, but launches the development container in your local device using Docker Desktop. To use this approach, you need to have the following tools pre-installed in your local device:
Make sure your Docker Desktop daemon is running on your local device. Then,
If your Dev Containers extension is installed correctly, you will be prompted to "re-open the project in a container" - just confirm to launch the container locally. Alternatively, you may need to trigger this step manually. See the Dev Containers Extension for more information.
Once your project launches in the local Docker desktop container, you should see the Visual Studio Code editor reflect that connection in the status bar (blue icon, bottom left). Congratulations! Your local dev environment is ready!
Clone the repo
git clone https://github.com/azure/contoso-chat
Open the repo in VS Code
cd contoso-chat
code .
Install the Prompt Flow Extension in VS Code
Install the Azure CLI for your device OS
Create a new local Python environment using either anaconda or venv for a managed environment.
Option 1: Using anaconda
conda create -n contoso-chat python=3.11
conda activate contoso-chat
pip install -r requirements.txt
Option 2: Using venv
python3 -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt
We setup our development ennvironment in the previous step. In this step, we'll provision Azure resources for our project, ready to use for developing our LLM Application.
Start by connecting your Visual Studio Code environment to your Azure account:
az login
.If you are running within a dev container, use these instructions to login instead:
az login --use-device-code
In either case, verify that the console shows a message indicating a successful authentication. Congratulations! Your VS Code session is now connected to your Azure subscription!
The project requires a number of Azure resources to be set up, in a specified order. To simplify this, an auto-provisioning script has been provided. (NOTE: It will use the current active subscription to create the resource. If you have multiple subscriptions, use az account set --subscription "<SUBSCRIPTION-NAME>"
first to set the desired active subscription.)
Run the provisioning script as follows:
./provision.sh
The script should set up a dedicated resource group with the following resources:
The script will set up an Azure AI Studio project with the following model deployments created by default, in a relevant region that supports them. Your Azure subscription must be enabled for Azure OpenAI access.
The Azure AI Search resource will have Semantic Ranker enabled for this project, which requires the use of a paid tier of that service. It may also be created in a different region, based on availability of that feature.
The script should automatically create a config.json
in your root directory, with the relevant Azure subscription, resource group, and AI workspace properties defined. These will be made use of by the Azure AI SDK for relevant API interactions with the Azure AI platform later.
If the config.json file is not created, simply download it from your Azure portal by visiting the Azure AI project resource created, and looking at its Overview page.
The default sample has an .env.sample
file that shows the relevant environment variables that need to be configured in this project. The script should create a .env
file that has these same variables but populated with the right values for your Azure resources.
If the file is not created, simply copy over .env.sample
to .env
- then populate those values manually from the respective Azure resource pages using the Azure Portal (for Azure CosmosDB and Azure AI Search) and the Azure AI Studio (for the Azure OpenAI values)
You will need to have your local Prompt Flow extension configured to have the following connection objects set up:
contoso-cosmos
to Azure Cosmos DB endpointcontoso-search
to Azure AI Search endpointaoai-connection
to Azure OpenAI endpointVerify if these were created by using the pf tool from the VS Code terminal as follows:
pf connection list
If the connections are not visible, create them by running the connections/create-connections.ipynb
notebook. Then run the above command to verify they were created correctly.
The auto-provisioning will have setup 2 of the 3 connections for you by default. First, verify this by
You should see contoso-search
and aoai-connection
pre-configured, else create them from the Azure AI Studio interface using the Create Connection workflow (and using the relevant values from your .env
file).
You will however need to create contoso-cosmos
manually from Azure ML Studio. This is a temporary measure for custom connections and may be automated in future. For now, do this:
.env
file:
Refresh main Connections list screen to verify that you now have all three required connections listed.
In this step we want to populate the required data for our application use case.
data/product_info/create-azure-search.ipynb
notebook.data/customer_info/create-cosmos-db.ipynb
notebook.We are now ready to begin building our prompt flow! The repository comes with a number of pre-written flows that provide the starting points for this project. In the following section, we'll explore what these are and how they work.
A prompt flow is a DAG (directed acyclic graph) that is made up of nodes that are connected together to form a flow. Each node in the flow is a python function tool that can be edited and customized to fit your needs.
Click on the contoso-chat/flow.dag.yaml
file in the Visual Studio Code file explorer.
You should get a view similar to what is shown below.
This will open up the prompt flow in the visual editor as shown: -
The prompt flow is a directed acyclic graph (DAG) of nodes, with a starting node (input), a terminating node (output), and an intermediate sub-graph of connected nodes as follows:
Node | Description |
---|---|
inputs | This node is used to start the flow and is the entry point for the flow. It has the input parameters customer_id and question , and chat_history . The customer_id is used to look up the customer information in the Cosmos DB. The question is the question the customer is asking. The chat_history is the chat history of the conversation with the customer. |
question_embedding | This node is used to embed the question text using the text-embedding-ada-002 model. The embedding is used to find the most relevant documents from the AI Search index. |
retrieve_documents | This node is used to retrieve the most relevant documents from the AI Search index with the question vector. |
customer_lookup | This node is used to get the customer information from the Cosmos DB. |
customer_prompt | This node is used to generate the prompt with the information retrieved and added to the customer_prompt.jinja2 template. |
llm_response | This node is used to generate the response to the customer using the GPT-35-Turbo model. |
outputs | This node is used to end the flow and return the response to the customer. |
Let's run the flow to see what happens. Note that the input node is pre-configured with a question. By running the flow, we anticipate that the output node should now provide the result obtained from the LLM when presented with the customer prompt that was created from the initial question with enhanced customer data and retrieved product context.
Run All
(play icon) at the top. When prompted, select "Run it with standard mode".Prompt Flow
tab in the Visual Studio Code terminal window for execution timesFor more details on running the prompt flow, follow the instructions here.
Congratulations!! You ran the prompt flow and verified it works!
If you like, you can try out other possible customer inputs to see what the output of the Prompt Flow might be. (This step is optional, and you can skip it if you like.)
Run All
(play icon) at the top. This time when prompted, select "Run it with interactive mode (text only)."Here are some questions you can try:
Now, we need to understand how well our prompt flow performs using defined metrics like groundedness, coherence etc. To evaluate the prompt flow, we need to be able to compare it to what we see as "good results" in order to understand how well it aligns with our expectations.
We may be able to evaluate the flow manually (e.g., using Azure AI Studio) but for now, we'll evaluate this by running the prompt flow using gpt-4 and comparing our performance to the results obtained there. To do this, follow the instructions and steps in the notebook evaluate-chat-prompt-flow.ipynb
under the eval
folder.
At this point, we've built, run, and evaluated, the prompt flow locally in our Visual Studio Code environment. We are now ready to deploy the prompt flow to a hosted endpoint on Azure, allowing others to use that endpoint to send user questions and receive relevant responses.
This process consists of the following steps:
Just follow the instructions and steps in the notebook push_and_deploy_pf.ipynb
under the deployment
folder. Once this is done, the deployment endpoint and key can be used in any third-party application to integrate with the deployed flow for real user experiences.
Login to Azure Shell
Follow the instructions to create a service principal here
Follow the instructions in steps 1 - 8 here to add create and add the user-assigned managed identity to the subscription and workspace.
Assign Data Science Role
and the Azure Machine Learning Workspace Connection Secrets Reader
to the service principal. Complete this step in the portal under the IAM.
Setup authentication with Github here
{
"clientId": <GUID>,
"clientSecret": <GUID>,
"subscriptionId": <GUID>,
"tenantId": <GUID>
}
SUBSCRIPTION
(this is the subscription) , GROUP
(this is the resource group name), WORKSPACE
(this is the project name), and KEY_VAULT_NAME
to GitHub.Follow the instructions to create a custom env with the packages needed here
upload existing docker
optionruntime\docker
Update the deployment.yml image to the newly created environemnt. You can find the name under Azure container registry
in the environment details page.
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.