Generate vector embeddings with Azure OpenAI and store them in Azure Cosmos DB for NoSQL
Azure OpenAI provides access to OpenAI’s advanced language models, including the text-embedding-ada-002
, text-embedding-3-small
, and text-embedding-3-large
models. By leveraging one of these models, you can generate vector representations of textual data, which can be stored in a vector store like Azure Cosmos DB for NoSQL. This facilitates efficient and accurate similarity searches, significantly enhancing a copilot’s ability to retrieve relevant information and provide contextually rich interactions.
In this lab, you will create an Azure OpenAI service and deploy an embedding model. You will then use Python code to create Azure OpenAI and Cosmos DB clients using their respective Python SDKs to generate vector representations of product descriptions and write them into your database.
🛑 The previous exercise in this module is a prerequisite for this lab. If you still need to complete that exercise, please finish it before continuing, as it provides the infrastructure required for this lab.
Create an Azure OpenAI service
Azure OpenAI provides REST API access to OpenAI’s powerful language models. These models can be easily adapted to your specific task including but not limited to content generation, summarization, image understanding, semantic search, and natural language to code translation.
-
In a new web browser window or tab, navigate to the Azure portal (
portal.azure.com
). -
Sign into the portal using the Microsoft credentials associated with your subscription.
-
Select Create a resource, search for Azure OpenAI, and then create a new Azure OpenAI resource with the following settings, leaving all remaining settings to their default values:
Setting Value Subscription Your existing Azure subscription Resource group Select an existing or create a new resource group Region Choose an available region that supports the text-embedding-3-small
model from the list of supporting regions.Name Enter a globally unique name Pricing Tier Choose Standard 0 📝 Your lab environments may have restrictions preventing you from creating a new resource group. If that is the case, use the existing pre-created resource group.
-
Wait for the deployment task to complete before continuing with the next task.
Deploy an embedding model
To use Azure OpenAI to generate embeddings, you must first deploy an instance of the desired embedding model within your service.
-
Navigate to your newly created Azure OpenAI service in the Azure portal (
portal.azure.com
). -
On the Overview page of the Azure OpenAI service, launch Azure AI Foundry by selecting the Go to Azure AI Foundry portal link on the toolbar.
-
In Azure AI Foundry, select Deployments from the left-hand menu.
-
On the Model deployments page, select Deploy model and select Deploy base model from the dropdown.
-
From the list of models, select
text-embedding-3-small
.💡 You can filter the list to display only Embeddings models using the inference tasks filter.
📝 If you do not see the
text-embedding-3-small
model, you may have selected an Azure region that does not currently support that model. In this case, you can use thetext-embedding-ada-002
model for this lab. Both models generate vectors with 1536 dimensions, so no changes are required to the container vector policy you defined on theProducts
container in Azure Cosmos DB. -
Select Confirm to deploy the model.
-
On the Model deployments page in Azure AI Foundry, note the Name of the
text-embedding-3-small
model deployment, as you will need this later in this exercise.
Deploy a chat completion model
In addition to the embedding model, you will need a chat completion model for your copilot. You will use OpenAI’s gpt-4o
large language model to generate responses from your copilot.
-
While still on the Model deployments page in Azure AI Foundry, select the Deploy model button again and choose Deploy base model from the dropdown.
-
Select the gpt-4o chat completion model from the list.
-
Select Confirm to deploy the model.
-
On the Model deployments page in Azure AI Foundry, note the Name of the
gpt-4o
model deployment, as you will need this later in this exercise.
Assign the Cognitive Services OpenAI User RBAC role
To allow your user identity to interact with the Azure OpenAI service, you can assign your account the Cognitive Services OpenAI User role. Azure OpenAI Service supports Azure role-based access control (Azure RBAC), an authorization system for managing individual access to Azure resources. Using Azure RBAC, you assign different team members different levels of permissions based on their needs for a given project.
📝 Microsoft Entra ID’s Role-Based Access Control (RBAC) for authenticating against Azure services like Azure OpenAI enhances security through precise access controls tailored to user roles, effectively reducing unauthorized access risks. Streamlining secure access management using Entra ID RBAC makes a more efficient and scalable solution for leveraging Azure services.
-
In the Azure portal (
portal.azure.com
), navigate to your Azure OpenAI resource. -
Select Access Control (IAM) on the left navigation pane.
-
Select Add, then select Add role assignment.
-
On the Role tab, select the Cognitive Services OpenAI User role, then select Next.
-
On the Members tab, select assign access to a user, group, or service principal, and select Select members.
-
In the Select members dialog, search for your name or email address and select your account.
-
On the Review + assign tab, select Review + assign to assign the role.
Create a Python virtual environment
Virtual environments in Python are essential for maintaining a clean and organized development space, allowing individual projects to have their own set of dependencies, isolated from others. This prevents conflicts between different projects and ensures consistency in your development workflow. By using virtual environments, you can manage package versions easily, avoid dependency clashes, and keep your projects running smoothly. It’s a best practice that keeps your coding environment stable and dependable, making your development process more efficient and less prone to issues.
-
Using Visual Studio Code, open the folder into which you cloned the lab code repository for Build copilots with Azure Cosmos DB learning module.
-
In Visual Studio Code, open a new terminal window and change directories to the
python/07-build-copilot
folder. -
Create a virtual environment named
.venv
by running the following command at the terminal prompt:python -m venv .venv
The above command will create a
.venv
folder under the07-build-copilot
folder, which will provide a dedicated Python environment for the exercises in this lab. -
Activate the virtual environment by selecting the appropriate command for your OS and shell from the table below and executing it at the terminal prompt.
Platform Shell Command to activate virtual environment POSIX bash/zsh source .venv/bin/activate
fish source .venv/bin/activate.fish
csh/tcsh source .venv/bin/activate.csh
pwsh .venv/bin/Activate.ps1
Windows cmd.exe .venv\Scripts\activate.bat
PowerShell .venv\Scripts\Activate.ps1
-
Install the libraries defined in
requirements.txt
:pip install -r requirements.txt
The
requirements.txt
file contains a set of Python libraries you will use throughout this lab.Library Version Description azure-cosmos
4.9.0 Azure Cosmos DB SDK for Python - Client library azure-identity
1.19.0 Azure Identity SDK for Python fastapi
0.115.5 Web framework for building APIs with Python openai
1.55.2 Provides access to the Azure OpenAI REST API from Python apps. pydantic
2.10.2 Data validation using Python type hints. requests
2.32.3 Send HTTP requests. streamlit
1.40.2 Transforms Python scripts into interactive web apps. uvicorn
0.32.1 An ASGI web server implementation for Python. httpx
0.27.2 A next-generation HTTP client for Python.
Add a Python function to vectorize text
The Python SDK for Azure OpenAI provides access to both synchronous and asynchronous classes that can be used to create embeddings for textual data. This functionality can be encapsulated in a function in your Python code.
-
In the Explorer pane within Visual Studio Code, navigate to the
python/07-build-copilot/api/app
folder and open themain.py
file located within it.📝 This file will serve as the entry point to a backend Python API you will build in the next exercise. In this exercise, you will provide a handful of async functions that can be used to import data with embeddings into Azure Cosmos DB that will be leveraged by the API.
-
To use the asychronous Azure OpenAI SDK for Python, import the library by adding the following code to the top of the
main.py
file:from openai import AsyncAzureOpenAI
-
You will be accessing Azure OpenAI and Cosmos DB asynchronously using Azure authentication and the Entra ID RBAC roles you previously assigned to your user identity. Add the following line below the
openai
import statement at the top of the file to import the required classes from theazure-identity
library:from azure.identity.aio import DefaultAzureCredential, get_bearer_token_provider
📝 To ensure you can securely interact with Azure services from your API, you will use the Azure Identity SDK for Python. This approach allows you to avoid having to store or interact with keys from code, instead leveraging the RBAC roles you assigned to your account for access to Azure Cosmos DB and Azure OpenAI in the previous exercises.
-
Create variables to store the Azure OpenAI API version and endpoint, replacing the
<AZURE_OPENAI_ENDPOINT>
token with the endpoint value for your Azure OpenAI service. Also, create a variable for the name of your embedding model deployment. Insert the following code below theimport
statements in the file:# Azure OpenAI configuration AZURE_OPENAI_ENDPOINT = "<AZURE_OPENAI_ENDPOINT>" AZURE_OPENAI_API_VERSION = "2024-10-21" EMBEDDING_DEPLOYMENT_NAME = "text-embedding-3-small"
If your embedding deployment name differs, update the value assigned to the variable accordingly.
💡 The API version of
2024-10-21
was the latest GA release version as of the time of this writing. You can use that or a new version, if one is available. The API specs documentation contains a table with the latest API versions.📝 The
EMBEDDING_DEPLOYMENT_NAME
is the Name value you noted after deploying thetext-embedding-3-small
model in Azure AI Foundry. If you need to refer back to it, launch Azure AI Foundry, navigate to the Deployments page and locate the deployment whose Model name istext-embedding-3-small
. Then, copy the Name field value of that item. If you deployed thetext-embedding-ada-002
model, use the name for that deployment. -
Use the Azure Identity SDK for Python’s
DefaultAzureCredential
class to create an asynchronous credential for accessing Azure OpenAI and Azure Cosmos DB using Microsoft Entra ID RBAC authentication by inserting the following code below the variable declarations:# Enable Microsoft Entra ID RBAC authentication credential = DefaultAzureCredential()
-
To handle the creation of embeddings, insert the following, which adds a function to generate embeddings using an Azure OpenAI client:
async def generate_embeddings(text: str): """Generates embeddings for the provided text.""" # Create an async Azure OpenAI client async with AsyncAzureOpenAI( api_version = AZURE_OPENAI_API_VERSION, azure_endpoint = AZURE_OPENAI_ENDPOINT, azure_ad_token_provider = get_bearer_token_provider(credential, "https://cognitiveservices.azure.com/.default") ) as client: response = await client.embeddings.create( input = text, model = EMBEDDING_DEPLOYMENT_NAME ) return response.data[0].embedding
Creation of the Azure OpenAI client does not require the
api_key
value because it is retrieving a bearer token using the Azure Identity SDK’sget_bearer_token_provider
class. -
The
main.py
file should now look similar to the following:from openai import AsyncAzureOpenAI from azure.identity.aio import DefaultAzureCredential, get_bearer_token_provider # Azure OpenAI configuration AZURE_OPENAI_ENDPOINT = "<AZURE_OPENAI_ENDPOINT>" AZURE_OPENAI_API_VERSION = "2024-10-21" EMBEDDING_DEPLOYMENT_NAME = "text-embedding-3-small" # Enable Microsoft Entra ID RBAC authentication credential = DefaultAzureCredential() async def generate_embeddings(text: str): """Generates embeddings for the provided text.""" # Create an async Azure OpenAI client async with AsyncAzureOpenAI( api_version = AZURE_OPENAI_API_VERSION, azure_endpoint = AZURE_OPENAI_ENDPOINT, azure_ad_token_provider = get_bearer_token_provider(credential, "https://cognitiveservices.azure.com/.default") ) as client: response = await client.embeddings.create( input = text, model = EMBEDDING_DEPLOYMENT_NAME ) return response.data[0].embedding
-
Save the
main.py
file.
Test the embedding function
To ensure the generate_embeddings
function in the main.py
file is working correctly, you will add a few lines of code at the bottom of the file to allow it to be run directly. These lines allow you to execute the generate_embeddings
function from the command line, passing in the text to embed.
-
Add a main guard block containing a call to
generate_embeddings
at the bottom of themain.py
file:if __name__ == "__main__": import asyncio import sys async def main(): print(await generate_embeddings(sys.argv[1])) # Close open async credential sessions await credential.close() asyncio.run(main())
📝 The
if __name__ == "__main__":
block is commonly referred to as the main guard or entry point in Python. It ensures that certain code is only executed when the script is run directly, and not when it is imported as a module in another script. This practice helps in organizing code and makes it more reusable and modular. -
Save the
main.py
file, which should now look like:from openai import AsyncAzureOpenAI from azure.identity.aio import DefaultAzureCredential, get_bearer_token_provider # Azure OpenAI configuration AZURE_OPENAI_ENDPOINT = "<AZURE_OPENAI_ENDPOINT>" AZURE_OPENAI_API_VERSION = "2024-10-21" EMBEDDING_DEPLOYMENT_NAME = "text-embedding-3-small" # Enable Microsoft Entra ID RBAC authentication credential = DefaultAzureCredential() async def generate_embeddings(text: str): """Generates embeddings for the provided text.""" # Create an async Azure OpenAI client async with AsyncAzureOpenAI( api_version = AZURE_OPENAI_API_VERSION, azure_endpoint = AZURE_OPENAI_ENDPOINT, azure_ad_token_provider = get_bearer_token_provider(credential, "https://cognitiveservices.azure.com/.default") ) as client: response = await client.embeddings.create( input = text, model = EMBEDDING_DEPLOYMENT_NAME ) return response.data[0].embedding if __name__ == "__main__": import asyncio import sys async def main(): print(await generate_embeddings(sys.argv[1])) # Close open async credential sessions await credential.close() asyncio.run(main())
-
In Visual Studio Code, open a new integrated terminal window.
-
Before running the API, which will send requests to Azure OpenAI, you must log into Azure using the
az login
command. At the terminal window, run:az login
-
Complete the login process in your browser.
-
At the terminal prompt, change directories to
python/07-build-copilot
. -
Ensure the intgrated terminal window is running within your Python virutal environment by activating your virtual environment using a command from the table below, selecting the appropriate command for your OS and shell.
Platform Shell Command to activate virtual environment POSIX bash/zsh source .venv/bin/activate
fish source .venv/bin/activate.fish
csh/tcsh source .venv/bin/activate.csh
pwsh .venv/bin/Activate.ps1
Windows cmd.exe .venv\Scripts\activate.bat
PowerShell .venv\Scripts\Activate.ps1
-
At the terminal prompt, change directories to
api/app
, then execute the following command:python main.py "Hello, world!"
-
Observe the output in the terminal window. You should see an array of floating point number, which is the vector representation of the “Hello, world!” string. It should look similiar to the following abbreviated output:
[-0.019184619188308716, -0.025279032066464424, -0.0017195191467180848, 0.01884828321635723...]
Build a function for writing data to Azure Cosmos DB
Using the Azure Cosmos DB SDK for Python, you can create a function that allows upserting documents into your database. An upsert operation will update a record if a match is found and insert a new record if one is not.
-
Return to the open
main.py
file in Visual Studio Code and import the asyncCosmosClient
class from the Azure Cosmos DB SDK for Python by inserting the following line just below theimport
statements already in the file:from azure.cosmos.aio import CosmosClient
-
Add another import statement to reference the
Product
class from the models module in theapi/app
folder. TheProduct
class defines the shape of products in the Cosmic Works dataset.from models import Product
-
Create a new group of variables containing configuration values associated with Azure Cosmos DB and add them to the
main.py
file below the Azure OpenAI variables you inserted previously. Ensure you replace the<AZURE_COSMOSDB_ENDPOINT>
token with the endpoint for your Azure Cosmos DB account.# Azure Cosmos DB configuration AZURE_COSMOSDB_ENDPOINT = "<AZURE_COSMOSDB_ENDPOINT>" DATABASE_NAME = "CosmicWorks" CONTAINER_NAME = "Products"
-
Add a function named
upsert_product
for upserting (update or insert) documents into Cosmos DB, inserting the following code below thegenerate_embeddings
function in themain.py
file:async def upsert_product(product: Product): """Upserts the provided product to the Cosmos DB container.""" # Create an async Cosmos DB client async with CosmosClient(url=AZURE_COSMOSDB_ENDPOINT, credential=credential) as client: # Load the CosmicWorks database database = client.get_database_client(DATABASE_NAME) # Retrieve the product container container = database.get_container_client(CONTAINER_NAME) # Upsert the product await container.upsert_item(product)
-
Save the
main.py
file, which should now look like:from openai import AsyncAzureOpenAI from azure.identity.aio import DefaultAzureCredential, get_bearer_token_provider from azure.cosmos.aio import CosmosClient from models import Product # Azure OpenAI configuration AZURE_OPENAI_ENDPOINT = "<AZURE_OPENAI_ENDPOINT>" AZURE_OPENAI_API_VERSION = "2024-10-21" EMBEDDING_DEPLOYMENT_NAME = "text-embedding-3-small" # Azure Cosmos DB configuration AZURE_COSMOSDB_ENDPOINT = "<AZURE_COSMOSDB_ENDPOINT>" DATABASE_NAME = "CosmicWorks" CONTAINER_NAME = "Products" # Enable Microsoft Entra ID RBAC authentication credential = DefaultAzureCredential() async def generate_embeddings(text: str): """Generates embeddings for the provided text.""" # Create an async Azure OpenAI client async with AsyncAzureOpenAI( api_version = AZURE_OPENAI_API_VERSION, azure_endpoint = AZURE_OPENAI_ENDPOINT, azure_ad_token_provider = get_bearer_token_provider(credential, "https://cognitiveservices.azure.com/.default") ) as client: response = await client.embeddings.create( input = text, model = EMBEDDING_DEPLOYMENT_NAME ) return response.data[0].embedding async def upsert_product(product: Product): """Upserts the provided product to the Cosmos DB container.""" # Create an async Cosmos DB client async with CosmosClient(url=AZURE_COSMOSDB_ENDPOINT, credential=credential) as client: # Load the CosmicWorks database database = client.get_database_client(DATABASE_NAME) # Retrieve the product container container = database.get_container_client(CONTAINER_NAME) # Upsert the product await container.upsert_item(product) if __name__ == "__main__": import sys print(generate_embeddings(sys.argv[1]))
Vectorize sample data
You are now ready to test both the generate_embeddings
and upsert_document
functions together. To do this, you will overwrite the if __name__ == "__main__"
main guard block with code that downloads a sample data file containing Cosmic Works product information from GitHub and then vectorizes the description
field of each product, and upserts the documents into the Products
container in your Azure Cosmos DB database.
📝 This approach is being used to demonstrate the techniques for generating with Azure OpenAI and storing embeddings in Azure Cosmos DB. In a real-world scenario, however, a more robust approach, such as using an Azure Function triggered by the Azure Cosmos DB change feed would be more appropiate for handling adding embeddings to existing and new documents.
-
In the
main.py
file, overwrite theif __name__ == "__main__":
code block with the following:if __name__ == "__main__": import asyncio from models import Product import requests async def main(): product_raw_data = "https://raw.githubusercontent.com/solliancenet/microsoft-learning-path-build-copilots-with-cosmos-db-labs/refs/heads/main/data/07/products.json?v=1" products = [Product(**data) for data in requests.get(product_raw_data).json()] # Call the generate_embeddings function, passing in an argument from the command line. for product in products: print(f"Generating embeddings for product: {product.name}", end="\r") product.embedding = await generate_embeddings(product.description) await upsert_product(product.model_dump()) print("All products with vectorized descriptions have been upserted to the Cosmos DB container.") # Close open credential sessions await credential.close() asyncio.run(main())
-
At the open integrated terminal prompt in Visual Studio Code, run the
main.py
file again using the command:python main.py
-
Wait for the code execution to complete, indicated by a message indicating all products with vectorized descriptions have been upserted to the Cosmos DB container. It may take up to 10 - 15 minutes for the vectorization and data upsert process to complete for the 295 records in the products dataset. If not all products are inserted, you can rerun
main.py
using the command above to add the remaining products.
Review upserted sample data in Cosmos DB
-
Return to the Azure portal (
portal.azure.com
) and navigate to your Azure Cosmos DB account. -
Select the Data Explorer for the left navigation menu.
-
Expand the CosmicWorks database and Products container, then select Items under the container.
-
Select several random documents within the container and ensure the
embedding
field is populated with a vector array.