Develop a multi-agent solution
In this exercise, you’ll create a project that orchestrates two AI agents using the Semantic Kernel SDK. An Incident Manager agent will analyze service log files for issues. If an issue is found, the Incident Manager will recommend a resolution action and a DevOps Assistant agent will receive the recommendation and invoke the corrective function to perform the resolution. The Incident Manager agent will then review the updated logs to make sure the resolution was successful.
Note: For this exercise, four sample log files are provided. The DevOps Assistant agent code only updates the sample log files with some example log messages.
This exercise should take approximately 30 minutes to complete.
Create an Azure AI Foundry project
Let’s start by creating an Azure AI Foundry project.
-
In a web browser, open the Azure AI Foundry portal at
https://ai.azure.com
and sign in using your Azure credentials. Close any tips or quick start panes that are opened the first time you sign in, and if necessary use the Azure AI Foundry logo at the top left to navigate to the home page, which looks similar to the following image: -
In the home page, select + Create project.
-
In the Create a project wizard, enter a suitable project name (for example,
my-ai-project
) and if an existing hub is suggested, choose the option to create a new one. Then review the Azure resources that will be automatically created to support your hub and project. -
Select Customize and specify the following settings for your hub:
- Hub name: A unique name - for example
my-ai-hub
- Subscription: Your Azure subscription
- Resource group: Create a new resource group with a unique name (for example,
my-ai-resources
), or select an existing one - Location: Select a region from the following:*
- australiaeast
- eastus
- eastus2
- francecentral
- swedencentral
- Connect Azure AI Services or Azure OpenAI: Create a new AI Services resource with an appropriate name (for example,
my-ai-services
) or use an existing one - Connect Azure AI Search: Skip connecting
* Model quotas are constrained at the tenant level by regional quotas. In the event of a quota limit being reached later in the exercise, there’s a possibility you may need to create another project in a different region.
- Hub name: A unique name - for example
-
Select Next and review your configuration. Then select Create and wait for the process to complete.
-
When your project is created, close any tips that are displayed and review the project page in Azure AI Foundry portal, which should look similar to the following image:
-
In the project overview page, in the Project details area, note the Project connection string. Later, you’ll use this connection string to connect to your project in a client application.
Deploy a generative AI model
Now you’re ready to deploy a generative AI language model to support your agents. In this example, you’ll use the same model to support both agents.
-
In the pane on the left for your project, in the My assets section, select the Models + endpoints page.
-
In the Models + endpoints page, in the Model deployments tab, in the + Deploy model menu, select Deploy base model.
-
Search for the gpt-4 model in the list, and then select and confirm it.
-
Deploy the model with the following settings by selecting Customize in the deployment details:
- Deployment name: A unique name for your model deployment - for example
gpt-4
(remember the name you choose - you’ll need it later) - Deployment type: Standard
- Model version: 0613
- Connected AI resource: Select your Azure OpenAI resource connection
- Tokens per Minute Rate Limit (thousands): 5K
- Content filter: DefaultV2
- Enable dynamic quota: Disabled
Note: Reducing the TPM helps avoid over-using the quota available in the subscription you are using. 5,000 TPM is sufficient for the data used in this exercise.
- Deployment name: A unique name for your model deployment - for example
-
Wait for the deployment to complete.
Create an AI Agent client app
Now you’re ready to create a client app that defines an agent and a custom function. Some code is provided for you in a GitHub repository.
Prepare the environment
-
Open a new browser tab (keeping the Azure AI Foundry portal open in the existing tab). Then in the new tab, browse to the Azure portal at
https://portal.azure.com
; signing in with your Azure credentials if prompted.Close any welcome notifications to see the Azure portal home page.
-
Use the [>_] button to the right of the search bar at the top of the page to create a new Cloud Shell in the Azure portal, selecting a PowerShell environment with no storage in your subscription.
The cloud shell provides a command line interface in a pane at the bottom of the Azure portal. You can resize or maximize this pane to make it easier to work in.
Note: If you have previously created a cloud shell that uses a Bash environment, switch it to PowerShell.
-
In the PowerShell pane, enter the following commands to clone the GitHub repo containing the code files for this exercise:
rm -r ai-agents -f git clone https://github.com/MicrosoftLearning/mslearn-ai-agents ai-agents
Tip: As you enter commands into the cloud shell, the output may take up a large amount of the screen buffer and the cursor on the current line may be obscured. You can clear the screen by entering the
cls
command to make it easier to focus on each task. -
Enter the following command to install the required version of Python in the cloud shell:
sh ~/ai-agents/Labfiles/update-python.sh
-
After the installation is complete, in the cloud shell toolbar, in the Settings menu, select Go to Classic version (this starts a new session, and is required to use the code editor).
Ensure you've switched to the classic version of the cloud shell and a new session has started before continuing.
-
Enter the following command to change the working directory to the folder containing the code files and list them all.
cd ai-agents/Labfiles/05-agent-orchestration/Python ls -a -l
The provided files include application code and a file for configuration settings.
Configure the application settings
-
In the cloud shell command line pane, enter the following command to install the libraries you’ll use:
pip install python-dotenv azure-identity semantic-kernel[azure]
Note: Installing semantic-kernel[azure] automatically installs a semantic kernel-compatible version of azure-ai-projects.
-
Enter the following command to edit the configuration file that is provided:
code .env
The file is opened in a code editor.
-
In the code file, replace the your_project_connection_string placeholder with the connection string for your project (copied from the project Overview page in the Azure AI Foundry portal), and the your_model_deployment placeholder with the name you assigned to your gpt-4 model deployment.
-
After you’ve replaced the placeholders, use the CTRL+S command to save your changes and then use the CTRL+Q command to close the code editor while keeping the cloud shell command line open.
Create AI agents
Now you’re ready to create the agents for your multi-agent solution! Let’s get started!
-
Enter the following command to edit the agent_chat.py file:
code agent_chat.py
- Review the code in the file, noting that it contains:
- Constants that define the names and instructions for your two agents.
- A main function where most of the code to implement your multi-agent solution will be added.
- A SelectionStrategy class which you’ll use to implement the logic required to determine which agent should be selected for each turn in the conversation.
- An ApprovalTerminationStrategy class, which you’ll use to implement the logic needed to determine when the conversation to end.
- A DevopsPlugin class that contains functions to perform devops operations.
- A LogFilePlugin class that contains functions to read and write log files.
First, you’ll create the Incident Manager agent, which will analyze service log files, identify potential issues, and recommend resolution actions or escalate issues when necessary.
-
Note the INCIDENT_MANAGER_INSTRUCTIONS string. These are the instructions for your agent.
-
In the main function, find the comment Create the incident manager agent on the Azure AI agent service, and add the following code to create an Azure AI Agent.
# Create the incident manager agent on the Azure AI agent service incident_agent_definition = await client.agents.create_agent( model=ai_agent_settings.model_deployment_name, name=INCIDENT_MANAGER, instructions=INCIDENT_MANAGER_INSTRUCTIONS )
This code creates the agent definition on your Azure AI Project client.
-
Find the comment Create a Semantic Kernel agent for the Azure AI incident manager agent, and add the following code to create a Semantic Kernel agent based on the Azure AI Agent definition.
# Create a Semantic Kernel agent for the Azure AI incident manager agent agent_incident = AzureAIAgent( client=client, definition=incident_agent_definition, plugins=[LogFilePlugin()] )
This code creates the Semantic Kernel agent with access to the LogFilePlugin. This plugin allows the agent to read the log file contents.
Now let’s create the second agent, which will respnd to issues and perform DevOps operations to resolve them.
-
At the top of the code file, take a moment to observe the DEVOPS_ASSISTANT_INSTRUCTIONS string. These are the instructions you’ll provide to the new DevOps assistant agent.
-
Find the comment Create the devops agent on the Azure AI agent service, and add the following code to create an Azure AI Agent definition:
# Create the devops agent on the Azure AI agent service devops_agent_definition = await client.agents.create_agent( model=ai_agent_settings.model_deployment_name, name=DEVOPS_ASSISTANT, instructions=DEVOPS_ASSISTANT_INSTRUCTIONS, )
-
Find the comment Create a Semantic Kernel agent for the devops Azure AI agent, and add the following code to create a Semantic Kernel agent based on the Azure AI Agent definition.
# Create a Semantic Kernel agent for the devops Azure AI agent agent_devops = AzureAIAgent( client=client, definition=devops_agent_definition, plugins=[DevopsPlugin()] )
The DevopsPlugin allows the agent to simulate devops tasks, such as restarting the service or rolling back a transaction.
Define group chat strategies
INow you need to provide the logic used to determine which agent should be selected to take the next turn in a conversation, and when the conversation should be ended.
Let’s start with the SelectionStrategy, identifies which agent should take the next turn.
-
In the SelectionStrategy class (below the main function), find the comment Select the next agent that should take the next turn in the chat, and add the following code to define a selection function:
# Select the next agent that should take the next turn in the chat async def select_agent(self, agents, history): """"Check which agent should take the next turn in the chat.""" # The Incident Manager should go after the User or the Devops Assistant if (history[-1].name == DEVOPS_ASSISTANT or history[-1].role == AuthorRole.USER): agent_name = INCIDENT_MANAGER return next((agent for agent in agents if agent.name == agent_name), None) # Otherwise it is the Devops Assistant's turn return next((agent for agent in agents if agent.name == DEVOPS_ASSISTANT), None)
This code runs on every turn to determine which agent should respond, checking the chat history to see who last responded.
Now let’s implement the ApprovalTerminationStrategy class to help signal when the goal is complete and the conversation can be ended.
-
In the ApprovalTerminationStrategy class, find the comment End the chat if the agent has indicated there is no action needed, and add the following code to define the termination function:
# End the chat if the agent has indicated there is no action needed async def should_agent_terminate(self, agent, history): """Check if the agent should terminate.""" return "no action needed" in history[-1].content.lower()
The kernel invokes this function after the agent’s response to determine if the completion criteria is complete. In this case, the goal is met when the incident manager responds with “No action needed.” This phrase is defined in the incident manager agent instructions.
Implement the group chat
Now that you have two agents, and strategies to help them take turns and end a chat, you can implement the group chat.
-
Back up in the main function, find the comment Add the agents to a group chat with a custom termination and selection strategy, and add the following code to create the group chat:
# Add the agents to a group chat with a custom termination and selection strategy chat = AgentGroupChat( agents=[agent_incident, agent_devops], termination_strategy=ApprovalTerminationStrategy( agents=[agent_incident], maximum_iterations=10, automatic_reset=True ), selection_strategy=SelectionStrategy(agents=[agent_incident,agent_devops]), )
In this code, you create an agent group chat object with the incident manager and devops agents. You also define the termination and selection strategies for the chat. Notice that the ApprovalTerminationStrategy is tied to the incident manager agent only, and not the devops agent. This makes the incident manager agent is responsible for signaling the end of the chat. The SelectionStrategy includes all agents that should take a turn in the chat.
Note that the automatic reset flag will automatically clear the chat when it ends. This way, the agent can continue analyzing the files without the chat history object using too many unnecessary tokens.
-
Find the comment Append the current log file to the chat, and add the following code to add the most recently read log file text to the chat:
# Append the current log file to the chat await chat.add_chat_message(logfile_msg) print()
-
Find the comment Invoke a response from the agents, and add the following code to invoke the group chat:
# Invoke a response from the agents async for response in chat.invoke(): if response is None or not response.name: continue print(f"{response.content}")
This is the code that triggers the chat. Since the log file text has been added as a message, the selection strategy will determine which agent should read and respond to it and then the conversation will continue between the agents until the conditions of the termination strategy are met or the maximum number of iterations is reached.
-
Use the CTRL+S command to save your changes to the code file. You can keep it open (in case you need to edit the code to fix any errors) or use the CTRL+Q command to close the code editor while keeping the cloud shell command line open.
Sign into Azure and run the app
Now you’re ready to run your code and watch your AI agents collaborate.
-
In the cloud shell command line pane, enter the following command to sign into Azure.
az login
You must sign into Azure - even though the cloud shell session is already authenticated.
-
When prompted, follow the instructions to open the sign-in page in a new tab and enter the authentication code provided and your Azure credentials. Then complete the sign in process in the command line, selecting the subscription containing your Azure AI Foundry hub if prompted.
-
After you have signed in, enter the following command to run the application:
python agent_chat.py
You should see some output similar to the following:
INCIDENT_MANAGER > /home/.../sample_logs/log1.log | Restart service ServiceX DEVOPS_ASSISTANT > Service ServiceX restarted successfully. INCIDENT_MANAGER > No action needed. INCIDENT_MANAGER > /home/.../sample_logs/log2.log | Rollback transaction for transaction ID 987654. DEVOPS_ASSISTANT > Transaction rolled back successfully. INCIDENT_MANAGER > No action needed. INCIDENT_MANAGER > /home/.../sample_logs/log3.log | Increase quota. DEVOPS_ASSISTANT > Successfully increased quota. (continued)
-
Verify that the log files are updated with resolution operation messages from the DevopsAssistant.
For example, log1.log should have the following log messages appended:
[2025-02-27 12:43:38] ALERT DevopsAssistant: Multiple failures detected in ServiceX. Restarting service. [2025-02-27 12:43:38] INFO ServiceX: Restart initiated. [2025-02-27 12:43:38] INFO ServiceX: Service restarted successfully.
Summary
In this exercise, you used the Azure AI Agent Service and Semantic Kernel SDK to create AI incident and devops agents that can automatically detect issues and apply resolutions. Great work!
Clean up
If you’ve finished exploring Azure AI Agent Service, you should delete the resources you have created in this exercise to avoid incurring unnecessary Azure costs.
-
Return to the browser tab containing the Azure portal (or re-open the Azure portal at
https://portal.azure.com
in a new browser tab) and view the contents of the resource group where you deployed the resources used in this exercise. -
On the toolbar, select Delete resource group.
-
Enter the resource group name and confirm that you want to delete it.