Get started with agent development in Microsoft Foundry
In this exercise, you'll use Microsoft Foundry to start developing an AI agent that provides information and expertise on the history of computing.
Note: Many components of Microsoft Foundry, including the Microsoft Foundry portal, are subject to continual development. This reflects the fast-moving nature of artificial intelligence technology. Some elements of your user experience may differ from the images and descriptions in this exercise!
This exercise should take approximately 20 minutes to complete.
Create a Microsoft Foundry project
Microsoft Foundry uses projects to organize models, resources, data, and other assets used to develop an AI solution.
-
In a web browser, open Microsoft Foundry{:target="_blank"} at
https://ai.azure.comand sign in using your Azure credentials. Close any tips or quick start panes that are opened the first time you sign in, and if necessary use the Foundry logo at the top left to navigate to the home page. -
If it is not already enabled, in the tool bar the top of the page, enable the New Foundry option. Then, if prompted, create a new project with a unique name; expanding the Advanced options area to specify the following settings for your project:
- Foundry resource: Enter a valid name for your Foundry resource.
- Subscription: Your Azure subscription
- Resource group: Create or select a resource group
- Region: Select any of the AI Foundry recommended regions
-
Select Create. Wait for your project to be created. It may take a few minutes. After creating or selecting a project in the new Foundry portal, it should open in a page similar to the following image:

Deploy a model
At the heart of every AI agent, there's a large language model (LLM). Let's find one in the Foundry models catalog.
-
Now you're ready to Start building. Select Find models (or on the Discover page, select the Models tab) to view the Microsoft Foundry model catalog.
Microsoft Foundry provides a large collection of models from Microsoft, OpenAI, and other providers, that you can use in your AI apps and agents.

-
Search for and select the
gpt-4.1-minimodel, and view the page for this model, which describes its features and capabilities.
-
Use the Deploy button to deploy the model using the default settings. Deployment may take a minute or so.
Tip: Model deployments are subject to regional quotas. If you don't have enough quota to deploy the model in your project's region, you can use a different model - such as gpt-4.1-nano, or gpt-4o-mini. Alternatively, you can create a new project in a different region.
-
When the model has been deployed, view the model playground page that is opened, in which you can chat with the model.

Chat with the model
You can use the playground to explore the model by chatting with it.
-
Use the button at the bottom of the left navigation pane to hide it and give yourself more room to work with.
-
In the Chat pane, enter a prompt such as
Who was Ada Lovelace?, and review the response.
-
Enter a follow-up prompt, such as
Tell me more about her work with Charles Babbage.and review the response.Note: Generative AI chat applications often include the conversation history in the prompt; so the context of the conversation is retained between messages. In this case, "her" is interpreted as referring to Ada Lovelace.
-
At the top-right of the chat pane, use the New chat button to restart the conversation. This removes all conversation history.
-
Enter a new prompt, such as
Tell me about the ELIZA chatbot.and view the response. -
Continue the conversation with prompts such as
How does it compare with modern LLMs?.
Specify instructions in a system prompt
To support specific use cases, you should use a system prompt to provide the model with instructions that guide its responses. You can use the system prompt to give the model a specific focus or role, and provide guidelines about format, style, and constraints about what the model should and should not include in its responses.
-
In the model playground, at the top-right of the chat pane, use the New chat button to restart the conversation and remove the conversation history.
-
In the pane on the left, in the Instructions text area, change the system prompt to:
You are an expert in the history of computing and AI. You only answer questions about significant people and events in the development of computing, and about notable vintage computers. Do not engage in conversations on any topic that is unrelated to computing history. -
Now enter a new user prompt related to computing history, such as
What was Alan Turing's contribution to the development of AI?Review the response, which should provide some general guidance about expense claims.
-
Try asking an "off-topic" question, such as
What's the capital of Spain?; and view the response.
Add a web_search tool
So far, the model has answered questions based on the data with which it was trained. While this is useful, that leaves out a lot of current information on the web; which might help the model give more relevant answers.
We can use tools to give models access to external data sources, and to perform custom tasks. Let's add a tool that enables the model to search the Web for up-to-date information.
-
In the pane on the left, under the instructions, expand the Tools section if it is not already expanded.
-
In the Add drop-down list, select Web search. Then read the information about the tool and add it.
-
In the chat pane, enter the prompt
Find a vintage computer store near Seattle(or your local city!) and review the response.The model should have searched the Web for vintage computer stores near the specific city.
Save the model configuration as an agent
While you can implement generative AI apps using a standalone model, to create a fully agentic AI experience, you need to encapsulate the model, its instructions, and any tool configuration that provides additional functionality, in an agent.
-
In the model playground, at the top right select Save as agent. Then, when prompted, name your new agent
computing-historian.When the agent is created, it opens in a new playground specifically for working with agents.

-
In the pane on the right, view the YAML tab, which contains the definition for your agent. Note that its definition includes the model, its parameter settings, and the instructions you specified - similar to this:
metadata: logo: Avatar_Default.svg microsoft.voice-live.enabled: "false" object: agent.version id: computing-historian:1 name: computing-historian version: "1" description: "" created_at: 1776550090 definition: kind: prompt model: gpt-4.1-mini instructions: You are an expert in the history of computing and AI. You only answer questions about significant people and events in the development of computing, and about notable vintage computers. Do not engage in conversations on any topic that is unrelated to computing history. temperature: 1 top_p: 1 tools: - type: web_search status: active -
Switch back to the Chat tab, and enter the prompt
Who are you?The response should indicate that the agent is "aware" of its role as a computing historian.
Preview the agent
Now you have a working agent, you can preview it in a basic web chat application.
-
At the top of the chat pane, in the Preview drop-down list, select Preview agent.
A preview chat interface is opened in a new browser tab.
-
Enter a prompt, such as
What can you tell me about the Altair 8800?and view the response from your agent.
Summary
In this exercise, you explored how to deploy and chat with a generative AI model in Microsoft Foundry portal. You then configured instructions and tools before saving the model as an agent.
Next steps
This is the first in a series of lab exercises; save your work and continue to the next exercise if you're ready.
Tip: If you have finished exploring Microsoft Foundry, you should delete the Azure resources created in this exercise to avoid unnecessary utilization charges.