Azure AI Foundry is Microsoft’s unified platform for building, evaluating, and deploying AI applications. Previously known as Azure AI Studio, Foundry brings together model hosting, prompt engineering, orchestration flows, and evaluation tooling into a single workspace. If you have worked with individual Azure AI services before, Foundry is where they converge.
What Azure AI Foundry Provides
The platform is organized around several core capabilities:
- Model Catalog — Access hundreds of models from OpenAI, Meta, Mistral, Cohere, and Microsoft. Deploy them as managed endpoints without managing infrastructure.
- Prompt Flow — A visual and code-based tool for building LLM orchestration pipelines. Connect prompts, tools, and code steps into repeatable flows.
- Evaluation — Built-in metrics for measuring response quality, groundedness, coherence, and safety. Compare model outputs across configurations.
- AI Services Integration — Direct access to Azure OpenAI, Speech, Vision, Document Intelligence, and Search from within the same project.
- Responsible AI — Content filters, safety evaluations, and monitoring are integrated into the deployment pipeline.
Setting Up Your First Project
Step 1: Create an AI Foundry Hub
A hub is the top-level resource that provides shared infrastructure (networking, identity, storage) for your AI projects.
- Go to the Azure portal and search for Azure AI Foundry.
- Click Create a hub. Select your subscription and resource group.
- Choose a region that supports the models you plan to use (East US and West Europe have the broadest model availability).
- The hub automatically provisions a Storage Account and Key Vault.
Step 2: Create a Project
Within the hub, create a project. A project scopes your work: it has its own model deployments, prompt flows, datasets, and evaluation runs.
- Open your hub in the Foundry portal at ai.azure.com.
- Click New project and give it a name.
- The project inherits the hub’s networking and identity settings.
Step 3: Deploy a Model
Navigate to the Model catalog within your project. For a first deployment, GPT-4o is a practical choice.
- Search for gpt-4o in the catalog.
- Click Deploy. Choose Serverless API for a pay-per-token deployment with no reserved capacity, or Managed Compute for dedicated throughput.
- Set the deployment name (e.g.,
gpt-4o-main). - Configure rate limits and content filters. The default content filter blocks high-severity harmful content.
- Click Deploy. The endpoint is ready in under a minute for serverless deployments.
Once deployed, you can test the model immediately in the Playground by sending prompts and adjusting parameters like temperature and max tokens.
Building a Prompt Flow
Prompt Flow lets you build multi-step LLM pipelines. A typical flow takes user input, enriches it with context, sends it to a model, and processes the response.
Create a Basic Flow
- Go to Prompt Flow in your project and click Create.
- Choose Standard flow (or start from a template like “Chat with your data”).
- The flow editor opens with a visual DAG (directed acyclic graph).
A minimal flow has three nodes:
- Input node — Defines the user’s question as a string input.
- LLM node — Contains a prompt template with placeholders and points to your deployed model.
- Output node — Returns the model’s response.
Configure the LLM Node
In the LLM node, write your prompt template using Jinja2 syntax:
system:You are a helpful assistant that answers questions about Azure cloud services.Keep answers concise and technical.
user:{{question}}Connect the node to your gpt-4o-main deployment. Set temperature to 0.7 for balanced responses.
Test the Flow
Click Run and provide a test input like “What is the difference between Azure Functions and Container Apps?” The flow executes each node in sequence, and you can inspect the output at every step.
Evaluating Your Flow
Evaluation is where Foundry adds significant value over manual testing. Create an evaluation run to measure your flow against a dataset.
- Prepare a test dataset (JSON or CSV) with input questions and expected answers.
- Go to Evaluation and click New evaluation.
- Select your flow and dataset. Choose metrics: Groundedness, Relevance, Coherence, Fluency.
- Run the evaluation. Foundry scores each response and provides aggregate metrics.
Use evaluation results to iterate on your prompts, adjust model parameters, or compare different models for the same task.
From Prototype to Production
Once your flow produces satisfactory results:
- Deploy the flow as a managed online endpoint with a single click.
- Monitor token usage, latency, and error rates in the built-in dashboard.
- Add content safety filters appropriate for your use case.
- Integrate the endpoint into your application using the REST API or Azure SDKs.
Azure AI Foundry removes the infrastructure overhead from AI development. You get model hosting, prompt engineering, evaluation, and deployment in one place. Start with the playground to experiment, build a prompt flow to formalize your logic, evaluate to measure quality, and deploy when ready.