Explore AI Models in Azure AI Foundry
Learn about different types of AI models and their capabilities by exploring the model catalog and testing them in interactive playgrounds. This tutorial guides you through discovering, evaluating, and comparing models for your specific use cases.What you’ll learn
By the end of this tutorial, you will understand:- How to navigate the Azure AI Foundry model catalog
- The differences between various AI model types
- How to test models with your own prompts
- How to compare model performance for your use case
- Best practices for model selection
Prerequisites
- Completion of the Getting Started tutorial
- An Azure AI Foundry project with basic access
- 45 minutes of focused time
Step 1: Navigate the model catalog
Let’s start by exploring what models are available in Azure AI Foundry.- In your Azure AI Foundry project, go to “Models”
- Click “Explore all models” to see the full catalog
- Notice the different categories:
- Language models - For text generation and understanding
- Vision models - For image processing and generation
- Audio models - For speech and audio processing
- Multimodal models - For combining text, images, and audio
Step 2: Compare language models
Let’s compare different language models to understand their strengths.Test a large, capable model (GPT-4o)
- Find GPT-4o in the catalog
- Click “Try in playground” (or deploy if needed)
- Test it with this complex reasoning prompt:
- Handles complex constraints
- Provides detailed, structured responses
- Shows strong reasoning capabilities
Test a smaller, efficient model (GPT-4o mini)
- Switch to GPT-4o mini in the playground
- Try the same dinner party prompt
- Compare the responses:
- Speed difference
- Level of detail
- Accuracy of the solution
Test an open-source alternative
- Try Llama 3.1 8B or similar open-source model
- Use the same prompt
- Observe the differences in:
- Response style
- Completeness
- Processing time
- Larger models often provide more detailed responses
- Smaller models are faster and more cost-effective
- Open-source models offer customization opportunities
Step 3: Explore specialized capabilities
Code generation
Test how different models handle programming tasks: Prompt for all models:- Code quality and completeness
- Documentation clarity
- Error handling approach
- Example quality
Creative writing
Test creative capabilities with this prompt:- Creativity and originality
- Emotional depth
- Narrative structure
- Writing style
Analysis and reasoning
Test analytical capabilities:- Mathematical accuracy
- Business reasoning
- Consideration of trade-offs
- Clarity of recommendations
Step 4: Test vision capabilities
If available, explore vision-enabled models:Upload an image and test understanding
- Go to GPT-4o (or another vision-enabled model)
-
Upload an image - try different types:
- A photograph of a scene
- A chart or graph
- A handwritten note
- An artwork
-
Ask questions like:
- “Describe what you see in detail”
- “What mood or emotion does this convey?”
- “If this is a graph, what insights can you draw?”
Test multimodal reasoning
Try combining image analysis with other tasks:Step 5: Evaluate for your specific use case
Now think about your own application needs:Define your requirements
Consider:- Response quality needs - How accurate must responses be?
- Speed requirements - Real-time or batch processing?
- Cost constraints - High volume or occasional use?
- Customization needs - Generic or domain-specific?
Test with your domain
Create prompts that match your actual use case: For customer service:Step 6: Document your findings
Create a comparison table for your testing:| Model | Speed | Quality | Cost | Best For |
|---|---|---|---|---|
| GPT-4o | Slow | Excellent | High | Complex reasoning |
| GPT-4o mini | Fast | Good | Low | High volume tasks |
| Llama 3.1 | Medium | Good | Medium | Customizable scenarios |
What you’ve learned
Through this hands-on exploration, you’ve discovered: ✅ Model variety - Different models excel at different tasks ✅ Performance trade-offs - Speed vs. quality vs. cost considerations ✅ Capability assessment - How to evaluate models for your specific needs ✅ Testing methodology - Systematic approaches to model comparisonKey insights to remember
No one-size-fits-all: Different tasks may require different models, even within the same application. Context matters: Model performance varies significantly based on prompt quality and task complexity. Cost vs. capability: Smaller models can often handle simpler tasks effectively at much lower cost. Testing is essential: Always test models with your actual use cases before making decisions.Next steps
Now that you understand model capabilities, you’re ready to:Fine-tune a Model
Customize a model for your specific domain and use case
Evaluate Performance
Set up systematic evaluation and monitoring
This tutorial focused on experiential learning - you learned about AI models by actually using them. The hands-on comparison gave you intuitive understanding of their capabilities and trade-offs. Use this knowledge to make informed decisions about which models to deploy in your applications.

