Forum Discussion
Understanding Azure AI Services: Navigating the Confusion
Azure AI Services offers powerful capabilities for integrating artificial intelligence into applications, but the structure and relationships between different services can be confusing. This guide clarifies the Azure AI ecosystem based on hands-on exploration and provides solutions to common issues.
The Azure AI Services Ecosystem
Azure's AI offerings are organized into several key services, each with specific purposes:
1. Azure AI Search (formerly Azure Cognitive Search)
Purpose: Provides search capabilities for databases, including vector databases.
Key Features:
- Requires an embedding model (like OpenAI Ada) for vector search
- Provides database storage for searchable content
- Offers multiple search algorithms: hybrid, semantic, or vector-based
Common Issues & Solutions:
- When using AI Search with cognitive skills, you might encounter throttling errors when too many skills execute in parallel
- For throttling issues, consider using the SearchIndexingBufferedSender for automatic retry or reset and rerun the indexer
- Invalid skill inputs can cause warnings like "Skill input was invalid" - ensure required inputs like text and languageCode are properly provided
2. Azure AI Services (the service itself)
Purpose: Provides a collection of AI capabilities including text, audio, and visual processing.
Key Features:
- Creates text, audio, and visual model services
- Includes access to an AI Foundry portal with limited models (62 chat completion models)
Confusion Point: Despite mentioning text, audio, and visual capabilities, the AI Foundry portal accessed through this service only provides 62 chat completion models, creating a disconnect between service description and actual capabilities.
3. Azure OpenAI Service
Purpose: Provides access to OpenAI models within Azure's secure environment.
Key Features:
- Dedicated to OpenAI models specifically
- Enterprise-grade security and compliance
Best Practices:
- Use provisioned throughput (PTU) for production workloads to avoid "noisy neighbor" problems like increased latency and throttling
- Consider implementing streaming for applications like chatbots to enhance perceived performance
- Use dedicated model deployments per consumer group to provide usage isolation
4. Azure AI Foundry
Purpose: Provides comprehensive access to AI models and capabilities.
Key Features:
- Access to 1800+ models across various categories
- Automatically creates resources for speech, vision, etc.
Architecture:
- Built on capabilities provided by other Azure services
- Has a hub-project structure where the hub is the top-level resource
Understanding the Confusion Points
Different Foundry Access Points with Varying Model Availability
One of the most confusing aspects is that there are multiple ways to access AI Foundry with significantly different model availability:
- Azure AI Services Foundry: Only provides access to 62 chat completion models
- Azure OpenAI Resource Foundry: Only provides access to OpenAI models
- Azure AI Foundry Resource Foundry: Provides access to all 1800+ models
Hub vs. Project Structure
The Azure AI Foundry portal has two levels of access:
- Hub: Home to infrastructure (virtual network setup, managed keys, identities, policies)
- Project: Workspaces that allow you to build and deploy AI systems
This structure allows developers to create projects that inherit hub security settings, but contributes to confusion about where resources are created and accessed.
Common Issues and Solutions
Access Control Issues
Issue: "Principal doesn't have access to API/Operation" error when using the Azure AI Foundry portal chat playground
Solution: Assign the appropriate roles based on the services you're using:
- For Azure OpenAI: Cognitive Services OpenAI Contributor and Cognitive Services User role
- For Azure AI Search: Search Index Data Contributor or Search Index Data Reader roles
Storage Access Problems
Issue: Unable to access blob storage associated with an AI Foundry hub despite being the owner
Solution:
- Turn on system-assigned managed identity for the AI service
- Use RBAC to assign the storage blob data contributor role to the managed identity
- Wait for role assignments to propagate (can take minutes to an hour)
Private Endpoint Connection Issues
Issue: 403 errors or "access forbidden" messages when connecting to an Azure AI Foundry project with a private endpoint
Solution:
- Check proxy configurations that might prevent communication
- Ensure your proxy allows connections to AML APIs
- Verify network settings of the storage account associated with your hub
Deployment Troubleshooting
Issue: "Use of Azure OpenAI models in Azure Machine Learning requires Azure OpenAI Services resources"
Solution: This error means you might not have access to the particular Azure OpenAI model in your subscription or region
Issue: "Out of quota" errors
Solution: Review and manage quotas for:
- Deploying and inferencing models
- Azure OpenAI Service
- Resources with Azure AI Foundry
Conclusion
Understanding the structure of Azure AI Services helps navigate the platform more effectively. While the naming conventions can be confusing (particularly with "Azure AI Services" being both a parent category and a specific service), each component serves specific purposes in the AI development ecosystem.
For most comprehensive AI projects, start with Azure AI Foundry for the broadest access to models and capabilities, then integrate with other services like Azure AI Search as needed for specific functionality.