transactable apps
169 TopicsGoverning AI apps and agents for Marketplace
Governing AI apps and agents Governance is what turns powerful AI functionality into a solution that enterprises can confidently adopt, operate, and scale. It establishes clear responsibility for actions taken by the system, defines explicit boundaries for acceptable behavior, and creates mechanisms to review, explain, and correct outcomes over time. Without this structure, AI systems can become difficult to manage as they grow more connected and autonomous. For publishers, governance is how trust is earned — and sustained — in enterprise environments. It signals that AI behavior is intentional, accountable, and aligned with customer expectations, not left to inference or assumption. As AI apps and agents operate across users, data, and systems, risk shifts away from what a model can generate and toward how its behavior is governed in real‑world conditions. Marketplace readiness reflects this shift. It is defined less by raw capability and more by control, accountability, and trust. This post is part of a series on building and publishing well-architected AI apps and agents on Microsoft Marketplace. What governance means for AI apps and agents Governance in AI systems is operational and continuous. It is not limited to documentation, checklists, or periodic reviews — it shapes how an AI app or agent behaves while it is running in real customer environments. For AI apps and agents, governance spans three closely connected dimensions: Policy What the system is allowed to do, what data it is allowed to access, what is restricted, and what is explicitly prohibited. Enforcement How those policies are applied consistently in production, even as context, inputs, and conditions change. Evidence How decisions and actions are traced, reviewed, and audited over time. Governance works when intent, behavior, and proof move together — turning expectations into outcomes that can be trusted and examined. These dimensions are interdependent. Policy without enforcement is aspiration. Enforcement without evidence is unverifiable. Governance in action Governance becomes real when responsibility is explicit. For AI apps and agents, this starts with clarity around who is responsible for what: Who the agent acts for — and how its use protects business value Ensuring the agent is used for its intended purpose, produces measurable value, and is not misused, over‑extended, or operating outside approved business contexts. Who owns data access and data quality decisions Governing how the agent consumes and produces data, whether access is appropriate, and whether the data used or generated is reliable, accurate, and aligned with business and integrity expectations. Who is accountable for outcomes when behavior deviates Defining responsibility when the agent’s behavior creates risk, degrades value, or produces unexpected outcomes — so corrective action is timely, intentional, and owned. When governance is left vague or undefined, accountability gaps surface and agent actions become difficult to justify and explain across the publisher, the customer, and the solution itself. In this model, responsibility is shared but distinct. The publisher is responsible for designing and implementing the governance capabilities within the solution — defining boundaries, enforcement points, and evidence mechanisms that protect business value by default. Marketplace customers expect to understand who is accountable before they adopt an AI solution, not after an incident forces the question. The customer is responsible for configuring, operating, and applying those capabilities within their own environment, aligning them to internal policies, risk tolerance, and day‑to‑day use. Governance works when both roles are clear: the publisher provides the structure, and the customer brings it to life in practice. Data governance for AI: beyond storage and access For Marketplace‑ready AI apps and agents, data governance must account for where data moves, not just where it resides. Understanding how data flows across systems, tools, and tenants is essential to maintaining trust as solutions scale. Data governance for AI apps and agents extends beyond where data is stored. These systems introduce new artifacts that influence behavior and outcomes, including prompts and responses, retrieval context and embeddings, and agent‑initiated actions and tool outputs. Each of these elements can carry sensitive information and shape downstream decisions. Effective data governance for AI apps and agents requires clear structure: Explicit data ownership — defining who owns the data and under what conditions it can be accessed or used Access boundaries and context‑aware authorization — ensuring access decisions reflect identity, intent, and environment, not just static permissions Retention, auditability, and deletion strategies — so data use remains traceable and aligned with customer expectations over time Relying on prompts or inferred intent to determine access is a governance gap, not a shortcut. Without explicit controls, data exposure becomes difficult to predict or explain. Runtime policy enforcement in production Policies are stress tested when the agent is responding to real prompts, touching real data, and taking actions that carry real consequences. For software companies building AI apps and agents for Microsoft Marketplace, runtime enforcement is also how you keep the system fit for purpose: aligned to its intended use, supported by evidence, and constrained when conditions change. At runtime, governance becomes enforceable through three clear lanes of behavior: Decisions that require human approval Use approval gates for higher‑impact steps (for example: executing a write operation, sending an external request, or performing an irreversible workflow). This protects the business value of the agent by preventing “helpful” behavior from turning into misuse. Actions that can proceed automatically — within defined limits Automation is earned through clarity: define the agent’s intended uses and keep tool access, data access, and action scope anchored to those uses. Fit‑for‑purpose isn’t a feeling — it’s something you support with defined performance metrics, known error types, and release criteria that you measure and re‑measure as the system runs. Behaviors that are never permitted — regardless of context or intent Block classes of behavior that violate policy (including jailbreak attempts that try to override instructions, expand tool scope, or access disallowed data). When an intended use is not supported by evidence — or new evidence shows it no longer holds — treat that as a governance trigger: remove or revise the intended use in customer‑facing materials, notify customers as appropriate, and close the gap or discontinue the capability. To keep runtime enforcement meaningful over time, pair it with ongoing evaluation: document how you’ll measure performance and error patterns, run those evaluations pre‑release and continuously, and decide how often re‑evaluation is needed as models, prompts, tools, and data shift. This is what keeps autonomy intentional. It allows AI apps and agents to operate usefully and confidently, while ensuring behavior remains aligned with defined expectations — and backed by evidence — as systems evolve and scale. Auditability, explainability, and evidence Guardrails are the points in the system where governance becomes observable: where decisions are evaluated, actions are constrained, and outcomes are recorded. As described in Designing AI guardrails for apps and agents in Marketplace, guardrails shape how AI systems reason, access data, and take action — consistently and by default. Guardrails may be embedded within the agent itself or implemented as a separate supervisory layer — another agent or policy service — that evaluates actions before they proceed. Guardrail responses exist on a spectrum. Some enforce in the moment — blocking an action or requiring approval before it proceeds — while others generate evidence for post‑hoc review. Marketplace‑ready AI apps and agents could implement both, with the response mode matched to the severity, reversibility, and business impact of the action in question. These expectations align with the governance and evidence requirements outlined in the Microsoft Responsible AI Standard v2 General Requirements. In practice, guardrails support auditability and explainability by: Constraining behavior at design time Establishing clear defaults around what the system can and cannot do, so intended use is enforced before the system ever reaches production. Evaluating actions at runtime Making decisions visible as they happen — which tools were invoked, which data was accessed, and why an action was allowed to proceed or blocked. When governance is unclear, even strong guardrails lose their effectiveness. Controls may exist, but without clear intent they become difficult to justify, unevenly applied across environments, or disconnected from customer expectations. Over time, teams lose confidence not because the system failed, but because they can’t clearly explain why it behaved the way it did. When governance and guardrails are aligned, the result is different. Behavior is intentional. Decisions are traceable. Outcomes can be explained without guesswork. Auditability stops being a reporting exercise and becomes a natural byproduct of how the system operates day to day. Aligning governance with Marketplace expectations Governance for AI apps and agents must operate continuously, across all in‑scope environments — in both the publisher’s and the customer’s tenants. Marketplace solutions don’t live in a single boundary, and governance cannot stop at deployment or certification. Runtime enforcement is what keeps governance active as systems run and evolve. In practice, this means: Blocking or constraining actions that violate policy — such as stopping jailbreak attempts that try to override system instructions, escalate tool access, or bypass safety constraints through crafted prompts Adapting controls based on identity, environment, and risk — applying stricter limits when an agent acts across tenants, accesses sensitive data, or operates with elevated permissions Aligning agent behavior with enterprise expectations in real time — ensuring actions taken on behalf of users remain within approved roles, scopes, and approval paths These controls matter because AI behavior is dynamic. The same agent may behave differently depending on context, inputs, and downstream integrations. Governance must be able to respond to those shifts as they happen. Runtime enforcement is distinct from monitoring. Enforcement determines what is allowed to continue. Monitoring explains what happened once it’s already done. Marketplace‑ready AI solutions need both, but governance depends on enforcement to keep behavior aligned while it matters most. Operational health through auditability and traceability Operational health is the combination of traceability (what happened) and intelligibility (how to use it responsibly). When both are present, governance becomes a quality signal customers can feel day to day — not because you promised it, but because the system consistently behaves in ways they can understand and trust. Healthy AI apps and agents are not only traceable — they are intelligible in the moments that matter. For Marketplace customers, operational trust comes from being able to understand what the system is intended to do, interpret its behavior well enough to make decisions, and avoid over‑relying on outputs simply because they are produced confidently. A practical way to ground this is to be explicit about who needs to understand the system: Decision makers — the people using agent outputs to choose an action or approve a step Impacted users — the people or teams affected by decisions informed by the system’s outputs Once those stakeholders are clear, governance shows up as three operational promises you can actually support: Clarity of intended use Customers can see what the agent is designed to do (and what it is not designed to do), so outputs are used in the right contexts. Interpretability of behavior When an agent produces an output or recommendation, stakeholders can interpret it effectively — not perfectly, but reasonably well — with the context they need to make informed decisions. Protection against automation bias Your UX, guidance, and operational cues help customers stay aware of the natural tendency to over‑trust AI output, especially in high‑tempo workflows. This is where auditability and traceability become more than logs. Well governed AI systems should still answer: Who initiated an action — a user, an agent acting on their behalf, or an automated workflow What data was accessed — under which identity, scope, and context What decision was made, and why — especially when downstream systems or people are affected The logs should show evidence that stakeholders can interpret those outputs in realistic conditions — and there is a method to evaluate this, with clear criteria for release and ongoing evaluation as the solution evolves. Explainability still needs balance. Customers deserve transparency into intended use, behavior boundaries, and how to interpret outcomes — without requiring you to expose proprietary prompts, internal logic, or implementation details. For more information on securing your AI apps and agents, visit Securing AI apps and agents on Microsoft Marketplace | Microsoft Community Hub. What's next in the journey Governance creates the conditions for AI apps and agents to operate with confidence over time. With clear policies, enforcement, and evidence in place, publishers are better prepared to focus on operational maturity — how solutions are observed, maintained, and evolved safely in production. The next post explores what it takes to keep AI apps and agents healthy as they run, change, and scale in real customer environments. Key resources See curated, step-by-step guidance to help you build, publish, or sell your app or agent (no matter where you start) in App Advisor Quick-Start Development Toolkit can connect you with code templates for AI solution patterns Microsoft AI Envisioning Day Events How to build and publish AI apps and agents for Microsoft Marketplace Get over $126K USD in benefits and technical consultations to help you replicate and publish your app with ISV Success92Views4likes0CommentsMoving from Private Plans to Private Offers — Should We Make the Switch?
Hi Azure Marketplace community, We, at https://marketplace.microsoft.com/en-us/product/saturaminc.qualdo_drx are currently using private plans to handle custom pricing for specific customers, and we're evaluating whether it makes sense to transition to private offers. Would love to hear from others who've made this move — or who've deliberately stayed on private plans. Here's where we're at: private plans have served us well for restricting visibility and offering tiered pricing to select tenants, but as our deal complexity has grown (more enterprise customers, negotiated terms, channel partners), we're starting to feel some of the limitations. A few things pushing us toward private offers: Custom pricing flexibility — Private offers let us set percentage discounts or absolute prices per customer without creating a new plan for every deal. As our customer base grows, managing individual plans is getting unwieldy. Multi-party / channel support — We work with some resellers and CSPs. Private offers seem to support that flow much better with multi-party private offers (MPPO). Are there scenarios where private plans are still the better choice over private offers? How are you handling the coexistence of both during a transition period? Any impact on reporting, billing, or reconciliation we should be aware of? We want to make sure we're not solving one problem and creating another. Appreciate any real-world experiences!. Thanks in Advance, Kavitha SrinivasanSolved81Views2likes4CommentsQuokka, Relativity, and Belinda CZ offer transactable partner solutions in Microsoft Marketplace
Microsoft partners like Quokka, Relativity ODA LLC, and Belinda CZ s.r.o. deliver transact-capable offers, which allow you to purchase directly from Microsoft Marketplace. Learn about these offers in this post.72Views2likes0CommentsSeamless private offers are critical to closing negotiated deals faster in Microsoft Marketplace
The Seamless Marketplace private offers: creation to customer use | Microsoft Community Hub article walks through the end‑to‑end private offer journey, from creation in Partner Center to customer purchase and solution activation. Learn how Marketplace private offers support flexible selling motions, reduce friction for buyers, and help partners deliver a more streamlined purchasing experience. If you’re looking to strengthen your Marketplace transactions and improve deal execution, this is a must‑read. To learn more and have questions answered join the live webinar on April 15, Seamless private offers: From creation to purchase and activation - Microsoft Marketplace Community, where we will share an end-to-end walkthrough of private offer execution with a live demo and Q&A.How to design DORA-compliant customer communication for digital resilience in the EU
In this guest blog post from Microsoft Marketplace partner Messageflow, Julia Matuszewska, AI & Technical Content Writer, Vercom, discusses how Vercom MessageFlow can aid financial institutions to reduce risk, better prepare for audits, and increase compliance by aligning with international regulations and best practices.129Views1like0CommentsSeamless Marketplace private offers: creation to customer use
Private offers are a core mechanism for bringing negotiated commercial terms into Microsoft Marketplace. They allow publishers and channel partners to offer negotiated pricing, flexible billing structures, and custom terms; while enabling customers to purchase through the same Microsoft governed procurement, billing, and subscription experience they already use for Azure purchases. As Marketplace adoption grows, private offers increasingly involve channel partners, including resellers, system integrators, and Cloud Solution Providers. While commercial relationships vary, the Marketplace lifecycle remains consistent. Understanding that lifecycle—and where responsibilities differ by selling model—is essential to executing private offers efficiently and at scale. Join us April 15 for Marketplace Partner Office Hours, where Microsoft Marketplace experts Stephanie Brice and Christine Brown walk through how to execute private offers end to end—from creation to customer purchase and activation—across direct and partner‑led selling models. The session will include a live demonstration and Q&A, with practical guidance on flexible billing, channel scenarios, and common pitfalls. This article walks through the private offer lifecycle to help partners establish a clear, repeatable operating model to successfully transact in Microsoft Marketplace. Why private offers are structured the way they are Private offers are designed to align with how enterprise customers already procure software through Microsoft. Customers purchase through governed billing accounts, defined Azure role-based access control (RBAC) enforced roles, and Azure subscriptions that support cost management and compliance. Rather than bypassing these controls, private offers integrate negotiated deals directly into Microsoft Marketplace. This allows customers to: Apply purchases to existing Microsoft agreements (Microsoft Customer Agreement (MCA) or Enterprise Agreement (EA)) Preserve internal approval workflows Manage Marketplace subscriptions alongside other Azure resources Private offers also support flexible billing schedules. This is especially important for enterprise customers managing budget cycles, approvals, and cash flow. Flexible billing allows partners to align charges to agreed timelines—such as billing on a specific day of the month or spreading payments across defined milestones—while still transacting through Microsoft Marketplace. Customers can align Marketplace charges with internal finance processes without requiring separate contracts or off‑platform invoicing. For publishers and partners, this design creates a predictable lifecycle that scales across direct and channel‑led motions. Each stage exists for a specific reason and understanding that intent helps reduce delays and rework. Learn more: Private offers overview One lifecycle, multiple selling models All private offers—regardless of selling model—follow the same three stages: Creation of a private offer based on a publicly transactable Marketplace offer Acceptance, purchase, and configuration of the private offer Activation or deployment, based on how the solution is delivered What varies by model is who creates the offer, who sets margin, and who owns the customer relationship—not how Microsoft Marketplace processes the transaction. 1. Creation: Starting with a transactable public offer Every private offer begins with a publicly transactable Marketplace offer enabled for Sell through Microsoft. Private offers inherit the structure, pricing model, and delivery architecture of that public offer and its associated plan. If a public offer is listed as Contact me or otherwise non‑transactable, it must be updated before any private offers—direct to customer or channel‑led—can be created. Creation flows by selling model: Customer private offers (CPO) The publisher creates a private offer in Partner Center for a specific customer, based on the Azure subscription (Customer Azure Billing ID) provided by the customer. The publisher defines negotiated pricing, duration, billing terms (including any flexible billing schedule), and custom conditions. Multiparty private offers (MPO) The publisher creates a private offer in Partner Center and extends it to a specific channel partner. The partner adds margin and completes the offer before sending it to the customer. Resale enabled offers (REO) The publisher authorizes a channel partner in Partner Center to resell a publicly transactable Marketplace offer. Once authorized, the channel partner can independently create private offers for customers without publisher involvement in each deal. Cloud Solution Provider (CSP) private offers A CSP hosts the customer’s Azure environment (typically for SMB customers) and acts on behalf of the customer. The publisher creates a private offer in Partner Center for a CSP partner, extending margin so the CSP can sell the solution to customers through the CSP motion. In all cases, the private offer remains anchored to the same underlying public Marketplace offer. 2. Acceptance and purchase: What happens in Marketplace Microsoft Marketplace provides a consistent purchasing experience while supporting different partner‑led models behind the scenes. Customer private offer, multiparty private offer, resale enabled private offer For these models, the customer experience is the same and includes three steps: Accepting the private offer The customer accepts the negotiated terms (price, duration, custom terms) in Azure portal. This is the legal acceptance step under the customer’s MCA or EA. Purchasing or subscribing The customer associates the offer to the appropriate billing account and Azure subscription. This enables billing and fulfillment. Configuring the solution After subscription, the customer is redirected to the partner’s landing page. This step connects the Marketplace purchase to the partner’s system, enabling provisioning, subscription activation, and setup. Learn more: Accept the private offer Purchase and subscribe to the private offer In large enterprises, acceptance and purchase are often completed by different roles, supporting governance and auditability. CSP private offers In the CSP model, the CSP partner—not the end customer—accepts and purchases the private offer on the customer’s behalf. Microsoft invoices the CSP partner, and the CSP bills the end customer under their existing CSP relationship. Key distinctions: The end customer does not interact with the Marketplace private offer CSP private offers do not decrement customer Microsoft Azure Consumption Commitment (MACC) because there is no MACC in the CSP agreement Customer pricing and billing occur outside Marketplace Learn more: ISV to CSP private offers 3. Activation or deployment: Defined by delivery model, not selling motion Activation or deployment is determined by how the solution is built, not whether the deal is direct to customer or channel‑led. SaaS offers The solution runs in the publisher’s environment. After subscription, activation occurs through the SaaS fulfillment process, typically involving customer onboarding or account configuration. No Azure resources are deployed into the customer’s tenant. Deployable offer types (virtual machines, containers, Azure managed applications) The solution runs in the customer’s Azure tenant. Deployment provisions resources into the selected Azure subscription according to the offer’s architecture. Channel partners may support onboarding or deployment, but Marketplace activation or deployment reflects the technical delivery model—not the commercial route. Setting expectations that scale Successful partners set expectations early by separating commercial steps from technical activation: The customer transacts under an Enterprise Agreement (EA) or Microsoft Customer Agreement (MCA) The private offer includes custom pricing and any flexible billing schedule based on the publicly transactable offer The customer accepts negotiated terms in Microsoft Marketplace The purchase and subscribe steps associate the offer to the billing account and Azure subscription, the configure step triggers the notification to activate or deploy the solution for customer use Billing starts based on SaaS fulfillment or Azure resource deployment Choosing the right model While the lifecycle is consistent, each model supports different strategies: Customer private offers allow the publisher to negotiate terms directly with the customer Multiparty private offers enable close channel collaboration while sharing margin Resale enabled offers support scale by empowering channel partners to transact independently CSP private offers align with customer segments led with this motion The right choice depends on partner strategy, not on how Marketplace processes the transaction. Learn more: Transacting on Microsoft Marketplace Bringing it all together Private offers turn negotiated agreements into scalable, governed transactions inside Microsoft Marketplace. Regardless of whether a deal is direct or channel‑led, the underlying lifecycle remains the same, rooted in a transactable public offer, executed through Microsoft‑managed purchasing, and activated based on how the solution is delivered. By understanding that lifecycle and intentionally choosing the right direct or channel model and billing structure, partners can reduce friction, set clearer expectations, and scale Marketplace transactions with confidence. When aligned correctly, private offers become more than a deal construct; they become a repeatable operating model for Marketplace growth.146Views1like0CommentsProduction ready architectures for AI apps and agents on Marketplace
Why “production‑ready” architecture matters for Marketplace AI apps and agents A working AI prototype is not the same as a production‑ready AI app in Microsoft Marketplace. Marketplace solutions are expected to operate reliably in real customer environments, alongside mission‑critical workloads and under enterprise constraints. As a result, AI apps published through Marketplace must meet a higher bar than “it works in a demo.” Production‑ready Marketplace AI apps must assume: Alignment with enterprise expectations and the Azure Well‑Architected Framework, including cost optimization, security, reliability, operational excellence, and performance efficiency Architectural decisions made early are difficult to reverse, especially once customers, tenants, and billing relationships are in place A higher trust bar from customers, who expect Marketplace solutions to be Microsoft‑vetted, certified, and safe to run in production Customers come to Marketplace expecting solutions that are ready to run, ready to scale, and ready to be supported—not experiments. This post focuses on the architectural principles and patterns required to meet those expectations. Specific services and implementation details are covered later in the series. This post is part of a series on building and publishing well-architected AI apps and Agents on Microsoft Marketplace. Aligning offer type and architecture early sets you up for success A strong indicator of a smooth Marketplace journey is early alignment between offer type and solution architecture. Offer type defines more than how an AI app is listed—it establishes clear roles and responsibilities between publishers and customers, which in turn shape architectural boundaries. Across all other offer types, architecture must clearly answer three questions: Who owns the runtime? Where does the AI execute? Who controls updates and ongoing operations? These decisions will vary depending on whether the solution resides in the customer’s or publisher’s tenant based on the attributes associated with the following transactable marketplace offer types: SaaS offers, where the AI runtime lives in the publisher’s environment and architecture must support multi‑tenancy, strong isolation, and centralized operations Container offers, where workloads run in the customer’s Kubernetes environment and architecture emphasizes portability and clear operational assumptions Virtual Machine offers, where preconfigured environments run in the customer’s subscription and architecture is more tightly coupled to the OS and infrastructure footprint Azure Managed Applications, where the solution is deployed into the customer's subscription and architecture must balance customer control with defined lifecycle boundaries. What makes this model distinctive is its flexibility: an Azure Managed Application can package containers, virtual machines, or a combination of both — making it a natural fit for solutions that require customer-controlled infrastructure without sacrificing publisher-managed operations. The packaging choice shapes the underlying architecture, but the managed application wrapper is what defines how the solution is deployed, updated, and governed within the customer's environment. Architecture decisions naturally reinforce Marketplace requirements and reduce certification and operational friction later. Key factors that benefit from early alignment include: Roles and responsibilities, such as who operates the AI runtime and who is responsible for uptime, patching, scaling, and ongoing operations Proximity to data, particularly for AI solutions that rely on customer‑specific or proprietary data, where placement affects performance, data movement, and compliance Core architectural building blocks of AI apps Designing a production‑ready AI app starts with treating the solution as a system, not a single service. AI apps—especially agent‑based solutions—are composed of multiple cooperating layers that together enable reasoning, action, and safe operation at scale. At a high level, most production‑ready AI apps include the following building blocks: Interaction layer, which serves as the entry point for users or systems and is responsible for authentication, request shaping, and consistent responses Orchestration layer, which coordinates reasoning, tool selection, workflow execution, and retrieval‑augmented generation (RAG) flows across multi‑step interactions Model endpoints, which provide inference and generation capabilities and introduce distinct latency, cost, and dependency characteristics Data sources, including vector stores, operational data, documents, and logs that the AI system reasons over Control planes, such as identity, configuration, policy enforcement, feature flags, and secrets management, which govern behavior without redeploying core logic Observability, which enables tracing, monitoring, and diagnosis of agent decisions, actions, and outcomes Networking, which connects components using a zero‑trust posture where every call is authenticated and outbound access is explicitly controlled Together, these components form the foundation of most Marketplace‑ready AI architectures. How they are composed—and where boundaries are drawn—varies by offer type, tenancy model, and customer requirements. Specific services, patterns, and implementation guidance for each layer are explored later in the series. Tenancy design choices as an early architectural decision One of the earliest and most consequential architectural decisions is where the AI solution is hosted. Does it run in the publisher’s tenant, or is it deployed into the customer’s tenant? This choice establishes foundational boundaries and is difficult to change later without significant redesign. If the solution runs in the publisher’s tenant, it is inherently multi‑tenant and must be designed with strong logical isolation across customers. If it runs in the customer’s tenant, deployments are typically single‑tenant by default, with isolation provided through infrastructure boundaries. Many Marketplace AI apps fall between these extremes, making it essential to define the tenancy model early. Common tenancy approaches include: Publisher‑hosted, multi‑tenant solutions, where a shared AI runtime serves multiple customers and requires strict isolation of customer data, inference requests, identity, and cost attribution Customer‑hosted, single‑tenant deployments, where each customer operates an isolated instance within their own Azure subscription, often preferred for regulated or tightly controlled environments Hybrid models, which combine centralized AI services with customer‑hosted data or execution layers and require carefully defined trust and access boundaries Tenancy decisions influence several core architectural dimensions, including: Identity and access boundaries, which define how users and agents authenticate and act across tenants Data isolation, including how customer data is stored, processed, and protected Model usage patterns, such as shared models versus tenant‑specific models Cost allocation and scale, including how usage is tracked and attributed per customer These considerations are not implementation details—they shape how the AI system behaves, scales, and is governed in production. Reference architecture guidance for multi‑tenant AI and machine learning solutions in the Azure Architecture Center explores these tradeoffs in more detail. Understanding your customer’s needs Designing a production‑ready AI architecture starts with understanding the environment your customers expect your solution to operate in. Marketplace customers vary widely in their security posture, compliance obligations, operational practices, and tolerance for change. Architectures that reflect those realities reduce friction during onboarding, certification, and long‑term operation. Key customer considerations that shape architecture include: Security and compliance expectations, such as industry regulations, internal governance policies, or regional data requirements Target environments, including whether customers expect solutions to run in their own Azure subscription or are comfortable consuming centrally hosted services Change and outage windows, where operational constraints or seasonal restrictions require predictable and controlled updates Architectural alignment with customer needs is not about designing for every edge case. It is about making intentional tradeoffs that reflect how customers will deploy, operate, and depend on your AI solution in production. Specific security controls, compliance enforcement mechanisms, and operational policies are explored later in the series. This section establishes the architectural mindset required to support them. Separating environments for safe iteration Production AI systems must evolve continuously while remaining stable for customers. Separating environments is how publishers enable safe iteration without destabilizing live usage—and how customers maintain confidence when adopting and operating AI solutions in their own environments. From the publisher’s perspective, environment separation enables: Iteration on prompts, models, and orchestration logic without impacting production customers Validation of behavior changes before rollout, especially for AI‑driven systems where small changes can produce materially different outcomes Controlled release strategies that reduce operational risk From the customer’s perspective, environment separation shapes how the solution fits into their own development and operational practices: Where the solution is deployed across development, staging, and production environments How deployments are repeated or promoted, particularly when the solution runs in the customer’s tenant Whether environments can be recreated predictably, or whether customers are forced to manually reconfigure deployments with each iteration When AI solutions are deployed into the customer’s tenant, environment design becomes especially important. Customers should not be required to reverse‑engineer deployment logic, recreate environments from scratch, or re‑establish trust boundaries every time the solution evolves. These concerns should be addressed architecturally, not deferred to operational workarounds. Environment separation is therefore not just a DevOps choice—it is an architectural decision. It influences identity boundaries, deployment topology, validation strategies, and the shared operational contract between publisher and customer. Designing for AI‑specific scalability patterns AI workloads do not scale like traditional web or CRUD‑based applications. While front‑end and API layers may follow familiar scaling patterns, AI systems introduce behaviors that require different architectural assumptions. Production‑ready AI architectures must account for: Bursty inference demand, where usage can spike unpredictably based on user behavior or downstream automation Long‑running or multi‑step agent workflows, which may span tools, data sources, and time Model‑driven latency and cost characteristics, which influence throughput and responsiveness independently of application logic As a result, scalability decisions often vary by layer. Horizontal scaling is typically most effective in interaction, orchestration, and retrieval components, while model endpoints may require separate capacity planning, isolation, or throttling strategies. Treating identity as an architectural boundary Identity is foundational to Marketplace AI apps, but architecture must plan for it explicitly. Identity decisions define trust boundaries across users, agents, and services, and shape how the solution scales, secures access, and meets compliance requirements. Key architectural considerations include: Microsoft Entra ID as a foundation, where identity is treated as a core control plane rather than a late‑stage integration How users sign in, including: Their own corporate Microsoft Entra ID tenant B2B scenarios where one Entra ID tenant trusts another B2C identity providers for customer‑facing experiences How tenants authenticate, particularly in multi‑tenant or cross‑organization scenarios How AI agents act on behalf of users, including delegated access, authorization scope, and auditability How services communicate securely, using a zero‑trust posture where every call is authenticated and authorized Treating identity as an architectural boundary helps ensure that trust relationships remain explicit, enforceable, and consistent across tenants and environments. This foundation is critical for supporting secure operation, compliance enforcement, and future tenant‑linking scenarios. Designing for observability and auditability Production‑ready AI apps must be observable and auditable by design. Marketplace customers expect visibility into how systems behave in production, and publishers need clear insight to diagnose issues, operate reliably, and meet enterprise trust and compliance expectations. Key architectural considerations include: End‑to‑end observability, covering user interactions, agent reasoning steps, tool invocations, and downstream service calls Clear audit trails, capturing who initiated an action, what the AI system did, and how decisions were executed—especially when agents act on behalf of users Tenant‑aware visibility, ensuring logs, metrics, and traces are correctly attributed without exposing data across tenants Operational transparency, enabling effective troubleshooting, incident response, and continuous improvement without ad‑hoc instrumentation For AI systems, observability goes beyond infrastructure health. It must also account for AI‑specific behavior, such as prompt execution, model selection, retrieval outcomes, and tool usage. Without this visibility, diagnosing failures, validating changes, or explaining outcomes becomes difficult in real customer environments. Auditability is equally critical. Identity, access, and action histories must be traceable to support security reviews, regulatory obligations, and customer trust—particularly in regulated or enterprise settings. Common architectural pitfalls in Marketplace AI apps Even experienced teams run into similar challenges when moving from an AI prototype to a production‑ready Marketplace solution. The following pitfalls often surface when architectural decisions are deferred or made implicitly. Common pitfalls include: Treating AI as a single service instead of a system, where model inference is implemented without considering orchestration, data access, identity, observability, and operational boundaries Hard‑coding tenant assumptions, such as assuming a single tenant, identity model, or deployment topology, which becomes difficult to unwind as customer requirements diversify Not planning for a resilient model strategy, leaving the architecture fragile when model versions change, capabilities evolve, or providers introduce breaking behavior Assuming data lives within the same boundary as the solution, when in practice it may reside in a different tenant, subscription, or control plane Tightly coupling prompt logic to application code, making it harder to iterate on AI behavior, validate changes, or manage risk without full redeployments Assuming issues can be fixed after go‑live, which underestimates the cost and complexity of changing architecture once customers, subscriptions, and trust relationships are in place While these pitfalls may be caused by a lack of technical skill on the customer’s side, they could typically emerge when architectural decisions are postponed in favor of speed, or when AI behavior is treated as an isolated concern rather than part of a production system. What’s next in the journey The architectural decisions made early—around offer type, tenancy, identity, environments, and observability—establish the foundation on which everything else is built. When these choices are intentional, they reduce friction as the solution evolves, scales, and adapts to real customer needs. The next set of posts builds on this foundation, exploring different dimensions of operating, securing, and evolving Marketplace AI apps in production. Key resources See curated, step-by-step guidance to help you build, publish, or sell your app or agent (no matter where you start) in App Advisor Quick-Start Development Toolkit can connect you with code templates for AI solution patterns Microsoft AI Envisioning Day Events How to build and publish AI apps and agents for Microsoft Marketplace Get over $126K USD in benefits and technical consultations to help you replicate and publish your app with ISV Success236Views7likes1Comment