biztalk migration
19 TopicsHybrid Logic Apps deployment on Rancher K3s Kubernetes cluster
K3s is a lightweight Kubernetes distribution, certified by the Cloud Native Computing Foundation (CNCF) and originally developed by Rancher. It is optimized for on-premises environments with limited resources, making it ideal for edge computing and lightweight hybrid scenarios. Unlike a full Kubernetes distribution, K3s reduces overhead while maintaining full Kubernetes API compatibility. This makes K3s an ideal choice for hosting Logic Apps Standard near your data sources—such as on-premises SQL Server or local file shares—when you have lightweight workloads. There are 5 steps which are followed to setup the Hybrid Logic Apps including infrastructure which is illustrated in the following diagram. Most of these 5 steps are same as discussed in the Hybrid Logic Apps doc except the K3s Setup part Set up your own infrastructure for Standard logic app workflows - Azure Logic Apps | Microsoft Learn. Step 1: Prepare the K3s Cluster Docker desktop setup - In this case, the host machine is Windows 11 so decided to user Docker with WSL2 to setup the containers. Install the docker desktop using WSL2 Docker Desktop: The #1 Containerization Tool for Developers | Docker and make sure we select WSL2 Install K3s on your infrastructure and create single node cluster using k3d. #Install choco , kubectl and Helm Set-ExecutionPolicy Bypass -Scope Process -Force; [System.Net.ServicePointManager]::SecurityProtocol = [System.Net.ServicePointManager]::SecurityProtocol -bor 3072; iex ((New-Object System.Net.WebClient).DownloadString('https://community.chocolatey.org/install.ps1')) powershell choco install kubernetes-cli -y choco install kubernetes-helm -y choco install k3d -y # open in new powershell window powershell k3d cluster create # deleting the default load balancer Traefik as it conflicts with 80 and 443 port - we can configure the load balancer to other ports if needed kubectl delete svc traefik -n kube-system kubectl delete deployment traefik -n kube-system Next two steps are same as given Set up your own infrastructure for Standard logic app workflows - Azure Logic Apps | Microsoft Learn Step 2: Connect the Kubernetes cluster to Azure Arc Step 3: Setup the Azure Container Apps extension and environment You need to skip the core DNS setup required for Azure Local as given in Update CoreDNS Step 4: Conduct the Storage Configuration for SQL and SMB SQL Database (Runtime Store): Hybrid Logic Apps use SQL database for runtime operations and run history. In this scenario I used on-premise SQL server using SQL Authentication. I setup the SQL Server 2022 on the Windows host machine, enabled SQL server authentication and added new SQL admin user. Please follow the link for more details.. The SQL connection string can be validated using following PowerShell script $connectionString = "Server=<server IP address>;Initial Catalog=<databaseName>;Persist Security Info=False;User ID=<sqluser>;Password=<password>;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=True;Connection Timeout=30;" try { $connection = New-Object System.Data.SqlClient.SqlConnection $connection.ConnectionString = $connectionString $connection.Open() Write-Host "✅ Connection successful" $connection.Close() } catch { Write-Host "❌ Connection failed: $($_.Exception.Message)" } SMB is used as local file share on Windows host machine; it is advised to use a new user for the Windows SMB share $Username = "k3suser" $Password = ConvertTo-SecureString "<password complex>" -AsPlainText -Force $FullName = "K3s user" $Description = "Created via PowerShell" # Create the user New-LocalUser -Name $Username -Password $Password -FullName $FullName -Description $Description Add-LocalGroupMember -Group "Users" -Member $Username Once the above user is created you can use Windows hosted machine to create Artifacts folder and allow read and write access. Please follow the link for more details Step 5: Create your Logic App (Hybrid) With all prerequisites and infrastructure in place for creating Hybrid Logic Apps, the next step is to build the Logic Apps using the specified connection string and SMB share path. This can be accomplished through the Azure Portal, as outlined below. Now you can create Logic Apps workflows using the designer and execute the Logic Apps workflow.🚀 General Availability: Enhanced Data Mapper Experience in Logic Apps (Standard)
We’re excited to announce the General Availability (GA) of the redesigned Data Mapper UX in the Azure Logic Apps (Standard) extension for Visual Studio Code. This release marks a major milestone in our journey to modernize and streamline data transformation workflows for integration developer. What's new The new UX, previously available in public preview, is now the default experience in the Logic Apps Standard extension. This GA release reflects direct feedback from our integration developer community. We’ve resolved blockers that we heard from customers and usability issues that impacted performance and stability, including: Opening V1 maps in V2: Seamlessly open and edit existing maps you have already created with latest visual capabilities. Load schemas on Mac: Addressed schema-related crashes on macOS for a smoother experience. Function documentation updates: Improved guidance and examples for built-in collection functions that apply on repeating nodes. Stay connected We would love to hear your feedback. Please use this form link to let us know if there are any missing gaps or scenarios that are not yet coveredLogic Apps Aviators Newsletter - September 25
In this issue: Ace Aviator of the Month News from our product group Community Playbook News from our community Ace Aviator of the Month September’s Ace Aviator: Kritika Singh Integration Architect & Sr. Consultant at Capgemini Norge AS What's your role and title? What are your responsibilities? I work as an Integration Architect & Sr. Consultant at Capgemini Norge AS. In my role I assist clients in addressing a wide range of integration challenges, with a particular emphasis on modernizing legacy systems, such as BizTalk, by transitioning them to cloud-native Azure iPaaS solutions. I’m responsible for architecting secure and scalable integration landscapes, designing and developing solutions, mentoring team members, and engaging with stakeholders and cross-functional teams. I work extensively with technologies such as Azure Logic Apps, Azure Functions, API Management, Service Bus, App Service Environment(ASEv3), Virtual Networks and CI/CD pipelines using GitHub. One of my proudest achievements was successfully delivering a complex BizTalk modernization project to Azure that required deep technical expertise and strategic coordination. Can you give us some insights into your day-to-day activities and what a typical day in your role looks like? As part of a distributed team across different locations and countries, my day starts with stand-ups involving both offshore and onshore team members to review progress and assign tasks. I spend time with developers, helping them navigate technical challenges and mentoring them through dedicated sessions—this has helped improve delivery quality and team confidence. I collaborate closely with cross-functional teams and stakeholders to align on requirements and solution design. Throughout the day, I work on resolving issues, improving existing solutions, work on innovation ideas and managing tasks. What motivates and inspires you to be an active member of the Aviators/Microsoft community? I am deeply passionate about technology—having worked with BizTalk throughout my career and, for the past 5+ years, diving into Azure iPaaS. Microsoft products evolve constantly, and that sparks my curiosity to explore, learn, and innovate every day. What truly drives me is the opportunity to give back to the community by sharing my learnings, challenges, and even failures. It’s all about growing together and inspiring others along the way. Looking back, what advice do you wish you had been given earlier that you'd now share with those looking to get into STEM/technology? Curiosity and consistency matter more than perfection. Don’t be afraid to ask questions, experiment, and fail, that’s where the real learning happens. Also, find a community that supports you; sharing your journey, both wins and setbacks, can inspire others and help you grow faster. What has helped you grow professionally? My professional growth has been shaped by a blend of curiosity, courage, and the right opportunities. Starting with BizTalk and evolving into Azure iPaaS, I’ve embraced every challenge as a chance to learn. What’s made the biggest difference is having the boldness to take on responsibility, the willingness to take risks, and the drive to keep growing. Sharing my journey with the community has not only helped others but also deepened my own learning. If you had a magic wand that could create a feature in Logic Apps, what would it be and why? If I had a magic wand to enhance Logic Apps, I’d bring in three powerful features to supercharge developer productivity and solution resilience: Seamless Version Control & Rollback Having Git-like capabilities built into Logic Apps—track every change, compare versions, and roll back instantly when needed. This would empower teams to experiment confidently and collaborate more effectively without fear of breaking production workflows. Effortless Disaster Recovery Setup Setting up DR should be as simple as a few clicks. A built-in, automated DR configuration for AIS would ensure business continuity, reduce downtime, and give developers peace of mind—especially in mission-critical environments. Native JSON Mapper(Not Liquid) A visual, intuitive JSON mapping tool would simplify complex data transformations, reduce manual coding, and speed up development. This would be a game-changer for integration scenarios, especially when working with dynamic schemas and APIs. Simplified Authorization like ClaimChecks for Logic Apps Standard (Beyond EasyAuth) A more developer-friendly authorization setup that minimizes manual configurations and integrates seamlessly with identity providers. This would make securing Logic Apps faster, easier, and more consistent across environments. News from our product group Logic Apps Live August 2025 Missed Logic Apps Live in August? You can watch it here. We had a recap on Logic Apps Hybrid, our special guest Kritika Singh talking about her learnings with BizTalk Migration to AIS, and updates on Data Mapper GA and Logic Apps Standard Deployment Center. Logic Apps Community Day 2025 We are bringing Logic Apps Community Day again this year, on October 30, 2025 (Pacific Time) and we want you to join us as we host a full day of learning where you will be the star! Call for Speakers is still open until September 07, 2025 – so hurry and submit your session! General Availability: Enhanced Data Mapper Experience in Logic Apps (Standard) We’re excited to announce the General Availability (GA) of the redesigned Data Mapper UX in the Azure Logic Apps (Standard) extension for Visual Studio Code. This release marks a major milestone in our journey to modernize and streamline data transformation workflows for integration developer. Announcing: Setup CD in Azure Logic Apps Standard with Deployment Center Looking to automate your Azure Logic Apps code deploymentsin a faster way? Deployment Center - a built-in feature in Azure Logic Apps Standard - in now available, with built-in support on your VS Code projects, making it easier to deploy Logic Apps from your source control repository. Deployment Center is designed to make deploying, updating, and managing your Logic Apps workflows simple and straightforward. Hybrid Logic Apps deployment on Rancher K3s Kubernetes cluster Explore how Hybrid Logic Apps run effortlessly on K3s—delivering the power of Hybrid Logic Apps without the complexity and heavy infrastructure demands of a full Kubernetes cluster! News from our community What gets returned to the LLM by my Logic App Agent Loop tool? Video by Michael Stephenson Michael has been experimenting with Logic App Agent loop and, following a discussion with Kent Weare about the interaction between the workflow and the LLM, aimed to understand what data is returned to the model, since the number of tokens influences cost. He summarizes his findings from that conversation in this brief video. SOAP 1.2 Calls from Logic Apps – Fixing Unsupported Media & WS-Addressing Errors Post by Prashant Singh Struggling with SOAP 1.2 in Azure Logic Apps? Learn how to fix Unsupported Media errors, decode MTOM responses, and handle WS-Addressing headers for seamless integration in this post by Prashant. Demystifying AI Agent Loops in Logic Apps: The Future of Integration (But Not Everywhere) Post by Al Ghoniem Explore how AI Agent Loops enhance Azure Logic Apps for non-deterministic tasks like anomaly detection and IT Ops triage—while knowing when traditional workflows are the better fit. Can I use AI to create and deploy an Azure Logic Apps with Business Central connector? Post by Stefano Demiliani Stefano is testing the boundaries of what AI can do, so you don’t have to – he ran a blind test showing that AI can deploy Azure resources well—but struggles with external connectors like Business Central. Learn what worked, what didn’t, and why better prompts matter. How to use ChatGPT Agent Mode with Azure! Video by Stephen W. Thomas And looks like August was the month to experiment with Azure Resources. Stephen did some research too and shows you how easily ChatGPT-5 Agent Mode can auto-provision resources in Azure. This video demonstrates how to use a single prompt to build a logic app and create a resource group. Follow his video along to see how to get ChatGPT-5 agent working for you! Integration Love Story - Andrew Wilson Video by Ahmed Bayoumy and Robin Wilde In this special fast-paced episode recorded at INTEGRATE, Ahmed and Robin sit down with the brilliant Andrew, newly awarded Microsoft MVP and Logic Apps Ace Aviator, to talk about his journey, passions, and why integration is the powerhouse behind every digital experience. Tips for Migrating SAP IDoc Reception Workloads from BizTalk to Azure Logic Apps Post by Francois Malgreve Learn how to reuse BizTalk XSLTs in Azure Logic Apps! In this post by Francois, you will learn hot to configure the SAP trigger with the right IDoc format and namespace settings—minimizing code changes and easing migration. Query Azure DevOps work items with Logic App and Managed Identity Post by Michael Stephenson Learn how to use a reusable Logic App and a user-assigned managed identity to securely query Azure DevOps work items using WIQL—ideal for building scalable, secure workflows. Understand Agent Loops in Azure Logic Apps Video by Srikanth Gunnala In this video, Srikanth explore Azure Logic Apps AI Agents — also known as Agent Workflows or Agent Loops — and how they’re redefining workflow automation with Azure OpenAI. You’ll learn what an AI agent is in Azure Logic Apps, how it works, and see a live demo of building an AI-powered, adaptive workflow. Automate Microsoft Fabric Cost Savings with Logic Apps Post by Sherry L. Robinson Learn how to pause and resume Microsoft Fabric capacity using Azure Logic Apps—cutting costs during off-hours with minimal code and seamless integration via REST API or Resource Manager, in this insightful post by Sherry. Use Graph API to send Emails in Logic Apps Post by Şahin Özdemir In this post, Şahin shows your hot to use Microsoft Graph API with service principals to securely send emails from Logic Apps using app registrations and access policies. This will be quite useful in cases where you can’t associate the calls to a user account, which is a requirement for the Office 365 connector.346Views0likes0CommentsLogic Apps Aviators Newsletter - August 25
In this issue: Ace Aviator of the Month News from our product group News from our community Ace Aviator of the Month August Ace Aviator: Jenny Anderson What's your role and title? What are your responsibilities? I’m an Integration Architect at Tietoevry Tech Services, where I work with large enterprise customers to develop integration solutions. For the past two years my main focus has been on cloud and hybrid integrations. I design integration architectures, advise on best practices including security and the chosen architecture, and collaborate closely with development teams to implement and maintain these solutions. Can you give us some insights into your day-to-day activities and what a typical day in your role looks like? My days usually start with scrum meetings across ongoing projects, which help me stay updated on progress, align with teams and prioritize my tasks for the day. After that, I often have customer meetings where I advise on integration strategies, provide architectural guidance or work on pre-sales engagements to scope out potential solutions. Recently, a big focus has been on BizTalk migrations, helping customers modernize their integration platforms by moving to Azure-based solutions. I try to dedicate my afternoons to hands-on technical work, which I really enjoy. Lately, that’s involved working with the new hybrid capabilities in Logic Apps. It’s a great mix of strategic consulting and deep technical implementation, which keeps the role dynamic and rewarding. What motivates and inspires you to be an active member of the Aviators/Microsoft community? I’ve always received a lot of support from the community especially when I was starting out in my career and I still benefit from it today. That generosity and openness made a big impact on me, so I feel it’s important to give back whenever I can. For me it’s a way to pay it forward and stay connected to a network that has helped me grow both technically and professionally. Looking back, what advice do you wish you had been given earlier that you'd now share with those looking to get into STEM/technology? Don’t overthink it, just start doing! In the beginning of my career, I assumed that everyone else knew everything, and that I couldn’t contribute or be part of certain areas because I didn’t know enough. But the truth is, no one knows everything, and that’s completely okay. The best way to learn is by doing and taking on challenges, making mistakes and growing from experience. I believe confidence comes from action, not from waiting until you feel “ready.” What has helped you grow professionally? One thing that has really helped me grow is surrounding myself with people who have different experiences or areas of expertise, whether at work, in communities, or through networking. I’ve learned a lot simply by asking questions, even the ones that might seem obvious. I also try to say yes to new opportunities, especially when they push me outside my comfort zone. Being an overthinker, I’ve developed a personal mantra: “Think 40%, do 60%.” It reminds me not to get stuck in planning or doubt, but to take action and learn along the way. That mindset has really helped me move forward. If you had a magic wand that could create a feature in Logic Apps, what would it be and why? If I could use a bit of magic in Logic Apps, I’d want AI to automagically handle all the data mappings. It’s honestly my least favorite part of integration work. It takes forever, it’s a bit dull and yet it’s always important. So, if AI could just step in and quietly take care of it, I wouldn’t complain. I’ve also heard a few customers ask for a disconnected control plane that can be hosted on-premises. That would be a big win for scenarios where cloud access is limited or compliance rules are extra strict. News from our product group Logic Apps Live July 2025 Missed Logic Apps Live in July? You can watch it here. We had a sneak peek into Logic Apps MCP Servers and Python support for Agent loop. Excinting topics and worth a watch! Troubleshoot Az Module within Logic App Standard Learn how to resolve Az Module installation failures in Logic Apps due to network restrictions or storage limits. Quick tests and fixes included to keep your workflows running smoothly. Introducing API Management Support in the Azure SRE Agent Azure’s SRE Agent now supports API Management, offering real-time diagnostics, backend health visualization, and intelligent remediation to keep your APIs reliable and scalable. Launch Your Private MCP Registry with Azure API Center. Discover how to create a secure, governed, and enterprise-ready MCP registry using Azure API Center—empowering AI innovation while maintaining control and visibility. Perform video analysis by using Azure Machine Learning and Computer Vision Replace manual video review with a scalable, AI-powered pipeline using Azure Machine Learning, Logic Apps, and Computer Vision. Boost accuracy and efficiency across industries like agriculture, traffic control, and manufacturing. Bringing Azure Logic Apps to on-prem, private, or public cloud with new Hybrid model | Azure Friday In this video Scott Hanselman and Harold Campos discuss the new Logic Apps Hybrid deployment model that allows customers to run their integration workloads in their own Kubernetes environments. This is ideal for customers initiating their journey to the cloud and hosting multiple on-premises workloads, who need to meet industry regulations, who wants to reuse their own Kubernetes infrastructure, or to avoid the natural latency introduced in hybrid configurations. News from our community Exposing Logic Apps as MCP Server in Azure API Management Video by Kent Weare On top of his PM work, Kent also finds time to keep his personal YouTube channel quite active. This time, he shows a walkthrough of creating an MCP Server using Logic Apps and API Management. The initial explanation of MCP and the various protocols alone make this video a great watch! Integration Love Story - Divya Swarnkar Video by Ahmed Bayoumy and Robin Wilde In this short episode of Integration Love Story, Ahmed and Robin chat with our own Divya Swarnkar, Product Manager at Microsoft who's been on an incredible journey from using Logic Apps as a customer to now helping build the product with the team behind the scenes. From BizTalk to Azure: A Guide for the Slightly Terrified Post by Sandro Pereira Explore the risks, timelines, and migration strategies as BizTalk nears end-of-life. Sandro shares the webinar recording – another tool to help you decide whether to stay or move to Azure Integration Services—without losing sleep. Azure Logic Apps Naming Conventions whitepaper Post by Sandro Pereira Boost clarity, scalability, and collaboration in Azure Logic Apps with this whitepaper. Learn best practices for naming triggers, actions, variables, and more - essential for automation, CI/CD, and long-term maintainability. You can create and use your own personal templates in Azure Logic Apps Post by Sandro Pereira It is not a newsletter, without at least a Friday Fact from Sandro! In this post, you can learn how to build, manage, and share reusable templates for consistent, efficient integration across projects. Speed up automation and standardize workflows you’re your own personal templates in Logic Apps.460Views0likes0CommentsAnnouncement: General Availability of Logic Apps Hybrid Deployment Model
We are thrilled to announce the General Availability of the Logic Apps Hybrid Deployment Model, a groundbreaking feature that offers unparalleled flexibility and control to our customers. This innovative deployment model allows you to run Logic Apps workloads on customer-managed infrastructure, providing you with the option to host your integration solutions on-premises, in a private cloud, or even in a third-party public cloud. With the Logic Apps Hybrid Deployment Model, you can tailor your integration solutions to meet your specific needs, whether it's for regulatory compliance, data privacy, or network restrictions. This model ensures that you have the freedom to choose the best environment for your workflows, while still leveraging the powerful capabilities of Azure Logic Apps. The Hybrid Deployment Model supports a semi-connected architecture, offering local processing of workflows, local storage, and local network access. This means that the data processed by the workflows remains in your local SQL Server, and you have the ability to connect to local networks. Additionally, the built-in connectors will execute in your local compute, giving you access to local data sources and higher throughput. Since we launched the public preview, we have received an overwhelmingly positive response from customers across various industries. Many customers, including those looking to migrate from BizTalk Server, have expressed interest in this offering due to its ability to co-locate integration platforms near key lines of business systems, avoiding dependencies on public internet to process transactions. Journey of the Hybrid Deployment Model Feature At the Integrate 2024 event, we announced the early access preview of the Hybrid Deployment model for Logic Apps Standard. This initial phase allowed interested parties to nominate themselves for early access and provided valuable feedback on the model's functionality and benefits. Following the private preview, we launched the public preview, which empowered our customers with additional flexibility and control. This phase allowed customers to build and deploy workflows on customer-managed infrastructure, offering the option to run Logic Apps on-premises, in a private cloud, or in a third-party public cloud. The public preview also introduced the semi-connected architecture, enabling local processing of workflows and access to local data sources. In October 2024, we refreshed the public preview and received an overwhelmingly positive response from customers across various industries. This feedback highlighted the model's ability to meet specific use cases, such as migrating from BizTalk Server and co-locating integration platforms near key lines of business systems. The public preview refresh also emphasized the model's alignment with our promise of providing customers with more options to meet their business needs. We are excited to see how our customers will leverage the Logic Apps Hybrid Deployment Model to meet their business needs and drive innovation. Thank you for your continued support and feedback. New features in the GA release: Open Telemetry support: Open telemetry is a vendor-neutral open-source Observability framework for instrumenting, generating, collecting, and exporting telemetry data. The support for Open Telemetry in Hybrid deployment model ensures the seamless logging in the semi-connected scenarios and provides the ability to choose any observability platform as a telemetry endpoint. More details here. To set up Open Telemetry capability from Azure portal, follow these steps: Open the host.json in the root directory of SMB file share path configured in your logic app. In the host.json file, at the root level, add the following telemetryMode setting with the OpenTelemetry value, for example: { "version": "2.0", "extensionBundle": { "id": "Microsoft.Azure.Functions.ExtensionBundle.Workflows", "version": "[1.*, 2.0.0)" }, "telemetryMode": "OpenTelemetry" } When you enable Open Telemetry in the host.json file, your logic app exports telemetry based on the Open Telemetry-supported app settings that you define in the environment. Add below app settings from portal by navigating to Containers-->Environment variables-->edit and deploy. App setting Description OTEL_EXPORTER_OTLP_ENDPOINT The online transaction processing (OTLP) exporter endpoint URL for where to send the telemetry data. OTEL_EXPORTER_OTLP_HEADERS (optional) A list of headers to apply to all outgoing data. Commonly used to pass authentication keys or tokens to your observability backend. If your Open Telemetry endpoint requires other Open Telemetry related settings, include these settings in the app settings too. Support for Zip deployment through VSCode: The support for Zip deployment in VSCode deployment has enhanced the deployment experience with more reliability. This feature uses Azure Entra authentication for deployment, hence the VSCode machine doesn’t require to have permissions on the SMB share and the user need not to provide SMB credentials in subsequent deployments. To use Zip deployment, follow below steps: create an app registration. In the VSCode deployment, provide Client ID, Object ID and Client secret values. If there are any concerns with creating App registration, you can continue to use SMB deployment option by choosing "Use SMBDeployment For Hybrid" in the Extensions configuration of VSCode If you would like to use zip deployment in an existing Logic App, you will need to manually add the app settings as indicated here. The Zip deployment APIs can be used in CI/CD pipelines as well for DevOps deployment. We will be publishing another blog with detailed steps on the DevOps process. Support for more regions: We are pleased to announce the expansion of our hybrid deployment support to additional regions, in response to valuable customer feedback. This enhancement aims to better meet the diverse geographic and operational requirements of your businesses. The hybrid deployment is now available in the following regions: Central US, East Asia, East US, North Central US, Southeast Asia, Sweden Central, UK South, West Europe, and West US. Logic Apps Rules Engine Support on Linux containers: In this release, we have added support for Azure Logic Apps Rules Engine to run on Linux containers which enables customers to use the Rules Engine capabilities in Hybrid Logic Apps. Improvements for Effective Scaling and Performance: We have introduced few improvements in the runtime storage and the scaling behaviour aimed at improving the performance and achieving effective scaling. Please refer to the following articles: Scaling mechanism in hybrid deployment model for Azure Logic Apps Standard | Microsoft Community Hub Hybrid deployment model for Logic Apps- Performance Analysis and Optimization recommendations | Microsoft Community Hub Diagnostic tool: To assist with troubleshooting the environment configuration issues, we have created a troubleshooting tool, which will help you review the health of all the components of the hybrid deployment and provide insights. You can find the script in our GitHub repository. Select the troubleshoot.ps1 file and copy it to a folder and run the script using PowerShell. This script should be run where you have access to kubectl. References: Create Standard logic app workflows for hybrid deployment - Azure Logic Apps | Microsoft Learn Set up your own infrastructure for Standard logic app workflows - Azure Logic Apps | Microsoft Learn Set up and view enhanced telemetry for Standard workflows - Azure Logic Apps | Microsoft Learn1.7KViews1like0CommentsLogic Apps Aviators Newsletter - June 25
In this issue: Ace Aviator of the Month News from our product group News from our community Ace Aviator of the Month April’s Ace Aviator: Andrew Wilson What's your role and title? What are your responsibilities? I am the Chief Consultancy Officer at Black Marble, a multi-award-winning software company with a big focus on the Microsoft stack. I work with a talented team of consultants to help our customers get the most out of Azure. My role is all about enabling organisations to modernise, integrate, and optimise their systems, always with an eye on DevOps best practices. I’m involved across most of the software development lifecycle, but my focus tends to lean toward consultations, gathering requirements, and architecting solutions that solve real-world problems. I work across a range of areas including application modernisation, BizTalk to Azure Integration Services (AIS) migrations, system integrations, and cloud optimisation. Over time, I've developed a strong focus on Azure, especially around AIS. In short, I help bridge the gap between technical possibilities and business needs, making sure the solutions we design are both practical and future-ready. Can you give us some insights into your day-to-day activities and what a typical day in your role looks like? No two days are quite the same which keeps things interesting! I usually kick things off with a quick plan for the day (and a bit of reshuffling for the week ahead) to make sure we’re focused on what matters most for both customers and the team. My time is a mix of customer-facing work, sales conversations with new prospects, and supporting existing clients, whether that’s through solution design, quick fixes, or hands-on consultancy. I’m often reviewing or writing proposals and architectures, and jumping in to support the team on delivery when needed. There’s always some active learning in the mix too, reading, experimenting, or spinning up quick ideas to explore better ways of doing things. We don’t work in silos at Black Marble, so I’ll often jump in where I can add value, whether or not I’m directly on the project. It’s a real team effort, and that collaboration is a big part of what makes the role so rewarding. What motivates and inspires you to be an active member of the Aviators/Microsoft community? I’ve always enjoyed the challenge of bringing systems and applications together, there’s something really satisfying about seeing everything click into place and knowing it’s driving real business value What makes the Aviators and wider Microsoft community special is that everyone shares that same excitement. It’s a group of people who genuinely care about solving problems, pushing technology forward, and learning from one another. Being part of that kind of community is motivating in itself, we’re all collaborating, sharing ideas, and helping shape a better, more connected future. It’s hard not to be inspired when you’re surrounded by people who are just as passionate about the work as you are. Looking back, what advice do you wish you had been given earlier that you'd now share with those looking to get into STEM/technology? Stay curious, always ask “why,” and don’t be afraid to get things wrong, because you will, and that’s how you learn. Some of the best breakthroughs come after a few missteps (and maybe a bit of head-scratching). It’s easy to look around and feel like others have it all figured out, don’t let that discourage you. Everyone’s journey is different, and what looks effortless on the outside often has a lot of trial and error behind it. One of the best things about STEM is its diversity, there are so many different roles, paths, and people in this space. Whether you’re hands-on with code, designing systems, or solving data challenges, there’s a place for you. It’s not a one-size-fits-all, and that’s what makes it exciting. Most importantly, share what you learn. Even if something’s been “done,” your take on it might be exactly what someone else needs to see to help them get started. And yes, imposter syndrome is real, but don’t let it silence you. You belong here just as much as anyone else. What has helped you grow professionally? A big part of my growth has come from simply committing to continuous learning, whether that’s diving into new tech, attending conferences like Integrate, or being part of user groups where ideas (and challenges) get shared openly. I’ve also learned to say yes to opportunities, even when they’ve felt a bit daunting at first. Pushing through the unknown, especially with the support of a great team and community, has led to some of my most rewarding experiences. And finally, I try to approach everything with the mindset that I’m someone others can count on. That sense of responsibility has helped me stay focused, accountable, and constantly improving. If you had a magic wand that could create a feature in Logic Apps, what would it be and why? Wow, what an exciting question! If I had a magic wand, the first thing I’d add is having the option to throw exceptions that can be caught by try-catch scope blocks, this would bring much-needed clarity and flexibility to error handling. It’s a feature that would really help build more resilient and maintainable solutions. Then, the ability to break or continue loops, sometimes you need that fine-tuned control to keep your workflows running smoothly without extra workarounds. And lastly, full GA support for unit and integration testing, because testing is the backbone of reliable software, and having that baked in would save so much time and stress down the line. News from our product group Logic Apps Live May 2025 Missed Logic Apps Live in May? You can watch it here. We focused on the Logic Apps big announcements from Microsoft Build 2025. There are a lot of great things to check! Announcing agent loop: Build AI Agents in Azure Logic Apps The era of intelligent business processes has arrived! Today, we are excited to announce agent loop, a groundbreaking new capability in Azure Logic Apps to build AI agents into your enterprise workflows. With agent loop, you can embed advanced AI decision-making directly into your processes – enabling your apps and automation to not just follow predefined steps, but to reason, adapt, and act autonomously towards goals. Agent Loop Demos We announced the public preview of agent loop at Build 2025. Agent Loop is a new feature in Logic Apps to build AI Agents for use cases that span across industry domains and patterns. In this article, share with you use cases implemented in Logic Apps using agent loop and other features. Announcement: Azure Logic Apps Document Indexer in Azure Cosmos DB We’re excited to announce the public preview of Azure Logic Apps as a document indexer for Azure Cosmos DB!00 With this release, you can now use Logic Apps connectors and templates to ingest documents directly into Cosmos DB’s vector store—powering AI workloads like Retrieval-Augmented Generation (RAG) with ease. Announcement: Logic Apps connectors in Azure AI Search for Integrated Vectorization We’re excited to announce that Azure Logic Apps connectors are now supported within AI Search as data sources for ingestion into Azure AI Search vector stores. This unlocks the ability to ingest unstructured documents from a variety of systems—including SharePoint, Amazon S3, Dropbox and many more —into your vector index using a low-code experience. Announcement: Power your Agents in Azure AI Foundry Agent Service with Azure Logic Apps We’re excited to announce the Public Preview of two major integrations that bring the power of Azure Logic Apps to AI Agents in Foundry – Logic Apps as Tools and AI Agent Service Connector. Learn more on our announcement post! Codeful Workflows: A New Authoring Model for Logic Apps Standard Codeful Workflows expand the authoring and execution models of a Logic Apps Standard, offering developers the ability to implement, test and run workflows using an imperative programming model both locally and in the cloud. Announcing the General Availability of the Azure Logic Apps Rules Engine we are announcing the General Availability of our Azure Logic Apps Rules Engine. A deterministic rules engine runtime based on the RETE algorithm that allows in-memory execution, prioritization, and reevaluation of business rules in Azure Logic Apps. Integration Environment Update – Unified experience to create and manage alerts We’re excited to announce the next milestone in our journey to simplify monitoring across Azure Integration Services. As a follow-up to our earlier preview release on unified monitoring and dashboards, we’re now making it easier than ever to configure alerts for your integration applications. Automate Invoice data extraction with Logic Apps and Document Intelligence This blog post demonstrates how you can use Azure Logic Apps, the new Analyze Document Details action, and Azure OpenAI to automatically convert invoice images into structured data and store them in Azure Cosmos DB. Log Ingestion to Azure Log Analytics Workspace with Logic App Standard Discover how to send logs to Azure Log Analytics Workspace using Logic App Standard for VNet integration. Learn about shared key authentication and HTTP action configuration for seamless log ingestion. Generating Webhook Action Callback URL with Primary or secondary Access Key Learn how to manage Webhook action callback URLs in Azure Logic Apps when regenerating access keys. Discover how to use the accessKeyType property to ensure seamless workflow execution and maintain security. Announcing the Public Preview of the Applications feature in Azure API management Discover the new Applications feature in Azure API Management, enabling OAuth-based access to APIs and products. Streamline secure API access with built-in OAuth 2.0 application-based authorization. GA: Inbound private endpoint for Standard v2 tier of Azure API Management Today, we are excited to announce the general availability of inbound private endpoint for Azure API management Standard v2 tier. Securely connect clients in your private network to the API Management gateway using Azure Private Link. Announcing the open Public Preview of the Premium v2 tier of Azure API Management Announcing the public preview of Azure API Management Premium v2 tier. Experience superior capacity, highest entity limits, and unlimited calls with enhanced security and networking flexibility. Announcing Federated Logging in Azure API Management Announcing federated logging in Azure API Management. Gain centralized monitoring for platform teams and autonomy for API teams, streamlining API management with robust security and operational visibility. Introducing Workspace Gateway Metrics and Autoscale in Azure API Management Introducing workspace gateway metrics and autoscale in Azure API Management. Efficiently monitor and scale your gateway infrastructure with real-time insights and automated scaling for enhanced reliability and cost efficiency. Introducing Model Logging, Import from AI Foundry, and extended model support in AI Gateway Introducing workspace gateway metrics and autoscale in Azure API Management. Efficiently monitor and scale your gateway infrastructure with real-time insights and automated scaling for enhanced reliability and cost efficiency. Expose REST APIs as MCP servers with Azure API Management and API Center (now in preview) Discover how to expose REST APIs as MCP servers with Azure API Management and API Center, now in preview. Enhance AI integration with secure, observable, and scalable API operations. Now in Public Preview: System events for data-plane in API Management gateway Announcing the public preview of new data-plane system events in Azure Event Grid for the Azure API Management managed gateway. Gain near-real-time visibility into critical operations, automate responses, and prevent disruptions. News from our community Agentic AI – A Potential Black Swan Moment in System Integration Video by Ahmed Bayoumy Discover how Agentic Logic Apps are revolutionizing system integration with AI-driven workflows. Learn how this innovative approach transforms business processes by understanding goals, deciding actions, and using predefined tools for smart orchestration. Microsoft Build: Behind the Scenes with Agent Loop Workflow A New Phase in AI Evolution Video by Ahmed Bayoumy Explore how Agent Loop brings “human in the loop” control to enterprise workflows, on this video by Ahmed, sharing insights directly from Microsoft Build 2025, in a chat with Kent Weare and Divya Swarnkar. Microsoft Build 2025: Azure Logic Apps is Now Your AI Agent Superpower! Post by Sagar Sharma Discover how Azure Logic Apps is transforming AI agent development with new capabilities unveiled at Microsoft Build 2025. Learn about Agent Loop, AI Foundry integration, Document Indexer, and more for intelligent, adaptive workflows. Everyone is talking about AI Agents — Here’s how to actually build one that works Post by Mateusz Partyka Learn how to build effective AI agents with practical strategies and insights. Discover tips on choosing the right tech stack, prototyping fast, managing model costs, and prompt engineering for optimal results. Agent Loop | Azure Logic Apps Just Got Smarter Post by Andrew Wilson Discover Agent Loop in Azure Logic Apps – now in preview - a revolutionary AI-powered integration feature. Enhance workflows with advanced decision-making, context retention, and adaptive actions for smarter automation. Step-by-Step Guide to Azure Logic Apps Agent Loop Post by Stephen W. Thomas Dive into the step-by-step guide for creating AI Agents with Azure Logic Apps Agent Loop – now in preview. Learn to leverage 1300+ connectors, set up OpenAI models, and build intelligent workflows with no-code integration. You can also follow Stephen’s video tutorial Confessions of a Control Freak: How I Learned to Love Low Code (with Logic Apps) Post by Peter Mugisha Discover how a self-confessed control freak learned to embrace low-code development with Azure Logic Apps. From skepticism to advocacy, explore the journey of efficient integration and streamlined workflows. Logic Apps Standard vs. Large Files: Common Hurdles and How to Beat Them Post by Şahin Özdemir Learn how to overcome common hurdles when handling large files in Logic Apps Standard. Discover strategies for scaling, offloading memory-intensive operations, and optimizing performance for efficient integration. There is a new-new Data Mapper for Logic App Standard Post by Sandro Pereira Discover the new Data Mapper for Logic App Standard, now in public preview. Enjoy a modern BizTalk-style mapper with code-first, schema-aware experience, supporting XSLT 3.0, XSD, and JSON schemas for efficient data mapping! A Friday Fact from Sandro Pereira. The name of When a HTTP request is received trigger affects the workflow URL Post by Sandro Pereira Discover how the name of the "When a HTTP request is received" trigger affects the workflow URL in Azure Logic Apps. Learn best practices to avoid integration issues and ensure consistent endpoint paths. Changing APIM Operations Doesn’t Update their PathTemplate Post by Luis Rigueira Learn how to handle PathTemplate issues in Azure Logic Apps Standard when switching APIM operations. Ensure correct endpoint paths to avoid misleading results and streamline your workflow. It is a Friday Fact, brought to you by Luis Rigueira!398Views0likes0Comments🧩 Use Index + Direct Access to pull data across loops in Data Mapper
When working with repeating structures in Logic Apps Data Mapper, you may run into situations where two sibling loops exist under the same parent. What if you need to access data from one loop while you’re inside the other? This is where the Direct Access function, used in combination with Index, can save the day. 🧪 Scenario In this pattern, we’re focusing on the schema nodes shown below: 📸 Source & Destination Schemas (with loops highlighted) In the source schema: Under the parent node VehicleTrips, we have two sibling arrays: Vehicle → contains VehicleRegistration Trips → contains trip-specific values like VehicleID, Distance, and Duration In the destination schema: We're mapping into the repeating node Looping/Trips/Trip It expects each trip’s data along with a flattened VehicleRegistration value that combines both: The current trip’s VehicleID The corresponding vehicle’s VehicleRegistration The challenge? These two pieces of data live in two separate sibling arrays. 🧰 Try it yourself 📎 Download the sample files from GitHub Place them into the following folders in your Logic Apps Standard project: Artifacts → Source, destination and dependency schemas (.xsd) Map Definitions → .lml map file Maps → The .xslt file generated when you save the map Then right-click the .lml file and select “Open with Data Mapper” in VS Code. 🛠️ Step-by-step Breakdown ✅ Step 1: Set up the loop over Trips Start by mapping the repeating Trips array from the source to the destination's Trip node. Within the loop, we map: Distance Duration These are passed through To String functions before mapping, as the destination schema expects them as string values. As you map the child nodes, you will notice a loop automatically added on parent nodes (Trips->Trip) 📸 Mapping Distance and Duration nodes (context: we’re inside Trips loop) 🔍 Step 2: Use Index and Direct Access to bring in sibling loop values Now we want to map the VehicleRegistration node at the destination by combining two values: VehicleID (from the current trip) VehicleRegistration (from the corresponding vehicle) ➡️ Note: Before we add the Index function, delete the auto-generated loop from Trips to Trip To fetch the matching VehicleRegistration: Use the Index function to capture the current position within the Trips loop 📸 Index setup for loop tracking Use the Direct Access function to retrieve VehicleRegistration from the Vehicle array. 📘 Direct Access input breakdown The Direct Access function takes three inputs: Index – from the Index function, tells which item to access Scope – set to Vehicle, the array you're pulling from Target Node – VehicleRegistration, the value you want This setup means: “From the Vehicle array, get the VehicleRegistration at the same index as the current trip.” 📸 Direct Access setup 🔧 Step 3: Concatenate and map the result Use the Concat function to combine: VehicleID (from Trips) VehicleRegistration (from Vehicle, via Direct Access) Map the result to VehicleRegistration in the destination. 📸 Concat result to VehicleRegistration ➡️ Note: Before testing, delete the auto-generated loop from Vehicle to Trip 📸 Final map connections view ✅ Step 4: Test the output Once your map is saved, open the Test panel and paste a sample payload. You should see each Trip in the output contain: The original Distance and Duration values (as strings) A VehicleRegistration field combining the correct VehicleID and VehicleRegistration from the sibling array 📸 Sample Trip showing the combined nodes 💬 Feedback or ideas? Have feedback or want to share a mapping challenge? Open an issue on GitHubHybrid deployment model for Logic Apps- Performance Analysis and Optimization recommendations
A few weeks ago, we announced the Public Preview Refresh release of Logic Apps hybrid deployment model that allows customers to run Logic Apps workloads on a customer managed infrastructure. This model provides the flexibility to execute workflows, either on-premises or in any cloud environment, thereby offering enhanced control over the operation of logic apps. By utilizing customer-managed infrastructure, organizations can adhere to regulatory compliance requirements and optimize performance according to their specific needs. As customers consider leveraging hybrid environments, understanding the performance of logic apps under various configurations and scenarios becomes critical. This document offers an in-depth performance evaluation of Azure Logic Apps within a hybrid deployment framework. It examines, several key factors such as CPU and memory allocation and scaling mechanisms, providing valuable insights aimed at maximizing the application’s efficiency and performance. Achieving Optimal Logic Apps Performance in Hybrid Deployments In this section, we will explore the key aspects that affect Logic Apps performance when deployed in a hybrid environment. Factors such as the underlying infrastructure of the Kubernetes environment, SQL configuration and scaling configuration can significantly impact the efficiency of workflows and the overall performance of the applications. The following blog entry provides details of the scaling mechanism of Hybrid deployment model - Scaling mechanism in hybrid deployment model for Azure Logic Apps Standard | Microsoft Community Hub Configure Container Resource allocation: When you create a Logic App, a default value of 0.5 vCPU and 1GiB of memory would be allocated. From the Azure Portal, you can modify this allocation from the Container blade. - Create Standard logic app workflows for hybrid deployment - Azure Logic Apps | Microsoft Learn Currently, the maximum allocation is set to 2vCPU and 4 GiB memory per app. In the future, there would be a provision made to choose higher allocations. For CPU intense/memory intense processing like custom code executions, select a higher value for these parameters. In the next section, we will be comparing the performance with different values of the CPU and memory allocation. This allocation would impact the billing calculation of the Logic App resource. Refer vCPU calculation for more details on the billing impact. Optimize the node count and size in the Kubernetes cluster. Kubernetes runs application workloads by placing containers into Pods to run on Nodes. A node may be a virtual or physical machine, depending on the cluster. A node pool is a group of nodes that share the same configuration (CPU, Memory, Networking, OS, maximum number of pods, etc.). You can choose the capacity (cores and memory), minimum node count and maximum node count for each node pool of the Kubernetes cluster. We recommend allocating a higher capacity for processing CPU intense, or memory intense applications Configure Scale rule settings: For a Logic App resource, we recommend you configure the maximum and minimum replicas which could be scaled out when a scale event occurs. A higher value for the max replicas helps in sudden spikes in the number of application requests. The interval with which the scaler checks for the scaling event and the cooldown period for the scaling event can also be configured from the Scale blade of Logic Apps resource. These parameters impact the scaling pattern. Optimize the SQL server configuration: The hybrid deployment model uses Microsoft SQL for runtime storage. As such, there are lot of SQL operations performed throughout the execution of the workflow and SQL capacity has a significant impact on the performance of the app. Microsoft SQL server could either be a SQL server on Windows, or an Azure SQL database. Few recommendations on the SQL configuration for better performance: If you are using, Azure SQL database, run it on a SQL elastic pool. If you are using SQL server on Windows, run with at least 4vCPU configuration. Scale out the SQL server once the CPU usage of the SQL server hits 60-70% of the total available CPU. Performance analysis: For this performance analysis exercise, we have used a typical enterprise integration scenario which includes the below components. Data transformation: XSLT transformation, validation, and XML parsing actions Data routing: File system connector for storing the transformed content in a file share. Message queuing: RabbitMQ connector for sending the transformation result to Rabbit MQ queue endpoint. Control operations: For-each loop for looping through multiple records, condition execution, Scope, and error handling blocks. Request response: The XML data transmitted via HTTP request, and the status returned as a response. Summary: For these tests, we used the following environment settings: Kubernetes cluster: AKS cluster with Standard D2sV3 (2vCPU, 8GiBmemory) Max replicas: 20 Cooldown period: 300 seconds Polling interval: 30 With the above environment and settings, we have performed multiple application tests with different configuration of SQL server, resource allocation and test durations using Azure load testing tool. In the following table, we have summarized the response time, throughput, and the total vCPU consumption for each of these configurations. You can check each scenario for detailed information. Configuration Results Scenario SQL CPU and Memory allocation per Logic App Test duration Load 90 th Percentile Response time Throughput Total vCPU consumed Scenario 1 SQL general purpose V2 1vCPU/2GiB Memory 10 minutes with 50 users 503 requests 68.62 seconds 0.84/s 3.42 Scenario 2 SQL Elastic pool-4000DTU 1vCPU/2GiB Memory 10 minutes with 50 users 1004 requests 40.74 seconds 1.65/s 3 Scenario 3 SQL Elastic pool-4000DTU 2vCPU/4GiB Memory 10 minutes with 50 users 997 requests 40.63 seconds 1.66/s 4 Scenario 4 SQL Elastic pool-4000DTU 2vCPU/4GiB Memory 30 minutes with 50 users 3421 requests 26.6Seconds 1.9/s 18.6 Scenario 5 SQL Elastic pool-4000DTU 0.5vCPU/1GiB Memory 30 minutes with 50 users 3055 requests 31.38 seconds 1.7/s 12.4 Scenario 6 SQL 2022 Enterprise on Standard D4s V3 VM 0.5vCPU/1GiB Memory 30 minutes with 50 users 4105 requests 27.15 seconds 2.28/s 10 Scenario 1: SQL general purpose V2 with 1vCPU and 2 GiB Memory – 10 minutes test with 50 users In this scenario, we conducted a load test for 10 minutes with 50 users with the Logic App configuration of: 1 vCPU and 2 GiB Memory and Azure SQL database running on General purpose V2 plan. There were 503 requests with multiple records in each payload and it achieved the 68.62 seconds as the 90 th percentile response time and a throughput of 0.84 requests per second. Scaling: The Kubernetes nodes scaled out to 12 nodes and in total 3.42 vCPUs used by the app for the test duration. SQL Metrics: The CPU usage of the SQL server reached 90% of CPU usage quite early and stayed above 90% for the remaining duration of the test. From our backend telemetry as well, we observed that the actions executions were faster, but there was latency between the actions, which indicates SQL bottlenecks. Scenario 2: SQL elastic pool, with 1vCPU and 2 GiB memory- 10 minutes test with 50 users In this scenario, we conducted a load test for 10 minutes with 50 users with the Logic App configuration of: 1 vCPU and 2 GiB Memory and Azure SQL database running on a SQL elastic pool with 4000 DTU. There were 1004 requests with multiple records in each payload and it achieved the 40.74 seconds as the 90 th percentile response time and a throughput of 1.65 requests per second. Scaling: The Kubernetes nodes scaled out to 15 nodes and in total 3 vCPUs used by the app for the test duration. SQL Metrics: The SQL server’s CPU utilization peaked to 2% of the elastic pool. Scenario 3: SQL elastic pool, with 2vCPU and 4 GiB memory- 10 minutes test with 50 users In this scenario, we conducted a load test for 10 minutes with 50 users with the Logic App configuration of 2 vCPU and 4 GiB Memory and Azure SQL database running on a SQL elastic pool with 4000 DTU. There were 997 requests with multiple records in each payload and it achieved the 40.63 seconds as the 90 th percentile response time and a throughput of 1.66 requests per second. Scaling: The Kubernetes nodes scaled out to 21 nodes and in total 4 vCPUs used by the app for the test duration. SQL Metrics: The SQL server’s CPU utilization peaked to 5% of the elastic pool. Scenario 4: SQL elastic pool, with 2vCPU and 4 GiB memory- 30 minutes test with 50 users In this scenario, we conducted a load test for 30 minutes with 50 users with the Logic App configuration of: 2 vCPU and 4 GiB Memory and Azure SQL database running on a SQL elastic pool with 4000 DTU. There were 3421 requests with multiple records in each payload and it achieved the 26.67 seconds as the 90 th percentile response time and a throughput of 1.90 requests per second. Scaling: The Kubernetes nodes scaled out to 20 nodes and in total 18.6 vCPUs used by the app for the test duration. SQL Metrics: The SQL server’s CPU utilization peaked to 4.7% of the elastic pool. Scenario 5: SQL Elastic pool, with 0.5vCPU and 1 GiB memory- 30 minutes test with 50 users In this scenario, we have conducted a load test for 30 minutes with 50 users with the Logic App configuration of 0.5 vCPU and 1 GiB Memory and Azure SQL database running on a SQL elastic pool with 4000 DTU. There were 3055 requests with multiple records in each payload and it achieved the 31.38 seconds as the 90 th percentile response time and a throughput of 1.70 requests per second. Scaling: The Kubernetes nodes scaled out to 18 nodes and in total 12.4 vCPUs used by the app for the test duration. SQL Metrics: The SQL server’s CPU utilization peaked to 8.6% of the elastic pool CPU. Scenario 6: SQL 2022 Enterprise Gen2 on Windows 2022 on Standard D4s v3 image, with 0.5vCPU and 1 GiB memory- 30 minutes test with 50 users In this scenario, we conducted a load test for 30 minutes with 50 users with the Logic App configuration of: 0.5 vCPU and 1 GiB Memory and Azure SQL database running on an on-premises SQL 2022 Enterprise Gen2 version running on a Windows 2022 OS with Standard D4s v3 image (4 vCPU and 16GIB memory) There were 4105 requests with multiple records in each payload and it achieved the 27.15 seconds as the 90 th percentile response time and a throughput of 2.28 requests per second. Scaling: The Kubernetes nodes scaled out to 8 nodes and in total 10 vCPUs used by the app for the test duration. SQL metrics: The CPU usage of the SQL server went above 90% after few minutes and there was latency on few runs. Findings and recommendations: The following are the findings and recommendations for this performance exercise. Consider that this load test was conducted using unique conditions. If you conduct a similar test, the results and findings might vary, depending on factors such as workflow complexity, configuration, resource allocation and network configuration. The KEDA scaler performs the scale-out and scale-in operations faster, as such, while the total vCPU usage remains quite low, though the nodes scaled out in the range of 1-20 nodes. The SQL configuration plays a crucial role in reducing the latency between the action executions. For a satisfactory load test, we recommend starting with at least 4vCPU configuration on SQL server and scale out once CPU usage of the SQL server hits 60-70% of the total available CPU. For critical applications, we recommend having a dedicated SQL database for better performance. Increasing the dedicated vCPU allocation of the Logic App resource is helpful for the SAP connector, Rules Engine, .NET Framework based custom code operations and for the applications with many complex workflows. As a general recommendation, regularly monitor performance metrics and adjust configurations to meet evolving requirements and follow the coding best practices of Logic Apps standard. Consider reviewing the following article, for recommendations to optimize your Azure Logic Apps workloads: https://techcommunity.microsoft.com/blog/integrationsonazureblog/logic-apps-standard-hosting--performance-tips/3956971Transform Your Integration Strategy with Azure Integration Services
Still on Microsoft BizTalk Server or other legacy integration solutions? If you're relying on BizTalk or other legacy systems, you're already feeling the pain: rising costs, performance bottlenecks, and limited scalability. These outdated systems are holding you back, but the good news? The time to modernize is NOW and we’ve got the event that will show you how. Join us at "Unleash AI Innovation with a Modern Integration Platform and an API-First Strategy", where industry leaders like Visa, LyondellBasell, and Metcash share how they reimagined their integration landscape with Azure. This is your blueprint for moving beyond the limitations of legacy systems and unlocking innovation with a cloud-native, AI-ready approach. By attending, you’ll gain exclusive insights into how leading organizations have turned integration challenges into competitive advantages, positioning themselves for future growth with Azure. Real Stories, Real Impact Visa: Revolutionizing Operations with Azure Logic Apps Visa’s journey from BizTalk to Azure Logic Apps isn’t just a story of modernization—it’s a game-changer. By automating complex workflows and managing over 100 HR systems, Visa reduced manual intervention and slashed infrastructure costs by 95%. With Azure, Visa is now set to scale operations and leverage AI for continued growth. Discover the details of how Visa is preparing for tomorrow’s challenges today. LyondellBasell: Scaling Beyond BizTalk LyondellBasell, a global leader in chemicals, broke free from BizTalk’s limitations. Azure Integration Services empowered them with hybrid connectivity and real-time visibility, enabling them to streamline workflows and boost developer efficiency by 50%. Learn how they unlocked faster decision-making and stronger business outcomes. Brisbane City Council: From Legacy to API-First Agility Brisbane City Council, Australia’s largest local government, faced limitations with legacy integration solution. With Azure, they cut data processing times from 1 hour to under 5 minutes. The shift to an API-first approach future-proofed their strategy. Find out how this local government innovated without disruption. Metcash: Peak Retail Performance, Zero Downtime After moving to Azure Integration Services, Metcash, an Australian wholesaler, processed 8.65 million API calls in 48 hours during a major retail event without a single minute of downtime. With Azure, they achieved unrivaled resilience and scalability, all while slashing costs. See how Metcash’s transformation can serve as your blueprint for success. What You’ll Learn: How to cut costs and modernize your integration landscape Ways to scale with cloud-native solutions and AI-driven automation Strategies to secure every API with enterprise-grade governance Real-world migration paths from BizTalk and other platforms Choose Your Region and Register Now US/Canada: Reserve your seat today! Day 1: Tuesday, 29 April 2025 | 9:00 AM – 12:30 PM PDT Day 2: Wednesday, 30 April 2025 | 9:00 AM – 12:30 PM PDT Australia/New Zealand: Reserve Your Seat Today! Day 1: Wednesday, 30 April 2025 | 9:00 AM – 12:30 PM AEDT Day 2: Thursday, 1 May 2025 | 9:00 AM – 12:30 PM AEDT Europe: Reserve Your Seat Today! Day 1: Tuesday, 29 April 2025 | 9:00am – 12:30pm BST Day 2: Wednesday, 30 April 2025 | 9:00am – 12:30pm BST671Views0likes0CommentsQ1’2025: Azure Integration Services Quarterly Highlights and Insights
From reinventing hybrid integration to unlocking AI-powered productivity and simplifying API management across ecosystems, the first quarter of 2025 was all about making integration smarter, faster, and more accessible for everyone. Whether you're a developer modernizing legacy workflows, an IT pro securing mission-critical APIs, or a business technologist building intelligent automations, Azure Integration Services and Azure API Management are moving at the speed of innovation. Here’s what stood out this quarter and how these updates can help accelerate your next move. Product Announcements We’ve been hard at work delivering powerful new capabilities across Azure Logic Apps and Azure API Management designed to help you move faster, modernize smarter, and innovate with confidence. Here’s what’s new: Azure Logic Apps Hybrid Deployment Model [Public Preview Refresh] The public preview of the Logic Apps hybrid deployment model has been refreshed, allowing you to run Logic Apps on your own infrastructure, whether that’s on-premises, at the edge, or across multi-cloud environments. New highlights include: .NET Framework custom code support on Linux containers – Bring over BizTalk-style transformations without a rewrite. SAP built-in connector now supported on Linux containers – Run SAP workflows anywhere you deploy Hybrid Logic Apps. New Rabbit MQ built-in connector – Seamlessly connect to on-premises message brokers for hybrid queueing scenarios. Read the full update. Data Mapper in Azure Logic Apps (Standard) [Public Preview Refresh] The updated Data Mapper in Logic Apps (Standard) is now even more powerful, bringing forward critical BizTalk-era capabilities with a modern UX: Run enterprise-grade XSLT logic from reusable templates Extract values from complex XML using XPath Build and reuse custom XML functions within your maps Dive into the details. Azure API Center Continues to Evolve Azure API Center continues to evolve as the single source of truth for your organization’s APIs. Two key updates now in public preview: Managed API Center Portal – A turnkey developer portal to explore, customize, and govern APIs with Microsoft Entra integration. Set up instructions for the API Center portal. ️Amazon API Gateway integration – Automatically import and keep APIs in sync from AWS into API Center. No manual updates. No drift. Learn about the integration. Developer tooling: A better Azure API Management experience in VS Code We’ve released a new version of the Azure API Management VS Code extension, focused on making it easier to draft, debug, and explain API policies fueled by GitHub Copilot and AI enhancements. It’s all about helping developers move faster and smarter. Learn more. Your AI and modernization playbooks for Logic Apps Bring AI into your workflows with the Logic Apps AI Playbook Ready to infuse intelligence into your apps, bots, or business processes? Our AI Playbook for Azure Logic Apps gives you a practical guide to building smarter workflows fast. Whether you're starting from scratch or scaling existing automations, you’ll find step-by-step examples to accelerate your journey: Build conversational apps using Azure OpenAI and Azure AI Search Automate document intelligence with built-in connectors to Azure AI Document Intelligence Power real-time retrieval with RAG (Retrieval-Augmented Generation) patterns Create intelligent agents using Semantic Kernel and Logic Apps orchestration Explore AI scenarios and examples. Still running mainframe workloads? We've got a playbook for that too. Modernizing mission-critical systems doesn’t have to mean starting over. With our Mainframe and Midrange Modernization Playbook, Logic Apps helps you replatform legacy workloads with less risk and more flexibility. Discover how to connect to existing systems, orchestrate modern services, and unlock agility without rewriting everything. Explore the mainframe modernization guide. Partnership Announcements Azure API Management comes with a powerful, fully integrated developer portal built to help teams publish APIs, onboard developers, and manage access with enterprise-grade scalability and security. For many organizations, this out-of-the-box portal offers exactly what they need. But some businesses need more advanced branding, tailored onboarding journeys, or completely bespoke developer experience. That’s why we’re partnering with Pronovix and ApiBoost, two industry leaders in custom API portals. These partnerships give customers even more flexibility to extend and personalize their API programs. Customer Success Stories Telefônica Brasil is transforming call center operations for 115M+ users using Azure OpenAI and Azure API Management. With AI-driven insights and secure API access for 33,000 employees, they’ve cut average handling time by 9% and accelerated innovation. Read the story. Intermountain Health has saved thousands of hours by integrating Azure API Management, OpenAI, and Arize AI into their cloud AI infrastructure. Responsible AI deployment meets real-time observability for better patient outcomes. Learn more. Delta Dental of California unified its integration architecture across APIs, messaging, and app delivery using Azure API Management, App Gateway, Redis, and more—streamlining operations and unlocking agility. Explore the case study. NTT Communications uses Azure Logic Apps to automate threat response as part of its Microsoft Sentinel-powered security stack, enabling faster correlation of security signals across systems. Read the story. What’s coming up: Events you won’t want to miss Unleash AI Innovation with a Modern Integration Platform and an API-First Strategy Join us for a dynamic two-day virtual event April 29-30 where we’ll explore how Azure Integration Services and Azure API Management are powering intelligent, connected experiences. Hear directly from Azure product leaders, industry analysts, and enterprise customers as they share how they’re using Azure to scale AI-driven innovation, modernize integrations, and build API-first strategies that unlock new business value. This global event is designed to fit your schedule, with live sessions tailored for key regions: US/Canada: Reserve your seat today! Day 1: Tuesday, 29 April 2025 | 9:00 AM – 12:30 PM PDT Day 2: Wednesday, 30 April 2025 | 9:00 AM – 12:30 PM PDT Australia/New Zealand: Reserve Your Seat Today! Day 1: Wednesday, 30 April 2025 | 9:00 AM – 12:30 PM AEDT Day 2: Thursday, 1 May 2025 | 9:00 AM – 12:30 PM AEDT Europe: Reserve Your Seat Today! Day 1: Tuesday, 29 April 2025 | 9:00am – 12:30pm BST Day 2: Wednesday, 30 April 2025 | 9:00am – 12:30pm BST Microsoft Build 2025 Whether you're joining us in Seattle or tuning in online, Microsoft Build is your front-row seat to the future of technology. Taking place May 19–22, this year’s conference will showcase groundbreaking innovations across cloud, AI, developer tools, and of course APIs and integration. Session catalog is live. Stay Connected! Q1 2025 set the stage for even bolder moves in the integration space. With hybrid flexibility, next-gen mapping, cross-cloud API management, and AI in every layer, Azure Integration Services and Azure API Management are ready to power what’s next. Stay tuned and subscribe to the Azure Tech Community blog.497Views0likes0Comments