events
17 TopicsAzure CycleCloud 8.8 and CCWS 1.2 at SC25 and Ignite
Azure CycleCloud 8.8: Advancing HPC & AI Workloads with Smarter Health Checks Azure CycleCloud continues to evolve as the backbone for orchestrating high-performance computing (HPC) and AI workloads in the cloud. With the release of CycleCloud 8.8, users gain access to a suite of new features designed to streamline cluster management, enhance health monitoring, and future-proof their HPC environments. Key Features in CycleCloud 8.8 1. ARM64 HPC Support The platform expands its hardware compatibility with ARM64 HPC support, opening new possibilities for energy-efficient and cost-effective compute clusters. This includes access to the newer generation of GB200 VMs as well as general ARM64 support, enabling new AI workloads at a scale never possible before 2. Slurm Topology-Aware Scheduling The integration of topology-aware scheduling for Slurm clusters allows CycleCloud users to optimize job placement based on network and hardware topology. This leads to improved performance for tightly coupled HPC workloads and better utilization of available resources. 3. Nvidia MNNVL and IMEX Support With expanded support for Nvidia MNNVL and IMEX, CycleCloud 8.8 ensures compatibility with the latest GPU technologies. This enables users to leverage cutting-edge hardware for AI training, inference, and scientific simulations. 4. HealthAgent: Event-Driven Health Monitoring and Alerting A standout feature in this release is the enhanced HealthAgent, which delivers event-driven health monitoring and alerting. CycleCloud now proactively detects issues across clusters, nodes, and interconnects, providing real-time notifications and actionable insights. This improvement is a game-changer for maintaining uptime and reliability in large-scale HPC deployments. Node Healthagent supports both impactful healthchecks which can only run while nodes are idle as well as non-impactful healthchecks that can run throughout the lifecycle of a job. This allows CycleCloud to alert on issues that not only happen while nodes are starting, but also issues that may result from failures for long-running nodes. Later releases of CycleCloud will also include automatic remediation for common failures, so stay tuned! 5. Enterprise Linux 9 and Ubuntu 24 support One common request has been wider support for the various Enterprise Linux (EL) 9 variants, including RHEL9, AlmaLinux 9, and Rocky Linux 9. CycleCloud 8.8 introduces support for those distributions as well as the latest Ubuntu HPC release. Why These Features Matter The CycleCloud 8.8 release marks a significant leap forward for organizations running HPC and AI workloads in Azure. The improved health check support—anchored by HealthAgent and automated remediation—means less downtime, faster troubleshooting, and greater confidence in cloud-based research and innovation. Whether you’re managing scientific simulations, AI model training, or enterprise analytics, CycleCloud’s latest features help you build resilient, scalable, and future-ready HPC environments. Key Features in CycleCloud Workspace for Slurm 1.2 Along with the release of CycleCloud 8.8 comes a new CycleCloud Workspace for Slurm (CCWS) release. This release includes the General Availability of features that were previously in preview, such as Open OnDemand, Cendio ThinLinc, and managed Grafana monitoring capabilities. In addition to previously announced features, CCWS 1.2 also includes support for a new Hub and Spoke deployment model. This allows customers to retain a central hub of shared resources that can be re-used between cluster deployments with "disposable" spoke clusters that branch from the hub. Hub and Spoke deployments enable customers who need to re-deploy clusters in order to upgrade their operating system, deploy new versions of software, or even reconfigure the overall architecture of Slurm clusters. Come visit us at SC25 and MS Ignite To learn more about these features, come visit us at the Microsoft booth at #SC25 in St. Louis, MO and #Microsoft #Ignite in San Francisco this week!Microsoft Discovery: The path to an agentic EDA environment
Generative AI has been the buzz across engineering, science and consumer applications, including EDA. It was the centerpiece of the keynotes at both SNUG and CadenceLive, and it will feature heavily at DAC. Very impressive task specific tools and capabilities powered by traditional and generative AI are being developed by both industry vendors and customers. However, all these solutions are point solutions addressing specific tasks. This leaves the question of how customers will tie it all together and how customers will be able to run and access the LLMs, AI and data resources needed to power these solutions. While our industry has experience developing, running, and maintaining high-performance EDA environments, an AI centric data center running GPUs and low latency interconnect like Infiniband, is not an environment many chip development companies already have or have experience executing. Unfortunately, because LLMs are so resource hungry, it’s difficult to “ease into” a deployment. The Agentic Platform for EDA At the Microsoft Build conference in May, Microsoft introduced the Microsoft Discovery Platform. This platform aims to accelerate R&D across several industry verticals, specifically Biology (Life science and drug discovery), Chemistry (materials and substance discovery), and Physics (semiconductors and multi-physics). Microsoft Discovery provides the platform and capabilities to help customers implement a complete agentic AI environment. Being a cloud-based solution means customers won’t need to manage the AI models or RAG solutions themselves. Running inside the customer’s cloud tenant, the AI models, the data they use, and the results they produce all remain under the customer's control and within the customer’s environment. No data goes back to the Internet and all learning remains with the customer. This gives customers the confidence that they can safely and easily deploy and use AI models while maintaining complete sovereignty over their data and IP. Customers are free to deploy any of the dozens of available AI models offered on Azure. Customers can also deploy and use Graph RAG solutions to improve context and get better LLM responses. This is all available without having to deploy additional hardware or manage a large, independent GPU deployment. Customers testing out generative AI solutions and starting to develop their flows, tools, and methodologies with this new technology can deploy and use these resources as needed. The Microsoft Discovery platform does not try to replace the EDA tools you already have. Instead, it allows you to incorporate those tools into an agentic environment. Without anthropomorphizing, these agents can be thought of as AI driven task engines that can reason and interact with each other or tools. They can be used to make decisions, analyze results, generate responses, take action, or even drive tools. Customers will be able to incorporate existing EDA tools into the platform and drive them with an agent. Microsoft Discovery will even be able to run agents from partners and help customers intelligently tie together multiple capabilities and help automate analysis and decision-making on the flow helping each engineering teams accomplish a greater number of tasks more quickly and achieve increased productivity. HPC Infrastructure for EDA Of course, to run EDA tools, customers need an effective environment to run those tools in. One of the things that has always been true in our industry but is often overlooked is that, as good as the algorithms in the tools are, they’re always limited by the infrastructure it runs on. No matter how fast your algorithm is, running on a slow processor means turn-around time is still going to be slow. No matter how fast your tools are and how new and shiny your servers are, if your file system is a bottleneck, your tool and server will have to wait for the data. The infrastructure you run on sets the speed limit for your job regardless of how fast an engine you have. Most of the AI solutions being discussed for EDA focus only on the engine and ignore the infrastructure. The Microsoft Discovery platform understands this and addresses the issue by having the Azure HPC environment at its core. The HPC core of the platform uses elements familiar to the EDA community. High performance file storage utilizes Azure NetApp Files (ANF). This shared file service uses the same NetApp technology and hardware that many in the EDA community already uses on-prem. ANF delivers unmatched performance for cloud-based file storage, especially for metadata heavy workloads, like EDA. This will help provide EDA workloads with a familiar pathway into the Discovery platform to make use of the AI capabilities for chip design. Customers will also have access to Azure’s fleet of high-performance compute, including the recently released Intel Emerald Rapids-based FXv2, which was developed with large, back-end EDA workloads in mind. FXv2 features 1.8TB of RAM and all core turbo clock speed of 4 GHz. Ideal for large STA, P&R, and PV workloads. For front-end and moderate sized back-end workloads, in addition to the existing HPC compute offerings, Microsoft recently updated the D and E series compute SKUs with Intel Emerald Rapids processors in the v6 versions of those systems, further pushing performance for smaller workloads. Design teams will have access to the required high-performance compute and storage resources to maximize their EDA tools while also taking advantage of the benefits of AI capabilities offered by the platform. The familiar EDA-friendly HPC environment makes migration of existing workloads easier and ensures that tools will run effectively and, more importantly, flows mesh more smoothly. Industry Standards and Interoperability Another aspect of the Microsoft Discovery platform that will be especially important for EDA customers is the fact that the platform will utilize A2A for agent-to-agent communication and MCP for agent-service communication. The reason this is important is because both A2A and MCP are industry standard protocols. Microsoft also expects to support the evolution of these and other newer standards that will emerge in this field, future-proofing your investment. Those of us who have been involved in the various standards and interoperability efforts in semiconductor and EDA over the years will understand that having the platform use industry standards-based interfaces makes adoption of new technology much easier for all users. With AI development rushing forward and everyone, customers and vendors alike, trying to capitalize on gen AI’s promises, there are already independent efforts by customers and vendors to develop capabilities quickly. In the past, this meant that everyone went off in different directions developing mutually exclusive solutions. Vendors would develop mutually exclusive solutions that customers would have to also develop customized solutions to leverage. The various solutions would all work slightly differently, making integration a painful process. The history of VMM, OVM, and UVM was an example of this. As the industry starts to develop AI and agentic environments, the same fragmentation is likely to also happen again. By starting with A2A and MCP, Microsoft is signaling for the industry to align around these industry standard protocols. This will make it easier for agents developed by customers and vendors to interoperate with each other and the Discovery platform. Vendor tools implementing a MCP server interface can directly communicate with customer agents using MCP as well as with the Discovery platform. This makes it easier for our industry to develop interoperable solutions. Similarly, agents that use the A2A protocol to interact with other agents can be more easily integrated if the other agents also communicate using A2A. If you’re going to be building agents for EDA or EDA tools or services that interact with agents, build them using A2A for inter-agent communication and MCP for agent-to-tool/service communication. Generative AI is likely to be the most transformative technology to impact EDA this decade. It likely will be at least as impactful, productivity wise, for us a synthesis, STA, and automatic place and route were in their own ways. To learn more about these innovations, come join the Microsoft team at the Design Automation Conference (DAC) in San Francisco on June 23. At DAC, the Microsoft team will go into depth about the Discovery platform and the larger impact that AI will have on the semiconductor industry. In his opening keynote discussion on Monday, Bill Chappell, Microsoft's CTO for the Microsoft Discovery and Quantum team will discuss AI's impact on science and the semiconductor industry. Serge Leef’s engineering track session will talk about generative AI in chip design, and don't miss Prashant Varshney's detailed explanation of the Microsoft Discovery platform in his Exhibitor Forum session. Visit the Microsoft booth (second floor, 2124) for more in-depth discussions with our team.Join Microsoft @ SC25: Experience HPC and AI Innovation
Supercomputing 2025 is coming to St. Louis, MO, November 16–21! Visit Microsoft Booth #1627 to explore cutting-edge HPC and AI solutions, connect with experts, and experience interactive demos that showcase the future of compute. Whether you’re attending technical sessions, stopping by for a coffee, or joining our partner events, we’ve got something for everyone. Booth Highlights Alpine Formula 1 Showcar: Snap a photo with a real Alpine F1 car and learn how high-performance computing drives innovation in motorsports. Silicon Wall: Discover silicon diversity—featuring chips from our partners AMD and NVIDIA, alongside Microsoft’s own first-party silicon: Maia, Cobalt, and Majorana. NVIDIA Weather Modeling Demo: See how AI and HPC predict extreme weather events with Tomorrow.io and NVIDIA technology. Coffee Bar with Barista: Enjoy a handcrafted coffee while you connect with our experts. Immersive Screens: Watch live demos and visual stories about HPC breakthroughs and AI innovation. Hardware Bar: Explore AMD EPYC™ and NVIDIA GB200 systems powering next-generation workloads. Whether you’re attending technical sessions, stopping by for a coffee and chat with our team, or joining our partner events, we’ve got something for everyone. Conference Details Conference week: Sun, Nov 16 – Fri, Nov 21 Expo hours (CST): Mon, Nov 17: 7:00–9:00 PM (Opening Night) Tue, Nov 18: 10:00 AM–6:00 PM Wed, Nov 19: 10:00 AM–6:00 PM Thu, Nov 20: 10:00 AM–3:00 PM Customer meeting rooms: Four Seasons Hotel Quick links RSVP — Microsoft + AMD Networking Reception (Tue, Nov 18): https://aka.ms/MicrosoftAMD-Mixer RSVP — Microsoft + NVIDIA Panel Luncheon (Wed, Nov 19): Luncheon is now closed as the event is fully booked. Earned Sessions (Technical Program) Sunday, Nov 16 Session Type Time (CST) Title Microsoft Contributor(s) Location Tutorial 8:30 AM–5:00 PM Delivering HPC: Procurement, Cost Models, Metrics, Value, and More Andrew Jones Room 132 Tutorial 8:30 AM–5:00 PM Modern High Performance I/O: Leveraging Object Stores Glenn Lockwood Room 120 Workshop 2:00–5:30 PM 14th International Workshop on Runtime and Operating Systems for Supercomputers (ROSS 2025) Torsten Hoefler Room 265 Monday, Nov 17 Session Type Time (CST) Title Microsoft Contributor(s) Location Early Career Program 3:30–4:45 PM Voices from the Field: Navigating Careers in Academia, Government, and Industry Joe Greenseid Room 262 Workshop 3:50–4:20 PM Towards Enabling Hostile Multi-tenancy in Kubernetes Ali Kanso; Elzeiny Mostafa; Gurpreet Virdi; Slava Oks Room 275 Workshop 5:00–5:30 PM On the Performance and Scalability of Cloud Supercomputers: Insights from Eagle and Reindeer Amirreza Rastegari; Prabhat Ram; Michael F. Ringenburg Room 267 Tuesday, Nov 18 Session Type Time (CST) Title Microsoft Contributor(s) Location BOF 12:15–1:15 PM High Performance Software Foundation BoF Joe Greenseid Room 230 Poster 5:30–7:00 PM Compute System Simulator: Modeling the Impact of Allocation Policy and Hardware Reliability on HPC Cloud Resource Utilization Jarrod Leddy; Huseyin Yildiz Second Floor Atrium Wednesday, Nov 19 Session Type Time (CST) Title Microsoft Contributor(s) Location BOF 12:15–1:15 PM The Future of Python on HPC Systems Michael Droettboom Room 125 BOF 12:15–1:15 PM Autonomous Science Network: Interconnected Autonomous Science Labs Empowered by HPC and Intelligent Agents Joe Tostenrude Room 131 Paper 1:30–1:52 PM Uno: A One‑Stop Solution for Inter‑ and Intra‑Data Center Congestion Control and Reliable Connectivity Abdul Kabbani; Ahmad Ghalayini; Nadeen Gebara; Terry Lam Rooms 260–267 Paper 2:14–2:36 PM SDR‑RDMA: Software‑Defined Reliability Architecture for Planetary‑Scale RDMA Communication Abdul Kabbani; Jie Zhang; Jithin Jose; Konstantin Taranov; Mahmoud Elhaddad; Scott Moe; Sreevatsa Anantharamu; Zhuolong Yu Rooms 260–267 Panel 3:30–5:00 PM CPUs Have a Memory Problem — Designing CPU‑Based HPC Systems with Very High Memory Bandwidth Joe Greenseid Rooms 231–232 Paper 4:36–4:58 PM SparStencil: Retargeting Sparse Tensor Cores to Scientific Stencil Computations Kun Li; Liang Yuan; Ting Cao; Mao Yang Rooms 260–267 Thursday, Nov 20 Session Type Time (CST) Title Microsoft Contributor(s) Location BOF 12:15–1:15 PM Super(computing)heroes Laura Parry Rooms 261–266 Paper 3:30–3:52 PM Workload Intelligence: Workload‑Aware IaaS Abstraction for Cloud Efficiency Anjaly Parayil; Chetan Bansal; Eli Cortez; Íñigo Goiri; Jim Kleewein; Jue Zhang; Pantea Zardoshti; Pulkit Misra; Raphael Ghelman; Ricardo Bianchini; Rodrigo Fonseca; Saravan Rajmohan; Xiaoting Qin Room 275 Paper 4:14–4:36 PM From Deep Learning to Deep Science: AI Accelerators Scaling Quantum Chemistry Beyond Limits Fusong Ju; Kun Li; Mao Yang Rooms 260–267 Friday, Nov 21 Session Type Time (CST) Title Microsoft Contributor(s) Location Workshop 9:00 AM–12:30 PM Eleventh International Workshop on Heterogeneous High‑performance Reconfigurable Computing (H2RC 2025) Torsten Hoefler Room 263 Booth Theater Sessions Monday, Nov 17 — 7:00 PM–9:00 PM Time (CST) Session Title Presenter(s) 8:00–8:20 PM Inside the World’s Most Powerful AI Data Center Chris Jones 8:30–8:50 PM Transforming Science and Engineering — Driven by Agentic AI, Powered by HPC Joe Tostenrude Tuesday, Nov 18 — 10:00 AM–6:00 PM Time (CST) Session Title Presenter(s) 11:00–11:50 AM Ignite Keynotes 12:00–12:20 PM Accelerating AI workloads with Azure Storage Sachin Sheth; Wolfgang De Salvador 12:30–12:50 PM Accelerate Memory Bandwidth‑Bound Workloads with Azure HBv5, now GA Jyothi Venkatesh 1:00–1:20 PM Radiation & Health Companion: AI‑Driven Flight‑Dose Awareness Olesya Sarajlic 1:30–1:50 PM Ascend HPC Lab: Your On‑Ramp to GPU‑Powered Innovation Daniel Cooke (Oakwood) 2:00–2:20 PM Azure AMD HBv5: Redefining CFD Performance and Value in the Cloud Rick Knoechel (AMD) 2:30–2:50 PM Empowering High Performance Life Sciences Workloads on Azure Qumulo 3:00–3:20 PM Transforming Science and Engineering — Driven by Agentic AI, Powered by HPC Joe Tostenrude 4:00–4:20 PM Unleashing AMD EPYC on Azure: Scalable HPC for Energy and Manufacturing Varun Selvaraj (AMD) 4:30–4:50 PM Automating HPC Workflows with Copilot Agents Xavier Pillons 5:00–5:20 PM Scaling the Future: NVIDIA’s GB300 NVL72 Rack for Next‑Generation AI Inference Kirthi Devleker (NVIDIA) 5:30–5:50 PM Enabling AI and HPC Workloads in the Cloud with Azure NetApp Files Andy Chan Wednesday, Nov 19 — 10:00 AM–6:00 PM Time (CST) Session Title Presenter(s) 10:30–10:50 AM AI‑Powered Digital Twins for Industrial Engineering John Linford (NVIDIA) 11:00–11:20 AM Advancing 5 Generations of HPC Innovation with AMD on Azure Allen Leibovitch (AMD) 11:30–11:50 AM Intro to LoRA Fine‑Tuning on Azure Christin Pohl 12:00–12:20 PM VAST + Microsoft: Building the Foundation for Agentic AI Lior Genzel (VAST Data) 12:30–12:50 PM Inside the World’s Most Powerful AI Data Center Chris Jones 1:00–1:20 PM Supervised GenAI Simulation – Stroke Prognosis (NVads V710 v5) Kurt Niebuhr 1:30–1:50 PM What You Don’t See: How Azure Defines VM Families Anshul Jain 2:00–2:20 PM Hammerspace Tier 0: Unleashing GPU Storage Performance on Azure Raj Sharma (Hammerspace) 2:30–2:50 PM GM Motorsports: Accelerating Race Performance with AI Physics on Rescale Bernardo Mendez (Rescale) 3:00–3:20 PM Hurricane Analysis and Forecasting on the Azure Cloud Salar Adili (Microsoft); Unni Kirandumkara (GDIT); Stefan Gary (Parallel Works) 3:30–3:50 PM Performance at Scale: Accelerating HPC & AI Workloads with WEKA on Azure Desiree Campbell; Wolfgang De Salvador 4:00–4:20 PM Pushing the Limits of Performance: Supercomputing on Azure AI Infrastructure Biju Thankachen; Ojasvi Bhalerao 4:30–4:50 PM Accelerating Momentum: Powering AI & HPC with AMD Instinct™ GPUs Jay Cayton (AMD) Thursday, Nov 20 — 10:00 AM–3:00 PM Time (CST) Session Title Presenter(s) 11:30–11:50 AM Intro to LoRA Fine‑Tuning on Azure Christin Pohl 12:00–12:20 PM Accelerating HPC Workflows with Ansys Access on Microsoft Azure Dr. John Baker (Ansys) 12:30–12:50 PM Accelerate Memory Bandwidth‑Bound Workloads with Azure HBv5, now GA Jyothi Venkatesh 1:00–1:20 PM Pushing the Limits: Supercomputing on Azure AI Infrastructure Biju Thankachen; Ojasvi Bhalerao 1:30–1:50 PM The High Performance Software Foundation Todd Gamblin (HPSF) 2:00–2:20 PM Heidi AI — Deploying Azure Cloud Environments for Higher‑Ed Students & Researchers James Verona (Adaptive Computing); Dr. Sameer Shende (UO/ParaTools) Partner Session Schedule Tuesday, Nov 18 Date Time (CST) Title Microsoft Contributor(s) Location Nov 18 11:00 AM–11:50 AM Cloud Computing for Engineering Simulation Joe Greenseid Ansys Booth Nov 18 1:00 PM–1:30 PM Revolutionizing Simulation with Artificial Intelligence Joe Tostenrude Ansys Booth Nov 18 4:30 PM–5:00 PM [HBv5] Jyothi Venkatesh AMD Booth Wednesday, Nov 19 Date Time (CST) Title Microsoft Contributor(s) Location Nov 19 11:30 AM–1:30 PM Accelerating Discovery: How HPC and AI Are Shaping the Future of Science (Lunch Panel) Andrew Jones (Moderator); Joe Greenseid (Panelist) Ruth's Chris Steak House Nov 19 1:00 PM–1:30 PM VAST and Microsoft Kanchan Mehrotra VAST Booth Demo Pods at Microsoft Booth Azure HPC & AI Infrastructure Explore how Azure delivers high-performance computing and AI workloads at scale. Learn about VM families, networking, and storage optimized for HPC. Agentic AI for Science See how autonomous agents accelerate scientific workflows, from simulation to analysis, using Azure AI and HPC resources. Hybrid HPC with Azure Arc Discover how Azure Arc enables hybrid HPC environments, integrating on-prem clusters with cloud resources for flexibility and scale. Ancillary Events (RSVP Required) Microsoft + AMD Networking Reception — Tuesday Night When: Tue, Nov 18, 6:30–10:00 PM (CST) Where: UMB Champions Club, Busch Stadium RSVP: https://aka.ms/MicrosoftAMD-Mixer Microsoft + NVIDIA Panel Luncheon — Wednesday When: Wed, Nov 19, 11:30 AM–1:30 PM (CST) Where: Ruth’s Chris Steak House Topic: Accelerating Discovery: How AI and HPC Are Shaping the Future of Science Panelists: Dan Ernst (NVIDIA); Rollin Thomas (NERSC); Joe Greenseid (Microsoft); Antonia Maar (Intersect360 Research); Fernanda Foertter (University of Alabama) RSVP: Luncheon is now closed as the event is fully booked. Conclusion We’re excited to connect with you at SC25! Whether you’re exploring our booth demos, attending technical sessions, or joining one of our partner events, this is your opportunity to experience how Microsoft is driving innovation in HPC and AI. Stop by Booth #1627 to see the Alpine F1 showcar, explore the Silicon Wall featuring AMD, NVIDIA, and Microsoft’s own chips, and enjoy a coffee from our barista while networking with experts. Don’t forget to RSVP for our Microsoft + AMD Network Reception and Microsoft + NVIDIA Panel Luncheon See you in St. Louis!Explore HPC & AI Innovation: Microsoft + AMD at HPC Roundtable 2025
The HPC Roundtable 2025 in Turin brings together industry leaders, engineers, and technologists to explore the future of high-performance computing (HPC) and artificial intelligence (AI) infrastructure. Hosted by DoITNow, the event features Microsoft and AMD as key participants, with sessions highlighting real-world innovations such as Polestar’s adoption of Microsoft Azure HPC for Computer-Aided Engineering (CAE). Attendees will gain insights into cloud-native HPC, hybrid compute environments, and the convergence of simulation and machine learning. The roundtable offers networking opportunities, strategic discussions, and showcases how Microsoft Azure and AMD are accelerating engineering innovation and intelligent workloads in automotive and other industries.Accelerating Genomic Pipelines with Nextflow on Microsoft Azure at the Nextflow Summit Boston 2025
Nextflow is one of the most widely adopted open-source workflow orchestrators in the scientific research domain. In genomics, a pipeline refers to a series of computational steps designed to analyze and interpret large-scale genomic data. These pipelines consist of multiple stages, each leveraging specialized bioinformatics tools and software to transform raw genomic inputs into meaningful biological insights. Running these pipelines efficiently requires significant computational power, parallel processing capabilities, and scalability—needs that are met by using a robust workflow orchestrator like Nextflow. This is precisely where Microsoft Azure HPC + AI services deliver strong value for scientists and bioinformaticians. With deep integrations between Nextflow and Azure cloud services, customers can build cloud-native infrastructures capable of handling multiple petabytes of genomic data from diverse sources. These integrations enable elastic pipeline scaling and seamless execution in hybrid or fully cloud-based environments. The Nextflow Summit Boston 2025 is an exciting gathering of scientists, bioinformaticians, data scientists, developers, and industry leaders from around the globe. It serves as a platform to shape the future of biotech research and development through community-driven innovation in data science and computational biology—all centered around a shared interest in Nextflow. This year, Microsoft is honored to participate in the summit with both a speaker session and a poster presentation, highlighting how Nextflow can run seamlessly in hybrid and fully cloud-native environments on Microsoft Azure. Join us on Thursday, May 15 at 4:00 PM for the session titled: “Running Nextflow on Microsoft Azure: Selecting Executors and Infrastructure Components” Presented by Wolfgang De Salvador, Venkat Malladi, and Jer-Ming Chia, this session offers a comprehensive overview of running Nextflow workflows in the cloud. It will explore the range of executor options available—including traditional schedulers, batch-managed services, and Kubernetes—as well as storage solutions spanning from file-based to object-based systems. This is a must-attend session for anyone looking to optimize infrastructure and boost performance for their Nextflow workflows on Azure. To register for Nextflow Summit Boston 2025 and access recordings after the event, visit the official Nextflow Summit website.Unpacking the Performance of Microsoft Azure ND GB200 v6 Virtual Machines
For a comprehensive understanding of our benchmarking methodologies and detailed performance results, please refer to our benchmarking guide available on the official Azure GitHub repository: Azure AI Benchmarking Guide. Breakdown of Benchmark Tests GEMM Performance General Matrix Multiply (GEMM) operations form the backbone of AI models. We measured that more than 60% of the time spent inferencing or training an AI model is spent doing matrix multiplication. Thus, measuring their speed is key to understand the performance of a GPU-based virtual machine. Azure benchmark assesses matrix-to-matrix multiplication efficiency using NVIDIA’s CuBLASLt library with FP8 precision, ensuring results reflect enterprise AI workloads. We measured the peak theoretical performance of the NVIDIA GB200 Blackwell GPU to be 4,856 TFLOPS, representing a 2.45x increase in performance compared to peak theoretical 1,979 TFLOPS on the NVIDIA H100 GPU. This finding is in-line with NVIDIA’s announcement of a 2.5x performance increase at GTC 2024. The true performance gain of the NVIDIA GB200 Blackwell GPU over its predecessors emerges in real-life conditions. For example, using 10,000 warm-up iterations and randomly initialized matrices demonstrated a sustained 2,744 TFLOPS for FP8 workloads, which, while expectedly lower than the theoretical peak, is still double that of the H100. The impact of these improvements on real workloads indicates up to a 3x speedup on average for end-to-end training and inference workloads based on our early results. High-Bandwidth Memory (HBM) Bandwidth Memory bandwidth is the metric that governs data movements. Our benchmarks showed a peak memory bandwidth of 7.35 TB/s, achieving 92% of its theoretical peak of 7.9 TB/s. This efficiency mirrors that of the H100, which also operated close to its theoretical maximum, while reaching 2.5x faster data transfers. This speedup ensures that data-intensive tasks, such as training large-scale neural networks, are executed efficiently. NVBandwidth The ND GB200 v6 architecture significantly enhances AI workload performance with NVLink C2C, enabling a direct, high-speed connection between the GPU and host system. This design reduces latency and improves data transfer efficiency, making AI workloads faster and more scalable. Our NVBandwidth tests measured CPU-to-GPU and GPU-to-CPU transfer rates to be nearly 4× faster than the ND H100 v5. This improvement minimizes bottlenecks in data-intensive applications and optimizes data movement efficiency over previous GPU-powered virtual machines. In addition, it allows the GPU to readily access additional memory when needed, which can be quickly accessed via the C2C link. NCCL Bandwidth NVIDIA’s Collective Communications Library (NCCL) enables high-speed communication between GPUs within and across nodes. We built our tests to measure the speed of communication between GPUs over NVLink within one virtual machine. Hig-speed communication is instrumental as most enterprise workloads consist of large-scale distributed models. The ND GB200 v6’s NVLink achieved a bandwidth of approximately 680 GB/s, aligning with NVIDIA’s projections. Conclusion The ND GB200 v6 virtual machine, powered by the NVIDIA GB200 Blackwell GPUs, showcases substantial advancements in computational performance, memory bandwidth, and data transfer speeds compared to the previous generations of virtual machines. These improvements are pivotal for efficiently managing the increasing demands of AI workloads like generative and agentic use-cases. Following our Benchmarking Guide will provide early access to performance reviews of the innovations announced at GTC 2025, helping customers drive the next wave of AI on Azure’s purpose-built AI infrastructure.