hybrid
11 TopicsUnlocking Enterprise AI: SQL Server 2025 and NVIDIA Nemotron RAG Accelerate AI
Today, most of the world’s data still remains untapped, sitting in databases, documents, and systems across organizations. Enterprises are racing to unlock this data’s value by building the next wave of generative AI applications—solutions that can answer questions, summarize documents, and drive smarter decisions. At the heart of these innovations are retrieval-augmented generation (RAG) pipelines, which enable users to interactively engage with large amount of data that continuously evolves. Yet, as promising as RAG pipelines are, enterprises face real challenges in making them work at scale. Handling both structured and unstructured data, processing massive volumes efficiently, and ensuring privacy and security are just a few hurdles. This is where the integration between SQL Server 2025 and NVIDIA Nemotron RAG models, deployed as NVIDIA NIM microservices, comes in, offering a new approach that streamlines AI deployment and delivers enterprise-grade performance—whether you’re running workloads in the cloud or on-premises. “As AI becomes core to every enterprise, organizations need efficient and compliant ways to bring intelligence to their data,” said Joey Conway, Senior Director of Generative AI software at NVIDIA. “With SQL Server 2025’s built-in AI and NVIDIA Nemotron RAG, deployed as NIM microservices, enterprises can deploy and run AI models close to their data on premises or in the cloud without complex integration, accelerating innovation while maintaining data sovereignty and control.” Overcoming the complexity of generating embeddings at scale Customer challenge Building responsive AI applications using RAG requires converting SQL data into vector embeddings—a process that feeds huge amounts of text through complex neural networks. This is inherently parallel and compute-intensive, often creating performance bottlenecks that prevent real-time data indexing. The result? Slow applications and poor user experiences. Moreover, enterprises need flexibility. Different embedding models excel at different tasks—semantic search, recommendations, classification—and each comes with its own tradeoffs in accuracy, speed, and cost. Businesses want to mix and match models, balance premium performance with budget constraints, and stay resilient against model deprecation or API changes. Furthermore, rapid experimentation and adaptation are key to staying ahead and thus developers want models that offer flexible customization and full transparency. The Solution: SQL Server 2025 + NVIDIA Nemotron RAG SQL Server 2025 brings AI closer to your data, allowing you to natively and securely connect to any model hosted anywhere. You can generate embeddings directly in SQL using extensions to T-SQL —no need for new languages, frameworks, or third-party tools. By connecting SQL Server 2025 to the llama-nemotron-embed-1b-v2 embedding model from NVIDIA, you eliminate bottlenecks and deliver the massive throughput needed for real-time embedding generation. llama-nemotron-embed-1b-v2 is a best in class embedding model that offers multilingual and cross-lingual text question-answering retrieval with long context support and optimized data storage. This model is part of NVIDIA Nemotron RAG models, a collection of extraction, embedding, reranking models, fine-tuned with the Nemotron RAG datasets and scripts, to achieve the best accuracy. These models offer flexible customization, enabling easy fine-tuning and rapid experimentation. They also offer full transparency with open access to models, datasets, and scripts. Llama-nemotron-embed-1b-v2 is the model of choice for embedding workflows, but this high-speed inference pipeline is not limited to this model and can potentially call any optimized AI model as an NVIDIA NIM microservice, seamlessly powering every stage of the RAG pipeline. From multimodal data ingestion and advanced retrieval to reranking, all operations run directly on your data within SQL Server. Such RAG systems can be applied across a wide range of use cases, enabling intelligent, context-aware applications across industries. Customer Benefits: With GPU acceleration and built-in AI of SQL Server 2025, you can achieve optimal inference, ensuring performance that meets the demands of modern applications. Our flexible approach lets you mix and match models to suit different use cases, striking the right balance between accuracy and cost. And with open models that enable vendor flexibility and rapid adaptation, you gain resilience to stay ahead of the curve in an ever-changing AI landscape. Streamlining AI Model Deployment with Enterprise-Grade Confidence Customer Challenge Integrating advanced AI models into enterprise workflows has historically been slow and complex. Specialized teams must manage intricate software dependencies, configure infrastructure, and handle ongoing maintenance—all while navigating the risks of deploying unsupported models in mission-critical environments. This complexity slows innovation, drains engineering resources, and increases risk. The Solution: Simplified, Secure Model Deployment with NVIDIA NIM This collaboration simplifies and de-risks AI deployment. The llama-nemotron-embed-1b-v2 model is available as an NVIDIA NIM microservice for secure, reliable deployment across multiple Azure compute platforms. Prebuilt NIM containers for a broad spectrum of AI models and can be deployed with a single command for easy integration into enterprise-grade AI applications using built-in REST APIs of SQL Server 2025 and just a few lines of code, regardless where you run SQL Server workloads and NVIDIA NIM, on premises or in the cloud. NIM containers package the latest AI models together with the best inference technology from NVIDIA and the community and all dependencies into a ready-to-run container, abstracting away the complexity of environment setup so customers can spin up AI services quickly. Furthermore, NVIDIA NIM is enterprise-grade and is continuously managed by NVIDIA with dedicated software branches, rigorous validation processes, and support. As a result, developers can confidently integrate state-of-the-art AI into their data applications. This streamlined approach significantly reduces development overhead and provides the reliability needed for mission-critical enterprise systems. NVIDIA NIM containers are discoverable and deployable via Microsoft Azure AI Foundry’s model catalog. Customer Benefits Rapid deployment with minimal setup means you can start leveraging AI without specialized engineering, and SQL Server 2025 makes it even easier with built-in support for AI workloads and native REST APIs. Enterprise-grade security and monitoring ensure safe, reliable operations, while SQL Server’s integration with Entra ID and advanced compliance features provide added protection. Direct integration into SQL workflows reduces complexity and risk, and with SQL Server’s hybrid flexibility, you can run seamlessly across on-premises and cloud environments—simplifying modernization while maintaining control. Innovating Without Compromise on Security or Flexibility Customer Challenge Organizations in regulated industries often face a tough choice: adopt powerful AI or maintain strict data residency and compliance. Moving sensitive data to external services is often not an option, and many companies run AI inference workloads both in the cloud and on-premises to balance scalability, privacy, regulatory compliance, and low-latency requirements. The Solution: Flexible, Secure Integration—On-Premises and Cloud SQL Server 2025 enables organizations in regulated environments to securely integrate locally hosted AI models, ensuring data residency and compliance while minimizing network overhead. This architecture boosts throughput by keeping sensitive data on-premises and leveraging SQL Server’s native extensibility for direct model invocation. With SQL Server 2025 and Nemotron RAG, deployed as NVIDIA NIM microservices, you get the best of both worlds. This solution can be seamlessly deployed in the cloud with serverless NVIDIA GPUs on Azure Container Apps (ACA) or on-premises with NVIDIA GPUs on Azure Local. Sensitive data never leaves your secure environment, allowing you to harness the full power of Nemotron models while maintaining complete data sovereignty and meeting the strictest compliance mandates. Customer Benefits SQL Server 2025 helps you maintain compliance by supporting data residency and meeting regulatory standard requirements across regions. Sensitive data stays protected on-premises with enterprise-grade security, including consistent access controls, ledger support, and advanced encryption to minimize risk. At the same time, SQL Server’s hybrid flexibility lets you deploy AI workloads wherever they’re needed—on-premises, in the cloud, or across a hybrid environment—while leveraging built-in AI features like vector search and secure integration with locally hosted models for performance and control. Conclusion: Powering the Next Wave of Enterprise AI The collaboration between Microsoft and NVIDIA is more than a technical integration. It’s designed to help enterprises overcome the toughest challenges in AI deployment. By streamlining vector embedding and vector search, delivering enterprise-grade performance, and enabling secure, flexible integration across cloud and on-premises environments, this joint solution empowers organizations to unlock the full value of their data. Whether you’re building conversational AI, automating document analysis, or driving predictive insights, SQL Server 2025 and NVIDIA Nemotron RAG models, deployed as NIM, provide the tools you need to innovate with confidence. The future of enterprise AI is here and it’s flexible, secure, and built for real business impact. Get started today: Learn more about SQL Server 2025 and download it today Learn more about our joint solution from NVIDIA’s Technical Blog GitHub: Microsoft SQL Server 2025 and NVIDIA Nemotron RAG627Views1like0CommentsReimagining Data Excellence: SQL Server 2025 Accelerated by Pure Storage
SQL Server 2025 is a leap forward as enterprise AI-ready database, unifying analytics, modern AI application development, and mission-critical engine capabilities like security, high availability and performance from ground to cloud. Pure Storage’s all-Flash solutions are engineered to optimize SQL Server workloads, offering faster query performance, reduced latency, and simplified management. Together it helps customers accelerate the modernization of their data estate.505Views2likes1CommentMicrosoft at PASS Data Community Summit 2025
Microsoft is excited to be back as the Sapphire Sponsor for this year’s PASS Data Community Summit, together with AMD. It’ll be a week packed with keynotes, learning pathways, sessions, breakfast, giveaways and more! Our experts and engineers will be there and ready to share all things SQL Server 2025, cover the latest from Azure SQL, Microsoft Fabric and more—delivered on a foundation of AMD-powered innovation. Whether your goals are modernizing for performance and AI readiness or building intelligent apps and agents, we’ll have you covered. Join us to “Connect, Share and Learn” alongside the rest of your peers at the PASS community. CVP of Azure Databases, Shireesh Thota, is back as the keynote speaker, and will be joined by leaders across Microsoft Data including Priya Sathy and Bob Ward—highlighting how Microsoft and AMD are shaping the future of data together. Join us on Day 2, Thursday, November 20 th . Come back and check as new sessions are added: Learning Pathways: Becoming Azure SQL DBA – High Availability and BCDR Dr. Dani Ljepava, Bob Ward, John Morehouse Learn how to evolve your Azure SQL DBA skills in the domain of High Availability (HA), Business Continuity and Disaster Recovery (BCDR) from the perspective of on-premises DBA’s. Becoming Azure SQL DBA – Security, Compliance, Threats, Connectivity Pam Lahoud, Joey Dantoni Evolve your Azure SQL DBA skills in the domain of security, compliance, authentication and connectivity, from the perspective of an on-premises DBA now supporting databases in Azure. Becoming Azure SQL DBA – Performance Monitoring, Tuning, and Alerting Erin Stellato, Pam Lahoud, Monica Morehouse (Rathbun) Extend your Azure SQL DBA skills in the domain of performance monitoring, tuning, and alerting from the perspective of on-premises DBA. Becoming Azure SQL DBA – Copilot and AI Davide Mauri, Erin Stellato Unlock the future of data productivity with a hands-on exploration of AI-powered Copilots! We’ll dive into Copilot in SQL Server Management Studio (SSMS), Microsoft Copilot in Azure with NL2SQL and SQL, and Copilot for SQL Databases in Microsoft Fabric. Becoming an Azure SQL DBA - New Opportunities for DBAs in Azure Bob Ward, Dr. Dani Ljepava, Erin Stellato, Pam Lahoud Explore new skill development opportunities as an Azure SQL DBA. Many traditional SQL Server DBA tasks have been partially or fully delegated to Microsoft, and this shift in responsibilities provides a great opportunity to invest in developing new cloud skills that will help you excel as an Azure SQL DBA champion. General Sessions Breakfast with the Microsoft Data Leadership Team JUST ADDED! Shireesh Thota, Priya Sathy, Bob Ward, Erin Stellato Get your day started early at PASS Data Community Summit with a free breakfast and a Q&A session with a panel of leaders across Microsoft hosted by Bob Ward. Tell us what is top of mind for you across SQL Server, Azure SQL, Microsoft Fabric and topics like AI. This is always one of the most popular sessions at the PASS Data Community Summit, so you won’t want to miss it! Lunch and Learn with Bob Ward JUST ADDED! Join us for a lunch and learn with Bob Ward as he covers all things SQL and dives into his latest release: SQL Server 2025 Unveiled: The AI-Ready Enterprise Database with Microsoft Fabric Integration. Space will be extremely limited, bring your lunch and join this exclusive session! Inside SQL Server 2025 Bob Ward Join Bob Ward and friends to go deep into the next major release of SQL Server, SQL Server 2025, the Enterprise AI-ready database. You will learn the fundamentals and a deep dive of all the new capabilities in the release so you can plan and make key decisions on when and how to upgrade. SQL Database in Fabric: The Unified Database for AI Apps and Analytics JUST ADDED! Amar Digamber Patil, Yo-Lei Chen, Basudep Sarkar Discover how SQL in Fabric brings transactional and analytical workloads together in one cloud-native database. In this session, we’ll show how developers and data teams can simplify AI-driven application development with near real-time insights and built-in AI, seamless OneLake integration, and end-to-end analytics—all in a single, unified experience. Building Scalable, Secure AI Apps with Azure SQL Hyperscale Build AI apps that run securely and scale with your needs with Azure SQL Database Hyperscale. We’ll cover native vector indexes for semantic search, read scale‑out for low‑latency RAG, using the model of your choice, from T‑SQL. Ecosystem sessions Smarter GitHub Copilot + SSMS 22 Erin Stellato Discover how GitHub Copilot is transforming the way you write T-SQL and optimize your SQL databases inside SQL Server Management Studio (SSMS) 22. In this session, we’ll showcase the newest SSMS 22 features alongside real-world demos of GitHub Copilot, highlighting how AI assistance can speed up query writing, reduce errors, and boost productivity. Booth Experiences What: Book signing "SQL Server 2025 Unveiled" with Bob Ward When: Wednesday, November 19th evening reception Where: Microsoft and AMD booth (#404) Join us as we celebrate the release of "SQL Server 2025 Unveiled" at the PASS Summit evening reception! Limited to the first 200 attendees. Now on demand! Get a jump start on your PASS experience. "The AI Opportunity: From Ground to Cloud" webinar is now available on demand! Get an early preview of the sessions at PASS Summit and a closer look into SQL Server 2025. Hosted by Redgate’s Kellyn Gorman and Bob Ward, sign up and watch today! --Exclusive Offer for PASS Attendees-- As a special offer from Microsoft, use the code AZURE150 to receive $150 off your 3-day conference pass. Don’t miss this opportunity to connect, grow, and learn with the community. Register today487Views0likes0CommentsAzure Arc enabled SQL Server - determine your Azure SQL target with ease and confidence
Selecting the best Azure SQL target for your Azure Arc enabled SQL Server with confidence is now easier. All of this while you continue to manage secure and govern your SQL Server estate from Azure.2.3KViews2likes0CommentsSQL Server enabled by Azure Arc, now assists in selecting the best Azure SQL target
To make the SQL Server migration journey more efficient, the SQL Server enabled by Azure Arc can now help our customers assess the readiness of their SQL Server workloads for Azure SQL migration and assist them in choosing the most suitable Azure SQL configuration.
5.5KViews3likes1CommentArc enabled Azure migration & modernization journey
This blog explains how SQL Servers connected to Azure Arc can make the migration and modernization process to Azure SQL easier and faster. SQL Server enabled by Azure Arc allows SQL Server instances hosted outside of Azure to access Azure services. This results in operational efficiencies and cost reductions. Moreover, it also offers a simple migration path to Azure with minimal or no downtime.7.7KViews5likes0CommentsThe Bridge - How Azure Arc brings cloud innovation to SQL Server anywhere
The ability to effectively manage data is more critical than ever and increasingly complex — with organizations hosting data on-premises, at the edge, and across multiple clouds. With the increasing need for businesses to maximize existing resources, one thing is clear: Today’s organizations need a bridge to bring cloud innovation to data estate across all environments – a consistent and efficient experience for managing, governing and security data. That bridge is Microsoft’s Azure Arc.
28KViews1like0Comments