The post NVIDIA Enhances AI Scalability with NIM Operator 3.0.0 Release appeared on BitcoinEthereumNews.com. Darius Baruo Sep 10, 2025 17:33 NVIDIA’s NIM Operator 3.0.0 introduces advanced features for scalable AI inference, enhancing Kubernetes deployments with multi-LLM and multi-node capabilities, and efficient GPU utilization. NVIDIA has unveiled the latest iteration of its NIM Operator, version 3.0.0, aimed at bolstering the scalability and efficiency of AI inference deployments. This release, as detailed in a recent NVIDIA blog post, introduces a suite of enhancements designed to optimize the deployment and management of AI inference pipelines within Kubernetes environments. Advanced Deployment Capabilities The NIM Operator 3.0.0 facilitates the deployment of NVIDIA NIM microservices, which cater to the latest large language models (LLMs) and multimodal AI models. These include applications across reasoning, retrieval, vision, and speech domains. The update supports multi-LLM compatibility, allowing the deployment of diverse models with custom weights from various sources, and multi-node capabilities, addressing the challenges of deploying massive LLMs across multiple GPUs and nodes. Collaboration with Red Hat An important facet of this release is NVIDIA’s collaboration with Red Hat, which has enhanced the NIM Operator’s deployment on KServe. This integration leverages KServe lifecycle management, simplifying scalable NIM deployments and offering features such as model caching and NeMo Guardrails, which are essential for building trusted AI systems. Efficient GPU Utilization The release also marks the introduction of Kubernetes’ Dynamic Resource Allocation (DRA) to the NIM Operator. DRA simplifies GPU management by allowing users to define GPU device classes and request resources based on specific workload requirements. This feature, although currently under technology preview, promises full GPU and MIG usage, as well as GPU sharing through time slicing. Seamless Integration with KServe NVIDIA’s NIM Operator 3.0.0 supports both raw and serverless deployments on KServe, enhancing inference service management through intelligent caching and NeMo microservices support. This integration… The post NVIDIA Enhances AI Scalability with NIM Operator 3.0.0 Release appeared on BitcoinEthereumNews.com. Darius Baruo Sep 10, 2025 17:33 NVIDIA’s NIM Operator 3.0.0 introduces advanced features for scalable AI inference, enhancing Kubernetes deployments with multi-LLM and multi-node capabilities, and efficient GPU utilization. NVIDIA has unveiled the latest iteration of its NIM Operator, version 3.0.0, aimed at bolstering the scalability and efficiency of AI inference deployments. This release, as detailed in a recent NVIDIA blog post, introduces a suite of enhancements designed to optimize the deployment and management of AI inference pipelines within Kubernetes environments. Advanced Deployment Capabilities The NIM Operator 3.0.0 facilitates the deployment of NVIDIA NIM microservices, which cater to the latest large language models (LLMs) and multimodal AI models. These include applications across reasoning, retrieval, vision, and speech domains. The update supports multi-LLM compatibility, allowing the deployment of diverse models with custom weights from various sources, and multi-node capabilities, addressing the challenges of deploying massive LLMs across multiple GPUs and nodes. Collaboration with Red Hat An important facet of this release is NVIDIA’s collaboration with Red Hat, which has enhanced the NIM Operator’s deployment on KServe. This integration leverages KServe lifecycle management, simplifying scalable NIM deployments and offering features such as model caching and NeMo Guardrails, which are essential for building trusted AI systems. Efficient GPU Utilization The release also marks the introduction of Kubernetes’ Dynamic Resource Allocation (DRA) to the NIM Operator. DRA simplifies GPU management by allowing users to define GPU device classes and request resources based on specific workload requirements. This feature, although currently under technology preview, promises full GPU and MIG usage, as well as GPU sharing through time slicing. Seamless Integration with KServe NVIDIA’s NIM Operator 3.0.0 supports both raw and serverless deployments on KServe, enhancing inference service management through intelligent caching and NeMo microservices support. This integration…

NVIDIA Enhances AI Scalability with NIM Operator 3.0.0 Release

2025/09/11 14:46


Darius Baruo
Sep 10, 2025 17:33

NVIDIA’s NIM Operator 3.0.0 introduces advanced features for scalable AI inference, enhancing Kubernetes deployments with multi-LLM and multi-node capabilities, and efficient GPU utilization.





NVIDIA has unveiled the latest iteration of its NIM Operator, version 3.0.0, aimed at bolstering the scalability and efficiency of AI inference deployments. This release, as detailed in a recent NVIDIA blog post, introduces a suite of enhancements designed to optimize the deployment and management of AI inference pipelines within Kubernetes environments.

Advanced Deployment Capabilities

The NIM Operator 3.0.0 facilitates the deployment of NVIDIA NIM microservices, which cater to the latest large language models (LLMs) and multimodal AI models. These include applications across reasoning, retrieval, vision, and speech domains. The update supports multi-LLM compatibility, allowing the deployment of diverse models with custom weights from various sources, and multi-node capabilities, addressing the challenges of deploying massive LLMs across multiple GPUs and nodes.

Collaboration with Red Hat

An important facet of this release is NVIDIA’s collaboration with Red Hat, which has enhanced the NIM Operator’s deployment on KServe. This integration leverages KServe lifecycle management, simplifying scalable NIM deployments and offering features such as model caching and NeMo Guardrails, which are essential for building trusted AI systems.

Efficient GPU Utilization

The release also marks the introduction of Kubernetes’ Dynamic Resource Allocation (DRA) to the NIM Operator. DRA simplifies GPU management by allowing users to define GPU device classes and request resources based on specific workload requirements. This feature, although currently under technology preview, promises full GPU and MIG usage, as well as GPU sharing through time slicing.

Seamless Integration with KServe

NVIDIA’s NIM Operator 3.0.0 supports both raw and serverless deployments on KServe, enhancing inference service management through intelligent caching and NeMo microservices support. This integration aims to reduce inference time and autoscaling latency, thereby facilitating faster and more responsive AI deployments.

Overall, the NIM Operator 3.0.0 is a significant step forward in NVIDIA’s efforts to streamline AI workflows. By automating deployment, scaling, and lifecycle management, the operator enables enterprise teams to more easily adopt and scale AI applications, aligning with NVIDIA’s broader AI Enterprise initiatives.

Image source: Shutterstock


Source: https://blockchain.news/news/nvidia-enhances-ai-scalability-nim-operator-3-0-0

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.
Share Insights

You May Also Like

Ethereum Foundation Moves Entire $650M+ Treasury to Safe Multisig

Ethereum Foundation Moves Entire $650M+ Treasury to Safe Multisig

The post Ethereum Foundation Moves Entire $650M+ Treasury to Safe Multisig appeared on BitcoinEthereumNews.com. EF completes full treasury migration to Safe smart accounts, joining Vitalik Buterin as key Safe user + Safe smart accounts cross 750M transactions milestone.   The Ethereum Foundation has completed the migration of its full treasury, over 160,000 ETH worth approximately $650 million to Safe{Wallet}, following months of successful DeFi testing. Safe{Wallet}, operated by Safe Labs (a fully owned subsidiary of the Safe Foundation), is the crypto industry’s trusted smart account standard for multisig wallets, securing billions of dollars in assets for institutions, DAOs, and projects. The move follows the Foundation’s June 2025 treasury policy announcement, which committed to actively participating in Ethereum’s DeFi ecosystem. Since February, the EF had been testing Safe with a separate DeFi-focused account, dogfooding protocols including Aave, Cowswap, and Morpho as part of their strategy to support applications built on Ethereum. After testing a 3-of-5 multisig configuration on January 20th, the Foundation has now consolidated its remaining ETH holdings into Safe, completing the transition from their previous custom-built multisig solution. This implementation enables the Ethereum Foundation to actively participate in DeFi via Safe while maintaining battle-tested security standards, marking another step toward Safe’s vision of moving the world’s GDP onchain through battle-tested self-custody infrastructure. “Safe has proven safe and has a great user experience, and we will transfer more of our funds here over time,” the Ethereum Foundation announced, indicating this is the beginning of a deeper commitment to the Safe smart account standard. Safe’s Momentum The timing is notable: Safe has just crossed 750 million transactions (751,062,286 as of today) with over 57.5 million Safes created across multiple chains. The protocol has emerged as crypto’s de facto standard for multisig wallets, securing billions in institutional and DAO treasuries. Safe also counts Ethereum co-founder Vitalik Buterin among its prominent users, who revealed in May 2024 that…
Share
2025/10/23 04:15
Share
Citadel’s Stake in Solana Treasury Firm DeFi Dev Corp Highlights Potential Crypto Exposure

Citadel’s Stake in Solana Treasury Firm DeFi Dev Corp Highlights Potential Crypto Exposure

The post Citadel’s Stake in Solana Treasury Firm DeFi Dev Corp Highlights Potential Crypto Exposure appeared on BitcoinEthereumNews.com. COINOTAG recommends • Exchange signup 💹 Trade with pro tools Fast execution, robust charts, clean risk controls. 👉 Open account → COINOTAG recommends • Exchange signup 🚀 Smooth orders, clear control Advanced order types and market depth in one view. 👉 Create account → COINOTAG recommends • Exchange signup 📈 Clarity in volatile markets Plan entries & exits, manage positions with discipline. 👉 Sign up → COINOTAG recommends • Exchange signup ⚡ Speed, depth, reliability Execute confidently when timing matters. 👉 Open account → COINOTAG recommends • Exchange signup 🧭 A focused workflow for traders Alerts, watchlists, and a repeatable process. 👉 Get started → COINOTAG recommends • Exchange signup ✅ Data‑driven decisions Focus on process—not noise. 👉 Sign up → Citadel’s investment in DeFi Dev Corp represents a 4.5% stake held by the firm, with CEO Ken Griffin owning another 4.5%, totaling significant exposure to Solana treasury operations through this leading DAT company. Citadel and subsidiaries control over 9% of DeFi Dev Corp shares, highlighting hedge fund interest in Solana-based treasuries. DeFi Dev Corp has increased its SOL per share by 375% since initial acquisitions. Solana treasuries now hold 20.31 million SOL, with 9 million staked for an average 7.7% yield. Discover Citadel’s 4.5% stake in DeFi Dev Corp and its impact on Solana treasuries. Explore SOL holdings growth and market insights for informed crypto investment decisions today. What is Citadel’s Stake in DeFi Dev Corp? Citadel’s investment in DeFi Dev Corp includes a 4.5% ownership through the firm itself, complemented by an additional 4.5% held directly by CEO Ken Griffin. This positions Citadel among the top shareholders in the Solana-focused treasury company. Various Citadel subsidiaries, such as Citadel Advisors LLC and Citadel Securities LLC, contribute further stakes totaling around 6%, as detailed in a recent ownership report. COINOTAG…
Share
2025/10/23 03:57
Share