NVIDIA NIM Revolutionizes Model Deployment, Now Available to Turn Millions of Developers Worldwide into Generative AI Developers

Applications of AI


  • Over 150 partners across all tiers of the AI ​​ecosystem are incorporating NIM inference microservices to accelerate deployment of enterprise AI applications from weeks to minutes
  • Members of the NVIDIA Developer Program have free access to NIM for research, development and testing.

COMPUTEX—NVIDIA announced today that it is now available for download to 28 million developers worldwide. NVIDIA NIM™ (an inference microservice that delivers models as optimized containers) deployed in the cloud, datacenter, or on a workstation, makes it easy to build generative AI applications like copilots and chatbots in minutes instead of weeks.

These new generative AI applications are becoming increasingly complex and often utilize multiple models with different capabilities for generating text, images, video, voice, etc. NVIDIA NIMs dramatically improve developer productivity by providing a simple and standardized way to add generative AI to applications.

NIM also allows businesses to maximize their infrastructure investments. For example, running Meta Llama 3-8B on NIM will generate up to 3x more generated AI tokens on faster infrastructure than without NIM. This allows businesses to increase efficiency and generate more responses using the same amount of computing infrastructure.

Approximately 200 technology partners, including Cadence Cloudra, Cohesiveness, DataStacks, NetAppScale AI and overview — is integrating NIM into its platform to speed up the adoption of generative AI for domain-specific applications such as co-pilots, code assistants and digital human avatars. Hugging Face NIM now available Metalama 3.

“Every business wants to incorporate generative AI into their operations, but not every company has a dedicated team of AI researchers,” said Jensen Huang, founder and CEO, NVIDIA. “Integrated into any platform, accessible to any developer and able to run anywhere, NVIDIA NIM is helping the technology industry bring generative AI to every organization.”

Through NIM, companies can deploy AI applications into production. NVIDIA AI Enterprise Software platform. Starting next month, NVIDIA Developer Program You can access NIM for free to conduct research, development, and testing on your preferred infrastructure.

Age 40 or older NIM Microservices Power AI Model Generation in Various Modalities
NIM containers are pre-built to speed up model deployment for GPU-accelerated inference. NVIDIA CUDA® software, NVIDIA Triton Inference Server™ and NVIDIA TensorRT™-LLM software.

Over 40 NVIDIA and community models are available as NIM endpoints. source:include Databricks DBRXGoogle's open model Gemma, Meta Llama 3, Microsoft Phi-3, Mistral Large, Mixtral 8x22B, and Snowflake Arctic.

Developers can now use NVIDIA NIM microservices for their Meta Llama 3 models. Hugging Face An AI platform that makes it easy for developers to access and run Llama 3 NIM with just a few clicks using NVIDIA GPU-powered Hugging Facial Inference Endpoints on the cloud of their choice.

Companies can use NIM to run applications that generate text, images, video, voice and digital humans. NVIDIA BioNemo™ NIM Microservices for Digital Biology enables researchers to build novel protein structures to accelerate drug discovery.

Dozens of healthcare companies Introducing NIM It powers generative AI inference across a range of applications including surgical planning, digital assistants, drug discovery, and clinical trial optimization.

new NVIDIA ACE NIM MicroservicesDevelopers can easily build and interact with interactive, lifelike digital humans in applications such as customer service, telemedicine, education, gaming, and entertainment.

hundreds AI Ecosystemem Partner NIM integration
Platform providers include: Orthodox, Red Hat, Nutanix and VMware (acquired by Broadcom) supports NIM Open source KServe Or enterprise solutions. AI application companies Hippocrates’ AI, Greening, Kinetica and Ladies We are also introducing NIM to power generative AI inference.

Leading AI tools and MLOps partners (Amazon SageMaker, Microsoft Azure AI, Dataiku, DataRobot, etc.) Deeply upsetDomino Data Lab, Run Chain, Llama Index, DuplicateRun.ai, Saturn CloudsSecuriti AI and Weights and Biases — is incorporating NIM into its platform to enable developers to build and deploy domain-specific generative AI applications with optimized inference.

Global system integrators and service delivery partners Accenture, Deloitte, Infosys, Latent View, QuantifySoftServe, TCS and Wipro have built the NIM competency to help enterprises globally rapidly develop and deploy AI strategies.

Businesses can run NIM-enabled applications anywhere, including in virtual environments. NVIDIA-Certified Systems™ Cisco, a global infrastructure manufacturer; Dell Technologies, Hewlett Packard Enterprise, Lenovo Supermicro, server manufacturers, etc. ASRock Rack, Asus, Gigabytes, Inglasys, Inventec, PegatronQCT, Wistron, Wiwynn. NIM microservices are also integrated. Amazon Web Services, Google Cloud, Azure and Oracle Cloud Infrastructure.

Industry giants strengthen the generationNative AI with NIM
Industry leaders Foxconn, Pegatron, AmdocsLowe's, Service Now Siemens is among some of the companies using NIM for generative AI applications in manufacturing, healthcare, financial services, retail, customer service and more.

  • Foxconn The world's largest electronics manufacturer is using NIM to develop domain-specific LLMs embedded into various internal systems and processes for AI factories for smart manufacturing, smart cities, and smart electric vehicles.
  • Pegatron The Taiwanese electronics manufacturer is leveraging NIM for Project TaME, a Taiwanese specialist blended model designed to foster the development of local LLMs for the industry.
  • Amdocs A leading global provider of software and services to communications and media companies. Using NIM Perform customer billing LLM with significantly reduced token costs, improved accuracy by up to 30%, reduced latency by 80% and near real-time response.
  • Lowe's Big Money® The 50th largest home improvement company is using generative AI for a variety of use cases. For example, the retailer is leveraging NVIDIA NIM inference microservices to improve employee and customer experiences.
  • Service Now ServiceNow, an AI platform for business transformation, announced earlier this year that it was one of the first platform providers to access NIM to enable fast, scalable and cost-effective development and deployment of LLM for its customers. The NIM microservices are integrated into the Now AI multimodal model and are available to customers who have installed ServiceNow's generative AI experience, Now Assist.
  • Siemens The global technology company, focused on industry, infrastructure, transportation, and healthcare, is integrating its operational technology with NIM microservices to power AI workloads in the field, and is also using NIM to build an on-premise version of Industrial Copilot for machine operators.

availability
Developers can try out NVIDIA's microservices. source: It's free to use and enables enterprises to deploy production-grade NIM microservices. NVIDIA AI Enterprise It runs on NVIDIA-certified systems and major cloud platforms. Starting next month, NVIDIA Developer Program You will have free access to NIM for research and testing.

clock Huang's COMPUTEX keynote speech Learn more about NVIDIA NIM.



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *