country_code

Entos Transforms Drug Discovery and Design With NVIDIA AI-Powered Molecular Simulation

NVIDIA Clara Discovery is powering the San Diego startup’s new machine learning approach for developing next-generation therapeutics.
by Abraham Stern
molecular simulation

More knowledge means more informed predictions. That’s the principle San Diego-based startup Entos is applying to revolutionize drug design with an AI-powered approach that enables a thousandfold acceleration in molecular properties prediction.

Drug discovery is a notoriously time-consuming and data-intensive process, but Entos’ OrbNet architecture changes that. It requires 30x less data to train a model for molecular drug discovery with quantum accuracy and 100x fewer experiments to find promising drug compounds — cutting through the waiting time and complexities associated with traditional therapeutic drug discovery methods.

The company, a member of the NVIDIA Inception program for startups revolutionizing industries with advancements in AI, data science and high performance computing, is advancing its work with NVIDIA Clara Discovery — a collection of state-of-the-art frameworks, applications and pretrained models built to unlock insights about how billions of potential drug molecules interact inside our bodies.

“Our physics-based approach means we include more qualities about the underlying quantum mechanics into the machine learning model,” said Tom Miller, CEO of Entos. “This enables us to make better predictions, even while using less data.”

Entos is focusing on identifying drug molecules that could deactivate proteins linked to certain forms of cancer. By including quantum mechanics calculations into its machine learning workflow, the startup can more quickly narrow the pool of potential compounds that bind with these target proteins.

Bringing AI Into the Drug Discovery Process

Machine learning is transforming the way scientists approach everything from climate science to drug discovery. Now the combination of AI and machine learning is leading to a new way of doing science, creating a hybrid of deep learning and physics-based simulation to transform the way drugs are discovered.

Drug discovery is a data-intensive process where researchers perform computationally dense calculations to simulate how molecules and proteins interact to identify the right therapeutics. Traditionally, these forms of quantum calculations are extremely expensive to perform and take weeks or months to complete.

These computational experiments benefit from the incorporation of AI and accelerated computing by allowing researchers to simulate the interaction of a drug with a protein at quantum accuracy. The simulations are far too computationally expensive to perform using traditional quantum mechanical calculations.

Entos optimizes its OrbNet drug discovery software on NVIDIA DGX A100 Tensor Core GPUs. The OrbNet AI model — developed jointly at Caltech by Entos CEO Tom Miller and Anima Anandkumar, director of machine learning research at NVIDIA — enables robotic synthesis and high-throughput experimentation to speed up therapeutic design.

“OrbNet uses a graph neural network built on domain-specific features that account for interactions between atoms. Further, we account for symmetries such as three-dimensional rotations,” said Anandkumar. “These design considerations make it possible to train OrbNet only on small molecules — those with less than 40 atoms — and directly apply the model on large protein molecules with a high degree of accuracy.”

By collaborating with NVIDIA experts, Miller’s team of scientists is able to perform high-throughput experimentation and “open new doors to what we can go after,” he said.

Unlocking the Potential of Covalent Bonds

Recent developments in machine learning are transforming the size and scale of chemical databases that researchers can trawl for promising drug compounds. AI models also allow scientists to study the innumerable chemical reactions of enzymes within the body in a new way.

Together, these advancements are enabling the study of entirely new classes of drugs that researchers were unable to investigate with previous methods.

One promising technique involves the creation of drug molecules that form covalent bonds with the target protein. If therapeutics can form these bonds with only the target protein, patients could be prescribed smaller doses and experience fewer side effects. Entos researchers plan to apply this method to disease areas including cancer, diabetes and cystic fibrosis.

Entos has forged partnerships with leaders in the pharmaceuticals, materials and chemical industries and raised $53 million in July to support its efforts to create meaningful therapies with high accuracy. The company credits active collaborations with the NVIDIA healthcare team as an asset in getting connected to technical resources and assistance in optimizing its applications on NVIDIA hardware architecture.

Hear from Entos and additional Inception startups at NVIDIA GTC, running online through Nov. 11. Tune in to a healthcare special address by Kimberly Powell, NVIDIA’s VP of healthcare.

Watch NVIDIA founder and CEO Jensen Huang’s GTC keynote address below. Subscribe to NVIDIA healthcare news.

New NVIDIA Nemotron 3 Super Delivers 5x Higher Throughput for Agentic AI

A new, open, 120-billion-parameter hybrid mixture-of-experts model optimized for NVIDIA Blackwell addresses the costs of long thinking and context explosion that slow autonomous agent workflows.
by Kari Briski

Launched today, NVIDIA Nemotron 3 Super is a 120‑billion‑parameter open model with 12 billion active parameters designed to run complex agentic AI systems at scale. 

Available now, the model combines advanced reasoning capabilities to efficiently complete tasks with high accuracy for autonomous agents.

AI-Native Companies: Perplexity offers its users access to Nemotron 3 Super for search and as one of 20 orchestrated models in Computer. Companies offering software development agents like CodeRabbit, Factory and Greptile are integrating the model into their AI agents along with proprietary models to achieve higher accuracy at lower cost. And life sciences and frontier AI organizations like Edison Scientific and Lila Sciences will power their agents for deep literature search, data science and molecular understanding.

Enterprise Software Platforms: Industry leaders such as Amdocs, Palantir, Cadence, Dassault Systèmes and Siemens are deploying and customizing the model to automate workflows in telecom, cybersecurity, semiconductor design and manufacturing. 

As companies move beyond chatbots and into multi‑agent applications, they encounter two constraints.

The first is context explosion. Multi‑agent workflows generate up to 15x more tokens than standard chat because each interaction requires resending full histories, including tool outputs and intermediate reasoning. 

Over long tasks, this volume of context increases costs and can lead to goal drift, where agents lose alignment with the original objective.

The second is the thinking tax. Complex agents must reason at every step, but using large models for every subtask makes multi-agent applications too expensive and sluggish for practical applications.

Nemotron 3 Super has a 1‑million‑token context window, allowing agents to retain full workflow state in memory and preventing goal drift.

Nemotron 3 Super has set new standards, claiming the top spot on Artificial Analysis for efficiency and openness with leading accuracy among models of the same size. 

The model also powers the NVIDIA AI-Q research agent to the No. 1 position on DeepResearch Bench and DeepResearch Bench II leaderboards, benchmarks that measure an AI system’s ability to conduct thorough, multistep research across large document sets while maintaining reasoning coherence. 

Hybrid Architecture

Nemotron 3 Super uses a hybrid mixture‑of‑experts (MoE) architecture that combines three major innovations to deliver up to 5x higher throughput and up to 2x higher accuracy than the previous Nemotron Super model. 

  • Hybrid Architecture: Mamba layers deliver 4x higher memory and compute efficiency, while transformer layers drive advanced reasoning.
  • MoE: Only 12 billion of its 120 billion parameters are active at inference. 
  • Latent MoE: A new technique that improves accuracy by activating four expert specialists for the cost of one to generate the next token at inference.
  • Multi-Token Prediction: Predicts multiple future words simultaneously, resulting in 3x faster inference.

On the NVIDIA Blackwell platform, the model runs in NVFP4 precision. That cuts memory requirements and pushes inference up to 4x faster than FP8 on NVIDIA Hopper, with no loss in accuracy. 

Open Weights, Data and Recipes

NVIDIA is releasing Nemotron 3 Super with open weights under a permissive license. Developers can deploy and customize it on workstations, in data centers or in the cloud.

The model was trained on synthetic data generated using frontier reasoning models. NVIDIA is publishing the complete methodology, including over 10 trillion tokens of pre- and post-training datasets, 15 training environments for reinforcement learning and evaluation recipes. Researchers can further use the NVIDIA NeMo platform to fine-tune the model or build their own. 

Use in Agentic Systems

Nemotron 3 Super is designed to handle complex subtasks inside a multi-agent system. 

A software development agent can load an entire codebase into context at once, enabling end-to-end code generation and debugging without document segmentation. 

In financial analysis it can load thousands of pages of reports into memory,  eliminating the need to re-reason across long conversations, which improves efficiency. 

Nemotron 3 Super has high-accuracy tool calling that ensures autonomous agents reliably navigate massive function libraries to prevent execution errors in high-stakes environments, like autonomous security orchestration in cybersecurity.

Availability

NVIDIA Nemotron 3 Super, part of the Nemotron 3 family, can be accessed at build.nvidia.com, Perplexity, OpenRouter and Hugging Face. Dell Technologies is bringing the model to the Dell Enterprise Hub on Hugging Face, optimized for on-premise deployment on the Dell AI Factory, advancing multi-agent AI workflows. HPE is also bringing NVIDIA Nemotron to its agents hub to help ensure scalable enterprise adoption of agentic AI. 

Enterprises and developers can deploy the model through several partners:

The model is packaged as an NVIDIA NIM microservice, allowing deployment from on-premises systems to the cloud.

Stay up to date on agentic AI, NVIDIA Nemotron and more by subscribing to NVIDIA AI news, joining the community, and following NVIDIA AI on LinkedIn, Instagram, X and Facebook.

Explore self-paced video tutorials and livestreams.

NVIDIA and ComfyUI Streamline Local AI Video Generation for Game Developers and Creators at GDC

AI-powered video generation becomes more accessible with ComfyUI’s App Mode view, NVIDIA RTX Video Super Resolution and new NVFP4 models.
by Michael Fukuyama

Game developers and artists are building cinematic worlds and iconic characters — raising the bar for immersive experiences on NVIDIA RTX AI PCs

At the Game Developers Conference (GDC) in San Francisco this week, NVIDIA announced a suite of updates that streamline AI video generation for concept development and storyboarding on RTX GPUs and the NVIDIA DGX Spark desktop supercomputer.

These announcements include:

  • ComfyUI’s new App View with a simplified interface, lowering the barrier for entry for the popular generative AI tool.
  • RTX Video Super Resolution available for ComfyUI, a real-time 4K upscaler ideal for video generation — also available for developers as a Python Wheel.
  • NVFP4 and FP8 model variants are available today for FLUX.2 Klein, with NVFP4 support for LTX-2.3 coming soon, delivering up to 2.5x performance gains and 60% lower memory usage for both models.

Frictionless Local AI: Collaborate, Optimize, Customize

Many of today’s popular AI applications are making it easier for beginners to try state-of-the-art models directly on their laptop or desktop.

For artists unfamiliar with node graphs, ComfyUI’s new App View presents workflows in a simplified interface. Users only need to enter a prompt, adjust simple parameters and hit generate. The full node-based experience remains available as Node View, and users can seamlessly switch between the two modes.

App View is compatible with the RTX optimizations in ComfyUI. Performance for RTX GPUs is 40% faster since September, and ComfyUI now supports NVFP4 and FP8 data formats natively. All combined, performance is 2.5x faster and VRAM is reduced by 60% with NVIDIA GeForce RTX 50 Series GPUs’ NVFP4 format, and performance is 1.7x faster and VRAM is reduced by 40% with FP8.

At CES in January, NVIDIA announced several models released with NVFP4 and FP8 support. And now more NVFP4 and FP8 models are available — LTX-2.3, with NVFP4 support coming soon, FLUX.2 Klein 4B, and FLUX.2 Klein 9B directly in ComfyUI. To get started, download the NVFP4 and FP8 checkpoints directly from Hugging Face, load the default workflows in ComfyUI via the Template Browser and replace the default model checkpoint with the newly downloaded checkpoint. 

App View mode is available today. Learn more on ComfyUI

Faster 4K Video Generation 

Getting high-quality video outputs often means juggling three constraints: speed, VRAM and control. While many artists ultimately want 4K quality, most prefer to generate smaller, faster previews first, and then upscale them. Today’s upscalers take minutes to upscale a 10‑second clip into 4K resolution.

Now, users can quickly upscale generated video to 4K with NVIDIA RTX Video Super Resolution, available as a node for ComfyUI. RTX Video can be accessed as a standalone node for building video workflows from scratch.

For AI developers, NVIDIA released a free Python package available via the PyPI repository, along  with sample code on GitHub and a VFX Python bindings guide, to get started quickly. The package provides programmatic access to the same AI upscaling technology that powers RTX Video, running directly on RTX GPU Tensor Cores to deliver 4K upscaling 30x faster than alternative popular local upscalers, and at a fraction of the VRAM cost. The package is powered by the NVIDIA Video Effects software development kit.

Generative AI model performance for LTX-2 and FLUX.2 Klein 9B on an NVIDIA RTX 5090 GPU. Performance testing done on RTX 5090. LTX-2: 512×768 resolution, 100 frames, 20 steps. FLUX.2 Klein 9B (base): 1024×1024 resolution, 20 steps.

Ready to get started with ComfyUI? Check out the latest NVIDIA Studio Sessions tutorial hosted by  visual effects artist Max Novak for a guided walkthrough:

#ICYMI: The Latest Updates for RTX AI PCs at GDC

🎉Join NVIDIA at GTC, March 16-19 in San Jose! Check out “Create Generative AI Workflow for Design and Visualization in ComfyUI” on March 17, for a training session led by NVIDIA 3D workflow specialists focused on building RTX-accelerated generative workflows for images, video, 3D, and PBR materials. Register today and explore the session catalog.

💡LTX Desktop is a fully local, open-source video editor running directly on the LTX engine, optimized for NVIDIA GPUs and compatible hardware.

🦥 LM Link connects separate devices running LM Studio, allowing models to run on remote machines as if they were local. It’s ideal for users wanting to run an agent on their laptop while still accessing free and private AI, powered by their DGX Spark or RTX desktop. Learn how to run LM Studio on DGX Spark.

🎮On Tuesday, March 31, as part of the next opt-in NVIDIA App beta, overrides for NVIDIA DLSS 4.5 Dynamic Multi Frame Generation and DLSS 4.5 Multi Frame Generation 6x Mode will be released for GeForce RTX 50 Series owners. Learn about NVIDIA news at GDC.

🤖Next month, a new NVIDIA RTX Remix update will introduce Advanced Particle VFX, enabling modders to create a wide array of particle effects that further improve image quality, detail and immersion.

🦄Topaz Labs has collaborated with NVIDIA to optimize NeuroStream for NVIDIA GPUs — a proprietary VRAM optimization that allows complex AI models to run on consumer hardware.

📃Microsoft has introduced support for VoiceMod, one of the first apps to enable Windows ML for GPU inference, significantly improving performance voice quality compared with CPUs. 

Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter. Follow NVIDIA Workstation on LinkedIn and X

See notice regarding software product information.

AI Is a 5-Layer Cake

by Jensen Huang