Editor’s note: This post is part of the AI Decoded series, which demystifies AI by making the technology more accessible, and showcases new hardware, software, tools and accelerations for GeForce RTX PC and NVIDIA RTX workstation users.
The AI Decoded series over the past year has broken down all things AI — from simplifying the complexities of large language models (LLMs) to highlighting the power of RTX AI PCs and workstations.
Recapping the latest AI advancements, this roundup highlights how the technology has changed the way people write, game, learn and connect with each other online.
NVIDIA GeForce RTX GPUs offer the power to deliver these experiences on PC laptops, desktops and workstations. They feature specialized AI Tensor Cores that can deliver more than 1,300 trillion operations per second (TOPS) of processing power for cutting-edge performance in gaming, creating, everyday productivity and more. For workstations, NVIDIA RTX GPUs deliver over 1,400 TOPS, enabling next-level AI acceleration and efficiency.
Unlocking Productivity and Creativity With AI-Powered Chatbots
AI Decoded earlier this year explored what LLMs are, why they matter and how to use them.
For many, tools like ChatGPT were their first introduction to AI. LLM-powered chatbots have transformed computing from basic, rule-based interactions to dynamic conversations. They can suggest vacation ideas, write customer service emails, spin up original poetry and even write code for users.
Introduced in March, ChatRTX is a demo app that lets users personalize a GPT LLM with their own content, such as documents, notes and images.
With features like retrieval-augmented generation (RAG), NVIDIA TensorRT-LLM and RTX acceleration, ChatRTX enables users to quickly search and ask questions about their own data. And since the app runs locally on RTX PCs or workstations, results are both fast and private.
NVIDIA offers the broadest selection of foundation models for enthusiasts and developers, including Gemma 2, Mistral and Llama-3. These models can run locally on NVIDIA GeForce and RTX GPUs for fast, secure performance without needing to rely on cloud services.
Download ChatRTX today.
Introducing RTX-Accelerated Partner Applications
AI is being incorporated into more and more apps and use cases, including games, content creation apps, software development and productivity tools.
This expansion is fueled by the wide selection of RTX-accelerated developer and community tools, software development kits, models and frameworks have made it easier than ever to run models locally in popular applications.
AI Decoded in October spotlighted how Brave Browser’s Leo AI, powered by NVIDIA RTX GPUs and the open-source Ollama platform, enables users to run local LLMs like Llama 3 directly on their RTX PCs or workstations.
This local setup offers fast, responsive AI performance while keeping user data private — without relying on the cloud. NVIDIA’s optimizations for tools like Ollama offer accelerated performance for tasks like summarizing articles, answering questions and extracting insights, all directly within the Brave browser. Users can switch between local and cloud models, providing flexibility and control over their AI experience.
For simple instructions on how to add local LLM support via Ollama, read Brave’s blog. Once configured to point to Ollama, Leo AI will use the locally hosted LLM for prompts and queries.
Agentic AI — Enabling Complex Problem-Solving
Agentic AI is the next frontier of AI, capable of using sophisticated reasoning and iterative planning to autonomously solve complex, multi-step problems.
AI Decoded explored how the AI community is experimenting with the technology to create smarter, more capable AI systems.
Partner applications like AnythingLLM showcase how AI is going beyond simple question-answering to improving productivity and creativity. Users can harness the application to deploy built-in agents that can tackle tasks like searching the web or scheduling meetings.
AnythingLLM lets users interact with documents through intuitive interfaces, automate complex tasks with AI agents and run advanced LLMs locally. Harnessing the power of RTX GPUs, it delivers faster, smarter and more responsive AI workflows — all within a single local desktop application. The application also works offline and is fast and private, capable of using local data and tools typically inaccessible with cloud-based solutions.
AnythingLLM’s Community Hub lets anyone easily access system prompts that can help them steer LLM behavior, discover productivity-boosting slash commands and build specialized AI agent skills for unique workflows and custom tools.
By enabling users to run agentic AI workflows on their own systems with full privacy, AnythingLLM is fueling innovation and making it easier to experiment with the latest technologies.
AI Decoded Wrapped
Over 600 Windows apps and games today are already running AI locally on more than 100 million GeForce RTX AI PCs and workstations worldwide, delivering fast, reliable and low-latency performance. Learn more about NVIDIA GeForce RTX AI PCs and NVIDIA RTX AI workstations.
Tune into the CES keynote delivered by NVIDIA founder and CEO Jensen Huang on Jan. 6. to discover how the latest in AI is supercharging gaming, content creation and development.
Generative AI is transforming gaming, videoconferencing and interactive experiences of all kinds. Make sense of what’s new and what’s next by subscribing to the AI Decoded newsletter.