Weather Data Source: Wetter vorhersage 30 tage

NVIDIA RTX 5090 Dominates AMD & Apple with Local OpenAI Models

Developers and creative professionals are increasingly drawn to running AI models locally, especially for greater control and enhanced privacy. One of the most popular choices these days is OpenAI’s new gpt-oss family of models. These models are lightweight yet powerful and can work well on everyday hardware, requiring only 16GB of memory. If you’re looking for the best experience with these models, NVIDIA’s GPUs, especially the latest RTX 5090, are leading the pack.

Why Local AI Models are Gaining Popularity

With countless companies and nations racing to develop their own unique AI solutions, many are also gravitating towards open-source models like OpenAI’s gpt-oss-20b. This model is similar in performance to the previously popular GPT-4o mini model, which captured a lot of attention over the past year. The gpt-oss-20b introduces several advanced features, such as improved reasoning abilities and expanded context lengths, making it easier to think through complex problems using less hardware.

However, to truly get the best performance out of these models, you’ll need a powerful graphics card. Enter the NVIDIA GeForce RTX 5090—this flagship card offers exceptional speed for gaming but also excels in professional workloads like AI processing. Thanks to its Blackwell architecture, thousands of CUDA cores, and massive 32GB memory, the RTX 5090 is perfectly suited for running local AI models.

The Role of Llama.cpp and NVIDIA Optimization

Llama.cpp is an open-source framework designed to enable efficient usage of large language models (LLMs). Thanks to collaborative optimizations with NVIDIA, running LLMs on RTX GPUs has never been better. Llama.cpp allows users to adjust quantization techniques and manage CPU loading more flexibly, which further boosts its performance.

In recent tests performed by Llama.cpp, the RTX 5090 achieved an impressive benchmark of 282 tokens per second (tok/s) while running the gpt-oss-20b model. For comparison, the Apple Mac M3 Ultra managed only 116 tok/s, while AMD’s 7900 XTX recorded a mere 102 tok/s. This remarkable performance is due in part to the built-in Tensor Cores within the RTX 5090, which are specifically designed to speed up AI tasks.

For those who might not be familiar, tok/s measures how quickly the model can read or produce text. So, a higher number means a more efficient processing ability.

User-Friendly AI Applications

For AI enthusiasts interested in using local LLMs, LM Studio is an excellent option built on the Llama.cpp framework. LM Studio is designed to streamline the experience of running and experimenting with large LLMs, removing the need to tackle complex command-line tools or deep technical setups. It introduces support for Retrieval-Augmented Generation (RAG), making it even easier to work with different models.

Another highly regarded open-source framework is Ollama. This user-friendly tool is great for experimenting with various AI models, including the OpenAI gpt-oss models. NVIDIA has worked closely with Ollama to ensure peak performance on its GPUs, allowing users to seamlessly manage model downloads and GPU acceleration. With built-in model management features, it simplifies working with multiple models in a local environment.

If you want to test the latest gpt-oss model effortlessly, Ollama makes that simple. Similar to Llama.cpp, other applications also harness the power of Ollama to run LLMs, such as AnythingLLM. This tool offers a straightforward local interface, making it fantastic for beginners.

Getting Started with gpt-oss-20b

If you own one of NVIDIA’s latest GPUs, you can easily try out gpt-oss-20b across several platforms. LM Studio provides a sleek interface, while AnythingLLM is designed for easy use on both Windows x64 and Windows on ARM. Ollama, although a bit less visually appealing, allows for quick setups if you already have some technical knowledge.

No matter which application you choose to work with for gpt-oss-20b, it’s clear that NVIDIA’s Blackwell GPUs, particularly the RTX 5090, deliver the best performance.

Conclusion

In summary, if you’re looking for a powerful setup to run local AI models, NVIDIA’s RTX 5090 is currently the best choice. Its outstanding performance, coupled with user-friendly applications like LM Studio and Ollama, makes it easier than ever to explore the capabilities of modern AI technology.

NVIDIA #RTX5090 #OpenAI #gptoss #LocalAI #AIModels #MachineLearning #AItechnology #CreativeSoftware #LocalLLMs

Original Text – https://www.pcworld.com/article/2916928/nvidia-rtx-5090-outperforms-amd-and-apple-running-local-openai-language-models.html