Llama 4 Herd is here and already works with Red Hat OpenShift AI
Summary:Meta released its newest version of the Llama model family this weekend–Llama 4As a leading commercial contributor to the vLLM project, Red Hat collaborated with the Meta team to help enable Day Zero support for Llama 4 inference with vLLMFurthermore, the Red Hat OpenShift AI team has enabled our customers to experiment with Llama 4 using the latest release of vLLM inside their OpenShift AI environments.Read on to get started with inferencing the Llama 4 Heard inside OpenShift AI!Over the April 5 weekend, Meta released its newest version of the Llama model family–Llama 4, enabling
Meet the latest OpenShift Superheroes
This week at Red Hat OpenShift Commons in London we had the chance to recognize a few OpenShift Superheroes. While each member of the OpenShift community is a hero, helping contribute to the project’s success and growth, some members really stand out. These contributors are the advocates and champions that make the community strong and successful.OpenShift Superheroes are made up of:Builders: contributing to the evolution of OpenShiftAdvocates: amplifying their OpenShift experience and learnings through events, blogs, meetups and more.Ambassadors: shares knowledge across diverse teams, indus
Red Hat named to Fast Company’s annual list of the World’s Most Innovative Companies of 2025
Today, we are proud to announce that Red Hat has been named to Fast Company’s prestigious list of the World’s Most Innovative Companies of 2025. The 2025 list recognizes 609 organizations across 58 sectors and regions, all businesses that are shaping industry and culture through innovations that set new standards and achieve remarkable milestones in all sectors of the economy.For context, here’s what Fast Company Editor-in-Chief Brendan Vaughan says about this year’s list:“Our list of the Most Innovative Companies offers both a comprehensive look at innovation today and a playbook fo
Get ready for OpenShift Commons Gathering in London
The upcoming Red Hat OpenShift Commons Gathering in London is taking place on April 1, co-located alongside KubeCon + CloudNativeCon Europe. OpenShift Commons Gatherings bring together users, partners, customers, contributors and upstream project leads to collaborate and work together across the OpenShift cloud-native ecosystem. Next month’s event is shaping up to have a great line up of end-users who will be sharing use cases, insights into their workloads and lessons learned along the way. The event will also feature an Ask-me-Anything area, breakout sessions, labs and the presentation of
Transitioning Red Hat’s Operational and IT leadership
Today, we are sharing news that Carolyn Nash, Red Hat’s senior vice president and chief operating officer (COO), will retire after a 35-year career spanning numerous leadership positions and more than eight years with Red Hat. Bobby Leibrock will step into a combined role of senior vice president, chief operating officer and chief financial officer, effective immediately. Carolyn will remain with Red Hat in an advisory capacity through June 2025 to assist with the transition. As part of this move, Marco Bill has been named senior vice president and chief information officer, reporting to Bob
Griot and Grits is preserving Black history through AI
Artificial Intelligence (AI) is quickly paving the way for technological innovation across countless industries and has infiltrated many of our day-to-day activities–especially in education. Educational institutions have increasingly adopted AI technology in the spirit of innovating teaching and learning practices and providing greater access to educational resources. History, for example, is a core subject that can benefit from the use of AI to enrich and synthesize the context of historical events, figures and movements as additional background information and sources are uncovered–and R
What’s new with the command line interface in RHEL AI 1.4?
How many times have you had to start running a program on your terminal, and then… you’re just stuck there? Suddenly, you can’t close out your terminal or multi-task.Well, no more–kind of!With the release of Red Hat Enterprise Linux (RHEL) AI 1.4, Red Hat is introducing background process management for developer preview. This feature enables you to run the synthetic data generation (SDG) process in the background by allowing you to detach it from the terminal, spin off child processes to remain running in the background, monitor the status, reattach the process to bring it back and te
Unlocking the Effective Context Length: Benchmarking the Granite-3.1-8b Model
Large language models (LLMs) are evolving rapidly enabling applications such as chatbots, code generation, and knowledge extraction. One crucial factor influencing their effectiveness is the context length - the number of tokens a model can look at once. While theoretical context lengths continue to grow, the practical, effective context length (ECL) determines real-world usability.In this blog post, we will explore the ECL of the Granite-3.1-8b instruct model and validate its capabilities across various tasks. The study takes its inspiration from the paper "Measuring Effective Context Length
Introducing RHEL AI 1.4: Powering the Next Wave of Generative AI Innovation
The field of generative AI (gen AI) is evolving rapidly, and Red Hat Enterprise Linux AI (RHEL AI) is at the forefront of this transformation. RHEL AI 1.4 introduces support for a new, powerful Granite 3.1 model with multi-language support and a larger context window. The latest version also adds capabilities for evaluating customized and trained models with Document-Knowledge bench (DK-bench), a benchmark designed to measure the knowledge a model has acquired. Finally, we’re introducing the developer preview of our new graphical UI that helps to simplify data ingestion and writing InstructL
Empowering non-technical users to contribute knowledge and enhance model responses
In every release of our products, the User Experience Design team at Red Hat (UXD) prioritizes implementing work that directly relates to customer feedback and insights. For a nascent product like RHEL AI, we have had to be creative about doing so. Feedback channels we used for this upcoming release to help us continuously improve the experience include upstream community engagement and direct feedback from users via Slack, interviews with internal team members, interviews with AI users and enthusiasts who may be interested in adopting RHEL AI, and examining our competitive landscape. This has
Evolving our middleware strategy
Editor’s note: Earlier today, president and chief executive officer of Red Hat, Matt Hicks, shared the following email with Red Hatters.--Hi all,Today, we’re sharing that Red Hat and IBM will join forces to secure the future of the Java application ecosystem for our customers. The Red Hat middleware engineering and product teams will join the IBM Data Security, IAM, and Runtimes organization, forming a single team.* This move supports and grows our combined customer footprint, and builds a unified product strategy for the future of Java applications and Integration solutions in the era of
Open source AI: Red Hat’s point-of-view
TL;DR - Red Hat views the minimum criteria for open source AI as open source-licensed model weights combined with open source software components.More than three decades ago, Red Hat saw the potential of how open source development and licenses can create better software to fuel IT innovation. Thirty-million lines of code later, Linux not only developed to become the most successful open source software but the most successful software to date. Our commitment to open source principles continues today, not only in our business model, but in our corporate culture as well. We believe that these c
Nominations now open for the OpenShift Superhero Awards
Each member of the OpenShift community is a hero, helping contribute to the project’s success and growth. But some members really stand out - they are the advocates and champions that make the community strong and successful. These are our superheroes.The vibrant OpenShift community is built on the contributions of its members and we are now giving users the opportunity to recognize those that help to advance the community. Nominations are currently open for the OpenShift Superhero Awards, acknowledging users for their contributions, participation and innovation.OpenShift superheroes are ma
The EU Cyber Resilience Act - what you need to know
Today marks a new milestone in European cybersecurity: the Cyber Resilience Act (CRA) has been published in the EU’s Official Journal, bringing significant changes for businesses operating in the EU. But what does this mean for companies and users alike, and how is Red Hat positioned to support your needs in the new landscape?The CRA is a robust new legislative framework aimed at enhancing the cybersecurity of (hardware and software) products with digital elements - everything from smart home devices to complex operating systems in critical national infrastructure. The CRA enters into force
Red Hat to Contribute Comprehensive Container Tools Collection to Cloud Native Computing Foundation
The continued importance of cloud-native applications in an AI and hybrid cloud-centric world demands an open, more accessible ecosystem of development tools. Today, we’re pleased to help drive cloud-native evolution further into the next-generation of IT with our intent to contribute a comprehensive set of container tools to the Cloud Native Computing Foundation (CNCF), including bootc, Buildah, Composefs, Podman, Podman Desktop and Skopeo.Upon acceptance by the CNCF, the contributed tools will become hosted projects – alongside technologies like Kubernetes, Prometheus, Helm and many more
The State of Platform Engineering in the Age of AI
Platform engineering has changed how organizations develop, deploy and manage applications by streamlining processes, improving efficiencies and fostering collaboration. As organizations now look to integrate generative AI (gen AI) as quickly and as often as possible - platform engineering is proving to be critical to setting organizations up for success in adopting and implementing these groundbreaking new technologies.Our inaugural State of Platform Engineering in the Age of AI report examines trends, challenges and best practices from industry practitioners to help us to better understand h
How to make generative AI more consumable
Think about some of the past trends in technology, and you’ll start to see some patterns emerge. For example, with cloud computing there’s no one-size-fits-all approach. Combinations of different approaches, such as on premise and different cloud providers, have led to organizations taking advantage of hybrid infrastructure benefits in deploying their enterprise applications. When we think about the future, a similar structure will be essential for the consumption of artificial intelligence (AI) across diverse applications and business environments. Flexibility will be crucial as no single
FAQ: Red Hat to acquire Neural Magic
FAQ: Red Hat to acquire Neural MagicWhat is being announced?On Nov. 12, 2024, Red Hat announced that it has signed a definitive agreement to acquire Neural Magic, subject to regulatory reviews and other customary closing conditions.What is Neural Magic?Neural Magic is a Somerville, Massachusetts-based early stage company developing technology that helps organizations optimize AI models for greater performance and efficiency. Specifically, Neural Magic is a leader in the vLLM community project with expertise in quantization and sparsification.What does Neural Magic provide?Neural Magic provides
Creating cost effective specialized AI solutions with LoRA adapters on Red Hat OpenShift AI
Picture this: you have a powerful language model in production, but it struggles to provide satisfying answers for specific, niche questions. You try retrieval-augmented generation (RAG) and carefully crafted prompt engineering, but the responses still need to be revised. The next step might seem to be full model fine tuning—updating every layer of the model to handle your specialized cases—but that demands significant time and compute resources.What is LoRA?Low-rank adaptation (LoRA) is a faster and less resource-intensive fine tuning technique that can shape how a large language model re
An Introduction to TrustyAI
TrustyAI is an open source community dedicated to providing a diverse toolkit for responsible artificial intelligence (AI) development and deployment. TrustyAI was founded in 2019 as part of Kogito, an open source business automation community, as a response to growing demand from users in highly regulated industries such as financial services and healthcare. With increasing global regulation of AI technologies, toolkits for responsible AI are an invaluable and necessary asset to any MLOps platform. Since 2021, TrustyAI has been independent of Kogito, and has grown in size and scope amidst the