Red Hat unveils enhancements for streamlined enterprise AI deployment

Red Hat unveils enhancements for streamlined enterprise AI deployment
Matt Hicks President and Chief Executive Officer — Red Hat
0Comments

Red Hat, a prominent provider of open-source solutions, has announced new advancements in its AI offerings. These include the introduction of the Red Hat AI Inference Server, third-party validated models, and integration with Llama Stack and Model Context Protocol (MCP) APIs. These developments are designed to enhance enterprise AI deployment across various environments.

Joe Fernandes, Vice President and General Manager of Red Hat’s AI Business Unit, stated: “Faster, more efficient inference is emerging as the newest decision point for gen AI innovation. Red Hat AI, with enhanced inference capabilities through Red Hat AI Inference Server and a new collection of validated third-party models, helps equip organizations to deploy intelligent applications where they need to, how they need to and with the components that best meet their unique needs.”

The new Red Hat AI Inference Server aims to provide faster and more cost-effective inference at scale within hybrid cloud settings. It integrates into recent releases of Red Hat OpenShift AI and Red Hat Enterprise Linux AI but is also available independently.

Additionally, the third-party validated models from Red Hat are now accessible on Hugging Face. These models have been optimized for performance on the Red Hat platform using compression techniques to improve speed while reducing resource consumption.

Incorporating standardized APIs for building and deploying AI applications is another focus area for Red Hat. The company is integrating Llama Stack by Meta and MCP by Anthropic into its systems. Currently available in developer preview within Red Hat AI, these tools aim to streamline access to various generative AI functions.

Red Hat’s enhancements also extend to updates in its OpenShift AI (v2.20) which now includes features such as an optimized model catalog and distributed training capabilities through KubeFlow Training Operator.

Michele Rosen from IDC remarked: “Organizations are moving beyond initial AI explorations and are focused on practical deployments. The key to their continued success lies in the ability to be adaptable with their AI strategies to fit various environments and needs.”

With these updates, Red Hat continues its vision of providing versatile infrastructure options that allow organizations greater flexibility in deploying any model across any cloud or accelerator environment.



Related

Brian Moynihan Chair of the Board and Chief Executive Officer

Bank of America announces $1 billion equity award for non-executive employees

Bank of America has announced it will distribute $1 billion in stock to employees through its Sharing Success Program.

D. Reid Wilson, Secretary at North Carolina Department of Environmental Quality

South Atlantic Fishery Management Council schedules stakeholder meetings across North Carolina

The North Carolina Department of Environmental Quality’s Division of Marine Fisheries is encouraging fishermen and stakeholders to attend a series of upcoming meetings hosted by the South Atlantic Fishery Management Council.

Harry K. Sideris, President and Chief Executive Officer

Duke Energy declares quarterly dividends for common and preferred shareholders

Duke Energy has announced that it will pay a quarterly cash dividend of $1.065 per share on its common stock.

Trending

The Weekly Newsletter

Sign-up for the Weekly Newsletter from North Wake News.