Graphics processing unit

Red Hat Supercharges AI Workloads with Intel Technologies from the Datacenter to the Edge

Retrieved on: 
Tuesday, May 7, 2024

Red Hat, Inc., the world's leading provider of open source solutions, today announced a collaboration with Intel to power enterprise AI usages on Red Hat OpenShift AI .

Key Points: 
  • Red Hat, Inc., the world's leading provider of open source solutions, today announced a collaboration with Intel to power enterprise AI usages on Red Hat OpenShift AI .
  • Red Hat and Intel are addressing these needs with a hybrid cloud approach wherein the AI stack is not confined to a single cloud or single datacenter.
  • Red Hat is collaborating with Intel to certify its hardware solutions on Red Hat OpenShift AI to ensure interoperability and enable comprehensive AI capabilities.
  • With this collaboration, Red Hat and Intel bring together the might of their respective ecosystems, spanning leading solution providers, software vendors, equipment manufacturers and more, to further extend and build AI capabilities, solutions and services on Intel® AI products with Red Hat OpenShift AI.

Astera Labs Announces Financial Results for the First Quarter of Fiscal Year 2024

Retrieved on: 
Tuesday, May 7, 2024

Astera Labs (Nasdaq: ALAB), a global leader in semiconductor-based connectivity solutions for cloud and AI infrastructure, today announced preliminary financial results for the first quarter of fiscal 2024, ended March 31, 2024.

Key Points: 
  • Astera Labs (Nasdaq: ALAB), a global leader in semiconductor-based connectivity solutions for cloud and AI infrastructure, today announced preliminary financial results for the first quarter of fiscal 2024, ended March 31, 2024.
  • “Astera Labs started the year strong, achieving record revenue in the first quarter, driven by the accelerating deployment of AI infrastructure,” said Jitendra Mohan, Astera Labs’ Chief Executive Officer.
  • Net proceeds to Astera Labs from the offering were $672.2 million after deducting underwriting discounts and commissions.
  • We also estimate the following:
    Astera Labs will host a conference call to review its financial results for the first quarter of fiscal 2024 and to discuss our financial outlook today at 1:30 p.m. Pacific Time.

Red Hat and Run:ai Optimize AI Workloads for the Hybrid Cloud

Retrieved on: 
Tuesday, May 7, 2024

Red Hat and Run:ai are working to meet this critical need for GPU resource optimization with Run:ai’s certified OpenShift Operator on Red Hat OpenShift AI, which helps users scale and optimize wherever their AI workloads are located.

Key Points: 
  • Red Hat and Run:ai are working to meet this critical need for GPU resource optimization with Run:ai’s certified OpenShift Operator on Red Hat OpenShift AI, which helps users scale and optimize wherever their AI workloads are located.
  • Run:ai’s cloud-native compute orchestration platform on Red Hat OpenShift AI helps:
    Address GPU scheduling issues for AI workloads with a dedicated workload scheduler to more easily prioritize mission-critical workloads and confirm that sufficient resources are allocated to support those workloads.
  • In the future, Red Hat and Run:ai plan to continue building on this collaboration with additional integration capabilities for Run:ai on Red Hat OpenShift AI.
  • Join the Red Hat Summit keynotes to hear the latest from Red Hat executives, customers and partners:

GigaIO and SourceCode Launch First Ever Suitcase-sized AI Supercomputer

Retrieved on: 
Monday, May 6, 2024

GigaIO and SourceCode today announced the launch of Gryf, the industry’s most powerful mobile AI platform for tactical edge environments, at GEOINT 2024.

Key Points: 
  • GigaIO and SourceCode today announced the launch of Gryf, the industry’s most powerful mobile AI platform for tactical edge environments, at GEOINT 2024.
  • Billed as “Gryf, co-designed by GigaIO and SourceCode,” the system is the result of a unique technology collaboration between the companies that is designed to engineer complex customer solutions.
  • GigaIO provides award-winning open workload-defined infrastructure for AI and accelerated computing, and Source­Code is a global provider of co-designed, compliant IT systems for next generation intelligent infrastructure.
  • FabreX can disaggregate and re-aggregate GPUs and other system components in order to enable optimal hardware configurations for each workload.

ATMECS Global Accelerates AI Innovation by Joining the NVIDIA Partner Network as Solution Advisor Consultant.

Retrieved on: 
Monday, May 6, 2024

ATMECS Global, renowned for its prowess in digital transformation and engineering solutions, proudly announces that it has joined the NVIDIA Partner Network as a Solution Advisor Consultant.

Key Points: 
  • ATMECS Global, renowned for its prowess in digital transformation and engineering solutions, proudly announces that it has joined the NVIDIA Partner Network as a Solution Advisor Consultant.
  • The milestone marks a pivotal step in ATMECS’s journey to redefine the AI and ML landscape, amplifying its commitment to innovation and excellence in delivering next-generation AI solutions.
  • We aim to unlock new possibilities and deliver cutting-edge AI solutions that drive business transformation and create a competitive edge for our clients."
  • With AI, ATMECS’s clients can transform how they operate and innovate, providing an unprecedented edge in the competitive AI marketplace.

Hyperscale cloud support for AI: Google leads in technology, Amazon Web Services excels in price

Retrieved on: 
Wednesday, May 8, 2024

GCP benefits from Google's status as a powerhouse of fundamental AI research, while Amazon Web Services benefits both from the enormous scale of its existing business and its excellence in day-to-day operations.

Key Points: 
  • GCP benefits from Google's status as a powerhouse of fundamental AI research, while Amazon Web Services benefits both from the enormous scale of its existing business and its excellence in day-to-day operations.
  • Customers looking to adopt the latest technology will be best served by GCP, while those focused on price will be best served by AWS.
  • By definition, inference is required when an AI application goes into production, with demand driven by end user needs.
  • The study provides analysis on pricing and availability of custom AI silicon, such as Google TPUs, flagship, mid-range, and entry-level GPUs, and CPU options that hyperscalers recommend for AI inference.

INPHER's Enterprise-Ready SecurAI Performantly Protects LLMs with NVIDIA Confidential Computing

Retrieved on: 
Tuesday, May 7, 2024

SAN FRANCISCO, May 7, 2024 /PRNewswire/ -- RSA Conference 2024 - Inpher, pioneers in privacy-enhanced computation, today announced the general availability release of SecurAI, a leading solution that protects the privacy and security of user inputs on large language models. This enterprise-ready release of SecurAI leverages the NVIDIA H100 Tensor Core GPU for maximum speed and performance.

Key Points: 
  • This enterprise-ready release of SecurAI leverages the NVIDIA H100 Tensor Core GPU for maximum speed and performance.
  • "Inpher's SecurAI coupled with NVIDIA confidential computing technology enables organizations to use confidential GPUs for improved data utility and business productivity."
  • Powered by confidential GPUs, this much anticipated release of SecurAI also integrates retrieval-augmented generation (RAG).
  • SecurAI also enables enterprise users to upload their own files based on their organizational needs by leveraging secure RAG.

SealingTech Powers Up Missions with Its SN 4500

Retrieved on: 
Tuesday, May 7, 2024

COLUMBIA, Md., May 7, 2024 /PRNewswire/ -- Sealing Technologies (SealingTech), a Parsons Company (NYSE: PSN) ramps up production of its SN 4500.

Key Points: 
  • COLUMBIA, Md., May 7, 2024 /PRNewswire/ -- Sealing Technologies (SealingTech), a Parsons Company (NYSE: PSN) ramps up production of its SN 4500.
  • SealingTech expands line of servers with a cost-effective, open-architecture mission flexible compact 1U form factor.
  • SealingTech expands its innovative line of servers with a cost-effective, open-architecture compact 1U form factor that's mission flexible as well as enterprise resilient.
  • As your mission and business needs evolve, the SN 4500 evolves with you.

Groq® is Selected to Provide Access to World's Fastest AI Inference Engine for the National AI Research Resource (NAIRR) Pilot

Retrieved on: 
Monday, May 6, 2024

MOUNTAIN VIEW, Calif., May 6, 2024 /PRNewswire/ -- Groq®, the leader in real-time AI inference, announced its participation in the National Artificial Intelligence Research Resource (NAIRR) Pilot today. The Pilot, a U.S. National Science Foundation-led program, marks the first step towards creating a shared national research infrastructure to connect U.S. researchers and educators to responsible and trustworthy AI research resources. In collaboration with 13 federal agencies and 25 private sector, nonprofit, and philanthropic organizations, Groq is powering the next phase of responsible AI research, discovery, and innovation by providing access to its LPU™ Inference Engine – the only solution delivering real-time AI inference today – via GroqCloud™.

Key Points: 
  • The Pilot, a U.S. National Science Foundation-led program, marks the first step towards creating a shared national research infrastructure to connect U.S. researchers and educators to responsible and trustworthy AI research resources.
  • In collaboration with 13 federal agencies and 25 private sector, nonprofit, and philanthropic organizations, Groq is powering the next phase of responsible AI research, discovery, and innovation by providing access to its LPU™ Inference Engine – the only solution delivering real-time AI inference today – via GroqCloud™.
  • Groq announces its participation in the National Artificial Intelligence Research Resource (NAIRR) Pilot.
  • It is an honor to provide the next generation of AI innovators with the real-time inference needed to run text-based applications and other AI workloads at scale."

ScaleFlux's Solid-State Drives Sustainability and Energy Efficiency in the New AI Culture

Retrieved on: 
Monday, May 6, 2024

MILPITAS, Calif., May 6, 2024 /PRNewswire-PRWeb/ -- The apocryphal tale of Benjamin Franklin venturing out into a storm with kite and key in hand set the stage for the 21st century's voracious appetite for electricity. Artificial Intelligence has proved to be a glutton for energy consumption. The amount of computing power necessary for AI machine learning is doubling every six months. (1) Data center behemoths like Amazon Web Services, Microsoft, and Google are looking to reduce their carbon footprint, increase efficiency, and reduce energy costs. To comply with this trend, the IT infrastructure is evolving towards a more distributed compute architecture and a broader variety of processor cores and types (such as Arm-based CPUs and system-on-chip processors). JB Baker, VP of Products for ScaleFlux says, "Large-scale AI workloads rely heavily on memory and networking, putting much of the current focus on alleviating bottlenecks in these areas. However, with the petabyte and exabyte scale of the datasets used in AI, power and efficiencies of the underlying storage infrastructure are quietly a major challenge. Improving the pipeline between storage, memory, and GPUs with solutions like solid-state drives (SSDs) enhanced with computational storage capabilities is critical in enabling companies to achieve their efficiency and energy sustainability goals."

Key Points: 
  • Solid-state drives and Arm-based solutions enable AI infrastructures to maintain high performance while remaining energy-efficient, bringing about the best of both worlds.
  • (1) Data center behemoths like Amazon Web Services, Microsoft, and Google are looking to reduce their carbon footprint, increase efficiency, and reduce energy costs.
  • Improving the pipeline between storage, memory, and GPUs with solutions like solid-state drives (SSDs) enhanced with computational storage capabilities is critical in enabling companies to achieve their efficiency and energy sustainability goals."
  • This cutting-edge technology offers unprecedented data transfer speeds and energy efficiency to meet the growing demands for higher bandwidth and energy-efficient storage solutions.