The Latest News in AI

We publish news articles on Forbes, which are copied here for your convenience.  

Following MSFT & GOOG Lead, AWS Embraces Nvidia And Upgrades Own Chips

As expected, Amazon AWS CEO Adam Selipsky announced updated AWS-engineered AI Training and Arm CPU chips to lower the cost of cloud services. But AWS has a problem: they have kept Nvidia at arm’s length to their detriment as Microsoft Azure increases market share. To...

Tenstorrent Could Reshape The AI And CPU Competitive Landscape

Now led by Jim Keller, the company has built a new leadership team and a new strategy. It has tremendous potential. Now it must execute. Introduction It is hard to believe the difference a year makes. In 2021, there were over 100 public and venture-backed startups...

NVIDIA Adds New Software That Can Double H100 Inference Performance

TensorRT-LLM adds a slew of new performance-enhancing features to all NVIDIA GPUs. Just ahead of the next round of MLPerf benchmarks, NVIDIA has announced a new TensorRT software for Large Language Models (LLMs) that can dramatically improve performance and efficiency...

INTEL Lays Down The Gauntlet For AMD And NVIDIA GPUs

The company’s Xe GPU for HPC looks pretty good, but will it be good enough to catch up to AMD and NVIDIA? Intel has adopted a "Domain-Specific Architecture" strategy espoused by John L. Hennessy, Alphabet Chairman and former President of Stanford...

Amazon EC2 Inf1 Instances Now Support Amazon SageMaker

image: AWS Last year at AWS reInvent, out of the 100s of announcements, I chose the top 5 for overall, long-term impact. One of those was Amazon’s EC2 Inf1 Instances that used its new Inferentia machine learning inference chip. I chose Inf1 Instances as a top 5 for a...

Perceive AI Launches 2nd Edge AI Chip For Low Power Applications

Company claims Ergo2 is up to four times faster than Perceive’s first-generation Ergo chip, and can handle much larger models such as NLP. Edge AI is coming into its own, with a variety of chips being launched that offer low cost, low power, and high performance....

Enhanced Memory Grace Hopper Superchip Could Shift Demand To NVIDIA CPU And Away From X86

The company’s new high bandwidth memory version is only available with the CPU-GPU Superchip. In addition, a new dual Grace-Hopper MGX Board offers 282GB of fast memory for large model inferencing. The AI landscape continues to change rapidly, and fast memory (HBM)...

IBM Research And The Full Stack Approach To Innovation

IBM’s vaunted research organization, one of the largest in the industry, thinks the path to innovation is to be taken as a team with clients and partners exploring the full stack needed to address a business challenge. We had a chance this week to hear how IBM...

Why Can’t NVIDIA Be Bested In MLPerf?

MLPerf, an industry consortium of over 70 companies and institutions, has released the second round of AI Inference processing results. These benchmarks now represent production applications from all major areas of AI deployment today. But only a few technologies...

Will Open AI’s o1 Reasoning Model Really Change The World?

OpenAI launched Strawberry — another name for its its 01 model — on Sept. 12, including the full function o1-preview and the more affordable o1-mini to demonstrate how AI can be greatly improved by breaking a query down into step-by-step reasoning. This...