The Latest News in AI

We publish news articles on Forbes, which are copied here for your convenience.  

AI Inference Is King; Do You Know Which Chip is Best?

Everyone is not just talking about AI inference processing; they are doing it. Analyst firm Gartner released a new report this week forecasting that global generative AI spending will hit $644 billion in 2025, growing 76.4% year-over-year. Meanwhile, MarketsandMarkets...

AMD Claims MI300X Is The World’s Fastest AI Hardware

The hardware looks quite capable, but the software optimization story has a long way to go to get close to Nvidia. But given the current demand/supply imbalance, I suspect AMD can sell all they can make. AMD launched the MI300 in San Jose to an anxious audience of...

Meta Builds World’s Largest AI Supercomputer With NVIDIA For AI Research And Production

There are big implications, for both companies, beyond just bragging rights. Facebook, I mean Meta, has always been one of the industry leaders when it comes to AI research and deployment. The company processes hundreds of trillions (yes, trillions with a “T”) of...

News Flash: NVIDIA Remains The Fastest Available AI Accelerator

Ok, while that is hardly surprising news, given the comfortable lead NVIDIA enjoys, Google’s upcoming TPUv4 out-performed NVIDIA on three of eight benchmarks on a chip-to-chip basis, while Graphcore’s performance and price/performance is in the ball park. Meanwhile,...

Microsoft Beats Google To The Punch (Bowl), Adds ChatGPT To Bing

That wasn’t that hard, now, was it? The Bingmasters blended the prose results from ChatGPT with advertising revenue generating web urls. I’m switching to Bing! For the last 2 months, everyone (well, investors and the media) has been speculating how ChatGPT could...

What NVIDIA Will Say At Hotchips‘22

“We are the Accelerated Data Center”: GPU’s, CPUs, Networking and Systems. As anyone not comatose knows, today’s modern data center workloads — like AI, HPC, and machine learning — absolutely demand acceleration. And the appetite for acceleration seems insatiable, in...

Cerebras Gets Into The Inference Market With A Bang

Cerebras’ Wafer-Scale Engine has only been used for AI training, but new software enables leadership inference processing performance and costs. Should Nvidia be afraid? As Cerebras prepares to go public, it has expanded its target markets and competitive stance by...

Is Jensen Huang Nvidia’s Chief Revenue Destruction Officer?

At this year's GTC event in San Jose, Nvidia CEO Jensen Huang held over 25,000 people in the palm of his hand, captivated by his vision of AI and how it could transform the world we live in. Some folks in the audience couldn't keep up and started fiddling with their...

What’s Next For On-Device AI? Ask Qualcomm

Qualcomm held a one day analyst event in San Diego updating us all on their AI research. Pretty amazing stuff, but the big news is yet to come with a new Oryon-based Snapdragon expected this Fall, and perhaps a new Cloud AI 100 next year. Qualcomm was pretty confident...

NVIDIA Keeps The Performance Crown For AI Inference For The 6th Time In A Row

In The Data Center And On The Edge, the bottom line is that the H100 (Hopper-based) GPU is up to four times faster than the NVIDIA A100 on the newly released ​MLPerf V2.1 benchmark suite. The A100 retains leadership in many benchmarks versus other available products...