The Latest News in AI

We publish news articles on Forbes, which are copied here for your convenience.  

Who Has The Fastest AI Inference, And Why Does It Matter?

A food fight erupted at the AI HW Summit earlier this year, where three companies all claimed to offer the fastest AI processing. All were faster than GPUs. Now Cerebras has claimed insanely fast AI performance with their latest software running on the company’s...

NVIDIA Again Claims The Title For The Fastest AI; Competitors Disagree

Intel Habana Labs and Graphcore add scale and software optimizations, while Google skips this round, choosing to put a stake in the ground for half-trillion parameter models. Every six months, the AI hardware community gathers virtually to strut their hardware stuff...

Qualcomm Launches Unified AI Stack For Cloud To Edge Intelligence

The new Qualcomm AI Stack lays the foundation for domain-specific SDK’s While advances in AI accelerators have improved silicon performance by over a thousand-fold, it’s software that turns bits of silicon into useful capabilities for consumers and businesses. After...

A Deeper Look Into Qualcomm’s MLPerf Results

Qualcomm not only bested the entire field for power efficiency, a 16-card server was the fastest tested. Qualcomm, perhaps best known for its leadership Snapdragon mobile platform, has further enhanced its AI bona fides with V1.1 MLPerf inference benchmark suite...

Cerebras Publishes 7 Trained Generative AI Models To Open Source

The AI company is the first to use Non-GPU tech to train GPT-based Large Language Models and make available to the AI community. The early days of an open AI community, sharing work and building on each other’s success, is over. As there is now much more money at...

Graphcore And Spell Collaborate For End-To-End ML Operations

Companies also announce free IPU Test Drive with Graphcore AI hardware and Spell software. Building, training, optimizing, and deploying AI models at scale can be excruciatingly difficult. Especially for large models that need to be optimized to run well at scale and...

Amazon EC2 Inf1 Instances Now Support Amazon SageMaker

image: AWS Last year at AWS reInvent, out of the 100s of announcements, I chose the top 5 for overall, long-term impact. One of those was Amazon’s EC2 Inf1 Instances that used its new Inferentia machine learning inference chip. I chose Inf1 Instances as a top 5 for a...

Graphcore Launches 3rd-Gen AI With Wafer-On-Wafer (WoW!) Technology

Startup also teases $120M Brain-Scale AI “Good Computer” For 2024 Remember Intel’s old Tick-Tock processor roadmap? The company alternated between two approaches, bringing out a new generation featuring architectural enhancements, then subsequently shrinking that chip...

Cadence Creates “True Hybrid” Cloud For Designers

Instead of having to choose to run on-prem or in the cloud, with attendant upload times and costs, the new True Hybrid offering from Cadence creates a dynamic environment for chip and system design. When creating a new chip, design teams currently have to chose...

Who Wins If The New Biden AI Export Rules Stand?

While Nvidia and the European Union have expressed their displeasure with the latest salvo of AI export restrictions from the Biden administration, a few companies and countries could actually benefit from them. But the industry as a whole will suffer, and so will...