by Karl Freund | Nov 19, 2024 | In the News
Startup launches “Corsair” AI platform with Digital In-Memory Computing, using on-chip SRAM memory that can produce 30,000 tokens/second at 2 ms/token latency for Llama3 70B in a single rack. Using Generative AI, called inference processing, is a memory-intensive...
by Karl Freund | Nov 19, 2024 | In the News
The company tackled inferencing the Llama-3.1 405B foundation model and just crushed it. And for the crowds at SC24 this week in Atlanta, the company also announced it is 700 times faster than Frontier, the worlds fastest supercomputer, on a molecular dynamics...
by Karl Freund | Nov 18, 2024 | In the News
I awoke Sunday morning to an article in The Information written to instigate fear, uncertainty and doubt amongst Nvidia investors and users. Don’t worry. Nvidia’s got this. The article circulating this weekend highlighted the thermal challenges some customers face...
by Karl Freund | Nov 18, 2024 | In the News
Nvidia has gone from a niche provider of HPC technology to becoming a dominant force in the industry. This year’s SC event reinforces that leadership. The annual SuperComputing event in North America is taking place this week in Atlanta, Georgia, and as usual the show...
by Karl Freund | Nov 14, 2024 | In the News
It should not surprise anyone: Nvidia is still the fastest AI and HPC accelerator across all MLPerf benchmarks. And while Google submitted results, AMD was a no-show. This blog has been corrected on 11/14 with a fresh TPU Trillium vs. Blackwell comparison. Say what...
by Karl Freund | Nov 1, 2024 | In the News
A food fight erupted at the AI HW Summit earlier this year, where three companies all claimed to offer the fastest AI processing. All were faster than GPUs. Now Cerebras has claimed insanely fast AI performance with their latest software running on the company’s...