2024.08.27
OpenAI, Intel, and Qualcomm talk AI compute at legendary Hot Chips conference
The science and engineering of making chips dedicated to processing artificial intelligence is as vibrant as ever, judging from a well-attended chip conference taking place this week at Stanford University called Hot Chips.
The Hot Chips show, currently in its 36th year, draws 1,500 attendees, just over half of whom participate via the online live feed and the rest at Stanford's Memorial Auditorium. For decades, the show has been a hotbed for discussion of the most cutting-edge chips from Intel, AMD, IBM, and many other vendors, with companies often using the show to unveil new products.
Also: Linus Torvalds talks AI, Rust adoption, and why the Linux kernel is 'the only thing that matters'
This year's conference received over a hundred submissions for presentation from all over the world. In the end, 24 talks were accepted, about as many as would fit in a two-day conference format. Two tutorial sessions took place on Sunday, with a keynote on Monday and Tuesday. There are also thirteen poster sessions.
The tech talks onstage and the poster presentations are highly technical and oriented toward engineers. The audience tends to spread out laptops and multiple screens as if spending the sessions in their personal offices.
Monday morning's session, featuring presentations from Qualcomm about its Oryon processor for the data center and Intel's Lunar Lake processor, drew a packed crowd and elicited plenty of audience questions.
In recent years, a big focus has been on chips designed to run neural network forms of AI better. This year's conference included a keynote by OpenAI's Trevor Cai, the company's head of hardware, about "Predictable scaling and infrastructure."
OpenAI infrastructure engineer Trevor Cai on the predictable scaling benefits of increasing computing power that have been OpenAI's focus since the beginning.
Cai, who has spent his time putting together OpenAI's compute infrastructure, said ChatGPT is the result of the company "spending years and billions of dollars predicting the next word better." That led to successive abilities such as "zero-shot learning."
"How did we know it would work?" Cai asked rhetorically. Because there are "scaling laws" that show ability can predictably increase as a "power law" of the compute used. Every time computing is doubled, the accuracy gets close to an "irreducible" entropy, he explained.
Other
Qualcomm Announces Snapdragon 7s Gen 3 For Mid-Range Devices With AI Capabilities
2024.08.21
SK Hynix Preps Large-Scale DRAM Price Hike, DDR5 Up To 20% Expensive
2024.08.22
Qualcomm, Motorola, Rohde & Schwarz show 5G Broadcast innovation
2024.08.22
SK Hynix Is Developing Next-Gen HBM With 30x Performance Uplift
2024.08.23
Samsung Completes NVIDIA's Quality Test for HBM3E Memory, Begins Shipments
2024.09.04
TSMC supplier says AI chip market growth to accelerate, dismisses Nvidia wipeout
2024.09.05
SK Hynix to start mass producing HBM3E 12-layer chips this month
2024.09.05
Qualcomm working on mixed-reality glasses with Google, Samsung
2024.09.06