🚗 #GateSquareCommunityChallenge# Round 2 — Which coin is not listed on Gate Launchpad❓
Time to prove if you’re a true Gate veteran!
💰 Join the challenge — 5 lucky winners will share $50 in GT!
👉 How to participate:
1️⃣ Follow Gate_Square
2️⃣ Like this post
3️⃣ Comment with your answer
🗓️ Deadline: October 8, 2025, 24:00 (UTC+8)
Frontier, the world's largest supercomputer, uses 3,072 AMD GPUs to train over a trillion parameter LLMs
Bit News According to a report by New Zhiyuan on January 13, AMD's software and hardware systems can also train GPT-3.5 level large models.
Frontier, the world's largest supercomputer at Oak Ridge National Laboratory, is home to 37,888 MI250X GPUs and 9,472 Epyc7A53CPUs. Recently, researchers trained a GPT-3.5-scale model using only about 8% of those GPUs. The researchers successfully used the ROCM software platform to successfully break through many difficulties of distributed training models on AMD hardware, and established the most advanced distributed training algorithm and framework for large models on AMD hardware using the ROCM platform.
Successfully provides a feasible technical framework for efficient training of LLMs on non-NVIDIA and non-CUDA platforms.
After the training, the researchers summarized the experience of training large models on Frontier into a paper detailing the challenges encountered and overcome.