Benjamin Carr, Ph.D. 👨🏻💻🧬<p><a href="https://hachyderm.io/tags/Frontier" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Frontier</span></a> trained a <a href="https://hachyderm.io/tags/ChatGPT" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ChatGPT</span></a>-sized <a href="https://hachyderm.io/tags/largelanguagemodel" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>largelanguagemodel</span></a> with only 3,000 of its 37,888 <a href="https://hachyderm.io/tags/Radeon" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Radeon</span></a> <a href="https://hachyderm.io/tags/GPU" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GPU</span></a> — the world's fastest <a href="https://hachyderm.io/tags/supercomputer" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>supercomputer</span></a> blasts through one trillion parameter model with only 8 percent of its <a href="https://hachyderm.io/tags/MI250X" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>MI250X</span></a> GPUs<br>The paper notes that the key challenge in training such a large <a href="https://hachyderm.io/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> is the amount of memory required, which was 14 terabytes at minimum. <a href="https://www.tomshardware.com/tech-industry/supercomputers/frontier-trained-a-chatgpt-sized-large-language-model-with-only-3000-of-its-37888-radeon-gpus-the-worlds-fastest-supercomputer-blasts-through-one-trillion-parameter-model-with-only-8-percent-of-its-mi250x-gpus" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">tomshardware.com/tech-industry</span><span class="invisible">/supercomputers/frontier-trained-a-chatgpt-sized-large-language-model-with-only-3000-of-its-37888-radeon-gpus-the-worlds-fastest-supercomputer-blasts-through-one-trillion-parameter-model-with-only-8-percent-of-its-mi250x-gpus</span></a></p>