sigmoid.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A social space for people researching, working with, or just interested in AI!

Server stats:

597
active users

#fp8

0 posts0 participants0 posts today
Hacker News<p>FP8 is ~100 tflops faster when the kernel name has "cutlass" in it</p><p><a href="https://twitter.com/cis_female/status/1943069934332055912" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">twitter.com/cis_female/status/</span><span class="invisible">1943069934332055912</span></a></p><p><a href="https://mastodon.social/tags/HackerNews" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HackerNews</span></a> <a href="https://mastodon.social/tags/FP8" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP8</span></a> <a href="https://mastodon.social/tags/tflops" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>tflops</span></a> <a href="https://mastodon.social/tags/cutlass" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>cutlass</span></a> <a href="https://mastodon.social/tags/performance" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>performance</span></a> <a href="https://mastodon.social/tags/optimization" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>optimization</span></a> <a href="https://mastodon.social/tags/AI" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AI</span></a></p>
Benjamin Carr, Ph.D. 👨🏻‍💻🧬<p><a href="https://hachyderm.io/tags/JackDongarra" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>JackDongarra</span></a> Makes a Stand for Traditional <a href="https://hachyderm.io/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a>: "US still doesn’t have a clear, long-term plan for what comes next.... U.S. risks falling behind."</p><p>Challenges to high-performance computing threaten <a href="https://hachyderm.io/tags/US" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>US</span></a> <a href="https://hachyderm.io/tags/innovation" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>innovation</span></a></p><p>The <a href="https://hachyderm.io/tags/AI" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AI</span></a> boom has led chip makers to focus on <a href="https://hachyderm.io/tags/FP16" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP16</span></a> and <a href="https://hachyderm.io/tags/FP8" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP8</span></a>, not the <a href="https://hachyderm.io/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> used by scientific research. If chip companies stop making the parts that <a href="https://hachyderm.io/tags/scientists" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>scientists</span></a> need, then it could become harder to do important research. <br><a href="https://theconversation.com/challenges-to-high-performance-computing-threaten-us-innovation-255188" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">theconversation.com/challenges</span><span class="invisible">-to-high-performance-computing-threaten-us-innovation-255188</span></a></p>
Hacker News<p>DeepSeek Open Sources DeepGEMM: Clean and efficient FP8 GEMM kernels — <a href="https://github.com/deepseek-ai/DeepGEMM" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="">github.com/deepseek-ai/DeepGEMM</span><span class="invisible"></span></a><br><a href="https://mastodon.social/tags/HackerNews" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HackerNews</span></a> <a href="https://mastodon.social/tags/DeepSeek" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DeepSeek</span></a> <a href="https://mastodon.social/tags/DeepGEMM" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DeepGEMM</span></a> <a href="https://mastodon.social/tags/FP8" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP8</span></a> <a href="https://mastodon.social/tags/AI" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AI</span></a> <a href="https://mastodon.social/tags/Kernels" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Kernels</span></a> <a href="https://mastodon.social/tags/OpenSource" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>OpenSource</span></a></p>
michabbb<p>Introducing Phind-405B and faster, high quality <a href="https://social.vivaldi.net/tags/AI" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AI</span></a> answers for everyone</p><p>🚀 Phind-405B: New flagship <a href="https://social.vivaldi.net/tags/llm" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>llm</span></a>, based on Meta Llama 3.1 405B, designed for programming &amp; technical tasks. <a href="https://social.vivaldi.net/tags/Phind405B" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Phind405B</span></a></p><p>⚡ 128K tokens, 32K context window at launch, 92% on HumanEval, great for web app design. <a href="https://social.vivaldi.net/tags/Programming" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Programming</span></a> <a href="https://social.vivaldi.net/tags/AIModel" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AIModel</span></a></p><p>💡 Trained on 256 H100 GPUs with FP8 mixed precision, 40% memory reduction. <a href="https://social.vivaldi.net/tags/DeepSpeed" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DeepSpeed</span></a> <a href="https://social.vivaldi.net/tags/FP8" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP8</span></a></p><p>⚡ Phind Instant Model: Super fast, 350 tokens/sec, based on Meta Llama 3.1 8B. <a href="https://social.vivaldi.net/tags/PhindInstant" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>PhindInstant</span></a></p><p>🚀 Runs on NVIDIA TensorRT-LLM with flash decoding, fused CUDA kernels. <a href="https://social.vivaldi.net/tags/NVIDIA" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>NVIDIA</span></a> <a href="https://social.vivaldi.net/tags/GPUs" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPUs</span></a></p><p>🔍 Faster Search: Prefetches results, saves up to 800ms latency, better embeddings. <a href="https://social.vivaldi.net/tags/FastSearch" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FastSearch</span></a></p><p>👨‍💻 Goal: Help developers experiment faster, new features coming soon! <a href="https://social.vivaldi.net/tags/DevTools" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DevTools</span></a> <a href="https://social.vivaldi.net/tags/Innovation" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Innovation</span></a></p><p><a href="https://www.phind.com/blog/introducing-phind-405b-and-better-faster-searches" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">phind.com/blog/introducing-phi</span><span class="invisible">nd-405b-and-better-faster-searches</span></a></p>
Charlie Blake<p>Glad to be on here! My <a href="https://sigmoid.social/tags/introduction" class="mention hashtag" rel="tag">#<span>introduction</span></a>:</p><p>I&#39;m an AI researcher in the UK, working at Graphcore - a semiconductor company who develop the <a href="https://sigmoid.social/tags/IPU" class="mention hashtag" rel="tag">#<span>IPU</span></a> (a <a href="https://sigmoid.social/tags/GPU" class="mention hashtag" rel="tag">#<span>GPU</span></a> alternative) 💻 I joined last year, having previously been at Oxford for my MSc.</p><p>My interests are in <a href="https://sigmoid.social/tags/numerics" class="mention hashtag" rel="tag">#<span>numerics</span></a> (especially <a href="https://sigmoid.social/tags/fp8" class="mention hashtag" rel="tag">#<span>fp8</span></a> 8️⃣), <a href="https://sigmoid.social/tags/LLMs" class="mention hashtag" rel="tag">#<span>LLMs</span></a>, mixture-of-expert models, and anything to do with <a href="https://sigmoid.social/tags/solitaire" class="mention hashtag" rel="tag">#<span>solitaire</span></a> ♣️ ♦️ </p><p>Thanks to <span class="h-card" translate="no"><a href="https://sigmoid.social/@thegradient" class="u-url mention">@<span>thegradient</span></a></span> for making this happen 😃</p>