sigmoid.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A social space for people researching, working with, or just interested in AI!

Server stats:

593
active users

Which B/16 reigns supreme? I've recently fine-tuned quite a few new ViT models and wanted to compare them. With new multi-weight support on the way I realized timm will soon have ~20 different B/16 (or close to). B/16 is the most common ViT model and easiest to compare across wide range of pretrain datasets and methods. In the lead is BEiT v2, but hot on its heels are fine-tuned LAION2B and OpenAI CLIP image towers. Check out a notebook at colab.research.google.com/driv

Ross Wightman

This is also my first time experimenting with ImageNet-X (facebookresearch.github.io/ima), a lot to unpack here, but I hope to explore more models in timm further with this soon...

And finally, many of these models are already in timm, but the CLIP tower weights are currently being added. Check out the progress in github.com/rwightman/pytorch-i and watch the hub huggingface.co/timm