Trending
Content tagged with "transformers"
Hacker News
Top stories from the Hacker News community• Updated 7 minutes ago
InfoQ
Latest articles from InfoQ• Updated 16 minutes ago
Transformers v5 Introduces a More Modular and Interoperable Core
Hugging Face has released the first candidate for Transformers v5, marking a significant evolution from v4 five years ago. The library has grown from a specialized model toolkit to a critical resource in AI development, achieving over three million installations daily and more than 1.2 billion total installs. By Robert Krzaczyński
Meta's Optimization Platform Ax 1.0 Streamlines LLM and System Optimization
Now stable, Ax is an open-source platform from Meta designed to help researchers and engineers apply machine learning to complex, resource-intensive experimentation. Over the past several years, Meta has used Ax to improve AI models, accelerate machine learning research, tune production infrastructure, and more. By Sergio De Simone
OpenAI's New GPT-5.1 Models are Faster and More Conversational
OpenAI recently released upgrades to their GPT-5 model. GPT‑5.1 Instant, the default chat model, has improvements to instruction following. GPT‑5.1 Thinking, the reasoning model, is faster and gives more understandable responses. GPT‑5.1-Codex-Max, the coding model, is trained to use compaction to perform long-running tasks. By Anthony Alford
Top posts from tech subreddits• Updated 16 minutes ago
Hugging Face Trending
Popular models from Hugging Face• Updated 34 minutes ago
GitHub Trending
Popular repositories from GitHub• Updated about 1 hour ago
awesome-chatgpt-prompts
Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.
Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
transformers
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
pytorch-image-models
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more