Trending
Content tagged with "transformers"
Hacker News
Top stories from the Hacker News community• Updated 10 minutes ago
InfoQ
Latest articles from InfoQ• Updated 7 minutes ago
Meta's Optimization Platform Ax 1.0 Streamlines LLM and System Optimization
Now stable, Ax is an open-source platform from Meta designed to help researchers and engineers apply machine learning to complex, resource-intensive experimentation. Over the past several years, Meta has used Ax to improve AI models, accelerate machine learning research, tune production infrastructure, and more. By Sergio De Simone
OpenAI's New GPT-5.1 Models are Faster and More Conversational
OpenAI recently released upgrades to their GPT-5 model. GPT‑5.1 Instant, the default chat model, has improvements to instruction following. GPT‑5.1 Thinking, the reasoning model, is faster and gives more understandable responses. GPT‑5.1-Codex-Max, the coding model, is trained to use compaction to perform long-running tasks. By Anthony Alford
Top posts from tech subreddits• Updated 7 minutes ago
Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters.
We cut agent token usage and speed by ~82% with one dumb trick: let AI use variables
Hugging Face Trending
Popular models from Hugging Face• Updated 7 minutes ago
GitHub Trending
Popular repositories from GitHub• Updated 22 minutes ago
awesome-chatgpt-prompts
Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.
Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
transformers
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
pytorch-image-models
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more