Trending

Content tagged with "transformers"

transformers

Hacker News

Top stories from the Hacker News community• Updated less than a minute ago

HN

Context engineering

chrisloy.dev
70
37
chrisloy
about 1 month ago
HN

Llamafile Returns

blog.mozilla.ai
112
19
aittalam
about 1 month ago
HN

Bertie the Brain

en.wikipedia.org
HN

A Look at Antml: The Anthropic Markup Language

karashiiro.leaflet.pub
33
9
ko_pivot
about 1 month ago

InfoQ

Latest articles from InfoQ• Updated 12 minutes ago

InfoQ

OpenAI's New GPT-5.1 Models are Faster and More Conversational

OpenAI recently released upgrades to their GPT-5 model. GPT‑5.1 Instant, the default chat model, has improvements to instruction following. GPT‑5.1 Thinking, the reasoning model, is faster and gives more understandable responses. GPT‑5.1-Codex-Max, the coding model, is trained to use compaction to perform long-running tasks. By Anthony Alford

infoq.com
Anthony Alford
3 days ago

Reddit

Top posts from tech subreddits• Updated 12 minutes ago

Hugging Face Trending

Popular models from Hugging Face• Updated 12 minutes ago

GitHub Trending

Popular repositories from GitHub• Updated 26 minutes ago

Megatron-LM

Ongoing research training transformer models at scale

whisper

Robust Speech Recognition via Large-Scale Weak Supervision

Swin-Transformer

This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".

transformers

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

whisper.cpp

Port of OpenAI's Whisper model in C/C++

awesome-chatgpt-prompts

This repo includes ChatGPT prompt curation to use ChatGPT and other LLM tools better.

pytorch-image-models

The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more

flash-attention

Fast and memory-efficient exact attention