Trending

Content tagged with "llm"

llm

Hacker News

Top stories from the Hacker News community• Updated 14 minutes ago

InfoQ

Latest articles from InfoQ• Updated 11 minutes ago

InfoQ

Article: NextGen Search - Where AI Meets OpenSearch Through MCP

In this article, authors Srikanth Daggumalli and Arun Lakshmanan discuss next-generation context-aware conversational search using OpenSearch and AI agents powered by Large Language Models (LLMs) and Model Context Protocol (MCP). By Srikanth Daggumalli, Arun Lakshmanan

infoq.com
InfoQ

TornadoVM 2.0 Brings Automatic GPU Acceleration and LLM support to Java

The TornadoVM project recently reached version 2.0, a major milestone for the open-source project that aims to provide a heterogeneous hardware runtime for Java. The project automatically accelerates Java programs on multi-core CPUs, GPUs, and FPGAs. This release is likely to be of particular interest to teams developing LLM solutions on the JVM. By Ben Evans

infoq.com
InfoQ

Podcast: Building a More Appealing CLI for Agentic LLMs Based on Learnings from the Textual Framework

Will McGugan, the maker of Textual and Rich frameworks, speaks about the reasoning of developing the two two libraries and the lesson learned. Also, he shares light on Toad, his current project, which he envisions being a more visually appealing way of interacting with agentic LLMs through command line. By Will McGugan

infoq.com
InfoQ

Podcast: Platform Engineering for AI: Scaling Agents and MCP at LinkedIn

QCon AI New York Chair Wes Reisz talks with LinkedIn’s Karthik Ramgopal and Prince Valluri about enabling AI agents at enterprise scale. They discuss how platform teams orchestrate secure, multi-agentic systems, the role of MCP, the use of foreground and background agents, improving developer experience, and reducing toil. By Karthik Ramgopal, Prince Valluri

infoq.com
InfoQ

OpenAI's New GPT-5.1 Models are Faster and More Conversational

OpenAI recently released upgrades to their GPT-5 model. GPT‑5.1 Instant, the default chat model, has improvements to instruction following. GPT‑5.1 Thinking, the reasoning model, is faster and gives more understandable responses. GPT‑5.1-Codex-Max, the coding model, is trained to use compaction to perform long-running tasks. By Anthony Alford

infoq.com
Anthony Alford
10 days ago
InfoQ

Replit Introduces New AI Integrations for Multi-Model Development

Replit has introduced Replit AI Integrations, a feature that lets users select third-party models directly inside the IDE and automatically generate the code needed to run inference. By Daniel Dominguez

infoq.com

Reddit

Top posts from tech subreddits• Updated 29 minutes ago

Hugging Face Trending

Popular models from Hugging Face• Updated 11 minutes ago

GitHub Trending

Popular repositories from GitHub• Updated 25 minutes ago

RWKV-LM

RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.

Python
13,338
900