Mistral reports on the environmental impact of LLMs

Working on a Programming Language in the Age of LLMs

Context Rot: How increasing input tokens impacts LLM performance

Enhancing COBOL Code Explanations: A Multi-Agents Approach Using LLMs

Evaluating publicly available LLMs on IMO 2025

How the Free Software Foundation battles the LLM bots

Bad Actors Are Grooming LLMs to Produce Falsehoods

Coding with LLMs in the summer of 2025 – an update

Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs

Do LLMs identify fonts?

LLM Inference Handbook

Any-LLM: A unified API to access any LLM provider

Show HN: Price Per Token – LLM API Pricing Data

Why LLMs Can't Write Q/Kdb+: Writing Code Right-to-Left

LLMs exploit our tolerance for sloppiness

Show HN: Any-LLM – Lightweight router to access any LLM Provider

Ask HN: What's Your Useful Local LLM Stack?

VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention

I'm Building LLM for Satellite Data EarthGPT.app

Everything around LLMs is still magical and wishful thinking

Smollm3: Smol, multilingual, long-context reasoner LLM

Overclocking LLM Reasoning: Monitoring and Controlling LLM Thinking Path Lengths

A non-anthropomorphized view of LLMs

LLMs should not replace therapists

The Sparse Frontier: Sparse Attention Trade-offs in Transformer LLMs

Optimizing Tool Selection for LLM Workflows with Differentiable Programming

The Right Way to Embed an LLM in a Group Chat

Local LLM Tool Calling: Which LLM Should You Use?

LLM-assisted writing in biomedical publications through excess vocabulary

LLMs as Compilers

More →