2 articles found
NVIDIA's controversial research argues that tiny language models outperform giant LLMs for agentic tasks and they're about to flip the AI industry on its head
A 9B-parameter model achieving six times the throughput of a 70B-parameter competitor raises questions about architectural innovation versus hardware dependency.