Wink Pings

oLLM: Running 80B LLMs on 8GB VRAM, the Open-Source Community's Latest Brute-Force Hack

A lightweight Python library called oLLM enables consumer-grade GPUs to run massive language models without quantization or compromise, using SSD swap space to bypass VRAM limitations.

The open-source community has pulled off another audacious stunt. The oLLM Python library allows consumer-grade GPUs with just 8GB of VRAM (like the RTX 3060) to run 80B-parameter models without quantization—maintaining full fp16/bf16 precision.

The approach is brutally simple: offload model parameters that don't fit in VRAM onto an SSD, loading them on demand. It's like giving your GPU an external hard drive, except this drive must handle gigabytes of data transfer per second. While some raised concerns about SSD wear, the developers sidestepped the issue, suggesting users test with spare drives.

Currently supports behemoths like gpt-oss-20B and Qwen3-next-80B—even Llama-3.1-8B seems like a light snack in comparison. The GitHub repo gained 335 stars in just three days, with users reporting ~3-second latency for 80B models in conversational scenarios—slower than cloud APIs but ten times faster than CPU.

AMD users shouldn't celebrate yet—ROCm support remains nonexistent. Mac users are also left out. When asked how it differs from Ollama, developers emphasized oLLM's specialization in ultra-long-context scenarios, ideal for local processing of large documents.

The irony? This solution perfectly illustrates the "death of Moore's Law" narrative—as hardware progress slows, software resorts to unorthodox methods to squeeze every drop of performance. But let's be honest: why pay $200/month for cloud services when a $200 graphics card can do the job?

![A logo with the text "oLLM" in large blue letters, featuring a smiling emoji face and a lightning bolt. Below the logo, text reads "LLM Inference for Large-Context Offline Workloads" and lists model names like gpt-oss-20B, Qwen3-next-80B, and Llama-3.1-8B.](https://wink.run/image?url=https%3A%2F%2Fpbs.twimg.com%2Fmedia%2FG18wXIkbkAEkYc_%3Fformat%3Djpg%26name%3Dlarge)

![A screenshot of a GitHub repository page shows "Mega4alik/ollm" on the left and a red-hatted man's avatar on the right. Five icons below display metrics: Contributor (1), Issues (3), Discussions (3), Stars (335), and Forks (11). The white-background page uses primarily black text.](https://wink.run/image?url=https%3A%2F%2Fpbs.twimg.com%2Fcard_img%2F1971582037636399104%2FEqIm_8Hb%3Fformat%3Djpg%26name%3Dlarge)

发布时间: 2025-09-29 01:11