It’s typically mentioned that supercomputers of some a long time in the past pack much less energy than right this moment’s sensible watches. Now we have now an organization, Tiiny AI Inc., claiming to have constructed the world’s smallest private AI supercomputer that may run a 120-billion-parameter massive language mannequin on-device — with out cloud connectivity, servers or GPUs.
The corporate mentioned the Arm-based product is powered by two expertise advances that make large-parameter LLMs viable on a compact system.
- TurboSparse, a neuron-level sparse activation approach, improves inference effectivity whereas sustaining full mannequin intelligence.
- PowerInfer, an open-source heterogeneous inference engine with greater than 8,000 GitHub stars, accelerates heavy LLM workloads dynamically distributing computation throughout CPU and NPU, enabling sever-grade efficiency at a fraction of conventional energy consumption.
Designed for energy-efficient private intelligence, Tiiny AI Pocket Lab runs inside a 65W energy envelope and allows large-model efficiency at a fraction of the power and carbon footprint of conventional GPU-based methods, Tiiny mentioned.
The corporate mentioned the bottleneck in right this moment’s AI ecosystem is not computing energy — it’s dependence on the cloud.

“Cloud AI has introduced outstanding progress, but it surely additionally created dependency, vulnerability, and sustainability challenges,” mentioned Samar Bhoj, GTM Director of Tiiny AI. “With Tiiny AI Pocket Lab, we imagine intelligence shouldn’t belong to knowledge facilities, however to folks. This is the first step towards making superior AI really accessible, personal, and private, by bringing the energy of huge fashions from the cloud to each particular person system.”

“The system represents a significant shift in the trajectory of the AI trade,” the firm mentioned. “As cloud-based AI more and more struggles with sustainability considerations, rising power prices, international outages, the prohibitive prices of long-context processing, and rising privateness dangers, Tiiny AI introduces another mannequin centered on private, moveable, and totally personal intelligence.”
The system has been verified by Guinness World Data underneath the class The Smallest MiniPC (100B LLM Domestically), in accordance to the firm.
The corporate calls itself a “U.S.-based deep tech AI startup,” whereas its announcement has a Hong Kong dateline. The corporate was shaped in 2024 and brings collectively engineers from MIT, Stanford, HKUST, SJTU, Intel, and Meta, AI inference and {hardware}–software program co-design. Their analysis has been printed in tutorial conferences together with SOSP, OSDI, ASPLOS, and EuroSys. In 2025, Tiiny AI secured a multi-million greenback seed spherical from main international traders in accordance to the firm.
Tiiny AI Pocket Lab is designed to help main private AI use circumstances, serving builders, researchers, creators, professionals, and college students. It allows multi-step reasoning, deep context understanding, agent workflows, content material technology, and safe processing of delicate information — even with out web entry. The system additionally offers true long-term private reminiscence by storing person knowledge, preferences, and paperwork domestically with bank-level encryption, providing a degree of privateness and persistence that cloud-based AI methods can’t present.
Tiiny AI Pocket Lab operates in the ‘golden zone’ of private AI (10B–100B parameters), which satisfies greater than 80 p.c of real-world wants, in accordance to the firm. It helps fashions scaling up to 120B LLM, delivering intelligence ranges comparable to GPT-4o. This allows PhD-level reasoning, multi-step evaluation, and deep contextual understanding — however with the safety of totally offline, on-device processing.
The corporate mentioned the system helps one-click set up of widespread open-source fashions together with OpenAI GPT-OSS, Llama, Qwen, DeepSeek, Mistral, and Phi, and allows deployment of open-source AI agents resembling OpenManus, ComfyUI, Flowise, Presenton, Libra, Bella, and SillyTavern. Customers obtain steady updates, together with official OTA {hardware} upgrades. The above options will probably be launched at CES in January 2026.
Disclaimer: This article is sourced from external platforms. OverBeta has not independently verified the information. Readers are advised to verify details before relying on them.