19:11 Business & Strategy2 days ago Your Agent Can Now Train Models — Merve Noyan, Hugging Face Merve Noyan from the Hugging Face open-source team delivers a broad survey of the current open-model landscape alongside several firs... 0 comments 1.9K views
22:54 Tutorials4 days ago This 100% uncensored AI model is insane… let’s run it David Ondrej walks through the rationale, setup, and practical use of uncensored large language models running locally in 2026. The v... 0 comments 25.7K views
11:12 Benchmarks5 days ago Qwen3.6 27B Gets 20% Faster with MTP and llama.cpp Locally Fahd Mirza demonstrates how to enable multi-token prediction (MTP) on Qwen3.6 27B using ik_llama.cpp — a community fork of the popula... 0 comments 3.3K views
09:01 Coding & Dev Tools2 weeks ago Running a 27B model at 130 tokens sec on a single GPU Locally with Luce DFlash LlamaDeFlash is a custom inference engine built from scratch in C++ and CUDA — no vLLM, no llama.cpp, no Python in the critical path... 0 comments 7.7K views
14:53 Coding & Dev Tools3 weeks ago This Mutant AI Model Should Not Exist: Qwopus-GLM-18B-Merged Locally Fahd Mirza walks through the creation and live testing of Qwopus-GLM-18B-Merged, a community-built model that stitches together two s... 0 comments 1.4K views
09:08 Tutorials3 weeks ago Open WebUI Desktop App – Install on Linux, Windows & Mac Open WebUI has shipped its first native desktop application for Windows, macOS, and Linux, and Fahd Mirza walks through the complete... 0 comments 1.2K views
15:26 Business & Strategy4 weeks ago Gemma, DeepMind’s Family of Open Models — Omar Sanseviero, Google DeepMind Omar Sanseviero, a researcher at Google DeepMind, delivers the first public conference talk on Gemma 4 just one week after its releas... 0 comments 2.1K views
15:26 Business & Strategy4 weeks ago Gemma, DeepMind’s Family of Open Models — Omar Sanseviero, Google DeepMind Omar Sanseviero, a researcher at Google DeepMind, delivers the first public conference talk on Gemma 4 just one week after its releas... 0 comments 6K views
14:56 Coding & Dev Tools1 month ago MiniMax M2.7 Running Locally on CPU + GPU – Everyone Can Do It Fahd Mirza walks through the complete process of running MiniMax M2.7 — a newly open-sourced 229-billion-parameter mixture-of-experts... 0 comments 2.7K views
11:54 Tutorials1 month ago Run GLM-5.1 Locally on CPU + GPU Easily: Step-by-Step Tutorial Fahd Mirza demonstrates how to run GLM-5.1 — the newly open-sourced flagship agentic model from Zhipu AI's GLM team — locally on a si... 0 comments 4.8K views