Gemma 4 31B Abliterated — quality-preserving guardrail removal for Google's most capable open model. Apache 2.0. Runs on Apple Silicon via MLX.
-
Updated
Apr 3, 2026 - Python
Gemma 4 31B Abliterated — quality-preserving guardrail removal for Google's most capable open model. Apache 2.0. Runs on Apple Silicon via MLX.
Benchmark Gemma 4 E2B on Apple Silicon: MLX (mlx-lm) vs GGUF (llama-server), with TTFT, tokens/sec, and memory.
Drop-in KV cache compression for MLX on Apple Silicon. Brings PolarQuant (Google, ICLR 2026) to mlx-lm with first-class Gemma 4 support: MatFormer, dual head_dim, hybrid sliding/global attention, cross-layer KV sharing. 3-bit → 4.8× smaller cache, 0.995 logit cosine @ 4-bit.
Hands-on benchmark of Google Gemma 4 in LM Studio, with prompts, Gradio evaluator, saved outputs, and a published companion report.
Agentic Coding MCP Server: Local Gemma 4 powered coding assistant integrated with Zed editor
Nairobi-first edge AI simulation for urban flood resilience. Orchestrating multimodal agents (Gemini 2.5/Gemma 4) to navigate the Mbagathi Basin during infrastructure failure. Offline-first, spatial-aware, and mission-critical.
Add a description, image, and links to the gemma-4 topic page so that developers can more easily learn about it.
To associate your repository with the gemma-4 topic, visit your repo's landing page and select "manage topics."