Generated about 17 hours ago.
TL;DR: The day’s AI coverage spanned model efficiency research, offline/open-model assistants, and the growing “agent/tooling” governance debate around MCP and provider routing.
Model research & tooling
-
Extreme low-bit quantization: The Salomi GitHub repo tests near-binary transformer quantization for GPT-2–class models and reports that strict 1.00 bpp post-hoc binary quantization doesn’t hold up; more credible results appear around ~1.2–1.35 bpp using approaches like Hessian-guided vector quantization and mixed precision.
-
Offline assistant on-device: AbodeLLM brings an Android AI assistant that runs open models (e.g., LLaMA, DeepSeek) locally, with optional multimodal inputs and expert controls.
-
Model marketplace/routing: OpenRouter lists Arcee AI’s “Trinity Large Thinking” with pricing and routing/fallback behavior across providers.
Agents, eval, and data governance
-
MCP trust gaps: A critique of Perplexity’s MCP stance argues the main issue isn’t just token overhead, but missing trust-aware controls for sensitive data after authorization (suggesting sensitivity metadata, trust-tier registries, and runtime enforcement).
-
Eval rigor: “The revenge of the data scientist” warns that LLM harnessing often repeats data-science pitfalls—weak experimental design, unreliable judges, and poor validation.
Politics and broader discourse
-
AI forecasts: Kronaxis claims a synthetic-voter method predicts UK local elections with ~75% winner accuracy on limited by-election validation, with country-level seat predictions.
-
Meta-critique of tech elites: The Nation argues Silicon Valley leaders promote an anti-intellectual narrative that dismisses deep-learning thinking while profiting from it.