AI

Summary

Generated about 15 hours ago.

TL;DR: April 3 centered on open AI models, agentic/assistant tooling, and practical deployment/search improvements—plus a safety edge-case reminder from autonomous driving.

Model releases & open weights

  • Google DeepMind released Gemma 4 open models, positioning them for agentic workflows and multimodal (audio/vision) reasoning, with multilingual support and fine-tuning options, plus efficiency variants for edge/local use.

AI tooling & faster systems

  • Cursor 3 launched as an agent-first developer workspace, adding a unified agents sidebar, improved PR diffs/workflows, deeper navigation (LSP), and an integrated browser.
  • Mintlify replaced RAG with a “virtual filesystem” for its AI documentation assistant, mapping doc sections to an in-memory structure and reusing an existing Chroma index—reporting a major drop in session boot time.
  • Community projects/gists advanced local and search setups: Ollama + Gemma 4 setup on Mac mini, and TurboQuant-style vector search quantizing embeddings (2–4 bits) for smaller/faster indexes.

Safety & real-world training limits

  • WIRED reported alleged Waymo violations involving school-bus stop signals, underscoring the difficulty of reliably training self-driving systems for rare, hard-to-detect safety devices (under NTSB investigation).

Stories

A School District Tried to Help Train Waymos to Stop for School Buses (wired.com) AI

WIRED reports that Austin Independent School District officials alleged Waymo robotaxis repeatedly passed school buses while their stop arms and red lights were active, despite software updates and a federal recall. The district and Waymo also held a mid-December data-collection event meant to improve recognition of school-bus signals, but violations continued into January and are still under investigation by the NTSB. The incident highlights challenges in training self-driving systems to reliably handle hard-to-detect safety devices and rare edge cases.

We replaced RAG with a virtual filesystem for our AI documentation assistant (mintlify.com) AI

Mintlify says it replaced RAG-based retrieval in its AI documentation assistant with a “virtual filesystem” that maps docs pages and sections to an in-memory directory tree and files. The assistant’s shell-like commands (e.g., ls, cd, cat, grep) are intercepted and translated into queries against the existing Chroma index, with page reassembly from chunks, caching, and RBAC-based pruning of inaccessible paths. By avoiding per-session sandbox startup and reusing the already-running Chroma database, the team reports cutting session boot time from about 46 seconds to ~100 milliseconds and reducing marginal compute cost.

Understanding young news audiences at a time of rapid change (reutersinstitute.politics.ox.ac.uk) AI

The Reuters Institute report synthesizes more than a decade of research on how 18–24-year-olds access and think about news amid major media and technology change. It finds young audiences have shifted from news websites to social and video platforms, pay more attention to individual creators than news brands, and consume news less frequently and with less interest—often saying it is irrelevant or hard to understand. The study also highlights greater openness to AI for news, alongside continued concerns about fairness and perceived impartiality, and it concludes publishers need to rethink both distribution and news relevance for younger people.

Cursor 3 (cursor.com) AI

Cursor has released Cursor 3, a redesigned, agent-first workspace intended to make it easier to manage work across multiple repositories and both local and cloud agents. The update adds a unified agents sidebar (including agents started from tools like GitHub and Slack), faster switching between local and cloud sessions, and improved PR workflows with a new diffs view. It also brings deeper code navigation (via full LSPs), an integrated browser, and support for installing plugins from the Cursor Marketplace.

Google releases Gemma 4 open models (deepmind.google) AI

Google DeepMind has released Gemma 4, a set of open models intended for building AI applications. The page highlights capabilities such as agentic workflows, multimodal (audio/vision) reasoning, multilingual support, and options for fine-tuning. It also describes efficiency-focused variants for edge devices and local use, along with safety and security measures and links to download the model weights via multiple platforms.

Show HN: TurboQuant for vector search – 2-4 bit compression (github.com) AI

Show HN spotlights py-turboquant (turbovec), an unofficial implementation of Google’s TurboQuant vector-search method that compresses high-dimensional embeddings to 2–4 bits per coordinate using a data-oblivious random rotation and math-derived Lloyd-Max quantization. The project is implemented in Rust with Python bindings via PyO3 and emphasizes zero training and fast indexing. Benchmarks on Apple Silicon and x86 compare favorably to FAISS (especially at 4-bit) in speed while achieving comparable or better recall, with much smaller index sizes than FP32.

ESP32-S31: Dual-Core RISC-V SoC with Wi-Fi 6, Bluetooth 5.4, and Advanced HMI (espressif.com) AI

Espressif announced the upcoming ESP32-S31, a dual-core 32-bit RISC-V SoC combining Wi‑Fi 6, Bluetooth 5.4 (including LE Audio and mesh), and IEEE 802.15.4 for Thread/Zigbee, plus a 1Gbps Ethernet MAC. The chip targets next-generation IoT devices with a 320MHz core, multimedia-oriented HMI features (camera/display/touch and graphics acceleration), security hardware (secure boot, encryption, side-channel and glitch protections, and TEE), and support for ESP-IDF and Matter-related frameworks.

Show HN: Apfel – The free AI already on your Mac (apfel.franzai.com) AI

Show HN project Apfel presents a free, on-device AI for macOS Apple Silicon that exposes Apple’s built-in language model as a terminal CLI, an OpenAI-compatible local HTTP server, and an interactive chat. The tool is designed to run inference locally with no API keys or network calls, and it supports features like streaming and JSON output for use with existing OpenAI client libraries. The post also highlights related companion tools in the “apfel family,” such as a GUI and clipboard-based actions.

A Recipe for Steganogravy (theo.lol) AI

The article describes a Python CLI concept for “steganogravy,” using neural linguistic steganography to hide a small payload in the introduction text of AI-generated recipe blog posts. It explains the basic arithmetic-coding approach, the need for encoder/decoder to match model settings and prompts, and practical limitations like inefficiency and tokenization divergence. The author also notes a filtering method to prevent decoding failures and illustrates recovery of a hidden message from the generated text.

April 2026 TLDR Setup for Ollama and Gemma 4 26B on a Mac mini (gist.github.com) AI

The gist provides a step-by-step guide for running Ollama on an Apple Silicon Mac mini, pulling the Gemma 4 12B model, and configuring it to start automatically with the model preloaded and kept alive. It includes commands to verify GPU/CPU usage, create a launch agent to periodically “warm” the model, and set OLLAMA_KEEP_ALIVE to prevent unloading due to inactivity. It also notes relevant Ollama updates such as the MLX backend and summarizes key memory considerations for a 24GB system.