AI news

Browse stored weekly and monthly summaries for this subject.

Summary

Generated about 22 hours ago.

TL;DR: March’s AI news centered on (1) scaling and governance—policy councils, safety evaluations, and automated research, (2) agent tooling plus reliability/security lessons, and (3) compute constraints and rising edge-hardware demand.

Policy, governance & safety

  • The U.S. President’s new science council (PCAST) is heavily weighted toward tech billionaires, with AI, quantum info, and nuclear as key areas.
  • Multiple reports highlight risks as AI agents grow more autonomous:
    • A red-teaming study (“Agents of Chaos”) documents real failures with persistent, tool-using LLM agents.
    • A Nature piece describes progress toward end-to-end automation of the AI research pipeline.
    • A Stanford arXiv paper flags evaluation gaps: vision-language models can invent plausible content for unseen images.

Agents, model releases & tooling

  • Anthropic’s Claude Code saw controversy and operational friction: a source-code leak allegation, usage-limit complaints, and discussion of mitigation approaches.
  • New/ongoing agent infrastructure themes included browser/agent runtimes (e.g., Rust-based “Pardus Browser”), containerized agent environments (“Coasts”), and local/Apple-Silicon inference previews (Ollama on MLX).
  • Model releases: Cohere launched Transcribe (open-source ASR); Google released TimesFM (200M time-series model, 16k context).

Compute & market signals

  • Semiconductor capacity constraints: TSMC is reportedly booked through 2028 for leading-edge nodes; downstream impact may affect advanced GPU/CPU availability.
  • Edge demand rose: Raspberry Pi profit increased, attributed to AI-driven use cases.
  • Market narrative: coverage noted a “sudden fall” in momentum for one of OpenAI’s most-hyped products, alongside broader commentary on how AI and bots are changing online activity.

Stories

Google's 200M-parameter time-series foundation model with 16k context (github.com) AI

Google Research has released TimesFM, a pretrained time-series foundation model for forecasting, with an updated TimesFM 2.5 checkpoint. The newer version uses 200M parameters (down from 500M), extends context length to 16k, and adds continuous quantile forecasting up to a 1k horizon via an optional quantile head. The GitHub repo includes instructions and example code for running the model with PyTorch or Flax, along with notes about ongoing support updates.