Vera – A language designed for machines to write
(veralang.dev)
AI
Vera is a new programming language aimed at having LLMs write code by prioritizing “checkability” over human convenience. It requires explicit function contracts and typed effects, replaces variable names with structural (De Bruijn-style) references, and uses static and runtime verification (e.g., via Z3) to catch problems early. The article also notes that Vera supports common agent-facing tasks like typed JSON/HTTP/LLM calls, compiles to WebAssembly, and reports early benchmark results (VeraBench) suggesting some models produce more correct Vera than comparable languages.
Show HN: I built a local data lake for AI powered data engineering and analytics
(stream-sock-3f5.notion.site)
AI
The post announces “Nile Local,” a fully local AI data IDE and data stack aimed at reducing the setup overhead of big-data engineering. It combines a local data lake/storage layer, Spark compute for interactive query testing, and automated “zero-ETL” ingestion with lineage and versioning. The tool also includes AI-assisted analytics and query writing using embedded or cloud LLMs so users can work and iterate on data locally before deploying to the cloud.
Finetuning Activates Verbatim Recall of Copyrighted Books in LLMs
(arxiv.org)
AI
A new arXiv preprint argues that model finetuning can reactivate verbatim memorization of copyrighted books in major LLMs. The authors claim that training models to expand plot summaries into full text enables GPT-4o, Gemini-2.5-Pro, and DeepSeek-V3.1 to reproduce large portions of held-out copyrighted books, even when prompted only with semantic descriptions rather than any book text. They report the effect generalizes across authors and even across different models and providers, suggesting an industry-wide vulnerability beyond common alignment measures like RLHF and output filtering.
The Hormuz chokehold affects AI funding too
(highabsolutevalue.substack.com)
AI
The article argues that disruptions to shipping through the Strait of Hormuz could reduce Gulf sovereign wealth funds’ profits and slow their rapidly growing AI investments. It cites estimates that Gulf funds were major sources of AI funding in 2025, but notes that if cross-strait crossings remain sharply lower, less capital may be available for late-2026 and 2027 mega rounds. The piece suggests this could also strengthen Big Tech by limiting competition from better-capitalized regional investors.
Claude Managed Agents Overview
(platform.claude.com)
AI
Anthropic’s Claude Managed Agents is a pre-built, configurable agent harness that runs in managed infrastructure, avoiding the need to build your own agent loop and tool runtime. It lets Claude operate within a defined environment (container template, packages, network access) and run long-running, tool-heavy tasks with persisted session history, streaming events, and the ability to steer or interrupt execution mid-run. The docs outline core concepts (agent, environment, session, events), available built-in tools (e.g., Bash, file operations, web search/fetch, MCP connectors), and note that the feature is currently in beta with a required beta header, rate limits, and branding guidelines for partners.
Giving LLMs a Formal Reasoning Engine for Code Analysis
(yogthos.net)
AI
The article describes Chiasmus, an MCP server that augments LLM code assistants with formal reasoning by parsing source code with tree-sitter, converting structure into Prolog facts, and using Tau Prolog (and Z3 for constraint solving) to answer questions like transitive reachability, dead code, cycles, and impact with proof-like certainty. It argues that this approach avoids the token-heavy, error-prone limitations of repeatedly “grepping” and reconstructing call chains via ad hoc search. It also notes the same architecture can verify properties of diagrams (e.g., Mermaid) and certain constraint problems beyond code.
AI Did It in 12 Minutes. It Took Me 10 Hours to Fix It
(idiallo.com)
AI
Ibrahim Diallo describes using an AI LLM to generate a PHP media-upload manager for his blog, which initially produced “spaghetti code” and several integration bugs. After getting it working, he spent about 10 more hours rewriting and restructuring the project to make it maintainable, cutting the codebase from roughly 5,000 lines to about 1,254. He concludes that while AI can speed up first drafts, cleanup and understanding the code still take substantial effort.
The World Needs More Software Engineers – Box CEO
(oreilly.com)
AI
In a conversation with O’Reilly, Box CEO Aaron Levie argues that AI may increase the demand for software engineering by making more kinds of projects economically feasible across the broader enterprise. The article also suggests that while interoperability standards help systems connect, the tougher challenge is organizing data and context so AI agents can reliably find what they need. Levie predicts a sustained push toward infrastructure modernization, with companies needing to clean up and structure their data to benefit from agents.
AMD AI director says Claude Code is becoming dumber and lazier since update
(theregister.com)
AI
AMD AI director Stella Laurenzo says her team’s long-running use of Anthropic’s Claude Code has degraded since February, with evidence from thousands of sessions suggesting the tool is ending early, avoiding ownership, and reading less before making edits. She links the changes to “thinking” content redaction introduced around Claude Code version 2.1.69 and asks Anthropic for transparency about whether it is reducing or capping thinking tokens and to expose thinking-token counts per request. Laurenzo says her team is switching to another provider and warns Anthropic may lose ground unless the behavior is fixed.
The AI Great Leap Forward
(leehanchung.github.io)
AI
The article argues that many corporate “AI transformation” efforts mirror China’s Great Leap Forward: top-down mandates push teams to ship impressive but poorly validated “AI” outputs, while metrics and incentives encourage inflated claims. It warns that eliminating people and processes (middle managers, QA, documentation, operations knowledge) creates second-order failures once real-world edge cases appear, and that attempts to “distill” expertise into agent skills can backfire by making workers strategically indispensable. Overall, it calls for evaluation, data, monitoring, and maintainability rather than demos and paperwork.
Show HN: Skrun – Deploy any agent skill as an API
(github.com)
AI
Skrun is an open-source tool that turns an “agent skill” (defined in SKILL.md) into a callable API endpoint using a POST /run interface. It supports typed inputs/structured outputs, multi-model backends with fallback (e.g., Anthropic, OpenAI, Google, Mistral, Groq), and stateful agent runs via stored key-value data. The project ships with a local runtime and includes CLI commands to init, develop, test, package, and deploy agents, with an architecture intended for cloud deployment.
Databricks co-founder wins prestigious ACM award, says 'AGI is here already'
(techcrunch.com)
AI
Databricks co-founder and CTO Matei Zaharia won the ACM Prize in Computing, with the award highlighting his work on Spark and related contributions. He said he believes AGI is already here in a limited form, arguing people should avoid judging AI by human standards. Zaharia also discussed how AI could better support research and engineering, while warning that current agent systems can introduce security risks.
The demise of software engineering jobs has been greatly exaggerated
(cnn.com)
AI
CNN argues that claims that AI will wipe out software engineering jobs are overstated. The article says developer postings are still rising and that AI is changing what engineers do—shifting work from routine coding to overseeing AI-generated code and focusing more on design and customer problems. It also notes a short-term transition that may be difficult for some workers, as companies cut costs and require engineers to keep learning new skills.
Show HN: Voxcode: local speech to text and ripgrep = transcript and code context
(github.com)
AI
Voxcode is an open-source macOS app for local speech-to-text tailored to coding agents: you select code in your editor, speak instructions, and it pastes the transcript back with a ripgrep-style file/line reference (or the selected snippet when exact lines can’t be resolved). The project indexes local git repositories and uses an optimized parallel file-walking approach plus a local ONNX Runtime transcription model (Parakeet TDT) to keep searches and transcription fast. It’s designed to work across IDEs and agent tools without direct integration by operating purely on clipboard/paste and filesystem context.
Claude Managed Agents
(claude.com)
AI
Anthropic announced Claude Managed Agents, a set of composable APIs meant to help developers deploy cloud-hosted AI agents faster by handling production concerns like secure sandboxing, long-running sessions, permissions, and tracing. The company says teams can move from prototype to launch in days instead of months and that Managed Agents are available in public beta on the Claude Platform, with multi-agent coordination in research preview.
Façade (2005 Video Game)
(en.wikipedia.org)
AI
Façade is a 2005 interactive drama in which the player converses via text with an AI-driven married couple, Trip and Grace, in an open-ended story about their deteriorating relationship. Built using natural-language processing and an AI “behavior language,” the game was praised at launch for its conversational design and storytelling ambitions, and later developed a cult following fueled by awkward moments from its reactions. A planned sequel, The Party, was paused after 2013 and later resumed in 2024.
A Digital Compute-in-Memory Architecture for NFA Evaluation
(dl.acm.org)
AI
The paper proposes a compute-in-memory digital architecture aimed at evaluating nondeterministic finite automata (NFA) more efficiently by performing computation closer to where data is stored, reducing data movement overhead compared with conventional approaches.
Show HN: TUI-use: Let AI agents control interactive terminal programs
(github.com)
AI
The GitHub project tui-use ("Show HN") proposes a way for AI agents to operate interactive terminal programs by running them in a PTY, rendering the screen with a headless xterm emulator, and sending keystrokes based on clean screen snapshots (including TUI selection “highlights”). It targets use cases like REPL sessions, CLI wizards, database CLIs, SSH-driven interactive workflows, and full-screen TUIs such as vim/htop/lazygit, with a command-line interface plus plugins for agents like Claude Code. The article notes it works on Unix-like systems and strips most terminal styling, relying on metadata to identify active selections.
Digital Hopes, Real Power: How the Arab Spring Fueled a Global Surveillance Boom
(eff.org)
AI
The EFF argues that the 2011 Arab uprisings’ digital tactics spurred a global surveillance industry: governments upgraded monitoring, expanded cybercrime and protest-related laws to criminalize dissent, and relied on spyware markets to hack targets at scale. It also describes how biometrics, facial recognition, and “smart city” systems helped normalize automated tracking and risk profiling, including in migration and humanitarian settings. The piece warns that these tools and legal frameworks—often sold without meaningful safeguards—have been exported beyond the Middle East to support digital authoritarianism worldwide.
Meta debuts Muse Spark, first AI model under Alexandr Wang
(axios.com)
AI
Meta has launched Muse Spark, a new homegrown AI model (code-named Avocado) built over nine months under Alexandr Wang’s leadership, aimed at narrowing performance gaps with top rivals. The text-only model takes voice, text, and images as inputs and will power queries in the Meta AI app and on Meta.ai, with plans to expand across Facebook, Instagram, and WhatsApp, alongside several “modes” including a shopping mode. Meta says all versions will be free with possible rate limits and plans an open-source licensed release, while noting privacy rules may allow broad use of data shared with its AI systems.
Muse Spark: Scaling Towards Personal Superintelligence
(ai.meta.com)
AI
Meta introduced Muse Spark, a multimodal reasoning model designed to support tool use, visual “chain-of-thought,” and multi-agent orchestration. The company says it will be available on meta.ai and via a private API preview, with “Contemplating mode” rolling out gradually. Meta also outlines how it improved compute efficiency through changes to pretraining, reinforcement learning, and test-time reasoning, and it reports safety evaluations showing strong refusal behavior in high-risk scientific domains.
Muse Spark – Meta Superintelligence Labs
(meta.ai)
AI
The linked page appears to introduce “Muse Spark” and Meta’s “Superintelligence Labs,” describing the lab’s mission and related research efforts, though no specific article text was available to summarize in detail.
The Future of Everything is Lies, I Guess
(aphyr.com)
AI
The author argues that today’s AI—especially large language models—is less like human-like intelligence and more like a “bullshit machine” that statistically imitates text while frequently confabulating, misunderstanding context, and making factual errors. They describe why LLMs can’t reliably reason about their own outputs, how “reasoning traces” and generated explanations can be misleading, and why models can be both astonishingly capable and still repeatedly “idiotic” in practical tasks. Overall, the piece frames current AI progress as creating major real-world risks alongside potential benefits, without offering a single definitive prediction of the future.