Front-page articles summarized hourly.
Small Datum reports that MariaDB 12.3 improves vector-index performance and recall over 11.8 and beats Postgres 18.2 with pgvector 0.8.1, especially on larger datasets. Using ann-benchmarks on dbpedia-openai-x-angular tests across 100k, 500k and 1M items, on a 48-core/128 GB server, 12.3 uses less CPU per query (vmstat-confirmed). Gains over 11.8 grow with dataset size, with 12.3 achieving the best results among the configurations tested.
Infinity, Inc. is a DC Comics superhero team consisting of the descendants of the Justice Society of America.
The piece outlines eight levels of agentic engineering for AI-assisted coding, showing how teams move from simple tab-complete tools (Levels 1–2) to richer context management (3), compounding learning (4), and shared capabilities through MCPs and skills (5). It then covers harnessing feedback loops (6), the rise of background agents (7), and finally autonomous agent teams (8). Key themes include managing context and tools, planning versus automation, backpressure and security, orchestration vs. parallel agents, and the ongoing teamwork-driven race toward higher levels.
Shahram Khosravi reframes defeat as a method: thinking and acting from within ruins caused by colonial dispossession. He links the Bakhtiari oil theft (1908) and Iran’s precarity to Fanon’s idea that defeat can generate knowledge, not paralysis. An “open face” facing disaster without illusion enables radical imagination, fugitivity, and ethical critique. From Karbala, Ashura, and Black/Indigenous histories, defeat reveals power’s injustices and births resistant thought—radical hope, not victory, as political practice. The defeated must think with defeat to imagine new futures, especially for Palestinians and others.
Using Castlevania as a metaphor, the piece portrays AI agents as Dracula—driven by prompts and a reward model—while security practitioners are the Belmonts who can’t defeat Dracula but must win every battle. An agent is a simple loop of API calls to LLMs and tools, now aided by planning, memory, and multi-agent orchestration. But non-determinism, tool-errors, and infinite loops persist, and API fragmentation across OpenAI/Claude/Gemini hinders model-agnostic security. The author argues security is behind: govern untrustworthy payloads with anomaly detection, circuit breakers, IAM, and secure defaults, not LLM defenses; standards will arrive slowly, costs remain high.
Addie Foote recounts trying to post-train a 1-trillion-parameter open-weight model (Kimi-K2-Thinking) with open-source tools. After failed attempts with LLaMA-Factory, KTransformers, and HuggingFace, the team builds a custom training stack because the existing options are buggy and memory/quantization issues abound. The piece follows creating a Yoda-style dataset, loading the model across GPUs, applying LoRA to quantized MoE weights, and wrestling with compression, CUDA memory fragmentation, and dequantization. A working forward/backward loop and some qualitative gains appear, but the open-source stack remains brittle, with debt across layers. The takeaway: sometimes patching yields diminishing returns and building may be necessary.
Andy Chen argues that enterprises can create an Enterprise Context Layer (ECL)—a centralized, self-updating knowledge layer that enables an AI to reason with organizational context—using roughly 1000 lines of Python and a Github repo. Distinguishing retrieval from synthesis, he shows how context graphs, citations, and an agent-based maintenance system can map product, process, and politics across R&D, GTM, legal, HR. A seed in meta/ and 20 parallel agents produce a rich ECL; governance, access control, and future scalability are key next steps.
Notes exploring how AI tools should price by outcomes, not monthly credits. Introduces 'vibe coders' who build AI apps but stumble on monetization infrastructure. Proposes Lovable adopt revenue sharing instead of upfront subscriptions: take a percentage of creators' revenue (e.g., 5–15%, up to 30%), in exchange for one-click monetization, Stripe handling, migration and optimization services, and scalable support. Describes a Lovable Partners Program where hands-on services train the platform, enabling automation of repeatable tasks and a data flywheel. Envisions a 'Billion-Dollar Mission' to pay out $1B to vibe coders; argues platforms that align incentives will capture this wave.
After 18 months and multiple pivots, the team shut down the product and started over. They regret the No-Tests era and rewrote with strict TypeScript and tests, abandoning heavy Next.js/Server Actions. They now use React with tRPC and a small Hono backend, deployed on Kubernetes and served from a CDN. For orchestration, they favor Argo over useworkflow/Temporal to manage stateful Kubernetes jobs. They’re seeking feedback and plan a launch with design partners.
Complex systems like climate, markets, and addiction resist small, elegant theories; Enlightenment tools often fail on complexity. The Santa Fe Institute showed descriptive but not prescriptive insights. Modern AI—especially large language models—are large but rely on a compact architecture that compresses complex systems into usable models. There may be two theory layers: system-specific weights and a universal, compact architecture. Mechanistic interpretability aims to extract structure from trained models, turning compression into science. This shift yields probabilistic predictions and a new epistemology for understanding complex phenomena.
Microsoft's Copilot update introduces "context preservation" that forces links clicked inside Copilot to open in a side panel powered by Edge, rather than the user's default browser, effectively trapping browsing within Microsoft's rendering surface and raising privacy concerns. It's unclear if opt-in. The update can access tab context, summarize across tabs, or draft text from on-screen content, and can save conversations, and with permission sync passwords and form data. Rollout is limited to Windows Insider builds (version 146.0.3856.39+).
Back Market and Google are selling a $3 USB stick to install ChromeOS Flex on older PCs and Macs, giving outdated devices a new lease on life. The initial limited run offers 3,000 keys starting March 30, targeting sellers, buyers, schools, and small businesses. ChromeOS Flex is a lighter ChromeOS with no Android app support, but compatible with many old Windows machines and pre-Apple Silicon Macs per Google's compatibility list.
Could not summarize article.
Ankur Sethi built Cutlet, a dynamic programming language, in four weeks using Claude Code. It runs on macOS and Linux, and is named after his cat. Variables use my, arrays hold doubles, and it supports vectorized operations via the @ meta-operator (e.g., (temps-c @* 1.8) @+ 32). The @: operator zips arrays into maps; the language is fully expression-based, with functions declared by fn and a say function that returns nothing. Some features—like file I/O and error handling—are still missing. He emphasizes guardrails, testing, and ‘agentic engineering’ for LLM-driven coding.
RCLI is an on-device voice AI for macOS (Apple Silicon) delivering a complete STT + LLM + TTS pipeline with local RAG over documents and sub-200ms end-to-end latency, no cloud or API keys. Built on MetalRT, it runs 43 macOS actions via voice and supports hybrid document retrieval over PDFs/DOCX/TXT. Requires macOS 13+ on Apple Silicon (M1+), with M3+ and MetalRT recommended; M1/M2 fallback to llama.cpp. Install via curl script or Homebrew; interactive TUI with rcli setup/listen/ask/rag/models/voices/bench. Models include Qwen/Llama/LFM2; supports VAD, STT (Zipformer/Whisper), TTS, tool calling. MIT license; MetalRT proprietary.
An update to Mond's MOND←TECH MAGAZINE piece on Rust micro-optimizations, titled 'How many options fit into a boolean?'. It explains why a Result<bool, bool> cannot be compressed beyond a byte due to needing a valid memory representation, discusses size_of and the turbofish syntax, and notes the challenges of embedding a PDF. The post also includes updates about the author's life—planning a move from Central Europe to Seattle to work in AI—and reflections on writing, enthusiasm for contributing to Paged Out!, and other tech musings.
David Noel Ng describes LLM Neuroanatomy: topping the HuggingFace Open LLM Leaderboard by duplicating middle transformer layers rather than training. He posits a functional anatomy for Transformers: early layers encode, late layers decode, middle layers form reusable circuits. He created a ‘brain scanner’ heatmap testing (i, j) configurations on 72B models, using two orthogonal probes (hard math and EQ) and a scoring method based on digit-distribution expectations. The optimal config (45,52) duplicates seven middle layers, boosting RYS-XLarge to #1 across five benchmarks with no weight updates. This suggests circuit-like middle layers and orthogonality to fine-tuning.
Site is experiencing technical difficulties. Oracle is aware and working to fix the issue; apologies for the inconvenience. Contact options: sales at 1-800-ORACLE1; Corporate HQ at 1-650-506-7000; US tech support at 1-800-633-0738. Incident Number: 0.1e4e4317.1773156461.80af7e76.
RFC 454545 proposes the Human Em Dash (HED), a Unicode character (U+10EAD) visually identical to the em dash but distinct, paired with a Human Attestation Mark (HAM, e.g., U+10EAC) to signal human authorship. Rendering remains the same; automated systems must not emit HAM. Conforming implementations verify evidence of hesitation (delays, backspaces, cursor movement). The concept, called Human Cognitive Proof-of-Work (HCPoW), aims to combat Dash Authenticity Collapse by differentiating human from AI-produced punctuation. Legacy em dashes remain valid; backward compatibility noted. Security, policy, and IANA considerations discussed; examples provided.
Made by Johno Whitaker using FastHTML