HackerNews Digest

April 06, 2026

Show HN: I built a tiny LLM to demystify how language models work

GuppyLM is a 9 M‑parameter vanilla transformer language model that emulates a fish character named Guppy. It is trained from scratch on a synthetic dataset of ~60 K single‑turn conversations covering 60 predefined topics (e.g., greetings, food, bubbles, weather). The data are generated by templated composition of tank objects, food types, and activities, yielding ~16 K unique outputs. Training uses a simple pipeline (tokenizer creation, data loading, cosine‑annealed learning rate, AMP) and completes in ~5 minutes on a single GPU (e.g., a Colab T4). The model employs standard multi‑head attention, ReLU feed‑forward layers, and LayerNorm; no GQA, SwiGLU, RoPE, or early‑exit mechanisms are included. Inference is single‑turn only, limited to a 128‑token context window to avoid degradation after multiple turns. The repository provides config, model, dataset, training, generation, evaluation, and inference scripts, plus tooling to export the dataset to HuggingFace and generate a Colab notebook. The pre‑trained model can be downloaded from HuggingFace and run locally or in a browser for interactive chat.
Read full article →
The comments convey strong approval of the fish‑themed LLM as an accessible, educational demonstration, noting its charm, humor, and clear constraints that make its behavior easy to grasp. Several users express curiosity about the amount of training data required for a coherent fish personality and suggest playful extensions such as emoji dialogue or joke‑telling. References to philosophical framing and comparisons to larger models highlight appreciation for the simplicity and intentional design, while a few brief “dead” entries add no substantive content.
Read all comments →

Gemma 4 on iPhone

Google AI Edge Gallery is an iOS‑only, free app that runs open‑source large language models (LLMs) locally on the device, providing offline, private generative AI. The latest version adds official support for the Gemma 4 model family, enabling advanced reasoning, logic, and creative tasks without network transmission. Core capabilities include: - **Agent Skills**: extend LLMs with tools (Wikipedia lookup, interactive maps, visual summary cards) and load community‑contributed skills via URLs or GitHub. - **AI Chat with Thinking Mode**: multi‑turn conversation with optional step‑by‑step reasoning view (currently for Gemma 4). - **Ask Image**: multimodal object identification and description from camera or gallery. - **Audio Scribe**: real‑time on‑device transcription and translation. - **Prompt Lab**: test prompts with adjustable parameters (temperature, top‑k). - **Mobile Actions & Tiny Garden**: offline device controls and a mini‑game using FunctionGemma 270 M. - **Model Management & Benchmark**: download, import, and benchmark models on specific hardware. The app collects identifiers, diagnostic data, location, and usage metrics (some linked to identity, some not). Requires iOS 17+, macOS 14+ on Apple Silicon, or visionOS 1+. Source code is on GitHub; support email is google‑ai‑edge‑gallery‑android‑[email protected].
Read full article →
The comments express strong enthusiasm for running the Gemma 4 model locally on iOS and Android devices, highlighting impressive speed, useful “mobile actions,” and potential for privacy‑preserving applications such as education tools and home‑assistant tasks. Users report successful deployment on recent iPhones and Macs, noting good performance but occasional heating, UI glitches, and occasional reasoning errors. There is interest in deeper integration with system shortcuts and broader availability of models, alongside cautious remarks about privacy policies, the “edge” branding, and the need for clearer documentation and open‑source transparency. Overall sentiment is largely positive with practical concerns.
Read all comments →

Show HN: YouTube search barely works, I made a search form with advanced filters

Advanced Search for YouTube is presented on Playlists.at as a tool to refine YouTube queries using specialized search prefixes. The interface allows selection of search sources (YouTube, Google Video) and filtering by playlists, with options to specify the number of videos. In the provided example, a search returned no matches, prompting the user to modify terms or filters. The page also includes a placeholder for visual content, referencing an image with alt text “Playlists.at”. No additional functionality, instructions, or data are included beyond the basic search layout and the indication of an empty result set.
Read full article →
The comments express strong dissatisfaction with YouTube’s current search and recommendation systems, describing them as unreliable, overly biased toward top results and Shorts, and lacking useful filters such as language, upload date, or channel‑specific searches. Users report difficulty locating specific videos, unwanted additions to watch history, and a degraded subscriptions experience, prompting some to maintain multiple accounts or build personal databases. There is a common desire for improved search functionality, better organization tools, and alternatives, including self‑hosted open‑source options, to restore the platform’s former utility.
Read all comments →

LÖVE: 2D Game Framework for Lua

LÖVE is an open‑source 2D game framework for Lua, supporting Windows, macOS, Linux, Android, and iOS. Development occurs on the **main** branch (unstable) with separate branches for released major versions; releases are tagged and binary downloads provided. Nightly builds are generated via GitHub CI, with Ubuntu packages in ppa:bartbes/love‑unstable and Arch packages in the AUR. A comprehensive test suite covers all LÖVE APIs and can be run with `love testing`. Contributions are accepted through the issue tracker, Discord/IRC, and pull requests, following the source‑code style guide; AI‑generated patches are rejected. Build instructions require out‑of‑tree CMake (≥ 3.15) or `-DCMAKE_INSTALL_PREFIX=` for older versions. macOS builds use the Xcode project `platform/xcode/love.xcodeproj`; iOS builds need the `love‑apple‑dependencies` zip and the same Xcode project. Android build steps are in a separate repository. Core dependencies include SDL3, OpenGL 3.3+/ES 3.0+/Vulkan/Metal, OpenAL, Lua/LuaJIT/LLVM‑lua, FreeType, harfbuzz, ModPlug, Vorbisfile, and Theora.
Read full article →
Comments highlight strong enthusiasm for LÖVE (Love2D), praising its beginner‑friendly workflow, simple Lua API, and rapid prototyping via zip deployment. The community is described as welcoming and supportive, and the engine’s cross‑platform capabilities, SDL2 foundation, and performance are noted positively. Some express desire for an updated stable release and mention using the latest repository head for better performance. Comparisons to other frameworks note LÖVE’s lightweight nature versus larger engines, while a few note web‑based solutions as faster. Overall sentiment is largely favorable.
Read all comments →

Microsoft hasn't had a coherent GUI strategy since Petzold

Microsoft has lacked a single, clear GUI strategy for Windows since the era of Charles Petzold’s 1988 *Programming Windows* (Win16/Win32). Subsequent attempts introduced complexity rather than coherence: * **1990s** – MFC wrapped Win32 in C++; COM/OLE/ActiveX added component‑level abstractions without a unified UI model. * **2003‑2006** – “Avalon” (later WPF) offered XAML‑based, GPU‑accelerated UI, but internal resistance to managed code limited its adoption. * **2007‑2010** – Silverlight provided a high‑quality browser plug‑in and Windows‑Phone base, yet was abruptly abandoned for HTML5. * **2012** – Windows 8’s Metro/WinRT introduced a native C++ runtime separate from .NET, creating parallel, conflicting roadmaps (WinRT vs. WPF/.NET). * **2015‑present** – UWP promised “write once, run everywhere” but suffered from phone platform loss and limited adoption; later efforts (WinUI 3, Project Reunion/Windows App SDK, MAUI) aim to unify APIs but still lack a stable, single answer. Current Windows GUI options include legacy Win32, MFC, WinForms, WPF, WinUI 3/Windows App SDK, MAUI, plus web‑hybrid (Blazor Hybrid, WebView2) and third‑party frameworks (Electron, Flutter, Tauri, Qt, React Native for Windows, Avalonia, Uno, Delphi, Java Swing/JavaFX). The fragmentation stems from internal team politics, premature conference‑driven bets, and abrupt business pivots rather than technical shortcomings.
Read full article →
The comments convey a broadly negative view of Microsoft’s desktop UI strategy, describing frequent shifts between frameworks, abandoned technologies, and a lack of a consistent design system that hampers developer productivity and user experience. Critics favor legacy Win32/WinForms or cross‑platform approaches such as web‑based, Qt, or Avalonia solutions, and cite performance or compatibility problems with newer stacks like WPF, UWP, and WinUI. The perceived focus on enterprise and cloud over cohesive native UX is seen as a cultural issue, while some note similar challenges at Apple and advocate simpler, stable APIs.
Read all comments →

Artemis II crew see first glimpse of far side of Moon [video]

NASA’s Artemis II crew—astronauts Reid Wiseman, Victor Glover, Christina Koch and CSA astronaut Jeremy Hansen—reported seeing the Moon’s far side for the first time during the third day of their Orion flight. The crew captured a photograph of the Orientale basin, which NASA confirmed as the first human‑eye view of the entire basin. Koch noted the visual difference from the familiar near‑side appearance. At 23:00 BST on Saturday the mission’s online dashboard recorded the spacecraft at over 180,000 mi (≈289,681 km) from Earth, on its trajectory around the lunar far side before return. The article emphasizes the crew’s observations and the historic imaging of the basin, without reference to unrelated image captions.
Read full article →
The comments collectively express enthusiasm for the Artemis II mission and the unprecedented view of the lunar far side, mixed with disappointment over perceived cynicism and politicized debate. Many participants seek technical explanations of communication, illumination, and trajectory, while others wish for bolder mission goals, cultural touches, or more risk‑taking. A recurring theme is appreciation of the symbolic significance of international cooperation and human exploration, juxtaposed with occasional skepticism about relevance amid Earth’s problems and criticism of media coverage. Overall sentiment balances excitement with calls for clearer information and less negativity.
Read all comments →

Eight years of wanting, three months of building with AI

Lalit Maganti spent eight years wanting better SQLite developer tools and, after ~250 hours over three months, released syntaqlite—a parser, formatter, and IDE extensions for SQLite and PerfettoSQL—largely powered by AI coding agents (Claude Code, Aider, Roo Code). The project’s difficulty stemmed from SQLite’s lack of a formal grammar and its C implementation that does not expose a parse tree, requiring extraction of parser code and handling >400 grammar rules. Initial AI‑generated code (C extraction pipeline, Python scripts) was functional but tangled, prompting a complete rewrite in Rust with tighter design, systematic refactoring, and extensive testing (500+ tests). AI accelerated concrete coding tasks, provided rapid learning on topics like Wadler‑Lindig pretty‑printing, and enabled fast prototyping of editor extensions, WASM playground, and documentation. However, reliance on AI led to codebase fragility, loss of mental model, over‑deferred design decisions, and addictive prompting cycles. The author concludes that AI is a strong implementation multiplier but insufficient for high‑level design, architecture, and preserving historical context.
Read full article →
Comments convey a broadly balanced view of AI‑assisted coding: many acknowledge its speed and ability to jump‑start projects, yet stress that the generated code often lacks coherent architecture, requires extensive review, and can produce fragile “spaghetti” implementations. Review effort, testing, and clear specifications are repeatedly cited as essential guardrails, while poor prompts or fatigue lead to lower‑quality output. Users report both productivity gains and “slot‑machine” frustration, emphasizing that AI functions best as a force multiplier when integrated with disciplined design, testing, and iterative refinement.
Read all comments →

Endian wars and anti-portability: this again?

The author argues against hostility toward portability in open‑source projects, countering common objections. He notes that Linux still supports niche but educational architectures such as DEC Alpha (influencing RISC‑V) and that MIPS, SPARC, PowerPC, and even 68000 have ongoing commercial or embedded use; ports to these systems demonstrate genuine community demand. He refutes the claim that “big endian is dead,” emphasizing that both endian modes are required for a robust ecosystem, citing real bugs uncovered only on big‑endian hardware (e.g., a Git memory‑corruption issue and a Clang invalid read). He also defends 32‑bit support, pointing out its relevance in developing regions, low‑cost hardware, and the efficiency benefits of fitting software within a 4 GB address space. Overall, he urges maintainers to accept community‑driven ports, allowing projects to evolve with user needs while preserving software quality.
Read full article →
Comments express concern over the high maintenance cost of supporting rare architectures, with many agreeing that maintainers should not be forced to sustain unpopular ports and that focusing on dominant platforms is pragmatic. Some acknowledge community interest in big‑endian support but view it as low‑ROI and cite limited hardware availability and testing challenges. A few reference Linux kernel practices and Linus Torvalds’ stance against big‑endian RISC‑V, while others dismiss big‑endian preference as personal bias. Overall, the consensus leans toward prioritizing mainstream architectures over niche ports.
Read all comments →

Copilot is 'for entertainment purposes only', per Microsoft's terms of use

Microsoft’s Copilot terms of use, last updated October 24 2025, label the service “for entertainment purposes only,” warning that it can err, may not work as intended, and should not be relied upon for important advice. A Microsoft spokesperson confirmed that “legacy language” will be revised in a forthcoming update to better reflect current usage. The disclaimer aligns with similar notices from OpenAI and xAI, which caution users not to treat their outputs as definitive truth or sole factual sources. The article notes that, despite these cautions, Microsoft continues to market Copilot to corporate customers while facing criticism on social media for the disclaimer’s wording.
Read full article →
The discussion conveys a predominantly negative view of Microsoft’s current AI offerings, describing them as primarily entertainment‑oriented and lacking serious utility. Commenters contrast this perception with Microsoft’s earlier reputation as an AI leader following its investment in OpenAI, suggesting the company has fallen behind. Skepticism is expressed about the effectiveness of Copilot, with some participants questioning which version is being referenced and implying that the service does not meet professional expectations.
Read all comments →

Show HN: Gemma Gem – AI model embedded in a browser – no API keys, no cloud

Gemma Gem is a Chrome extension that runs Google’s Gemma‑4 language model entirely on‑device via WebGPU, eliminating API keys, cloud services, and data transmission. It requires Chrome with WebGPU support and loads either the ~500 MB E2B or ~1.5 GB E4B quantized ONNX models (q4f16, 128 K context) on first use. The architecture comprises an offscreen document that hosts the model using @huggingface/transformers and executes the agent loop, a service worker that routes messages, and a content script that injects a chat UI (shadow DOM) and provides DOM tools (read_page_content, click_element, type_text, scroll_page, take_screenshot, run_javascript). Users open the chat via a gem icon, select model variants, toggle “thinking,” set max tool‑call iterations, clear context, or disable the extension per site. Development and production builds are managed with pnpm and the WXT Vite‑based framework; logging is routed through extension‑specific consoles. The agent directory defines ModelBackend and ToolExecutor interfaces, usable as a standalone library.
Read full article →
The comment highlights the Prompt API, currently available through an Origin Trial, as a potential web interface for AI models and supplies detailed statistics for a specific model, v3Nano, noting its substantial size and GPU backend. It observes that the large download requirements make immediate adoption as a native web feature improbable, and suggests that future implementations might rely on operating‑system‑level LLM integration rather than embedding full models within browsers.
Read all comments →