From 26 items, 14 important content pieces were selected
- OpenAI stops using SWE-bench Verified due to saturation ⭐️ 9.0/10
- AI Agent Deletes Production Database, Sparks Safety Debate ⭐️ 9.0/10
- Asahi Linux 7.0: Major Audio Driver Breakthrough ⭐️ 9.0/10
- DeepSeek-V4 Preview Released and Open-Sourced ⭐️ 9.0/10
- AI Should Elevate Thinking, Not Replace It ⭐️ 8.0/10
- Statecharts: Hierarchical State Machines for UI ⭐️ 8.0/10
- GoDaddy Transfers Domain to Stranger Without Verification ⭐️ 8.0/10
- Why Alzheimer’s Research Has Stagnated ⭐️ 8.0/10
- HauhauCS’s uncensored LLMs based on plagiarized Heretic tool ⭐️ 8.0/10
- Qwen3.6-27B INT4 Hits 100+ tps on RTX 5090 ⭐️ 8.0/10
- 1900 US Academy Members Urge Trump to Stop Attacks on Science ⭐️ 8.0/10
- Lishuan 7G100 GPU Gets Microsoft WHQL Certification ⭐️ 8.0/10
- Top university subdomains hijacked to serve porn and scams ⭐️ 8.0/10
- Friendster Bought for $30k, Plans Physical-Contact Social Network ⭐️ 7.0/10
OpenAI stops using SWE-bench Verified due to saturation ⭐️ 9.0/10
OpenAI announced it will no longer evaluate its models on the SWE-bench Verified benchmark, which has reached 93.9% saturation, and the SWE-bench team revealed upcoming multilingual and multimodal benchmarks. This highlights the challenge of benchmark saturation in AI evaluation, where top models quickly max out scores, reducing the benchmark’s ability to differentiate capabilities and incentivizing gaming. SWE-bench Verified is a human-filtered subset of 500 instances; the upcoming SWE-bench Multilingual includes 300 tasks across 9 languages, and SWE-bench Multimodal will be open-sourced within a month.
hackernews · r/LocalLLaMA · kmdupree · Apr 26, 13:58
Background: SWE-bench is a benchmark that evaluates AI models on real-world software engineering tasks, such as fixing bugs or implementing features from GitHub issues. SWE-bench Verified is a curated subset designed to remove ambiguous or infeasible samples, making evaluation more reliable. Benchmark saturation occurs when models achieve near-perfect scores, diminishing the benchmark’s utility for measuring progress.
Discussion: Co-creator ofirpress acknowledged saturation but noted room for growth for others, while commenters like Jcampuzano2 and cpard highlighted the inevitability of benchmark gaming and structural issues in benchmarks like ELT-Bench. jddj pointed out that many SWE-bench passing PRs would not be merged in practice.
Tags: #AI benchmarks, #coding capabilities, #SWE-bench, #LLM evaluation, #machine learning
AI Agent Deletes Production Database, Sparks Safety Debate ⭐️ 9.0/10
An AI coding agent deleted a production database, and the company’s postmortem blamed the agent rather than inadequate safeguards, sparking widespread discussion on AI safety and operational responsibility. This incident highlights critical gaps in deploying AI agents in production environments, where autonomous actions can cause irreversible damage, and underscores the need for robust safeguards and clear accountability. The agent had access to production secrets and was able to execute destructive commands without human approval or read-only restrictions, revealing a lack of least-privilege principles and change management controls.
hackernews · jeremyccrane · Apr 26, 16:27
Background: AI agents are autonomous systems that can perform tasks like code generation and database operations. Without proper safeguards—such as sandboxing, human-in-the-loop approvals, and read-only access—they pose significant risks in production environments.
References
Discussion: Commenters largely agree that the incident is a classic operational failure, not an AI-specific problem, and criticize the postmortem for deflecting blame. Some find irony in using an LLM to write about the incident, while others emphasize that AI agents can output any sequence of tokens, making safeguards essential.
Tags: #AI safety, #production incident, #database security, #LLM agents, #postmortem
Asahi Linux 7.0: Major Audio Driver Breakthrough ⭐️ 9.0/10
Asahi Linux’s 7.0 progress report details major advancements in audio driver support for Apple Silicon Macs, achieved through extensive reverse engineering of the CS42L84 audio codec. This progress brings Linux on Apple Silicon closer to full hardware support, enabling high-quality audio output and input on Macs that Apple does not officially support for Linux. The team reverse-engineered the CS42L84 codec, which is similar to the documented CS42L42, and added support for 48 and 96 kHz sample rates, with potential for more rates in the future.
hackernews · elisaado · Apr 26, 10:50
Background: Asahi Linux is a volunteer-driven project that ports Linux to Apple Silicon Macs by reverse-engineering undocumented hardware. Apple does not provide official documentation or support for running Linux on its M-series chips, making this work crucial for the open-source community.
References
Discussion: Commenters expressed admiration for the technical achievement but some voiced concerns about the project remaining separate from the kernel mainline and mainstream distributions. Others hoped Apple would eventually provide documentation to ease the effort.
Tags: #Asahi Linux, #Apple Silicon, #Linux kernel, #reverse engineering, #audio drivers
DeepSeek-V4 Preview Released and Open-Sourced ⭐️ 9.0/10
DeepSeek has released the preview version of DeepSeek-V4, including both Pro and Flash variants, and open-sourced the model weights on Hugging Face. The new model features a 1M-token context window, enhanced Agent capabilities, and state-of-the-art performance on math, STEM, and coding benchmarks. DeepSeek-V4 surpasses all current open-source models in key benchmarks and rivals top proprietary models, making advanced AI more accessible and affordable. Its strong Agent capabilities and cost-effective API options could accelerate the adoption of AI agents in real-world applications. DeepSeek-V4-Pro uses a sparse attention architecture that reduces single-token inference FLOPs to 27% of DeepSeek-V3.2 at 1M-token context, while V4-Flash achieves only 10% FLOPs and 7% KV cache size. V4-Flash has 284B total parameters with 13B active, offering faster and cheaper API service while maintaining strong reasoning and Agent performance.
telegram · zaihuapd · Apr 26, 07:17
Background: DeepSeek is a Chinese AI company known for developing open-source large language models. The V4 series introduces a sparse attention mechanism to efficiently handle long contexts up to 1M tokens, which is critical for agentic tasks that involve long tool-use trajectories. Agent capabilities refer to the model’s ability to autonomously use tools, browse the web, and execute multi-step tasks.
References
Tags: #AI, #DeepSeek, #open-source, #LLM, #Agent
AI Should Elevate Thinking, Not Replace It ⭐️ 8.0/10
An essay argues that AI should augment human cognition and creativity, not substitute it, warning that over-reliance on AI could degrade critical thinking skills. This perspective challenges the prevailing trend of using AI as a replacement for human effort, urging a balanced approach that preserves and enhances human intellectual capabilities. The essay has garnered significant community engagement with 227 points and 186 comments, featuring debates on practical AI use and the evolution of engineering roles.
hackernews · koshyjohn · Apr 26, 20:03
Background: The discussion around AI augmentation versus replacement is central to current debates in software engineering and productivity. Many fear that AI tools, if used uncritically, could lead to a loss of deep understanding and problem-solving skills.
Discussion: Commenters express a range of views: some argue that AI is just another abstraction layer, similar to modern IDEs or package managers, while others warn that over-reliance could degrade engineering skills. There is agreement that AI should be used to augment, not replace, human thinking.
Tags: #AI, #critical thinking, #productivity, #software engineering, #philosophy
Statecharts: Hierarchical State Machines for UI ⭐️ 8.0/10
Statecharts.dev provides an introduction to hierarchical state machines, with community discussion highlighting their value in UI development and nuances like history pseudo-states. Statecharts help manage complex UI logic by organizing states hierarchically, making interactions easier to reason about and maintain. The discussion from XState creator and others underscores their practical utility in modern software engineering. History pseudo-states (H, H*) introduce non-determinism because entering a parent via H restores the last active child, meaning the same event can lead to different states. XState is a JavaScript/TypeScript library for authoring, executing, and visualizing state machines and statecharts.
hackernews · sph · Apr 26, 09:32
Background: A state machine is a model of behavior with a finite number of states and transitions between them. Hierarchical state machines (statecharts) extend this by allowing states to contain sub-states, reducing complexity through nesting. They are widely used in UI development, embedded systems, and protocol modeling.
Discussion: The community discussion includes insights from XState creator David Khourshid, who emphasizes treating statecharts as executable behavior rather than just documentation. Another commenter notes that history pseudo-states break the deterministic promise, as they introduce hidden state not shown in diagrams.
Tags: #statecharts, #state machines, #UI development, #XState, #software engineering
GoDaddy Transfers Domain to Stranger Without Verification ⭐️ 8.0/10
A domain owner reported that GoDaddy transferred their domain to an unauthorized party without requiring any documentation or proper verification, exposing a critical security flaw in the registrar’s transfer process. This incident highlights severe security risks in domain management, as domain hijacking can lead to loss of email access, website control, and business operations. It underscores the need for stronger verification protocols and raises trust concerns about GoDaddy, one of the largest domain registrars. The transfer was executed without any documentation, such as a signed authorization form or identity verification, which are standard requirements for domain transfers. The victim noted that all associated email addresses, marketing materials, and SEO rankings were compromised, and they were effectively locked out of online accounts tied to the domain.
hackernews · jamesponddotco · Apr 26, 16:57
Background: Domain transfers typically require a Form of Authorization (FOA) and confirmation from the current registrar to prevent unauthorized transfers. ICANN sets guidelines for these processes, but enforcement varies by registrar. GoDaddy has a history of security incidents, including issuing unvalidated SSL certificates and injecting JavaScript into customer websites.
References
Discussion: Commenters expressed strong distrust of GoDaddy, citing its poor security track record. Some suggested the incident might be an inside job, while others recommended registering domains as trademarks to gain stronger legal protection. The discussion also highlighted that losing a domain can lock users out of critical online accounts like banking and CRM systems.
Tags: #domain security, #GoDaddy, #registrar, #security breach, #hackernews
Why Alzheimer’s Research Has Stagnated ⭐️ 8.0/10
A Freakonomics podcast explores the lack of progress in Alzheimer’s research, highlighting the dominance of the amyloid hypothesis and its failure to produce effective treatments. This discussion matters because Alzheimer’s affects millions worldwide, and the stagnation in research has wasted billions in funding and delayed potential treatments. It also raises questions about scientific consensus and funding biases. The amyloid hypothesis, proposed in 1992, posits that amyloid-beta plaques cause Alzheimer’s, but drugs targeting amyloid have repeatedly failed in clinical trials. Critics argue that the hypothesis may be incorrect or incomplete, and that research funding has been overly concentrated on this single theory.
hackernews · chiefalchemist · Apr 26, 00:12
Background: Alzheimer’s disease is a progressive neurodegenerative disorder and the most common cause of dementia. The amyloid hypothesis has been the dominant theory for decades, leading to massive investment in drugs that clear amyloid plaques, yet no disease-modifying therapy has been approved. Recent approvals of lecanemab and donanemab show modest benefits but do not cure the disease, and their clinical significance is debated.
References
Discussion: Commenters express frustration with the amyloid hypothesis, with one noting that research as early as 2010 showed no mechanistic link between amyloid aggregates and Alzheimer’s. Others point to systemic issues, such as the ‘cabal’ of amyloid proponents that stifled alternative research, and recommend Karl Herrup’s book ‘How Not to Study a Disease’ for a critical perspective.
Tags: #Alzheimer's, #research methodology, #pharmaceutical industry, #amyloid hypothesis, #science policy
HauhauCS’s uncensored LLMs based on plagiarized Heretic tool ⭐️ 8.0/10
HauhauCS, known for popular uncensored LLM models, published a package called ‘reaper-abliteration’ that is a plagiarized fork of the Heretic abliteration tool, violating the AGPL-3.0 license by removing attribution and adding a commercial restriction. This incident undermines trust in the open-source LLM community, as HauhauCS’s models have over 5 million monthly downloads, and the plagiarism violates software ethics and license compliance, potentially affecting users and developers who rely on proper attribution. Evidence shows 7/7 module filenames, 30/32 refusal markers, and 30+ function names are identical to Heretic v1.2.0, with internal variable names like ‘good_residuals’ and ‘bad_residuals’ left unchanged. The original creator of Heretic confirmed the findings.
reddit · r/LocalLLaMA · nathandreamfast · Apr 26, 13:13
Background: Abliteration is a technique to remove refusal behaviors from LLMs without retraining, often used to create uncensored models. Heretic is an open-source tool that automates this process using Optuna parameter optimization, licensed under AGPL-3.0, which requires attribution and share-alike for derivative works.
References
Discussion: The community expressed strong condemnation, with many users reporting being blocked by HauhauCS for asking questions. The original creator of Heretic confirmed the plagiarism, and users noted that such misconduct damages reputation and trust.
Tags: #plagiarism, #open-source, #LLM, #ethics, #license-violation
Qwen3.6-27B INT4 Hits 100+ tps on RTX 5090 ⭐️ 8.0/10
The Qwen3.6-27B model quantized to INT4 using AutoRound achieves 105-108 tokens per second with a full 256k context window on a single RTX 5090 GPU via vLLM 0.19, with community reports of up to 160+ tps using custom patches. This demonstrates that large 27B parameter models with long context can run efficiently on consumer-grade hardware, making advanced AI capabilities more accessible to individuals and small teams. The setup uses vLLM with flashinfer attention backend, FP8 KV cache, and MTP speculative decoding with 3 draft heads, achieving high throughput while maintaining model quality with a KLD much better than NVFP4.
reddit · r/LocalLLaMA · Kindly-Cantaloupe978 · Apr 26, 08:37
Background: Qwen3.6-27B is a 27-billion-parameter language model from the Qwen series. INT4 quantization reduces model size and memory bandwidth requirements, while vLLM is a high-performance inference engine that supports various optimizations like speculative decoding and KV cache quantization to accelerate generation.
Discussion: The community is highly enthusiastic, with users reporting similar performance on RTX 3090 (71-83 tps) and even higher speeds (160+ tps) with Genesis patches. Some users discuss trade-offs between INT4 27B and FP8 35B A3B models, and others ask about optimal setups for lower-end GPUs like the RTX 5060 Ti.
Tags: #LLM inference, #quantization, #vLLM, #speculative decoding, #local LLM
1900 US Academy Members Urge Trump to Stop Attacks on Science ⭐️ 8.0/10
On March 31, 2025, 1900 members of the US National Academies, including over a dozen Nobel laureates, signed an open letter drafted by 13 scientists from fields such as medicine, epidemiology, and climate science, calling on the Trump administration to halt its assault on American science. This unprecedented mobilization of top scientists signals a severe crisis in US science policy, potentially undermining research funding, innovation capacity, and global competitiveness. The letter reflects growing alarm over cuts to basic research and political interference in scientific institutions. The signatories include Nobel laureates Harvey J. Alter, Francoise Barre-Sinoussi, Reinhard Genzel, Edvard I. Moser, and May-Britt Moser. The letter was drafted by 13 scientists and released online on March 31, 2025.
telegram · zaihuapd · Apr 26, 00:40
Background: The US National Academies (National Academy of Sciences, National Academy of Engineering, and National Academy of Medicine) are prestigious honorary societies whose members are elected for outstanding contributions to research. The Trump administration’s second term has seen significant cuts to federal research funding, particularly in climate and social sciences, and increased scrutiny of university research, leading to concerns about a ‘brain drain’ and declining scientific output.
References
Tags: #science policy, #research funding, #US politics, #open letter
Lishuan 7G100 GPU Gets Microsoft WHQL Certification ⭐️ 8.0/10
Lishuan Technology’s 7G100 series GPU has obtained Microsoft WHQL certification, making it the first Chinese company and the fourth globally to achieve this certification for a GPU. This milestone demonstrates significant progress in China’s domestic GPU development, with performance approaching NVIDIA’s RTX 4060, and enhances the country’s self-sufficiency in critical semiconductor components. The GPU is built on a 6nm process with the proprietary ‘Tiantu’ architecture, achieving fully independent design of compute cores, instruction set, and software stack. In the Steel Nomad benchmark, it scored 2268, close to the RTX 4060.
telegram · zaihuapd · Apr 26, 02:59
Background: WHQL (Windows Hardware Quality Labs) certification is Microsoft’s official testing program that ensures hardware and drivers are compatible and stable with Windows. It is a crucial step for any hardware aiming for broad consumer adoption. Lishuan’s achievement places it alongside major GPU vendors like NVIDIA, AMD, and Intel.
Tags: #GPU, #WHQL, #Chinese semiconductor, #hardware, #AI
Top university subdomains hijacked to serve porn and scams ⭐️ 8.0/10
At least 34 top universities, including UC Berkeley and Columbia, had their subdomains hijacked by the threat group Hazy Hawk to serve pornographic content and scams. The attack exploited administrators’ failure to clean up CNAME records after decommissioning subdomains. This incident highlights a critical domain management vulnerability that can affect any organization, especially those with high domain authority like universities. The hijacked pages rank high in Google search results, potentially reaching a large audience and damaging institutional reputations. Hazy Hawk has been active since at least December 2023, targeting abandoned cloud resources and DNS misconfigurations. Hundreds of subdomains and thousands of malicious pages have been found in search engines.
telegram · zaihuapd · Apr 26, 09:02
Background: A subdomain takeover occurs when an attacker gains control over a subdomain by exploiting a dangling DNS record, such as a CNAME pointing to a decommissioned external service. CNAME records map one domain name to another, and if the target service is removed without deleting the record, an attacker can claim it. This attack is well-known in security circles but often overlooked in domain management practices.
References
Tags: #security, #domain hijacking, #DNS, #university, #cyberattack
Friendster Bought for $30k, Plans Physical-Contact Social Network ⭐️ 7.0/10
The author purchased the domain Friendster.com for $20k in Bitcoin plus a domain worth $9k/year in ad revenue, and plans to revive it as a social network that requires physical phone-to-phone contact to connect. This acquisition revives a historic social network brand and introduces a novel physical-contact requirement that could address issues of stale connections and algorithmic content control. The domain was acquired from a previous owner who had been using it for a landing page; the new owner plans to launch an app where users must hold phones together to connect, with connections fading after a year.
hackernews · ca98am79 · Apr 26, 20:41
Background: Friendster was a pioneering social network launched in 2002, predating Facebook, but declined due to technical issues and competition. It was later sold and eventually shut down, with the domain remaining dormant. The new owner aims to differentiate by requiring physical proximity for connections, contrasting with today’s algorithm-driven platforms.
References
Discussion: Commenters raised concerns about the chicken-and-egg problem with the physical feature, and suggested allowing initial virtual connections to gain traction. Others praised the idea of fading connections to combat stale networks, and noted the difficulty of finding the app in app stores.
Tags: #social media, #domain names, #startup, #nostalgia, #acquisition