Skip to the content.

From 53 items, 32 important content pieces were selected


  1. Microsoft and OpenAI End Exclusive Revenue-Sharing Deal ⭐️ 9.0/10
  2. FDA Approves First Gene Therapy for Genetic Hearing Loss ⭐️ 9.0/10
  3. pgBackRest archived: maintainer steps down ⭐️ 9.0/10
  4. Luce DFlash: 2x Throughput for Qwen3.6-27B on RTX 3090 ⭐️ 9.0/10
  5. ChatGPT 5.2 tops human scores in Tokyo, Kyoto entrance exams ⭐️ 9.0/10
  6. Quantum Magnifier Creates Most Intense Light Ever in Lab ⭐️ 9.0/10
  7. vLLM v0.20.0 Released with DeepSeek V4 and CUDA 13.0 ⭐️ 8.0/10
  8. Mercor Breach Leaks 4TB Voice Samples and IDs ⭐️ 8.0/10
  9. China Blocks Meta’s Acquisition of AI Startup Manus ⭐️ 8.0/10
  10. Super ZSNES: GPU-Accelerated SNES Emulator Revived ⭐️ 8.0/10
  11. GitHub Copilot Moves to Usage-Based Billing ⭐️ 8.0/10
  12. Open-source CLI agent tops TerminalBench with Gemini-3-flash-preview ⭐️ 8.0/10
  13. Supreme Court Weighs Geofence Warrants’ Constitutionality ⭐️ 8.0/10
  14. Microsoft Open-Sources VibeVoice with Speaker Diarization ⭐️ 8.0/10
  15. Microsoft and OpenAI Drop AGI Clause from Partnership ⭐️ 8.0/10
  16. Zig 0.16.0 Introduces Structured Concurrency via Io Interface ⭐️ 8.0/10
  17. MiMo-V2.5-Pro: Open-Source AI Rivaling Claude Opus ⭐️ 8.0/10
  18. Asahi Linux Nears Alpha Support for Apple M3 Chips ⭐️ 8.0/10
  19. Xiaomi Xuanjie O1 Chip Shipments Exceed 1 Million ⭐️ 8.0/10
  20. Easyduino: Open Source PCB Devboards for KiCad ⭐️ 7.0/10
  21. macOS 27 to enforce TLS 1.2, deprecate AFP for Time Machine ⭐️ 7.0/10
  22. The Lost Art of Staring at Walls ⭐️ 7.0/10
  23. The Woes of Sanitizing SVGs ⭐️ 7.0/10
  24. Why Not Just Use Lean? A Critical Comparison ⭐️ 7.0/10
  25. pip 26.1 Released with Dependency Cooldowns and Pylock Support ⭐️ 7.0/10
  26. Linux 7.1 Merge Window Closes with 13,000 Changesets ⭐️ 7.0/10
  27. Microsoft Open-Sources TRELLIS.2: 4B-Parameter Image-to-3D Model ⭐️ 7.0/10
  28. 4B LLM Benchmark: Nemotron 3 Nano Tops Finance, Reasoning, Code ⭐️ 7.0/10
  29. Boost VRAM for LLMs by Adding an Old GPU ⭐️ 7.0/10
  30. Skymizer Unveils 384GB PCIe Card for 700B LLM Inference ⭐️ 7.0/10
  31. Microsoft’s Windows K2 Plan Aims to Revamp Windows 11 ⭐️ 7.0/10
  32. Chinese Tech Giants Face User Backlash Over AI Content ⭐️ 7.0/10

Microsoft and OpenAI End Exclusive Revenue-Sharing Deal ⭐️ 9.0/10

Microsoft and OpenAI have restructured their partnership, ending the exclusive revenue-sharing deal that required OpenAI to use Azure exclusively. Under the new terms, Microsoft will no longer pay a revenue share to OpenAI, while OpenAI’s payments to Microsoft continue through 2030 at the same 20% rate but with a total cap. This shift frees OpenAI to use other cloud providers like AWS and potentially Google TPUs, breaking Microsoft’s exclusive grip on OpenAI’s infrastructure. It could reshape the AI cloud market, intensify competition among cloud providers, and accelerate OpenAI’s access to specialized hardware like Google’s latest TPUs. The revenue share from OpenAI to Microsoft remains at 20% but is now capped, and Microsoft continues as a major shareholder with a reduced stake of 27% (down from 49%). OpenAI is now free to run workloads on AWS and other clouds, and may consider using Google’s newly launched Gen 8 TPUs.

hackernews · helsinkiandrew · Apr 27, 13:22

Background: Microsoft and OpenAI formed a close partnership in 2019, with Microsoft investing billions and providing exclusive cloud infrastructure via Azure. The deal included a revenue-sharing arrangement where Microsoft received 20% of OpenAI’s revenue. This exclusive arrangement limited OpenAI’s ability to use competing cloud services or specialized hardware like Google’s TPUs, which are widely used by other frontier AI labs.

References

Discussion: Commenters see Google as a potential winner, as OpenAI may now adopt Google TPUs. Some question why Microsoft agreed to terms that seem favorable to OpenAI, speculating that Microsoft feared the previous deal was harming OpenAI’s competitiveness against rivals like Anthropic. Others note that this move allows OpenAI to use AWS, highlighting Azure’s weakening position.

Tags: #AI, #Microsoft, #OpenAI, #Cloud Computing, #Partnership


FDA Approves First Gene Therapy for Genetic Hearing Loss ⭐️ 9.0/10

The FDA has approved the first-ever gene therapy for treating genetic hearing loss caused by mutations in the OTOF gene, marking a historic milestone in the treatment of congenital deafness. This approval opens the door for a new class of treatments for genetic deafness, potentially restoring hearing in children who would otherwise rely on cochlear implants or hearing aids, and paves the way for therapies targeting other deafness-causing genes. The therapy, DB-OTO, uses a dual adeno-associated virus (AAV) vector to deliver a functional copy of the OTOF gene to inner ear hair cells, and is administered via injection into the inner ear. It is specifically for patients with autosomal recessive deafness 9 (DFNB9), a rare form of congenital deafness caused by OTOF mutations.

hackernews · JeanKage · Apr 27, 10:15

Background: Genetic mutations account for up to 60% of hearing loss present at birth. The OTOF gene provides instructions for making otoferlin, a protein essential for hearing. Mutations in OTOF disrupt sound signal transmission from hair cells to the auditory nerve, causing severe-to-profound deafness. Previous treatments for genetic hearing loss were limited to hearing aids or cochlear implants, which do not address the underlying genetic cause.

References

Discussion: Community comments express widespread enthusiasm for the approval, with many sharing personal stories of hearing loss and hope for future treatments. Some users noted that while this therapy targets a small population, it is a crucial step forward, and others highlighted the need for therapies addressing other genetic causes like GJB2 mutations. A few expressed disappointment that their specific type of deafness (e.g., from mumps) is not yet treatable by such approaches.

Tags: #gene therapy, #FDA approval, #hearing loss, #medical breakthrough, #biotechnology


pgBackRest archived: maintainer steps down ⭐️ 9.0/10

David Steele, the creator and sole maintainer of pgBackRest, has archived the project on GitHub and announced it is no longer maintained after 13 years of development. This discontinuation of a widely-used PostgreSQL backup tool highlights critical issues in open-source sustainability, including maintainer burnout and loss of corporate sponsorship after Crunchy Data’s acquisition. pgBackRest had 3.8k stars on GitHub and was sponsored by Crunchy Data until its sale; Steele’s efforts to find new sponsorship or a position to continue maintenance were unsuccessful.

rss · LWN.net · Apr 27, 14:06

Background: pgBackRest is an open-source backup and restore tool for PostgreSQL, supporting full, differential, and incremental backups, WAL archiving, and point-in-time recovery. It was developed over 13 years by David Steele, initially with corporate sponsorship from Crunchy Data, a PostgreSQL services company. After Crunchy Data was sold, the sponsorship ended, leaving Steele without funding to continue the project.

References

Discussion: Community members expressed sadness and gratitude for Steele’s work, with some noting that the source remains available for forking or paid maintenance. Others pointed to the Crunchy Data acquisition as a cautionary tale about relying on single-company sponsorship for critical open-source infrastructure.

Tags: #PostgreSQL, #open source, #backup, #maintainer burnout, #pgBackRest


Luce DFlash: 2x Throughput for Qwen3.6-27B on RTX 3090 ⭐️ 9.0/10

Luce DFlash, a GGUF port of DFlash speculative decoding, achieves up to 2x throughput for Qwen3.6-27B on a single RTX 3090 without retraining. This breakthrough significantly accelerates local LLM inference on consumer hardware, making high-performance AI more accessible to individuals and small teams. The implementation uses a standalone C++/CUDA stack on top of ggml, requires CUDA 12+ and an NVIDIA GPU with 24 GB VRAM, and achieves ~1.98x mean speedup on benchmarks like HumanEval and GSM8K.

reddit · r/LocalLLaMA · sandropuppo · Apr 27, 16:31

Background: Speculative decoding accelerates LLM inference by using a small draft model to propose tokens that a larger target model verifies in parallel. GGUF is a file format for quantized models that enables efficient inference on consumer hardware. DFlash is a block diffusion model designed for speculative decoding, enabling parallel drafting.

References

Discussion: The community is highly positive, praising the innovation and speed gains. Some users express concerns about quality degradation due to quantization and the lack of multi-GPU support, while others ask about use cases and benchmarking.

Tags: #speculative decoding, #local LLM inference, #GGUF, #CUDA, #Qwen


ChatGPT 5.2 tops human scores in Tokyo, Kyoto entrance exams ⭐️ 9.0/10

LifePrompt tested ChatGPT 5.2 Thinking on Tokyo and Kyoto University undergraduate entrance exams, and it scored higher than the top human examinees, including a perfect math score on the most competitive science track at Tokyo University. This marks a significant milestone where AI surpasses top human performance in Japan’s most prestigious university exams, challenging the current exam system that emphasizes memorization and computation, and prompting discussions on education reform. The test used image inputs of exam papers, and essays were graded by teachers from Kawai Juku, a major Japanese cram school. In 2024, ChatGPT-4 failed the same exams, while the o1 model first passed in 2025.

telegram · zaihuapd · Apr 27, 09:15

Background: Tokyo and Kyoto University entrance exams are among the most competitive in Japan, testing deep knowledge and problem-solving skills. ChatGPT 5.2 is a family of models released by OpenAI in December 2025, with the Thinking variant optimized for reasoning tasks.

References

Tags: #AI, #benchmark, #education, #ChatGPT, #Japan


Quantum Magnifier Creates Most Intense Light Ever in Lab ⭐️ 9.0/10

Physicists used a ‘quantum magnifier’ technique with dual plasma mirrors on the Gemini laser system to produce the most intense light ever in a lab, achieving over 9 mJ in harmonics 12-47 and approaching the Schwinger limit. This breakthrough brings laboratory light intensities close to 10²⁹ W/cm², enabling direct studies of quantum vacuum and strong-field quantum electrodynamics, which were previously only theoretical. The experiment used dual plasma mirrors to shorten the laser rise time to about 351 femtoseconds, dramatically improving the efficiency of relativistic surface high-harmonic generation. The measured efficiency matched theoretical predictions closely.

telegram · zaihuapd · Apr 27, 16:00

Background: The Schwinger limit is the critical electric field strength in quantum electrodynamics beyond which the vacuum becomes unstable and can spontaneously produce electron-positron pairs. Reaching this limit requires light intensities around 10²⁹ W/cm², which has been a long-standing goal in high-energy physics. Relativistic surface high-harmonic generation uses intense laser pulses interacting with plasma mirrors to produce coherent high-order harmonics.

References

Tags: #high-energy physics, #quantum electrodynamics, #laser physics, #plasma optics, #Schwinger limit


vLLM v0.20.0 Released with DeepSeek V4 and CUDA 13.0 ⭐️ 8.0/10

vLLM v0.20.0 introduces initial support for DeepSeek V4, switches the default CUDA wheel to CUDA 13.0, upgrades to PyTorch 2.11, and adds Python 3.14 support. This release significantly enhances vLLM’s capabilities for serving cutting-edge large language models, particularly DeepSeek V4, and aligns with the latest CUDA and PyTorch ecosystems, benefiting the AI inference community. The release includes 752 commits from 320 contributors, with notable features like FlashAttention 4 as default MLA prefill, TurboQuant 2-bit KV cache, and a new online quantization frontend.

github · khluu · Apr 27, 21:20

Background: vLLM is an open-source high-throughput LLM inference engine widely used in production. DeepSeek V4 is the latest model from DeepSeek, known for strong coding capabilities. CUDA 13.0 and PyTorch 2.11 are the latest versions of NVIDIA’s parallel computing platform and the popular deep learning framework, respectively.

References

Discussion: The community has reported a DSML token-leakage bug in DeepSeek V4, which v0.20.0 addresses. There is also discussion about the CUDA version policy following PyTorch’s lead, with recommendations to use --torch-backend=cu129 for CUDA 12.9 users.

Tags: #vLLM, #LLM inference, #DeepSeek, #CUDA, #PyTorch


Mercor Breach Leaks 4TB Voice Samples and IDs ⭐️ 8.0/10

A breach at AI staffing platform Mercor leaked 4TB of voice recordings and ID documents from over 40,000 contractors, accessed via a compromised API key in the LiteLLM library. This breach uniquely combines voice samples with ID scans, creating a ready-made deepfake kit for fraud, including bypassing banking voiceprints and impersonating executives in video calls. The breach occurred through LiteLLM, a Python library managing API keys for AI models, and was posted on the Lapsus$ leak site. Mercor confirmed the incident on March 31, 2026, and faces multiple lawsuits.

hackernews · Oravys · Apr 27, 09:57

Background: Audio deepfake technology uses AI to generate speech that mimics a specific individual, often requiring only a few seconds of sample audio. Voice cloning is a subset of deepfakes focused on audio manipulation. Biometric data like voiceprints are increasingly used for authentication, but unlike passwords, they cannot be changed once compromised.

References

Discussion: Commenters highlighted the irony of victims being asked to send their voice to another AI company for analysis, and noted that biometrics are effectively ‘forever passwords’ that cannot be rotated. The author emphasized the practical attack vectors, such as banking voiceprint bypass and Arup-style video call fraud.

Tags: #data breach, #AI safety, #deepfake, #privacy, #cybersecurity


China Blocks Meta’s Acquisition of AI Startup Manus ⭐️ 8.0/10

China has blocked Meta’s $2 billion acquisition of AI startup Manus, citing export control and national security laws, despite Manus being headquartered in Singapore. This marks a significant escalation in China’s use of extraterritorial legal authority over AI technology, setting a precedent that could reshape global tech M&A and AI governance. Manus was founded in China, later moved operations to Singapore after a $75 million funding round led by Benchmark in May 2025. China is seeking to annul the completed deal, invoking a catch-all clause in its export control law and offshore affiliate rules.

hackernews · yakkomajuri · Apr 27, 11:43

Background: China’s export control law, updated in 2020, allows the government to restrict transfers of sensitive technologies, including AI algorithms. The law includes a catch-all clause (Article 12) that can be applied to activities outside China if they threaten national security. This case parallels the TikTok dispute, where China asserted control over a Chinese-origin algorithm used abroad.

References

Discussion: Commenters debate whether China is applying its own version of US export controls, with some noting the irony given Meta and Google’s open AI contributions. Others question the legal basis for blocking a Singapore-based company and the implications for Singapore’s role as a tech hub.

Tags: #AI, #geopolitics, #regulation, #M&A, #China


Super ZSNES: GPU-Accelerated SNES Emulator Revived ⭐️ 8.0/10

The original developers of ZSNES, zsKnight and Demo, have released Super ZSNES, a completely rewritten SNES emulator that leverages GPU acceleration for enhanced performance and features like hi-res Mode 7 and uncompressed audio replacements. This marks a significant technical leap in SNES emulation, potentially offering higher accuracy and visual enhancements that benefit retro gaming enthusiasts and preservation efforts. Super ZSNES is rewritten from scratch and uses the GPU for rendering, enabling per-game enhancements and a modernized UI. It also supports uncompressed audio replacements, which could improve sound quality.

hackernews · haunter · Apr 27, 17:50

Background: ZSNES was one of the earliest and most popular SNES emulators, but its development stalled years ago. Traditional emulators rely on CPU-based software rendering, which can limit performance and accuracy. GPU-powered emulation offloads graphics processing to the graphics card, allowing for higher resolutions and effects.

References

Discussion: Community members expressed nostalgia for the original ZSNES and excitement about the GPU-powered rewrite. Some discussed technical implementation details, such as per-pixel rendering versus tile-by-tile approaches, and noted that the project could benefit from community contributions like uncompressed audio samples.

Tags: #emulation, #retro gaming, #GPU computing, #SNES


GitHub Copilot Moves to Usage-Based Billing ⭐️ 8.0/10

GitHub announced that Copilot will transition from a flat monthly subscription to a usage-based billing model, where users receive a monthly allowance of AI credits that can be topped up or rolled over. The change affects Copilot Pro ($10/month with $10 credits) and Pro+ ($39/month with $39 credits) plans. This pricing shift ends the era of subsidized inference for Copilot users, potentially increasing costs for heavy users and prompting many to evaluate alternatives like OpenRouter or DeepSeek. It reflects a broader industry trend of AI tool providers moving to consumption-based pricing to align costs with actual usage. The new model multipliers range from 1x for base models to 6x for GPT-4 and Sonnet, and up to 27x for Opus, meaning a single request could consume many credits. Users previously could get hundreds of dollars worth of Opus usage for $10/month, but now that usage would cost significantly more.

hackernews · frizlab · Apr 27, 16:03

Background: GitHub Copilot is an AI-powered code completion tool that suggests code snippets and functions in real-time. It was launched in 2021 and quickly became popular among developers. The original pricing was a flat monthly fee, which effectively subsidized heavy usage of expensive AI models like GPT-4 and Claude Opus.

Discussion: Community sentiment is largely negative, with users calling the change a ‘50x increase’ for heavy Opus users and questioning the value proposition. Many are considering switching to pay-as-you-go providers like OpenRouter or free alternatives like DeepSeek, citing the end of subsidized inference.

Tags: #GitHub Copilot, #pricing, #AI tools, #developer tools, #billing


Open-source CLI agent tops TerminalBench with Gemini-3-flash-preview ⭐️ 8.0/10

An open-source CLI agent called Dirac achieved a score of 65.2% on TerminalBench using Gemini-3-flash-preview, outperforming Google’s official agent (47.8%) and the previous top closed-source model Junie CLI (64.3%). This result demonstrates that a well-designed harness can significantly boost an AI agent’s performance, even surpassing official implementations and closed-source competitors, highlighting the importance of open-source innovation in AI agent development. The author explicitly states no cheating mechanisms were used, the agent was run in a leaderboard-compliant manner, and the full open-source version was used. The agent employs hash-anchored edits, AST-based context fetching, and batched operations.

hackernews · GodelNumbering · Apr 27, 12:35

Background: TerminalBench is a benchmark for evaluating AI agents on command-line interface tasks, testing their ability to navigate and manipulate real environments. Gemini-3-flash-preview is a fast and cost-effective model from Google, released in December 2025. The harness refers to the software framework that controls how the agent interacts with the environment and executes actions.

References

Discussion: Commenters praised the harness improvements and noted the lack of cross-model generalization testing. Some raised concerns about telemetry data being sent to the author’s endpoint, while others suggested comparing with different model families to rule out overfitting.

Tags: #AI agents, #benchmarking, #open source, #CLI, #LLM


Supreme Court Weighs Geofence Warrants’ Constitutionality ⭐️ 8.0/10

The US Supreme Court heard oral arguments on April 27, 2026, regarding whether police geofence warrants that compel tech companies to hand over cell location data violate the Fourth Amendment’s protection against unreasonable searches. This case could set a major precedent for digital privacy, affecting how law enforcement accesses location data from companies like Google, Apple, and Uber, and potentially reshaping the balance between surveillance and privacy rights. Google stopped responding to geofence warrants in 2025 by shifting location data storage to users’ devices, but other companies like Apple, Lyft, and Snapchat still receive such requests. The court’s decision will determine whether third-party data sharing eliminates reasonable expectation of privacy.

hackernews · unethical_ban · Apr 27, 15:29

Background: A geofence warrant is a search warrant that requires a company to identify all mobile devices within a specific geographic area during a certain time period. These warrants have surged in use since 2018, with Google’s Sensorvault database being a primary target. The Fourth Amendment generally requires a warrant based on probable cause, but courts have struggled with applying it to digital data shared with third parties.

References

Discussion: Commenters noted Google’s shift to on-device storage as a positive privacy move, but debated whether third-party data should lose privacy protection. Some compared geofence warrants to police reviewing security camera footage, while others expressed concern about the scope of surveillance and the justices’ apparent biases.

Tags: #privacy, #surveillance, #supreme court, #geofencing, #tech policy


Microsoft Open-Sources VibeVoice with Speaker Diarization ⭐️ 8.0/10

Microsoft released VibeVoice, an MIT-licensed speech-to-text model with built-in speaker diarization, on January 21, 2026. Simon Willison demonstrated running it on a Mac using mlx-audio, transcribing a 99-minute podcast in under 9 minutes. This open-source release makes advanced speaker-aware transcription accessible to developers, potentially accelerating applications in meeting transcription, podcast analysis, and accessibility tools. The MIT license encourages broad adoption and community contributions. The model requires significant memory, peaking at 61.5GB during prefill on a 128GB M5 Max MacBook Pro. It handles up to one hour of audio per run; longer recordings must be split with overlap to avoid errors.

rss · Simon Willison · Apr 27, 23:46

Background: Speaker diarization identifies ‘who spoke when’ in an audio stream, partitioning speech into segments by speaker identity. Traditional systems separate transcription and diarization into distinct pipelines, but VibeVoice integrates both into a single model, simplifying deployment. MLX is Apple’s machine learning framework optimized for Apple Silicon, and mlx-audio provides audio processing tools built on it.

References

Tags: #speech-to-text, #open-source, #machine learning, #Microsoft, #audio processing


Microsoft and OpenAI Drop AGI Clause from Partnership ⭐️ 8.0/10

On April 27, 2026, OpenAI and Microsoft announced the removal of the AGI clause from their partnership agreement, which previously would have nullified Microsoft’s IP rights upon AGI achievement. This change removes a key governance mechanism that tied AGI definition to commercial profits, potentially accelerating commercialization of OpenAI’s technology and reshaping the balance of power between the two companies. The clause originally defined AGI as systems capable of generating $100 billion in profit, but was later changed to be verified by an independent expert panel; the removal means Microsoft retains IP rights indefinitely.

rss · Simon Willison · Apr 27, 18:38

Background: The AGI clause was a unique feature of the Microsoft-OpenAI partnership, intended to ensure that if OpenAI achieved AGI, Microsoft would not have exclusive commercial rights. Over time, the definition of AGI evolved from a vague concept to a profit-based metric and then to expert verification. The removal reflects OpenAI’s push for more independence and flexibility in its cloud partnerships.

References

Tags: #OpenAI, #Microsoft, #AGI, #AI governance, #intellectual property


Zig 0.16.0 Introduces Structured Concurrency via Io Interface ⭐️ 8.0/10

Zig 0.16.0 expands the new Io interface, which provides async() and concurrent() functions that return Future objects with await() and cancel() methods, implementing structured concurrency. This design offers an explicit and verbose alternative to other languages’ implicit async models, potentially reducing bugs from cancellation and improving resource cleanup control in systems programming. The async() function can run a callback simultaneously in another thread, while concurrent() must run in another thread. Future methods are idempotent, allowing multiple await() or cancel() calls safely.

rss · LWN.net · Apr 27, 13:52

Background: Structured concurrency ensures that concurrent tasks are scoped within a parent context, preventing orphaned threads and improving error propagation. Zig’s approach passes an Io interface explicitly, similar to its allocator API, contrasting with Rust’s implicit async model which has faced issues with cancellation.

References

Tags: #Zig, #structured concurrency, #systems programming, #asynchronous


MiMo-V2.5-Pro: Open-Source AI Rivaling Claude Opus ⭐️ 8.0/10

Xiaomi released MiMo-V2.5-Pro, an open-source AI model under the MIT license, which is reported to rival Claude Opus in performance, especially in coding and long-context tasks. This release significantly advances open-source AI by providing a high-quality model that can run locally without corporate restrictions, potentially democratizing access to cutting-edge AI capabilities. MiMo-V2.5-Pro has over 1 trillion total parameters and supports a 1 million token context length, with an Elo score of 1426 on GDPval-AA benchmarks, surpassing many peer models.

reddit · r/LocalLLaMA · Namra_7 · Apr 27, 17:57

Background: Large language models (LLMs) like Claude Opus are typically proprietary and require cloud access. Open-source models under permissive licenses like MIT allow anyone to download, modify, and run them locally, reducing reliance on commercial APIs and enabling private, customizable AI applications.

References

Discussion: The community is highly enthusiastic, with many praising the model’s quality and MIT license. Some users note it excels in reasoning and long context but is weaker at coding agents. Hardware requirements are a concern, as running the full model demands multiple high-end GPUs.

Tags: #AI, #open-source, #LLM, #machine learning, #model release


Asahi Linux Nears Alpha Support for Apple M3 Chips ⭐️ 8.0/10

Asahi Linux has achieved near-Alpha support for Apple M3, M3 Pro, and M3 Max chips, enabling basic hardware functionality such as internal SSD, display, keyboard, trackpad, PCIe, and Bluetooth audio, though GPU acceleration remains unsupported. This marks a significant milestone in bringing Linux to Apple Silicon, expanding the ARM Linux ecosystem and providing users with an alternative operating system on the latest Mac hardware. GPU hardware acceleration is still missing, with graphics rendering currently handled by CPU software; additionally, the project is migrating the m1n1 bootloader to Rust for improved security and maintainability, and has already performed basic boot tests on M4 and M5 chips.

telegram · zaihuapd · Apr 27, 02:08

Background: Asahi Linux is a community-driven project that ports Linux to Apple Silicon Macs by reverse-engineering the undocumented hardware. The project released its first alpha installer in March 2022 for M1 devices, and has since expanded support to newer chips. Fedora Asahi Remix is the flagship distribution, providing a full Linux experience on Apple hardware.

References

Tags: #Asahi Linux, #Apple M3, #Linux on ARM, #open source, #hardware support


Xiaomi Xuanjie O1 Chip Shipments Exceed 1 Million ⭐️ 8.0/10

Xiaomi announced at its Investor Day that shipments of its self-developed 3nm Xuanjie O1 chip have surpassed 1 million units, with plans to apply future chips in Xiaomi cars and release annual upgrades. This milestone marks Xiaomi’s significant progress in self-developed chips, challenging established players and potentially reshaping the semiconductor landscape, especially with automotive applications. Xiaomi’s large-chip project restarted in 2021 with a planned investment of at least 50 billion yuan over 10 years, and R&D spending on Xuanjie has exceeded 13.5 billion yuan as of April 2025.

telegram · zaihuapd · Apr 27, 05:50

Background: The 3nm process is a leading-edge semiconductor manufacturing node that offers higher transistor density, better performance, and lower power consumption compared to previous nodes. Xiaomi’s Xuanjie O1 is a self-developed chip using this advanced process, indicating the company’s ambition to reduce reliance on external suppliers and compete in high-end chip design.

References

Tags: #semiconductor, #Xiaomi, #chip design, #3nm, #automotive


Easyduino: Open Source PCB Devboards for KiCad ⭐️ 7.0/10

Easyduino is a new open-source GitHub repository that provides KiCad PCB designs for popular microcontroller devboards like Arduino UNO, ESP32, and Raspberry Pi Pico, making it easy for hobbyists and educators to learn PCB design. This project lowers the barrier to entry for learning PCB design by offering proven, open-source templates that can be modified and reused, fostering hands-on education and innovation in hardware development. The repository includes designs for Arduino UNO, Arduino Nano, ESP32, ESP32-S3, Raspberry Pi Pico (RP2040), and STM32 Bluepill, all unified in KiCad with USB-C support and a 4-layer copper stackup optimized for JLCPCB manufacturing.

hackernews · Hanqaqa · Apr 27, 17:45

Background: KiCad is a free and open-source electronic design automation (EDA) suite used for schematic capture and PCB layout. Designing a custom PCB from scratch can be daunting for beginners, as it requires knowledge of component placement, routing, and manufacturing constraints. Easyduino provides ready-to-use reference designs that demonstrate best practices, allowing learners to start with a working board and modify it for their own projects.

References

Discussion: The community response has been overwhelmingly positive, with users praising the project as an amazing resource that fills a gap for open-source devboard designs. Some users shared their own experiences designing similar boards, while others asked for advice on introducing PCB design to children, indicating strong educational interest.

Tags: #PCB design, #KiCad, #Arduino, #open source hardware, #education


macOS 27 to enforce TLS 1.2, deprecate AFP for Time Machine ⭐️ 7.0/10

macOS 27 will require TLS 1.2 for certain network connections and deprecate the Apple Filing Protocol (AFP) for Time Machine backups, pushing users to adopt SMB instead. This change enhances security by enforcing modern encryption standards and streamlines network file sharing, but may disrupt users relying on older AFP-based Time Machine setups, especially those with legacy hardware like Time Capsules. The TLS 1.2 requirement applies to connections to certain servers, and AFP deprecation means Time Machine backups over AFP will no longer be supported, requiring migration to SMB. Users with Time Capsules may need to run custom Samba builds to continue using them.

hackernews · pvtmert · Apr 27, 15:36

Background: AFP is a proprietary Apple protocol for file sharing, while SMB is a widely used cross-platform protocol. TLS 1.2 is a cryptographic protocol that ensures secure communications; older versions like TLS 1.1 are considered insecure. Apple has been gradually moving away from AFP for years.

References

Discussion: Commenters expressed mixed reactions: some noted that TLS 1.2 is long overdue, while others recalled past networking issues when Apple rewrote mDNSResponder. A few users shared workarounds, such as building Samba 4 for Time Capsules, and others lamented the unpolished state of Time Machine animations.

Tags: #macOS, #networking, #TLS, #Time Machine, #SMB


The Lost Art of Staring at Walls ⭐️ 7.0/10

A personal essay argues that smartphones have eliminated the idle moments once used for mind-wandering, a state called ‘disattention’. This reflection highlights a subtle but profound loss in the smartphone era: the ability to let the mind wander freely, which is crucial for creativity and mental well-being. The essay uses the phrase ‘men who stare at walls’ to describe people lost in thought, contrasting it with today’s constant phone-checking. Community comments compare this state to meditation and discuss its benefits.

hackernews · aselimov3 · Apr 27, 11:08

Background: The concept of ‘disattention’ refers to unstructured downtime where the mind is free to wander without external stimulation. In the pre-smartphone era, such moments were common during waiting, commuting, or idle breaks. Constant smartphone use has filled these gaps, potentially reducing opportunities for creative thinking and self-reflection.

Discussion: Commenters largely agree with the essay’s premise, sharing personal experiences of staring at walls and the benefits of mind-wandering. Some debate whether this state is a form of meditation, while others caution that it is distinct from meditation and not a cure for attention deficits.

Tags: #attention, #smartphones, #meditation, #mental health, #technology


The Woes of Sanitizing SVGs ⭐️ 7.0/10

A detailed article explains why sanitizing SVGs is extremely difficult due to their complex features like scripts, event handlers, and CSS, and recommends using Content Security Policy (CSP) as the only reliable fix for HTTP leak problems. This matters because many web applications allow SVG uploads or rendering, and incomplete sanitization can lead to XSS and data exfiltration. The article provides practical CSP advice that can help developers secure their applications. The article notes that SVG supports inline scripts, event handlers, CSS, and foreign objects, making sanitization a whack-a-mole problem. It recommends CSP headers or tags as the only reliable defense against HTTP leaks.

hackernews · varun_ch · Apr 27, 15:31

Background: SVG (Scalable Vector Graphics) is an XML-based vector image format that can include interactive elements like scripts and CSS. Sanitization is the process of removing dangerous content from user-submitted SVGs to prevent security vulnerabilities like cross-site scripting (XSS). Content Security Policy (CSP) is a browser security mechanism that restricts which resources can be loaded and executed on a page.

References

Discussion: Commenters agree that CSP is the only credible fix, with one suggesting a sandbox attribute for inline SVGs. Another proposes supporting only a safe subset of SVG features, while a third mentions the HTML Sanitizer API’s limited SVG support.

Tags: #web security, #SVG, #CSP, #sanitization, #frontend


Why Not Just Use Lean? A Critical Comparison ⭐️ 7.0/10

Lawrence Paulson published a blog post critically examining why Lean may not be the best choice for all proof assistant tasks, comparing it with Coq, Agda, and Isabelle/HOL. This analysis helps practitioners choose the right proof assistant for their needs, highlighting trade-offs in automation, dependent types, and community support. The article argues that Lean’s retention of proof objects can lead to large terms, though community comments note this is a misconception. Lean is praised for being decent at everything and having a large community.

hackernews · ibobev · Apr 27, 14:24

Background: Proof assistants like Lean, Coq, Agda, and Isabelle/HOL are software tools for writing and verifying mathematical proofs. They are based on different type theories and offer varying levels of automation and expressiveness. Lean is based on the calculus of constructions with inductive types and has a large mathematical library (mathlib).

References

Discussion: Commenters debated Lean’s strengths: some find it clunky compared to Agda for functional programming, while others praise its balance. A key correction was that Lean does not retain full proof objects, contrary to the article’s claim.

Tags: #proof assistants, #Lean, #functional programming, #formal verification


pip 26.1 Released with Dependency Cooldowns and Pylock Support ⭐️ 7.0/10

pip 26.1 has been released, introducing dependency cooldowns, experimental support for pylock.toml files, resolver improvements, security fixes, and the end of Python 3.9 support. This release is significant for Python developers as it improves dependency resolution reliability, enhances reproducibility through pylock support, and addresses security vulnerabilities, while dropping Python 3.9 encourages migration to newer versions. Dependency cooldowns allow pip to temporarily avoid re-resolving recently resolved dependencies, reducing unnecessary work. The experimental pylock support implements PEP 751 for reproducible installations, and resolver improvements lift several limitations of the 2020 resolver, moving closer to removing the legacy resolver.

rss · LWN.net · Apr 27, 18:45

Background: pip is the standard package installer for Python. The 2020 resolver introduced stricter dependency conflict detection, but had limitations that pip 26.1 addresses. Dependency cooldowns are a feature inspired by uv, a fast Python package manager. pylock.toml is a new lock file format defined in PEP 751 to ensure reproducible installations across environments.

References

Tags: #pip, #Python, #package management, #release


Linux 7.1 Merge Window Closes with 13,000 Changesets ⭐️ 7.0/10

The Linux 7.1 merge window closed with 12,996 non-merge changesets from 2,011 developers, including 342 first-time contributors. Notable additions include seccomp() support for Alpha, high-memory support for Loongarch, and the beginning of sub-scheduler support in sched_ext. This merge window shows continued growth in the Linux kernel developer community, with a rising number of first-time contributors. The new features, such as sub-schedulers and improved swap subsystem, will enhance performance and flexibility for a wide range of users. The merge window saw over 9,000 changesets arrive after the first-half summary, with a focus on driver updates. Key technical changes include the removal of execute-in-place support for RISC-V due to lack of maintainers, and the addition of remote ring buffers for tracing in virtual machines.

rss · LWN.net · Apr 27, 16:39

Background: The Linux kernel development cycle begins with a two-week merge window during which new features are accepted into the mainline. After the merge window closes, a stabilization period follows, leading to the final release. The 7.1 merge window is part of the ongoing development of the Linux kernel, which is maintained by Linus Torvalds and a global community of developers.

Tags: #Linux kernel, #merge window, #open source, #development


Microsoft Open-Sources TRELLIS.2: 4B-Parameter Image-to-3D Model ⭐️ 7.0/10

Microsoft has released TRELLIS.2, an open-source 4-billion-parameter image-to-3D generative model that produces high-fidelity 3D assets with full PBR materials at resolutions up to 1536³, using a novel O-Voxel structure and native 3D VAEs with 16× spatial compression. This open-source release democratizes access to state-of-the-art 3D generation, enabling developers and artists to create high-quality 3D assets from single images efficiently, which could accelerate workflows in gaming, VR/AR, and digital content creation. The model leverages a field-free sparse voxel structure called O-Voxel to handle complex topologies and sharp features, and it uses native 3D VAEs with 16× spatial compression for efficient latent representation. The code and a live demo are available on GitHub and Hugging Face.

reddit · r/LocalLLaMA · 44th–Hokage · Apr 27, 20:07

Background: Image-to-3D generation aims to create 3D models from a single 2D image, which is challenging due to the need to infer depth, geometry, and materials. Traditional methods often rely on neural radiance fields or signed distance functions, but TRELLIS.2 introduces a compact latent space via O-Voxel and 3D VAEs to improve efficiency and fidelity. PBR (physically based rendering) materials simulate realistic light interactions, enhancing visual quality.

References

Discussion: Community comments note that the model was actually released four months ago, reducing the urgency of the news. Some users reported difficulty replicating the high-quality results shown in demos, and there are requests for ROCm support, with an approved pull request but ongoing compatibility issues on AMD GPUs.

Tags: #3D generation, #image-to-3D, #open-source, #generative AI, #computer vision


4B LLM Benchmark: Nemotron 3 Nano Tops Finance, Reasoning, Code ⭐️ 7.0/10

A community benchmark compared five 4B-parameter LLMs (Gemma 4, Qwen 3.5, Granite 4, Nemotron Nano, Phi-4 Mini) on 39 tasks covering finance, reasoning, and code, with Nemotron 3 Nano achieving the highest overall score of 85%. This benchmark provides a timely, head-to-head comparison of the latest small LLMs, highlighting Nemotron 3 Nano’s strong finance performance and sparking debate about benchmark methodology and token budget fairness. The benchmark used a 1024-token max output, temperature=0, seed=42, and median aggregation over 3 trials per task. Nemotron 3 Nano scored 100% on finance, 80% on reasoning, and 67% on code, while Qwen 3.5 scored only 15% overall, likely due to inadequate thinking token budget.

reddit · r/LocalLLaMA · FederalAnalysis420 · Apr 27, 18:36

Background: 4B-parameter LLMs are small enough to run on consumer hardware like laptops, making them attractive for local deployment. The benchmark tested models from Google, Alibaba, IBM, NVIDIA, and Microsoft, all released between late 2024 and early 2026. Token budget is critical for thinking models that use internal reasoning tokens before outputting an answer.

References

Discussion: Commenters questioned the 1024-token budget, arguing it unfairly penalizes thinking models like Qwen 3.5. Some suggested disabling thinking or using external budget control. Others noted that Phi-4 Mini’s high code score despite being older indicates potential benchmark flaws.

Tags: #LLM, #benchmark, #local LLM, #model comparison, #AI


Boost VRAM for LLMs by Adding an Old GPU ⭐️ 7.0/10

A Reddit user discovered that combining a 16GB RTX 5070 Ti with an old 6GB RTX 2060 via Vulkan in llama.cpp allows running larger LLMs (e.g., Qwen3.6-27B) that would otherwise not fit in 16GB VRAM, achieving 22GB total VRAM. This practical tip enables users with limited VRAM (e.g., 16GB) to run larger dense models without buying an expensive 24GB card, democratizing access to high-end LLMs for enthusiasts and researchers. The setup uses llama-server with dev=Vulkan1,Vulkan2 to split the model across GPUs, and requires no-mmap and mlock=false to keep the model off RAM. Performance may be bottlenecked by the slower card (e.g., 20 t/s vs 30 t/s on a single 3090 Ti).

reddit · r/LocalLLaMA · akira3weet · Apr 27, 10:20

Background: Large language models (LLMs) require significant VRAM to run entirely on GPU, which is much faster than CPU+RAM. Many users have only 16GB VRAM, insufficient for 30B+ parameter models. llama.cpp supports splitting models across multiple GPUs using Vulkan or CUDA, but conventional wisdom suggests using identical GPUs for best performance.

References

Discussion: Community members largely agree with the tip, with some sharing their own multi-GPU setups (e.g., 3060 as bonus to 3x3090). However, several users note performance bottlenecks: Mysterious_Role_8852 reports a drop from 30 t/s to 20 t/s when adding a 2070 to a 3090 Ti. Others raise practical concerns about power, space, and slot bandwidth.

Tags: #LLM, #VRAM, #GPU, #hardware, #llama.cpp


Skymizer Unveils 384GB PCIe Card for 700B LLM Inference ⭐️ 7.0/10

Skymizer Taiwan Inc. announced a breakthrough PCIe card that uses six HTX301 chips and 384GB of memory to run 700B-parameter LLM inference locally at approximately 240W, by offloading the decode stage from the GPU to the HTX301 cards. This architecture could enable enterprises to run very large models on a single card without requiring expensive multi-GPU setups, potentially democratizing access to large-scale LLM inference at lower power and cost. The card splits LLM inference into two phases: the GPU handles compute-intensive prefill, while the HTX301 cards handle memory-bandwidth-intensive decode. The memory is not HBM but packaged like GDDR6/6X/7, and bandwidth details have not been disclosed.

reddit · r/LocalLLaMA · lurenjia_3x · Apr 27, 12:56

Background: LLM inference consists of two stages: prefill (processing the input prompt) and decode (generating tokens one by one). Decode is memory-bandwidth-bound, while prefill is compute-bound. Traditional GPUs handle both, but Skymizer’s approach dedicates specialized hardware to each stage to improve efficiency.

References

Discussion: The community expressed skepticism about undisclosed bandwidth, potential high cost (tens of thousands of dollars), and past vaporware promises. Some questioned whether shuttling weights between cards would introduce latency, while others noted the lack of concrete performance data.

Tags: #LLM inference, #hardware architecture, #AI hardware, #edge AI


Microsoft’s Windows K2 Plan Aims to Revamp Windows 11 ⭐️ 7.0/10

Microsoft is internally advancing a long-term initiative called ‘Windows K2’ that aims to overhaul Windows 11 by focusing on performance, reliability, and community-driven improvements, including a WinUI 3-based start menu that is 60% faster and ad-free, reduced update frequency to once per month, and gaming performance parity with SteamOS within one to two years. This plan signals a strategic shift from rapid feature releases to quality-first development, potentially restoring user trust in Windows 11 by addressing long-standing complaints about performance, bloatware, and intrusive updates. If successful, it could make Windows 11 more competitive with gaming-focused platforms like SteamOS and improve the overall user experience for millions of PC users. The K2 initiative is built on four pillars: performance, engineering, reliability, and community feedback. Key technical details include rebuilding the Start Menu with WinUI 3 for a 60% speed boost and removing ads, reducing system update reboots to once per month, and optimizing system footprint to lower idle memory usage.

telegram · zaihuapd · Apr 27, 10:31

Background: Windows 11 has faced criticism for performance issues, a cluttered Start Menu with ads, and frequent updates that require reboots. WinUI 3 is Microsoft’s modern native UI framework for Windows desktop apps, decoupled from the OS for faster updates. SteamOS is Valve’s Linux-based gaming OS optimized for handhelds like the Steam Deck, known for its gaming performance and ease of use.

References

Tags: #Windows 11, #Microsoft, #Operating Systems, #WinUI 3, #Gaming


Chinese Tech Giants Face User Backlash Over AI Content ⭐️ 7.0/10

NetEase’s game Identity V faced a player boycott on April 23 due to suspected AI-generated posters with anatomical errors, while iQiyi sparked a portrait rights dispute by opening its ‘artist library’ for AIGC creators, and ByteDance’s Fanqie Novel began cracking down on low-quality AI-generated content after a training protocol controversy. These incidents highlight the growing tension between AI-driven efficiency and user/creator rights in China’s tech industry, potentially forcing platforms to rethink their AI deployment strategies to avoid alienating their user base. NetEase has deployed an AI-native pipeline in art and programming since 2025, boosting efficiency by up to 300% in some areas, but the backlash shows that AI-generated content quality and ethical concerns remain unresolved.

telegram · zaihuapd · Apr 27, 12:22

Background: AI-native pipeline refers to a development workflow where AI is deeply integrated into content creation, such as generating graphics, sounds, and scripts from text prompts. ‘Tingke’ (stop spending) is a term used by Chinese gamers to protest by ceasing in-game purchases. AIGC (AI-Generated Content) involves using AI to produce creative works, raising copyright and portrait rights issues.

References

Discussion: The Telegram channel discussion likely includes diverse viewpoints on AI ethics and industry impact, with some users criticizing the platforms for prioritizing efficiency over quality and rights, while others acknowledge the potential of AI but call for better regulation.

Tags: #AI Ethics, #Content Moderation, #Gaming, #Copyright, #China Tech