From 46 items, 17 important content pieces were selected
- Meta’s smart glasses workers report extensive data access as company plans facial recognition launch ⭐️ 8.0/10
- First in-utero stem cell therapy for fetal spina bifida repair proven safe in study ⭐️ 8.0/10
- Linux kernel developers debate approaches to implement atomic buffered writes ⭐️ 8.0/10
- Motorola partners with GrapheneOS Foundation to enhance smartphone security ⭐️ 8.0/10
- Qwen3.5-9B Released: A 9B Parameter Vision-Language Model with Hybrid Gated DeltaNet/Attention Architecture ⭐️ 8.0/10
- StepFun releases two base models and training framework for Step 3.5 Flash ⭐️ 8.0/10
- Researchers Reverse-Engineer Apple M4 Neural Engine, Reveal 38 TOPS Marketing as Misleading ⭐️ 8.0/10
- Encrypted Client Hello (ECH) Protocol Completes Final IETF Approval, RFC 9849 Nears Publication ⭐️ 8.0/10
- Technical deep dive explains TCP zero-copy networking in Linux kernel ⭐️ 7.0/10
- The ‘Exploitation Paradox’ in Open Source: How Loopholes Threaten FOSS Freedoms ⭐️ 7.0/10
- Qwen releases new small-scale 3.5 models (0.8B, 2B, 9B) for resource-constrained hardware ⭐️ 7.0/10
- Qwen 3.5 0.8B multimodal model runs locally in browser via WebGPU and Transformers.js ⭐️ 7.0/10
- Qwen3.5-0.8B runs locally on a 7-year-old Samsung phone at 12 tokens/second. ⭐️ 7.0/10
- Qwen3.5’s 9B and 4B models achieve benchmark performance surpassing older, much larger models. ⭐️ 7.0/10
- Qwen 3.5 2B model demonstrates exceptional OCR capabilities for diverse text types. ⭐️ 7.0/10
- LM Studio’s parser silently breaks Qwen3.5 tool calling and reasoning, connecting year-long bug reports ⭐️ 7.0/10
- Xiaomi’s Humanoid Robot Deployed in Auto Factory for Die-Cast Part Assembly ⭐️ 7.0/10
Meta’s smart glasses workers report extensive data access as company plans facial recognition launch ⭐️ 8.0/10
According to a New York Times report based on internal documents, Meta plans to introduce facial recognition to its Ray-Ban Meta smart glasses during a distracting political environment when critics are focused elsewhere. Simultaneously, workers involved with the glasses have reported having extensive access to user data, including images and audio captured by the devices. This represents a significant escalation in the privacy risks associated with wearable technology, as facial recognition on always-on glasses could enable pervasive, real-time surveillance in public and private spaces. The strategic timing of the feature’s launch, coupled with insider reports of broad data access, raises serious ethical questions about corporate transparency and user consent in the age of ambient computing. The internal document cited by The Times explicitly states Meta intends to launch the feature “during a dynamic political environment where many civil society groups that we would expect to attack us would have their resources focused on other concerns.” The current Ray-Ban Meta glasses, which require connection to a smartphone app, feature 12 MP cameras and Meta AI with multimodal computer vision capabilities, as per their technical specifications.
hackernews · sandbach · Mar 2, 22:32
Background: Smart glasses like the Ray-Ban Meta are wearable devices that blend cameras, microphones, and displays into eyewear, often powered by AI for tasks like translation or object identification. Facial recognition is a biometric technology that uses algorithms to identify individuals based on their facial features, and its integration into consumer wearables has been controversial due to privacy and surveillance concerns. Previous attempts like Google Glass faced significant public backlash over similar privacy issues, often being labeled as “creepy” technology.
Discussion: The community expresses strong concern and cynicism, drawing parallels to the failure of Google Glass due to social stigma and privacy fears. Comments highlight perceived corporate hypocrisy, noting Meta CEO Mark Zuckerberg’s own practice of taping his laptop webcam while his company pushes always-on cameras. Others connect the issue to broader geopolitical surveillance, sarcastically suggesting there are “no downsides” to proliferating networked sensors, and urge consumers to “vote with your dollars.”
Tags: #privacy, #facial-recognition, #surveillance, #meta, #ethics
First in-utero stem cell therapy for fetal spina bifida repair proven safe in study ⭐️ 8.0/10
A pioneering clinical study has demonstrated the safety of the first-ever in-utero stem cell therapy for repairing fetal spina bifida, a major congenital neural tube defect. The therapy, part of the ongoing “CuRe Trial: Cellular Therapy for In Utero Repair of Myelomeningocele,” represents a novel approach beyond traditional fetal surgery. This breakthrough matters because it offers a potential path to a more complete repair of the spinal cord defect before birth, which could significantly improve long-term neurological outcomes and quality of life for affected children. It represents a major step forward in fetal regenerative medicine, moving beyond merely closing the physical gap to potentially restoring neural tissue. The therapy is administered in utero, typically between 19 and 26 weeks of gestation, and involves using stem cells, potentially amniotic fluid-derived mesenchymal stem cells (AFMSCs), to aid in the repair. While the initial findings focus on safety, the full clinical trial (CuRe Trial) is expected to continue until around 2030 to evaluate long-term efficacy.
hackernews · gmays · Mar 2, 14:54
Background: Spina bifida is a congenital neural tube defect where the spinal column fails to close completely during fetal development, often leading to nerve damage, paralysis, and other complications. Traditional fetal surgery for the most severe form, myelomeningocele, involves opening the uterus to surgically close the back lesion, which carries risks for both mother and fetus. In-utero stem cell therapy is an experimental strategy aiming to use regenerative cells to repair the defect, potentially offering functional improvement beyond physical closure.
References
Discussion: Community comments express profound hope and emotional resonance, with individuals sharing personal connections to spina bifida and related conditions. Contributors highlight the wide spectrum of severity, the lifelong impact on families, and the historical context of limited options, viewing this research as a transformative advance that could prevent suffering and improve quality of life.
Tags: #medical-research, #stem-cells, #fetal-surgery, #spina-bifida, #biotech
Linux kernel developers debate approaches to implement atomic buffered writes ⭐️ 8.0/10
A discussion initiated in February 2026 by Pankaj Raghav highlights that while ext4 and XFS now support atomic direct I/O, atomic buffered I/O remains unimplemented despite multiple proposals. The conversation reveals ongoing disagreement about the necessity and complexity of this feature, with PostgreSQL cited as a primary potential user. Atomic buffered writes are crucial for ensuring data integrity in applications like databases that write multi-block data, preventing partial “torn writes” that can corrupt data. Implementing this feature would benefit important workloads like PostgreSQL that rely on buffered I/O for performance or memory management reasons, addressing a long-standing gap in Linux filesystem capabilities. Two major patch sets have been proposed but stalled: one from John Garry in 2024 and another more recent one from Ojaswin Mujoo. The main concerns center on the added complexity to I/O paths and debates about whether the feature is truly needed, with some developers suggesting applications should migrate to direct I/O instead.
rss · LWN.net · Mar 2, 22:27
Background: Atomic writes ensure that multi-block data operations either complete entirely or fail completely, preventing “torn writes” where only part of the data is written. Direct I/O bypasses the kernel’s page cache, while buffered I/O uses it for performance. Some filesystems like ext4 and XFS already support atomic direct I/O, but extending this to buffered I/O is more complex due to interactions with the page cache and writeback mechanisms.
Discussion: The discussion reveals a split in the developer community: PostgreSQL developer Andres Freund argues that many users cannot benefit from direct I/O due to memory constraints or performance reasons, creating a legitimate need for atomic buffered writes. However, developers like Christoph Hellwig contend that helping PostgreSQL move off buffered I/O would be preferable to adding complex kernel special cases.
Tags: #linux-kernel, #filesystems, #storage, #io, #atomic-operations
Motorola partners with GrapheneOS Foundation to enhance smartphone security ⭐️ 8.0/10
In March 2026, Motorola announced a partnership with the GrapheneOS Foundation to collaborate on strengthening smartphone security and engineering future devices with GrapheneOS compatibility, including plans to release a smartphone with GrapheneOS pre-installed. This partnership marks a significant step toward mainstream adoption of enhanced mobile security, as a major smartphone manufacturer aligns with a leading privacy-focused Android distribution, potentially setting new industry standards for secure devices. GrapheneOS is a free, open-source Android-based OS focused on privacy and security, historically developed for Google Pixel devices. The partnership indicates Motorola’s commitment to offering devices with enhanced security out-of-the-box, moving beyond the niche custom ROM community.
rss · LWN.net · Mar 2, 14:58
Background: GrapheneOS is a security-hardened, privacy-focused operating system built on the Android Open Source Project (AOSP). It is developed by the non-profit GrapheneOS Foundation and is known for its strong security enhancements, such as improved sandboxing and verified boot, while maintaining compatibility with Android apps. Custom Android distributions like GrapheneOS and LineageOS are typically installed by users to replace the manufacturer’s OS, but partnerships with device makers for pre-installation are rare.
References
Tags: #mobile-security, #android, #grapheneos, #industry-partnership
Qwen3.5-9B Released: A 9B Parameter Vision-Language Model with Hybrid Gated DeltaNet/Attention Architecture ⭐️ 8.0/10
The Qwen team has released Qwen3.5-9B, a novel 9-billion parameter vision-language model. It features a hybrid architecture that combines Gated DeltaNet layers with standard Gated Attention layers in an 8 × (3 × (Gated DeltaNet → FFN) → 1 × (Gated Attention → FFN)) pattern and demonstrates strong performance on benchmarks, reportedly surpassing some larger predecessor models. This release is significant because it provides a highly capable, mid-sized model that is efficient enough to run on consumer-grade hardware like 16GB GPUs, making advanced vision-language AI more accessible. Its novel hybrid architecture, which outperforms larger models in some benchmarks, represents a meaningful step forward in model efficiency and design for the open-source AI community. The model natively supports a context length of 262,144 tokens and is extensible up to 1,010,000 tokens. It was trained using multi-step training (MTP), and quantized versions in the GGUF format are already available, which is crucial for deployment on resource-constrained hardware.
reddit · r/LocalLLaMA · jacek2023 · Mar 2, 12:33
Background: Gated DeltaNet is a recently proposed architecture that combines the gating mechanism from Mamba2 with the delta rule from DeltaNet, aiming to improve upon linear attention models for better efficiency and performance. The GGUF (Generic GPT Unified Format) is a binary file format specifically designed for storing and efficiently running quantized large language models, enabling them to run on less powerful hardware. Multi-step training (MTP) is a technique where models are trained to predict multiple future tokens simultaneously, which can improve inference speed and model generalization.
References
Discussion: The community reaction is overwhelmingly positive, with excitement focused on the model’s accessibility for users with 16GB GPUs and its surprisingly strong benchmark performance compared to larger models. Key discussion points include requests for various quantization options (“QUANTS PLEASE”), questions about which specific quantized version is best for 16GB VRAM, and speculation about the architectural innovations that allow a 9B model to outperform much larger predecessors.
Tags: #large-language-models, #computer-vision, #model-architecture, #open-source, #quantization
StepFun releases two base models and training framework for Step 3.5 Flash ⭐️ 8.0/10
StepFun has released two base models for its Step 3.5 Flash large language model and has also open-sourced its SteptronOSS training framework. This release provides the foundational components and the training pipeline used to develop the model. This is a significant open-source contribution that provides researchers and developers with the building blocks and methodology behind a frontier-level AI model. Releasing the training framework, in particular, enhances transparency and reproducibility, allowing the community to study, adapt, and build upon StepFun’s work. The Step 3.5 Flash model is described as offering ‘Open Frontier-Level Intelligence’ with 11 billion active parameters. The released base models likely serve as the core architecture or checkpoints from which the full model is fine-tuned or scaled.
reddit · r/LocalLLaMA · tarruda · Mar 2, 20:57
Background: StepFun is an AI lab known for developing large-scale models, having previously launched the trillion-parameter Step-2 LLM and other models like Step-1.5V. A ‘base model’ typically refers to a pre-trained, general-purpose language model before it is specialized for specific tasks. A training framework is the software infrastructure and methodology used to train such models, encompassing data handling, optimization algorithms, and distributed computing strategies.
References
Discussion: The community reaction is overwhelmingly positive and enthusiastic. Commenters praise the release as “amazing” and “huge for OSS,” specifically highlighting the value of open-sourcing the training pipeline (SteptronOSS). One user expressed hope for a future model update, indicating ongoing interest in StepFun’s developments.
Tags: #AI, #open-source, #machine-learning, #model-training, #LLM
Researchers Reverse-Engineer Apple M4 Neural Engine, Reveal 38 TOPS Marketing as Misleading ⭐️ 8.0/10
Researchers, collaborating with AI under the name maderix, successfully reverse-engineered Apple’s M4 Neural Engine (ANE) by bypassing the CoreML framework and directly accessing the private _ANEClient interface. Their benchmarks revealed the ANE’s actual FP16 peak performance is 19 TFLOPS, not the marketed 38 TOPS, and discovered architectural details including 32MB of SRAM and significant performance gains when bypassing CoreML. This independent verification exposes a significant discrepancy between Apple’s marketing claims and the hardware’s actual capabilities, challenging industry-standard performance reporting practices. The findings have major implications for AI hardware benchmarking accuracy and for developers optimizing neural network workloads on Apple Silicon, especially for mobile and edge AI applications. The research found that the claimed 38 TOPS figure is derived by doubling the FP16 performance, an industry convention, but the hardware does not deliver double the throughput for INT8 operations compared to FP16. Notably, bypassing CoreML can increase throughput for small operations by 2-4x, and the ANE’s convolution operations are three times faster than its matrix multiplication.
telegram · zaihuapd · Mar 2, 08:00
Background: Apple’s Neural Engine (ANE) is a dedicated AI accelerator core integrated into its M-series chips, designed to efficiently handle machine learning tasks like image recognition and natural language processing. TOPS (Tera Operations Per Second) and TFLOPS (Tera Floating-Point Operations Per Second) are common metrics for measuring AI accelerator performance, with INT8 operations often theoretically offering double the throughput of FP16 due to lower precision. CoreML is Apple’s framework for deploying machine learning models on its platforms.
References
Tags: #hardware-reverse-engineering, #ai-accelerators, #apple-silicon, #benchmarking, #neural-networks
Encrypted Client Hello (ECH) Protocol Completes Final IETF Approval, RFC 9849 Nears Publication ⭐️ 8.0/10
The Encrypted Client Hello (ECH) protocol has completed the final AUTH48 approval stage by all authors, IANA, and area directors in late February 2026, after seven years and 25 draft revisions. It has been assigned RFC number 9849 and is pending the resolution of one final GitHub technical issue (#1308) before official publication. This final approval marks the culmination of a major effort to close a significant privacy gap left by TLS 1.3 in 2018. The widespread implementation support from major browsers and platforms like Chrome, Firefox, Android, and Cloudflare means ECH is poised to significantly enhance user privacy across the internet by encrypting previously exposed handshake metadata. The protocol’s core function is to encrypt previously plaintext metadata in the TLS handshake, specifically the Server Name Indication (SNI) and Application-Layer Protocol Negotiation (ALPN). Major implementations are already in place, with Chrome, Firefox, and Android supporting ECH on the client side, and Cloudflare having deployed server-side support by the end of 2024.
telegram · zaihuapd · Mar 2, 10:28
Background: In a traditional TLS handshake, the initial ClientHello message is sent in plaintext, revealing the Server Name Indication (SNI) which indicates the specific website a client intends to connect to. This SNI leak allows network observers, like Internet Service Providers or those on the same network, to see which domains a user is visiting, even if the subsequent connection is encrypted. TLS 1.3, finalized in 2018, greatly improved security and performance but did not address this SNI privacy leak, which ECH was designed to solve.
References
Tags: #TLS, #Privacy, #IETF, #Network Security, #Protocols
Technical deep dive explains TCP zero-copy networking in Linux kernel ⭐️ 7.0/10
Toke Høiland-Jørgensen published a detailed technical overview explaining how TCP zero-copy networking works in the Linux kernel, specifically focusing on the memory management and asynchronous notification mechanisms involved. The article describes how the sendmsg() syscall operates asynchronously, returning immediately while the kernel later notifies userspace when memory buffers can be reused. This matters because TCP zero-copy is a critical performance optimization that reduces CPU overhead and memory bandwidth consumption in high-throughput networking applications like web servers, databases, and streaming services. Understanding these kernel mechanisms helps developers optimize network-intensive applications and contributes to the broader ecosystem of high-performance computing. The implementation requires userspace applications to keep memory buffers unmodified until the kernel sends a completion notification, as data is transferred directly from userspace to the network device. This optimization is enabled by the MSG_ZEROCOPY flag for socket send calls and is currently supported for TCP, UDP, and VSOCK sockets.
rss · LWN.net · Mar 2, 20:12
Background: Traditional network I/O involves copying data from user-space buffers to kernel-space buffers before transmission, which consumes CPU cycles and memory bandwidth. Zero-copy networking aims to eliminate these redundant copies by allowing data to be transferred directly from user memory to the network interface card (NIC). In Linux, this is achieved through mechanisms like MSG_ZEROCOPY and careful management of buffer lifecycles between the kernel and userspace.
References
Tags: #linux-kernel, #networking, #systems-programming, #performance-optimization, #tcp
The ‘Exploitation Paradox’ in Open Source: How Loopholes Threaten FOSS Freedoms ⭐️ 7.0/10
At CfgMgmtCamp 2026 in Ghent, lawyer and FOSS licensing expert Richard Fontana presented the concept of the ‘exploitation paradox’ in open source, describing the recurring pattern where actors exploit legal and governance loopholes to restrict freedoms or gain advantage. He argued that attempts to close these loopholes and maintain freedom must look beyond traditional licensing approaches. 这很重要,因为它指出了FOSS治理中一个系统性的脆弱点,这种脆弱性随着每次技术变革(从Linux到云计算再到AI)而反复出现,威胁着定义开源的核心自由。理解这一悖论对于社区、公司和政策制定者来说至关重要,他们需要在不断变化的技术和商业环境中保护开源原则。 Fontana noted that foundational definitions of freedom (like the FSF’s four freedoms and the OSI’s Open Source Definition) remain static while the technical, social, and economic ‘infrastructure’ of software evolves, creating tension. He also highlighted that gatekeepers of these definitions are often reluctant to revise them, making open source a ‘conservative domain’ in this respect.
rss · LWN.net · Mar 2, 15:28
Background: Free and Open Source Software (FOSS) is built on principles that grant users freedoms to use, study, modify, and share software. The Free Software Foundation’s ‘four essential freedoms’ and the Open Source Initiative’s Open Source Definition are foundational documents that establish these norms. CfgMgmtCamp is an annual conference focused on configuration management and infrastructure, though it has expanded to include broader topics in open source and DevOps.
Tags: #open-source, #governance, #software-licensing, #community, #foss
Qwen releases new small-scale 3.5 models (0.8B, 2B, 9B) for resource-constrained hardware ⭐️ 7.0/10
Alibaba Cloud’s Qwen team has released new small-scale models in the Qwen 3.5 family, specifically the 0.8B, 2B, and 9B parameter versions, designed to deliver strong performance on limited hardware. The community immediately began quantizing and testing these models, with various quantized versions appearing on Hugging Face shortly after the announcement. This release significantly expands access to capable language models for users with consumer-grade or edge hardware, such as older GPUs, mobile devices, and single-board computers like the Raspberry Pi. It represents a major step in the democratization of AI, allowing more developers and hobbyists to run powerful models locally without expensive cloud infrastructure. Early community testing suggests the 9B model’s performance sits between that of 20B and 120B parameter open-source models. Users report that these 3.5 variants may share a tendency with some previous Qwen versions to “overthink,” and a prompt engineering tip suggests adjusting the template to turn off “thinking” and setting temperature to around 0.45 for more accurate responses, especially in vision tasks.
reddit · r/LocalLLaMA · Illustrious-Swim9663 · Mar 2, 12:32
Background: Qwen is a family of large language models developed by Alibaba Cloud, released under the permissive Apache 2.0 license. Model quantization is a critical optimization technique for deploying models on edge devices, reducing model size and computational requirements by representing weights and activations with lower-precision data types (e.g., from 32-bit floats to 8-bit integers). Small Language Models (SLMs), typically under 10B parameters, are designed for efficient inference on consumer hardware and have been shown in benchmarks to sometimes match or exceed the performance of much larger models when fine-tuned for specific tasks.
References
Discussion: The community reaction is overwhelmingly positive, with users celebrating the release as a “Christmas” for those with limited GPU resources. Discussion is highly practical, focusing on immediate quantization efforts, performance comparisons (e.g., the 9B model being placed between 20B and 120B models), and prompt engineering tips to optimize output. There is also curiosity about the capabilities of the smallest models (0.8B, 2B) and their potential to run on devices like the Raspberry Pi.
Tags: #llm, #model-release, #edge-ai, #quantization, #open-source
Qwen 3.5 0.8B multimodal model runs locally in browser via WebGPU and Transformers.js ⭐️ 7.0/10
A developer has created a demo that runs the Qwen 3.5 0.8B multimodal model entirely locally within a web browser using the WebGPU API and the Transformers.js library. This follows the release of the Qwen 3.5 Small model family, which includes several sizes (0.8B, 2B, 4B, 9B) designed for on-device use. This demonstrates the feasibility of running sophisticated, multimodal AI models directly on user devices without relying on cloud servers, which enhances privacy, reduces latency, and enables AI features in offline or low-connectivity scenarios. It showcases the convergence of advanced browser APIs and efficient model architectures, pushing the boundary of what’s possible for on-device AI in web applications. The demo’s primary performance bottleneck is the vision encoder component of the model. The developer notes that while this is a limitation, the fact that it runs at all is a significant achievement. The demo is publicly accessible online, and the specific model variant used is part of the newly released Qwen 3.5 Small collection.
reddit · r/LocalLLaMA · xenovatech · Mar 2, 17:46
Background: WebGPU is a modern web API that provides low-level, high-performance access to a device’s Graphics Processing Unit (GPU) for general-purpose computation and graphics, enabling complex tasks like machine learning inference directly in the browser. Transformers.js is a JavaScript library from Hugging Face that ports the functionality of the popular Python transformers library to the browser, allowing pre-trained AI models to run client-side without a server. Qwen 3.5 is a series of multimodal large language models (VLMs) from Alibaba Cloud that can process and understand both text and images, with the ‘Small’ variants specifically optimized for efficient deployment on consumer hardware.
References
Discussion: The discussion includes technical insights and troubleshooting. One commenter identifies the vision encoder as a common WebGPU bottleneck and suggests an alternative approach using quantized GGUF models via llama.cpp’s WebAssembly port for better performance. Other comments include a user reporting an unresponsive ‘start’ button in the demo, a request for the source code, and a question clarifying which specific Qwen variant was used, alongside some humorous or off-topic remarks.
Tags: #WebGPU, #On-Device AI, #Multimodal Models, #Browser ML, #Qwen
Qwen3.5-0.8B runs locally on a 7-year-old Samsung phone at 12 tokens/second. ⭐️ 7.0/10
A user successfully ran the newly released Qwen3.5-0.8B large language model locally on a 7-year-old Samsung Galaxy S10E smartphone, achieving an inference speed of 12 tokens per second using llama.cpp and Termux. This demonstrates that a capable, conversational AI model can now function on aging, low-end mobile hardware. This achievement highlights the rapid progress in model efficiency and compression, making advanced AI accessible on resource-constrained devices without relying on cloud services. It paves the way for more private, low-latency, and cost-effective AI applications on ubiquitous mobile hardware, potentially democratizing AI access. The phone is powered by a Qualcomm Snapdragon 855 chipset, and the performance is attributed to efficient quantization (likely Q4_0 or Q8) and the optimized NEON SIMD path within llama.cpp for ARM processors. While the model is functional, a community member noted it primarily understands English and may have limited utility in other languages.
reddit · r/LocalLLaMA · HighFlyingB1rd · Mar 2, 21:21
Background: Qwen3.5 is a series of open-weight large language models from Alibaba’s Qwen team, with variants ranging from small 0.6B parameter models to massive dense and mixture-of-experts (MoE) models. Llama.cpp is an open-source, C/C++-based inference engine designed to run LLMs efficiently on various hardware, including CPUs, by leveraging quantization to reduce model size and memory requirements. Termux is a terminal emulator and Linux environment application for Android that allows users to run command-line tools and software, making it a popular platform for on-device AI experimentation.
References
Discussion: The community expressed amazement at the progress, noting that a 0.8B model holding a coherent conversation was unexpected a year ago. Technical discussions focused on the quantization method used (Q4_0 vs. Q8), the role of NEON SIMD optimizations in achieving performance on older ARM chips, and installation details for llama.cpp. Some users shared their own test results on different phones, while others questioned the practical utility of smaller models for non-English tasks.
Tags: #edge-ai, #llama.cpp, #model-efficiency, #mobile-computing, #qwen
Qwen3.5’s 9B and 4B models achieve benchmark performance surpassing older, much larger models. ⭐️ 7.0/10
Alibaba’s Qwen team has released new 9-billion and 4-billion parameter models under the Qwen3.5 series, with benchmark results showing the 9B model outperforming older Qwen 30B and 80B models on certain tasks like general knowledge and reasoning. This represents a significant leap in performance-per-parameter efficiency for small language models (SLMs). This demonstrates rapid progress in model efficiency, where smaller, newer models can outperform their much larger predecessors, making powerful AI more accessible for edge deployment and reducing computational costs. It challenges the traditional notion that model capability scales linearly with parameter count and highlights the importance of architectural innovations and training data quality. While impressive in general knowledge and multimodal tasks, the new models reportedly score lower in reasoning and coding benchmarks compared to some open-source GPT models, indicating they may be specialized or have trade-offs. The community has also noted a lack of direct comparison to the previous high-performing Qwen3 4B 2507 model and criticized the presented charts for poor readability.
reddit · r/LocalLLaMA · Nunki08 · Mar 2, 12:44
Background: Qwen is a series of large language models developed by Alibaba Cloud. The ‘3.5’ designation indicates a generation within this series. Parameters refer to the internal variables a model learns during training, and traditionally, more parameters have been associated with greater capability, but also higher computational requirements. Small Language Models (SLMs) like these 4B and 9B models are designed to be efficient enough to run on less powerful hardware, such as personal computers or edge devices. Benchmarking involves testing models on standardized tasks to measure and compare their performance across areas like reasoning, coding, and knowledge.
References
Discussion: The community expressed astonishment at the 9B model’s performance against larger predecessors, with questions focusing on the technical ‘how’—speculating about compression or vectorization techniques. There is strong interest in practical comparisons, such as performance trade-offs between running a quantized larger model versus a high-precision smaller one. Criticisms include poor chart readability and a desire for more direct comparisons with specific previous models like the Qwen3 4B 2507.
Tags: #llm-benchmarks, #model-efficiency, #qwen, #small-language-models, #performance-comparison
Qwen 3.5 2B model demonstrates exceptional OCR capabilities for diverse text types. ⭐️ 7.0/10
The Qwen 3.5 2B vision-language model has been observed to perform impressively on optical character recognition (OCR) tasks, handling text at various angles and qualities, from clear scans to low-quality photos, and supporting structured output. Users report it outperforms its predecessor, Qwen 3.5 0.8B, on specific tasks like reading passport Machine Readable Zones (MRZ), which previously caused repetitive output errors. This development is significant because it shows a small, efficient model (2B parameters) can deliver robust OCR performance locally, potentially reducing reliance on cloud services or specialized, larger models for document processing. It opens up practical applications for edge deployment, such as processing identity documents, game UIs, and handwritten text on consumer hardware. The model reportedly handles challenging cases like handwritten text and Arabic documents with tables, and it appears to have resolved a repetition bug that affected the Qwen 3.5 0.8B model on passport MRZ lines. Its performance is being actively compared to other small dedicated OCR models like GLM-OCR and DeepSeek-OCR-2, which are also around 2B parameters.
reddit · r/LocalLLaMA · deadman87 · Mar 2, 15:34
Background: Qwen is a family of large language models developed by Alibaba Cloud, with many variants released as open-weight models. The Qwen 3.5 2B is a vision-language model (VLM) that uses an early fusion architecture to process multimodal inputs. OCR (Optical Character Recognition) is the technology that converts images of text into machine-encoded text. The Machine Readable Zone (MRZ) is a standardized area at the bottom of passports and IDs containing encoded personal data, often a challenging test for OCR systems due to its specific font and format.
References
Discussion: The community discussion is focused on comparative performance testing and practical applications. Users are sharing experiences comparing Qwen 3.5 2B to other models like GLM-OCR, DeepSeek-OCR-2, and earlier Qwen VL variants, noting its surprising accuracy on handwritten text. Specific use cases raised include processing Arabic legal documents with tables, comic book lettering, and game UIs, with users seeking recommendations for the best model for each scenario.
Tags: #OCR, #Vision-Language-Models, #Qwen, #LocalLLM, #Document-Processing
LM Studio’s parser silently breaks Qwen3.5 tool calling and reasoning, connecting year-long bug reports ⭐️ 7.0/10
A user compiled multiple critical bug reports showing that LM Studio’s server parser contains a cluster of interacting bugs that silently corrupt tool calling and reasoning output for models like Qwen3.5 and DeepSeek-R1. The core issue is that the parser incorrectly scans inside model reasoning blocks (like <think>) for tool call patterns, creating a recursive failure trap. This matters because LM Studio is a popular tool for running local LLMs, and these silent bugs make advanced models appear less capable than they actually are, misleading users and hindering the adoption of local, agentic AI workflows. It highlights the importance of robust parsing logic in tools that interface with reasoning models that use structured output formats. The parser fails to distinguish between a model’s prose discussion of tool call syntax within a <think> block and an actual tool call attempt, leading to parse errors that are fed back to the model and cause infinite recursion. This issue is not limited to tool calling but also corrupts general reasoning output, and it affects multiple reasoning-capable models beyond just Qwen3.5.
reddit · r/LocalLLaMA · One-Cheesecake389 · Mar 2, 15:52
Background: LM Studio is a graphical user interface and local server for running open-source large language models (LLMs) on personal computers. Models like Qwen3.5 and DeepSeek-R1 are “reasoning” models that often structure their internal thought process within special tags like <think> before producing a final answer or action, such as a tool/function call. Tool calling is a feature where an LLM can request the execution of external functions (like API calls) using a specific syntax, which the hosting application must parse correctly to execute.
Discussion: The community sentiment is largely supportive of the bug report, with users validating the issue’s significance and sharing their own frustrating experiences. Key viewpoints include appreciation for connecting isolated reports, criticism of LM Studio’s development practices (“vibe coding” without sufficient testing), and practical advice to use alternatives like llama.cpp’s server as a stopgap. Several users confirmed experiencing the issue specifically with Qwen3.5 models and are seeking workarounds.
Tags: #local-llm, #bug-report, #tool-calling, #lm-studio, #qwen
Xiaomi’s Humanoid Robot Deployed in Auto Factory for Die-Cast Part Assembly ⭐️ 7.0/10
Xiaomi has deployed its humanoid robot in an automotive die-casting workshop, where it autonomously performed the task of installing self-tapping nuts on die-cast parts. The robot operated continuously for three hours, achieving a 90.2% success rate for bilateral installation and meeting a production cycle time requirement of 76 seconds. This deployment represents a significant step towards the practical application of humanoid robots in complex, real-world manufacturing environments, moving beyond controlled lab demonstrations. It demonstrates the potential for advanced AI-driven robots to address labor-intensive and precise tasks in industries like automotive manufacturing, which could reshape future factory automation strategies. The task was powered by the Xiaomi-Robotics-0 model, utilizing end-to-end data-driven control and reinforcement learning. It integrated multimodal sensory information including vision, touch, and joint perception to handle precise assembly under complex working conditions.
telegram · zaihuapd · Mar 2, 08:30
Background: Die-casting is a manufacturing process used to produce metal parts with high precision and complexity, commonly for automotive components. End-to-end data-driven control in robotics refers to systems where sensor inputs are directly mapped to control outputs using learned models, often trained on large datasets. Multimodal AI in this context combines different sensory inputs like vision and touch to give robots a more comprehensive understanding of their environment for manipulation tasks.
References
Tags: #robotics, #industrial-automation, #artificial-intelligence, #manufacturing, #reinforcement-learning