OrangeBot.AI Digest — 2026-04-20
85 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- John Ternus to become Apple CEO (www.apple.com)
- AI Resistance: some recent anti-AI stuff that’s worth discussing (stephvee.ca)
- At long last, InfoWars is ours (theonion.com)
- We accepted surveillance as default (vivianvoss.net)
- Not buying another Kindle (www.androidauthority.com)
- Deezer says 44% of songs uploaded to its platform daily are AI-generated (techcrunch.com)
- Kimi K2.6: Advancing open-source coding (www.kimi.com)
- Qwen3.6-Max-Preview: Smarter, Sharper, Still Evolving (qwen.ai)
- All phones sold in the EU to have replaceable batteries from 2027 (www.theolivepress.es)
- Sauna effect on heart rate (tryterra.co)
- Atlassian enables default data collection to train AI (letsdatascience.com)
- ggsql: A Grammar of Graphics for SQL (opensource.posit.co)
- Tesla concealed fatal accidents to continue testing autonomous driving (www.rts.ch)
- M 7.4 earthquake – 100 km ENE of Miyako, Japan (earthquake.usgs.gov)
- NSA is using Anthropic's Mythos despite blacklist (www.axios.com)
GitHub Trending(10)
Product Hunt(15)
- Co-Tasker
Book local pros for quick & affordable help
- CONA
E-commerce accounting that runs itself
- Silex
Swiss legal AI, built by lawyers for lawyers
- Tetractys
AI for biomanufacturers
- Auxilius.ai
Turn compliance into code with agentic AI
- Getpin
Pin business, get interest, be found
- Telagri
Smarter agri-lending with real-time field visibility
- Sirputis
Where engineering meets seaweed
- MIRA vision
AI-powered pathology analysis with synthetic data
- OptionV
MacOS menubar clipboard manager
- Vertai Technology
(site is a launching-soon placeholder)
- Mav9
Time to allocate capital
- MetaNeural
Enable high-risk industries w/ AI-led XR training solutions
- Claude Desktop Buddy
Bring Claude into the physical world with maker hardware
- Dune
Context-aware Mac keypad to automate workflows + meetings
Hugging Face(15)
- Elucidating the SNR-t Bias of Diffusion Probabilistic Models
Diffusion Probabilistic Models have demonstrated remarkable performance across a wide range of generative tasks. However, we have observed that these models often suffer from a Signal-to-Noise Ratio-timestep (SNR-t) bias. This bias refers to the misalignment between the SNR of the denoising sample and its corresponding timestep during the inference phase. Specifically, during training, the SNR of a sample is strictly coupled with its timestep. However, this correspondence is disrupted during inference, leading to error accumulation and impairing the generation quality. We provide comprehensive empirical evidence and theoretical analysis to substantiate this phenomenon and propose a simple yet effective differential correction method to mitigate the SNR-t bias. Recognizing that diffusion models typically reconstruct low-frequency components before focusing on high-frequency details during the reverse denoising process, we decompose samples into various frequency components and apply differential correction to each component individually. Extensive experiments show that our approach significantly improves the generation quality of various diffusion models (IDDPM, ADM, DDIM, A-DPM, EA-DPM, EDM, PFGM++, and FLUX) on datasets of various resolutions with negligible computational overhead. The code is at https://github.com/AMAP-ML/DCW.
- Maximal Brain Damage Without Data or Optimization: Disrupting Neural Networks via Sign-Bit Flips
Deep Neural Networks (DNNs) can be catastrophically disrupted by flipping only a handful of parameter bits. We introduce Deep Neural Lesion (DNL), a data-free and optimizationfree method that locates critical parameters, and an enhanced single-pass variant, 1P-DNL, that refines this selection with one forward and backward pass on random inputs. We show that this vulnerability spans multiple domains, including image classification, object detection, instance segmentation, and reasoning large language models. In image classification, flipping just two sign bits in ResNet-50 on ImageNet reduces accuracy by 99.8%. In object detection and instance segmentation, one or two sign flips in the backbone collapse COCO detection and mask AP for Mask R-CNN and YOLOv8-seg models. In language modeling, two sign flips into different experts reduce Qwen3-30B-A3B-Thinking from 78% to 0% accuracy. We also show that selectively protecting a small fraction of vulnerable sign bits provides a practical defense against such attacks.
- PersonaVLM: Long-Term Personalized Multimodal LLMs
Multimodal Large Language Models (MLLMs) serve as daily assistants for millions. However, their ability to generate responses aligned with individual preferences remains limited. Prior approaches enable only static, single-turn personalization through input augmentation or output alignment, and thus fail to capture users' evolving preferences and personality over time (see Fig.1). In this paper, we introduce PersonaVLM, an innovative personalized multimodal agent framework designed for long-term personalization. It transforms a general-purpose MLLM into a personalized assistant by integrating three key capabilities: (a) Remembering: It proactively extracts and summarizes chronological multimodal memories from interactions, consolidating them into a personalized database. (b) Reasoning: It conducts multi-turn reasoning by retrieving and integrating relevant memories from the database. (c) Response Alignment: It infers the user's evolving personality throughout long-term interactions to ensure outputs remain aligned with their unique characteristics. For evaluation, we establish Persona-MME, a comprehensive benchmark comprising over 2,000 curated interaction cases, designed to assess long-term MLLM personalization across seven key aspects and 14 fine-grained tasks. Extensive experiments validate our method's effectiveness, improving the baseline by 22.4% (Persona-MME) and 9.8% (PERSONAMEM) under a 128k context, while outperforming GPT-4o by 5.2% and 2.0%, respectively. Project page: https://PersonaVLM.github.io.
- Web Retrieval-Aware Chunking (W-RAC) for Efficient and Cost-Effective Retrieval-Augmented Generation Systems
Retrieval-Augmented Generation (RAG) systems critically depend on effective document chunking strategies to balance retrieval quality, latency, and operational cost. Traditional chunking approaches, such as fixed-size, rule-based, or fully agentic chunking, often suffer from high token consumption, redundant text generation, limited scalability, and poor debuggability, especially for large-scale web content ingestion. In this paper, we propose Web Retrieval-Aware Chunking (W-RAC), a novel, cost-efficient chunking framework designed specifically for web-based documents. W-RAC decouples text extraction from semantic chunk planning by representing parsed web content as structured, ID-addressable units and leveraging large language models (LLMs) only for retrieval-aware grouping decisions rather than text generation. This significantly reduces token usage, eliminates hallucination risks, and improves system observability.Experimental analysis and architectural comparison demonstrate that W-RAC achieves comparable or better retrieval performance than traditional chunking approaches while reducing chunking-related LLM costs by an order of magnitude.
- Qwen3.5-Omni Technical Report
In this work, we present Qwen3.5-Omni, the latest advancement in the Qwen-Omni model family. Representing a significant evolution over its predecessor, Qwen3.5-Omni scales to hundreds of billions of parameters and supports a 256k context length. By leveraging a massive dataset comprising heterogeneous text-vision pairs and over 100 million hours of audio-visual content, the model demonstrates robust omni-modality capabilities. Qwen3.5-Omni-plus achieves SOTA results across 215 audio and audio-visual understanding, reasoning, and interaction subtasks and benchmarks, surpassing Gemini-3.1 Pro in key audio tasks and matching it in comprehensive audio-visual understanding. Architecturally, Qwen3.5-Omni employs a Hybrid Attention Mixture-of-Experts (MoE) framework for both Thinker and Talker, enabling efficient long-sequence inference. The model facilitates sophisticated interaction, supporting over 10 hours of audio understanding and 400 seconds of 720P video (at 1 FPS). To address the inherent instability and unnaturalness in streaming speech synthesis, often caused by encoding efficiency discrepancies between text and speech tokenizers, we introduce ARIA. ARIA dynamically aligns text and speech units, significantly enhancing the stability and prosody of conversational speech with minimal latency impact. Furthermore, Qwen3.5-Omni expands linguistic boundaries, supporting multilingual understanding and speech generation across 10 languages with human-like emotional nuance. Finally, Qwen3.5-Omni exhibits superior audio-visual grounding capabilities, generating script-level structured captions with precise temporal synchronization and automated scene segmentation. Remarkably, we observed the emergence of a new capability in omnimodal models: directly performing coding based on audio-visual instructions, which we call Audio-Visual Vibe Coding.
- Cut Your Losses! Learning to Prune Paths Early for Efficient Parallel Reasoning
Parallel reasoning enhances Large Reasoning Models (LRMs) but incurs prohibitive costs due to futile paths caused by early errors. To mitigate this, path pruning at the prefix level is essential, yet existing research remains fragmented without a standardized framework. In this work, we propose the first systematic taxonomy of path pruning, categorizing methods by their signal source (internal vs. external) and learnability (learnable vs. non-learnable). This classification reveals the unexplored potential of learnable internal methods, motivating our proposal of STOP (Super TOken for Pruning). Extensive evaluations across LRMs ranging from 1.5B to 20B parameters demonstrate that STOP achieves superior effectiveness and efficiency compared to existing baselines. Furthermore, we rigorously validate the scalability of STOP under varying compute budgets - for instance, boosting GPT-OSS-20B accuracy on AIME25 from 84% to nearly 90% under fixed compute budgets. Finally, we distill our findings into formalized empirical guidelines to facilitate optimal real-world deployment. Code, data and models are available at https://bijiaxihh.github.io/STOP
- (1D) Ordered Tokens Enable Efficient Test-Time Search
Tokenization is a key component of autoregressive (AR) generative models, converting raw data into more manageable units for modeling. Commonly, tokens describe local information, such as regions of pixels in images or word pieces in text, and AR generation predicts these tokens in a fixed order. A worthwhile question is whether token structures affect the ability to steer the generation through test-time search, where multiple candidate generations are explored and evaluated by a verifier. Using image generation as our testbed, we hypothesize that recent 1D ordered tokenizers with coarse-to-fine structure can be more amenable to search than classical 2D grid structures. This is rooted in the fact that the intermediate states in coarse-to-fine sequences carry semantic meaning that verifiers can reliably evaluate, enabling effective steering during generation. Through controlled experiments, we find that AR models trained on coarse-to-fine ordered tokens exhibit improved test-time scaling behavior compared to grid-based counterparts. Moreover, we demonstrate that, thanks to the ordered structure, pure test-time search over token sequences (i.e., without training an AR model) can perform training-free text-to-image generation when guided by an image-text verifier. Beyond this, we systematically study how classical search algorithms (best-of-N, beam search, lookahead search) interact with different token structures, as well as the role of different verifiers and AR priors. Our results highlight the impact of token structure on inference-time scalability and provide practical guidance for test-time scaling in AR models.
- Repurposing 3D Generative Model for Autoregressive Layout Generation
We introduce LaviGen, a framework that repurposes 3D generative models for 3D layout generation. Unlike previous methods that infer object layouts from textual descriptions, LaviGen operates directly in the native 3D space, formulating layout generation as an autoregressive process that explicitly models geometric relations and physical constraints among objects, producing coherent and physically plausible 3D scenes. To further enhance this process, we propose an adapted 3D diffusion model that integrates scene, object, and instruction information and employs a dual-guidance self-rollout distillation mechanism to improve efficiency and spatial accuracy. Extensive experiments on the LayoutVLM benchmark show LaviGen achieves superior 3D layout generation performance, with 19% higher physical plausibility than the state of the art and 65% faster computation. Our code is publicly available at https://github.com/fenghora/LaviGen.
- Where does output diversity collapse in post-training?
Post-trained language models produce less varied outputs than their base counterparts. This output diversity collapse undermines inference-time scaling methods that rely on varied samples, and risks homogenizing model outputs on creative and value-laden tasks. Prior work attributes collapse to specific post-training methods, without separating the role of training data composition from the method, or the generation format from the model weights. We trace output diversity through three parallel post-training lineages of Olmo 3, Think (chain-of-thought distillation), Instruct (broad multi-source data), and RL-Zero, across 15 tasks and four text diversity metrics. We find that the location of collapse co-varies with data composition: the Think lineage loses most semantic diversity at supervised fine-tuning, and the effect of DPO is larger in Instruct than in Think. Suppressing chain-of-thought reasoning at inference in Think models drops accuracy on hard tasks, yet leaves answer-level diversity unchanged, showing that the collapse is embedded in the model weights by training data, not imposed by the generation format. Decomposing diversity loss on six verifiable tasks into a quality-control component (removal of incorrect outputs) and a residual component (genuine narrowing among correct outputs) reveals that the split is task-dependent, and Think models retain more correct-answer diversity than Instruct despite collapsing more in aggregate. Our results indicate that diversity collapse is determined during training by data composition and cannot be addressed at inference time alone.
- QuantCode-Bench: A Benchmark for Evaluating the Ability of Large Language Models to Generate Executable Algorithmic Trading Strategies
Large language models have demonstrated strong performance on general-purpose programming tasks, yet their ability to generate executable algorithmic trading strategies remains underexplored. Unlike standard code benchmarks, trading-strategy generation requires simultaneous mastery of domain-specific financial logic, knowledge of a specialized API, and the ability to produce code that is not only syntactically correct but also leads to actual trades on historical data. In this work, we present QuantCode-Bench, a benchmark for the systematic evaluation of modern LLMs in generating strategies for the Backtrader framework from textual descriptions in English. The benchmark contains 400 tasks of varying difficulty collected from Reddit, TradingView, StackExchange, GitHub, and synthetic sources. Evaluation is conducted through a multi-stage pipeline that checks syntactic correctness, successful backtest execution, the presence of trades, and semantic alignment with the task description using an LLM judge. We compare state-of-the-art models in two settings: single-turn, where the strategy must be generated correctly on the first attempt, and agentic multi-turn, where the model receives iterative feedback and may repair its errors. We analyze the failure modes across different stages of the pipeline and show that the main limitations of current models are not related to syntax, but rather to the correct operationalization of trading logic, proper API usage, and adherence to task semantics. These findings suggest that trading strategy generation constitutes a distinct class of domain-specific code generation tasks in which success requires not only technical correctness, but also alignment between natural-language descriptions, financial logic, and the observable behavior of the strategy on data.
- Can Large Language Models Reinvent Foundational Algorithms?
LLMs have shown strong potential to advance scientific discovery. Whether they possess the capacity for foundational innovation, however, remains an open question. In this work, we focus on a prerequisite for foundational innovation: can LLMs reinvent foundational algorithms in computer science? Our Unlearn-and-Reinvent pipeline applies LLM unlearning to remove a specific foundational algorithm, such as Dijkstra's or Euclid's algorithm, from an LLM's pretrained knowledge, and then tests whether the model can reinvent it in a controlled environment. To enable effective unlearning, we adopt a GRPO-based, on-policy unlearning method. Across 10 target algorithms, 3 strong open-weight models, and 3 hint levels, our experiments demonstrate that (1) the strongest model Qwen3-4B-Thinking-2507 successfully reinvents 50% of the algorithms with no hint, 70% at hint level 1, and 90% at hint level 2; (2) a few high-level hints can enhance the reinvention success rate, but even step-by-step hints fail for those complicated algorithms; and (3) test-time reinforcement learning enables successful reinvention for the Strassen algorithm at hint level 2. Through analyses of output trajectories and ablation studies, we find that generative verifier in the reinvention phase plays a critical role in sustaining models' reasoning strength, helping to avoid the ``thought collapse'' phenomenon. These findings offer insights into both the potential and current limits of LLMs' innovative thinking.
- Learning Adaptive Reasoning Paths for Efficient Visual Reasoning
Visual reasoning models (VRMs) have recently shown strong cross-modal reasoning capabilities by integrating visual perception with language reasoning. However, they often suffer from overthinking, producing unnecessarily long reasoning chains for any tasks. We attribute this issue to Reasoning Path Redundancy in visual reasoning: many visual questions do not require the full reasoning process. To address this, we propose AVR, an adaptive visual reasoning framework that decomposes visual reasoning into three cognitive functions: visual perception, logical reasoning, and answer application. It further enables models to dynamically choose among three response formats: Full Format, Perception-Only Format, and Direct Answer. AVR is trained with FS-GRPO, an adaptation of Group Relative Policy Optimization that encourages the model to select the most efficient reasoning format while preserving correctness. Experiments on multiple vision-language benchmarks show that AVR reduces token usage by 50--90\% while maintaining overall accuracy, especially in perception-intensive tasks. These results demonstrate that adaptive visual reasoning can effectively mitigate overthinking in VRMs. Code and data are available at: https://github.com/RunRiotComeOn/AVR.
- TIPSv2: Advancing Vision-Language Pretraining with Enhanced Patch-Text Alignment
Recent progress in vision-language pretraining has enabled significant improvements to many downstream computer vision applications, such as classification, retrieval, segmentation and depth prediction. However, a fundamental capability that these models still struggle with is aligning dense patch representations with text embeddings of corresponding concepts. In this work, we investigate this critical issue and propose novel techniques to enhance this capability in foundational vision-language models. First, we reveal that a patch-level distillation procedure significantly boosts dense patch-text alignment -- surprisingly, the patch-text alignment of the distilled student model strongly surpasses that of the teacher model. This observation inspires us to consider modifications to pretraining recipes, leading us to propose iBOT++, an upgrade to the commonly-used iBOT masked image objective, where unmasked tokens also contribute directly to the loss. This dramatically enhances patch-text alignment of pretrained models. Additionally, to improve vision-language pretraining efficiency and effectiveness, we modify the exponential moving average setup in the learning recipe, and introduce a caption sampling strategy to benefit from synthetic captions at different granularities. Combining these components, we develop TIPSv2, a new family of image-text encoder models suitable for a wide range of downstream applications. Through comprehensive experiments on 9 tasks and 20 datasets, we demonstrate strong performance, generally on par with or better than recent vision encoder models. Code and models are released via our project page at https://gdm-tipsv2.github.io/ .
- GTA-2: Benchmarking General Tool Agents from Atomic Tool-Use to Open-Ended Workflows
The development of general-purpose agents requires a shift from executing simple instructions to completing complex, real-world productivity workflows. However, current tool-use benchmarks remain misaligned with real-world requirements, relying on AI-generated queries, dummy tools, and limited system-level coordination. To address this, we propose GTA-2, a hierarchical benchmark for General Tool Agents (GTA) spanning atomic tool use and open-ended workflows. Built on real-world authenticity, it leverages real user queries, deployed tools, and multimodal contexts. (i) GTA-Atomic, inherited from our prior GTA benchmark, evaluates short-horizon, closed-ended tool-use precision. (ii) GTA-Workflow introduces long-horizon, open-ended tasks for realistic end-to-end completion. To evaluate open-ended deliverables, we propose a recursive checkpoint-based evaluation mechanism that decomposes objectives into verifiable sub-goals, enabling unified evaluation of both model capabilities and agent execution frameworks (i.e., execution harnesses). Experiments reveal a pronounced capability cliff: while frontier models already struggle on atomic tasks (below 50%), they largely fail on workflows, with top models achieving only 14.39% success. Further analysis shows that checkpoint-guided feedback improves performance, while advanced frameworks such as Manus and OpenClaw substantially enhance workflow completion, highlighting the importance of execution harness design beyond the underlying model capacity. These findings provide guidance for developing reliable personal and professional assistants. Dataset and code will be available at https://github.com/open-compass/GTA.
- Hierarchical Codec Diffusion for Video-to-Speech Generation
Video-to-Speech (VTS) generation aims to synthesize speech from a silent video without auditory signals. However, existing VTS methods disregard the hierarchical nature of speech, which spans coarse speaker-aware semantics to fine-grained prosodic details. This oversight hinders direct alignment between visual and speech features at specific hierarchical levels during property matching. In this paper, leveraging the hierarchical structure of Residual Vector Quantization (RVQ)-based codec, we propose HiCoDiT, a novel Hierarchical Codec Diffusion Transformer that exploits the inherent hierarchy of discrete speech tokens to achieve strong audio-visual alignment. Specifically, since lower-level tokens encode coarse speaker-aware semantics and higher-level tokens capture fine-grained prosody, HiCoDiT employs low-level and high-level blocks to generate tokens at different levels. The low-level blocks condition on lip-synchronized motion and facial identity to capture speaker-aware content, while the high-level blocks use facial expression to modulate prosodic dynamics. Finally, to enable more effective coarse-to-fine conditioning, we propose a dual-scale adaptive instance layer normalization that jointly captures global vocal style through channel-wise normalization and local prosody dynamics through temporal-wise normalization. Extensive experiments demonstrate that HiCoDiT outperforms baselines in fidelity and expressiveness, highlighting the potential of discrete modelling for VTS. The code and speech demo are both available at https://github.com/Jiaxin-Ye/HiCoDiT.
Techmeme(15)
- John Ternus, senior VP of Hardware Engineering, will become Apple's next CEO on September 1; Tim Cook will become executive chairman of Apple's board (CNBC)
CNBC : John Ternus, senior VP of Hardware Engineering, will become Apple's next CEO on September 1; Tim Cook will become executive chairman of Apple's board — Apple said on Monday that John Ternus is succeeding Tim Cook as CEO, with Cook assuming the role of executive chairman on Sept. 1.
- OpenAI rolls out Chronicle, which builds memories from screen captures to make Codex more aware of context, as a research preview for Pro subscribers on macOS (Zac Hall/9to5Mac)
Zac Hall / 9to5Mac : OpenAI rolls out Chronicle, which builds memories from screen captures to make Codex more aware of context, as a research preview for Pro subscribers on macOS — Last week, OpenAI released an all-new version of Codex for Mac that includes the best example of AI-driven computer use yet.
- Apple says Johny Srouji, who most recently served as senior VP of Hardware Technologies, will assume an expanded role leading Hardware Engineering (Apple)
Apple : Apple says Johny Srouji, who most recently served as senior VP of Hardware Technologies, will assume an expanded role leading Hardware Engineering — Apple today announced that, effective immediately, Apple executive Johny Srouji will become chief hardware officer.
- Amazon agrees to invest up to $25B in Anthropic, on top of the $8B that it has already invested; Anthropic commits to spend $100B+ on AWS over the next 10 years (Ashley Capoot/CNBC)
Ashley Capoot / CNBC : Amazon agrees to invest up to $25B in Anthropic, on top of the $8B that it has already invested; Anthropic commits to spend $100B+ on AWS over the next 10 years — Amazon has agreed to invest up to $25 billion in Anthropic, on top of the $8 billion that it's poured into the artificial intelligence startup …
- John Ternus, senior VP of Hardware Engineering, will become Apple's next CEO on September 1; Tim Cook will become executive chairman of Apple's board (Apple)
Apple : John Ternus, senior VP of Hardware Engineering, will become Apple's next CEO on September 1; Tim Cook will become executive chairman of Apple's board — Apple® announced that Tim Cook will become executive chairman of Apple's board of directors and John Ternus, senior vice president …
- Microsoft pauses new GitHub Copilot signups for Pro, Pro+, and Student tiers, tightens usage limits, removes Opus models from Pro, and limits Opus 4.7 to Pro+ (The GitHub Blog)
The GitHub Blog : Microsoft pauses new GitHub Copilot signups for Pro, Pro+, and Student tiers, tightens usage limits, removes Opus models from Pro, and limits Opus 4.7 to Pro+ — As shared in our recent blog post, we're making the following changes to Copilot plans for individuals as part of our ongoing efforts …
- Docs: Microsoft plans to eventually move GitHub Copilot from request- to token-based billing, as the week-over-week cost to run it has nearly doubled since Jan. (Edward Zitron/Ed Zitron's Where's ...)
Edward Zitron / Ed Zitron's Where's Your Ed At : Docs: Microsoft plans to eventually move GitHub Copilot from request- to token-based billing, as the week-over-week cost to run it has nearly doubled since Jan. — Executive Summary: — Internal documents reveal that Microsoft plans to temporarily suspend individual account signups …
- RaveDAO's RAVE has lost $6.6B+ in market cap and its price has sunk ~98% since Saturday, after ZachXBT called on exchanges to probe if it was being manipulated (André Beganski/Decrypt)
André Beganski / Decrypt : RaveDAO's RAVE has lost $6.6B+ in market cap and its price has sunk ~98% since Saturday, after ZachXBT called on exchanges to probe if it was being manipulated — RaveDAO's token cratered this weekend after blockchain sleuth ZachXBT called on exchanges to investigate trading tied to its surging token.
- A profile of far-right influencer Nick Fuentes, who has been kicked off most mainstream social media but made ~$900K from "fanatical" donors since early 2025 (Washington Post)
Washington Post : A profile of far-right influencer Nick Fuentes, who has been kicked off most mainstream social media but made ~$900K from “fanatical” donors since early 2025 — Summary — Kristine Kasubienski's donation appeared on viewers' screens four hours into the live stream of Nick Fuentes …
- Filing in California's antitrust lawsuit: CA accuses Amazon of price-fixing by pressuring brands to ask competing retailers to increase prices on some products (David McCabe/New York Times)
David McCabe / New York Times : Filing in California's antitrust lawsuit: CA accuses Amazon of price-fixing by pressuring brands to ask competing retailers to increase prices on some products — The state claimed the e-commerce giant pressured brands like Levi's and Hanes to ask competing retailers to raise prices on certain products.
- A new media company called MTS launches for "monitoring the situation" across tech, business, politics, and culture, with backing from a16z and others (a16z)
a16z : A new media company called MTS launches for “monitoring the situation” across tech, business, politics, and culture, with backing from a16z and others — Thrilled to announce our investment in a new media company called MTS (short for Monitoring the Situation).
- WhatsApp says it is testing a new subscription called WhatsApp Plus, which includes features like expanded pinned chats, custom lists, and new chat themes (Ivan Mehta/TechCrunch)
Ivan Mehta / TechCrunch : WhatsApp says it is testing a new subscription called WhatsApp Plus, which includes features like expanded pinned chats, custom lists, and new chat themes — WhatsApp is testing a new subscription tier, similar to Instagram Plus and Snapchat+, that lets users pay for customized icons, themes, ringtones, and more.
- Moonshot introduces Kimi K2.6, an open-weight model that it says shows strong improvements in long-horizon coding tasks, available under a modified MIT License (Kimi AI)
Kimi AI : Moonshot introduces Kimi K2.6, an open-weight model that it says shows strong improvements in long-horizon coding tasks, available under a modified MIT License — We are open sourcing our latest model, Kimi K2.6, featuring state-of-the-art coding, long-horizon execution, and agent swarm capabilities.
- Deezer says AI-generated tracks now account for 44% of daily uploads, totaling ~75K tracks per day and 2M+ per month, but account for just 1-3% of consumption (Aisha Malik/TechCrunch)
Aisha Malik / TechCrunch : Deezer says AI-generated tracks now account for 44% of daily uploads, totaling ~75K tracks per day and 2M+ per month, but account for just 1-3% of consumption — Deezer announced on Monday that AI-generated tracks now represent 44% of all new music uploaded to its platform.
- Sources: UK-based CuspAI, which aims to use AI for discovering new materials, is in discussions to raise at least $200M, taking its valuation to over $1B (Yazhou Sun/Bloomberg)
Yazhou Sun / Bloomberg : Sources: UK-based CuspAI, which aims to use AI for discovering new materials, is in discussions to raise at least $200M, taking its valuation to over $1B — CuspAI, a British startup that aims to use artificial intelligence for discovering new materials, is in discussions to raise …
Solidot(15)
- 从 2027 年起欧盟销售的智能手机和平板必须能更换电池
根据欧盟的新规定,从 2027 年起欧洲销售的智能手机和平板电脑必须配备可更换电池。此举旨在减少电子垃圾。欧盟地区每年售出约 1.5 亿部智能手机和 2400 万台平板电脑,相当于每年产生约 500 万吨电子垃圾,只有不到四成的电子垃圾被妥善回收。强制电池可更换的规定于 2027 年 2 月 18 日生效,它还规定任何便捷式电子产品的替换电池必须在产品最后一台投放市场后至少五年内继续供应。电池必须能由消费者自行替换,如果需要专用工具则必须在出售时免费提供。欧盟的新规定还要求操作系统的更新必须持续至少五年。
- 诺奖得主对人类再生存 50 年感到悲观
美国理论物理学家 David Gross 因与学生 Frank Anthony Wilczek 发现了量子色动力学中的渐近自由而在 2004 年共同获得诺贝尔物理学奖,2026 年 4 月 18 日他因为其一生对理论物理学的开创性贡献而获得了基础物理学突破奖的特别奖,获得了 300 万美元奖金。他在接受采访时被问到理论物理学是否可能在 50 年内实现大一统时表示,人类能再生存 50 年的概率非常小。他说,人类每年爆发核战争的概率大约为 2%,过去十年大国之间没有签署任何条约,人类正陷入一场惊人的军备竞赛。最近的一系列事件都增加了核战争的风险,2% 的概率已是保守估计了。当今世界有 9 个拥有核武器的国家,其中三个是核超级大国,情况比两个核大国复杂得多,国家之间的协议和规范都在瓦解。他认为人类能再生存一百年的概率微乎其微,再生存两百年的概率更是趋向于无限小了。所以费米提出的“银河系所有文明所有智慧生命都去哪里?为什么不与人类交流?”的悖论的答案是它们已经自我毁灭了。
- GitHub 上项目的伪造星数
在最大的源代码托管平台 GitHub,一个项目的星数曾经是衡量其受欢迎程度的重要指标。因为重要,因此伪造星数或者付费刷星数也日益商业化。卡内基梅隆、北卡州立大学和 Socket 的研究人员在 ICSE 2026 上发表了一项研究,使用工具 StarScout 分析了 20TB GitHub 元数据,涵盖 2019 年到 2024 年 67 亿个事件和 3.26 亿星数,识别了 600 万被怀疑刷的虚假星数,涉及 30.1 万个账户创建的 18,617 个库。付费刷星数在 2024 年急剧恶化,到 7 月 16.66% 有 50 或以上星数的项目涉嫌刷星数。到了 2025 年 1 月,涉嫌刷星数的项目有 90.42% 被官方移除,涉嫌的账号有 57.07% 被关闭。AI 和 LLM(大模型)的项目超过区块链/加密货币,成为刷星数最多的非恶意项目类别。调查发现有几十家网站、以及 Fiverr 卖家和 Telegram 频道提供付费刷星数的服务,价格最低 0.03 美元/星,最高 0.8-0.9 美元/星。清华大学的一项研究发现 QQ 和微信推广群也提供了刷星数的付费服务。
- WireGuard For Windows v1.0 释出
WireGuard 作者 Jason Donenfeld 在邮件列表上宣布 WireGuard For Windows 以及 Windows 下内核模式实现 WireGuardNT 释出 v1.0。WireGuard 是开源 VPN 协议和自由开源软件,旨在获得比 IPsec 和 OpenVPN 更好的性能。项目在 2015 年发布了最早的版本,2020 年其 Linux 版本达到稳定生产阶段,正式合并到内核主线。Windows 版本从测试阶段到成熟又经历了五年时间。
- Brave 推出付费版 Brave Origin,Linux 版免费
Brave 推出了付费版浏览器 Brave Origin,该版本移除了原版内置的变现功能如 Rewards。Origin 可单独下载,或作为现有版本的升级,一次购买即可解锁,可以在多个设备上激活。Origin 的 Linux 版本是免费的,这可能会让 Windows 付费版用户困惑:为什么他们要为别人免费获得的东西付费?
- Sruthi Chandran 当选为 DPL
2026 年 Debian 项目领导人(DPL)选举结束,唯一的候选人、来自印度的图书管理员 Sruthi Chandran 当选,她将于 4 月 21 日上任。Sruthi Chandran 从图书管理员转为自由软件爱好者和 Debian 开发者,自 2016 年以来一直参与 Debian 的 Ruby、JavaScript、Go 和字体软件包的开发,但近期开发不再活跃,她还是 Community Team 代表,Outreach 团队成员,DebConf Committee 成员。她希望能帮助提升 Debian 社区的多元性,推动多元性议题的讨论。
- 亚马逊 Fire 电视棒不再支持侧载
亚马逊正式声明,新款 Fire 电视棒不再支持从其官方应用商店之外的任何来源侧载 Android 应用。从去年发布的 Fire TV Stick 4K Select 开始,未来的 Fire 电视棒都将运行基于 Linux 的操作系统 Vega,不支持侧载非亚马逊官方应用商店之外的应用。上周亚马逊发布 Fire TV Stick HD 时在其产品页面加入了声明:为增强安全性,本设备禁止侧载或安装来自未知来源的应用。仅可下载来自亚马逊应用商店的应用。亚马逊自有品牌设备如今都运行 Vega。旧型号 Fire 电视棒仍然运行基于 Android 的操作系统,不会升级到 Vega,也不会限制侧载,此前 Fire 电视棒一直被批评助长了盗版,因为用户可以侧载能播放盗版内容的应用。
- 偏头痛发作增加与空气污染相关
研究团队对 7032 名生活在以色列南部内盖夫沙漠地区的偏头痛患者进行了平均 10 年的随访,分析其每日暴露于交通、工业及沙尘暴带来的空气污染情况以及气象条件,并与患者因急性偏头痛就医时间进行对比,同时追溯发作前 7 天内的环境变化。研究还通过药房记录统计了曲坦类药物(专门用于缓解偏头痛发作的药物)的使用情况。结果显示,约 32% 的参与者至少因急性偏头痛就医一次,47% 的人曾购买曲坦类药物。在偏头痛就诊人数最多的日子,PM10、PM2.5 及二氧化氮(NO2)浓度均显著高于平均水平;而就诊人数最少时,污染水平也相对较低。短期暴露于高水平 NO2 的人群,因偏头痛就医的概率增加 41%;暴露于较强紫外线辐射的人群,就医概率增加 23%。从长期来看,持续暴露于高 NO2 和 PM2.5 水平的人群,偏头痛药物使用率分别增加 10% 和 9%。进一步分析表明,高温、低湿环境会增强 NO2 对偏头痛的影响,而寒冷潮湿条件则会加剧 PM2.5 的作用。
- Blue Origin 第二次回收 New Glenn 火箭但上面级未能成功入轨
贝佐斯(Jeff Bezos)旗下的 Blue Origin 公司于上周日执行了 New Glenn 重型火箭的第三次飞行,火箭第一级顺利回收,但上面级未能成功将 AST SpaceMobile 的蜂窝宽带通信卫星送入正确的轨道,这次发射被认为失败了。New Glenn 是直径 7 米的两级构型火箭。其第一级火箭由 7个 BE-4 发动机提供动力,它被设计为可重复使用。第二级/上面级则为一次性使用。New Glenn 于 2025 年 1 月 16 日执行了首次发射。去年 11 月的第二次飞行成功回收了第一级。此次发射使用的第一级就是去年底回收的,它被命名为 Never Tell Me The Odds。AST SpaceMobile 的 6 吨重通信卫星原计划发射到高度 460 公里倾角为 49 度的轨道,但被送入了一个低得多的轨道,无法靠自带的推进器维持运行。
- 欧盟推动远程办公以缓解能源危机
欧盟委员会将鼓励远程办公,并为公共交通提供补贴,以减少化石燃料使用;目前,各国正艰难应对中东战事引发的能源价格冲击。欧委会将于本周向成员国提出一系列措施,以降低能源需求、提高能效,并推动向清洁电力转型。这些举措旨在为高企的能源价格提供“即时缓解”。这些建议基于俄罗斯入侵乌克兰后引发的上一轮能源危机期间实施的各项措施。它们旨在减少对化石燃料的依赖,并鼓励使用绿色能源。
- 蚂蚁也有清洁工
鱼类中有“清洁鱼”——即帮助清洁大鱼身上寄生虫的小鱼,蚂蚁中也有给大蚂蚁清洁身体的小蚂蚁。根据发表在《Ecology and Evolution》期刊上的一项研究,昆虫学家观察到体型巨大的收获蚁涌到体型很小的锥蚁(cone ants)巢穴入口,双腿伸开,大颚张开,静静等待小蚂蚁爬上它们身上,舔舐啃啃。一只收获蚁身上有时会有五只小锥蚁。收获蚁会彼此清理身上的垃圾和寄生虫,但可能这种清洁并不彻底,因此需要小蚂蚁进行更深入的清理工作。但其他科学家提出了不同的意见,有昆虫学家认为它们是在交换信息素彼此安抚,或者交换有益的微生物为两种蚂蚁创造更健康的微生物群落。
- NASA 关闭旅行者 1 号的 LECP 仪器以节省电力维持运行
NASA JPL 工程师于 4 月 17 日向旅行者 1 号发送指令,关闭低能带电粒子仪(LECP)以节省电力维持探测器的运行。旅行者 1 号与旅行者 2 号都携带了 10 个科学仪器,过去 49 年已有 7 个仪器关闭,剩下 3 个仪器——LECP,三轴磁通门磁强计(MAG)和电浆波系统(PWS)还在运行中。LECP 被用于测量低能带电粒子,包括来自太阳系和银河系的离子、电子和宇宙射线。LECP 提供了星际介质结构的关键数据,探测到日球层外的压力锋面和粒子密度变化区域。旅行者 1 号是距离地球最遥远的人造探测器,也是唯一能探测此类信息的探测器。旅行者号使用放射性同位素热电机作为动力,每年损失约 4 瓦的功率,今年 2 月 27 日工程师检测到功率的意外下降,如果功率再下降将会触发低电压故障保护系统,为保护探测器而关闭组件。而恢复组件将是一个漫长且充满风险的过程。工程团队因此决定主动关闭仪器以节省电力,他们在三个仍正常工作的仪器中选择了 LECP。工程师估计关闭 LECP 能为旅行者 1 号提供一年的喘息时间,计划在此期间完善名为“Big Bang”的节能方案,以进一步延长旅行者号的运行时间。
- 人形机器人打破人类半马世界纪录
在周日(4 月 19 日)举行的 2026 北京亦庄人形机器人半程马拉松赛上,机器人首次跑赢了人类半马世界纪录,前三名成绩全部优于人类纪录。荣耀旗下的齐天大圣队、雷霆闪电队、星火燎原队分别夺得冠军、亚军、季军,净用时分别为 50 分 26 秒、50 分 56 秒、53 分 01 秒,均优于乌干达名将基普利莫在今年 3 月里斯本半程马拉松赛中创造的 57 分 20 秒人类男子半马世界纪录。这一成绩与去年相比进步显著:一年前在亦庄举行的首届人形机器人半程马拉松上,天工队的人形机器人以 2 小时 40 分 42 秒的成绩夺冠。一年后,参赛队伍数量从 20 支增至百余支,是去年的五倍。整体完赛成绩大幅提升,人形机器人续航更稳、步态更顺、算法更稳。途中仍有部分机器人摔倒,或撞上赛道护栏。技术层面同样有进步:去年工程师们还可以跟着机器人一路小跑“陪跑",今年很多机器人的速度大幅提升,工作人员不得不改坐高尔夫球车跟在后面。去年,天工是赛场上唯一采用"半自主"方式参赛的选手,其余几乎全部依赖遥控。今年,约四成队伍已实现自主导航。为鼓励这一趋势,赛事规则规定遥控操作组成绩须乘以 1.2 加权系数,变相惩罚依赖人工遥控的队伍。
- 卫星无人机图像显示美国四成数据中心可能延期
硅谷科技巨头斥资数千亿美元建造 AI 数据中心,然而大规模的数据中心建设面临巨大的施工和电力挑战。卫星无人机图像显示美国近四成数据中心项目可能无法按计划在今年内完工。分析显示,微软、甲骨文和 OpenAI 等公司的数据中心项目很可能延期三个月以上完工。数据中心项目延误的原因包括劳动力、电力和设备长期短缺,以及获得必要许可的繁琐流程。参与 OpenAI 项目的建筑公司高管提到,他们缺乏足够的技工如电工和管道工去同时开展多个数据中心项目。因巨大的电力需求,科技巨头甚至在建自己的发电厂,大量使用安装在半挂式卡车上的移动式燃气发电机,以及最初为飞机和军舰设计的涡轮发动机。
- 内存芯片短缺可能持续到 2030 年
SK 集团会长崔泰源表示,全球内存芯片晶圆短缺问题可能会一直持续到 2030 年。SK 集团子公司 SK 海力士是最大的 HBM 芯片供应商,其市场份额高达 57%,它同时也是 DRAM 市场的第二大厂商,占据了 32% 的份额。英伟达的 AI 芯片主要使用的是 HBM 内存。崔泰源称 AI 芯片消耗了大量的 HBM,而 HBM 的生产会消耗大量的晶圆。增加晶圆的生产需要至少四到五年时间,目前的短缺可能会一直持续到 2030 年,该公司预计晶圆缺口将超过 20%。他还表示 SK 海力士将努力制定稳定 DRAM 价格的策略。