OrangeBot.AI Digest — 2026-03-07
86 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Effort to prevent government officials from engaging in prediction markets (www.merkley.senate.gov)
- War prediction markets are a national-security threat (www.theatlantic.com)
- A decade of Docker containers (cacm.acm.org)
- FLASH radiotherapy's bold approach to cancer treatment (spectrum.ieee.org)
- The yoghurt delivery women combatting loneliness in Japan (www.bbc.com)
- Tinnitus Is Connected to Sleep (www.sciencealert.com)
- PC processors entered the Gigahertz era today in the year 2000 with AMD's Athlon (www.tomshardware.com)
- Files are the interface humans and agents interact with (madalitso.me)
- US economy sheds 92,000 jobs in February in sharp slide (www.ft.com)
- Boy I was wrong about the Fediverse (matduggan.com)
- Ki Editor - an editor that operates on the AST (ki-editor.org)
- Uploading Pirated Books via BitTorrent Qualifies as Fair Use, Meta Argues (torrentfreak.com)
- QGIS 4.0 (changelog.qgis.org)
- Sarvam 105B, the first competitive Indian open source LLM (www.sarvam.ai)
- Helix: A post-modern text editor (helix-editor.com)
GitHub Trending(11)
Product Hunt(15)
- GetBeel
Let AI collect invoices and do reconciliation in automatic
- NotchPad
The secure notepad and clipboard manager for your Mac.
- Thinking Line
AI-powered doodle video and vector generator
- LTX Desktop
Local open-source LTX video editor optimized for GPUs
- TestSprite 2.1
Agentic testing for the AI-native team.
- Olmo Hybrid
7B open model mixing transformers and linear RNNs
- 21st Agents SDK
SDK to add an Claude Code AI agent to your app
- Variant
Endless designs for your ideas, just scroll.
- Codex Security
Our application security agent
- Tailwind Form Builder
Create responsive HTML forms in minutes. No login required.
- FasterGH
Fast GitHub browsing experience
- Copperlane
Turn hours of loan processing into seconds
- Saydi
Real time voice translation for persona & work
- ChatGPT for Excel
Build and update spreadsheets with ChatGPT in real time
- Imbue
We build AI that works for humans
Hugging Face(15)
- MOOSE-Star: Unlocking Tractable Training for Scientific Discovery by Breaking the Complexity Barrier
While large language models (LLMs) show promise in scientific discovery, existing research focuses on inference or feedback-driven training, leaving the direct modeling of the generative reasoning process, P(hypothesis|background) (P(h|b)), unexplored. We demonstrate that directly training P(h|b) is mathematically intractable due to the combinatorial complexity (O(N^k)) inherent in retrieving and composing inspirations from a vast knowledge base. To break this barrier, we introduce MOOSE-Star, a unified framework enabling tractable training and scalable inference. In the best case, MOOSE-Star reduces complexity from exponential to logarithmic (O(log N)) by (1) training on decomposed subtasks derived from the probabilistic equation of discovery, (2) employing motivation-guided hierarchical search to enable logarithmic retrieval and prune irrelevant subspaces, and (3) utilizing bounded composition for robustness against retrieval noise. To facilitate this, we release TOMATO-Star, a dataset of 108,717 decomposed papers (38,400 GPU hours) for training. Furthermore, we show that while brute-force sampling hits a ''complexity wall,'' MOOSE-Star exhibits continuous test-time scaling.
- SkillNet: Create, Evaluate, and Connect AI Skills
Current AI agents can flexibly invoke tools and execute complex tasks, yet their long-term advancement is hindered by the lack of systematic accumulation and transfer of skills. Without a unified mechanism for skill consolidation, agents frequently ``reinvent the wheel'', rediscovering solutions in isolated contexts without leveraging prior strategies. To overcome this limitation, we introduce SkillNet, an open infrastructure designed to create, evaluate, and organize AI skills at scale. SkillNet structures skills within a unified ontology that supports creating skills from heterogeneous sources, establishing rich relational connections, and performing multi-dimensional evaluation across Safety, Completeness, Executability, Maintainability, and Cost-awareness. Our infrastructure integrates a repository of over 200,000 skills, an interactive platform, and a versatile Python toolkit. Experimental evaluations on ALFWorld, WebShop, and ScienceWorld demonstrate that SkillNet significantly enhances agent performance, improving average rewards by 40% and reducing execution steps by 30% across multiple backbone models. By formalizing skills as evolving, composable assets, SkillNet provides a robust foundation for agents to move from transient experience to durable mastery.
- DARE: Aligning LLM Agents with the R Statistical Ecosystem via Distribution-Aware Retrieval
Large Language Model (LLM) agents can automate data-science workflows, but many rigorous statistical methods implemented in R remain underused because LLMs struggle with statistical knowledge and tool retrieval. Existing retrieval-augmented approaches focus on function-level semantics and ignore data distribution, producing suboptimal matches. We propose DARE (Distribution-Aware Retrieval Embedding), a lightweight, plug-and-play retrieval model that incorporates data distribution information into function representations for R package retrieval. Our main contributions are: (i) RPKB, a curated R Package Knowledge Base derived from 8,191 high-quality CRAN packages; (ii) DARE, an embedding model that fuses distributional features with function metadata to improve retrieval relevance; and (iii) RCodingAgent, an R-oriented LLM agent for reliable R code generation and a suite of statistical analysis tasks for systematically evaluating LLM agents in realistic analytical scenarios. Empirically, DARE achieves an NDCG at 10 of 93.47%, outperforming state-of-the-art open-source embedding models by up to 17% on package retrieval while using substantially fewer parameters. Integrating DARE into RCodingAgent yields significant gains on downstream analysis tasks. This work helps narrow the gap between LLM automation and the mature R statistical ecosystem.
- AgentVista: Evaluating Multimodal Agents in Ultra-Challenging Realistic Visual Scenarios
Real-world multimodal agents solve multi-step workflows grounded in visual evidence. For example, an agent can troubleshoot a device by linking a wiring photo to a schematic and validating the fix with online documentation, or plan a trip by interpreting a transit map and checking schedules under routing constraints. However, existing multimodal benchmarks mainly evaluate single-turn visual reasoning or specific tool skills, and they do not fully capture the realism, visual subtlety, and long-horizon tool use that practical agents require. We introduce AgentVista, a benchmark for generalist multimodal agents that spans 25 sub-domains across 7 categories, pairing realistic and detail-rich visual scenarios with natural hybrid tool use. Tasks require long-horizon tool interactions across modalities, including web search, image search, page navigation, and code-based operations for both image processing and general programming. Comprehensive evaluation of state-of-the-art models exposes significant gaps in their ability to carry out long-horizon multimodal tool use. Even the best model in our evaluation, Gemini-3-Pro with tools, achieves only 27.3% overall accuracy, and hard instances can require more than 25 tool-calling turns. We expect AgentVista to accelerate the development of more capable and reliable multimodal agents for realistic and ultra-challenging problem solving.
- RoboPocket: Improve Robot Policies Instantly with Your Phone
Scaling imitation learning is fundamentally constrained by the efficiency of data collection. While handheld interfaces have emerged as a scalable solution for in-the-wild data acquisition, they predominantly operate in an open-loop manner: operators blindly collect demonstrations without knowing the underlying policy's weaknesses, leading to inefficient coverage of critical state distributions. Conversely, interactive methods like DAgger effectively address covariate shift but rely on physical robot execution, which is costly and difficult to scale. To reconcile this trade-off, we introduce RoboPocket, a portable system that enables Robot-Free Instant Policy Iteration using single consumer smartphones. Its core innovation is a Remote Inference framework that visualizes the policy's predicted trajectory via Augmented Reality (AR) Visual Foresight. This immersive feedback allows collectors to proactively identify potential failures and focus data collection on the policy's weak regions without requiring a physical robot. Furthermore, we implement an asynchronous Online Finetuning pipeline that continuously updates the policy with incoming data, effectively closing the learning loop in minutes. Extensive experiments demonstrate that RoboPocket adheres to data scaling laws and doubles the data efficiency compared to offline scaling strategies, overcoming their long-standing efficiency bottleneck. Moreover, our instant iteration loop also boosts sample efficiency by up to 2times in distributed environments a small number of interactive corrections per person. Project page and videos: https://robo-pocket.github.io.
- HiFi-Inpaint: Towards High-Fidelity Reference-Based Inpainting for Generating Detail-Preserving Human-Product Images
Human-product images, which showcase the integration of humans and products, play a vital role in advertising, e-commerce, and digital marketing. The essential challenge of generating such images lies in ensuring the high-fidelity preservation of product details. Among existing paradigms, reference-based inpainting offers a targeted solution by leveraging product reference images to guide the inpainting process. However, limitations remain in three key aspects: the lack of diverse large-scale training data, the struggle of current models to focus on product detail preservation, and the inability of coarse supervision for achieving precise guidance. To address these issues, we propose HiFi-Inpaint, a novel high-fidelity reference-based inpainting framework tailored for generating human-product images. HiFi-Inpaint introduces Shared Enhancement Attention (SEA) to refine fine-grained product features and Detail-Aware Loss (DAL) to enforce precise pixel-level supervision using high-frequency maps. Additionally, we construct a new dataset, HP-Image-40K, with samples curated from self-synthesis data and processed with automatic filtering. Experimental results show that HiFi-Inpaint achieves state-of-the-art performance, delivering detail-preserving human-product images.
- Large Multimodal Models as General In-Context Classifiers
Which multimodal model should we use for classification? Previous studies suggest that the answer lies in CLIP-like contrastive Vision-Language Models (VLMs), due to their remarkable performance in zero-shot classification. In contrast, Large Multimodal Models (LMM) are more suitable for complex tasks. In this work, we argue that this answer overlooks an important capability of LMMs: in-context learning. We benchmark state-of-the-art LMMs on diverse datasets for closed-world classification and find that, although their zero-shot performance is lower than CLIP's, LMMs with a few in-context examples can match or even surpass contrastive VLMs with cache-based adapters, their "in-context" equivalent. We extend this analysis to the open-world setting, where the generative nature of LMMs makes them more suitable for the task. In this challenging scenario, LMMs struggle whenever provided with imperfect context information. To address this issue, we propose CIRCLE, a simple training-free method that assigns pseudo-labels to in-context examples, iteratively refining them with the available context itself. Through extensive experiments, we show that CIRCLE establishes a robust baseline for open-world classification, surpassing VLM counterparts and highlighting the potential of LMMs to serve as unified classifiers, and a flexible alternative to specialized models.
- Interactive Benchmarks
Standard benchmarks have become increasingly unreliable due to saturation, subjectivity, and poor generalization. We argue that evaluating model's ability to acquire information actively is important to assess model's intelligence. We propose Interactive Benchmarks, a unified evaluation paradigm that assesses model's reasoning ability in an interactive process under budget constraints. We instantiate this framework across two settings: Interactive Proofs, where models interact with a judge to deduce objective truths or answers in logic and mathematics; and Interactive Games, where models reason strategically to maximize long-horizon utilities. Our results show that interactive benchmarks provide a robust and faithful assessment of model intelligence, revealing that there is still substantial room to improve in interactive scenarios. Project page: https://github.com/interactivebench/interactivebench
- DreamWorld: Unified World Modeling in Video Generation
Despite impressive progress in video generation, existing models remain limited to surface-level plausibility, lacking a coherent and unified understanding of the world. Prior approaches typically incorporate only a single form of world-related knowledge or rely on rigid alignment strategies to introduce additional knowledge. However, aligning the single world knowledge is insufficient to constitute a world model that requires jointly modeling multiple heterogeneous dimensions (e.g., physical commonsense, 3D and temporal consistency). To address this limitation, we introduce DreamWorld, a unified framework that integrates complementary world knowledge into video generators via a Joint World Modeling Paradigm, jointly predicting video pixels and features from foundation models to capture temporal dynamics, spatial geometry, and semantic consistency. However, naively optimizing these heterogeneous objectives can lead to visual instability and temporal flickering. To mitigate this issue, we propose Consistent Constraint Annealing (CCA) to progressively regulate world-level constraints during training, and Multi-Source Inner-Guidance to enforce learned world priors at inference. Extensive evaluations show that DreamWorld improves world consistency, outperforming Wan2.1 by 2.26 points on VBench. Code will be made publicly available at https://github.com/ABU121111/DreamWorld{mypink{Github}}.
- SageBwd: A Trainable Low-bit Attention
Low-bit attention, such as SageAttention, has emerged as an effective approach for accelerating model inference, but its applicability to training remains poorly understood. In prior work, we introduced SageBwd, a trainable INT8 attention that quantizes six of seven attention matrix multiplications while preserving fine-tuning performance. However, SageBwd exhibited a persistent performance gap to full-precision attention (FPA) during pre-training. In this work, we investigate why this gap occurs and demonstrate that SageBwd matches full-precision attention during pretraining. Through experiments and theoretical analysis, we reach a few important insights and conclusions: (i) QK-norm is necessary for stable training at large tokens per step, (ii) quantization errors primarily arise from the backward-pass score gradient dS, (iii) reducing tokens per step enables SageBwd to match FPA performance in pre-training, and (iv) K-smoothing remains essential for training stability, while Q-smoothing provides limited benefit during pre-training.
- Timer-S1: A Billion-Scale Time Series Foundation Model with Serial Scaling
We introduce Timer-S1, a strong Mixture-of-Experts (MoE) time series foundation model with 8.3B total parameters, 0.75B activated parameters for each token, and a context length of 11.5K. To overcome the scalability bottleneck in existing pre-trained time series foundation models, we perform Serial Scaling in three dimensions: model architecture, dataset, and training pipeline. Timer-S1 integrates sparse TimeMoE blocks and generic TimeSTP blocks for Serial-Token Prediction (STP), a generic training objective that adheres to the serial nature of forecasting. The proposed paradigm introduces serial computations to improve long-term predictions while avoiding costly rolling-style inference and pronounced error accumulation in the standard next-token prediction. Pursuing a high-quality and unbiased training dataset, we curate TimeBench, a corpus with one trillion time points, and apply meticulous data augmentation to mitigate predictive bias. We further pioneer a post-training stage, including continued pre-training and long-context extension, to enhance short-term and long-context performance. Evaluated on the large-scale GIFT-Eval leaderboard, Timer-S1 achieves state-of-the-art forecasting performance, attaining the best MASE and CRPS scores as a pre-trained model. Timer-S1 will be released to facilitate further research.
- MASQuant: Modality-Aware Smoothing Quantization for Multimodal Large Language Models
Post-training quantization (PTQ) with computational invariance for Large Language Models~(LLMs) have demonstrated remarkable advances, however, their application to Multimodal Large Language Models~(MLLMs) presents substantial challenges. In this paper, we analyze SmoothQuant as a case study and identify two critical issues: Smoothing Misalignment and Cross-Modal Computational Invariance. To address these issues, we propose Modality-Aware Smoothing Quantization (MASQuant), a novel framework that introduces (1) Modality-Aware Smoothing (MAS), which learns separate, modality-specific smoothing factors to prevent Smoothing Misalignment, and (2) Cross-Modal Compensation (CMC), which addresses Cross-modal Computational Invariance by using SVD whitening to transform multi-modal activation differences into low-rank forms, enabling unified quantization across modalities. MASQuant demonstrates stable quantization performance across both dual-modal and tri-modal MLLMs. Experimental results show that MASQuant is competitive among the state-of-the-art PTQ algorithms. Source code: https://github.com/alibaba/EfficientAI.
- Locality-Attending Vision Transformer
Vision transformers have demonstrated remarkable success in classification by leveraging global self-attention to capture long-range dependencies. However, this same mechanism can obscure fine-grained spatial details crucial for tasks such as segmentation. In this work, we seek to enhance segmentation performance of vision transformers after standard image-level classification training. More specifically, we present a simple yet effective add-on that improves performance on segmentation tasks while retaining vision transformers' image-level recognition capabilities. In our approach, we modulate the self-attention with a learnable Gaussian kernel that biases the attention toward neighboring patches. We further refine the patch representations to learn better embeddings at patch positions. These modifications encourage tokens to focus on local surroundings and ensure meaningful representations at spatial positions, while still preserving the model's ability to incorporate global information. Experiments demonstrate the effectiveness of our modifications, evidenced by substantial segmentation gains on three benchmarks (e.g., over 6% and 4% on ADE20K for ViT Tiny and Base), without changing the training regime or sacrificing classification performance. The code is available at https://github.com/sinahmr/LocAtViT/.
- RealWonder: Real-Time Physical Action-Conditioned Video Generation
Current video generation models cannot simulate physical consequences of 3D actions like forces and robotic manipulations, as they lack structural understanding of how actions affect 3D scenes. We present RealWonder, the first real-time system for action-conditioned video generation from a single image. Our key insight is using physics simulation as an intermediate bridge: instead of directly encoding continuous actions, we translate them through physics simulation into visual representations (optical flow and RGB) that video models can process. RealWonder integrates three components: 3D reconstruction from single images, physics simulation, and a distilled video generator requiring only 4 diffusion steps. Our system achieves 13.2 FPS at 480x832 resolution, enabling interactive exploration of forces, robot actions, and camera controls on rigid objects, deformable bodies, fluids, and granular materials. We envision RealWonder opens new opportunities to apply video models in immersive experiences, AR/VR, and robot learning. Our code and model weights are publicly available in our project website: https://liuwei283.github.io/RealWonder/
- On-Policy Self-Distillation for Reasoning Compression
Reasoning models think out loud, but much of what they say is noise. We introduce OPSDC (On-Policy Self-Distillation for Reasoning Compression), a method that teaches models to reason more concisely by distilling their own concise behavior back into themselves. The entire approach reduces to one idea: condition the same model on a "be concise" instruction to obtain teacher logits, and minimize per-token reverse KL on the student's own rollouts. No ground-truth answers, no token budgets, no difficulty estimators. Just self-distillation. Yet this simplicity belies surprising sophistication: OPSDC automatically compresses easy problems aggressively while preserving the deliberation needed for hard ones. On Qwen3-8B and Qwen3-14B, we achieve 57-59% token reduction on MATH-500 while improving accuracy by 9-16 points absolute. On AIME 2024, the 14B model gains 10 points with 41% compression. The secret? Much of what reasoning models produce is not just redundant-it is actively harmful, compounding errors with every unnecessary token.
Techmeme(15)
- Iran targeting commercial datacenters in the UAE and Bahrain signals a new frontier in asymmetric warfare and raises doubts over the Gulf as a global AI hub (Daniel Boffey/The Guardian)
Daniel Boffey / The Guardian : Iran targeting commercial datacenters in the UAE and Bahrain signals a new frontier in asymmetric warfare and raises doubts over the Gulf as a global AI hub — Iran's targeting of commercial datacentres in the UAE and Bahrain signals a new frontier in asymmetric warfare
- Leading the Future, a pro-AI PAC backed by Palantir-cofounder Joe Lonsdale, hit pro-regulation Democrat Alex Bores with attack ads over Bores' work for Palantir (Nancy Scola/Politico)
Nancy Scola / Politico : Leading the Future, a pro-AI PAC backed by Palantir-cofounder Joe Lonsdale, hit pro-regulation Democrat Alex Bores with attack ads over Bores' work for Palantir — NEW YORK CITY — Alex Bores had known for a while that a political operation called Leading the Future, funded …
- Caitlin Kalinowski, OpenAI's head of hardware and robotics, resigns over concerns about domestic surveillance and autonomous weapons after OpenAI's DOD contract (Sharon Goldman/Fortune)
Sharon Goldman / Fortune : Caitlin Kalinowski, OpenAI's head of hardware and robotics, resigns over concerns about domestic surveillance and autonomous weapons after OpenAI's DOD contract — Caitlin Kalinowski, who had been leading hardware and robotic engineering teams at OpenAI since November 2024, announced she has left the company.
- Sources: Palmer Luckey's ModRetro, which wants to sell updated versions of 1990s consoles like the Nintendo 64, is in talks to raise funding at a $1B valuation (Financial Times)
Financial Times : Sources: Palmer Luckey's ModRetro, which wants to sell updated versions of 1990s consoles like the Nintendo 64, is in talks to raise funding at a $1B valuation — From AI weapons to Game Boys, defence tech billionaire is in talks to raise funds for new gaming venture ModRetro
- China's commerce ministry warns of global chip shortages due to "new conflicts" between Dutch chipmaker Nexperia and its Chinese subsidiary (Eduardo Baptista/Reuters)
Eduardo Baptista / Reuters : China's commerce ministry warns of global chip shortages due to “new conflicts” between Dutch chipmaker Nexperia and its Chinese subsidiary — China's commerce ministry on Saturday raised the possibility of another global semiconductor supply chain crisis due to “new conflicts” …
- Financial figures shared with prospective investors: Anduril expects its revenue to grow 100% in 2026 to ~$4.3B, and its operating loss to rise by ~50% to $1.2B (Cory Weinberg/The Information)
Cory Weinberg / The Information : Financial figures shared with prospective investors: Anduril expects its revenue to grow 100% in 2026 to ~$4.3B, and its operating loss to rise by ~50% to $1.2B — Weapons maker Anduril expects to roughly double revenue this year to about $4.3 billion, while its operating loss would rise …
- A profile of Simile, which offers "agentic twins" modeled on real people to provide answers for polling and market research for companies like CVS and Gallup (Belle Lin/Wall Street Journal)
Belle Lin / Wall Street Journal : A profile of Simile, which offers “agentic twins” modeled on real people to provide answers for polling and market research for companies like CVS and Gallup — AI startup Simile offers ‘agentic twins’ modeled on real people to provide answers for polling and market research
- Roblox says it paid out $1.5B to game creators in 2025 and the top 1,000 earned $1.3M on average; 50%+ of creators list high school as their highest education (Cecilia D'Anastasio/Bloomberg)
Cecilia D'Anastasio / Bloomberg : Roblox says it paid out $1.5B to game creators in 2025 and the top 1,000 earned $1.3M on average; 50%+ of creators list high school as their highest education — Roblox creators are raking in hundreds of thousands a month from simple games like ‘Fisch,’ even as major studios cut jobs and cancel blockbusters.
- A look at the rivalry between Kalshi CEO Tarek Mansour and Polymarket CEO Shayne Coplan, who have competing visions for how prediction markets should grow (Bobby Allyn/NPR)
Bobby Allyn / NPR : A look at the rivalry between Kalshi CEO Tarek Mansour and Polymarket CEO Shayne Coplan, who have competing visions for how prediction markets should grow — There's one word Kalshi CEO Tarek Mansour is loath to say: Polymarket. — Like a politician swiping his never-to-be-spoken …
- Sources: online DTC luxury brand Quince is in talks to raise funding at a $10B+ valuation, up from $4.5B in July; its annualized revenue run rate has hit ~$2B (The Information)
The Information : Sources: online DTC luxury brand Quince is in talks to raise funding at a $10B+ valuation, up from $4.5B in July; its annualized revenue run rate has hit ~$2B — Online retailer Quince is in talks with investors to raise a funding round that would value the company at above $10 billion after the investment …
- China's smartphone market is seeing one of its widest price hikes as memory chip crunch deepens; Meizu announces it will suspend developing new mobile hardware (Nikkei Asia)
Nikkei Asia : China's smartphone market is seeing one of its widest price hikes as memory chip crunch deepens; Meizu announces it will suspend developing new mobile hardware — TOKYO — China's smartphone market is seeing one of its biggest and widest price hikes ever, affecting almost all brands and product categories …
- Google's TIG documented 90 zero-day vulnerabilities exploited in 2025, up from 78 in 2024; commercial spyware vendors and China-linked groups led the abuse (Jessica Lyons/The Register)
Jessica Lyons / The Register : Google's TIG documented 90 zero-day vulnerabilities exploited in 2025, up from 78 in 2024; commercial spyware vendors and China-linked groups led the abuse — Of the 90 zero-days GTIG tracked in 2025, 43 hit enterprise tech — Zero-day exploitation targeting enterprise tech products reached …
- US investors are pushing Asian fund managers to create special vehicles so they can invest in Asia while avoiding US investment restrictions on Chinese tech (Financial Times)
Financial Times : US investors are pushing Asian fund managers to create special vehicles so they can invest in Asia while avoiding US investment restrictions on Chinese tech — Institutional clients increasingly want so-called parallel funds that exclude certain assets — US investors are increasingly …
- The US' February jobs report shows the tech sector's post-2022 job losses are now outpacing past downturns in 2008 and 2020 (Business Insider)
Business Insider : The US' February jobs report shows the tech sector's post-2022 job losses are now outpacing past downturns in 2008 and 2020 — - Tech industries are losing their strength. — One economist said tech job losses outpaced the past two recessions. — Still, there is weakness in other areas of the job market.
- Alphabet gives Sundar Pichai a new three-year pay deal worth up to $692M, with stock incentives worth as much as $350M linked to the growth of Waymo and Wing (Stephen Morris/Financial Times)
Stephen Morris / Financial Times : Alphabet gives Sundar Pichai a new three-year pay deal worth up to $692M, with stock incentives worth as much as $350M linked to the growth of Waymo and Wing — Alphabet boss gets bumper package of stock options tied to the share price and growth of Waymo in the next three years
Solidot(15)
- 在卫星照片披露美国军事基地损失之后 Planet Labs 停止发布卫星照片
过去几天 Planet Labs 公司的卫星图像展示了美国在中东军事基地的战损情况,包括移动雷达 THAAD 遭到攻击的画面。Planet 周五宣布停止发布部分地区的卫星照片。Planet 运营着数百颗地球成像卫星,能每天对地球上的每一块陆地进行一次观测。其客户包括智库、非政府组织、学术机构、新闻媒体以及农业、林业和能源等行业的商业用户。它还与美国军方和情报机构签署了出售卫星照片的合同。Planet 宣布对特定地区的卫星照片强制推迟 96 小时发布。
- AI 翻译工具会将“幻觉”加入到维基百科文章
维基百科编辑实施了新政策,限制使用 AI 翻译工具将英文条目翻译到其它语言的贡献者,原因是他们发现 AI 工具会在译文中加入“幻觉”——即原文不存在的内容。问题与非营利组织 Open Knowledge Association (OKA)有关,该组织主要依赖来自“全球南方(Global South)”的廉价劳工充当翻译合同工,将英文维基百科文章翻译到其他语言。部分译者开始使用 Google Gemini 和 ChatGPT 等工具加快翻译,但编辑在审阅译文时发现了大量错误,包括事实错误、缺少引用以及引用不相关来源。
- 苹果禁止美国用户下载字节跳动的其它应用
拥有中国 App Store 账户的美国 iPhone 用户报告他们无法再下载或更新字节跳动的其它应用。用户会看到警告,“此应用在您所在的国家或地区不可用”。苹果是在遵守 2024 年美国国会通过的 TikTok 法案——《The Protecting Americans from Foreign Adversary Controlled Applications Act》,法案主要针对 TikTok,但也涵盖了字节跳动的其它应用如 TikTok 中国版抖音、AI 智能助手豆包以及阅读平台番茄小说。
- 小行星 2024 YR4 不会撞击月球
去年引发广泛关注的小行星 2024 YR4 一度成为近 20 年发现的最危险的小行星,它撞击地球的可能性已经排除,但有 4% 的概率会在 2032 年 12 月 22 日撞击月球。根据天文学家利用韦伯太空望远镜近红外相机 NIRCam 对其的最新观测,它也不会撞击月球了。2024 YR4 会安全从距离月球 2 万多公里处掠过。
- OpenWrt 25.12.0 释出
面向路由器等嵌入式设备的发行版 OpenWrt 释出了 v25.12.0。OpenWrt 25.12.0 以 2025 年 4 月 1 日去世的 Dave Täht 名字命名,他是 Bufferbloat 项目联合创始人,致力于降低网络延迟,他的工作让无数人的网速更快更稳定可靠。新版本的主要变化包括:包管理器从 opkg 切换到 apk,opkg 不再维护;attended.sysupgrade LuCI 大幅简化了升级流程;保留 Shell 历史;支持大量新设备,总数超过 2200 种,等等。
- 研究发现减少服用 GLP-1 的次数仍然能维持体重减轻
GLP-1 减肥药已经改变了无数肥胖者的生活,但停用减肥药的人通常会恢复大部分减轻的体重。GLP-1 减肥药价格不菲,如果需要终身服药那么这就给服用者带来了经济上的巨大负担。发表在《Obesity》的一项研究探讨了如何在减肥的同时减少减肥药支出的可能性。小规模实验发现,减少 GLP-1 减肥药的服用间隔仍然维持减肥效果。30 名服用减肥药的人参与了研究,23 人将服药间隔改为两周或至少 10 天,另外 7 人服药间隔更长。研究人员发现,几乎所有人减肥后的 BMI 指数都保持稳定,只有 5 人体重略有回升。有 4 人在体重再次增加后恢复到了原来的服药方案。研究人员表示需要扩大规模验证这项发现。
- 美国国会将国际空间站寿命延长到 2032 年,要求 NASA 转向商业空间站
最近修订的参议院授权法案(Senate authorization bill)将国际空间站的寿命从 2030 年延长到 2032 年,同时要求 NASA 转向商业空间站,加速用商业空间站取代国际空间站。法案要求在批准 60 天内 NASA 必须公开近地轨道商业空间站的需求;90 天内发布最终的征求建议书征求行业的回应;180 天内与两家或两家以上的商业供应商签订合同。Axiom Space、Blue Origin、Vast 和 Voyager 等私营公司正在敲定商业空间站的设计方案,它们都希望 NASA 提供更多需求信息,包括宇航员在空间站驻留时间、所需的科学设备类型等等。
- LGPL 授权代码用 AI 重写后改用 MIT 授权
重新授权开源项目的许可协议在开源领域是非常困难的,因为这通常需要所有曾贡献过一行代码的人一致同意,这对历史悠久的项目而言是几乎不可能完成的任务。Python 字符编码检测器项目 chardet 移植自用 C++ 开发的 Mozilla 项目,采用了与原项目相同的 LGPL 许可证,LGPL 许可对商业使用不是太友好。维护者最近在 Claude Code 的帮助下重写了库发布了 v7.0.0 版本,将许可协议从 LGPL 更改为 MIT。项目原作者 a2mark 认为此举构成了潜在的 GPL 违反,因为开发者已经接触过原代码,并非是净室实现,因此完全重写代码的说法是没有意义的。
- 大模型提示注入漏洞导致四千开发者机器被入侵
2026 年 2 月 17 日,有人在 npm 上发布了 cline@2.3.0,它与之前的版本基本上相同,唯一的区别是在 package.json 中加入了一行代码:"postinstall": "npm install -g openclaw@latest"。在之后的八小时内,所有安装或更新 Cline 的开发者都在未经许可的情况下,在计算机上全局安装了 OpenClaw——拥有完整系统访问权限的 AI 智能体。在该软件包被撤回之前,其下载量大约 4000 次。有意思的不是有效载荷,而是攻击者最初是如何获取到 npm 令牌的:将一个提示信息注入到 GitHub 问题标题中,AI 分类机器人读取了提示信息,将其解释为指令并执行。
- 微软确认开发代号为 Project Helix 的下一代 Xbox
微软新上任的游戏业务负责人 Asha Sharma 确认了该公司正在开发代号为 Project Helix 的下一代 Xbox 游戏机。关于新主机的有效信息很少,但看起来它可能类似 Valve 的 Linux 游戏机 Steam Machine,模糊了游戏机和 PC 的界限,能同时运行 Xbox 和 PC 游戏。为了维持向后兼容性,新主机可能会继续使用 AMD 的 SoC,结合 Xbox 硬件与 PC 架构。Project Helix 可能会标志着游戏机生态系统结构的重大转变,从封闭的硬件平台转向更接近于统一的 PC-主机环境。
- 日本首次批准 iPS 细胞再生医疗产品
日本厚生劳动省以附带条件和限期的方式批准制造和销售使用诱导多能干细胞(iPS 细胞)的再生医疗产品。此次获批的是用于重度心力衰竭的 ReHeart 和用于帕金森病的 Amchepry。此次批准在 iPS 细胞实际应用于再生医疗方面开创了全球先河。ReHeart 预计价格在 1000 万日元(约合人民币 44 万元)以上,Amchepry 价格也不菲。此次批准期限为 7 年,如果能通过治疗确认有效性,就将转为无条件批准。ReHeart 用于血管堵塞导致血液难以到达心脏的“缺血性心肌病”引发的重度心力衰竭。它的原理是将来自他人 iP S细胞的心肌细胞培育成薄膜状并贴在心脏表面,使之生成新的血管。该药由源于大阪大学的初创企业“Cuorips”(东京)研发。Amchepry 的对象是脑内释放神经传导物质多巴胺的神经细胞减少引发身体僵硬及手足颤抖的“帕金森病”。原理是将他人的iPS细胞培育成释放多巴胺的神经前体细胞,并移植到头部。这可能有助于根治。它由住友制药(大阪市)研发。
- 关注面向创业公司和投资机构的 GTC 2026
3月16 - 19日,NVIDIA初创加速计划将携手创业生态合作伙伴、优秀会员企业代表及创投联盟资深投资人在GTC 2026上带来全新会议特辑。会议包含3场针对中国创业者的精彩演讲。特辑内容将围绕中国创业生态格局、前沿技术趋势、2025年中国AI市场前景,以及重点行业投资方向等议题展开,全景呈现当前AI创业、前沿技术领域的热门话题。GTC现场还将设置Inception Startup Pavilion创业企业展区、投资人AI Day、创业公司和投资机构路演等环节。
- 十分之一的 Firefox 崩溃是比特翻转导致的
Mozilla Staff Platform Engineer Gabriele Svelto 称十分之一的 Firefox 崩溃是比特翻转导致的。比特翻转(Bitflips)是指储存在电子设备上的个别比特发生翻转的事件,比如从 0 变为 1 或反之亦然。导致比特翻转的自然因素主要包括宇宙射线、功率波动和温度等。Firefox 去年部署了浏览器崩溃后在用户电脑上运行的内存测试工具。上周 Firefox 收到了 47 万份崩溃报告。崩溃报告是用户自愿递交的,因此实际崩溃数量通常会是报告的数倍。47 万份崩溃报告中约 2.5 万份检测到可能是比特翻转导致的。意味着每 20 次崩溃中就有一次可能是由内存不稳定或间歇性出错导致的。由于检测方法非常保守,实际数量至少是两倍,即十分之一。Gabriele Svelto 指出硬件不稳定的用户比硬件稳定的用户更可能遭遇崩溃。他表示今天的笔记本电脑和智能手机的内存通常是焊在设备上,要更换基本上不可能。
- Epic CEO Tim Sweeney 同意在 2032 年前停止批评 Play Store
Google 和 Epic 就《堡垒之夜》的佣金比例分歧达成和解,Google 同意降低所有应用的佣金比例。这场诉讼对 Epic CEO Tim Sweeney 而言是大获全胜,他完全实现了自己的目标。当然他也“被迫”做出了让步:根据和解协议,Sweeney 同意在 2032 年之前停止批评 Google 的应用商店政策,禁止进一步迫使 Google 再次修改应用商店政策,必须公开支持 Google 修改后的政策,甚至可能需要在世界各地的法庭出庭为与 Google 达成的这项协议辩护。Sweeney 已经通过其社交媒体账号赞美了 Google。
- 维基百科因安全失误短暂进入只读模式
维基媒体基金会的安全工程师在执行安全审查时加载随机用户脚本,结果加载了来自 ruwiki 的一个恶意脚本,恶意脚本利用安全工程师的高访问权限在维基百科快速扩散,为了遏制破坏,维基媒体基金会项目被迫进入只读模式两小时,短暂禁用了大部分用户脚本。维基媒体基金会称,恶意脚本导致 Meta-Wiki 页面被删除,但被删除的网页已经恢复,它不认为恶意代码造成了永久性破坏,也没有用户信息在此次事件中泄露。维基媒体基金会表示正与社区协商,开发针对用户脚本的安全缓解措施,以最大限度的降低未来发生此类事件的风险。