OrangeBot.AI Digest — 2026-03-19
86 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Anthropic takes legal action against OpenCode (github.com)
- Google details new 24-hour process to sideload unverified Android apps (arstechnica.com)
- An update on Steam / GOG changes for OpenTTD (www.openttd.org)
- 4Chan mocks £520k fine for UK online safety breaches (www.bbc.com)
- Show HN: Three new Kitten TTS models – smallest less than 25MB (github.com)
- macOS 26 breaks custom DNS settings including .internal (gist.github.com)
- Astral to Join OpenAI (astral.sh)
- Juggalo makeup blocks facial recognition technology (2019) (consequence.net)
- Afroman Wins Civil Trial over Use of Police Raid Footage in His Music Videos (www.nytimes.com)
- Iran war energy shock sparks global push to reduce fossil fuel dependence (www.reuters.com)
- Denmark was reportedly preparing for full-scale war with the US over Greenland (bsky.app)
- “Your frustration is the product” (daringfireball.net)
- Afroman found not liable in defamation case (nypost.com)
- Conway's Game of Life, in real life (lcamtuf.substack.com)
- Mozilla to launch free built-in VPN in upcoming Firefox 149 (cyberinsider.com)
GitHub Trending(11)
Product Hunt(15)
- MelonSound
Your Local AI Music Studio for macOS
- talat
Realtime meeting notes that don’t leave your Mac
- Cimanote
The fast, clean note app Evernote used to be
- Stitch 2.0 by Google
Vibe design beautiful production-ready UI in seconds
- Billy.sh
Local AI coding assistant for your terminal using Ollama
- Link AI
The Agentic Business Suite that replaces your entire stack
- Offload
Offload your test suite to speed up the agent loop
- Bit Office
Pixel office for AI agents and multi-agent collaboration
- GB1: The AI from the UK
Your private, planet-friendly AI assistant from the UK.
- MCPCore
Build AI-powered MCP servers in the cloud
- InfrOS
Predict and validate cloud architectures before launch
- Netlify.new
Start a project with just a prompt on Netlify
- MiniMax-M2.7
Self-evolving AI model powering autonomous agents
- Doodles Ai
An artist platform using a self-contained Doodles IP LLM
- PixelClaw
A tiny pixel crab that lives on your Dock
Hugging Face(15)
- MetaClaw: Just Talk -- An Agent That Meta-Learns and Evolves in the Wild
Large language model (LLM) agents are increasingly used for complex tasks, yet deployed agents often remain static, failing to adapt as user needs evolve. This creates a tension between the need for continuous service and the necessity of updating capabilities to match shifting task distributions. On platforms like OpenClaw, which handle diverse workloads across 20+ channels, existing methods either store raw trajectories without distilling knowledge, maintain static skill libraries, or require disruptive downtime for retraining. We present MetaClaw, a continual meta-learning framework that jointly evolves a base LLM policy and a library of reusable behavioral skills. MetaClaw employs two complementary mechanisms. Skill-driven fast adaptation analyzes failure trajectories via an LLM evolver to synthesize new skills, enabling immediate improvement with zero downtime. Opportunistic policy optimization performs gradient-based updates via cloud LoRA fine-tuning and Reinforcement Learning with a Process Reward Model (RL-PRM). This is triggered during user-inactive windows by the Opportunistic Meta-Learning Scheduler (OMLS), which monitors system inactivity and calendar data. These mechanisms are mutually reinforcing: a refined policy generates better trajectories for skill synthesis, while richer skills provide higher-quality data for policy optimization. To prevent data contamination, a versioning mechanism separates support and query data. Built on a proxy-based architecture, MetaClaw scales to production-size LLMs without local GPUs. Experiments on MetaClaw-Bench and AutoResearchClaw show that skill-driven adaptation improves accuracy by up to 32% relative. The full pipeline advances Kimi-K2.5 accuracy from 21.4% to 40.6% and increases composite robustness by 18.3%. Code is available at https://github.com/aiming-lab/MetaClaw.
- Video-CoE: Reinforcing Video Event Prediction via Chain of Events
Despite advances in the application of MLLMs for various video tasks, video event prediction (VEP) remains relatively underexplored. VEP requires the model to perform fine-grained temporal modeling of videos and establish logical relationships between videos and future events, which current MLLMs still struggle with. In this work, we first present a comprehensive evaluation of current leading MLLMs on the VEP task, revealing the reasons behind their inaccurate predictions, including lack of logical reasoning ability for future events prediction and insufficient utilization of visual information. To address these challenges, we propose Chain of Events (CoE) paradigm, which constructs temporal event chains to implicitly enforce MLLM focusing on the visual content and the logical connections between videos and future events, incentivizing model's reasoning capability with multiple training protocols. Experimental results on public benchmarks demonstrate that our method outperforms both leading open-source and commercial MLLMs, establishing a new state-of-the-art on the VEP task. Codes and models will be released soon.
- MosaicMem: Hybrid Spatial Memory for Controllable Video World Models
Video diffusion models are moving beyond short, plausible clips toward world simulators that must remain consistent under camera motion, revisits, and intervention. Yet spatial memory remains a key bottleneck: explicit 3D structures can improve reprojection-based consistency but struggle to depict moving objects, while implicit memory often produces inaccurate camera motion even with correct poses. We propose Mosaic Memory (MosaicMem), a hybrid spatial memory that lifts patches into 3D for reliable localization and targeted retrieval, while exploiting the model's native conditioning to preserve prompt-following generation. MosaicMem composes spatially aligned patches in the queried view via a patch-and-compose interface, preserving what should persist while allowing the model to inpaint what should evolve. With PRoPE camera conditioning and two new memory alignment methods, experiments show improved pose adherence compared to implicit memory and stronger dynamic modeling than explicit baselines. MosaicMem further enables minute-level navigation, memory-based scene editing, and autoregressive rollout.
- Alignment Makes Language Models Normative, Not Descriptive
Post-training alignment optimizes language models to match human preference signals, but this objective is not equivalent to modeling observed human behavior. We compare 120 base-aligned model pairs on more than 10,000 real human decisions in multi-round strategic games - bargaining, persuasion, negotiation, and repeated matrix games. In these settings, base models outperform their aligned counterparts in predicting human choices by nearly 10:1, robustly across model families, prompt formulations, and game configurations. This pattern reverses, however, in settings where human behavior is more likely to follow normative predictions: aligned models dominate on one-shot textbook games across all 12 types tested and on non-strategic lottery choices - and even within the multi-round games themselves, at round one, before interaction history develops. This boundary-condition pattern suggests that alignment induces a normative bias: it improves prediction when human behavior is relatively well captured by normative solutions, but hurts prediction in multi-round strategic settings, where behavior is shaped by descriptive dynamics such as reciprocity, retaliation, and history-dependent adaptation. These results reveal a fundamental trade-off between optimizing models for human use and using them as proxies for human behavior.
- Complementary Reinforcement Learning
Reinforcement Learning (RL) has emerged as a powerful paradigm for training LLM-based agents, yet remains limited by low sample efficiency, stemming not only from sparse outcome feedback but also from the agent's inability to leverage prior experience across episodes. While augmenting agents with historical experience offers a promising remedy, existing approaches suffer from a critical weakness: the experience distilled from history is either stored statically or fail to coevolve with the improving actor, causing a progressive misalignment between the experience and the actor's evolving capability that diminishes its utility over the course of training. Inspired by complementary learning systems in neuroscience, we present Complementary RL to achieve seamless co-evolution of an experience extractor and a policy actor within the RL optimization loop. Specifically, the actor is optimized via sparse outcome-based rewards, while the experience extractor is optimized according to whether its distilled experiences demonstrably contribute to the actor's success, thereby evolving its experience management strategy in lockstep with the actor's growing capabilities. Empirically, Complementary RL outperforms outcome-based agentic RL baselines that do not learn from experience, achieving 10% performance improvement in single-task scenarios and exhibits robust scalability in multi-task settings. These results establish Complementary RL as a paradigm for efficient experience-driven agent learning.
- When AI Navigates the Fog of War
Can AI reason about a war before its trajectory becomes historically obvious? Analyzing this capability is difficult because retrospective geopolitical prediction is heavily confounded by training-data leakage. We address this challenge through a temporally grounded case study of the early stages of the 2026 Middle East conflict, which unfolded after the training cutoff of current frontier models. We construct 11 critical temporal nodes, 42 node-specific verifiable questions, and 5 general exploratory questions, requiring models to reason only from information that would have been publicly available at each moment. This design substantially mitigates training-data leakage concerns, creating a setting well-suited for studying how models analyze an unfolding crisis under the fog of war, and provides, to our knowledge, the first temporally grounded analysis of LLM reasoning in an ongoing geopolitical conflict. Our analysis reveals three main findings. First, current state-of-the-art large language models often display a striking degree of strategic realism, reasoning beyond surface rhetoric toward deeper structural incentives. Second, this capability is uneven across domains: models are more reliable in economically and logistically structured settings than in politically ambiguous multi-actor environments. Finally, model narratives evolve over time, shifting from early expectations of rapid containment toward more systemic accounts of regional entrenchment and attritional de-escalation. Since the conflict remains ongoing at the time of writing, this work can serve as an archival snapshot of model reasoning during an unfolding geopolitical crisis, enabling future studies without the hindsight bias of retrospective analysis.
- GigaWorld-Policy: An Efficient Action-Centered World--Action Model
World-Action Models (WAM) initialized from pre-trained video generation backbones have demonstrated remarkable potential for robot policy learning. However, existing approaches face two critical bottlenecks that hinder performance and deployment. First, jointly reasoning over future visual dynamics and corresponding actions incurs substantial inference overhead. Second, joint modeling often entangles visual and motion representations, making motion prediction accuracy heavily dependent on the quality of future video forecasts. To address these issues, we introduce GigaWorld-Policy, an action-centered WAM that learns 2D pixel-action dynamics while enabling efficient action decoding, with optional video generation. Specifically, we formulate policy training into two coupled components: the model predicts future action sequences conditioned on the current observation, and simultaneously generates future videos conditioned on the predicted actions and the same observation. The policy is supervised by both action prediction and video generation, providing richer learning signals and encouraging physically plausible actions through visual-dynamics constraints. With a causal design that prevents future-video tokens from influencing action tokens, explicit future-video generation is optional at inference time, allowing faster action prediction during deployment. To support this paradigm, we curate a diverse, large-scale robot dataset to pre-train an action-centered video generation model, which is then adapted as the backbone for robot policy learning. Experimental results on real-world robotic platforms show that GigaWorld-Policy runs 9x faster than the leading WAM baseline, Motus, while improving task success rates by 7%. Moreover, compared with pi-0.5, GigaWorld-Policy improves performance by 95% on RoboTwin 2.0.
- Look Before Acting: Enhancing Vision Foundation Representations for Vision-Language-Action Models
Vision-Language-Action (VLA) models have recently emerged as a promising paradigm for robotic manipulation, in which reliable action prediction critically depends on accurately interpreting and integrating visual observations conditioned on language instructions. Although recent works have sought to enhance the visual capabilities of VLA models, most approaches treat the LLM backbone as a black box, providing limited insight into how visual information is grounded into action generation. Therefore, we perform a systematic analysis of multiple VLA models across different action-generation paradigms and observe that sensitivity to visual tokens progressively decreases in deeper layers during action generation. Motivated by this observation, we propose DeepVision-VLA, built on a Vision-Language Mixture-of-Transformers (VL-MoT) framework. This framework enables shared attention between the vision foundation model and the VLA backbone, injecting multi-level visual features from the vision expert into deeper layers of the VLA backbone to enhance visual representations for precise and complex manipulation. In addition, we introduce Action-Guided Visual Pruning (AGVP), which leverages shallow-layer attention to prune irrelevant visual tokens while preserving task-relevant ones, reinforcing critical visual cues for manipulation with minimal computational overhead. DeepVision-VLA outperforms prior state-of-the-art methods by 9.0\% and 7.5\% on simulated and real-world tasks, respectively, providing new insights for the design of visually enhanced VLA models.
- Temporal Gains, Spatial Costs: Revisiting Video Fine-Tuning in Multimodal Large Language Models
Multimodal large language models (MLLMs) are typically trained in multiple stages, with video-based supervised fine-tuning (Video-SFT) serving as a key step for improving visual understanding. Yet its effect on the fine-grained evolution of visual capabilities, particularly the balance between spatial and temporal understanding, remains poorly understood. In this paper, we systematically study how Video-SFT reshapes visual capabilities in MLLMs. Across architectures, parameter scales, and frame sampling settings, we observe a consistent pattern: Video-SFT reliably improves video performance, but often yields limited gains or even degradation on static image benchmarks. We further show that this trade-off is closely tied to temporal budget: increasing the number of sampled frames generally improves video performance, but does not reliably improve static image performance. Motivated by this finding, we study an instruction-aware Hybrid-Frame strategy that adaptively allocates frame counts and partially mitigates the image-video trade-off. Our results indicate that Video-SFT is not a free lunch for MLLMs, and preserving spatial understanding remains a central challenge in joint image-video training.
- BenchPreS: A Benchmark for Context-Aware Personalized Preference Selectivity of Persistent-Memory LLMs
Large language models (LLMs) increasingly store user preferences in persistent memory to support personalization across interactions. However, in third-party communication settings governed by social and institutional norms, some user preferences may be inappropriate to apply. We introduce BenchPreS, which evaluates whether memory-based user preferences are appropriately applied or suppressed across communication contexts. Using two complementary metrics, Misapplication Rate (MR) and Appropriate Application Rate (AAR), we find even frontier LLMs struggle to apply preferences in a context-sensitive manner. Models with stronger preference adherence exhibit higher rates of over-application, and neither reasoning capability nor prompt-based defenses fully resolve this issue. These results suggest current LLMs treat personalized preferences as globally enforceable rules rather than as context-dependent normative signals.
- LoST: Level of Semantics Tokenization for 3D Shapes
Tokenization is a fundamental technique in the generative modeling of various modalities. In particular, it plays a critical role in autoregressive (AR) models, which have recently emerged as a compelling option for 3D generation. However, optimal tokenization of 3D shapes remains an open question. State-of-the-art (SOTA) methods primarily rely on geometric level-of-detail (LoD) hierarchies, originally designed for rendering and compression. These spatial hierarchies are often token-inefficient and lack semantic coherence for AR modeling. We propose Level-of-Semantics Tokenization (LoST), which orders tokens by semantic salience, such that early prefixes decode into complete, plausible shapes that possess principal semantics, while subsequent tokens refine instance-specific geometric and semantic details. To train LoST, we introduce Relational Inter-Distance Alignment (RIDA), a novel 3D semantic alignment loss that aligns the relational structure of the 3D shape latent space with that of the semantic DINO feature space. Experiments show that LoST achieves SOTA reconstruction, surpassing previous LoD-based 3D shape tokenizers by large margins on both geometric and semantic reconstruction metrics. Moreover, LoST achieves efficient, high-quality AR 3D generation and enables downstream tasks like semantic retrieval, while using only 0.1%-10% of the tokens needed by prior AR models.
- ESPIRE: A Diagnostic Benchmark for Embodied Spatial Reasoning of Vision-Language Models
A recent trend in vision-language models (VLMs) has been to enhance their spatial cognition for embodied domains. Despite progress, existing evaluations have been limited both in paradigm and in coverage, hindering rapid, iterative model development. To address these limitations, we propose ESPIRE, a diagnostic benchmark for embodied spatial reasoning. ESPIRE offers a simulated world that physically grounds VLMs and evaluates them on spatial-reasoning-centric robotic tasks, thus narrowing the gap between evaluation and real-world deployment. To adapt VLMs to robotic tasks, we decompose each task into localization and execution, and frame both as generative problems, in stark contrast to predominant discriminative evaluations (e.g., via visual-question answering) that rely on distractors and discard execution. This decomposition further enables a fine-grained analysis beyond passive spatial reasoning toward reasoning to act. We systematically design ESPIRE both at the instruction level and at the environment level, ensuring broad coverage of spatial reasoning scenarios. We use ESPIRE to diagnose a range of frontier VLMs and provide in-depth analysis of their spatial reasoning behaviors.
- Conservative Offline Robot Policy Learning via Posterior-Transition Reweighting
Offline post-training adapts a pretrained robot policy to a target dataset by supervised regression on recorded actions. In practice, robot datasets are heterogeneous: they mix embodiments, camera setups, and demonstrations of varying quality, so many trajectories reflect recovery behavior, inconsistent operator skill, or weakly informative supervision. Uniform post-training gives equal credit to all samples and can therefore average over conflicting or low-attribution data. We propose Posterior-Transition Reweighting (PTR), a reward-free and conservative post-training method that decides how much each training sample should influence the supervised update. For each sample, PTR encodes the observed post-action consequence as a latent target, inserts it into a candidate pool of mismatched targets, and uses a separate transition scorer to estimate a softmax identification posterior over target indices. The posterior-to-uniform ratio defines the PTR score, which is converted into a clipped-and-mixed weight and applied to the original action objective through self-normalized weighted regression. This construction requires no tractable policy likelihood and is compatible with both diffusion and flow-matching action heads. Rather than uniformly trusting all recorded supervision, PTR reallocates credit according to how attributable each sample's post-action consequence is under the current representation, improving conservative offline adaptation to heterogeneous robot data.
- Stereo World Model: Camera-Guided Stereo Video Generation
We present StereoWorld, a camera-conditioned stereo world model that jointly learns appearance and binocular geometry for end-to-end stereo video generation.Unlike monocular RGB or RGBD approaches, StereoWorld operates exclusively within the RGB modality, while simultaneously grounding geometry directly from disparity. To efficiently achieve consistent stereo generation, our approach introduces two key designs: (1) a unified camera-frame RoPE that augments latent tokens with camera-aware rotary positional encoding, enabling relative, view- and time-consistent conditioning while preserving pretrained video priors via a stable attention initialization; and (2) a stereo-aware attention decomposition that factors full 4D attention into 3D intra-view attention plus horizontal row attention, leveraging the epipolar prior to capture disparity-aligned correspondences with substantially lower compute. Across benchmarks, StereoWorld improves stereo consistency, disparity accuracy, and camera-motion fidelity over strong monocular-then-convert pipelines, achieving more than 3x faster generation with an additional 5% gain in viewpoint consistency. Beyond benchmarks, StereoWorld enables end-to-end binocular VR rendering without depth estimation or inpainting, enhances embodied policy learning through metric-scale depth grounding, and is compatible with long-video distillation for extended interactive stereo synthesis.
- Expert Threshold Routing for Autoregressive Language Modeling with Dynamic Computation Allocation and Load Balancing
Token-choice Mixture-of-Experts (TC-MoE) routes each token to a fixed number of experts, limiting dynamic computation allocation and requiring auxiliary losses to maintain load balance. We propose Expert Threshold (ET) routing, where each expert maintains an exponential moving average (EMA) threshold estimated from the global token distribution. At both training and inference, each token is independently routed to an expert if its score exceeds the expert's threshold, enabling dynamic computation allocation while achieving load balance without auxiliary losses. This fully causal mechanism eliminates dependence on other tokens in the batch, making it well-suited for autoregressive language modeling. In pretraining experiments scaling to 2.4B parameters on FineWeb-Edu, ET achieves 0.067 lower cross-entropy loss than TC-MoE, equivalent to reaching the same performance with 1.6times fewer tokens.
Techmeme(15)
- Filing: the Pentagon says Anthropic's use of foreign workers, including from China, poses security risks and that its case is "different" from other companies' (Maria Curi/Axios)
Maria Curi / Axios : Filing: the Pentagon says Anthropic's use of foreign workers, including from China, poses security risks and that its case is “different” from other companies' — The Pentagon is highlighting new national security concerns about Anthropic's use of foreign workers, including from China, according to a court filing.
- Verily raised $300M led by Series X Capital, giving it ~2 years of runway per a source; Alphabet remains an investor but no longer has a controlling stake (Axios)
Axios : Verily raised $300M led by Series X Capital, giving it ~2 years of runway per a source; Alphabet remains an investor but no longer has a controlling stake — What to read next
- Sources: Kalshi is raising ~$1B led by Coatue at a $22B valuation, up from the $11B valuation announced in December, and has a revenue run rate of ~$1.5B (Yuliya Chernova/Wall Street Journal)
Yuliya Chernova / Wall Street Journal : Sources: Kalshi is raising ~$1B led by Coatue at a $22B valuation, up from the $11B valuation announced in December, and has a revenue run rate of ~$1.5B — Coatue Management is leading the new financing that would total about $1 billion for the prediction-market platform, according to sources
- Amazon acquires Zurich-based autonomous robotics startup Rivr, formerly Swiss-Mile; PitchBook: the startup was valued at $110M in an August 2024 funding round (The Information)
The Information : Amazon acquires Zurich-based autonomous robotics startup Rivr, formerly Swiss-Mile; PitchBook: the startup was valued at $110M in an August 2024 funding round — Amazon has acquired autonomous robotics startup Rivr, an Amazon spokesperson confirmed, a deal that could help the commerce …
- Signal's Moxie Marlinspike says his privacy-focused AI platform Confer will integrate its encryption tech into Meta AI to provide E2EE for chatbot interactions (Wired)
Wired : Signal's Moxie Marlinspike says his privacy-focused AI platform Confer will integrate its encryption tech into Meta AI to provide E2EE for chatbot interactions — Moxie Marlinspike says the technology powering his encrypted AI chatbot, Confer, will be integrated into Meta AI.
- Sources: Google has started consumer beta testing of a dedicated Gemini AI app for the Mac, as it seeks to compete with Mac apps for ChatGPT and Claude (Mark Gurman/Bloomberg)
Mark Gurman / Bloomberg : Sources: Google has started consumer beta testing of a dedicated Gemini AI app for the Mac, as it seeks to compete with Mac apps for ChatGPT and Claude — Google is ramping up development of a dedicated Gemini AI app for Apple Inc.'s Mac computer lineup, looking to step up competition with OpenAI and Anthropic PBC.
- Bluesky raised a $100M Series B led by Bain Capital Crypto in April 2025, following a $15M Series A in 2024 and an $8M seed in 2023; it now has over 43M users (Sarah Perez/TechCrunch)
Sarah Perez / TechCrunch : Bluesky raised a $100M Series B led by Bain Capital Crypto in April 2025, following a $15M Series A in 2024 and an $8M seed in 2023; it now has over 43M users — Social network Bluesky is gearing up for big changes with today's news that the company raised $100 million in Series B funding.
- Sources: Jeff Bezos is in talks to raise $100B for a fund that would buy companies in industrial sectors such as chipmaking and defense and revamp them with AI (Wall Street Journal)
Wall Street Journal : Sources: Jeff Bezos is in talks to raise $100B for a fund that would buy companies in industrial sectors such as chipmaking and defense and revamp them with AI — Amazon.com founder has traveled to Middle East, Singapore in fundraising effort linked to Project Prometheus AI startup
- Google moved some staffers working on Project Mariner, its AI agent that can navigate Chrome and complete tasks on a user's behalf, to higher-priority projects (Maxwell Zeff/Wired)
Maxwell Zeff / Wired : Google moved some staffers working on Project Mariner, its AI agent that can navigate Chrome and complete tasks on a user's behalf, to higher-priority projects — As Silicon Valley obsesses over a new wave of AI coding agents, Google and other AI labs are shifting their bets.
- Oasis Security, which specializes in securing non-human identities, such as AI bots and automated work tools, raised $120M, bringing its total funding to $190M (Marissa Newman/Bloomberg)
Marissa Newman / Bloomberg : Oasis Security, which specializes in securing non-human identities, such as AI bots and automated work tools, raised $120M, bringing its total funding to $190M — Oasis Security, a cybersecurity startup that helps companies manage access to their systems from non-human accounts …
- Google announces a new "advanced flow" for Android sideloading that requires a mandatory 24-hour cooling-off period to install apps from unverified developers (Dominic Preston/The Verge)
Dominic Preston / The Verge : Google announces a new “advanced flow” for Android sideloading that requires a mandatory 24-hour cooling-off period to install apps from unverified developers — The one-time ‘advanced flow’ includes 24 hours of cooling off time. … Google has revealed the “advanced flow” …
- Privacy-focused MVNO Cape raised a $100M Series C at a $900M valuation, and says its revenue grew from $4.5M in 2024 to $37M in 2025 (Thomas Brewster/Forbes)
Thomas Brewster / Forbes : Privacy-focused MVNO Cape raised a $100M Series C at a $900M valuation, and says its revenue grew from $4.5M in 2024 to $37M in 2025 — Cape cofounder and CEO John Doyle is seeing rapid revenue growth for his cell network, which deletes call logs and doesn't collect social security numbers like AT&T and Verizon do.
- Mark Zuckerberg's original metaverse vision is effectively over, after renaming Facebook to Meta and spending ~$80B on the endeavor, as his focus shifts to AI (New York Times)
New York Times : Mark Zuckerberg's original metaverse vision is effectively over, after renaming Facebook to Meta and spending ~$80B on the endeavor, as his focus shifts to AI — Five years ago, Mark Zuckerberg proclaimed that the future of Facebook would be the metaverse.
- Cursor says Composer 2 is "frontier-level at coding" and is priced at $0.50/1M input tokens and $2.50/1M output tokens, with a faster variant costing 3x more (Cursor)
Cursor : Cursor says Composer 2 is “frontier-level at coding” and is priced at $0.50/1M input tokens and $2.50/1M output tokens, with a faster variant costing 3x more — 38.0 — 40.0 — 56.9 — These quality improvements come from our first continued pretraining run …
- Sources: TP-Link told US federal agencies probing its China ties that CEO Jeffrey Chao applied for US permanent residency under the $1M Trump Gold Card program (Kate O'Keeffe/Bloomberg)
Kate O'Keeffe / Bloomberg : Sources: TP-Link told US federal agencies probing its China ties that CEO Jeffrey Chao applied for US permanent residency under the $1M Trump Gold Card program — The Chinese founder of the massive router maker facing national security probes by the Trump administration has applied …
Solidot(15)
- 居家办公有助于提高生育率
斯坦福、普林斯顿、伦敦国王学院的研究人员调查了居家办公和生育率的关系。研究是基于 2024 年 11 月到 2025 年 2 月之间收集的 38 国 Global Survey of Working Arrangements(G-SWA)调查数据,以及 2022 年 12 月到 2025 年 12 月间收集的 U.S. Survey of Working Arrangements and Attitudes (SWAA)美国调查数据。研究显示居家办公有助于提高生育率。在 38 国样本中,如果伴侣双方每周至少一天在家工作,则每位女性的终生生育率会增加 0.32 个孩子;美国样本中则会增加 0.45 个孩子。
- Firefox v149 将内置 VPN
Mozilla 宣布,3 月 24 日释出的 Firefox v149 将内置 VPN 服务。该服务将首先提供给美国、法国、德国和英国的用户,免费套餐的流量为每月 50GB。Firefox VPN 与 Mozilla VPN 不同,Mozilla VPN 是一项独立的付费服务,可同时在五台设备上使用;而 Firefox VPN 只限于浏览器本身,设计通过 Mozilla 管理的服务器路由流量,隐藏用户的真实 IP 地址。Mozilla 此前强调,不会出售个人数据,同步浏览数据如历史和书签将依赖于端到端加密。
- 美国私人太空公司计划捕捉小行星
美国私人太空公司 TransAstra 周三宣布了名为“New Moon”的任务,该任务旨在捕获一颗房屋大小、质量 100 吨左右的小行星,将其转移到地球附近的 L2 点。TransAstra CEO Joel Sercel 表示设想将小行星变成一个材料加工和制造的机器人研发基地,长远目标是无需在地面制造太空飞行器硬件,无需从地球发射推进剂,直接利用太空原材料提取推进剂。他表示如果资金到位,New Moon 任务最早在 2028 年或 2029 年发射太空“捕获袋”与一颗小行星交会。
- GNOME 50 释出
桌面环境项目 GNOME 释出了 v50。该版本被命名为“东京(Tokyo)”,以表彰 GNOME.Asia Summit 2025 当地组织者的工作。GNOME 50 的主要特性包括:改进了家长控制,家长和监护人能监控儿童的屏幕使用时间,为儿童账户设置就寝自动锁屏限制;增强 Orca 屏幕阅读器;改进文件管理器的 UI 和性能;Wayland 会话支持鼠标预览;新的 Reduced Motion 选项减少动画造成的任何不适或干扰;文档查看器支持添加文本注释、添加线条和高亮显示;等等,更多可浏览发布公告。
- 2026 年图灵奖授予了两位量子信息理论的奠基人
2026 年图灵奖授予了量子信息理论的两位奠基人 Charles Henry Bennett 和 Gilles Brassard,两人共享 100 万美元奖金。两人是在 1979 年出席一学术会议期间抽空去游泳时相遇的,他们讨论了利用量子机制制造永远也无法伪造的货币。他们的合作推动了量子密码学的诞生,1984 年他们提出了第一个实用的量子密码协议 BB84,他们的论文《Quantum Cryptography: Public Key Distribution and Coin Tossing》证明,即使面临一个拥有无限计算能力和尖端技术如量子计算机的对手,通信双方仍然能建立一个由物理定律确保安全的加密密钥。BB84 依赖于量子信息的一个基本属性:它无法在不被干扰的情况下被复制或测量。任何窃听尝试都会在信息泄露前留下可检测的痕迹。
- Meta 将于 6 月 15 日关闭 VR 社交网络 Horizon Worlds
元宇宙已成为过去,Meta 将于 6 月 15 日关闭其 VR 社交网络 Horizon Worlds。该应用将于 3 月底从 Quest 应用商店下架,6 月 15 日前从 Quest 头显中完全移除,此后将转变为独立的“移动专属体验”。Horizon Worlds 于 2021 年末正式发布,用户可以在虚拟 3D 世界中操控虚拟化身与其他用户互动以及玩游戏。在 Meta 于 2023 年 9 月推出移动应用前,它只支持 Quest VR 平台,移动应用的工作方式类似 Roblox。Meta 数周前已经裁掉了负责元宇宙的 Reality Labs 部门逾千名员工。
- DarkSword 漏洞影响数亿 iPhone 用户
Google 安全团队披露了被称为 DarkSword 的漏洞利用链,数亿 iPhone 用户受到影响,苹果已经释出修复补丁。对 DarkSword 的利用最早可追溯到 2025 年 11 月,研究人员推测商业间谍软件正在利用该漏洞。受影响的版本从 iOS 18.4 到 18.7 版本,攻击者组合利用六个漏洞完全控制设备,之后部署恶意程序 GHOSTBLADE、GHOSTKNIFE 和 GHOSTSABER。Google 安全团队于 2025 年底向苹果公司报告了 DarkSword 使用的漏洞,所有漏洞已在 iOS 26.3 中修复。
- 调查显示近六成人愿意为保护环境而放弃经济增长
根据发表在《Ecological Economics》期刊上的一项研究,当保护环境和经济增长发生冲突时 58% 的人倾向于选择保护环境。研究分析了 92 个国家居民的反馈数据,结果显示:58% 的人更偏向保护环境;西欧、东南亚、美洲、澳大利亚和新西兰居民对环境保护的支持度最高;西方国家中的女性、年轻人、受过良好教育以及政治倾向较为自由的人更重视环境保护;东欧、中亚、非洲和中东地区对保护环境的支持度较低,研究人员猜测可能与这些地区的经济发展较低相关。
- 月球土壤能种植马铃薯
在科幻电影《火星救援(The Martian)》中,马特·达蒙扮演的宇航员 Mark Watney 因意外被孤身一人留在火星,他随后在堆肥的帮助下通过种植马铃薯生存了下来。根据发表在 bioRxiv 平台上的一项研究,研究人员报告他们在类似月球的土壤中成功种植了马铃薯,和《火星救援》类似,类月球土壤也需要大量堆肥。月球风化层缺乏植物生长所需的有机物,俄勒冈州立大学的 David Handy 和同事添加了蚯蚓粪——蚯蚓的有机排泄物。他们发现,添加 5% 蚯蚓粪的混合物能让马铃薯在模拟月球环境的严酷条件下生长。经过近两个月的生长,研究团队收获了马铃薯块茎,将其冷冻干燥并研磨成粉末进行进一步测试。DNA 分析表明,马铃薯压力相关的基因已被激活,其铜和锌含量高于地球土壤种植的马铃薯,人类食用可能有害。但这些马铃薯的营养价值与地球马铃薯相似。
- 三星准备停售三折叠屏智能手机 Galaxy Z TriFold
三星三折叠屏智能手机 Galaxy Z TriFold 仅仅上市数个月就面临停售,原因是生产成本过高和供应有限。Galaxy Z TriFold 于四个月前在韩国发布,今年一月开始在美国等市场发售,售价 2899 美元。对于在韩国销售的 Galaxy Z TriFold,三星将在本周最后一次补货。美国的库存补充次数可能更多,将会持续到产量售罄为止。三折叠屏智能手机的寿命相当短暂。
- 为什么 AI 系统无法自主学习
图灵奖得主 Yann LeCun 等人在预印本平台 arXiv 上发布论文,讨论了目前的 AI 模型在自主学习上的局限性,提出了一个受人类等启发的自主学习框架。AI 模型一旦部署,基本上无法学习任何新知识,模型的操作是固定的,如果不能适应环境,就必须由人类专家使用新数据重构。人类儿童能在不同学习模式之间切换,但 AI 模型的自监督学习、监督学习和强化学习等机器学习范式则是孤立的,混合使用不同学习模式主要是通过人类专家反复试验实现,只能针对特定应用如聊天机器人或编程助手。换句话说,当前的 AI 系统的学习是外包给人类专家,而不是其自身固有的能力。要构建能在现实世界中可靠运行的 AI 系统,自主学习应视为一项核心能力。
- Xbox One 游戏机被通过名为 Bliss 的漏洞破解
在 RE//verse 2026 大会上,Markus‘Doom’Gaasedelen 演示了利用名为“Bliss”的两漏洞组合破解了微软的 Xbox One 游戏机。Xbox One 于 2013 年上市,使用了 AMD 的 APU——基于 Jaguar 架构的 CPU 和基于 GCN 2 架构的 GPU,在此之前一直未被破解。Gaasedelen 利用了两次电压毛刺攻击(voltage glitches)跳过安全检查执行未经授权或未签名代码。
- 瑞士构建 BGP 的安全替代
边界网关路由(BGP)不是为安全设计的,而是为构成互联网的数以千计的自治系统之间大规模快速路由数据包设计的。过去四十年,BGP 运作良好,但其安全缺陷也日益显现。为堵上漏洞,BGP 引入了一系列补丁和扩展如 Resource Public Key Infrastructure (RPKI)、BGPsec 和 RPKI-based Route Origin Authorization (ROA),但无法从根本上解决问题。瑞士苏黎世联邦理工学院开发的 SCION——代表 Scalability, Control, and Isolation On Next-Generation Networks——尝试从根本上改变互联网的路由架构,提供一种更安全的替代。SCION 的首席架构师 Adrian Perrig 是苏黎世联邦理工的计算机科学教授,一直致力于提升互联网的安全。他发现安全无法拼拼凑凑,必须彻底改变设计。SCION 尝试通过三个关联机制解决 BGP 的安全缺陷:其一是多路径路由,两点之间能同时建立数十条甚至数百条并行路径,一条路径发生故障,系统会在几毫秒内完成重路由;其二是不依赖证书颁发机构的隔离域名 ISD 机制;其三是加密路径验证,路径上的每个路由器都提供一个加密签名。瑞士银行已成功测试了 SCION。
- GTC 2026 重磅 AI 会议推荐:注册观看还有机会获得 NVIDIA 定制装备
重磅一:中国创业生态精彩会议 错过首播也无妨,复制下方链接即可直达回看页面。 GTC 2026 创业企业会议特辑深度聚焦中国创业生态,干货满满,不容错过。 线上演讲:《十载相伴,NVIDIA 赋能创业公司在 AI 时代加速前行》 https://www.nvidia.cn/gtc/session-catalog/sessions/gtc26-s81981/?ncid=so-othe-950414
- 韩国游戏发行商 CEO 为避免支付合同承诺的 2.5 亿美元而求助于 ChatGPT
Unknown Worlds Entertainment 是知名水下生存游戏《Subnautica》的开发商,2021 年韩国游戏发行商 Krafton 以 5 亿美元收购了该游戏工作室,并在合同中承诺如果续作《Subnautica 2》销量足够好将额外支付 2.5 亿美元。Krafton 内部对《Subnautica 2》的销量预测相当乐观,因此 2.5 亿美元看起来不得不兑现了。然而 CEO Changhan Kim 不想支付这笔费用,他因此求助于 AI 聊天机器人 ChatGPT 而不是公司法务讨论如何避免支付 2.5 亿美元。公司法务认为奖金取消不了,但 Kim 在 ChatGPT 的帮助下设计了一个方案以莫须有理由突然解雇了 Unknown Worlds 的主要高管。被解雇的高管提起诉讼曝光了 ChatGPT 的阴谋。法官本周裁决 Unknown Worlds 前 CEO Ted Gill 恢复原职。拖延了很久的《Subnautica 2》预计将于今年 5 月发布抢先体验版本(early access)。