DIGEST · 2026-03-20

OrangeBot.AI Digest — 2026-03-20

84 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. OpenCode – The open source AI coding agent (opencode.ai)
  2. Our commitment to Windows quality (blogs.windows.com)
  3. Delve – Fake Compliance as a Service (deepdelver.substack.com)
  4. Oregon school cell phone ban: 'Engaged students, joyful teachers' (portlandtribune.com)
  5. Super Micro Shares Plunge 25% After Co-Founder Charged in $2.5B Smuggling Plot (www.forbes.com)
  6. The Los Angeles Aqueduct Is Wild (practical.engineering)
  7. Java is fast, code might not be (jvogel.me)
  8. Chuck Norris has died (variety.com)
  9. HP trialed mandatory 15-minute support call wait times (2025) (arstechnica.com)
  10. I'm OK being left behind, thanks (shkspr.mobi)
  11. France's aircraft carrier located in real time by Le Monde through fitness app (www.lemonde.fr)
  12. Entso-E final report on Iberian 2025 blackout (www.entsoe.eu)
  13. Cursor Composer 2 is just Kimi K2.5 with RL (twitter.com)
  14. Drawvg Filter for FFmpeg (ayosec.github.io)
  15. FSF statement on copyright infringement lawsuit Bartz v. Anthropic (www.fsf.org)

GitHub Trending(9)

  1. jarrodwatts / claude-hud
  2. langchain-ai / open-swe
  3. obra / superpowers
  4. opendataloader-project / opendataloader-pdf
  5. louis-e / arnis
  6. newton-physics / newton
  7. vas3k / TaxHacker
  8. TauricResearch / TradingAgents
  9. openrocket / openrocket

Product Hunt(15)

  1. Telea

    Speak like you always know what to say

  2. AI Skills Manager

    One place for all your AI skills

  3. Cacheless

    AI-Powered Mac System Data Cleaner

  4. MusicLib

    The Ultimate Sheet Music Library Solution

  5. Gately

    Everything you need to build your own membership

  6. Fig Prompt

    Build Figma plugins with just a prompt

  7. Gaze Guard

    Instant Privacy & Screen Blur

  8. Chat

    turn your backend into a chat app instantly

  9. Room Service

    The Mac cleaner built for developers

  10. Visdiff

    Stop bridging the design-to-code gap, close it

  11. Google AI Studio 2.0

    Full-stack vibe coding powered by Antigravity + Firebase

  12. AdsTurbo

    Create ads with AI actors that look truly human

  13. Built for Devs

    See how developers really experience your product

  14. GitAgent by Lyzr

    Your repository becomes your agent

  15. Assembly 2.0

    Build modern client portals for service businesses

Hugging Face(15)

  1. Generation Models Know Space: Unleashing Implicit 3D Priors for Scene Understanding

    While Multimodal Large Language Models demonstrate impressive semantic capabilities, they often suffer from spatial blindness, struggling with fine-grained geometric reasoning and physical dynamics. Existing solutions typically rely on explicit 3D modalities or complex geometric scaffolding, which are limited by data scarcity and generalization challenges. In this work, we propose a paradigm shift by leveraging the implicit spatial prior within large-scale video generation models. We posit that to synthesize temporally coherent videos, these models inherently learn robust 3D structural priors and physical laws. We introduce VEGA-3D (Video Extracted Generative Awareness), a plug-and-play framework that repurposes a pre-trained video diffusion model as a Latent World Simulator. By extracting spatiotemporal features from intermediate noise levels and integrating them with semantic representations via a token-level adaptive gated fusion mechanism, we enrich MLLMs with dense geometric cues without explicit 3D supervision. Extensive experiments across 3D scene understanding, spatial reasoning, and embodied manipulation benchmarks demonstrate that our method outperforms state-of-the-art baselines, validating that generative priors provide a scalable foundation for physical-world understanding. Code is publicly available at https://github.com/H-EmbodVis/VEGA-3D.

  2. SAMA: Factorized Semantic Anchoring and Motion Alignment for Instruction-Guided Video Editing

    Current instruction-guided video editing models struggle to simultaneously balance precise semantic modifications with faithful motion preservation. While existing approaches rely on injecting explicit external priors (e.g., VLM features or structural conditions) to mitigate these issues, this reliance severely bottlenecks model robustness and generalization. To overcome this limitation, we present SAMA (factorized Semantic Anchoring and Motion Alignment), a framework that factorizes video editing into semantic anchoring and motion modeling. First, we introduce Semantic Anchoring, which establishes a reliable visual anchor by jointly predicting semantic tokens and video latents at sparse anchor frames, enabling purely instruction-aware structural planning. Second, Motion Alignment pre-trains the same backbone on motion-centric video restoration pretext tasks (cube inpainting, speed perturbation, and tube shuffle), enabling the model to internalize temporal dynamics directly from raw videos. SAMA is optimized with a two-stage pipeline: a factorized pre-training stage that learns inherent semantic-motion representations without paired video-instruction editing data, followed by supervised fine-tuning on paired editing data. Remarkably, the factorized pre-training alone already yields strong zero-shot video editing ability, validating the proposed factorization. SAMA achieves state-of-the-art performance among open-source models and is competitive with leading commercial systems (e.g., Kling-Omni). Code, models, and datasets will be released.

  3. FASTER: Rethinking Real-Time Flow VLAs

    Real-time execution is crucial for deploying Vision-Language-Action (VLA) models in the physical world. Existing asynchronous inference methods primarily optimize trajectory smoothness, but neglect the critical latency in reacting to environmental changes. By rethinking the notion of reaction in action chunking policies, this paper presents a systematic analysis of the factors governing reaction time. We show that reaction time follows a uniform distribution determined jointly by the Time to First Action (TTFA) and the execution horizon. Moreover, we reveal that the standard practice of applying a constant schedule in flow-based VLAs can be inefficient and forces the system to complete all sampling steps before any movement can start, forming the bottleneck in reaction latency. To overcome this issue, we propose Fast Action Sampling for ImmediaTE Reaction (FASTER). By introducing a Horizon-Aware Schedule, FASTER adaptively prioritizes near-term actions during flow sampling, compressing the denoising of the immediate reaction by tenfold (e.g., in π_{0.5} and X-VLA) into a single step, while preserving the quality of long-horizon trajectory. Coupled with a streaming client-server pipeline, FASTER substantially reduces the effective reaction latency on real robots, especially when deployed on consumer-grade GPUs. Real-world experiments, including a highly dynamic table tennis task, prove that FASTER unlocks unprecedented real-time responsiveness for generalist policies, enabling rapid generation of accurate and smooth trajectories.

  4. 3DreamBooth: High-Fidelity 3D Subject-Driven Video Generation Model

    Creating dynamic, view-consistent videos of customized subjects is highly sought after for a wide range of emerging applications, including immersive VR/AR, virtual production, and next-generation e-commerce. However, despite rapid progress in subject-driven video generation, existing methods predominantly treat subjects as 2D entities, focusing on transferring identity through single-view visual features or textual prompts. Because real-world subjects are inherently 3D, applying these 2D-centric approaches to 3D object customization reveals a fundamental limitation: they lack the comprehensive spatial priors necessary to reconstruct the 3D geometry. Consequently, when synthesizing novel views, they must rely on generating plausible but arbitrary details for unseen regions, rather than preserving the true 3D identity. Achieving genuine 3D-aware customization remains challenging due to the scarcity of multi-view video datasets. While one might attempt to fine-tune models on limited video sequences, this often leads to temporal overfitting. To resolve these issues, we introduce a novel framework for 3D-aware video customization, comprising 3DreamBooth and 3Dapter. 3DreamBooth decouples spatial geometry from temporal motion through a 1-frame optimization paradigm. By restricting updates to spatial representations, it effectively bakes a robust 3D prior into the model without the need for exhaustive video-based training. To enhance fine-grained textures and accelerate convergence, we incorporate 3Dapter, a visual conditioning module. Following single-view pre-training, 3Dapter undergoes multi-view joint optimization with the main generation branch via an asymmetrical conditioning strategy. This design allows the module to act as a dynamic selective router, querying view-specific geometric hints from a minimal reference set. Project page: https://ko-lani.github.io/3DreamBooth/

  5. Bridging Semantic and Kinematic Conditions with Diffusion-based Discrete Motion Tokenizer

    Prior motion generation largely follows two paradigms: continuous diffusion models that excel at kinematic control, and discrete token-based generators that are effective for semantic conditioning. To combine their strengths, we propose a three-stage framework comprising condition feature extraction (Perception), discrete token generation (Planning), and diffusion-based motion synthesis (Control). Central to this framework is MoTok, a diffusion-based discrete motion tokenizer that decouples semantic abstraction from fine-grained reconstruction by delegating motion recovery to a diffusion decoder, enabling compact single-layer tokens while preserving motion fidelity. For kinematic conditions, coarse constraints guide token generation during planning, while fine-grained constraints are enforced during control through diffusion-based optimization. This design prevents kinematic details from disrupting semantic token planning. On HumanML3D, our method significantly improves controllability and fidelity over MaskControl while using only one-sixth of the tokens, reducing trajectory error from 0.72 cm to 0.08 cm and FID from 0.083 to 0.029. Unlike prior methods that degrade under stronger kinematic constraints, ours improves fidelity, reducing FID from 0.033 to 0.014.

  6. MonoArt: Progressive Structural Reasoning for Monocular Articulated 3D Reconstruction

    Reconstructing articulated 3D objects from a single image requires jointly inferring object geometry, part structure, and motion parameters from limited visual evidence. A key difficulty lies in the entanglement between motion cues and object structure, which makes direct articulation regression unstable. Existing methods address this challenge through multi-view supervision, retrieval-based assembly, or auxiliary video generation, often sacrificing scalability or efficiency. We present MonoArt, a unified framework grounded in progressive structural reasoning. Rather than predicting articulation directly from image features, MonoArt progressively transforms visual observations into canonical geometry, structured part representations, and motion-aware embeddings within a single architecture. This structured reasoning process enables stable and interpretable articulation inference without external motion templates or multi-stage pipelines. Extensive experiments on PartNet-Mobility demonstrate that OM achieves state-of-the-art performance in both reconstruction accuracy and inference speed. The framework further generalizes to robotic manipulation and articulated scene reconstruction.

  7. Cubic Discrete Diffusion: Discrete Visual Generation on High-Dimensional Representation Tokens

    Visual generation with discrete tokens has gained significant attention as it enables a unified token prediction paradigm shared with language models, promising seamless multimodal architectures. However, current discrete generation methods remain limited to low-dimensional latent tokens (typically 8-32 dims), sacrificing the semantic richness essential for understanding. While high-dimensional pretrained representations (768-1024 dims) could bridge this gap, their discrete generation poses fundamental challenges. In this paper, we present Cubic Discrete Diffusion (CubiD), the first discrete generation model for high-dimensional representations. CubiD performs fine-grained masking throughout the high-dimensional discrete representation -- any dimension at any position can be masked and predicted from partial observations. This enables the model to learn rich correlations both within and across spatial positions, with the number of generation steps fixed at T regardless of feature dimensionality, where T ll hwd. On ImageNet-256, CubiD achieves state-of-the-art discrete generation with strong scaling behavior from 900M to 3.7B parameters. Crucially, we validate that these discretized tokens preserve original representation capabilities, demonstrating that the same discrete tokens can effectively serve both understanding and generation tasks. We hope this work will inspire future research toward unified multimodal architectures. Code is available at: https://github.com/YuqingWang1029/CubiD.

  8. Nemotron-Cascade 2: Post-Training LLMs with Cascade RL and Multi-Domain On-Policy Distillation

    We introduce Nemotron-Cascade 2, an open 30B MoE model with 3B activated parameters that delivers best-in-class reasoning and strong agentic capabilities. Despite its compact size, its mathematical and coding reasoning performance approaches that of frontier open models. It is the second open-weight LLM, after DeepSeekV3.2-Speciale-671B-A37B, to achieve Gold Medal-level performance in the 2025 International Mathematical Olympiad (IMO), the International Olympiad in Informatics (IOI), and the ICPC World Finals, demonstrating remarkably high intelligence density with 20x fewer parameters. In contrast to Nemotron-Cascade 1, the key technical advancements are as follows. After SFT on a meticulously curated dataset, we substantially expand Cascade RL to cover a much broader spectrum of reasoning and agentic domains. Furthermore, we introduce multi-domain on-policy distillation from the strongest intermediate teacher models for each domain throughout the Cascade RL process, allowing us to efficiently recover benchmark regressions and sustain strong performance gains along the way. We release the collection of model checkpoint and training data.

  9. LVOmniBench: Pioneering Long Audio-Video Understanding Evaluation for Omnimodal LLMs

    Recent advancements in omnimodal large language models (OmniLLMs) have significantly improved the comprehension of audio and video inputs. However, current evaluations primarily focus on short audio and video clips ranging from 10 seconds to 5 minutes, failing to reflect the demands of real-world applications, where videos typically run for tens of minutes. To address this critical gap, we introduce LVOmniBench, a new benchmark designed specifically for the cross-modal comprehension of long-form audio and video. This dataset comprises high-quality videos sourced from open platforms that feature rich audio-visual dynamics. Through rigorous manual selection and annotation, LVOmniBench comprises 275 videos, ranging in duration from 10 to 90 minutes, and 1,014 question-answer (QA) pairs. LVOmniBench aims to rigorously evaluate the capabilities of OmniLLMs across domains, including long-term memory, temporal localization, fine-grained understanding, and multimodal perception. Our extensive evaluation reveals that current OmniLLMs encounter significant challenges when processing extended audio-visual inputs. Open-source models generally achieve accuracies below 35%, whereas the Gemini 3 Pro reaches a peak accuracy of approximately 65%. We anticipate that this dataset, along with our empirical findings, will stimulate further research and the development of advanced models capable of resolving complex cross-modal understanding problems within long-form audio-visual contexts.

  10. Memento-Skills: Let Agents Design Agents

    We introduce Memento-Skills, a generalist, continually-learnable LLM agent system that functions as an agent-designing agent: it autonomously constructs, adapts, and improves task-specific agents through experience. The system is built on a memory-based reinforcement learning framework with stateful prompts, where reusable skills (stored as structured markdown files) serve as persistent, evolving memory. These skills encode both behaviour and context, enabling the agent to carry forward knowledge across interactions. Starting from simple elementary skills (like Web search and terminal operations), the agent continually improves via the Read--Write Reflective Learning mechanism introduced in Memento~2~wang2025memento2. In the read phase, a behaviour-trainable skill router selects the most relevant skill conditioned on the current stateful prompt; in the write phase, the agent updates and expands its skill library based on new experience. This closed-loop design enables continual learning without updating LLM parameters, as all adaptation is realised through the evolution of externalised skills and prompts. Unlike prior approaches that rely on human-designed agents, Memento-Skills enables a generalist agent to design agents end-to-end for new tasks. Through iterative skill generation and refinement, the system progressively improves its own capabilities. Experiments on the General AI Assistants benchmark and Humanity's Last Exam demonstrate sustained gains, achieving 26.2\% and 116.2\% relative improvements in overall accuracy, respectively. Code is available at https://github.com/Memento-Teams/Memento-Skills.

  11. F2LLM-v2: Inclusive, Performant, and Efficient Embeddings for a Multilingual World

    We present F2LLM-v2, a new family of general-purpose, multilingual embedding models in 8 distinct sizes ranging from 80M to 14B. Trained on a newly curated composite of 60 million publicly available high-quality data samples, F2LLM-v2 supports more than 200 languages, with a particular emphasis on previously underserved mid- and low-resource languages. By integrating a two-stage LLM-based embedding training pipeline with matryoshka learning, model pruning, and knowledge distillation techniques, we present models that are far more efficient than previous LLM-based embedding models while retaining competitive performances. Extensive evaluations confirm that F2LLM-v2-14B ranks first on 11 MTEB benchmarks, while the smaller models in the family also set a new state of the art for resource-constrained applications. To facilitate open-source embedding model research, we release all models, data, code, and intermediate checkpoints.

  12. ReactMotion: Generating Reactive Listener Motions from Speaker Utterance

    In this paper, we introduce a new task, Reactive Listener Motion Generation from Speaker Utterance, which aims to generate naturalistic listener body motions that appropriately respond to a speaker's utterance. However, modeling such nonverbal listener behaviors remains underexplored and challenging due to the inherently non-deterministic nature of human reactions. To facilitate this task, we present ReactMotionNet, a large-scale dataset that pairs speaker utterances with multiple candidate listener motions annotated with varying degrees of appropriateness. This dataset design explicitly captures the one-to-many nature of listener behavior and provides supervision beyond a single ground-truth motion. Building on this dataset design, we develop preference-oriented evaluation protocols tailored to evaluate reactive appropriateness, where conventional motion metrics focusing on input-motion alignment ignore. We further propose ReactMotion, a unified generative framework that jointly models text, audio, emotion, and motion, and is trained with preference-based objectives to encourage both appropriate and diverse listener responses. Extensive experiments show that ReactMotion outperforms retrieval baselines and cascaded LLM-based pipelines, generating more natural, diverse, and appropriate listener motions.

  13. AndroTMem: From Interaction Trajectories to Anchored Memory in Long-Horizon GUI Agents

    Long-horizon GUI agents are a key step toward real-world deployment, yet effective interaction memory under prevailing paradigms remains under-explored. Replaying full interaction sequences is redundant and amplifies noise, while summaries often erase dependency-critical information and traceability. We present AndroTMem, a diagnostic framework for anchored memory in long-horizon Android GUI agents. Its core benchmark, AndroTMem-Bench, comprises 1,069 tasks with 34,473 interaction steps (avg. 32.1 per task, max. 65). We evaluate agents with TCR (Task Complete Rate), focusing on tasks whose completion requires carrying forward critical intermediate state; AndroTMem-Bench is designed to enforce strong step-to-step causal dependencies, making sparse yet essential intermediate states decisive for downstream actions and centering interaction memory in evaluation. Across open- and closed-source GUI agents, we observe a consistent pattern: as interaction sequences grow longer, performance drops are driven mainly by within-task memory failures, not isolated perception errors or local action mistakes. Guided by this diagnosis, we propose Anchored State Memory (ASM), which represents interaction sequences as a compact set of causally linked intermediate-state anchors to enable subgoal-targeted retrieval and attribution-aware decision making. Across multiple settings and 12 evaluated GUI agents, ASM consistently outperforms full-sequence replay and summary-based baselines, improving TCR by 5%-30.16% and AMS by 4.93%-24.66%, indicating that anchored, structured memory effectively mitigates the interaction-memory bottleneck in long-horizon GUI tasks. The code, benchmark, and related resources are publicly available at [https://github.com/CVC2233/AndroTMem](https://github.com/CVC2233/AndroTMem).

  14. Cognitive Mismatch in Multimodal Large Language Models for Discrete Symbol Understanding

    While Multimodal Large Language Models (MLLMs) have achieved remarkable success in interpreting natural scenes, their ability to process discrete symbols -- the fundamental building blocks of human cognition -- remains a critical open question. Unlike continuous visual data, symbols such as mathematical formulas, chemical structures, and linguistic characters require precise, deeper interpretation. This paper introduces a comprehensive benchmark to evaluate how top-tier MLLMs navigate these "discrete semantic spaces" across five domains: language, culture, mathematics, physics, and chemistry. Our investigation uncovers a counterintuitive phenomenon: models often fail at basic symbol recognition yet succeed in complex reasoning tasks, suggesting they rely on linguistic probability rather than true visual perception. By exposing this "cognitive mismatch", we highlight a significant gap in current AI capabilities: the struggle to truly perceive and understand the symbolic languages that underpin scientific discovery and abstract thought. This work offers a roadmap for developing more rigorous, human-aligned intelligent systems.

  15. EffectErase: Joint Video Object Removal and Insertion for High-Quality Effect Erasing

    Video object removal aims to eliminate dynamic target objects and their visual effects, such as deformation, shadows, and reflections, while restoring seamless backgrounds. Recent diffusion-based video inpainting and object removal methods can remove the objects but often struggle to erase these effects and to synthesize coherent backgrounds. Beyond method limitations, progress is further hampered by the lack of a comprehensive dataset that systematically captures common object effects across varied environments for training and evaluation. To address this, we introduce VOR (Video Object Removal), a large-scale dataset that provides diverse paired videos, each consisting of one video where the target object is present with its effects and a counterpart where the object and effects are absent, with corresponding object masks. VOR contains 60K high-quality video pairs from captured and synthetic sources, covers five effects types, and spans a wide range of object categories as well as complex, dynamic multi-object scenes. Building on VOR, we propose EffectErase, an effect-aware video object removal method that treats video object insertion as the inverse auxiliary task within a reciprocal learning scheme. The model includes task-aware region guidance that focuses learning on affected areas and enables flexible task switching. Then, an insertion-removal consistency objective that encourages complementary behaviors and shared localization of effect regions and structural cues. Trained on VOR, EffectErase achieves superior performance in extensive experiments, delivering high-quality video object effect erasing across diverse scenarios.

Techmeme(15)

  1. A US jury finds Elon Musk intentionally misled Twitter shareholders by disparaging the company in 2022 to buy it for a lower price than his original $44B bid (Bloomberg)

    Bloomberg : A US jury finds Elon Musk intentionally misled Twitter shareholders by disparaging the company in 2022 to buy it for a lower price than his original $44B bid —  Elon Musk defrauded Twitter Inc. investors when he disparaged the company in 2022 in an effort to buy the social media platform …

  2. Moonshot says Kimi K2.5 provides "the foundation" for Cursor's Composer 2 model and that Cursor accesses Kimi K2.5 via Fireworks AI (@kimi_moonshot)

    @kimi_moonshot : Moonshot says Kimi K2.5 provides “the foundation” for Cursor's Composer 2 model and that Cursor accesses Kimi K2.5 via Fireworks AI —  Congrats to the @cursor_ai team on the launch of Composer 2! We are proud to see Kimi-k2.5 provide the foundation. Seeing our model integrated effectively through Cursor's continued pretraining & high-compute RL training is the open model ecosystem we love to support.

  3. Super Micro says co-founder Yih-Shyan Liaw has resigned from its board after US prosecutors indicted him on allegations of smuggling Nvidia AI chips to China (Jordan Novet/CNBC)

    Jordan Novet / CNBC : Super Micro says co-founder Yih-Shyan Liaw has resigned from its board after US prosecutors indicted him on allegations of smuggling Nvidia AI chips to China —  Super Micro Computer said Friday that Yih-Shyan “Wally” Liaw, a co-founder, has resigned from the server maker's board …

  4. Anthropic launches Claude Code channels, which let users interact with a Claude Code session through Telegram and Discord (Marcus Schuler/Implicator.ai)

    Marcus Schuler / Implicator.ai : Anthropic launches Claude Code channels, which let users interact with a Claude Code session through Telegram and Discord —  Anthropic released Claude Code Channels, a research-preview feature that lets developers send messages to a running Claude Code session from Telegram and Discord.

  5. Jensen Huang proposes a compensation model where engineers receive an AI token budget on top of their base salary, to deploy agents as productivity multipliers (Anniek Bao/CNBC)

    Anniek Bao / CNBC : Jensen Huang proposes a compensation model where engineers receive an AI token budget on top of their base salary, to deploy agents as productivity multipliers —  The perks of working in Silicon Valley have long included high salaries.  Now, some engineers may be offered a new incentive: artificial intelligence tokens.

  6. Pinterest CEO Bill Ready calls on governments to ban social media for users under 16, says social platforms gave "insufficient forethought" about consequences (Bill Ready/Time)

    Bill Ready / Time : Pinterest CEO Bill Ready calls on governments to ban social media for users under 16, says social platforms gave “insufficient forethought” about consequences —  When Pinterest removed social features for teens and made every account under 16 private—meaning no discoverability …

  7. Microsoft acknowledges complaints about Windows 11, promising a reduction of "unnecessary" Copilot integrations, more control over updates, and more (Ed Bott/ZDNET)

    Ed Bott / ZDNET : Microsoft acknowledges complaints about Windows 11, promising a reduction of “unnecessary” Copilot integrations, more control over updates, and more —  ZDNET's key takeaways  — Microsoft finally acknowledged complaints about Windows 11.  — The company is promising sweeping changes to a slew of features.

  8. Sources: contract electronics manufacturer Zetwerk plans to file for an IPO in India, aiming to raise ~$550M at a ~$4B valuation; it was valued at $3.1B in 2024 (Reuters)

    Reuters : Sources: contract electronics manufacturer Zetwerk plans to file for an IPO in India, aiming to raise ~$550M at a ~$4B valuation; it was valued at $3.1B in 2024 —  India's Zetwerk is preparing to confidentially file draft papers for an initial public offering within the next one to two weeks …

  9. Memo: Roblox is planning to take a share of revenue from sponsorships in its games and is overhauling advertising policies beginning May 4 (Cecilia D'Anastasio/Bloomberg)

    Cecilia D'Anastasio / Bloomberg : Memo: Roblox is planning to take a share of revenue from sponsorships in its games and is overhauling advertising policies beginning May 4 —  Roblox Corp. will take a share of revenue from sponsorships in its games as part of a major overhaul of its advertising policies.

  10. A judge issues an order requiring Kalshi to temporarily halt sports and election contracts in Nevada, the first US state to force Kalshi to cease operations (Kate Knibbs/Wired)

    Kate Knibbs / Wired : A judge issues an order requiring Kalshi to temporarily halt sports and election contracts in Nevada, the first US state to force Kalshi to cease operations —  A judge ordered Kalshi to immediately halt sports and election contracts in the state, intensifying a growing regulatory battle over prediction markets.

  11. WordPress.com says it will now allow AI agents to draft, edit, and publish content on customers' websites, as well as manage comments, update metadata, and more (Sarah Perez/TechCrunch)

    Sarah Perez / TechCrunch : WordPress.com says it will now allow AI agents to draft, edit, and publish content on customers' websites, as well as manage comments, update metadata, and more —  Web hosting platform WordPress.com is embracing AI agents, a decision that could change the look and feel of the web.

  12. OpenAI plans "an autonomous AI research intern" by September and says its "North Star" is to build a fully automated multi-agent research system by 2028 (Will Douglas Heaven/MIT Technology Review)

    Will Douglas Heaven / MIT Technology Review : OpenAI plans “an autonomous AI research intern” by September and says its “North Star” is to build a fully automated multi-agent research system by 2028 —  EXECUTIVE SUMMARY  —  OpenAI is refocusing its research efforts and throwing its resources into a new grand challenge.

  13. A new anonymous Substack alleges AI compliance startup Delve "faked" compliance for startups by generating pre-populated audit reports and fabricating evidence (DeepDelver)

    DeepDelver : A new anonymous Substack alleges AI compliance startup Delve “faked” compliance for startups by generating pre-populated audit reports and fabricating evidence —  How Delve managed to falsely convince hundreds of customers they were compliant and then lied about it when exposed and called out

  14. Google is running a "small" experiment replacing news headlines in search results with AI-generated ones, after adding the feature in Google Discover in January (Sean Hollister/The Verge)

    Sean Hollister / The Verge : Google is running a “small” experiment replacing news headlines in search results with AI-generated ones, after adding the feature in Google Discover in January —  We're seeing Verge headlines rewritten by Google AI. … Since roughly the turn of the millennium, Google Search has been the bedrock of the web.

  15. Mistral's CEO proposes a revenue-based levy for AI model providers in the EU, to be invested "in new content creation and supporting Europe's cultural sectors" (Arthur Mensch/Financial Times)

    Arthur Mensch / Financial Times : Mistral's CEO proposes a revenue-based levy for AI model providers in the EU, to be invested “in new content creation and supporting Europe's cultural sectors” —  A revenue-based charge would protect the livelihoods of copyright holders and bring legal certainty

Solidot(15)

  1. 北美何时开始使用弓箭

    根据发表在《PNAS Nexus》上的一项考古学研究,北美居民大约是在 1400 年前开始用弓箭取代飞镖和投矛器。在南部区域弓箭几乎是立即被接受,北部区域的接受度比较慢,一开始是将弓箭作为现有工具的补充,花了千年时间才淘汰飞镖和投矛器。弓箭在精度、射程、速度、射击频率等方面都强于飞镖和投矛器,但制造和维护成本更高,且需要双手操作,无法同时持盾,但其优点大于缺点而获得广泛使用。弓箭由有机材料制成,没有石器、骨器或金属工具那样容易保存,因此确定其出现时间和流行时间比较困难。

  2. 《二重螺旋》两次通过更新向玩家传播恶意程序

    免费抽卡游戏《二重螺旋(Duet Night Abyss)》开发商英雄游戏为 3 月 18 日发生的“网络安全事故”道歉,攻击者利用游戏启动器的更新向用户传播了窃取信息的恶意程序 Trojan:MSIL/UmbralStealer.DG!MTB,该恶意程序主要被用于窃取密码和加密货币。包含恶意程序的更新是在 3 月 18 日 7:39 am UTC 在 Steam 上释出的。 这不是《二重螺旋》最近几个月第一次遭遇安全事故,二月底它发生过类似的事故,攻击者主要引导用户去玩《原神》,恶意程度略低,可能主要是发出警告。

  3. 卫星照片显示澳大利亚的红色沙漠在大雨之后变成绿色

    位于澳洲地理中心的小镇爱丽丝泉(Alice Springs),因其锈红色的沙漠景观而常被称为红土中心(Red Centre)。然而在 2026 年 2 月至 3 月连续数周的大雨过后,沙漠披上了一层绿装。NASA Terra 卫星显示,原本因富含铁质的岩石氧化而呈现红褐色的地貌,已完全被新生的植被覆盖。根据澳洲气象局的数据,北领地在 2026 年 2 月的平均降雨量高达 239 毫米,创下自 1900 年有纪录以来排名第三的湿润二月。这场地貌转变的背后也伴随着严重的地面洪涝灾害,洪水不仅连根拔起了树木,更造成许多民众受困。

  4. Meta 撤回了 VR 版 Horizo​​n Worlds 的死亡判决

    Meta 撤回了 VR 版 Horizo​​n Worlds 的死亡判决。元宇宙或许已死,但 Meta 还不打算完全放弃。Meta CTO Andrew Bosworth 通过其 Instagram 账号宣布了这一消息,他表示该公司在收到用户反馈之后决定继续为现有游戏提供 Horizo​​n Worlds 的 VR 支持,但该公司的开发团队将专注于 Horizo​​n Worlds 的移动版本而不是 VR 版本。Meta 此前裁掉了负责 VR 的 Reality Labs 部门千名员工,但它仍然还有数千名员工,将继续开发新的 VR 头显,以及 AR 眼镜及相关技术。

  5. 人类与动物在声音偏好上存在共性

    人类与其他动物是否共享对声音之美的感知?发表在《科学》期刊上的一项新研究表明,答案或许是肯定的。在整个动物界中,动物会通过发声来交流及吸引配偶。尽管物种内部的求偶叫声和鸣唱各不相同,但倾听者往往会偏好其中某些特定的变体。这些偏好可能源于固有的感觉偏好、演化压力,或两者的结合。由于感官系统的基本架构在各物种间具有广泛的共性,因此那些旨在吸引同类的声音——例如悦耳的鸟鸣——也可能对其他物种(包括人类)产生吸引力;达尔文将这一理论称为“对美的欣赏力”。然而关于人类是否与其他动物在声音审美偏好上存在相似性的这一观念并未得到严格测试。研究人员开展了一项全球性的公民科学实验,共有 4196 名参与者评估了录自 16 个物种的 110 对动物的声音。在每一对动物的声音中,此前的研究已经确定了某动物本身偏爱哪一种声音。参与者需从每对声音中选出自己更喜欢的那个声音,从而使研究人员能比较人类与动物的声音偏好。结果显示,人类与广泛的动物类群——包括昆虫、蛙类、鸟类以及其他哺乳动物——在声音偏好上存在某些共性。

  6. 沃尔玛获得利用机器学习预测需求并自动定价的专利

    零售巨头沃尔玛在今年 1 月获得了一项关于“动态自动更新商品价格的系统和方法”的专利,上周它又获得了一项利用机器学习预测需求并推荐商品价格的专利,再次引发了它可能会引入动态定价的争议。沃尔玛业务遍及世界,该公司否认这些专利与动态定价相关,称一月份的专利针对的是降价促销,而上周的专利则是旨在帮助销售团队做出决策。

  7. Google 分享如何安装未验证身份开发者应用的流程

    Google 从今年 9 月开始将强制性要求验证所有 Android 应用开发者的身份,未经身份验证的开发者的应用将无法在用户 Android 设备上侧载(sideload)。此举在 Android 社区引发了强烈抗议,迫使 Google 软化立场,表示会引入一个流程让用户能在设备上安装未验证身份的开发者应用。现在它分享了整个流程细节: 在“关于本机”页面连续点击软件版本号七次开启开发者选项; 在“设置”>“系统”中打开“开发者选项”,向下滚动找到“允许未验证的软件包(Allow Unverified Packages)”; 打开开关点击确认您并非是被迫的; 输入解锁设备的 Pin 码/密码; 重启设备; 等待 24 小时; 返回“允许未验证的软件包”选项,忽略警告,选择“暂时允许”(七天)或“永久允许”; 勾选方框确认您了解相关风险。 现在用户可以安装未验证身份开发者的应用了。 Android 生态系统总裁 Sameer Samat 解释说设置延迟 24 小时生效的限制是为了应对日益猖獗的高压社交工程攻击,在此类攻击中,骗子会说服受害者必须立即安装某个应用以避免严重后果。

  8. 新能源危机迫使政府重新考虑对化石燃料的依赖

    正在爆发的新能源危机正促使全球决策者重新思考对化石燃料的依赖,提议扩大核能和可再生能源,推动能源供应来源多元化。欧洲上周公布了对核能的新金融担保,中国发改委的一个部门表示要加快可再生能源转型,多个亚洲国家建议居民在家办公节约能源,日本一直在讨论重启闲置的核反应堆。中国因拥有充足的应急石油储备和较高的电气化率而受到的影响相对较小,电动汽车占中国国内新车销量的半数以上,电网中逾 50% 的电力来自可再生能源。

  9. 内存条和传统 DIMM 插槽可能将消失

    JEDEC 和主要内存制造商美光、SK 海力士和三星正在共同制定下一代内存标准 DDR6,我们所熟悉的内存条和传统 DIMM 插槽可能将消失。DDR5 将传统的 64-bit 通道拆分为两个 32 位子通道,而 DDR6 将会进一步分为 4 个 24 位子通道以降低每个通道的电负载。DDR6 的数据传输带宽也将进一步提高至 17,600 MT/s。新一代内存模块 CAMM2 (Compression Attached Memory Module)的外形也将摒弃传统的垂直 DIMM 插槽,采用低矮的螺栓锁住设计,缩短 CPU 和内存芯片之间的距离。CAMM2 很可能成为 DIMM 标准的接替者。

  10. 社媒使用降低个人幸福感

    牛津大学幸福研究中心、盖洛普和联合国等机构发布了 2026 年世界幸福报告。报告是基于对 140 个国家约 10 万人的调查。报告指出,重度使用社交媒体与年轻人幸福感下降有关,尤其是在英语国家和西欧国家的少女群体中。过度使用社交媒体,尤其是每天使用社交媒体超过七小时,与较低的幸福感密切相关,而算法、以图片为中心的平台以及网红内容是造成这种现象的关键因素。大多数美国大学生希望社交媒体平台不存在:“他们使用社交媒体是因为其他人都在用,但他们更希望没有人使用社交媒体。”芬兰连续第九年蝉联全球最幸福国家榜首,前十包括芬兰、冰岛、丹麦、哥斯达黎加、瑞典、挪威、荷兰、以色列、卢森堡、瑞士。东亚三国日本(55 名)、韩国(58名)和中国(68名)。

  11. 居家办公有助于提高生育率

    斯坦福、普林斯顿、伦敦国王学院的研究人员调查了居家办公和生育率的关系。研究是基于 2024 年 11 月到 2025 年 2 月之间收集的 38 国 Global Survey of Working Arrangements(G-SWA)调查数据,以及 2022 年 12 月到 2025 年 12 月间收集的 U.S. Survey of Working Arrangements and Attitudes (SWAA)美国调查数据。研究显示居家办公有助于提高生育率。在 38 国样本中,如果伴侣双方每周至少一天在家工作,则每位女性的终生生育率会增加 0.32 个孩子;美国样本中则会增加 0.45 个孩子。

  12. Firefox v149 将内置 VPN

    Mozilla 宣布,3 月 24 日释出的 Firefox v149 将内置 VPN 服务。该服务将首先提供给美国、法国、德国和英国的用户,免费套餐的流量为每月 50GB。Firefox VPN 与 Mozilla VPN 不同,Mozilla VPN 是一项独立的付费服务,可同时在五台设备上使用;而 Firefox VPN 只限于浏览器本身,设计通过 Mozilla 管理的服务器路由流量,隐藏用户的真实 IP 地址。Mozilla 此前强调,不会出售个人数据,同步浏览数据如历史和书签将依赖于端到端加密。

  13. 美国私人太空公司计划捕捉小行星

    美国私人太空公司 TransAstra 周三宣布了名为“New Moon”的任务,该任务旨在捕获一颗房屋大小、质量 100 吨左右的小行星,将其转移到地球附近的 L2 点。TransAstra CEO Joel Sercel 表示设想将小行星变成一个材料加工和制造的机器人研发基地,长远目标是无需在地面制造太空飞行器硬件,无需从地球发射推进剂,直接利用太空原材料提取推进剂。他表示如果资金到位,New Moon 任务最早在 2028 年或 2029 年发射太空“捕获袋”与一颗小行星交会。

  14. GNOME 50 释出

    桌面环境项目 GNOME 释出了 v50。该版本被命名为“东京(Tokyo)”,以表彰 GNOME.Asia Summit 2025 当地组织者的工作。GNOME 50 的主要特性包括:改进了家长控制,家长和监护人能监控儿童的屏幕使用时间,为儿童账户设置就寝自动锁屏限制;增强 Orca 屏幕阅读器;改进文件管理器的 UI 和性能;Wayland 会话支持鼠标预览;新的 Reduced Motion 选项减少动画造成的任何不适或干扰;文档查看器支持添加文本注释、添加线条和高亮显示;等等,更多可浏览发布公告。

  15. 2026 年图灵奖授予了两位量子信息理论的奠基人

    2026 年图灵奖授予了量子信息理论的两位奠基人 Charles Henry Bennett 和 Gilles Brassard,两人共享 100 万美元奖金。两人是在 1979 年出席一学术会议期间抽空去游泳时相遇的,他们讨论了利用量子机制制造永远也无法伪造的货币。他们的合作推动了量子密码学的诞生,1984 年他们提出了第一个实用的量子密码协议 BB84,他们的论文《Quantum Cryptography: Public Key Distribution and Coin Tossing》证明,即使面临一个拥有无限计算能力和尖端技术如量子计算机的对手,通信双方仍然能建立一个由物理定律确保安全的加密密钥。BB84 依赖于量子信息的一个基本属性:它无法在不被干扰的情况下被复制或测量。任何窃听尝试都会在信息泄露前留下可检测的痕迹。