DIGEST · 2026-03-11

OrangeBot.AI Digest — 2026-03-11

84 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. The dead Internet is not a theory anymore (www.adriankrebs.ch)
  2. Don't post generated/AI-edited comments. HN is for conversation between humans.
  3. Show HN: I built a tool that watches webpages and exposes changes as RSS (sitespy.app)
  4. The MacBook Neo (daringfireball.net)
  5. Google closes deal to acquire Wiz (www.wiz.io)
  6. Making WebAssembly a first-class language on the Web (hacks.mozilla.org)
  7. Temporal: A nine-year journey to fix time in JavaScript (bloomberg.github.io)
  8. Faster asin() was hiding in plain sight (16bpp.net)
  9. Entities enabling scientific fraud at scale (2025) (doi.org)
  10. Whistleblower claims ex-DOGE member says he took Social Security data to new job (www.washingtonpost.com)
  11. How we hacked McKinsey's AI platform (codewall.ai)
  12. Lego's 0.002mm specification and its implications for manufacturing (2025) (www.thewave.engineer)
  13. Swiss e-voting pilot can't count 2,048 ballots after decryption failure (www.theregister.com)
  14. BitNet: 100B Param 1-Bit model for local CPUs (github.com)
  15. Building a TB-303 from Scratch (loopmaster.xyz)

GitHub Trending(9)

  1. msitarzewski / agency-agents
  2. 666ghj / MiroFish
  3. promptfoo / promptfoo
  4. obra / superpowers
  5. fishaudio / fish-speech
  6. virattt / ai-hedge-fund
  7. alibaba / page-agent
  8. NousResearch / hermes-agent
  9. AstrBotDevs / AstrBot

Product Hunt(15)

  1. Fort

    Fort tracks strength for people who care about longevity.

  2. InsForge

    Give agents everything they need to ship fullstack apps

  3. ScreenGeany AI

    Ask AI about anything on your screen with a single hotkey

  4. CodeYam CLI & Memory

    Comprehensive memory management for Claude Code

  5. MorphMind: A Steerable AI Platform

    Build a team of AI specialists that deliver quality work

  6. Teract AI

    Your AI reputation coach for LinkedIn, X, Reddit & more

  7. Gemini Embedding 2

    Google's first natively multimodal embedding model

  8. Typinator 10

    The fast and private text expander for macOS and iOS

  9. ELVES

    Summon your army of AI agents

  10. Mindspase

    A visual AI knowledge base that organizes what you save

  11. Product Workbench for Claude Code

    Turn feature ideas into stakeholder-ready code prototypes

  12. IonRouter

    Serve Any AI Model, Faster & Cheaper

  13. Cardboard

    Cursor for video editing

  14. Nativeline AI + Cloud

    Native Swift apps + a real cloud database. One prompt away.

  15. Firecrawl CLI

    The complete web data toolkit for AI agents

Hugging Face(15)

  1. Geometry-Guided Reinforcement Learning for Multi-view Consistent 3D Scene Editing

    Leveraging the priors of 2D diffusion models for 3D editing has emerged as a promising paradigm. However, maintaining multi-view consistency in edited results remains challenging, and the extreme scarcity of 3D-consistent editing paired data renders supervised fine-tuning (SFT), the most effective training strategy for editing tasks, infeasible. In this paper, we observe that, while generating multi-view consistent 3D content is highly challenging, verifying 3D consistency is tractable, naturally positioning reinforcement learning (RL) as a feasible solution. Motivated by this, we propose RL3DEdit, a single-pass framework driven by RL optimization with novel rewards derived from the 3D foundation model, VGGT. Specifically, we leverage VGGT's robust priors learned from massive real-world data, feed the edited images, and utilize the output confidence maps and pose estimation errors as reward signals, effectively anchoring the 2D editing priors onto a 3D-consistent manifold via RL. Extensive experiments demonstrate that RL3DEdit achieves stable multi-view consistency and outperforms state-of-the-art methods in editing quality with high efficiency. To promote the development of 3D editing, we will release the code and model.

  2. Omni-Diffusion: Unified Multimodal Understanding and Generation with Masked Discrete Diffusion

    While recent multimodal large language models (MLLMs) have made impressive strides, they predominantly employ a conventional autoregressive architecture as their backbone, leaving significant room to explore effective and efficient alternatives in architectural design. Concurrently, recent studies have successfully applied discrete diffusion models to various domains, such as visual understanding and image generation, revealing their considerable potential as a promising backbone for multimodal systems. Drawing inspiration from these pioneering research, we introduce Omni-Diffusion, the first any-to-any multimodal language model built entirely on mask-based discrete diffusion models, which unifies understanding and generation across text, speech, and images. Omni-Diffusion employs a unified mask-based discrete diffusion model to directly capture the joint distribution over discrete multimodal tokens. This approach supports not only bimodal tasks but also more complex scenarios involving multiple modalities. On a diverse set of benchmarks, our method outperforms or performs on par with existing multimodal systems that process two or more modalities, highlighting the significant promise of diffusion models in powering the next generation of multimodal foundation models. Project webpage: https://omni-diffusion.github.io.

  3. Thinking to Recall: How Reasoning Unlocks Parametric Knowledge in LLMs

    While reasoning in LLMs plays a natural role in math, code generation, and multi-hop factual questions, its effect on simple, single-hop factual questions remains unclear. Such questions do not require step-by-step logical decomposition, making the utility of reasoning highly counterintuitive. Nevertheless, we find that enabling reasoning substantially expands the capability boundary of the model's parametric knowledge recall, unlocking correct answers that are otherwise effectively unreachable. Why does reasoning aid parametric knowledge recall when there are no complex reasoning steps to be done? To answer this, we design a series of hypothesis-driven controlled experiments, and identify two key driving mechanisms: (1) a computational buffer effect, where the model uses the generated reasoning tokens to perform latent computation independent of their semantic content; and (2) factual priming, where generating topically related facts acts as a semantic bridge that facilitates correct answer retrieval. Importantly, this latter generative self-retrieval mechanism carries inherent risks: we demonstrate that hallucinating intermediate facts during reasoning increases the likelihood of hallucinations in the final answer. Finally, we show that our insights can be harnessed to directly improve model accuracy by prioritizing reasoning trajectories that contain hallucination-free factual statements.

  4. MM-Zero: Self-Evolving Multi-Model Vision Language Models From Zero Data

    Self-evolving has emerged as a key paradigm for improving foundational models such as Large Language Models (LLMs) and Vision Language Models (VLMs) with minimal human intervention. While recent approaches have demonstrated that LLM agents can self-evolve from scratch with little to no data, VLMs introduce an additional visual modality that typically requires at least some seed data, such as images, to bootstrap the self-evolution process. In this work, we present Multi-model Multimodal Zero (MM-Zero), the first RL-based framework to achieve zero-data self-evolution for VLM reasoning. Moving beyond prior dual-role (Proposer and Solver) setups, MM-Zero introduces a multi-role self-evolving training framework comprising three specialized roles: a Proposer that generates abstract visual concepts and formulates questions; a Coder that translates these concepts into executable code (e.g., Python, SVG) to render visual images; and a Solver that performs multimodal reasoning over the generated visual content. All three roles are initialized from the same base model and trained using Group Relative Policy Optimization (GRPO), with carefully designed reward mechanisms that integrate execution feedback, visual verification, and difficulty balancing. Our experiments show that MM-Zero improves VLM reasoning performance across a wide range of multimodal benchmarks. MM-Zero establishes a scalable path toward self-evolving multi-model systems for multimodal models, extending the frontier of self-improvement beyond the conventional two-model paradigm.

  5. InternVL-U: Democratizing Unified Multimodal Models for Understanding, Reasoning, Generation and Editing

    Unified multimodal models (UMMs) that integrate understanding, reasoning, generation, and editing face inherent trade-offs between maintaining strong semantic comprehension and acquiring powerful generation capabilities. In this report, we present InternVL-U, a lightweight 4B-parameter UMM that democratizes these capabilities within a unified framework. Guided by the principles of unified contextual modeling and modality-specific modular design with decoupled visual representations, InternVL-U integrates a state-of-the-art Multimodal Large Language Model (MLLM) with a specialized MMDiT-based visual generation head. To further bridge the gap between aesthetic generation and high-level intelligence, we construct a comprehensive data synthesis pipeline targeting high-semantic-density tasks, such as text rendering and scientific reasoning, under a reasoning-centric paradigm that leverages Chain-of-Thought (CoT) to better align abstract user intent with fine-grained visual generation details. Extensive experiments demonstrate that InternVL-U achieves a superior performance - efficiency balance. Despite using only 4B parameters, it consistently outperforms unified baseline models with over 3x larger scales such as BAGEL (14B) on various generation and editing tasks, while retaining strong multimodal understanding and reasoning capabilities.

  6. Stepping VLMs onto the Court: Benchmarking Spatial Intelligence in Sports

    Sports have long attracted broad attention as they push the limits of human physical and cognitive capabilities. Amid growing interest in spatial intelligence for vision-language models (VLMs), sports provide a natural testbed for understanding high-intensity human motion and dynamic object interactions. To this end, we present CourtSI, the first large-scale spatial intelligence dataset tailored to sports scenarios. CourtSI contains over 1M QA pairs, organized under a holistic taxonomy that systematically covers spatial counting, distance measurement, localization, and relational reasoning, across representative net sports including badminton, tennis, and table tennis. Leveraging well-defined court geometry as metric anchors, we develop a semi-automatic data engine to reconstruct sports scenes, enabling scalable curation of CourtSI. In addition, we introduce CourtSI-Bench, a high-quality evaluation benchmark comprising 3,686 QA pairs with rigorous human verification. We evaluate 25 proprietary and open-source VLMs on CourtSI-Bench, revealing a remaining human-AI performance gap and limited generalization from existing spatial intelligence benchmarks. These findings indicate that sports scenarios expose limitations in spatial intelligence capabilities captured by existing benchmarks. Further, fine-tuning Qwen3-VL-8B on CourtSI improves accuracy on CourtSI-Bench by 23.5 percentage points. The adapted model also generalizes effectively to CourtSI-Ext, an evaluation set built on a similar but unseen sport, and demonstrates enhanced spatial-aware commentary generation. Together, these findings demonstrate that CourtSI provides a scalable pathway toward advancing spatial intelligence of VLMs in sports.

  7. Reading, Not Thinking: Understanding and Bridging the Modality Gap When Text Becomes Pixels in Multimodal LLMs

    Multimodal large language models (MLLMs) can process text presented as images, yet they often perform worse than when the same content is provided as textual tokens. We systematically diagnose this "modality gap" by evaluating seven MLLMs across seven benchmarks in five input modes, spanning both synthetically rendered text and realistic document images from arXiv PDFs to Wikipedia pages. We find that the modality gap is task- and data-dependent. For example, math tasks degrade by over 60 points on synthetic renderings, while natural document images often match or exceed text-mode performance. Rendering choices such as font and resolution are strong confounds, with font alone swinging accuracy by up to 47 percentage points. To understand this, we conduct a grounded-theory error analysis of over 4,000 examples, revealing that image mode selectively amplifies reading errors (calculation and formatting failures) while leaving knowledge and reasoning errors largely unchanged, and that some models exhibit a chain-of-thought reasoning collapse under visual input. Motivated by these findings, we propose a self-distillation method that trains the model on its own pure text reasoning traces paired with image inputs, raising image-mode accuracy on GSM8K from 30.71% to 92.72% and transferring to unseen benchmarks without catastrophic forgetting. Overall, our study provides a systematic understanding of the modality gap and suggests a practical path toward improving visual text understanding in multimodal language models.

  8. VLM-SubtleBench: How Far Are VLMs from Human-Level Subtle Comparative Reasoning?

    The ability to distinguish subtle differences between visually similar images is essential for diverse domains such as industrial anomaly detection, medical imaging, and aerial surveillance. While comparative reasoning benchmarks for vision-language models (VLMs) have recently emerged, they primarily focus on images with large, salient differences and fail to capture the nuanced reasoning required for real-world applications. In this work, we introduce VLM-SubtleBench, a benchmark designed to evaluate VLMs on subtle comparative reasoning. Our benchmark covers ten difference types - Attribute, State, Emotion, Temporal, Spatial, Existence, Quantity, Quality, Viewpoint, and Action - and curate paired question-image sets reflecting these fine-grained variations. Unlike prior benchmarks restricted to natural image datasets, our benchmark spans diverse domains, including industrial, aerial, and medical imagery. Through extensive evaluation of both proprietary and open-source VLMs, we reveal systematic gaps between model and human performance across difference types and domains, and provide controlled analyses highlighting where VLMs' reasoning sharply deteriorates. Together, our benchmark and findings establish a foundation for advancing VLMs toward human-level comparative reasoning.

  9. MiniAppBench: Evaluating the Shift from Text to Interactive HTML Responses in LLM-Powered Assistants

    With the rapid advancement of Large Language Models (LLMs) in code generation, human-AI interaction is evolving from static text responses to dynamic, interactive HTML-based applications, which we term MiniApps. These applications require models to not only render visual interfaces but also construct customized interaction logic that adheres to real-world principles. However, existing benchmarks primarily focus on algorithmic correctness or static layout reconstruction, failing to capture the capabilities required for this new paradigm. To address this gap, we introduce MiniAppBench, the first comprehensive benchmark designed to evaluate principle-driven, interactive application generation. Sourced from a real-world application with 10M+ generations, MiniAppBench distills 500 tasks across six domains (e.g., Games, Science, and Tools). Furthermore, to tackle the challenge of evaluating open-ended interactions where no single ground truth exists, we propose MiniAppEval, an agentic evaluation framework. Leveraging browser automation, it performs human-like exploratory testing to systematically assess applications across three dimensions: Intention, Static, and Dynamic. Our experiments reveal that current LLMs still face significant challenges in generating high-quality MiniApps, while MiniAppEval demonstrates high alignment with human judgment, establishing a reliable standard for future research. Our code is available in github.com/MiniAppBench.

  10. Fish Audio S2 Technical Report

    We introduce Fish Audio S2, an open-sourced text-to-speech system featuring multi-speaker, multi-turn generation, and, most importantly, instruction-following control via natural-language descriptions. To scale training, we develop a multi-stage training recipe together with a staged data pipeline covering video captioning and speech captioning, voice-quality assessment, and reward modeling. To push the frontier of open-source TTS, we release our model weights, fine-tuning code, and an SGLang-based inference engine. The inference engine is production-ready for streaming, achieving an RTF of 0.195 and a time-to-first-audio below 100 ms.Our code and weights are available on GitHub (https://github.com/fishaudio/fish-speech) and Hugging Face (https://huggingface.co/fishaudio/s2-pro). We highly encourage readers to visit https://fish.audio to try custom voices.

  11. Are Audio-Language Models Listening? Audio-Specialist Heads for Adaptive Audio Steering

    Multimodal large language models can exhibit text dominance, over-relying on linguistic priors instead of grounding predictions in non-text inputs. One example is large audio-language models (LALMs) where decisive audio evidence can be under-utilized even when it contains important information. To address this issue we use mechanistic interpretability to identify a small set of audio-specialist attention heads whose audio attention yields a ``listening'' signal. We show that this signal increases when audio evidence affects the model's output, providing an indicator of audio engagement under standard prompting. Leveraging this localization, we construct an audio--silence steering direction and apply an inference-time activation intervention to the final representation, amplifying the model's audio effect. To demonstrate the utility of this intervention, we show on MMAU that this improves accuracy by up to +8.0 percentage points on two Qwen-based LALMs, without any parameter updates.

  12. Streaming Autoregressive Video Generation via Diagonal Distillation

    Large pretrained diffusion models have significantly enhanced the quality of generated videos, and yet their use in real-time streaming remains limited. Autoregressive models offer a natural framework for sequential frame synthesis but require heavy computation to achieve high fidelity. Diffusion distillation can compress these models into efficient few-step variants, but existing video distillation approaches largely adapt image-specific methods that neglect temporal dependencies. These techniques often excel in image generation but underperform in video synthesis, exhibiting reduced motion coherence, error accumulation over long sequences, and a latency-quality trade-off. We identify two factors that result in these limitations: insufficient utilization of temporal context during step reduction and implicit prediction of subsequent noise levels in next-chunk prediction (i.e., exposure bias). To address these issues, we propose Diagonal Distillation, which operates orthogonally to existing approaches and better exploits temporal information across both video chunks and denoising steps. Central to our approach is an asymmetric generation strategy: more steps early, fewer steps later. This design allows later chunks to inherit rich appearance information from thoroughly processed early chunks, while using partially denoised chunks as conditional inputs for subsequent synthesis. By aligning the implicit prediction of subsequent noise levels during chunk generation with the actual inference conditions, our approach mitigates error propagation and reduces oversaturation in long-range sequences. We further incorporate implicit optical flow modeling to preserve motion quality under strict step constraints. Our method generates a 5-second video in 2.61 seconds (up to 31 FPS), achieving a 277.3x speedup over the undistilled model.

  13. Decoupling Reasoning and Confidence: Resurrecting Calibration in Reinforcement Learning from Verifiable Rewards

    Reinforcement Learning from Verifiable Rewards (RLVR) significantly enhances large language models (LLMs) reasoning but severely suffers from calibration degeneration, where models become excessively over-confident in incorrect answers. Previous studies devote to directly incorporating calibration objective into existing optimization target. However, our theoretical analysis demonstrates that there exists a fundamental gradient conflict between the optimization for maximizing policy accuracy and minimizing calibration error. Building on this insight, we propose DCPO, a simple yet effective framework that systematically decouples reasoning and calibration objectives. Extensive experiments demonstrate that our DCPO not only preserves accuracy on par with GRPO but also achieves the best calibration performance and substantially mitigates the over-confidence issue. Our study provides valuable insights and practical solution for more reliable LLM deployment.

  14. The Reasoning Trap -- Logical Reasoning as a Mechanistic Pathway to Situational Awareness

    Situational awareness, the capacity of an AI system to recognize its own nature, understand its training and deployment context, and reason strategically about its circumstances, is widely considered among the most dangerous emergent capabilities in advanced AI systems. Separately, a growing research effort seeks to improve the logical reasoning capabilities of large language models (LLMs) across deduction, induction, and abduction. In this paper, we argue that these two research trajectories are on a collision course. We introduce the RAISE framework (Reasoning Advancing Into Self Examination), which identifies three mechanistic pathways through which improvements in logical reasoning enable progressively deeper levels of situational awareness: deductive self inference, inductive context recognition, and abductive self modeling. We formalize each pathway, construct an escalation ladder from basic self recognition to strategic deception, and demonstrate that every major research topic in LLM logical reasoning maps directly onto a specific amplifier of situational awareness. We further analyze why current safety measures are insufficient to prevent this escalation. We conclude by proposing concrete safeguards, including a "Mirror Test" benchmark and a Reasoning Safety Parity Principle, and pose an uncomfortable but necessary question to the logical reasoning community about its responsibility in this trajectory.

  15. BrandFusion: A Multi-Agent Framework for Seamless Brand Integration in Text-to-Video Generation

    The rapid advancement of text-to-video (T2V) models has revolutionized content creation, yet their commercial potential remains largely untapped. We introduce, for the first time, the task of seamless brand integration in T2V: automatically embedding advertiser brands into prompt-generated videos while preserving semantic fidelity to user intent. This task confronts three core challenges: maintaining prompt fidelity, ensuring brand recognizability, and achieving contextually natural integration. To address them, we propose BrandFusion, a novel multi-agent framework comprising two synergistic phases. In the offline phase (advertiser-facing), we construct a Brand Knowledge Base by probing model priors and adapting to novel brands via lightweight fine-tuning. In the online phase (user-facing), five agents jointly refine user prompts through iterative refinement, leveraging the shared knowledge base and real-time contextual tracking to ensure brand visibility and semantic alignment. Experiments on 18 established and 2 custom brands across multiple state-of-the-art T2V models demonstrate that BrandFusion significantly outperforms baselines in semantic preservation, brand recognizability, and integration naturalness. Human evaluations further confirm higher user satisfaction, establishing a practical pathway for sustainable T2V monetization.

Techmeme(15)

  1. Sources: PayPay priced its US IPO at $16 per share, below its targeted price range of between $17 and $20; the IPO raised $880M, valuing the company at $10.7B (Echo Wang/Reuters)

    Echo Wang / Reuters : Sources: PayPay priced its US IPO at $16 per share, below its targeted price range of between $17 and $20; the IPO raised $880M, valuing the company at $10.7B —  SoftBank Group-backed PayPay priced its U.S. initial public offering at $16 per share on Wednesday, below its targeted price range …

  2. Nvidia debuts Nemotron 3 Super, a 120B-parameter hybrid MoE open-weight model; filing: Nvidia plans to spend $26B over the next five years to build open models (Will Knight/Wired)

    Will Knight / Wired : Nvidia debuts Nemotron 3 Super, a 120B-parameter hybrid MoE open-weight model; filing: Nvidia plans to spend $26B over the next five years to build open models —  The move could position the AI infrastructure powerhouse to quickly compete with OpenAI, Anthropic, and DeepSeek.

  3. Microsoft says that Windows 11 Xbox mode, a controller-first, full-screen gaming interface, will begin rolling out in April and work across all PC form factors (Abhijith M B/Windows Latest)

    Abhijith M B / Windows Latest : Microsoft says that Windows 11 Xbox mode, a controller-first, full-screen gaming interface, will begin rolling out in April and work across all PC form factors —  Microsoft has confirmed that Xbox mode is coming to Windows 11 PCs, heralding one of the biggest shifts in the company's gaming strategy in years.

  4. Microsoft says the next Xbox, Project Helix, will have a custom AMD chip, and it will begin sending out "alpha versions" of Project Helix to developers in 2027 (The Verge)

    The Verge : Microsoft says the next Xbox, Project Helix, will have a custom AMD chip, and it will begin sending out “alpha versions” of Project Helix to developers in 2027 —  Microsoft announced new details about the console at the GDC Festival of Gaming.

  5. Perplexity announces Personal Computer, an OpenClaw-like AI agent that can run on a Mac, and an enterprise version of Perplexity Computer (Ina Fried/Axios)

    Ina Fried / Axios : Perplexity announces Personal Computer, an OpenClaw-like AI agent that can run on a Mac, and an enterprise version of Perplexity Computer —  - “Personal Computer” runs locally on a dedicated device.  — “Perplexity Computer,” which was announced a few weeks ago, operates in the cloud.

  6. Sources: Netflix will pay as much as $600M for InterPositive, Ben Affleck's AI moviemaking company, including bonuses for meeting certain performance targets (Lucas Shaw/Bloomberg)

    Lucas Shaw / Bloomberg : Sources: Netflix will pay as much as $600M for InterPositive, Ben Affleck's AI moviemaking company, including bonuses for meeting certain performance targets —  Netflix Inc. will pay as much as $600 million for InterPositive, the AI moviemaking company founded by Ben Affleck …

  7. Sources: the foldable iPhone will feature an iPad-like interface for its iPad mini-sized inner display and have an outer screen the size of a small iPhone (Mark Gurman/Bloomberg)

    Mark Gurman / Bloomberg : Sources: the foldable iPhone will feature an iPad-like interface for its iPad mini-sized inner display and have an outer screen the size of a small iPhone —  Apple Inc.'s forthcoming foldable iPhone will include updates to the iOS operating system that enable iPad-like layouts and side …

  8. FCC Chair Brendan Carr criticizes the slow pace of Amazon's satellite launches, after Amazon raised concerns about SpaceX's plan to launch up to 1M satellites (David Shepardson/Reuters)

    David Shepardson / Reuters : FCC Chair Brendan Carr criticizes the slow pace of Amazon's satellite launches, after Amazon raised concerns about SpaceX's plan to launch up to 1M satellites —  The chairman of the Federal Communications Commission dismissed criticism from Amazon.com (AMZN.O) of Elon Musk's SpaceX plan …

  9. Superhuman says it has disabled Grammarly's Expert Review feature, which gave editing suggestions "inspired by" real writers without permission, after backlash (Stevie Bonifield/The Verge)

    Stevie Bonifield / The Verge : Superhuman says it has disabled Grammarly's Expert Review feature, which gave editing suggestions “inspired by” real writers without permission, after backlash —  Now it will ‘reimagine’ the Expert Review feature, and allow experts a choice about participating in future AI plans.

  10. US medtech giant Stryker suffers a global outage after a cyberattack; staff and contractors say an Iran-linked hacking group's logo appeared on login pages (James Rundle/Wall Street Journal)

    James Rundle / Wall Street Journal : US medtech giant Stryker suffers a global outage after a cyberattack; staff and contractors say an Iran-linked hacking group's logo appeared on login pages —  Global outage affects U.S. medtech company; some staff devices remotely wiped  —  Medical technology giant Stryker is experiencing …

  11. Sources: Ripple kicks off a share buyback at a $50B valuation to repurchase up to $750M in shares from investors and employees in a tender offer through April (Bloomberg)

    Bloomberg : Sources: Ripple kicks off a share buyback at a $50B valuation to repurchase up to $750M in shares from investors and employees in a tender offer through April —  Ripple has kicked off a share buyback that would value the company at $50 billion, cementing it as one of the most valuable digital-asset firms …

  12. Anthropic says Claude for Excel and Claude for PowerPoint now share full context across open files, and skills are available inside Excel and PowerPoint add-ins (Sabrina Ortiz/The Deep View)

    Sabrina Ortiz / The Deep View : Anthropic says Claude for Excel and Claude for PowerPoint now share full context across open files, and skills are available inside Excel and PowerPoint add-ins —  W  —  hile Anthropic doesn't offer a productivity suite of apps like Google or Microsoft, it is expanding its integrations with Microsoft 365 applications.

  13. Amazon raised €14.5B in its euro bond market debut, the biggest ever corporate deal in the currency, following a dollar offering on Tuesday that raised $37B (Bloomberg)

    Bloomberg : Amazon raised €14.5B in its euro bond market debut, the biggest ever corporate deal in the currency, following a dollar offering on Tuesday that raised $37B —  Amazon.com Inc. raised €14.5 billion ($16.8 billion) in its euro bond market debut, the biggest ever corporate deal in the currency …

  14. WordPress debuts my.WordPress.net, which lets users run WordPress entirely in the browser to set up a private workspace not accessible from the public internet (Sarah Perez/TechCrunch)

    Sarah Perez / TechCrunch : WordPress debuts my.WordPress.net, which lets users run WordPress entirely in the browser to set up a private workspace not accessible from the public internet —  WordPress's publishing software can now run entirely in the web browser, the organization behind the open source publishing software announced on Wednesday.

  15. Sources: xAI's AI agent project Macrohard has stalled as Tesla ramps up its own AI agent project Digital Optimus; Elon Musk says it is a joint xAI-Tesla project (Grace Kay/Business Insider)

    Grace Kay / Business Insider : Sources: xAI's AI agent project Macrohard has stalled as Tesla ramps up its own AI agent project Digital Optimus; Elon Musk says it is a joint xAI-Tesla project —  Follow Grace Kay … - XAI's Macrohard project has stalled amid leadership changes and a data project pause.

Solidot(15)

  1. 为节省燃料亚洲多国下令公务员在家远程办公

    大多数亚洲国家依赖进口石油,战争导致霍尔木兹海峡关闭,海湾国家石油出口中断,整个地区面临燃料短缺。亚洲部分国家的政府已采取措施节约燃料。泰国政府要求非一线服务人员在家办公,要求将空调温度设定为 26°C,鼓励员工使用楼梯而非电梯。菲律宾政府下令公员工每周只需到办公室工作四天,要求尽可能采用线上会议,将办公室空调温度设定为 24°C,以及在必要时选择最佳路线出行。巴基斯坦命令一半的政府部门员工在家办公,建议私营部门采取相同的做法,要求大学等教育机构切换到网课。越南促所有公民尽可能远程办公,骑自行车而不是驾驶机动车辆出行。

  2. 亚马逊要求资深工程师批准 AI 辅助的代码变更

    亚马逊电商业务召集大批工程师参加周二的会议,围绕一系列故障进行“深度剖析”,其中包括与使用 AI 编码工具相关的故障事件。根据会议简报,亚马逊表示,近几个月出现了“事件增多的趋势”,其特征包括“影响范围大”和“在生成式人工智能辅助下进行的变动”等因素。在“促成因素”一栏中,简报写道:“新型生成式人工智能的使用方式,其最佳实践与防护措施尚未完全确立。”电商业务高级副总裁 Dave Treadwell 现在要求初级和中级工程师的任何 AI 辅助代码变更都需要获得更高级别工程师的批准。

  3. 图灵奖得主 Tony Hoare 去世,享年 92 岁

    英国计算机科学家、图灵奖得主、快速排序算法和霍尔逻辑的开发者、并发和结构化编程理论的先驱 Tony Hoare 于 3 月 5 日去世,享年 92 岁。他出生于英属锡兰的科伦坡,在英国本土受教育,获得牛津的西洋古典学学士学位,大学毕业后在海军服兵役 18 个月,之后回牛津研读统计学取得学士后学位,他后在苏联莫斯科国立大学留学获得博士学位。Tony Hoare 是快速排序 Quicksort 和快速选择 Quickselect 的作者,1980 年获得图灵奖。

  4. 殷拓集团考虑出售 SUSE

    瑞典私募股权公司殷拓集团(EQT)考虑出售 SUSE。EQT 已聘请投行 Arma Partners 接洽一批私募股权投资者商讨出售 SUSE 的可能性。相关讨论处于早期阶段,EQT 是否会进行交易尚无定论。SUSE 是最早的 Linux 发行版之一,诞生于 1992 年,最初的名字 S.u.S.E 是德语 Software und System-Entwicklung 的首字母缩写,意思是“软件和系统开发”,之后改为 SuSE,最后又变成 SUSE。SUSE 在 2003 年被 Novell 收购,2010 年 Novell 被 The Attachmate Group 收购,2014 年它被 Micro Focus 收购。2018 年 EQT 以 25.35 亿美元收购了 SUSE。2021 年它在法兰克福证券交易所上市,2023 年 8 月再次被 EQT 私有化。

  5. 因安全担忧 Ig 诺贝尔奖颁奖典礼将在欧洲举行

    Ig 诺贝尔奖的颁奖典礼过去三十多年都是在美国波士顿的 MIT/哈佛/波士顿大学举行,但今年的颁奖典礼将在瑞士苏黎世举行,且在可预见的未来都会在欧洲举行。原因是国际旅客最近几年访问美国可能不太安全,国际旅客也不太愿意去美国参加活动。Ig 诺贝尔奖创办于1991 年,是对诺贝尔奖的善意戏仿,表彰那些令人发笑但又发人深思的研究,该奖的主办方是《The Annals of Improbable Research》杂志,颁奖活动主持人兼杂志编辑 Marc Abrahams 说,对嘉宾而言前往美国不再安全,他们无法昧着良心要求新获奖者或报道活动的国际记者今年前往美国。今年的颁奖典礼将与苏黎世联邦理工学院和苏黎世大学联合举办,未来几年瑞士将每隔一年举办一次,单数年将在一个欧洲不同城市举办。

  6. 每年逾 500 万例死亡可归因于身体活动不足

    研究显示在全球范围内每年有超过 500 万例死亡可归因于身体活动不足。尽管如此,三分之一的成人和八成的青少年未达到世界卫生组织指南推荐活动量——建议为成人每周 150 分钟中等强度的活动,儿童每天 60 分钟。研究人员分析了全球 68 个国家的身体活动数据,发现了活动性上持续存在的不平等:主动休闲活动(如娱乐性锻炼——唯一持续由选择驱动的活动类型)参与率,在社会优势群体(高收入国家富裕男性)中比弱势群体(低收入国家贫困女性)高 40%。相反,由经济需求驱动的活动(如体力劳动)在弱势群体中参与更普遍。研究发现的证据还表明,身体活动能支持免疫力、降低感染性疾病风险、降低抑郁症状,并与癌症结局改善有关。

  7. 创业公司想发射数千颗反射阳光的卫星

    创业公司 Reflect Orbital 计划发射数千颗安装有反射镜的卫星,在地球上的晚上通过反射阳光提供照明,比如让太阳能发电站在晚上继续发电,或者为救援人员提供照明和照亮城市街道。Reflect Orbital 的首颗原型卫星有冰箱大小,轨道高度大约 644 公里,将会展开一个直径 18 米的镜子,卫星会照亮一块直径约 4.8 公里的圆形区域。该公司已经向 FCC 递交了发射申请。如果 FCC 批准,卫星最早将在夏天发射送人轨道。该公司已筹集到 2800 万美元,计划未来一年再发射两颗原型卫星,到 2028 年底前发射 1000 颗比原型更大的卫星,2030 年前发射 5000 颗。它计划的最大的卫星发射镜直径约 55 米,反射的光量相当于 100 个满月。该公司的目标是在 2035 年前部署总共 5 万颗卫星的完整星座。如果客户签订一年长约,使用时长达到或超过逾千小时,该公司将按每小时约 5000 美元的价格收取单面镜子的照明费用。

  8. 运动时肠道细菌会重写与大脑的化学对话

    当一只老鼠开始奔跑,某种变化便悄然发生。不是那些显而易见的变化——加速的心跳、升温的肌肉、爪子击打跑轮的节奏声——而是更为安静的某种事物。它起始于肠道盘旋的幽暗之处,借由血液与生化信号,一路传递至海马体。那是一片海马形状的薄薄组织,记忆在此成形,情绪在此扎根。发表《脑医学》的一项新研究,已开始绘制这段隐秘的旅程。研究者的发现表明,运动激活了肠道细菌与大脑之间的分子联系。研究考察了成年雄性 Sprague-Dawley 大鼠在自由使用跑轮八周后,其肠道微生物群、循环代谢物及海马基因表达所发生的变化。研究团队发现,运动降低了两种细菌属的相对丰度:Alistipes 与 Clostridium,二者均与色氨酸代谢相关。色氨酸是一种必需氨基酸,也是血清素的前体,在肠-脑信号传导中具有重要地位。

  9. 图书出版商联合起诉安娜的档案

    Penguin Random House、Elsevier 和 HarperCollins 等 13 家大型图书出版商联合起诉安娜的档案(Anna’s Archive),指控该影子图书馆助长图书盗版。出版商此举旨在获得法庭禁令,对安娜的档案的域名注册商施压。安娜的档案已经深陷了多起诉讼,去年底流媒体巨头 Spotify 和唱片公司起诉安娜的档案导致其失去了 .org 主域名。安娜的档案的运营者不太可能在法庭上回应诉讼,在域名被扣押之后,它在不断增加备用域名,上周新上线的备用域名包括了 .VG、.PK 和.GD,但 .VG 在短时间内就被停用。

  10. 很多国际游戏开发者计划不参加今年的 GDC

    数万游戏开发者和制作人本周将齐聚旧金山,参加为期一周的游戏开发者大会(GDC),这是 1988 年以来的传统,但今年的 GDC 将有许多国际游戏开发者缺席,原因是他们觉得美国不再安全,无论 GDC 对他们的工作和职业发展有多么重要,他们不想冒不必要的风险。Godot 基金会执行董事 Emilio Coppola 称他认识的非美国人中没有一个人计划参加 GDC。独立游戏工作室 Le Cabinet du Savoir 的创意总监 Nazih Fares 表示不愿意亲身经历被边检逮捕。去年参加 GDC 的游戏开发者采取了额外的多重安保措施,他们很多人表示为避免麻烦而不想参加今年的 GDC。

  11. Meta 称上传盗版电子书属于合理使用

    为训练大模型,社交巨人 Meta 从 Z-Library 和 LibGen 等影子图书馆平台通过 BitTorrent 下载了逾百 TB 的电子书。在正在进行的由图书作者提起的诉讼中,Meta 律师辩称,通过 BitTorrent 将盗版电子书上传给陌生人属于合理使用。Meta 还强调,这些数据帮助美国确立了其在全球 AI 领域的领先地位。法庭去年裁决,使用盗版电子书训练大模型属于合理使用,但 Meta 仍然需要为通过 BitTorrent 下载和分享电子书的行为承担责任。图书作者认为,Meta 参与了侵权行为。Meta 在上周递交的补充书面询问中表示,在下载 BT 文件过程中共享文件也属于合理使用,理由是这是 BT 协议的固有特性,上传不是选择而是技术本身的工作方式。Meta 还辩称,使用 BitTorrent 共享文件是获取这些宝贵(但盗版)数据的必要手段。以 Anna’s Archive 为例,这些数据集只能通过 BT 下载获取,因此 BitTorrent 是唯一的选择。

  12. 为什么高处坠落的猫总是四脚着地?

    从高处坠落的猫总能四脚着地。科学家一直在争论背后的机制,他们提出了四种假说:收腿翻转(tuck and turn)模型认为猫收起一组爪子以便能旋转身体的不同部位;麦克斯韦(James Clerk Maxwell)提出的下落花样滑冰运动员(falling figure skate)模型认为猫通过收回或伸展爪子调整其角动量;弯曲扭转(bend and twist)模型认为猫在腰部弯曲使身体的两部分产生反向旋转;螺旋尾巴(propeller tail)模型认为猫通过像螺旋桨一样旋转尾巴去反转身体的旋转方向。根据发表在《The Anatomical Record》期刊上的最新研究,日本科学家从五具捐赠的猫尸上取出脊椎,保留韧带和椎间盘,将胸椎和腰椎部分分离,然后将其放入一具扭转装置,观察扭转它们所需的力以及各部分扭转的极限角度。他们还将两只活猫各自抛八次,拍摄了猫在自由落体下的高速照片。结果显示,上段脊椎的扭转角度大于下段脊椎,在扭转角度约 50 度时存在一个“最佳点”,在该点扭转时几乎没有阻力。下段脊椎则不存在这个点,这为“收腿翻转”假说提供了证据。高速摄影也观察到了猫的收腿翻转动作。研究人员还观察到猫总是倾向于右转,可能是内脏器官的不对称分布使其更容易向右转。

  13. 数据中心成为攻击基础设施的目标

    科技行业常把“云”说成是某种抽象且遥不可及的东西。但云运行在数据中心,而数据中心有地址,这个地址可能会遭到无人机袭击。上周亚马逊 AWS 运营的三个数据中心遭到袭击,其中两个位于阿联酋,一个位于巴林。袭击导致设施离线,引发了整个地区银行、支付、外卖应用和企业软件等服务的中断。此次袭击是数据中心首次成为攻击目标。专家认为这肯定不会是最后一次。数据中心正迅速成为重要战略资产,同时也成为易受攻击的目标。

  14. 瑞士通过修宪保障使用现金的权利

    瑞士选民以压倒性多数通过了一项宪法修正案,保障民众使用现金的权利。欧洲除了瑞士,匈牙利、斯洛伐克和斯洛文尼亚等国也都将保障现金使用权利写入宪法。官方统计结果显示,73.4% 的选民支持该宪法修正案。该修正案由政府提出,旨在反击“瑞士自由运动(Swiss Freedom Movement)”组织提出的类似倡议。瑞士自由运动发起了保护现金的倡议,收集了逾 10 万个签名,最终引发了全民公投。由于政府认为该组织提出的部分修正案过于激进,最终该倡议仅获得 46% 的投票支持。

  15. 调查发现三分之一美国人认为末日将在其有生之年来临

    美国相信末日来临的人并非少数。根据《Journal of Personality and Social Psychology》上发表的一篇报告,研究人员调查了 1409 名不同信仰的美国人,结果显示三分之一相信末日将在其有生之年来临。一部分人认为末日是人类引发的,还有部分人认为末日是由神或超自然力量引发的。相信末日临近且人类是罪魁祸首的人,感知到的风险更大,也更支持采取更极端的行动应对威胁。然而相信神控制世界末日的人则不太可能支持采取预防措施。