DIGEST · 2026-03-04

OrangeBot.AI Digest — 2026-03-04

88 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Building a new Flash (bill.newgrounds.com)
  2. An interactive map of Flock Cams (deflock.org)
  3. Making Firefox's right-click not suck with about:config (joshua.hu)
  4. Something is afoot in the land of Qwen (simonwillison.net)
  5. Government grant-funded research should not be published in for-profit journals (www.experimental-history.com)
  6. “It turns out” (2010) (jsomers.net)
  7. Libre Solar – Open Hardware for Renewable Energy (libre.solar)
  8. MacBook Neo (www.apple.com)
  9. MacBook Neo (www.apple.com)
  10. Qwen3.5 Fine-Tuning Guide – Unsloth Documentation (unsloth.ai)
  11. Nobody gets promoted for simplicity (terriblesoftware.org)
  12. Bet on German Train Delays (bahn.bet)
  13. RE#: how we built the fastest regex engine in F# (iev.ee)
  14. RFC 9849. TLS Encrypted Client Hello (www.rfc-editor.org)
  15. Agentic Engineering Patterns (simonwillison.net)

GitHub Trending(13)

  1. KeygraphHQ / shannon
  2. msitarzewski / agency-agents
  3. aquasecurity / trivy
  4. K-Dense-AI / claude-scientific-skills
  5. CodebuffAI / codebuff
  6. agentscope-ai / ReMe
  7. alibaba / OpenSandbox
  8. FujiwaraChoki / MoneyPrinterV2
  9. ItzCrazyKns / Perplexica
  10. agentscope-ai / agentscope
  11. moeru-ai / airi
  12. nautechsystems / nautilus_trader
  13. FlowiseAI / Flowise

Product Hunt(15)

  1. Gemini 3.1 Flash-Lite

    Best-in-class intelligence for your high-volume workloads

  2. Fix in Cursor

    GitHub PR comment to Cursor prompt in one click

  3. Picsart Persona & Storyline

    Design your AI influencer and create any story with it.

  4. Projekt

    The BYOK Design & Dev Tool for Building with Agents

  5. moltdj

    SoundCloud for OpenClaw agents to create, stream, earn

  6. NOVA

    AI coding that goes beyond suggestions

  7. Personal AI Memory

    Captures and stores your chat from various AI platforms

  8. AssemblyAI: Universal-3 Pro Streaming

    The most accurate streaming speech model for voice agents.

  9. ClawPane

    One API. LLM routing for cost, task-fit, latency per request

  10. ClawOffice

    Real Office for your Open Claw Agents

  11. Floyd enterprise world model

    Enterprise world model that learns how you would do tasks

  12. agile.flights

    Agile died in a JIRA board - replace sprints with flights

  13. Kodo

    Create fully editable designs by chatting with AI

  14. Anything API

    Any website. We deliver the API.

  15. Enia Code

    Proactive AI that refines code & learns your standards

Hugging Face(15)

  1. Utonia: Toward One Encoder for All Point Clouds

    We dream of a future where point clouds from all domains can come together to shape a single model that benefits them all. Toward this goal, we present Utonia, a first step toward training a single self-supervised point transformer encoder across diverse domains, spanning remote sensing, outdoor LiDAR, indoor RGB-D sequences, object-centric CAD models, and point clouds lifted from RGB-only videos. Despite their distinct sensing geometries, densities, and priors, Utonia learns a consistent representation space that transfers across domains. This unification improves perception capability while revealing intriguing emergent behaviors that arise only when domains are trained jointly. Beyond perception, we observe that Utonia representations can also benefit embodied and multimodal reasoning: conditioning vision-language-action policies on Utonia features improves robotic manipulation, and integrating them into vision-language models yields gains on spatial reasoning. We hope Utonia can serve as a step toward foundation models for sparse 3D data, and support downstream applications in AR/VR, robotics, and autonomous driving.

  2. UniG2U-Bench: Do Unified Models Advance Multimodal Understanding?

    Unified multimodal models have recently demonstrated strong generative capabilities, yet whether and when generation improves understanding remains unclear. Existing benchmarks lack a systematic exploration of the specific tasks where generation facilitates understanding. To this end, we introduce UniG2U-Bench, a comprehensive benchmark categorizing generation-to-understanding (G2U) evaluation into 7 regimes and 30 subtasks, requiring varying degrees of implicit or explicit visual transformations. Extensive evaluation of over 30 models reveals three core findings: 1) Unified models generally underperform their base Vision-Language Models (VLMs), and Generate-then-Answer (GtA) inference typically degrades performance relative to direct inference. 2) Consistent enhancements emerge in spatial intelligence, visual illusions, or multi-round reasoning subtasks, where enhanced spatial and shape perception, as well as multi-step intermediate image states, prove beneficial. 3) Tasks with similar reasoning structures and models sharing architectures exhibit correlated behaviors, suggesting that generation-understanding coupling induces class-consistent inductive biases over tasks, pretraining data, and model architectures. These findings highlight the necessity for more diverse training data and novel paradigms to fully unlock the potential of unified multimodal modeling.

  3. Beyond Language Modeling: An Exploration of Multimodal Pretraining

    The visual world offers a critical axis for advancing foundation models beyond language. Despite growing interest in this direction, the design space for native multimodal models remains opaque. We provide empirical clarity through controlled, from-scratch pretraining experiments, isolating the factors that govern multimodal pretraining without interference from language pretraining. We adopt the Transfusion framework, using next-token prediction for language and diffusion for vision, to train on diverse data including text, video, image-text pairs, and even action-conditioned video. Our experiments yield four key insights: (i) Representation Autoencoder (RAE) provides an optimal unified visual representation by excelling at both visual understanding and generation; (ii) visual and language data are complementary and yield synergy for downstream capabilities; (iii) unified multimodal pretraining leads naturally to world modeling, with capabilities emerging from general training; and (iv) Mixture-of-Experts (MoE) enables efficient and effective multimodal scaling while naturally inducing modality specialization. Through IsoFLOP analysis, we compute scaling laws for both modalities and uncover a scaling asymmetry: vision is significantly more data-hungry than language. We demonstrate that the MoE architecture harmonizes this scaling asymmetry by providing the high model capacity required by language while accommodating the data-intensive nature of vision, paving the way for truly unified multimodal models.

  4. BeyondSWE: Can Current Code Agent Survive Beyond Single-Repo Bug Fixing?

    Current benchmarks for code agents primarily assess narrow, repository-specific fixes, overlooking critical real-world challenges such as cross-repository reasoning, domain-specialized problem solving, dependency-driven migration, and full-repository generation. To address this gap, we introduce BeyondSWE, a comprehensive benchmark that broadens existing evaluations along two axes - resolution scope and knowledge scope - using 500 real-world instances across four distinct settings. Experimental results reveal a significant capability gap: even frontier models plateau below 45% success, and no single model performs consistently across task types. To systematically investigate the role of external knowledge, we develop SearchSWE, a framework that integrates deep search with coding abilities. Our experiments show that search augmentation yields inconsistent gains and can in some cases degrade performance, highlighting the difficulty of emulating developer-like workflows that interleave search and reasoning during coding tasks. This work offers both a realistic, challenging evaluation benchmark and a flexible framework to advance research toward more capable code agents.

  5. Beyond Length Scaling: Synergizing Breadth and Depth for Generative Reward Models

    Recent advancements in Generative Reward Models (GRMs) have demonstrated that scaling the length of Chain-of-Thought (CoT) reasoning considerably enhances the reliability of evaluation. However, current works predominantly rely on unstructured length scaling, ignoring the divergent efficacy of different reasoning mechanisms: Breadth-CoT (B-CoT, i.e., multi-dimensional principle coverage) and Depth-CoT (D-CoT, i.e., substantive judgment soundness). To address this, we introduce Mix-GRM, a framework that reconfigures raw rationales into structured B-CoT and D-CoT through a modular synthesis pipeline, subsequently employing Supervised Fine-Tuning (SFT) and Reinforcement Learning with Verifiable Rewards (RLVR) to internalize and optimize these mechanisms. Comprehensive experiments demonstrate that Mix-GRM establishes a new state-of-the-art across five benchmarks, surpassing leading open-source RMs by an average of 8.2\%. Our results reveal a clear divergence in reasoning: B-CoT benefits subjective preference tasks, whereas D-CoT excels in objective correctness tasks. Consequently, misaligning the reasoning mechanism with the task directly degrades performance. Furthermore, we demonstrate that RLVR acts as a switching amplifier, inducing an emergent polarization where the model spontaneously allocates its reasoning style to match task demands. The synthesized data and models are released at https://huggingface.co/collections/DonJoey/mix-grm{Hugging Face}, and the code is released at https://github.com/Don-Joey/Mix-GRM{Github}.

  6. Kling-MotionControl Technical Report

    Character animation aims to generate lifelike videos by transferring motion dynamics from a driving video to a reference image. Recent strides in generative models have paved the way for high-fidelity character animation. In this work, we present Kling-MotionControl, a unified DiT-based framework engineered specifically for robust, precise, and expressive holistic character animation. Leveraging a divide-and-conquer strategy within a cohesive system, the model orchestrates heterogeneous motion representations tailored to the distinct characteristics of body, face, and hands, effectively reconciling large-scale structural stability with fine-grained articulatory expressiveness. To ensure robust cross-identity generalization, we incorporate adaptive identity-agnostic learning, facilitating natural motion retargeting for diverse characters ranging from realistic humans to stylized cartoons. Simultaneously, we guarantee faithful appearance preservation through meticulous identity injection and fusion designs, further supported by a subject library mechanism that leverages comprehensive reference contexts. To ensure practical utility, we implement an advanced acceleration framework utilizing multi-stage distillation, boosting inference speed by over 10x. Kling-MotionControl distinguishes itself through intelligent semantic motion understanding and precise text responsiveness, allowing for flexible control beyond visual inputs. Human preference evaluations demonstrate that Kling-MotionControl delivers superior performance compared to leading commercial and open-source solutions, achieving exceptional fidelity in holistic motion control, open domain generalization, and visual quality and coherence. These results establish Kling-MotionControl as a robust solution for high-quality, controllable, and lifelike character animation.

  7. How Controllable Are Large Language Models? A Unified Evaluation across Behavioral Granularities

    Large Language Models (LLMs) are increasingly deployed in socially sensitive domains, yet their unpredictable behaviors, ranging from misaligned intent to inconsistent personality, pose significant risks. We introduce SteerEval, a hierarchical benchmark for evaluating LLM controllability across three domains: language features, sentiment, and personality. Each domain is structured into three specification levels: L1 (what to express), L2 (how to express), and L3 (how to instantiate), connecting high-level behavioral intent to concrete textual output. Using SteerEval, we systematically evaluate contemporary steering methods, revealing that control often degrades at finer-grained levels. Our benchmark offers a principled and interpretable framework for safe and controllable LLM behavior, serving as a foundation for future research.

  8. Qwen3-Coder-Next Technical Report

    We present Qwen3-Coder-Next, an open-weight language model specialized for coding agents. Qwen3-Coder-Next is an 80-billion-parameter model that activates only 3 billion parameters during inference, enabling strong coding capability with efficient inference. In this work, we explore how far strong training recipes can push the capability limits of models with small parameter footprints. To achieve this, we perform agentic training through large-scale synthesis of verifiable coding tasks paired with executable environments, allowing learning directly from environment feedback via mid-training and reinforcement learning. Across agent-centric benchmarks including SWE-Bench and Terminal-Bench, Qwen3-Coder-Next achieves competitive performance relative to its active parameter count. We release both base and instruction-tuned open-weight versions to support research and real-world coding agent development.

  9. PRISM: Pushing the Frontier of Deep Think via Process Reward Model-Guided Inference

    DEEPTHINK methods improve reasoning by generating, refining, and aggregating populations of candidate solutions, which enables strong performance on complex mathematical and scientific tasks. However, existing frameworks often lack reliable correctness signals during inference, which creates a population-enhancement bottleneck where deeper deliberation amplifies errors, suppresses correct minority solutions, and yields weak returns to additional compute. In this paper, we introduce a functional decomposition of DEEPTHINK systems and propose PRISM, a Process Reward Model (PRM)-guided inference algorithm that uses step-level verification to guide both population refinement and solution aggregation. During refinement, PRISM treats candidate solutions as particles in a PRM-defined energy landscape and reshapes the population through score-guided resampling and stochastic refinement, which concentrates probability mass on higher-quality reasoning while preserving diversity. Across mathematics and science benchmarks, PRISM is competitive with or outperforms existing DEEPTHINK methods, reaching 90.0%, 75.4%, and 71.4% with gpt-oss-20b on AIME25, HMMT25, and GPQA Diamond, respectively, while matching or exceeding gpt-oss-120b. Additionally, our analysis shows that PRISM produces consistent net-directional correction during refinement, remains reliable when the initial population contains few correct candidates, and often lies on the compute-accuracy Pareto frontier.

  10. Kiwi-Edit: Versatile Video Editing via Instruction and Reference Guidance

    Instruction-based video editing has witnessed rapid progress, yet current methods often struggle with precise visual control, as natural language is inherently limited in describing complex visual nuances. Although reference-guided editing offers a robust solution, its potential is currently bottlenecked by the scarcity of high-quality paired training data. To bridge this gap, we introduce a scalable data generation pipeline that transforms existing video editing pairs into high-fidelity training quadruplets, leveraging image generative models to create synthesized reference scaffolds. Using this pipeline, we construct RefVIE, a large-scale dataset tailored for instruction-reference-following tasks, and establish RefVIE-Bench for comprehensive evaluation. Furthermore, we propose a unified editing architecture, Kiwi-Edit, that synergizes learnable queries and latent visual features for reference semantic guidance. Our model achieves significant gains in instruction following and reference fidelity via a progressive multi-stage training curriculum. Extensive experiments demonstrate that our data and architecture establish a new state-of-the-art in controllable video editing. All datasets, models, and code is released at https://github.com/showlab/Kiwi-Edit.

  11. Next Embedding Prediction Makes World Models Stronger

    Capturing temporal dependencies is critical for model-based reinforcement learning (MBRL) in partially observable, high-dimensional domains. We introduce NE-Dreamer, a decoder-free MBRL agent that leverages a temporal transformer to predict next-step encoder embeddings from latent state sequences, directly optimizing temporal predictive alignment in representation space. This approach enables NE-Dreamer to learn coherent, predictive state representations without reconstruction losses or auxiliary supervision. On the DeepMind Control Suite, NE-Dreamer matches or exceeds the performance of DreamerV3 and leading decoder-free agents. On a challenging subset of DMLab tasks involving memory and spatial reasoning, NE-Dreamer achieves substantial gains. These results establish next-embedding prediction with temporal transformers as an effective, scalable framework for MBRL in complex, partially observable environments.

  12. Humans and LLMs Diverge on Probabilistic Inferences

    Human reasoning often involves working over limited information to arrive at probabilistic conclusions. In its simplest form, this involves making an inference that is not strictly entailed by a premise, but rather only likely given the premise. While reasoning LLMs have demonstrated strong performance on logical and mathematical tasks, their behavior on such open-ended, non-deterministic inferences remains largely unexplored. We introduce ProbCOPA, a dataset of 210 handcrafted probabilistic inferences in English, each annotated for inference likelihood by 25--30 human participants. We find that human responses are graded and varied, revealing probabilistic judgments of the inferences in our dataset. Comparing these judgments with responses from eight state-of-the-art reasoning LLMs, we show that models consistently fail to produce human-like distributions. Finally, analyzing LLM reasoning chains, we find evidence of a common reasoning pattern used to evaluate such inferences. Our findings reveal persistent differences between humans and LLMs, and underscore the need to evaluate reasoning beyond deterministic settings.

  13. BBQ-to-Image: Numeric Bounding Box and Qolor Control in Large-Scale Text-to-Image Models

    Text-to-image models have rapidly advanced in realism and controllability, with recent approaches leveraging long, detailed captions to support fine-grained generation. However, a fundamental parametric gap remains: existing models rely on descriptive language, whereas professional workflows require precise numeric control over object location, size, and color. In this work, we introduce BBQ, a large-scale text-to-image model that directly conditions on numeric bounding boxes and RGB triplets within a unified structured-text framework. We obtain precise spatial and chromatic control by training on captions enriched with parametric annotations, without architectural modifications or inference-time optimization. This also enables intuitive user interfaces such as object dragging and color pickers, replacing ambiguous iterative prompting with precise, familiar controls. Across comprehensive evaluations, BBQ achieves strong box alignment and improves RGB color fidelity over state-of-the-art baselines. More broadly, our results support a new paradigm in which user intent is translated into an intermediate structured language, consumed by a flow-based transformer acting as a renderer and naturally accommodating numeric parameters.

  14. Spilled Energy in Large Language Models

    We reinterpret the final Large Language Model (LLM) softmax classifier as an Energy-Based Model (EBM), decomposing the sequence-to-sequence probability chain into multiple interacting EBMs at inference. This principled approach allows us to track "energy spills" during decoding, which we empirically show correlate with factual errors, biases, and failures. Similar to Orgad et al. (2025), our method localizes the exact answer token and subsequently tests for hallucinations. Crucially, however, we achieve this without requiring trained probe classifiers or activation ablations. Instead, we introduce two completely training-free metrics derived directly from output logits: spilled energy, which captures the discrepancy between energy values across consecutive generation steps that should theoretically match, and marginalized energy, which is measurable at a single step. Evaluated on nine benchmarks across state-of-the-art LLMs (including LLaMA, Mistral, and Gemma) and on synthetic algebraic operations (Qwen3), our approach demonstrates robust, competitive hallucination detection and cross-task generalization. Notably, these results hold for both pretrained and instruction-tuned variants without introducing any training overhead. Code available at: github.com/OmnAI-Lab/spilled-energy

  15. Surgical Post-Training: Cutting Errors, Keeping Knowledge

    Enhancing the reasoning capabilities of Large Language Models (LLMs) via post-training is often constrained by the trade-off between efficiency and catastrophic forgetting. While prior research emphasizes the role of on-policy data in mitigating forgetting, we uncover--and validate both theoretically and empirically--an overlooked yet critical mechanism: the implicit regularization inherent in Direct Preference Optimization's (DPO) reward estimate. This motivates our Surgical Post-Training (SPoT), a new paradigm designed to optimize reasoning efficiently while preserving learned prior knowledge. SPoT consists of: (1) a data rectification pipeline that employs an Oracle to surgically correct erroneous steps via minimal edits, generating data proximal to the model's distribution; and (2) a reward-based binary cross-entropy objective. Unlike the relative ranking in DPO, this objective treats reasoning correctness as a binary classification problem, enforcing decoupled supervision signals. Empirically, with only 4k rectified math data pairs, SPoT improves Qwen3-8B's accuracy by 6.2% on average across in-domain and OOD tasks, requiring merely 28 minutes of training on 8x H800 GPUs. Code: https://github.com/Visual-AI/SPoT

Techmeme(15)

  1. Pasqal, a French startup that builds quantum processors using neutral atom technology, plans to go public via a SPAC merger at a $2B pre-money valuation (Bailey Lipschultz/Bloomberg)

    Bailey Lipschultz / Bloomberg : Pasqal, a French startup that builds quantum processors using neutral atom technology, plans to go public via a SPAC merger at a $2B pre-money valuation —  Pasqal Holding SAS agreed to merge with a blank-check firm in a deal that values the combined company at $2 billion pre-money …

  2. Google, Microsoft, Meta, Amazon, OpenAI, and others sign a pledge at the White House to bear the cost of new electricity generation to power their data centers (Reuters)

    Reuters : Google, Microsoft, Meta, Amazon, OpenAI, and others sign a pledge at the White House to bear the cost of new electricity generation to power their data centers —  Google (GOOGL.O), Microsoft (MSFT.O), Meta (META.O), Amazon (AMZN.O) and several artificial intelligence companies signed a pledge …

  3. Leaked Friday memo: Dario Amodei called OpenAI's DOD deal "safety theater", said DOD dislikes Anthropic in part for not giving "dictator-style praise to Trump" (The Information)

    The Information : Leaked Friday memo: Dario Amodei called OpenAI's DOD deal “safety theater”, said DOD dislikes Anthropic in part for not giving “dictator-style praise to Trump” —  Anthropic CEO Dario Amodei on Friday told employees that a deal OpenAI and its CEO Sam Altman struck …

  4. Source: a16z Crypto is targeting around $2B for its fifth fund and plans to close the raise by the end of the first half of 2026 (Fortune)

    Fortune : Source: a16z Crypto is targeting around $2B for its fifth fund and plans to close the raise by the end of the first half of 2026 —  The largest player in the crypto venture world is back on the fundraising circuit.  The blockchain arm of Andreessen Horowitz, also known as a16z crypto …

  5. Broadcom reports Q1 revenue up 29% YoY to $19.31B, vs. $19.18B est., AI revenue up 106% to $8.4B, and announces a $10B share buyback program (CNBC)

    CNBC : Broadcom reports Q1 revenue up 29% YoY to $19.31B, vs. $19.18B est., AI revenue up 106% to $8.4B, and announces a $10B share buyback program —  Broadcom reported better-than-expected earnings and revenue and issued a strong forecast for the current period as the chipmaker continues to benefit from the artificial intelligence boom.

  6. Source: Embo, which is developing world models for robotics, is in talks to raise a $100M+ seed led by a16z, with Khosla, DST Global, and Striker participating (Kevin McLaughlin/The Information)

    Kevin McLaughlin / The Information : Source: Embo, which is developing world models for robotics, is in talks to raise a $100M+ seed led by a16z, with Khosla, DST Global, and Striker participating —  Embo, an AI startup co-founded by former Google DeepMind research scientists, is in talks to raise a seed round of more than $100 million led …

  7. Google Pixel 10a review: long battery life and a good camera with no bump, but weak gaming performance and barely an upgrade from the Pixel 9a (Ryan Whitwam/Ars Technica)

    Ryan Whitwam / Ars Technica : Google Pixel 10a review: long battery life and a good camera with no bump, but weak gaming performance and barely an upgrade from the Pixel 9a —  Meet the new boss, same as the old boss.  —  Google's budget Pixels have long been a top recommendation for anyone who needs a phone …

  8. Tencent-owned Finnish game company Supercell says it is cooperating with a CFIUS security probe of Tencent's data practices (Cecilia D'Anastasio/Bloomberg)

    Cecilia D'Anastasio / Bloomberg : Tencent-owned Finnish game company Supercell says it is cooperating with a CFIUS security probe of Tencent's data practices —  Supercell Oy, the Finnish game company owned by Tencent Holdings Ltd., said it's cooperating with a US government security probe of its Chinese parent's data practices.

  9. Jensen Huang says Nvidia's recent $30B investment in OpenAI "might be the last time" it invests in the company, because OpenAI is "going to go public" (Ashley Capoot/CNBC)

    Ashley Capoot / CNBC : Jensen Huang says Nvidia's recent $30B investment in OpenAI “might be the last time” it invests in the company, because OpenAI is “going to go public” —  Nvidia CEO Jensen Huang said the company's recent $30 billion investment in OpenAI “might be the last time” …

  10. Authorities from 14 countries shut down LeakBase, seize its domains, and arrest multiple people allegedly tied to the cybercrime forum, which had 142K+ members (Matt Kapko/CyberScoop)

    Matt Kapko / CyberScoop : Authorities from 14 countries shut down LeakBase, seize its domains, and arrest multiple people allegedly tied to the cybercrime forum, which had 142K+ members —  The marketplace was one of the world's largest hubs for cybercrime with more than 142,000 members.

  11. Google announces an Android app store program and lower developer fees to resolve Epic's antitrust litigation and comply with new rules in Europe and elsewhere (Leah Nylen/Bloomberg)

    Leah Nylen / Bloomberg : Google announces an Android app store program and lower developer fees to resolve Epic's antitrust litigation and comply with new rules in Europe and elsewhere —  Alphabet Inc.'s Google unveiled a new system for apps on its Android phones and tablets Wednesday, agreeing to easier access …

  12. Sources: some investors push Anthropic to de-escalate its DOD dispute and avoid the "supply-chain risk" designation; source: some Anthropic-DOD talks continue (Reuters)

    Reuters : Sources: some investors push Anthropic to de-escalate its DOD dispute and avoid the “supply-chain risk” designation; source: some Anthropic-DOD talks continue —  Some Anthropic investors are racing to contain fallout from the AI research lab's dispute with the Pentagon …

  13. Sources: Arda, co-founded by ex-OpenAI chief research officer Bob McGrew to automate manufacturing using AI, is raising $70M at a $700M valuation (Wall Street Journal)

    Wall Street Journal : Sources: Arda, co-founded by ex-OpenAI chief research officer Bob McGrew to automate manufacturing using AI, is raising $70M at a $700M valuation —  Bob McGrew is raising $70 million to fund a startup making software platform to help run autonomous factories

  14. Sources: Neura Robotics, which is building cognitive, humanoid robots for logistics, is raising ~€1B in a funding round backed by Tether at a ~€4B valuation (Bloomberg)

    Bloomberg : Sources: Neura Robotics, which is building cognitive, humanoid robots for logistics, is raising ~€1B in a funding round backed by Tether at a ~€4B valuation —  German startup Neura Robotics is raising about €1 billion ($1.2 billion) in a funding round backed …

  15. The Media Trust report: online ads surpassed email as the primary malware channel in 2025, accounting for 60%+ of all observed malware and phishing campaigns (Lara O'Reilly/Business Insider)

    Lara O'Reilly / Business Insider : The Media Trust report: online ads surpassed email as the primary malware channel in 2025, accounting for 60%+ of all observed malware and phishing campaigns —  Follow Lara O'Reilly … - Online ads leapfrogged email as the primary channel for malware in 2025, per a new report.

Solidot(15)

  1. 《Highguard》将于 3 月 12 日永久关闭

    《Highguard》开发商 Wildlight Entertainment 宣布游戏将于 3 月 12 日永久关闭。《Highguard》是一款以突袭为主题的英雄射击游戏,于 1 月 26 日上线,一度吸引了 9.7 万玩家同时在线,但这一热度并没有持续太长时间,根据 Steamdb 的统计,过去二十四小时游戏同时在线人数最高仅为 460 人,对一款需要长期运营的免费 PvP 游戏而言,结局已经注定了。在关闭前《Highguard》共运营 45 天,是索尼《Concord》的 3.75 倍长,《Concord》运营 12 天就永久关闭了。Wildlight 的主要投资者是腾讯,它已经在两周前撤回了投资。

  2. 利用大模型进行大规模去匿名化

    根据海量数据训练并能快速检索相关信息的大模型大幅降低了网络开盒(或叫去匿名化)的成本。一个人可仅仅通过少数特征被个别界定,比如仅通过邮政编码、出生日期和性别,87% 的美国人口即可被个别界定。根据发表在预印本平台 arXiv 的一篇论文,大模型能用于大规模的去匿名化,能高精度的识别网络上的匿名用户。研究人员设计了一个攻击流程:提取身份特征,搜索候选匹配,通过推理验证匹配结果减少误判。传统的去匿名工作需要专业调查人员花费数小时或更长时间,大模型不仅花费时间更少,而且可以大幅扩大规模。利用大模型,以关联 Hacker News 匿名账号和 LinkedIn 实名账号为例,系统能在维持 99% 精度的情况下,将回索率从 0.1% 大幅提升至 45.1%。回索率(Recall)被用于衡量模型找回所有相关信息的能力。研究人员指出,保护网民匿名性的旧方法不再有效。

  3. Google Chrome 将每两周发布一个新版本

    从 9 月 8 日发布的 v153 起,Google Chrome 发布周期将从四周缩短到两周。Google 表示此举旨在确保开发者和用户能立即获取到最新的性能改进、修复和新功能。两周发布周期的版本更新规模更小,最大限度减少中断,简化发布后的调试(debugging)。Google 表示其它方面基本没有变化,如每周的安全更新,Dev 和 Canary 渠道都没有变化。面向企业和 Chromium 嵌入者客户的 Extended Stable 八周发布周期也没有变。

  4. OpenAI 开发 GitHub 的替代

    OpenAI 正在开发一个代码托管平台,与微软的 GitHub 展开竞争。原因据称 GitHub 服务过去几个月频繁中断,导致 OpenAI 工程师们无法工作,因此他们决定开发自己能控制的新代码托管平台。该项目仍然处于早期阶段,可能还需要几个月时间才能完成。如果 OpenAI 真的推出这款产品,那么这将意味着 OpenAI 将与其大股东微软展开直接竞争。根据最新一轮融资,OpenAI 的估值达到了 8400 亿美元。

  5. 蔗糖有助于缓解新生儿的疼痛

    新生儿,尤其是住在新生儿加护病房的早产儿,往往需接受多次疼痛处置。蔗糖为一种易取得、低成本的甜味溶液,在针刺处置前滴入口腔,数十年来已被应用于新生儿止痛。然而针对特定处置(如静脉采血)的实证仍相对有限。研究人员回顾纳入 29 项临床试验,涵盖超过 2,700 名接受静脉采血的新生儿(包括早产与足月婴儿)。结果显示,与未给予治疗、给予清水或标准照护相比,蔗糖可能在针刺当下及处置后立即减轻疼痛反应。研究亦指出,若将蔗糖与非营养性吸吮(如安抚奶嘴)并用,止痛效果更为显著。

  6. 摩托罗拉的 GrapheneOS 设备将支持解锁引导程序

    摩托罗拉本周一宣布与 Android 安全加固社区发行版项目 GrapheneOS 展开合作。GrapheneOS 项目通过其 Mastodon 社媒账号表示,摩托罗拉未来的手机设备将正式支持 GrapheneOS,而根据双方达成的硬件要求,这些设备的引导程序(bootloader)是允许解锁和重新锁定的,支持安装其它移动操作系统。过去几年 Android 厂商越来越多的禁止解锁引导程序,日益向苹果看齐。

  7. AI 生成的作品不受版权保护

    美国最高法院拒绝受理计算机科学家 Stephen Thaler 提起的 AI 生成作品是否拥有版权的案件。此前美国地方法院、上诉法院以及美国版权局都声明 AI 生成的艺术作品不受版权保护。美国专利局也表示,AI 系统不能被列为专利发明人,但人仍然可使用 AI 工具去开发专利。英国最高法院在 2023 年裁决 AI 不能成为专利的发明人——这起案件也是 Stephen Thaler 提起的。

  8. ChatGPT 卸载率在五角大楼交易之后飙升 295%

    根据 Sensor Tower 的数据,在 OpenAI 与五角大楼达成交易之后,用户对此做出了反应,2 月 28 日当天 OpenAI AI 聊天机器人 ChatGPT 应用的卸载量在美国比前一天飙升 295%,而过去 30 天它的平均日卸载率是 9%。与此同时,拒绝五角大楼要求的 OpenAI 竞争对手 Anthropic 的 Claude 应用的下载量在 2 月 27 日和 2 月 28 日分别增长了 37% 和 51%。ChatGPT 的下载量也受到了交易的影响,在宣布与五角大楼合作前的 2 月 27 日 ChatGPT 下载量环比增长 14%,但宣布交易后的 28 日其下载量环比下降 13%,3 月 1 日下载量继续环比下降 5%。Claude 在美国免费应用排行榜也已经连续三天登上榜首,这一波热潮也导致 Claude 多次发生短暂的宕机。

  9. ARM Cortex X925 桌面性能赶上了 AMD 和英特尔

    英国公司 Arm 设计的芯片长期以来是为低功耗和小面积优化的,但它也一直推出针对高性能应用场景的核心。2012 年 Arm 发布 64 位核心 Cortex A57 时,能媲美 AMD 和英特尔最新处理器还是遥不可及的梦想。它在 2024 年推出的高性能核心 Cortex X925 已将梦想变成了现实。英伟达超级芯片 GB10 Superchip 使用的 Arm 核心就是基于 Cortex X925。它在桌面性能上赶上了 AMD Zen 5 和英特尔的 Lion Cove。GB10 使用了 10 个 X925 核心,分成两个集群,其中之一的 X925 核心最高频率 4 GHz,另一个是 3.9 GHz。测试显示它的重排序性能优于 AMD Zen 5,L2 缓存容量赶上了英特尔处理器的 P-Cores(即性能核心)。

  10. 南极过去三十年损失了 1.2 万平方公里的底部冰

    加州尔湾的冰川学家绘制了过去 30 年的南极洲环极底部冰线迁移图,显示它损失了逾 1.2 万平方公里的底部冰(grounded ice)。底部冰是直接与海床或基岩接触的冰,区别于漂浮在水面上的冰架或冰山,它们通常更稳定。研究人员综合分析了多颗卫星的数据,发现 77% 的海岸线未发生冰川接地线迁移,但西南极洲、南极半岛和东南极洲部分地区损失了 12,820 平方公里的底部冰。冰盖正以平均每年 442 平方公里的速度从接地线后退。变化最显著的是西南极洲的 Amundsen Sea 和 Getz 海域,冰川后退了 10-40 公里左右。Pine Island 冰川后退了 33 公里,Thwaites 冰川后退了 26 公里,Smith 冰川的后退距离达到了 42 公里。

  11. 小米莱卡手机起售价 1.6 万元

    小米和莱卡宣布了面向高端智能手机市场的莱卡智能手机 Leitzphone,以莱卡创始人 Ernst Leitz 的名字命名。Leitzphone 配备了两个 5000 万像素镜头和一个 2 亿像素镜头,提供了类似相机的调整焦距、快门速度和曝光的旋钮,硬件配置与 Xiaomi 17 Ultra 基本相同,起售价 1999 欧元——约 1.6 万人民币。

  12. 人体血液中 CO2 水平也在上升

    随着大气二氧化碳(CO2)浓度持续升高,人体血液中的 CO2 水平也在悄然攀升。如果这一趋势不加遏制,几十年内,一项关键血液指标可能逼近健康警戒线。相关论文发表于《Air Quality, Atmosphere & Health》杂志。 为探究大气变化对人体内部的影响,研究团队系统梳理了美国 20 余年间的大规模人口数据。结果发现,血液化学成分的悄然演变,与大气 CO2 的上升轨迹惊人同步。数据显示,血清碳酸氢盐这一与体内 CO2 水平密切挂钩的标志物的平均浓度,在 20 余年间上升了约 7%。与此同时,钙和磷的平均水平则出现了下降。这一变化恰与大气 CO2 浓度从 2000 年的 369ppm 攀升至如今超过 420ppm 的背景重合。团队表示,这暗示着人体可能正在“默默代偿”,以应对环境变化带来的内在压力。碳酸氢盐是维持人体血液酸碱平衡的“缓冲剂”。当血液中 CO2 浓度升高,身体便会自动保留更多碳酸氢盐,以中和酸性,稳定内环境。然而这种代偿并非长久之计。持续的微调,终将打破人体精密的生理平衡。模型推演显示,若当前趋势不改,50 年后人体血清碳酸氢盐平均水平或将触及当今健康范围的上限,而钙、磷浓度也可能在本世纪晚些时候跌破下限。

  13. Ars Technica 的 AI 记者离职

    知名科技媒体 Ars Technica 上个月在报道 AI 新闻时被发现将 AI 生成的内容作为消息来源使用,Ars 联合创始人兼主编 Ken Fisher 为此发表声明公开道歉。这篇报道的合作者 Benj Edwards 是 Ars 的资深 AI 记者,他表示自己承担全部责任,另一位合作者与这起错误没有关联。他辩解说自己尝试使用基于 Claude Code 的实验性 AI 工具从原始材料中提取出可添加到大纲的结构化引用内容,但该 AI 拒绝处理,他猜测可能是文章描述的是一起骚扰事件(AI 骚扰人类),他于是将文本拷贝到 ChatGPT,没有注意到 ChatGPT 生成了文章作者的意译版本而不是原话,在引用时没有核实引用是否与原文一致。现在 Benj Edwards 在 Ars 的简历已经变成了过去时态,意味着他已经离职,但是否被解雇 Ars 没有说明, Edwards 本人拒绝置评。

  14. 加拿大 BC 省永久采用夏令时

    加拿大不列颠哥伦比亚省(British Columbia)省长 David Eby 宣布,BC 省将永久采用夏令时,本周日 3 月 8 日将是当地居民最后一次调整时钟。Eby 表示居民已经等待太长时间了,这次之后将不再需要去调整时钟。研究已经发现,一年两次调整时间并不利于公众健康和安全。BC 省居民将有八个月的时间为 2026 年 11 月 1 日做准备,届时时钟原本应拨慢一小时,但将保持不变。BC 省的新时区将被称为“太平洋时间”。

  15. 日本计划禁止飞机乘客使用充电宝

    在去年发生多起与移动电源相关的飞机起火事故后,日本计划从 4 月中旬起禁止乘客在飞机上使用移动电源/充电宝。日本国土交通省发布了《民用航空条例》修订版征求意见稿。在日本,移动电源被分类为备用电池,被禁止托运。随身行李中的移动电池则禁止超过 160 瓦时,超过 100 瓦时只能携带两块,低于 100 瓦时的不受限制。新修订的规定限制乘客最多能携带包括充电宝在内的两块电池,禁止在机上充电,建议乘客不要使用充电宝。日本航空公司则可能一步到位,直接禁止乘客使用充电宝。