DIGEST · 2026-03-02

OrangeBot.AI Digest — 2026-03-02

84 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. The workers behind Meta's smart glasses can see everything (www.svd.se)
  2. Welcome (back) to Macintosh (take.surf)
  3. British Columbia to end time changes, adopt year-round daylight time (www.cbc.ca)
  4. First in-utero stem cell therapy for fetal spina bifida repair is safe: study (health.ucdavis.edu)
  5. Anthropic Cowork feature creates 10GB VM bundle on macOS without warning (github.com)
  6. New iPad Air, powered by M4 (www.apple.com)
  7. OpenClaw surpasses React to become the most-starred software project on GitHub (www.star-history.com)
  8. Inside the M4 Apple Neural Engine, Part 1: Reverse Engineering (maderix.substack.com)
  9. Microslop Manifesto (microslop.com)
  10. Microsoft bans the word "Microslop" on its Discord, then locks the server (www.windowslatest.com)
  11. U.S. science agency moves to restrict foreign scientists from its labs (www.science.org)
  12. Mondrian Entered the Public Domain. The Estate Disagrees (copyrightlately.com)
  13. Jolla phone – a full-stack European alternative (commerce.jolla.com)
  14. /e/OS is a complete, fully “deGoogled” mobile ecosystem (e.foundation)
  15. How to talk to anyone and why you should (www.theguardian.com)

GitHub Trending(9)

  1. ruvnet / wifi-densepose
  2. moeru-ai / airi
  3. anthropics / prompt-eng-interactive-tutorial
  4. ruvnet / ruflo
  5. alibaba / OpenSandbox
  6. microsoft / markitdown
  7. K-Dense-AI / claude-scientific-skills
  8. superset-sh / superset
  9. servo / servo

Product Hunt(15)

  1. Rankfender

    AI visibility and automated SEO optimization platform

  2. ChatWithAds

    From Data to AI-Assisted Decision, In One Conversation.

  3. Mosaic

    Zapier for Video Editing

  4. JDoodleClaw

    The most user-friendly OpenClaw. Securely hosted.

  5. Voca AI

    The AI project manager that runs in the background

  6. WEIR AI

    Track your identity online to protect it or earn from it

  7. GojiberryAI

    AI agents turning high-intent leads into booked demos

  8. NothingHere

    A MacOS panic button where one key press cleans your screen

  9. Crawler.sh

    Free Local AEO & SEO Spider and a Markdown content extractor

  10. Aura

    Semantic version control for AI coding agents on top of Git

  11. Didit v3

    One platform for KYC, biometrics & fraud. 70% lower costs.

  12. Unfold

    Extend macOS Quick Look to folders, archives & code files

  13. CtrlAI

    Transparent proxy that secures AI agents with guardrails

  14. Expressive Mode for ElevenAgents

    AI voice agents that adapt tone, timing & emotion by context

  15. KatClaw™

    Your AI assistant to automate without scripts on Mac

Hugging Face(15)

  1. dLLM: Simple Diffusion Language Modeling

    Although diffusion language models (DLMs) are evolving quickly, many recent models converge on a set of shared components. These components, however, are distributed across ad-hoc research codebases or lack transparent implementations, making them difficult to reproduce or extend. As the field accelerates, there is a clear need for a unified framework that standardizes these common components while remaining flexible enough to support new methods and architectures. To address this gap, we introduce dLLM, an open-source framework that unifies the core components of diffusion language modeling -- training, inference, and evaluation -- and makes them easy to customize for new designs. With dLLM, users can reproduce, finetune, deploy, and evaluate open-source large DLMs such as LLaDA and Dream through a standardized pipeline. The framework also provides minimal, reproducible recipes for building small DLMs from scratch with accessible compute, including converting any BERT-style encoder or autoregressive LM into a DLM. We also release the checkpoints of these small DLMs to make DLMs more accessible and accelerate future research.

  2. Enhancing Spatial Understanding in Image Generation via Reward Modeling

    Recent progress in text-to-image generation has greatly advanced visual fidelity and creativity, but it has also imposed higher demands on prompt complexity-particularly in encoding intricate spatial relationships. In such cases, achieving satisfactory results often requires multiple sampling attempts. To address this challenge, we introduce a novel method that strengthens the spatial understanding of current image generation models. We first construct the SpatialReward-Dataset with over 80k preference pairs. Building on this dataset, we build SpatialScore, a reward model designed to evaluate the accuracy of spatial relationships in text-to-image generation, achieving performance that even surpasses leading proprietary models on spatial evaluation. We further demonstrate that this reward model effectively enables online reinforcement learning for the complex spatial generation. Extensive experiments across multiple benchmarks show that our specialized reward model yields significant and consistent gains in spatial understanding for image generation.

  3. Recovered in Translation: Efficient Pipeline for Automated Translation of Benchmarks and Datasets

    The reliability of multilingual Large Language Model (LLM) evaluation is currently compromised by the inconsistent quality of translated benchmarks. Existing resources often suffer from semantic drift and context loss, which can lead to misleading performance metrics. In this work, we present a fully automated framework designed to address these challenges by enabling scalable, high-quality translation of datasets and benchmarks. We demonstrate that adapting test-time compute scaling strategies, specifically Universal Self-Improvement (USI) and our proposed multi-round ranking method, T-RANK, allows for significantly higher quality outputs compared to traditional pipelines. Our framework ensures that benchmarks preserve their original task structure and linguistic nuances during localization. We apply this approach to translate popular benchmarks and datasets into eight Eastern and Southern European languages (Ukrainian, Bulgarian, Slovak, Romanian, Lithuanian, Estonian, Turkish, Greek). Evaluations using both reference-based metrics and LLM-as-a-judge show that our translations surpass existing resources, resulting in more accurate downstream model assessment. We release both the framework and the improved benchmarks to facilitate robust and reproducible multilingual AI development.

  4. CUDA Agent: Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

    GPU kernel optimization is fundamental to modern deep learning but remains a highly specialized task requiring deep hardware expertise. Despite strong performance in general programming, large language models (LLMs) remain uncompetitive with compiler-based systems such as torch.compile for CUDA kernel generation. Existing CUDA code generation approaches either rely on training-free refinement or fine-tune models within fixed multi-turn execution-feedback loops, but both paradigms fail to fundamentally improve the model's intrinsic CUDA optimization ability, resulting in limited performance gains. We present CUDA Agent, a large-scale agentic reinforcement learning system that develops CUDA kernel expertise through three components: a scalable data synthesis pipeline, a skill-augmented CUDA development environment with automated verification and profiling to provide reliable reward signals, and reinforcement learning algorithmic techniques enabling stable training. CUDA Agent achieves state-of-the-art results on KernelBench, delivering 100\%, 100\%, and 92\% faster rate over torch.compile on KernelBench Level-1, Level-2, and Level-3 splits, outperforming the strongest proprietary models such as Claude Opus 4.5 and Gemini 3 Pro by about 40\% on the hardest Level-3 setting.

  5. Mode Seeking meets Mean Seeking for Fast Long Video Generation

    Scaling video generation from seconds to minutes faces a critical bottleneck: while short-video data is abundant and high-fidelity, coherent long-form data is scarce and limited to narrow domains. To address this, we propose a training paradigm where Mode Seeking meets Mean Seeking, decoupling local fidelity from long-term coherence based on a unified representation via a Decoupled Diffusion Transformer. Our approach utilizes a global Flow Matching head trained via supervised learning on long videos to capture narrative structure, while simultaneously employing a local Distribution Matching head that aligns sliding windows to a frozen short-video teacher via a mode-seeking reverse-KL divergence. This strategy enables the synthesis of minute-scale videos that learns long-range coherence and motions from limited long videos via supervised flow matching, while inheriting local realism by aligning every sliding-window segment of the student to a frozen short-video teacher, resulting in a few-step fast long video generator. Evaluations show that our method effectively closes the fidelity-horizon gap by jointly improving local sharpness, motion and long-range consistency. Project website: https://primecai.github.io/mmm/.

  6. LK Losses: Direct Acceptance Rate Optimization for Speculative Decoding

    Speculative decoding accelerates autoregressive large language model (LLM) inference by using a lightweight draft model to propose candidate tokens that are then verified in parallel by the target model. The speedup is significantly determined by the acceptance rate, yet standard training minimizes Kullback-Leibler (KL) divergence as a proxy objective. While KL divergence and acceptance rate share the same global optimum, small draft models, having limited capacity, typically converge to suboptimal solutions where minimizing KL does not guarantee maximizing acceptance rate. To address this issue, we propose LK losses, special training objectives that directly target acceptance rate. Comprehensive experiments across four draft architectures and six target models, ranging from 8B to 685B parameters, demonstrate consistent improvements in acceptance metrics across all configurations compared to the standard KL-based training. We evaluate our approach on general, coding and math domains and report gains of up to 8-10% in average acceptance length. LK losses are easy to implement, introduce no computational overhead and can be directly integrated into any existing speculator training framework, making them a compelling alternative to the existing draft training objectives.

  7. CiteAudit: You Cited It, But Did You Read It? A Benchmark for Verifying Scientific References in the LLM Era

    Scientific research relies on accurate citation for attribution and integrity, yet large language models (LLMs) introduce a new risk: fabricated references that appear plausible but correspond to no real publications. Such hallucinated citations have already been observed in submissions and accepted papers at major machine learning venues, exposing vulnerabilities in peer review. Meanwhile, rapidly growing reference lists make manual verification impractical, and existing automated tools remain fragile to noisy and heterogeneous citation formats and lack standardized evaluation. We present the first comprehensive benchmark and detection framework for hallucinated citations in scientific writing. Our multi-agent verification pipeline decomposes citation checking into claim extraction, evidence retrieval, passage matching, reasoning, and calibrated judgment to assess whether a cited source truly supports its claim. We construct a large-scale human-validated dataset across domains and define unified metrics for citation faithfulness and evidence alignment. Experiments with state-of-the-art LLMs reveal substantial citation errors and show that our framework significantly outperforms prior methods in both accuracy and interpretability. This work provides the first scalable infrastructure for auditing citations in the LLM era and practical tools to improve the trustworthiness of scientific references.

  8. Compositional Generalization Requires Linear, Orthogonal Representations in Vision Embedding Models

    Compositional generalization, the ability to recognize familiar parts in novel contexts, is a defining property of intelligent systems. Although modern models are trained on massive datasets, they still cover only a tiny fraction of the combinatorial space of possible inputs, raising the question of what structure representations must have to support generalization to unseen combinations. We formalize three desiderata for compositional generalization under standard training (divisibility, transferability, stability) and show they impose necessary geometric constraints: representations must decompose linearly into per-concept components, and these components must be orthogonal across concepts. This provides theoretical grounding for the Linear Representation Hypothesis: the linear structure widely observed in neural representations is a necessary consequence of compositional generalization. We further derive dimension bounds linking the number of composable concepts to the embedding geometry. Empirically, we evaluate these predictions across modern vision models (CLIP, SigLIP, DINO) and find that representations exhibit partial linear factorization with low-rank, near-orthogonal per-concept factors, and that the degree of this structure correlates with compositional generalization on unseen combinations. As models continue to scale, these conditions predict the representational geometry they may converge to. Code is available at https://github.com/oshapio/necessary-compositionality.

  9. InfoNCE Induces Gaussian Distribution

    Contrastive learning has become a cornerstone of modern representation learning, allowing training with massive unlabeled data for both task-specific and general (foundation) models. A prototypical loss in contrastive training is InfoNCE and its variants. In this work, we show that the InfoNCE objective induces Gaussian structure in representations that emerge from contrastive training. We establish this result in two complementary regimes. First, we show that under certain alignment and concentration assumptions, projections of the high-dimensional representation asymptotically approach a multivariate Gaussian distribution. Next, under less strict assumptions, we show that adding a small asymptotically vanishing regularization term that promotes low feature norm and high feature entropy leads to similar asymptotic results. We support our analysis with experiments on synthetic and CIFAR-10 datasets across multiple encoder architectures and sizes, demonstrating consistent Gaussian behavior. This perspective provides a principled explanation for commonly observed Gaussianity in contrastive representations. The resulting Gaussian model enables principled analytical treatment of learned representations and is expected to support a wide range of applications in contrastive learning.

  10. Accelerating Masked Image Generation by Learning Latent Controlled Dynamics

    Masked Image Generation Models (MIGMs) have achieved great success, yet their efficiency is hampered by the multiple steps of bi-directional attention. In fact, there exists notable redundancy in their computation: when sampling discrete tokens, the rich semantics contained in the continuous features are lost. Some existing works attempt to cache the features to approximate future features. However, they exhibit considerable approximation error under aggressive acceleration rates. We attribute this to their limited expressivity and the failure to account for sampling information. To fill this gap, we propose to learn a lightweight model that incorporates both previous features and sampled tokens, and regresses the average velocity field of feature evolution. The model has moderate complexity that suffices to capture the subtle dynamics while keeping lightweight compared to the original base model. We apply our method, MIGM-Shortcut, to two representative MIGM architectures and tasks. In particular, on the state-of-the-art Lumina-DiMOO, it achieves over 4x acceleration of text-to-image generation while maintaining quality, significantly pushing the Pareto frontier of masked image generation. The code and model weights are available at https://github.com/Kaiwen-Zhu/MIGM-Shortcut.

  11. Ref-Adv: Exploring MLLM Visual Reasoning in Referring Expression Tasks

    Referring Expression Comprehension (REC) links language to region level visual perception. Standard benchmarks (RefCOCO, RefCOCO+, RefCOCOg) have progressed rapidly with multimodal LLMs but remain weak tests of visual reasoning and grounding: (i) many expressions are very short, leaving little reasoning demand; (ii) images often contain few distractors, making the target easy to find; and (iii) redundant descriptors enable shortcut solutions that bypass genuine text understanding and visual reasoning. We introduce Ref-Adv, a modern REC benchmark that suppresses shortcuts by pairing linguistically nontrivial expressions with only the information necessary to uniquely identify the target. The dataset contains referring expressions on real images, curated with hard distractors and annotated with reasoning facets including negation. We conduct comprehensive ablations (word order perturbations and descriptor deletion sufficiency) to show that solving Ref-Adv requires reasoning beyond simple cues, and we evaluate a broad suite of contemporary multimodal LLMs on Ref-Adv. Despite strong results on RefCOCO, RefCOCO+, and RefCOCOg, models drop markedly on Ref-Adv, revealing reliance on shortcuts and gaps in visual reasoning and grounding. We provide an in depth failure analysis and aim for Ref-Adv to guide future work on visual reasoning and grounding in MLLMs.

  12. Memory Caching: RNNs with Growing Memory

    Transformers have been established as the de-facto backbones for most recent advances in sequence modeling, mainly due to their growing memory capacity that scales with the context length. While plausible for retrieval tasks, it causes quadratic complexity and so has motivated recent studies to explore viable subquadratic recurrent alternatives. Despite showing promising preliminary results in diverse domains, such recurrent architectures underperform Transformers in recall-intensive tasks, often attributed to their fixed-size memory. In this paper, we introduce Memory Caching (MC), a simple yet effective technique that enhances recurrent models by caching checkpoints of their memory states (a.k.a. hidden states). Memory Caching allows the effective memory capacity of RNNs to grow with sequence length, offering a flexible trade-off that interpolates between the fixed memory (i.e., O(L) complexity) of RNNs and the growing memory (i.e., O(L^2) complexity) of Transformers. We propose four variants of MC, including gated aggregation and sparse selective mechanisms, and discuss their implications on both linear and deep memory modules. Our experimental results on language modeling, and long-context understanding tasks show that MC enhances the performance of recurrent models, supporting its effectiveness. The results of in-context recall tasks indicate that while Transformers achieve the best accuracy, our MC variants show competitive performance, close the gap with Transformers, and performs better than state-of-the-art recurrent models.

  13. LongVideo-R1: Smart Navigation for Low-cost Long Video Understanding

    This paper addresses the critical and underexplored challenge of long video understanding with low computational budgets. We propose LongVideo-R1, an active, reasoning-equipped multimodal large language model (MLLM) agent designed for efficient video context navigation, avoiding the redundancy of exhaustive search. At the core of LongVideo-R1 lies a reasoning module that leverages high-level visual cues to infer the most informative video clip for subsequent processing. During inference, the agent initiates traversal from top-level visual summaries and iteratively refines its focus, immediately halting the exploration process upon acquiring sufficient knowledge to answer the query. To facilitate training, we first extract hierarchical video captions from CGBench, a video corpus with grounding annotations, and guide GPT-5 to generate 33K high-quality chain-of-thought-with-tool trajectories. The LongVideo-R1 agent is fine-tuned upon the Qwen-3-8B model through a two-stage paradigm: supervised fine-tuning (SFT) followed by reinforcement learning (RL), where RL employs a specifically designed reward function to maximize selective and efficient clip navigation. Experiments on multiple long video benchmarks validate the effectiveness of name, which enjoys superior tradeoff between QA accuracy and efficiency. All curated data and source code are provided in the supplementary material and will be made publicly available. Code and data are available at: https://github.com/qiujihao19/LongVideo-R1

  14. SenCache: Accelerating Diffusion Model Inference via Sensitivity-Aware Caching

    Diffusion models achieve state-of-the-art video generation quality, but their inference remains expensive due to the large number of sequential denoising steps. This has motivated a growing line of research on accelerating diffusion inference. Among training-free acceleration methods, caching reduces computation by reusing previously computed model outputs across timesteps. Existing caching methods rely on heuristic criteria to choose cache/reuse timesteps and require extensive tuning. We address this limitation with a principled sensitivity-aware caching framework. Specifically, we formalize the caching error through an analysis of the model output sensitivity to perturbations in the denoising inputs, i.e., the noisy latent and the timestep, and show that this sensitivity is a key predictor of caching error. Based on this analysis, we propose Sensitivity-Aware Caching (SenCache), a dynamic caching policy that adaptively selects caching timesteps on a per-sample basis. Our framework provides a theoretical basis for adaptive caching, explains why prior empirical heuristics can be partially effective, and extends them to a dynamic, sample-specific approach. Experiments on Wan 2.1, CogVideoX, and LTX-Video show that SenCache achieves better visual quality than existing caching methods under similar computational budgets.

  15. Shared Nature, Unique Nurture: PRISM for Pluralistic Reasoning via In-context Structure Modeling

    Large Language Models (LLMs) are converging towards a singular Artificial Hivemind, where shared Nature (pre-training priors) result in a profound collapse of distributional diversity, limiting the distinct perspectives necessary for creative exploration and scientific discovery. To address this, we propose to equip models with inference-time Nurture (individualized epistemic trajectories) using Epistemic Evolution paradigm, progressing through explore, internalize, and express. We instantiate this via PRISM (Pluralistic Reasoning via In-context Structure Modeling), a model-agnostic system that augments LLM with dynamic On-the-fly Epistemic Graphs. On three creativity benchmarks, PRISM achieves state-of-the-art novelty and significantly expands distributional diversity. Moreover, we evaluate the real-world utility via a challenging rare-disease diagnosis benchmark. Results demonstrate that PRISM successfully uncovers correct long-tail diagnoses that standard LLM miss, confirming that its divergence stems from meaningful exploration rather than incoherent noise. Overall, this work establishes a new paradigm for Pluralistic AI, moving beyond monolithic consensus toward a diverse ecosystem of unique cognitive individuals capable of collective, multi-perspective discovery.

Techmeme(15)

  1. Filing: PayPay is seeking to raise up to $1.1B at a valuation of up to $13.4B in its US IPO, selling nearly 55M shares priced between $17 and $20 apiece (Arasu Kannagi Basil/Reuters)

    Arasu Kannagi Basil / Reuters : Filing: PayPay is seeking to raise up to $1.1B at a valuation of up to $13.4B in its US IPO, selling nearly 55M shares priced between $17 and $20 apiece —  PayPay and a selling shareholder are aiming to raise as much as $1.1 billion in an initial public offering in the United States …

  2. Source: Cursor's annualized revenue topped $2B in February, doubling from three months earlier, and about 60% of the revenue is coming from corporate customers (Rachel Metz/Bloomberg)

    Rachel Metz / Bloomberg : Source: Cursor's annualized revenue topped $2B in February, doubling from three months earlier, and about 60% of the revenue is coming from corporate customers —  Cursor's annualized revenue topped $2 billion in February, according to a person familiar with the matter …

  3. The US Treasury Department, State Department, and federal housing agency are ending use of Anthropic products; State Department says it will switch to OpenAI (Reuters)

    Reuters : The US Treasury Department, State Department, and federal housing agency are ending use of Anthropic products; State Department says it will switch to OpenAI —  The U.S. Treasury Department, State Department and the federal housing agency are terminating all use of Anthropic products …

  4. Sources: US considers limiting Chinese companies to 75K Nvidia H200 chips each, less than half of what some want to buy; AMD MI325 chips also count toward a cap (Bloomberg)

    Bloomberg : Sources: US considers limiting Chinese companies to 75K Nvidia H200 chips each, less than half of what some want to buy; AMD MI325 chips also count toward a cap —  US officials are considering caps on the number of AI accelerators Nvidia Corp. can export to any one Chinese company …

  5. Iranians turn to Starlink, decentralized messaging apps, and VPNs to circumvent an internet blackout; NetBlocks says connectivity is at 1% of ordinary levels (Bloomberg)

    Bloomberg : Iranians turn to Starlink, decentralized messaging apps, and VPNs to circumvent an internet blackout; NetBlocks says connectivity is at 1% of ordinary levels —  Iranians are finding ways to circumvent a fresh internet blackout imposed by their government and are sharing footage of US and Israeli airstrikes with the world.

  6. New York-based Ease Health, which is building an AI-native OS for behavioral health providers, emerged from stealth and raised a $41M Series A led by a16z (Vignesh R/Tech Funding News)

    Vignesh R / Tech Funding News : New York-based Ease Health, which is building an AI-native OS for behavioral health providers, emerged from stealth and raised a $41M Series A led by a16z —  Behavioural health providers across the US are struggling with outdated software.  Most clinics use separate systems for admissions, patient records, and billing.

  7. Alibaba releases the open-weight Qwen3.5 Small Model Series in 0.8B, 2B, 4B, and 9B sizes, claiming the 9B model rivals OpenAI's gpt-oss-120b on some benchmarks (Carl Franzen/VentureBeat)

    Carl Franzen / VentureBeat : Alibaba releases the open-weight Qwen3.5 Small Model Series in 0.8B, 2B, 4B, and 9B sizes, claiming the 9B model rivals OpenAI's gpt-oss-120b on some benchmarks —  Earlier today, e-commerce giant Alibaba's Qwen Team of AI researchers, focused primarily on developing and releasing to the world …

  8. Sources: SoftBank-owned Japanese payments app PayPay delays its IPO roadshow that was scheduled to launch Monday, as markets were rattled by the strikes in Iran (Echo Wang/Reuters)

    Echo Wang / Reuters : Sources: SoftBank-owned Japanese payments app PayPay delays its IPO roadshow that was scheduled to launch Monday, as markets were rattled by the strikes in Iran —  SoftBank's (9984.T) PayPay has delayed its highly anticipated IPO roadshow that was scheduled to launch Monday as markets …

  9. ByteDance shares details about its "Project Swan" headset, set to launch this year with a 4,000-PPI micro-OLED display, and unveils Pico OS 6, its new XR OS (David Heaney/UploadVR)

    David Heaney / UploadVR : ByteDance shares details about its “Project Swan” headset, set to launch this year with a 4,000-PPI micro-OLED display, and unveils Pico OS 6, its new XR OS —  ByteDance's Pico announced the key display and compute specs of its “Project Swan” headset, coming later this year …

  10. A profile of Palmer Luckey and his startup Anduril, which has more than $6B in global contracts, had roughly $2B in revenue last year, and is valued at ~$31B (New York Times)

    New York Times : A profile of Palmer Luckey and his startup Anduril, which has more than $6B in global contracts, had roughly $2B in revenue last year, and is valued at ~$31B —  Within minutes of arriving at a security conference with U.S. defense and intelligence officials in December …

  11. Anthropic's $60B+ in funding, half of which came just last month, from over 200 investors is now at risk due to the company's contract dispute with the Pentagon (Dan Primack/Axios)

    Dan Primack / Axios : Anthropic's $60B+ in funding, half of which came just last month, from over 200 investors is now at risk due to the company's contract dispute with the Pentagon —  Anthropic has raised more than $60 billion from over 200 “venture capital” investors, half of which came just last month.

  12. Sources: Reflection AI, which is developing open foundation models, seeks to raise $2B+ at a $20B+ valuation, after raising $2B at an $8B valuation in October (Financial Times)

    Financial Times : Sources: Reflection AI, which is developing open foundation models, seeks to raise $2B+ at a $20B+ valuation, after raising $2B at an $8B valuation in October —  Reflection AI's new funding talks come as Trump administration seeks US rivals to China's DeepSeek

  13. X announces a "Paid Partnership" label that creators can apply to their posts to indicate they're ads; until now, creators relied on hashtags to label posts (Sarah Perez/TechCrunch)

    Sarah Perez / TechCrunch : X announces a “Paid Partnership” label that creators can apply to their posts to indicate they're ads; until now, creators relied on hashtags to label posts —  Social network X on Monday announced the introduction of a new “Paid Partnership” label that creators can apply to their posts to indicate they're advertisements.

  14. SCOTUS declines to hear a dispute over copyrights for AI-generated material, in a case where a computer scientist was denied a copyright for AI-generated art (Blake Brittain/Reuters)

    Blake Brittain / Reuters : SCOTUS declines to hear a dispute over copyrights for AI-generated material, in a case where a computer scientist was denied a copyright for AI-generated art —  The U.S. Supreme Court declined on Monday to take up the issue of whether art generated by artificial intelligence …

  15. The Anthropic-DOD skirmish is the first major public debate on control over frontier AI, and institutions behaved erratically, maliciously, and without clarity (Dean W. Ball/Hyperdimensional)

    Dean W. Ball / Hyperdimensional : The Anthropic-DOD skirmish is the first major public debate on control over frontier AI, and institutions behaved erratically, maliciously, and without clarity —  On Anthropic and the Department of War  —  I.  —  A little more than a decade ago, I sat with my father and watched him die.

Solidot(15)

  1. NIST 限制外国科学家进入其实验室

    过去几周在美国国家标准与技术研究院(NIST)工作的数百名外国科学家被限制进入实验室,除非有联邦雇员陪同,否则不得在晚上和周末进入实验室。部分国家的科学家最早将在本月底失去访问权限。拟议中的规则尚无书面版本,仅通过会议传达。最新变化是基于 NIST 在 2025 年更新的研究安全规则,它将中国、俄罗斯、伊朗、朝鲜、古巴、委内瑞拉和叙利亚的科学家视为“高风险”人群,中国等国的研究人员已被告知,他们的实验室访问权限将于 3 月 31 日前接受审查,如果在 NIST 工作逾 3 年或从事量子技术或 AI 等敏感项目而构成“高风险”,其访问权限将被终止。低风险国家的研究人员也面临从 9 月或 12 月起失去访问权限。NIST 研究人员不从事机密研究,NIST 前主任 Patrick Gallagher 表示看不出这么做会带来什么安全上的好处。

  2. 亚马逊 AWS 中东数据中心遭遇火灾和断电

    亚马逊 AWS 披露其位于中东数据中心的一处遭遇断电一处遭遇“物体”撞击后起火。它没有披露是什么东西撞击了数据中心设施。AWS 表示它位于阿联酋的一个数据中心于 7:30 a.m. ET 遭遇撞击,撞击产生了火花和火灾,消防部门在灭火过程中切断了数据中心和发电机的电源。

  3. 为何女性的疼痛持续时间更长

    医生通常认为免疫系统会通过引起炎症加剧疼痛,而炎症通常表现为红肿。最新研究显示,免疫细胞在帮助缓解疼痛方面也可能至关重要,男女之间的免疫细胞功能差异可能会影响疼痛消退的速度。研究人员调查了名为 IL-10(interleukin-10)的分子,测量了小鼠皮肤损伤后和交通事故急诊伤者体内的 IL-10 水平,发现 IL-10 的作用不仅是缓解炎症,还能直接与疼痛感受神经细胞通信将其关闭。也就是 IL-10 有助于消除疼痛。IL-10 由免疫系统的一种白细胞单核细胞(Monocyte)产生,这种细胞会在血液中循环转移到受伤组织。在男性体内,单核细胞更容易产生 IL-10,而在女性体内不太明显。原因是睾酮会影响单核细胞产生的 IL-10 的数量,而男性体内有更高的睾酮水平。

  4. 小鼠研究发现器官同步衰老但存在性别差异

    研究人员构建了迄今最详尽的图谱,展示了衰老如何影响21种哺乳动物组织中的数千种细胞亚型。他们通过分析不同年龄段小鼠的近 700 万个单细胞,确定了随时间推移最易受损的细胞,以及促使其衰老的因素。研究人员对 32 只小鼠21个器官中提取的数百万个单细胞进行分析。这些小鼠处于 3 个年龄段:1 个月(年轻成年小鼠)、5 个月(中年小鼠)、21 个月(老年小鼠)。研究人员识别出了超过 1800 种不同的细胞亚型,其中包括许多此前未被完整描述过的罕见类型。随后研究团队追踪了各年龄阶段小鼠不同类型细胞的数量变化情况。数十年来,科学家们一直认为衰老主要改变的是细胞功能,而非细胞数量。而研究团队的分析结果对这一观点提出了挑战。他们发现,约 1/4 的细胞类型在数量上随时间推移发生了显著变化,比如某些肌肉细胞和肾细胞的数量大幅减少,而免疫细胞数量则大幅增加。这些变化在不同器官的细胞中具有同步性,相似的细胞状态在不同器官中几乎同时出现和消失。这种模式表明,血液中循环中的共同信号可能有助协调全身的衰老过程。大约 40% 与衰老相关的改变因性别而异。例如女性衰老过程中,表现出更广泛的免疫激活情况。

  5. 2026 年 2 月 Steam 统计显示简体中文用户占逾半数份额

    Valve 公布了 2026 年 2 月的 Steam 硬件和软件调查。数据显示了一个异常:简体中文用户出现爆炸式增长,单月增长 30.74% 至 54.60%,英文用户减少 14.74% 至 22.27%。另一个异常是 Windows 11 下降 10.43% 占 56.28%,64 位 Windows 10 增长 12.46% 至 40.25%。异常现象的一个解释也许是 2 月中旬是大部分简体中文用户的春节假期。其它数据包括:Windows 操作系统份额增长 1.99% 至 96.61%,Linux 下降 1.15% 占 2.23%,macOS 下降 0.85% 占 1.16%。

  6. 摩托罗拉手机宣布与 GrapheneOS 合作

    联想旗下的摩托罗拉手机宣布与 Android 安全加固社区发行版项目 GrapheneOS 展开合作。GrapheneOS 此前主要支持 Google 的 Pixel 系列手机,但摩托罗拉哪些型号的手机会支持 GrapheneOS 官方新闻稿没有给出信息,只是表示未来会公布。摩托罗拉同时推出了企业级分析平台 Moto Analytics,Moto Secure 支持私密图像数据功能,启用后会自动从设备上所有新拍摄的图像中移除敏感元数据如位置。

  7. 古代天文学著作发现伽利略的手写笔记

    历史学家 Ivan Malara 在意大利佛罗伦萨国家中央图书馆翻阅 16 世纪印刷的天文著作——托勒密的《天文学大成》。该书的地心说模型统治西方天文学长达 14 个世纪。他在翻阅时注意到了书上留下的大量批注,其笔迹非常类似意大利 Tuscan 天文学家伽利略。这一发现为天文学从地心说到日心说的转变提供了新的见解。伽利略的笔记可能写于 1590 年左右,也就是在他使用望远镜观测月球和木星前 20 年,揭示了他对托勒密著作既崇敬又批判性剖析的态度。

  8. 长期海洋暖化会导致海洋生物数量锐减

    根据发表在《Nature Ecology & Evolution》期刊上的一项研究,长期海洋变暖会导致海洋生物数量锐减,每十年升温 0.1C,鱼类数量会下降 7.2%。研究人员分析了1993-2021 年间北半球 33000 个鱼类种群的逐年变化,区分海床升温的十年均速与海洋热浪等短期变化的影响。他们发现长期升温导致的生物量下降幅度在一年内最高达到 19.8%。海洋生物易受化石燃料污染导致的大气温度变化影响,科学家对此反复发出警告。

  9. 微软官方 Copilot Discord 服务器封禁 Microslop,用户创造变体迫使微软锁定服务器

    微软 CEO 纳德拉(Satya Nadella)关于 AI 的评论导致网民为微软起了 Microslop 的绰号,绰号的流行和随处可见促使微软在官方 Copilot Discord 服务器将其封禁,用户输入 Microslop 后会收到警告称根据服务器规定其输入包含了不合适的短语。但用户很快找到了应对之策,创造了无数 Microslop 的变体,比如用数字“0”代替字母“o”的“Microsl0p”。在猫与老鼠的文字游戏中,微软显然是失败的一方,它被迫锁定了 Copilot Discord 服务器。

  10. Linux 项目延长 LTS 版本的支持时间

    稳定版内核维护者 Greg Kroah-Hartman 和 Sasha Levin 宣布延长多个 LTS(长期支持)版本的支持时间:Linux 6.6 的终止支持时间从目前的 2026 年 12 月延长到 2027 年 12 月,支持时间为四年;Linux 6.12 的终止支持时间从 2026 年 12 月延长到 2028 年 12 月;Linux 6.18 的终止支持时间从 2027 年 12 月延长到 2028 年 12 月,还有三年支持时间。Linux 5.10 和 5.15 都将于今年 12 月停止支持,使用这些内核版本的用户是时候考虑升级了。

  11. 美国人听播客的比例超过了谈话广播

    根据 Edison Research 的《Share of Ear》调查,播客超过 AM/FM 谈话广播成为美国最受欢迎的谈话类音频内容媒介。过去十年用户收听播客的比例一直在增加,收听谈话广播(talk radio)的比例一直在缩小,今年播客的收听时长首次超过了谈话广播,达到了 40%,而广播是 39%。调查涵盖的播客包括了视频播客,此前 YouTube 表示 2025 年用户每月在电视等客厅设备上观看播客的时长达到了 7 亿小时,而前一年是 4 亿小时。

  12. 美国加州和科罗拉多州计划要求在操作系统层级验证用户年龄

    美国多个州都要求成人网站验证访客年龄,但验证年龄的常用方法如扫描脸部或提供身份证件都存在泄露隐私的问题。加州以及科罗拉多州计划要求在操作系统层级验证年龄,然后通过 API 与应用共享。加州去年通过了 AB 1043 法案,要求操作系统开发商创建一种让设备所有者注册其年龄段的方法,该法律将于 2027 年 1 月 1 日生效。科罗拉多州的议员提出了类似的法案 SB26-051,该法案的共同提出者参议员 Matt Ball 表示,他们的目的通过一个以注重隐私的年龄验证框架为儿童的网络安全提供周全的保障。

  13. Anthropic 的 Claude 在苹果美国区免费应用榜跃居第一

    本周五,美国总统特朗普下令联邦机构立刻停用 Anthropic 的 Claude 助手,原因是 Anthropic 在安全原则上坚守其立场。相比之下,其竞争对手 OpenAI 看起来完全没有任何立场,此举在美国用户中间引发了卸载 OpenAI 的 ChatGPT 安装 Claude 的热潮,这一趋势推动 Claude 周六跃居苹果 App Store 美国区免费应用榜榜首,超过了 ChatGPT,ChatGPT 屈居第二,Google 的 Gemini 排名第四。根据分析公司 Sensor Tower 的数据,一个月前的 1 月 30 日 Claude 还排在排行榜的第 131 名,2 月的大部分时间徘徊在前 20 名左右,而 ChatGPT 通常是第一名。Anthropic 还发布了记忆导入功能,方便 ChatGPT 用户改用 Claude。

  14. 当你需要帮助时狗的反应类似 2 岁小孩但猫只会旁观

    根据发表在《Animal Behaviour》期刊上的一项研究,匈牙利研究人员对比了人在需要帮助时 18-24 个月的幼儿、以及宠物狗和猫的反应。结果显示,狗的自发性亲社会行为与幼儿类似,而猫则是冷眼旁观。在实验中,熟人如父母或主人假装在寻找一个藏起来的东西,四分之三的情况下狗和幼儿会提供帮助。猫只有在符合自身利益时才会参与进来提供帮助。

  15. Metacritic 不收录 AI 生成的评测

    Metacritic 是一家聚合电影、电视、音乐专辑、游戏评测及其评分的网站,它会基于相关评分给出一个加权平均值。Metacritic 成立于 2001 年,至今有 25 年历史,它的综合评分获得了广泛认可。网站联合创始人 Marc Doyle 在一份声明中表示 Metacritic 不会收录 AI 生成的评测。在这之前,一家叫 Videogamer.com 的英国老牌游戏网站(曾经很受欢迎)在被博彩公司收购之后解雇了大部分员工,然后用 AI 生成了新游戏的评测和评分。该网站评测作者的肖像是用 ChatGPT 生成的。Metacritic 在获悉之后删除了该网站的评测。