DIGEST · 2026-03-10

OrangeBot.AI Digest — 2026-03-10

88 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. After outages, Amazon to make senior engineers sign off on AI-assisted changes (arstechnica.com)
  2. Agents that run while I sleep (www.claudecodecamp.com)
  3. Yann LeCun raises $1B to build AI that understands the physical world (www.wired.com)
  4. Amazon is holding a mandatory meeting about AI breaking its systems (twitter.com)
  5. Debian decides not to decide on AI-generated contributions (lwn.net)
  6. Tony Hoare has died (blog.computationalcomplexity.org)
  7. Meta acquires Moltbook (www.axios.com)
  8. Show HN: How I Topped the HuggingFace Open LLM Leaderboard on Two Gaming GPUs (dnhkng.github.io)
  9. Rebasing in Magit (entropicthoughts.com)
  10. Intel Demos Chip to Compute with Encrypted Data (spectrum.ieee.org)
  11. Online age-verification tools for child safety are surveilling adults (www.cnbc.com)
  12. I put my whole life into a single database (howisfelix.today)
  13. Yann LeCun's AI startup raises $1B in Europe's largest ever seed round (www.ft.com)
  14. The Gervais Principle, or the Office According to “The Office” (2009) (www.ribbonfarm.com)
  15. Redox OS has adopted a Certificate of Origin policy and a strict no-LLM policy (gitlab.redox-os.org)

GitHub Trending(13)

  1. msitarzewski / agency-agents
  2. 666ghj / MiroFish
  3. NousResearch / hermes-agent
  4. promptfoo / promptfoo
  5. GoogleCloudPlatform / generative-ai
  6. virattt / ai-hedge-fund
  7. karpathy / nanochat
  8. obra / superpowers
  9. alibaba / page-agent
  10. sepinf-inc / IPED
  11. openclaw / openclaw
  12. pbakaus / impeccable
  13. bytedance / deer-flow

Product Hunt(15)

  1. Crikket

    The open source bug reporting and feedback tool

  2. Book Reading Habit

    Finally read the books you buy

  3. On Demand Ads by beehiv

    Premium sponsors, ready when you are

  4. Agent Skills

    Find skills for Claude Code, Cursor, Copilot & more

  5. VENTUNO Q

    Dual-brain edge AI computer by Qualcomm and Arduino

  6. Shipper 2.0

    Build web/mobile apps, sites and extensions by talking to AI

  7. GapHunt

    Find product gaps & build from bad reviews

  8. Pulse

    Lightweight real-time polls - open source & self-hosted

  9. Your Next Store

    AI-first platform for building commerce stores, fast

  10. Chronicle 2.0

    AI presentations without the AI slop

  11. Spine Swarm

    Manage a team of AI agents that do real work

  12. Claude Code Review

    Multi-agent review catching bugs early in AI-generated code

  13. humans fix ai

    Real developers help vibecoders with AI-built apps

  14. Fish Audio S2

    Real Expressive AI Voices

  15. Visual Translate by Vozo

    Translate text in your videos without recreating visuals

Hugging Face(15)

  1. Lost in Stories: Consistency Bugs in Long Story Generation by LLMs

    What happens when a storyteller forgets its own story? Large Language Models (LLMs) can now generate narratives spanning tens of thousands of words, but they often fail to maintain consistency throughout. When generating long-form narratives, these models can contradict their own established facts, character traits, and world rules. Existing story generation benchmarks focus mainly on plot quality and fluency, leaving consistency errors largely unexplored. To address this gap, we present ConStory-Bench, a benchmark designed to evaluate narrative consistency in long-form story generation. It contains 2,000 prompts across four task scenarios and defines a taxonomy of five error categories with 19 fine-grained subtypes. We also develop ConStory-Checker, an automated pipeline that detects contradictions and grounds each judgment in explicit textual evidence. Evaluating a range of LLMs through five research questions, we find that consistency errors show clear tendencies: they are most common in factual and temporal dimensions, tend to appear around the middle of narratives, occur in text segments with higher token-level entropy, and certain error types tend to co-occur. These findings can inform future efforts to improve consistency in long-form narrative generation. Our project page is available at https://picrew.github.io/constory-bench.github.io/.

  2. Holi-Spatial: Evolving Video Streams into Holistic 3D Spatial Intelligence

    The pursuit of spatial intelligence fundamentally relies on access to large-scale, fine-grained 3D data. However, existing approaches predominantly construct spatial understanding benchmarks by generating question-answer (QA) pairs from a limited number of manually annotated datasets, rather than systematically annotating new large-scale 3D scenes from raw web data. As a result, their scalability is severely constrained, and model performance is further hindered by domain gaps inherent in these narrowly curated datasets. In this work, we propose Holi-Spatial, the first fully automated, large-scale, spatially-aware multimodal dataset, constructed from raw video inputs without human intervention, using the proposed data curation pipeline. Holi-Spatial supports multi-level spatial supervision, ranging from geometrically accurate 3D Gaussian Splatting (3DGS) reconstructions with rendered depth maps to object-level and relational semantic annotations, together with corresponding spatial Question-Answer (QA) pairs. Following a principled and systematic pipeline, we further construct Holi-Spatial-4M, the first large-scale, high-quality 3D semantic dataset, containing 12K optimized 3DGS scenes, 1.3M 2D masks, 320K 3D bounding boxes, 320K instance captions, 1.2M 3D grounding instances, and 1.2M spatial QA pairs spanning diverse geometric, relational, and semantic reasoning tasks. Holi-Spatial demonstrates exceptional performance in data curation quality, significantly outperforming existing feed-forward and per-scene optimized methods on datasets such as ScanNet, ScanNet++, and DL3DV. Furthermore, fine-tuning Vision-Language Models (VLMs) on spatial reasoning tasks using this dataset has also led to substantial improvements in model performance.

  3. LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory

    Feedforward geometric foundation models achieve strong short-window reconstruction, yet scaling them to minutes-long videos is bottlenecked by quadratic attention complexity or limited effective memory in recurrent designs. We present LoGeR (Long-context Geometric Reconstruction), a novel architecture that scales dense 3D reconstruction to extremely long sequences without post-optimization. LoGeR processes video streams in chunks, leveraging strong bidirectional priors for high-fidelity intra-chunk reasoning. To manage the critical challenge of coherence across chunk boundaries, we propose a learning-based hybrid memory module. This dual-component system combines a parametric Test-Time Training (TTT) memory to anchor the global coordinate frame and prevent scale drift, alongside a non-parametric Sliding Window Attention (SWA) mechanism to preserve uncompressed context for high-precision adjacent alignment. Remarkably, this memory architecture enables LoGeR to be trained on sequences of 128 frames, and generalize up to thousands of frames during inference. Evaluated across standard benchmarks and a newly repurposed VBR dataset with sequences of up to 19k frames, LoGeR substantially outperforms prior state-of-the-art feedforward methods--reducing ATE on KITTI by over 74%--and achieves robust, globally consistent reconstruction over unprecedented horizons.

  4. Believe Your Model: Distribution-Guided Confidence Calibration

    Large Reasoning Models have demonstrated remarkable performance with the advancement of test-time scaling techniques, which enhances prediction accuracy by generating multiple candidate responses and selecting the most reliable answer. While prior work has analyzed that internal model signals like confidence scores can partly indicate response correctness and exhibit a distributional correlation with accuracy, such distributional information has not been fully utilized to guide answer selection. Motivated by this, we propose DistriVoting, which incorporates distributional priors as another signal alongside confidence during voting. Specifically, our method (1) first decomposes the mixed confidence distribution into positive and negative components using Gaussian Mixture Models, (2) then applies a reject filter based on positive/negative samples from them to mitigate overlap between the two distributions. Besides, to further alleviate the overlap from the perspective of distribution itself, we propose SelfStepConf, which uses step-level confidence to dynamically adjust inference process, increasing the separation between the two distributions to improve the reliability of confidences in voting. Experiments across 16 models and 5 benchmarks demonstrate that our method significantly outperforms state-of-the-art approaches.

  5. How Far Can Unsupervised RLVR Scale LLM Training?

    Unsupervised reinforcement learning with verifiable rewards (URLVR) offers a pathway to scale LLM training beyond the supervision bottleneck by deriving rewards without ground truth labels. Recent works leverage model intrinsic signals, showing promising early gains, yet their potential and limitations remain unclear. In this work, we revisit URLVR and provide a comprehensive analysis spanning taxonomy, theory and extensive experiments. We first classify URLVR methods into intrinsic versus external based on reward sources, then establish a unified theoretical framework revealing that all intrinsic methods converge toward sharpening the model's initial distribution This sharpening mechanism succeeds when initial confidence aligns with correctness but fails catastrophically when misaligned. Through systematic experiments, we show intrinsic rewards consistently follow a rise-then-fall pattern across methods, with collapse timing determined by model prior rather than engineering choices. Despite these scaling limits, we find intrinsic rewards remain valuable in test-time training on small datasets, and propose Model Collapse Step to measure model prior, serving as a practical indicator for RL trainability. Finally, we explore external reward methods that ground verification in computational asymmetries, showing preliminary evidence they may escape the confidence-correctness ceiling. Our findings chart boundaries for intrinsic URLVR while motivating paths toward scalable alternatives.

  6. CARE-Edit: Condition-Aware Routing of Experts for Contextual Image Editing

    Unified diffusion editors often rely on a fixed, shared backbone for diverse tasks, suffering from task interference and poor adaptation to heterogeneous demands (e.g., local vs global, semantic vs photometric). In particular, prevalent ControlNet and OmniControl variants combine multiple conditioning signals (e.g., text, mask, reference) via static concatenation or additive adapters which cannot dynamically prioritize or suppress conflicting modalities, thus resulting in artifacts like color bleeding across mask boundaries, identity or style drift, and unpredictable behavior under multi-condition inputs. To address this, we propose Condition-Aware Routing of Experts (CARE-Edit) that aligns model computation with specific editing competencies. At its core, a lightweight latent-attention router assigns encoded diffusion tokens to four specialized experts--Text, Mask, Reference, and Base--based on multi-modal conditions and diffusion timesteps: (i) a Mask Repaint module first refines coarse user-defined masks for precise spatial guidance; (ii) the router applies sparse top-K selection to dynamically allocate computation to the most relevant experts; (iii) a Latent Mixture module subsequently fuses expert outputs, coherently integrating semantic, spatial, and stylistic information to the base images. Experiments validate CARE-Edit's strong performance on contextual editing tasks, including erasure, replacement, text-driven edits, and style transfer. Empirical analysis further reveals task-specific behavior of specialized experts, showcasing the importance of dynamic, condition-aware processing to mitigate multi-condition conflicts.

  7. CoCo: Code as CoT for Text-to-Image Preview and Rare Concept Generation

    Recent advancements in Unified Multimodal Models (UMMs) have significantly advanced text-to-image (T2I) generation, particularly through the integration of Chain-of-Thought (CoT) reasoning. However, existing CoT-based T2I methods largely rely on abstract natural-language planning, which lacks the precision required for complex spatial layouts, structured visual elements, and dense textual content. In this work, we propose CoCo (Code-as-CoT), a code-driven reasoning framework that represents the reasoning process as executable code, enabling explicit and verifiable intermediate planning for image generation. Given a text prompt, CoCo first generates executable code that specifies the structural layout of the scene, which is then executed in a sandboxed environment to render a deterministic draft image. The model subsequently refines this draft through fine-grained image editing to produce the final high-fidelity result. To support this training paradigm, we construct CoCo-10K, a curated dataset containing structured draft-final image pairs designed to teach both structured draft construction and corrective visual refinement. Empirical evaluations on StructT2IBench, OneIG-Bench, and LongText-Bench show that CoCo achieves improvements of +68.83%, +54.8%, and +41.23% over direct generation, while also outperforming other generation methods empowered by CoT. These results demonstrate that executable code is an effective and reliable reasoning paradigm for precise, controllable, and structured text-to-image generation. The code is available at: https://github.com/micky-li-hd/CoCo

  8. HiAR: Efficient Autoregressive Long Video Generation via Hierarchical Denoising

    Autoregressive (AR) diffusion offers a promising framework for generating videos of theoretically infinite length. However, a major challenge is maintaining temporal continuity while preventing the progressive quality degradation caused by error accumulation. To ensure continuity, existing methods typically condition on highly denoised contexts; yet, this practice propagates prediction errors with high certainty, thereby exacerbating degradation. In this paper, we argue that a highly clean context is unnecessary. Drawing inspiration from bidirectional diffusion models, which denoise frames at a shared noise level while maintaining coherence, we propose that conditioning on context at the same noise level as the current block provides sufficient signal for temporal consistency while effectively mitigating error propagation. Building on this insight, we propose HiAR, a hierarchical denoising framework that reverses the conventional generation order: instead of completing each block sequentially, it performs causal generation across all blocks at every denoising step, so that each block is always conditioned on context at the same noise level. This hierarchy naturally admits pipelined parallel inference, yielding a 1.8 wall-clock speedup in our 4-step setting. We further observe that self-rollout distillation under this paradigm amplifies a low-motion shortcut inherent to the mode-seeking reverse-KL objective. To counteract this, we introduce a forward-KL regulariser in bidirectional-attention mode, which preserves motion diversity for causal inference without interfering with the distillation loss. On VBench (20s generation), HiAR achieves the best overall score and the lowest temporal drift among all compared methods.

  9. \$OneMillion-Bench: How Far are Language Agents from Human Experts?

    As language models (LMs) evolve from chat assistants to long-horizon agents capable of multi-step reasoning and tool use, existing benchmarks remain largely confined to structured or exam-style tasks that fall short of real-world professional demands. To this end, we introduce \OneMillion-Bench OneMillion-Bench, a benchmark of 400 expert-curated tasks spanning Law, Finance, Industry, Healthcare, and Natural Science, built to evaluate agents across economically consequential scenarios. Unlike prior work, the benchmark requires retrieving authoritative sources, resolving conflicting evidence, applying domain-specific rules, and making constraint decisions, where correctness depends as much on the reasoning process as the final answer. We adopt a rubric-based evaluation protocol scoring factual accuracy, logical coherence, practical feasibility, and professional compliance, focused on expert-level problems to ensure meaningful differentiation across agents. Together, \$OneMillion-Bench provides a unified testbed for assessing agentic reliability, professional depth, and practical readiness in domain-intensive scenarios.

  10. NLE: Non-autoregressive LLM-based ASR by Transcript Editing

    While autoregressive (AR) LLM-based ASR systems achieve strong accuracy, their sequential decoding limits parallelism and incurs high latency. We propose NLE, a non-autoregressive (NAR) approach that formulates speech recognition as conditional transcript editing, enabling fully parallel prediction. NLE extracts acoustic embeddings and an initial hypothesis from a pretrained speech encoder, then refines the hypothesis using a bidirectional LLM editor trained with a latent alignment objective. An interleaved padding strategy exploits the identity mapping bias of Transformers, allowing the model to focus on corrections rather than full reconstruction. On the Open ASR leaderboard, NLE++ achieves 5.67% average WER with an RTFx (inverse real-time factor) of 1630. In single-utterance scenarios, NLE achieves 27x speedup over the AR baseline, making it suitable for real-time applications.

  11. TDM-R1: Reinforcing Few-Step Diffusion Models with Non-Differentiable Reward

    While few-step generative models have enabled powerful image and video generation at significantly lower cost, generic reinforcement learning (RL) paradigms for few-step models remain an unsolved problem. Existing RL approaches for few-step diffusion models strongly rely on back-propagating through differentiable reward models, thereby excluding the majority of important real-world reward signals, e.g., non-differentiable rewards such as humans' binary likeness, object counts, etc. To properly incorporate non-differentiable rewards to improve few-step generative models, we introduce TDM-R1, a novel reinforcement learning paradigm built upon a leading few-step model, Trajectory Distribution Matching (TDM). TDM-R1 decouples the learning process into surrogate reward learning and generator learning. Furthermore, we developed practical methods to obtain per-step reward signals along the deterministic generation trajectory of TDM, resulting in a unified RL post-training method that significantly improves few-step models' ability with generic rewards. We conduct extensive experiments ranging from text-rendering, visual quality, and preference alignment. All results demonstrate that TDM-R1 is a powerful reinforcement learning paradigm for few-step text-to-image models, achieving state-of-the-art reinforcement learning performances on both in-domain and out-of-domain metrics. Furthermore, TDM-R1 also scales effectively to the recent strong Z-Image model, consistently outperforming both its 100-NFE and few-step variants with only 4 NFEs. Project page: https://github.com/Luo-Yihong/TDM-R1

  12. Scaling Agentic Capabilities, Not Context: Efficient Reinforcement Finetuning for Large Toolspaces

    Agentic systems operating over large tool ecosystems must plan and execute long-horizon workflows under weak or non-verifiable supervision. While frontier models mitigate these challenges through scale and large context budgets, small language models (SLMs) remain brittle: eager tool loading saturates context, execution errors compound over time, and sparse rewards limit learning. We introduce ATLAS, a reinforcement finetuning framework that enables SLMs to operate effectively in large-scale toolspace environments by learning how to acquire context and how to execute actions. Our approach makes two key contributions. First, we treat context control and execution structure as learnable decisions, combining iterative tool loading with programmatic tool orchestration to bound context growth and stabilize long-horizon trajectories. Second, we propose rubric-based reinforcement finetuning, which decomposes task success into structured, task-aligned criteria and enables scalable training using small judge models. Across MCP benchmarks, these design choices yield large and consistent gains over generic RL baselines, allowing a 4B SLM to approach frontier-agent performance under far tighter parameter and context budgets.

  13. Training-free Latent Inter-Frame Pruning with Attention Recovery

    Current video generation models suffer from high computational latency, making real-time applications prohibitively costly. In this paper, we address this limitation by exploiting the temporal redundancy inherent in video latent patches. To this end, we propose the Latent Inter-frame Pruning with Attention Recovery (LIPAR) framework, which detects and skips recomputing duplicated latent patches. Additionally, we introduce a novel Attention Recovery mechanism that approximates the attention values of pruned tokens, thereby removing visual artifacts arising from naively applying the pruning method. Empirically, our method increases video editing throughput by 1.45times, on average achieving 12.2 FPS on an NVIDIA A6000 compared to the baseline 8.4 FPS. The proposed method does not compromise generation quality and can be seamlessly integrated with the model without additional training. Our approach effectively bridges the gap between traditional compression algorithms and modern generative pipelines.

  14. Unlocking Data Value in Finance: A Study on Distillation and Difficulty-Aware Training

    Large Language Models (LLMs) have demonstrated strong general capabilities, yet their deployment in finance remains challenging due to dense domain-specific terminology, stringent numerical reasoning requirements, and low tolerance for factual errors. We conduct a controlled empirical study showing that in specialized vertical domains, performance is largely determined by the quality and difficulty/verifiability profile of post-training data. We introduce ODA-Fin-SFT-318k, constructed via multi-stage distillation and verification to produce high-quality Chain-of-Thought supervision, and ODA-Fin-RL-12k, curated for hard-but-verifiable tasks that balance reward precision and task diversity. Using standard SFT and RL pipelines, we show that high-quality CoT distillation establishes a robust foundation during SFT, while difficulty- and verifiability-aware sampling improves RL generalization. Evaluated on nine benchmarks spanning general financial tasks, sentiment analysis, and numerical reasoning, our ODA-Fin-RL-8B consistently surpasses open-source state-of-the-art (SOTA) financial LLMs of comparable size. We release our ODA-Fin-SFT-318k and ODA-Fin-RL-12k datasets, along with trained models to advance data-centric financial AI research.

  15. Concept-Guided Fine-Tuning: Steering ViTs away from Spurious Correlations to Improve Robustness

    Vision Transformers (ViTs) often degrade under distribution shifts because they rely on spurious correlations, such as background cues, rather than semantically meaningful features. Existing regularization methods, typically relying on simple foreground-background masks, which fail to capture the fine-grained semantic concepts that define an object (e.g., ``long beak'' and ``wings'' for a ``bird''). As a result, these methods provide limited robustness to distribution shifts. To address this limitation, we introduce a novel finetuning framework that steers model reasoning toward concept-level semantics. Our approach optimizes the model's internal relevance maps to align with spatially grounded concept masks. These masks are generated automatically, without manual annotation: class-relevant concepts are first proposed using an LLM-based, label-free method, and then segmented using a VLM. The finetuning objective aligns relevance with these concept regions while simultaneously suppressing focus on spurious background areas. Notably, this process requires only a minimal set of images and uses half of the dataset classes. Extensive experiments on five out-of-distribution benchmarks demonstrate that our method improves robustness across multiple ViT-based models. Furthermore, we show that the resulting relevance maps exhibit stronger alignment with semantic object parts, offering a scalable path toward more robust and interpretable vision models. Finally, we confirm that concept-guided masks provide more effective supervision for model robustness than conventional segmentation maps, supporting our central hypothesis.

Techmeme(15)

  1. Sandbar, which is developing the Stream ring, a $249+ AI-powered wearable that transcribes audio notes, raised a $23M Series A, bringing total funding to $36M (Alex Konrad/Upstarts Media)

    Alex Konrad / Upstarts Media : Sandbar, which is developing the Stream ring, a $249+ AI-powered wearable that transcribes audio notes, raised a $23M Series A, bringing total funding to $36M —  Sandbar CEO Mina Fahmi has raised $23M in Series A funding to fuel the launch of his AI note-taking ring, called Stream.

  2. Niantic Spatial partners with Coco Robotics to integrate a visual positioning system trained on data from Pokemon Go and Ingress into a fleet of delivery robots (Will Douglas Heaven/MIT Technology Review)

    Will Douglas Heaven / MIT Technology Review : Niantic Spatial partners with Coco Robotics to integrate a visual positioning system trained on data from Pokemon Go and Ingress into a fleet of delivery robots —  Pokémon Go was the world's first augmented-reality megahit.  Released in 2016 by the Google spinout Niantic …

  3. Leaked memo: a top Senate administrator gave aides the green light to use ChatGPT, Gemini, and Copilot for official Senate work, including preparing briefings (Catie Edmondson/New York Times)

    Catie Edmondson / New York Times : Leaked memo: a top Senate administrator gave aides the green light to use ChatGPT, Gemini, and Copilot for official Senate work, including preparing briefings —  New guidelines said Senate aides could use A.I. tools for official work, including research, drafting and editing documents …

  4. Internal doc: the State Department moved its internal chatbot from Claude Sonnet 4.5 to GPT-4.1, following Trump's directive to cancel Anthropic contracts (Nextgov/FCW)

    Nextgov/FCW : Internal doc: the State Department moved its internal chatbot from Claude Sonnet 4.5 to GPT-4.1, following Trump's directive to cancel Anthropic contracts —  The agency moved its chatbot to operate on OpenAI's GPT 4.1, internal document shows.  —  The State Department has shifted …

  5. Filing: Microsoft files an amicus brief in support of Anthropic and advocates for a temporary restraining order to block the DOD's supply chain risk designation (Ashley Capoot/CNBC)

    Ashley Capoot / CNBC : Filing: Microsoft files an amicus brief in support of Anthropic and advocates for a temporary restraining order to block the DOD's supply chain risk designation —  Microsoft threw its support behind Anthropic on Tuesday, saying a judge should issue a restraining order that would block …

  6. Amazon expands its healthcare AI assistant Health AI to its website and app; it was previously only available on the app for One Medical (Aisha Malik/TechCrunch)

    Aisha Malik / TechCrunch : Amazon expands its healthcare AI assistant Health AI to its website and app; it was previously only available on the app for One Medical —  Amazon announced on Tuesday that it's expanding access to its healthcare AI assistant to its website and app.  The assistant, called Health AI …

  7. Sources: General Catalyst is in talks with investors to raise about $10B; it raised $8B in capital in 2024 and had more than $40B in AUM as of last summer (Natasha Mascarenhas/Bloomberg)

    Natasha Mascarenhas / Bloomberg : Sources: General Catalyst is in talks with investors to raise about $10B; it raised $8B in capital in 2024 and had more than $40B in AUM as of last summer —  General Catalyst, a venture capital firm that has recently transformed into a broader financial services company …

  8. Nitra, which offers a platform powered by AI agents to manage medical practices, raised a $50M Series B, bringing its total funding to $205M (Catherina Gioino/Fortune)

    Catherina Gioino / Fortune : Nitra, which offers a platform powered by AI agents to manage medical practices, raised a $50M Series B, bringing its total funding to $205M —  Tim Hwang has spent his career moving between politics, policy, and startups.  He worked on Barack Obama's 2008 presidential campaign …

  9. Oracle reports Q3 revenue up 22% YoY to $17.19B, vs. $16.91B est., and cloud revenue up 44% to $8.9B, vs. $8.85B est.; ORCL jumps 8%+ after hours (Jordan Novet/CNBC)

    Jordan Novet / CNBC : Oracle reports Q3 revenue up 22% YoY to $17.19B, vs. $16.91B est., and cloud revenue up 44% to $8.9B, vs. $8.85B est.; ORCL jumps 8%+ after hours —  Oracle shares rose as much as 10% in extended trading on Tuesday after the software vendor reported quarterly results that surpassed Wall Street projections …

  10. MoffettNathanson: YouTube became the world's largest media company in 2025 with an estimated $62B in revenue, passing $60.9B earned by Disney's media business (Alex Weprin/The Hollywood Reporter)

    Alex Weprin / The Hollywood Reporter : MoffettNathanson: YouTube became the world's largest media company in 2025 with an estimated $62B in revenue, passing $60.9B earned by Disney's media business —  The influential financial research firm MoffettNathanson suggests that the Google-owned video platform passed Disney's media business in 2025.

  11. Sources: US Social Security's inspector general is investigating claims an ex-DOGE engineer took sensitive data on a thumb drive to his new private employer (Washington Post)

    Washington Post : Sources: US Social Security's inspector general is investigating claims an ex-DOGE engineer took sensitive data on a thumb drive to his new private employer —  The Social Security inspector general's office is investigating allegations that the former DOGE engineer took sensitive data …

  12. Nielsen's Gracenote sues OpenAI for copyright infringement, saying OpenAI copied Gracenote's data and relational framework used to connect metadata (Sara Fischer/Axios)

    Sara Fischer / Axios : Nielsen's Gracenote sues OpenAI for copyright infringement, saying OpenAI copied Gracenote's data and relational framework used to connect metadata —  - To date, there hasn't been a major media copyright lawsuit that focuses on the theft of a proprietary sequence or structure behind a dataset.

  13. Sandbar, which is developing the Stream Ring, a $249+ AI-powered wearable that transcribes audio notes, raised a $23M Series A, bringing total funding to $36M (Ivan Mehta/TechCrunch)

    Ivan Mehta / TechCrunch : Sandbar, which is developing the Stream Ring, a $249+ AI-powered wearable that transcribes audio notes, raised a $23M Series A, bringing total funding to $36M —  Sandbar, a startup by former Meta employees Mina Fahmi and Kirak Hong, attracted much attention last year when it showed off its note-taking wearable, the Stream ring.

  14. Slide, which develops data backup and disaster recovery tech for managed service providers, raised a $70M Series B led by GC, bringing its total funding to $95M (CJ Fairfield/CRN)

    CJ Fairfield / CRN : Slide, which develops data backup and disaster recovery tech for managed service providers, raised a $70M Series B led by GC, bringing its total funding to $95M —  'We're using this additional funding to continue to put our foot on the accelerator so that we can grow the business …

  15. The Senate confirms Army Lt. Gen. Joshua Rudd to lead the NSA and US Cyber Command, filling a vacancy created when Gen. Timothy Haugh was fired in April 2025 (Martin Matishak/The Record)

    Martin Matishak / The Record : The Senate confirms Army Lt. Gen. Joshua Rudd to lead the NSA and US Cyber Command, filling a vacancy created when Gen. Timothy Haugh was fired in April 2025 —  Army Lt. Gen. Joshua Rudd was confirmed by the Senate on Tuesday to be the next leader of U.S. Cyber Command and director …

Solidot(15)

  1. 很多国际游戏开发者计划不参加今年的 GDC

    数万游戏开发者和制作人本周将齐聚旧金山,参加为期一周的游戏开发者大会(GDC),这是 1988 年以来的传统,但今年的 GDC 将有许多国际游戏开发者缺席,原因是他们觉得美国不再安全,无论 GDC 对他们的工作和职业发展有多么重要,他们不想冒不必要的风险。Godot 基金会执行董事 Emilio Coppola 称他认识的非美国人中没有一个人计划参加 GDC。独立游戏工作室 Le Cabinet du Savoir 的创意总监 Nazih Fares 表示不愿意亲身经历被边检逮捕。去年参加 GDC 的游戏开发者采取了额外的多重安保措施,他们很多人表示为避免麻烦而不想参加今年的 GDC。

  2. Meta 称上传盗版电子书属于合理使用

    为训练大模型,社交巨人 Meta 从 Z-Library 和 LibGen 等影子图书馆平台通过 BitTorrent 下载了逾百 TB 的电子书。在正在进行的由图书作者提起的诉讼中,Meta 律师辩称,通过 BitTorrent 将盗版电子书上传给陌生人属于合理使用。Meta 还强调,这些数据帮助美国确立了其在全球 AI 领域的领先地位。法庭去年裁决,使用盗版电子书训练大模型属于合理使用,但 Meta 仍然需要为通过 BitTorrent 下载和分享电子书的行为承担责任。图书作者认为,Meta 参与了侵权行为。Meta 在上周递交的补充书面询问中表示,在下载 BT 文件过程中共享文件也属于合理使用,理由是这是 BT 协议的固有特性,上传不是选择而是技术本身的工作方式。Meta 还辩称,使用 BitTorrent 共享文件是获取这些宝贵(但盗版)数据的必要手段。以 Anna’s Archive 为例,这些数据集只能通过 BT 下载获取,因此 BitTorrent 是唯一的选择。

  3. 为什么高处坠落的猫总是四脚着地?

    从高处坠落的猫总能四脚着地。科学家一直在争论背后的机制,他们提出了四种假说:收腿翻转(tuck and turn)模型认为猫收起一组爪子以便能旋转身体的不同部位;麦克斯韦(James Clerk Maxwell)提出的下落花样滑冰运动员(falling figure skate)模型认为猫通过收回或伸展爪子调整其角动量;弯曲扭转(bend and twist)模型认为猫在腰部弯曲使身体的两部分产生反向旋转;螺旋尾巴(propeller tail)模型认为猫通过像螺旋桨一样旋转尾巴去反转身体的旋转方向。根据发表在《The Anatomical Record》期刊上的最新研究,日本科学家从五具捐赠的猫尸上取出脊椎,保留韧带和椎间盘,将胸椎和腰椎部分分离,然后将其放入一具扭转装置,观察扭转它们所需的力以及各部分扭转的极限角度。他们还将两只活猫各自抛八次,拍摄了猫在自由落体下的高速照片。结果显示,上段脊椎的扭转角度大于下段脊椎,在扭转角度约 50 度时存在一个“最佳点”,在该点扭转时几乎没有阻力。下段脊椎则不存在这个点,这为“收腿翻转”假说提供了证据。高速摄影也观察到了猫的收腿翻转动作。研究人员还观察到猫总是倾向于右转,可能是内脏器官的不对称分布使其更容易向右转。

  4. 数据中心成为攻击基础设施的目标

    科技行业常把“云”说成是某种抽象且遥不可及的东西。但云运行在数据中心,而数据中心有地址,这个地址可能会遭到无人机袭击。上周亚马逊 AWS 运营的三个数据中心遭到袭击,其中两个位于阿联酋,一个位于巴林。袭击导致设施离线,引发了整个地区银行、支付、外卖应用和企业软件等服务的中断。此次袭击是数据中心首次成为攻击目标。专家认为这肯定不会是最后一次。数据中心正迅速成为重要战略资产,同时也成为易受攻击的目标。

  5. 瑞士通过修宪保障使用现金的权利

    瑞士选民以压倒性多数通过了一项宪法修正案,保障民众使用现金的权利。欧洲除了瑞士,匈牙利、斯洛伐克和斯洛文尼亚等国也都将保障现金使用权利写入宪法。官方统计结果显示,73.4% 的选民支持该宪法修正案。该修正案由政府提出,旨在反击“瑞士自由运动(Swiss Freedom Movement)”组织提出的类似倡议。瑞士自由运动发起了保护现金的倡议,收集了逾 10 万个签名,最终引发了全民公投。由于政府认为该组织提出的部分修正案过于激进,最终该倡议仅获得 46% 的投票支持。

  6. 调查发现三分之一美国人认为末日将在其有生之年来临

    美国相信末日来临的人并非少数。根据《Journal of Personality and Social Psychology》上发表的一篇报告,研究人员调查了 1409 名不同信仰的美国人,结果显示三分之一相信末日将在其有生之年来临。一部分人认为末日是人类引发的,还有部分人认为末日是由神或超自然力量引发的。相信末日临近且人类是罪魁祸首的人,感知到的风险更大,也更支持采取更极端的行动应对威胁。然而相信神控制世界末日的人则不太可能支持采取预防措施。

  7. digiKam 9.0.0 释出

    照片管理系统 DigiKam 释出了v9.0.0 版本。新版本改进了性能、易用性和工作流效率,对用户界面进行了现代化、增强了元数据管理以及扩展对新相机型号和文件格式的支持。其它变化包括:新工具 Survey 帮助摄影师更高效的查看工作流项目;更先进的搜索和排序选项;批量编辑地理位置坐标,等等。

  8. 太空风暴让人类难以探测到 ET 的信号

    SETI 研究所的最新研究发现,太空风暴让人类难以探测到 ET 发射的无线电信号。行星信号发射源附近的恒星活动和等离子体湍流会使原本极其窄带的信号展宽,致使信号功率分散到更多频率上,因此在传统的窄带搜索中更难探测到。SETI 实验过去几十年一直致力于探测天体物理自然过程不太可能存在的频率尖峰。这项发表在《Astrophysical Journal》期刊上的研究显示,即使最初发射的信号非常窄,但信号离开其行星系统之后会展宽,因此会低于探测阈值。假设 ET 存在,且可能正在试图联系我们,但不可预测的太空天气干扰了信号,人类根本无法接收到。

  9. 一百万颗卫星会如何影响天空?

    SpaceX 计划向太空发射 1 百万颗卫星,理由是建造太空数据中心。暂不讨论太空数据中心的可行性(没什么可行性),这一百万颗卫星会如何影响我们每一个人?SpaceX 已经向 FCC 递交了发射提议,对该计划的公众意见征集于上周五结束,逾千条公众意见绝大多数持反对立场,要求 FCC 停止推进该计划。SpaceX 已经向地球轨道发射了上万颗卫星,一百万颗则是已有数量的一百倍。SpaceX 平均每周发射两次,它的卫星不断上天,也不断坠落。越来越多的证据表明,火箭发射会将污染物排放到空气中,影响大气层,造成潜在的温室效应,可能加剧对臭氧层的威胁。如果 SpaceX 的一百万颗卫星全部脱离轨道,意味着每三分钟就有一颗卫星重返大气层。轨道上的卫星越多,发生碰撞的可能性也越大。卫星也会影响天文观测,该公司一直试图与国际天文学联合会合作减少卫星对天文观测的影响,但一百万颗完全不同的量级,天文学家震惊不已。

  10. FBI 通过 Proton Mail 识别抗议者身份

    FBI 通过瑞士邮件服务商 Proton Mail 提供的信息识别了亚特兰大抗议组织 Defend the Atlanta Forest/Stop Cop City 领导人的身份。Proton Mail 坚称它必须遵守瑞士的法律。Stop Cop City 官方 FB 账号使用的邮箱是 defendtheatlantaforest@protonmail.com,FBI 援引《司法互助条约(Mutual Legal Assistance Treaty 或 MLAT)》,请求瑞士司法部向 Proton Mail 索取信息。瑞士与逾 30 个国家签订了 MLAT。Proton 向瑞士司法机构提供了信息,然后由瑞士转交给了 FBI。Proton AG 通讯主管 Edward Shone 表示,该公司没有直接向 FBI 提供信息,相关信息是 FBI 通过瑞士司法部获取到的。

  11. 卫星揭示北美和非洲桥梁面临稳定性风险

    根据发表在《Nature Communication》期刊上的一项研究,休斯顿大学等机构的研究人员利用卫星分析了全世界 744 座桥梁,评估其状况。研究结果显示,北美桥梁的状况普遍最差,其次是非洲桥梁。研究分析的很多桥梁已接近其设计使用寿命的上限。北美桥梁的建设高峰是在 1960 年代,很多已建成数十年,接近或超过其最初的设计寿命。研究人员利用名为多时相干涉合成孔径雷达(Multi-temporal InSAR, MTInSAR)的卫星遥感方法去监测桥梁结构中的微小位移。

  12. 耳鸣与睡眠密切相关

    幻影知觉(phantom percept)是大脑愚弄我们以为看到、听到、感觉到或闻到了实际上不存在的东西。耳鸣是最常见的幻影知觉,尽管有很多假说,但至今未找到确切的病因或疗法,全世界有 15% 的人口受到耳鸣的困扰。很多耳鸣患者都表示睡眠质量差,睡眠模式紊乱,但耳鸣与睡眠这一重要生理功能之间的潜在关联直到最近才被人所认识。牛津大学的神经学家提出,深度睡眠或非快速眼动睡眠(non-REM)期间出现的大幅度自发性脑电波可能抑制导致耳鸣的脑电活动。雪貂实验显示,耳鸣雪貂在进入 non-REM 睡眠后过度活跃的脑电活动会减弱。研究结果表明,深度睡眠可能有助于缓解耳鸣,有可能揭示了大脑调节异常活动的自然机制。

  13. 研究人员模拟月球土壤种植收获鹰嘴豆

    根据发表在《Scientific Reports》期刊上的一项研究,德州农工的研究人员模拟月球土壤种植收获了鹰嘴豆,但鹰嘴豆的食用安全性尚未确定。月球土壤贫瘠,缺乏营养元素,富含重金属,为克服这些问题,研究人员利用蚯蚓通过废弃物产生的堆肥去提供必需的微生物和营养物质,使用共生真菌 Arbuscular Mycorrhizal Fungi 去促进植物生长和减少对有毒重金属的吸收。结果显示添加堆肥和共生真菌的模拟月壤混合物能像普通地球土壤那样种植和收获鹰嘴豆。研究人员接下来将分析鹰嘴豆的营养成分,检测是否含有重金属,确保人类的食用安全。

  14. 新法律要求成人网站验证澳大利亚人年龄

    在禁止 16 岁以下儿童使用社交媒体三个月后,澳大利亚新法律要求有成人内容的网站验证该国访客的年龄,确保其年满 18 岁,违规者将面临罚款。澳大利亚网络安全监管机构表示,此举旨在保护儿童免受有害内容的侵害。eSafety 专员 Julie Inman Grant 表示,我们不允许儿童进入酒吧、酒专卖店、成人用品店或赌场,但对于网络空间,却没有此类保障措施。平台可能需要使用面部识别技术、数字身份和信用卡信息去验证访客。根据新规,搜索引擎、应用商店、社交媒体和游戏平台、色情网站以及包括聊天机器人在内的 AI 系统必须采取切实有效的措施,防止儿童接触成人内容。

  15. 印尼和印度卡纳塔克邦将禁止 16 岁以下儿童使用社媒

    继澳大利亚之后,印度科技重镇卡纳塔克邦(Karnataka)与印度尼西亚相继宣布将禁止 16 岁以下青少年使用社交媒体及高风险数字平台。印尼通信与数码部长梅蒂雅(Meutya Hafid)星期五发声明说,政府将从 3 月 28 日起,分阶段注销 16 岁以下青少年在“高风险平台”上的账户。首批受影响的平台包括 YouTube、TikTok、Facebook、Instagram、Threads、X、Bigo Live 以及游戏平台 Roblox。梅蒂雅说,颁布禁令是因为青少年面临网络色情、网络欺凌、网络诈骗和网络成瘾的威胁。印度卡纳塔克邦立法议员星期五在邦预算会议上提出禁止16岁以下青少年使用社媒应用的法案。卡纳塔克邦将成为全印度首个实行这项禁令的邦。议员里兹万说:“青少年在未了解后果的情况下,就开始使用社交媒体。我们将与社会人士探讨,如何在社媒落实年龄限制。”