OrangeBot.AI Digest — 2026-04-18
85 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Thoughts and feelings around Claude Design (samhenri.gold)
- Traders placed over $1B in perfectly timed bets on the Iran war (www.theguardian.com)
- The electromechanical angle computer inside the B-52 bomber's star tracker (www.righto.com)
- Opus 4.7 to 4.6 Inflation is ~45% (tokens.billchambers.me)
- Amazon is discontinuing Kindle for PC on June 30th (goodereader.com)
- Sumida Aquarium Posts 2026 Penguin Relationship Chart, with Drama and Breakups (www.sumida-aquarium.com)
- Migrating from DigitalOcean to Hetzner (isayeter.com)
- Why Japan has such good railways (worksinprogress.co)
- State of Kdenlive (kdenlive.org)
- The quiet disappearance of the free-range childhood (bigthink.com)
- It's OK to compare floating-points for equality (lisyarus.github.io)
- America Lost the Mandate of Heaven (geohot.github.io)
- Michael Rabin has died (en.wikipedia.org)
- Amiga Graphics Archive (amiga.lychesis.net)
- The simple geometry behind any road (sandboxspirit.com)
GitHub Trending(10)
Product Hunt(15)
- CraftBot
Self-hosted proactive AI assistant that lives locally
- Claude Design by Anthropic Labs
Make prototypes, slides & one-pagers by talking to Claude
- Android CLI
Build high quality Android apps 3x faster using any agent
- Notebooks in Gemini
Keep every project, chat, and file in one focused space
- ChatGPT Shopping
Richer, more visually immersive shopping experiences
- Grok Voice API
Fast, accurate STT and TTS APIs at the best price
- Claude Code Rendering
Mouse support and flicker-free rendering for Claude Code
- GPT‑Rosalind
Purpose-built model for scientific research & drug discovery
- Lounge
macOS Tahoe menu bar manager with notch-aware icon control
- CapyPlan
Your no stress chearleader for tiny tasks
- Is Your Site Agent-Ready? by Cloudflare
Scan your website to see how ready it is for AI agents.
- .MD This Page
Convert any page to clean Markdown instantly
- React Email 6.0 by Resend
Build, customize, and ship emails — all from your own app
- Vercel Flags
Feature flags, targeting rules, rollouts. All from Vercel.
- Hipocampus
AI operators that own team workflows
Hugging Face(15)
- HY-World 2.0: A Multi-Modal World Model for Reconstructing, Generating, and Simulating 3D Worlds
We introduce HY-World 2.0, a multi-modal world model framework that advances our prior project HY-World 1.0. HY-World 2.0 accommodates diverse input modalities, including text prompts, single-view images, multi-view images, and videos, and produces 3D world representations. With text or single-view image inputs, the model performs world generation, synthesizing high-fidelity, navigable 3D Gaussian Splatting (3DGS) scenes. This is achieved through a four-stage method: a) Panorama Generation with HY-Pano 2.0, b) Trajectory Planning with WorldNav, c) World Expansion with WorldStereo 2.0, and d) World Composition with WorldMirror 2.0. Specifically, we introduce key innovations to enhance panorama fidelity, enable 3D scene understanding and planning, and upgrade WorldStereo, our keyframe-based view generation model with consistent memory. We also upgrade WorldMirror, a feed-forward model for universal 3D prediction, by refining model architecture and learning strategy, enabling world reconstruction from multi-view images or videos. Also, we introduce WorldLens, a high-performance 3DGS rendering platform featuring a flexible engine-agnostic architecture, automatic IBL lighting, efficient collision detection, and training-rendering co-design, enabling interactive exploration of 3D worlds with character support. Extensive experiments demonstrate that HY-World 2.0 achieves state-of-the-art performance on several benchmarks among open-source approaches, delivering results comparable to the closed-source model Marble. We release all model weights, code, and technical details to facilitate reproducibility and support further research on 3D world models.
- RAD-2: Scaling Reinforcement Learning in a Generator-Discriminator Framework
High-level autonomous driving requires motion planners capable of modeling multimodal future uncertainties while remaining robust in closed-loop interactions. Although diffusion-based planners are effective at modeling complex trajectory distributions, they often suffer from stochastic instabilities and the lack of corrective negative feedback when trained purely with imitation learning. To address these issues, we propose RAD-2, a unified generator-discriminator framework for closed-loop planning. Specifically, a diffusion-based generator is used to produce diverse trajectory candidates, while an RL-optimized discriminator reranks these candidates according to their long-term driving quality. This decoupled design avoids directly applying sparse scalar rewards to the full high-dimensional trajectory space, thereby improving optimization stability. To further enhance reinforcement learning, we introduce Temporally Consistent Group Relative Policy Optimization, which exploits temporal coherence to alleviate the credit assignment problem. In addition, we propose On-policy Generator Optimization, which converts closed-loop feedback into structured longitudinal optimization signals and progressively shifts the generator toward high-reward trajectory manifolds. To support efficient large-scale training, we introduce BEV-Warp, a high-throughput simulation environment that performs closed-loop evaluation directly in Bird's-Eye View feature space via spatial warping. RAD-2 reduces the collision rate by 56% compared with strong diffusion-based planners. Real-world deployment further demonstrates improved perceived safety and driving smoothness in complex urban traffic.
- DR^{3}-Eval: Towards Realistic and Reproducible Deep Research Evaluation
Deep Research Agents (DRAs) aim to solve complex, long-horizon research tasks involving planning, retrieval, multimodal understanding, and report generation, yet their evaluation remains challenging due to dynamic web environments and ambiguous task definitions. We propose DR^{3}-Eval, a realistic and reproducible benchmark for evaluating deep research agents on multimodal, multi-file report generation. DR^{3}-Eval is constructed from authentic user-provided materials and paired with a per-task static research sandbox corpus that simulates open-web complexity while remaining fully verifiable, containing supportive documents, distractors, and noise. Moreover, we introduce a multi-dimensional evaluation framework measuring Information Recall, Factual Accuracy, Citation Coverage, Instruction Following, and Depth Quality, and validate its alignment with human judgments. Experiments with our developed multi-agent system DR^{3}-Agent based on multiple state-of-the-art language models demonstrate that DR^{3}-Eval is highly challenging and reveals critical failure modes in retrieval robustness and hallucination control. Our code and data are publicly available.
- How to Fine-Tune a Reasoning Model? A Teacher-Student Cooperation Framework to Synthesize Student-Consistent SFT Data
A widely adopted strategy for model enhancement is to use synthetic data generated by a stronger model for supervised fine-tuning (SFT). However, for emerging reasoning models like Qwen3-8B, this approach often fails to improve reasoning capabilities and can even lead to a substantial drop in performance. In this work, we identify substantial stylistic divergence between teacher generated data and the distribution of student as a major factor impacting SFT. To bridge this gap, we propose a Teacher-Student Cooperation Data Synthesis framework (TESSY), which interleaves teacher and student models to alternately generate style and non-style tokens. Consequently, TESSY produces synthetic sequences that inherit the advanced reasoning capabilities of the teacher while maintaining stylistic consistency with the distribution of the student. In experiments on code generation using GPT-OSS-120B as the teacher, fine-tuning Qwen3-8B on teacher-generated data leads to performance drops of 3.25% on LiveCodeBench-Pro and 10.02% on OJBench, whereas TESSY achieves improvements of 11.25% and 6.68%.
- GlobalSplat: Efficient Feed-Forward 3D Gaussian Splatting via Global Scene Tokens
The efficient spatial allocation of primitives serves as the foundation of 3D Gaussian Splatting, as it directly dictates the synergy between representation compactness, reconstruction speed, and rendering fidelity. Previous solutions, whether based on iterative optimization or feed-forward inference, suffer from significant trade-offs between these goals, mainly due to the reliance on local, heuristic-driven allocation strategies that lack global scene awareness. Specifically, current feed-forward methods are largely pixel-aligned or voxel-aligned. By unprojecting pixels into dense, view-aligned primitives, they bake redundancy into the 3D asset. As more input views are added, the representation size increases and global consistency becomes fragile. To this end, we introduce GlobalSplat, a framework built on the principle of align first, decode later. Our approach learns a compact, global, latent scene representation that encodes multi-view input and resolves cross-view correspondences before decoding any explicit 3D geometry. Crucially, this formulation enables compact, globally consistent reconstructions without relying on pretrained pixel-prediction backbones or reusing latent features from dense baselines. Utilizing a coarse-to-fine training curriculum that gradually increases decoded capacity, GlobalSplat natively prevents representation bloat. On RealEstate10K and ACID, our model achieves competitive novel-view synthesis performance while utilizing as few as 16K Gaussians, significantly less than required by dense pipelines, obtaining a light 4MB footprint. Further, GlobalSplat enables significantly faster inference than the baselines, operating under 78 milliseconds in a single forward pass. Project page is available at https://r-itk.github.io/globalsplat/
- ASGuard: Activation-Scaling Guard to Mitigate Targeted Jailbreaking Attack
Large language models (LLMs), despite being safety-aligned, exhibit brittle refusal behaviors that can be circumvented by simple linguistic changes. As tense jailbreaking demonstrates that models refusing harmful requests often comply when rephrased in past tense, a critical generalization gap is revealed in current alignment methods whose underlying mechanisms are poorly understood. In this work, we introduce Activation-Scaling Guard (ASGuard), an insightful, mechanistically-informed framework that surgically mitigates this specific vulnerability. In the first step, we use circuit analysis to identify the specific attention heads causally linked to the targeted jailbreaking such as a tense-changing attack. Second, we train a precise, channel-wise scaling vector to recalibrate the activation of tense vulnerable heads. Lastly, we apply it into a "preventative fine-tuning", forcing the model to learn a more robust refusal mechanism. Across four LLMs, ASGuard effectively reduces the attack success rate of targeted jailbreaking while preserving general capabilities and minimizing over refusal, achieving a Pareto-optimal balance between safety and utility. Our findings underscore how adversarial suffixes suppress the propagation of the refusal-mediating direction, based on mechanistic analysis. Furthermore, our work showcases how a deep understanding of model internals can be leveraged to develop practical, efficient, and targeted methods for adjusting model behavior, charting a course for more reliable and interpretable AI safety.
- HiVLA: A Visual-Grounded-Centric Hierarchical Embodied Manipulation System
While end-to-end Vision-Language-Action (VLA) models offer a promising paradigm for robotic manipulation, fine-tuning them on narrow control data often compromises the profound reasoning capabilities inherited from their base Vision-Language Models (VLMs). To resolve this fundamental trade-off, we propose HiVLA, a visual-grounded-centric hierarchical framework that explicitly decouples high-level semantic planning from low-level motor control. In high-level part, a VLM planner first performs task decomposition and visual grounding to generate structured plans, comprising a subtask instruction and a precise target bounding box. Then, to translate this plan into physical actions, we introduce a flow-matching Diffusion Transformer (DiT) action expert in low-level part equipped with a novel cascaded cross-attention mechanism. This design sequentially fuses global context, high-resolution object-centric crops and skill semantics, enabling the DiT to focus purely on robust execution. Our decoupled architecture preserves the VLM's zero-shot reasoning while allowing independent improvement of both components. Extensive experiments in simulation and the real world demonstrate that HiVLA significantly outperforms state-of-the-art end-to-end baselines, particularly excelling in long-horizon skill composition and the fine-grained manipulation of small objects in cluttered scenes.
- Dive into Claude Code: The Design Space of Today's and Future AI Agent Systems
Claude Code is an agentic coding tool that can run shell commands, edit files, and call external services on behalf of the user. This study describes its comprehensive architecture by analyzing the publicly available TypeScript source code and further comparing it with OpenClaw, an independent open-source AI agent system that answers many of the same design questions from a different deployment context. Our analysis identifies five human values, philosophies, and needs that motivate the architecture (human decision authority, safety and security, reliable execution, capability amplification, and contextual adaptability) and traces them through thirteen design principles to specific implementation choices. The core of the system is a simple while-loop that calls the model, runs tools, and repeats. Most of the code, however, lives in the systems around this loop: a permission system with seven modes and an ML-based classifier, a five-layer compaction pipeline for context management, four extensibility mechanisms (MCP, plugins, skills, and hooks), a subagent delegation mechanism with worktree isolation, and append-oriented session storage. A comparison with OpenClaw, a multi-channel personal assistant gateway, shows that the same recurring design questions produce different architectural answers when the deployment context changes: from per-action safety classification to perimeter-level access control, from a single CLI loop to an embedded runtime within a gateway control plane, and from context-window extensions to gateway-wide capability registration. We finally identify six open design directions for future agent systems, grounded in recent empirical, architectural, and policy literature.
- Switch-KD: Visual-Switch Knowledge Distillation for Vision-Language Models
Vision-Language Models (VLMs) have shown remarkable capabilities in joint vision-language understanding, but their large scale poses significant challenges for deployment in resource-constrained scenarios. Knowledge Distillation (KD) offers a viable way to improve model capabilities without increasing model size or data requirements, making deployment more efficient. However, applying KD to VLMs is challenged by modality-specific supervision: although multimodal knowledge in VLMs is fused within the language space, current methods supervise each modality separately without explicitly addressing multimodal alignment, leading to inconsistent multimodal knowledge transfer. To address this, we propose Switch-KD, a visual-switch distillation framework that unifies vision-language knowledge transfer within a shared text-probability space. Switch-KD comprises two key components: (1) Visual-Switch Distillation, which switches the student's visual outputs into the teacher's language pathway to construct cross-modal probabilistic references for implicit visual knowledge transfer; and (2) Dynamic Bi-directional Logits Difference (DBiLD) loss, which adaptively aligns informative probability regions while preserving the distributional structures of teacher and student through bidirectional supervision. Guided by Switch-KD, a 0.5B TinyLLaVA effectively distills rich multimodal knowledge from its 3B teacher, yielding an average improvement of 3.6 points across 10 multimodal benchmarks without any architectural modification.
- UniDoc-RL: Coarse-to-Fine Visual RAG with Hierarchical Actions and Dense Rewards
Retrieval-Augmented Generation (RAG) extends Large Vision-Language Models (LVLMs) with external visual knowledge. However, existing visual RAG systems typically rely on generic retrieval signals that overlook the fine-grained visual semantics essential for complex reasoning. To address this limitation, we propose UniDoc-RL, a unified reinforcement learning framework in which an LVLM agent jointly performs retrieval, reranking, active visual perception, and reasoning. UniDoc-RL formulates visual information acquisition as a sequential decision-making problem with a hierarchical action space. Specifically, it progressively refines visual evidence from coarse-grained document retrieval to fine-grained image selection and active region cropping, allowing the model to suppress irrelevant content and attend to information-dense regions. For effective end-to-end training, we introduce a dense multi-reward scheme that provides task-aware supervision for each action. Based on Group Relative Policy Optimization (GRPO), UniDoc-RL aligns agent behavior with multiple objectives without relying on a separate value network. To support this training paradigm, we curate a comprehensive dataset of high-quality reasoning trajectories with fine-grained action annotations. Experiments on three benchmarks demonstrate that UniDoc-RL consistently surpasses state-of-the-art baselines, yielding up to 17.7% gains over prior RL-based methods.
- Representations Before Pixels: Semantics-Guided Hierarchical Video Prediction
Accurate future video prediction requires both high visual fidelity and consistent scene semantics, particularly in complex dynamic environments such as autonomous driving. We present Re2Pix, a hierarchical video prediction framework that decomposes forecasting into two stages: semantic representation prediction and representation-guided visual synthesis. Instead of directly predicting future RGB frames, our approach first forecasts future scene structure in the feature space of a frozen vision foundation model, and then conditions a latent diffusion model on these predicted representations to render photorealistic frames. This decomposition enables the model to focus first on scene dynamics and then on appearance generation. A key challenge arises from the train-test mismatch between ground-truth representations available during training and predicted ones used at inference. To address this, we introduce two conditioning strategies, nested dropout and mixed supervision, that improve robustness to imperfect autoregressive predictions. Experiments on challenging driving benchmarks demonstrate that the proposed semantics-first design significantly improves temporal semantic consistency, perceptual quality, and training efficiency compared to strong diffusion baselines. We provide the implementation code at https://github.com/Sta8is/Re2Pix
- LeapAlign: Post-Training Flow Matching Models at Any Generation Step by Building Two-Step Trajectories
This paper focuses on the alignment of flow matching models with human preferences. A promising way is fine-tuning by directly backpropagating reward gradients through the differentiable generation process of flow matching. However, backpropagating through long trajectories results in prohibitive memory costs and gradient explosion. Therefore, direct-gradient methods struggle to update early generation steps, which are crucial for determining the global structure of the final image. To address this issue, we introduce LeapAlign, a fine-tuning method that reduces computational cost and enables direct gradient propagation from reward to early generation steps. Specifically, we shorten the long trajectory into only two steps by designing two consecutive leaps, each skipping multiple ODE sampling steps and predicting future latents in a single step. By randomizing the start and end timesteps of the leaps, LeapAlign leads to efficient and stable model updates at any generation step. To better use such shortened trajectories, we assign higher training weights to those that are more consistent with the long generation path. To further enhance gradient stability, we reduce the weights of gradient terms with large magnitude, instead of completely removing them as done in previous works. When fine-tuning the Flux model, LeapAlign consistently outperforms state-of-the-art GRPO-based and direct-gradient methods across various metrics, achieving superior image quality and image-text alignment.
- TRACER: Trace-Based Adaptive Cost-Efficient Routing for LLM Classification
Every call to an LLM classification endpoint produces a labeled input-output pair already retained in production logs. These pairs constitute a free, growing training set: a lightweight surrogate trained on them can absorb a significant portion of future traffic at near-zero marginal inference cost. The open questions are when the surrogate is reliable enough to deploy, what it handles versus defers, and how that boundary evolves as data accumulates. We introduce TRACER (Trace-based Adaptive Cost-Efficient Routing), an open-source system that trains ML surrogates on an LLM's own production traces and governs deployment through a parity gate: the surrogate is activated only when its agreement with the LLM exceeds a user-specified threshold α. To make the routing boundary transparent, TRACER generates interpretability artifacts describing which input regions the surrogate handles, where it plateaus, and why it defers. On a 77-class intent benchmark with a Sonnet 4.6 teacher, TRACER achieves 83-100% surrogate coverage depending on the quality target α; on a 150-class benchmark, the surrogate fully replaces the teacher. On a natural language inference task, the parity gate correctly refuses deployment because the embedding representation cannot support reliable separation. The system is available as open-source software.
- OneHOI: Unifying Human-Object Interaction Generation and Editing
Human-Object Interaction (HOI) modelling captures how humans act upon and relate to objects, typically expressed as <person, action, object> triplets. Existing approaches split into two disjoint families: HOI generation synthesises scenes from structured triplets and layout, but fails to integrate mixed conditions like HOI and object-only entities; and HOI editing modifies interactions via text, yet struggles to decouple pose from physical contact and scale to multiple interactions. We introduce OneHOI, a unified diffusion transformer framework that consolidates HOI generation and editing into a single conditional denoising process driven by shared structured interaction representations. At its core, the Relational Diffusion Transformer (R-DiT) models verb-mediated relations through role- and instance-aware HOI tokens, layout-based spatial Action Grounding, a Structured HOI Attention to enforce interaction topology, and HOI RoPE to disentangle multi-HOI scenes. Trained jointly with modality dropout on our HOI-Edit-44K, along with HOI and object-centric datasets, OneHOI supports layout-guided, layout-free, arbitrary-mask, and mixed-condition control, achieving state-of-the-art results across both HOI generation and editing. Code is available at https://jiuntian.github.io/OneHOI/.
- Reinforcement Learning via Value Gradient Flow
We study behavior-regularized reinforcement learning (RL), where regularization toward a reference distribution (the dataset in offline RL or the base model in LLM RL finetuning) is essential to prevent value over-optimization caused by erroneous out-of-distribution extrapolation. Existing methods either rely on reparameterized policy gradient, which are difficult to scale to large generative models, or on reject sampling, which can be overly conservative when attempting to move beyond the behavior support. In this paper, we propose Value Gradient Flow (VGF), a scalable new paradigm for behavior-regularized RL. VGF casts behavior-regularized RL as an optimal transport problem that maps the reference distribution to the value-induced optimal policy distribution. We solve this transport problem via discrete gradient flow, where value gradients guide particles initialized from the reference distribution. Our analysis shows that VGF imposes regularization implicitly by controlling the transport budget. VGF eliminates explicit policy parameterization while remaining expressive and flexible, this enables adaptive test-time scaling by adjusting the transport budget. Extensive experiments demonstrate that VGF significantly outperforms prior methods, achieving state-of-the-art results on offline RL benchmarks (D4RL, OGBench) and LLM RL tasks. Code and runs can be found at https://ryanxhr.github.io/vgf.
Techmeme(15)
- Anthropic's Mythos adds to concerns about rising workloads for open-source maintainers, as many have already been dealing with a "crazy" number of bug reports (Chris Stokel-Walker/Bloomberg)
Chris Stokel-Walker / Bloomberg : Anthropic's Mythos adds to concerns about rising workloads for open-source maintainers, as many have already been dealing with a “crazy” number of bug reports — Anthropic's Mythos and similar AI tools can identify threats and vulnerabilities faster than small teams can fix them, putting the internet at risk.
- Airbnb launches a pilot in NYC, LA, and other cities that lets users to select from a range of boutique hotels alongside private homes in a bid to boost growth (Stephanie Stacey/Financial Times)
Stephanie Stacey / Financial Times : Airbnb launches a pilot in NYC, LA, and other cities that lets users to select from a range of boutique hotels alongside private homes in a bid to boost growth — Company expands accommodation options alongside home rentals but analysts warn of stiff competition
- Salesforce announces Headless 360, an initiative that will give AI agents access to Salesforce's platform capabilities through APIs, MCP tools or CLI commands (Michael Nuñez/VentureBeat)
Michael Nuñez / VentureBeat : Salesforce announces Headless 360, an initiative that will give AI agents access to Salesforce's platform capabilities through APIs, MCP tools or CLI commands — Salesforce on Wednesday unveiled the most ambitious architectural transformation in its 27-year history, introducing “Headless 360” …
- A profile of OpenTable CEO Debby Soo, who shifted its focus from diners to restaurants; it now seats ~2B diners a year across 65K restaurants, an all-time high (Brent Crane/Bloomberg)
Brent Crane / Bloomberg : A profile of OpenTable CEO Debby Soo, who shifted its focus from diners to restaurants; it now seats ~2B diners a year across 65K restaurants, an all-time high — The pioneering reservation app was losing marquee restaurant groups before its chief executive made some big changes.
- Global DRAM supply is likely to meet only 60% of demand through 2027; memory to hit ~40% of low-end smartphone manufacturing costs by mid-2026, up from 20% now (Nikkei Asia)
Nikkei Asia : Global DRAM supply is likely to meet only 60% of demand through 2027; memory to hit ~40% of low-end smartphone manufacturing costs by mid-2026, up from 20% now — TOKYO/SEOUL — A shortage of memory chips appears likely to continue until around 2027, with the top U.S. and South Korean …
- Autonomous vehicle startups raised a record $21.4B across 34 deals through April 15, up from $5.9B raised across 99 investments globally in all of 2025 (Mary Ann Azevedo/Crunchbase News)
Mary Ann Azevedo / Crunchbase News : Autonomous vehicle startups raised a record $21.4B across 34 deals through April 15, up from $5.9B raised across 99 investments globally in all of 2025 — Funding to autonomous vehicle startups has seen a massive resurgence in 2026, more than tripling so far this year compared to all of 2025 globally, Crunchbase data shows.
- Chinese lidar maker Hesai, a primary supplier for Nvidia's ADAS, announces EXT lidar, calling it the industry's first to integrate spatial and color detection (Reuters)
Reuters : Chinese lidar maker Hesai, a primary supplier for Nvidia's ADAS, announces EXT lidar, calling it the industry's first to integrate spatial and color detection — Hesai (2525.HK), China's leading maker of lidar, has developed a version of the technology used in autonomous driving that is capable …
- Some Mac Mini and Mac Studio models are unavailable or facing up to 12-week wait times in the US, with analysts citing strong demand from AI agent power users (Nicole Nguyen/Wall Street Journal)
Nicole Nguyen / Wall Street Journal : Some Mac Mini and Mac Studio models are unavailable or facing up to 12-week wait times in the US, with analysts citing strong demand from AI agent power users — The scarcity of Apple's littlest Mac comes at a time of high interest from AI power users and a potential product refresh
- The EC awards a six-year, €180M sovereign cloud contract to four European providers as part of a push to reduce the EU's dependence on non-European tech (Leo Marchandon/Reuters)
Leo Marchandon / Reuters : The EC awards a six-year, €180M sovereign cloud contract to four European providers as part of a push to reduce the EU's dependence on non-European tech — The European Commission on Friday awarded a 180 million euro ($212 million) tender for sovereign cloud services to four European providers …
- Sources suggest Anthropic is holding off from a wider Mythos release until it can reliably serve it to customers; Anthropic has suffered outages in recent weeks (Financial Times)
Financial Times : Sources suggest Anthropic is holding off from a wider Mythos release until it can reliably serve it to customers; Anthropic has suffered outages in recent weeks — Dario Amodei met with Susie Wiles despite lawsuits over whether AI lab is a national security threat
- What some leaders think of using universal income to mitigate AI-fueled layoffs: Musk calls it the "best way", OpenAI's policy doc mentions a Public Wealth Fund (Siladitya Ray/Forbes)
Siladitya Ray / Forbes : What some leaders think of using universal income to mitigate AI-fueled layoffs: Musk calls it the “best way”, OpenAI's policy doc mentions a Public Wealth Fund — Topline — Elon Musk on Friday touted what he described as “Universal HIGH INCOME” as a solution to deal …
- Sources: Recursive Superintelligence, a four-month-old start-up developing self-teaching AI and founded by ex-DeepMind and OpenAI engineers, has raised $500M+ (Financial Times)
Financial Times : Sources: Recursive Superintelligence, a four-month-old start-up developing self-teaching AI and founded by ex-DeepMind and OpenAI engineers, has raised $500M+ — Group founded by former engineers at DeepMind and OpenAI secures $4bn valuation in deal with Google's venture arm and Nvidia
- White House says a meeting between Chief of Staff Susie Wiles and Dario Amodei had been "productive and constructive"; source: Scott Bessent joined the meeting (Axios)
Axios : White House says a meeting between Chief of Staff Susie Wiles and Dario Amodei had been “productive and constructive”; source: Scott Bessent joined the meeting — Treasury Secretary Scott Bessent joined a meeting on Friday between White House Chief of Staff Susie Wiles …
- A deep dive into Dwarkesh Patel's interview with Jensen Huang, including Huang's takes on Nvidia's moat and chip sales to China, and reactions to the interview (Zvi Mowshowitz/Don't Worry About the Vase)
Zvi Mowshowitz / Don't Worry About the Vase : A deep dive into Dwarkesh Patel's interview with Jensen Huang, including Huang's takes on Nvidia's moat and chip sales to China, and reactions to the interview — Some podcasts are self-recommending on the 'yep, I'm going to be breaking this one down' level. This was one of those. So here we go.
- Bill Peebles, the researcher behind Sora, is leaving OpenAI, along with Srinivas Narayanan, OpenAI's CTO of enterprise applications (Rebecca Bellan/TechCrunch)
Rebecca Bellan / TechCrunch : Bill Peebles, the researcher behind Sora, is leaving OpenAI, along with Srinivas Narayanan, OpenAI's CTO of enterprise applications — OpenAI is losing two of the architects of its most ambitious moonshots. Kevin Weil, who led the company's science research initiative, and Bill Peebles …
Solidot(15)
- Grinex 交易所声称遭敌对国家黑客入侵
注册于吉尔吉斯斯坦的加密货币交易所 Grinex 宣布暂停运营,它声称遭到敌对国家政府黑客的入侵,被盗走逾 1300 万美元加密货币。攻击针对的目标是该交易所的俄罗斯用户。Grinex 称,攻击的数字痕迹和性质表明,攻击者拥有前所未有的资源和技术,此类资源和技术通常只有敌对国家政府机构才拥有。初步数据表明攻击是经过协调的,旨在直接损害俄罗斯的金融主权。Grinex 被广泛视为是 Garantex 的新名字,Grinex 去年遭到了美国财政部的制裁。区块链研究公司 Elliptic 称,Grinex 与俄罗斯关系密切,是俄罗斯卢布兑换加密资产的最大交易所之一,迄今交易总额逾 60 亿美元。
- 大白鲨面临过热风险
大白鲨需要维持自身体温高于周围的海水,但在气候变化导致海洋变暖的时代,它们面临过热风险。鲨鱼属于中温动物(mesothermic),它们能利用代谢产生的热使体温高于周围海水,这具有演化上的优势,它们能拥有更高的游动速度、更强的捕食能力以及能长距离迁徙。然而在温暖的水域它们面临过热的风险,即身体产生热量的速度超过了散热的速度。中温动物在海洋暖化的时代不得不减缓游动速度、改变血液流动方式或潜入更冷的水域,捕食因人类过度捕捞而日益减少的食物。根据研究人员的计算,一吨重的温血鲨鱼难以在水温高于 17 摄氏度的水域生存。
- 暗能量巡天绘制出迄今最大的高分辨率 3D 宇宙地图
暗能量光谱仪(Dark Energy Spectroscopic Instrument, DESI)为期五年的暗物质巡天绘制出迄今最大的高分辨率 3D 宇宙地图。DESI 安装于美国亚利桑那州 Kitt Peak 国家天文台的 Nicholas U Mayall 4 米望远镜上,搭载可同时观测 5000 个天体光谱的系统,透过量测星系与类星体的红移重建 3D 宇宙分布。DESI 早期资料显示暗能量可能随时间演变,而非固定不变的宇宙学常数。该结果若被完整五年资料证实,将意味着现有宇宙学模型需要修正,甚至可能牵动对基本物理定律的重新理解。目前 DESI 已测量的星系与类星体数量达到过去所有观测总和的六倍,最终累积超过 4700 万个星系与类星体,以及约 2000 万颗恒星,提供前所未有的统计精度,使宇宙在不同时期的膨胀速率与星系分布差异得以被量测,进而检验暗能量是否随时间改变。
- 微软正式将 FAT32 分区大小从 32GB 增加到 2TB
微软最近释出了预览版 Windows 11 Insider Preview Build 26300.8170,其中一项变化是将 FAT32 分区大小从 32GB 增加到 2TB。FAT32 分区大小限制在 32GB 是微软开发者随手设置的,几十年来一直没变,导致当存储卡和 U 盘容量超过 32GB 时,用户只能选择 exFAT 或 NTFS。现在微软终于移除了这一限制,但该大小限制仍然只在命令行里移除。
- 拼多多美团等被罚 36 亿
市场监管总局发表公告,对上海寻梦信息技术有限公司(拼多多)、北京三快科技有限公司(美团)、北京京东叁佰陆拾度电子商务有限公司(京东)、上海拉扎斯信息科技有限公司(原饿了么,现淘宝闪购)、北京抖音科技有限公司(抖音)、浙江淘宝网络有限公司(淘宝)、浙江天猫网络有限公司(天猫)等7家电商平台“幽灵外卖”系列案,依据《中华人民共和国食品安全法》第一百三十一条、《中华人民共和国电子商务法》第八十三条的规定作出行政处罚决定,责令7家电商平台改正违法行为,暂停新增蛋糕店铺 3 至 9 个月不等,并处以罚没款共计 35.97 亿元。同时,依据《中华人民共和国食品安全法实施条例》第七十五条的规定,对 7 家平台企业法定代表人和食品安全总监合计处以罚款 1968.74 万元。其中拼多多被罚 1521930222.91元——即 15.2 亿。
- 英伟达 CEO 反对进一步限制向中国出口芯片
英伟达执行长黄仁勋在 Dwarkesh Podcast 节目中反驳美国强化对中国芯片设备出口管制的主张,反对进一步限制对中国出口。他指出,中国具备庞大能源资源,可透过扩大产能弥补制程差距,因此中国无法自主发展 AI 芯片的说法“毫无根据”。黄仁勋强调,美国不该放弃全球第二大的算力市场,若迫使中国加速建立本土 AI 技术系统,将损害美国科技领先地位。他批评,现行政策已间接推动中国芯片产业成长,并指出华为去年营收已创下历史新高。
- 美国科技巨头成功在欧盟法律中将数据中心环境影响列为保密信息
微软以及成员包括亚马逊、Google 和 Meta 的游说组织 DigitalEurope 被发现成功在欧盟法律中争取到一则保密条款,阻止公众获取数据中心环境影响的相关信息。法律学者认为该保密条款可能违反了欧盟的透明度规定。该保密条款是在 2024 年添加到 EU Energy Efficiency Directive 修订版中。欧盟委员会在 2023 年发布了第一版草案,按程序听取利益攸关者的反馈。2024 年初微软和 DigitalEurope 提出了他们的反馈意见:将数据中心的环境足迹信息列为机密和商业敏感信息。2024 年 3 月欧盟委员会发布终稿时逐字逐句的加入了微软和 DigitalEurope 的意见。
- 乌克兰军方开始大规模使用地面武装机器人
当人们还在争论是否应该武装机器人时,乌克兰已经开始将此类地面机器人大规模投入战场。乌克兰总统泽连斯基(Volodymyr Zelenskyy)称该国的地面机器人和无人机成功演示了独自突破俄军阵地并迫使俄军士兵投降。他的说法尚未得到独立验证,但他发布了一则宣传视频,称乌克兰军用机器人过去三个月完成了逾 22000 次任务。他的声明可能指的是去年乌克兰第三独立突击旅的一次任务:无人机配合自杀性地面机器人攻击了俄军阵地,在防御工事被摧毁后,俄军士兵向该部队的机器人投降。乌克兰部署了越来越多的配备机枪和榴弹发射器的地面机器人,有时机器人还被改装成了移动炸弹。乌克兰公司 DevDroid 研发的 Droid TW 12.7 就是一辆配备 M2 勃朗宁机枪的履带式机器人,其最高速度与成人行进速度相当,最远能达到 25 公里,能通过 Starlink 进行卫星通信。
- Firefox 加入了对 Web Serial API 的支持
Firefox Nightly 版加入了对 Web Serial API 的支持,而六年前 Mozilla 以不安全为由反对支持该 API。Web Serial API 允许浏览器与通过串行端口通信的设备交互,此类设备包括 3D 打印机,微控制器如 Arduino 和 ESP32,智能家居面板如 ESPHome,以及通过 USB 或蓝牙模拟串行端口的设备通信。Google Chrome 自 2021 年起加入了对 Web Serial API 的支持,基于 Chromium 的浏览器如 Edge、Opera 和 Vivaldi 也都支持该 API。Mozilla 杰出工程师 Martin Thomso 在 2020 年表示,对于如此强大的功能,无法为用户提供充分的保护,即使用户同意。串行端口是物理连接赋予高度信任的时代的遗物,许多设备允许通过该接口连接的设备在没有任何身份验证的情况下获得管理权限,这一权限甚至超过了 root。两年后 Mozilla 被要求重新考虑其立场,Firefox CTO Bobby Holley 表示 Mozilla 愿意采用和 WebMIDI 相同的附加组件守门机制(add-on-gating mechanism)支持 WebSerial API。Mozilla 目前仍然反对 WebUSB 和 WebHID,而苹果 WebKit 团队仍然对 WebSerial、WebUSB 和 WebHID 持反对态度。
- 大自然仍然在铸造人类基因
一万年对于现代人类的演化历史而言不过是一瞬间,因此科学家认为过去万年就人类演化而言变化甚微。然而根据发表《自然》期刊上的一项研究,科学家分析了 15836 具古代人类遗骸的 DNA,发现了 479 个过去万年受自然选择青睐的基因突变。研究人员认为可能还有数千种基因突变经历了自然选择。研究人员发现,导致麸质过敏腹泻性乳糜泻的突变出现在 4000 年前,意味着它比埃及金字塔建造的时间还要晚。今天全世界可能有 8000 万人患有乳糜泻,这是一种自身免疫性疾病,患者的免疫系统会攻击麸质并损害肠道。由于某种原因,携带这种突变的人比没有这种突变的人有更多的后代。研究人员还发现欧洲居民身上增加吸烟率的基因突变在减少,原因可能与吸烟的危害无关,因为欧洲人吸烟的历史只有 460 年。研究人员承认他们不知道是什么原因导致的。
- 威尼斯如何应对海平面上升
威尼斯是世界文化遗产城市,坐落于威尼斯潟湖内,过去 150 年间饱受洪水冲击。这座城市目前的防洪设施包括位于潟湖入口的三座可移动屏障。发表在《Scientific Reports》期刊上的一项研究评估威尼斯如何应对海平面上升的策略。研究人员估计,如采取更多措施,现有的可动防洪屏障或许能应对最高约 1.25 米的海平面上升。在低排放场景下,因气候变化和地面沉降,这一阈值可能在 2300 年被突破。当海平面上升0.5米时(在高排放场景下,可能发生于 2100 年之前),有必要修建堤坝。封闭潟湖的策略在海平面上升 0.5 米后也是可行的,这将保护城市抵御最高达 10 米的海平面上升。研究人员提出,在海平面上升超过 4.5 米后,迁移城市或将成为必要,预计将发生在 2300 年后。研究人员估算威尼斯现有防洪系统的总成本约为 60 亿欧元,估计建设堤坝的成本将在 5 亿至 45 亿欧元之间。用超级堤坝封闭潟湖最初成本将超过 300 亿欧元,搬迁城市的成本则可能高达 1000 亿欧元。
- SpaceX 将发射 ESA 的 Rosalind Franklin 火星漫游车
NASA 宣布,因种种原因多次推迟的 ESA Rosalind Franklin 火星漫游车将使用 SpaceX 的重型火箭 Falcon Heavy 发射到火星,最早发射时间是 2028 年。Rosalind Franklin 漫游车的历史可上溯到 1997 年,最初是 NASA 和 ESA 合作的项目,原计划 2018 年发射,但因为 NASA 在奥巴马政府任期内削减预算,美国退出了该项目。ESA 因此改与俄罗斯进行合作,计划 2020 年发射,但由于新冠疫情以及降落伞测试失败等问题,发射时间推迟到 2022 年。2022 年的俄乌战争促使 ESA 终止了与俄罗斯的合作。这一次美国再次伸出了援助之手,双方于 2024 年签署了合作协议。
- Discourse 强调会继续开源
日程安排平台 Cal.com 最近宣布从开源转为闭源,理由是 AI 工具更容易从开源代码中发现漏洞,而安全性依赖于模糊,因此闭源有助于提高安全。开源论坛软件 Discourse 对此做出了回应,强调会继续开源,同时表示不敢苟同其对软件安全的看法。Discourse 认为 AI 工具并不需要源代码去发现漏洞,它们针对的是编译后的二进制文件和黑盒 API。闭源并不会让软件更安全。世界最重要的互联网基础设施运行在以 Linux 为代表的开源软件之上,开源代码时刻暴露在无数人的注视之下。它遭受无情的攻击,但也在无止境的加固。这就是安全领域开源真正的意义所在:透明性不是消除风险,但能带来更强大的防御能力。开源带来了一种紧迫感:当代码公开时,你会预料到代码会被仔细审查,因此会更早更积极投入资源,在攻击者前面发现和修复问题。闭源只是给你带来虚幻的安全感。
- 美国主流媒体封禁互联网档案馆的存档机器人
互联网档案馆时光机器(Wayback Machine)存档的内容被媒体广泛使用,然而包括 NYT 和 USA Today 等美国几十家主流新闻网站最近都屏蔽了互联网档案馆的存档爬虫 ia_archiverbot,社交新闻平台 Reddit 也屏蔽了该爬虫,《卫报》没有屏蔽但进行了限制。《卫报》解释称这是为了防止 AI 公司滥用存档目的的内容抓取。NYT 给出的理由类似,称 AI 公司正利用互联网档案馆存档的纽约时报内容训练其模型。AI 公司大量收集互联网内容,而时光机器拥有数十年历史的资料库,被认为是一个极具吸引力的数据源。互联网档案馆运营了 30 年,存档了逾万亿网页。主流网站对其的限制可能削弱其保存工作。互联网档案馆正与 NYT 等媒体进行对话,希望它们最终会改变其做法。
- 新研究再次证实 AI 有害大脑
研究人员在预印本平台 ArXiv 上发表论文《AI assistance reduces persistence and hurts independent performance》,再次证实 AI 有害大脑。研究人员招募了 350 名美国人,任务是解决一些分数方程。半数参与者被随机分配到 AI 组,他们可从一个基于 OpenAI GPT-5 构建的专用聊天机器人获取帮助,另一半必须独立完成。考试进行到一半时,AI 组的访问权限被切断。此举导致 AI 组的正确答案数量急剧下降,很多人干脆放弃考试。这一结果——成绩和毅力双双下降——在一项包含 670 名参与者的更大规模实验中得到了重复验证。研究人员指出,AI 辅助能提高即时表现,但会带来巨大的认知代价。仅仅使用 AI 十分钟就会让人对这项技术产生依赖,一旦停止使用,会导致表现下降和倦怠。