DIGEST · 2026-04-19

OrangeBot.AI Digest — 2026-04-19

85 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Swiss authorities want to reduce dependency on Microsoft (www.swissinfo.ch)
  2. The Bromine Chokepoint (warontherocks.com)
  3. Notion leaks email addresses of all editors of any public page (twitter.com)
  4. Vercel says internal systems hit in breach (decipher.sc)
  5. Notes from the SF peptide scene (12gramsofcarbon.com)
  6. The creative software industry has declared war on Adobe (www.theverge.com)
  7. Vercel April 2026 security incident (www.bleepingcomputer.com)
  8. Airline worker arrested after sharing photos of bomb damage in WhatsApp group (www.lbc.co.uk)
  9. Show HN: Shader Lab, like Photoshop but for shaders (eng.basement.studio)
  10. The seven programming ur-languages (2022) (madhadron.com)
  11. Changes in the system prompt between Claude Opus 4.6 and 4.7 (simonwillison.net)
  12. Archive of BYTE magazine, starting with issue #1 in 1975 (archive.org)
  13. Ask HN: How did you land your first projects as a solo engineer/consultant?
  14. SPEAKE(a)R: Turn Speakers to Microphones for Fun and Profit [pdf] (2017) (www.usenix.org)
  15. The RAM shortage could last years (www.theverge.com)

GitHub Trending(10)

  1. Fincept-Corporation / FinceptTerminal
  2. thunderbird / thunderbolt
  3. tractorjuice / arc-kit
  4. openai / openai-agents-python
  5. pingdotgg / t3code
  6. paperless-ngx / paperless-ngx
  7. ruvnet / RuView
  8. EvoMap / evolver
  9. BasedHardware / omi
  10. Donchitos / Claude-Code-Game-Studios

Product Hunt(15)

  1. Perplexity Personal Computer

    Local files. Native apps. Voice control. Always on.

  2. Avina

    GTM Agents to Find and Reach Your Next Customer

  3. Gemini app for Mac

    Option + Space and Gemini is right there

  4. Verdent 2.0

    Your AI Technical Cofounder

  5. Fixa.dev

    A cloud-native AI agent that can build literally anything

  6. Tell

    Mac widgets, made fun.

  7. Assemble

    One /go command for AI work that remembers — zero runtime

  8. Nibbo

    Family hub with a 3D pet that grows as you get things done

  9. Paperweight

    Cleanup your email and manage your digital footprint

  10. Wyndo

    Weather app that tells you when to walk, bike or eat outside

  11. Vantage in Google Labs

    Practice & assess future-ready skills with AI-simulated team

  12. AGG Loop

    Secure, forever-free localhost tunnels (ex-Deposure).

  13. Creator OS

    Stop missing comments on Instagram.

  14. Vercel Flags

    Feature flags, targeting rules, rollouts. All from Vercel.

  15. Hipocampus

    AI operators that own team workflows

Hugging Face(15)

  1. HY-World 2.0: A Multi-Modal World Model for Reconstructing, Generating, and Simulating 3D Worlds

    We introduce HY-World 2.0, a multi-modal world model framework that advances our prior project HY-World 1.0. HY-World 2.0 accommodates diverse input modalities, including text prompts, single-view images, multi-view images, and videos, and produces 3D world representations. With text or single-view image inputs, the model performs world generation, synthesizing high-fidelity, navigable 3D Gaussian Splatting (3DGS) scenes. This is achieved through a four-stage method: a) Panorama Generation with HY-Pano 2.0, b) Trajectory Planning with WorldNav, c) World Expansion with WorldStereo 2.0, and d) World Composition with WorldMirror 2.0. Specifically, we introduce key innovations to enhance panorama fidelity, enable 3D scene understanding and planning, and upgrade WorldStereo, our keyframe-based view generation model with consistent memory. We also upgrade WorldMirror, a feed-forward model for universal 3D prediction, by refining model architecture and learning strategy, enabling world reconstruction from multi-view images or videos. Also, we introduce WorldLens, a high-performance 3DGS rendering platform featuring a flexible engine-agnostic architecture, automatic IBL lighting, efficient collision detection, and training-rendering co-design, enabling interactive exploration of 3D worlds with character support. Extensive experiments demonstrate that HY-World 2.0 achieves state-of-the-art performance on several benchmarks among open-source approaches, delivering results comparable to the closed-source model Marble. We release all model weights, code, and technical details to facilitate reproducibility and support further research on 3D world models.

  2. DR^{3}-Eval: Towards Realistic and Reproducible Deep Research Evaluation

    Deep Research Agents (DRAs) aim to solve complex, long-horizon research tasks involving planning, retrieval, multimodal understanding, and report generation, yet their evaluation remains challenging due to dynamic web environments and ambiguous task definitions. We propose DR^{3}-Eval, a realistic and reproducible benchmark for evaluating deep research agents on multimodal, multi-file report generation. DR^{3}-Eval is constructed from authentic user-provided materials and paired with a per-task static research sandbox corpus that simulates open-web complexity while remaining fully verifiable, containing supportive documents, distractors, and noise. Moreover, we introduce a multi-dimensional evaluation framework measuring Information Recall, Factual Accuracy, Citation Coverage, Instruction Following, and Depth Quality, and validate its alignment with human judgments. Experiments with our developed multi-agent system DR^{3}-Agent based on multiple state-of-the-art language models demonstrate that DR^{3}-Eval is highly challenging and reveals critical failure modes in retrieval robustness and hallucination control. Our code and data are publicly available.

  3. RAD-2: Scaling Reinforcement Learning in a Generator-Discriminator Framework

    High-level autonomous driving requires motion planners capable of modeling multimodal future uncertainties while remaining robust in closed-loop interactions. Although diffusion-based planners are effective at modeling complex trajectory distributions, they often suffer from stochastic instabilities and the lack of corrective negative feedback when trained purely with imitation learning. To address these issues, we propose RAD-2, a unified generator-discriminator framework for closed-loop planning. Specifically, a diffusion-based generator is used to produce diverse trajectory candidates, while an RL-optimized discriminator reranks these candidates according to their long-term driving quality. This decoupled design avoids directly applying sparse scalar rewards to the full high-dimensional trajectory space, thereby improving optimization stability. To further enhance reinforcement learning, we introduce Temporally Consistent Group Relative Policy Optimization, which exploits temporal coherence to alleviate the credit assignment problem. In addition, we propose On-policy Generator Optimization, which converts closed-loop feedback into structured longitudinal optimization signals and progressively shifts the generator toward high-reward trajectory manifolds. To support efficient large-scale training, we introduce BEV-Warp, a high-throughput simulation environment that performs closed-loop evaluation directly in Bird's-Eye View feature space via spatial warping. RAD-2 reduces the collision rate by 56% compared with strong diffusion-based planners. Real-world deployment further demonstrates improved perceived safety and driving smoothness in complex urban traffic.

  4. How to Fine-Tune a Reasoning Model? A Teacher-Student Cooperation Framework to Synthesize Student-Consistent SFT Data

    A widely adopted strategy for model enhancement is to use synthetic data generated by a stronger model for supervised fine-tuning (SFT). However, for emerging reasoning models like Qwen3-8B, this approach often fails to improve reasoning capabilities and can even lead to a substantial drop in performance. In this work, we identify substantial stylistic divergence between teacher generated data and the distribution of student as a major factor impacting SFT. To bridge this gap, we propose a Teacher-Student Cooperation Data Synthesis framework (TESSY), which interleaves teacher and student models to alternately generate style and non-style tokens. Consequently, TESSY produces synthetic sequences that inherit the advanced reasoning capabilities of the teacher while maintaining stylistic consistency with the distribution of the student. In experiments on code generation using GPT-OSS-120B as the teacher, fine-tuning Qwen3-8B on teacher-generated data leads to performance drops of 3.25% on LiveCodeBench-Pro and 10.02% on OJBench, whereas TESSY achieves improvements of 11.25% and 6.68%.

  5. GlobalSplat: Efficient Feed-Forward 3D Gaussian Splatting via Global Scene Tokens

    The efficient spatial allocation of primitives serves as the foundation of 3D Gaussian Splatting, as it directly dictates the synergy between representation compactness, reconstruction speed, and rendering fidelity. Previous solutions, whether based on iterative optimization or feed-forward inference, suffer from significant trade-offs between these goals, mainly due to the reliance on local, heuristic-driven allocation strategies that lack global scene awareness. Specifically, current feed-forward methods are largely pixel-aligned or voxel-aligned. By unprojecting pixels into dense, view-aligned primitives, they bake redundancy into the 3D asset. As more input views are added, the representation size increases and global consistency becomes fragile. To this end, we introduce GlobalSplat, a framework built on the principle of align first, decode later. Our approach learns a compact, global, latent scene representation that encodes multi-view input and resolves cross-view correspondences before decoding any explicit 3D geometry. Crucially, this formulation enables compact, globally consistent reconstructions without relying on pretrained pixel-prediction backbones or reusing latent features from dense baselines. Utilizing a coarse-to-fine training curriculum that gradually increases decoded capacity, GlobalSplat natively prevents representation bloat. On RealEstate10K and ACID, our model achieves competitive novel-view synthesis performance while utilizing as few as 16K Gaussians, significantly less than required by dense pipelines, obtaining a light 4MB footprint. Further, GlobalSplat enables significantly faster inference than the baselines, operating under 78 milliseconds in a single forward pass. Project page is available at https://r-itk.github.io/globalsplat/

  6. ASGuard: Activation-Scaling Guard to Mitigate Targeted Jailbreaking Attack

    Large language models (LLMs), despite being safety-aligned, exhibit brittle refusal behaviors that can be circumvented by simple linguistic changes. As tense jailbreaking demonstrates that models refusing harmful requests often comply when rephrased in past tense, a critical generalization gap is revealed in current alignment methods whose underlying mechanisms are poorly understood. In this work, we introduce Activation-Scaling Guard (ASGuard), an insightful, mechanistically-informed framework that surgically mitigates this specific vulnerability. In the first step, we use circuit analysis to identify the specific attention heads causally linked to the targeted jailbreaking such as a tense-changing attack. Second, we train a precise, channel-wise scaling vector to recalibrate the activation of tense vulnerable heads. Lastly, we apply it into a "preventative fine-tuning", forcing the model to learn a more robust refusal mechanism. Across four LLMs, ASGuard effectively reduces the attack success rate of targeted jailbreaking while preserving general capabilities and minimizing over refusal, achieving a Pareto-optimal balance between safety and utility. Our findings underscore how adversarial suffixes suppress the propagation of the refusal-mediating direction, based on mechanistic analysis. Furthermore, our work showcases how a deep understanding of model internals can be leveraged to develop practical, efficient, and targeted methods for adjusting model behavior, charting a course for more reliable and interpretable AI safety.

  7. HiVLA: A Visual-Grounded-Centric Hierarchical Embodied Manipulation System

    While end-to-end Vision-Language-Action (VLA) models offer a promising paradigm for robotic manipulation, fine-tuning them on narrow control data often compromises the profound reasoning capabilities inherited from their base Vision-Language Models (VLMs). To resolve this fundamental trade-off, we propose HiVLA, a visual-grounded-centric hierarchical framework that explicitly decouples high-level semantic planning from low-level motor control. In high-level part, a VLM planner first performs task decomposition and visual grounding to generate structured plans, comprising a subtask instruction and a precise target bounding box. Then, to translate this plan into physical actions, we introduce a flow-matching Diffusion Transformer (DiT) action expert in low-level part equipped with a novel cascaded cross-attention mechanism. This design sequentially fuses global context, high-resolution object-centric crops and skill semantics, enabling the DiT to focus purely on robust execution. Our decoupled architecture preserves the VLM's zero-shot reasoning while allowing independent improvement of both components. Extensive experiments in simulation and the real world demonstrate that HiVLA significantly outperforms state-of-the-art end-to-end baselines, particularly excelling in long-horizon skill composition and the fine-grained manipulation of small objects in cluttered scenes.

  8. Dive into Claude Code: The Design Space of Today's and Future AI Agent Systems

    Claude Code is an agentic coding tool that can run shell commands, edit files, and call external services on behalf of the user. This study describes its comprehensive architecture by analyzing the publicly available TypeScript source code and further comparing it with OpenClaw, an independent open-source AI agent system that answers many of the same design questions from a different deployment context. Our analysis identifies five human values, philosophies, and needs that motivate the architecture (human decision authority, safety and security, reliable execution, capability amplification, and contextual adaptability) and traces them through thirteen design principles to specific implementation choices. The core of the system is a simple while-loop that calls the model, runs tools, and repeats. Most of the code, however, lives in the systems around this loop: a permission system with seven modes and an ML-based classifier, a five-layer compaction pipeline for context management, four extensibility mechanisms (MCP, plugins, skills, and hooks), a subagent delegation mechanism with worktree isolation, and append-oriented session storage. A comparison with OpenClaw, a multi-channel personal assistant gateway, shows that the same recurring design questions produce different architectural answers when the deployment context changes: from per-action safety classification to perimeter-level access control, from a single CLI loop to an embedded runtime within a gateway control plane, and from context-window extensions to gateway-wide capability registration. We finally identify six open design directions for future agent systems, grounded in recent empirical, architectural, and policy literature.

  9. UniDoc-RL: Coarse-to-Fine Visual RAG with Hierarchical Actions and Dense Rewards

    Retrieval-Augmented Generation (RAG) extends Large Vision-Language Models (LVLMs) with external visual knowledge. However, existing visual RAG systems typically rely on generic retrieval signals that overlook the fine-grained visual semantics essential for complex reasoning. To address this limitation, we propose UniDoc-RL, a unified reinforcement learning framework in which an LVLM agent jointly performs retrieval, reranking, active visual perception, and reasoning. UniDoc-RL formulates visual information acquisition as a sequential decision-making problem with a hierarchical action space. Specifically, it progressively refines visual evidence from coarse-grained document retrieval to fine-grained image selection and active region cropping, allowing the model to suppress irrelevant content and attend to information-dense regions. For effective end-to-end training, we introduce a dense multi-reward scheme that provides task-aware supervision for each action. Based on Group Relative Policy Optimization (GRPO), UniDoc-RL aligns agent behavior with multiple objectives without relying on a separate value network. To support this training paradigm, we curate a comprehensive dataset of high-quality reasoning trajectories with fine-grained action annotations. Experiments on three benchmarks demonstrate that UniDoc-RL consistently surpasses state-of-the-art baselines, yielding up to 17.7% gains over prior RL-based methods.

  10. Switch-KD: Visual-Switch Knowledge Distillation for Vision-Language Models

    Vision-Language Models (VLMs) have shown remarkable capabilities in joint vision-language understanding, but their large scale poses significant challenges for deployment in resource-constrained scenarios. Knowledge Distillation (KD) offers a viable way to improve model capabilities without increasing model size or data requirements, making deployment more efficient. However, applying KD to VLMs is challenged by modality-specific supervision: although multimodal knowledge in VLMs is fused within the language space, current methods supervise each modality separately without explicitly addressing multimodal alignment, leading to inconsistent multimodal knowledge transfer. To address this, we propose Switch-KD, a visual-switch distillation framework that unifies vision-language knowledge transfer within a shared text-probability space. Switch-KD comprises two key components: (1) Visual-Switch Distillation, which switches the student's visual outputs into the teacher's language pathway to construct cross-modal probabilistic references for implicit visual knowledge transfer; and (2) Dynamic Bi-directional Logits Difference (DBiLD) loss, which adaptively aligns informative probability regions while preserving the distributional structures of teacher and student through bidirectional supervision. Guided by Switch-KD, a 0.5B TinyLLaVA effectively distills rich multimodal knowledge from its 3B teacher, yielding an average improvement of 3.6 points across 10 multimodal benchmarks without any architectural modification.

  11. LeapAlign: Post-Training Flow Matching Models at Any Generation Step by Building Two-Step Trajectories

    This paper focuses on the alignment of flow matching models with human preferences. A promising way is fine-tuning by directly backpropagating reward gradients through the differentiable generation process of flow matching. However, backpropagating through long trajectories results in prohibitive memory costs and gradient explosion. Therefore, direct-gradient methods struggle to update early generation steps, which are crucial for determining the global structure of the final image. To address this issue, we introduce LeapAlign, a fine-tuning method that reduces computational cost and enables direct gradient propagation from reward to early generation steps. Specifically, we shorten the long trajectory into only two steps by designing two consecutive leaps, each skipping multiple ODE sampling steps and predicting future latents in a single step. By randomizing the start and end timesteps of the leaps, LeapAlign leads to efficient and stable model updates at any generation step. To better use such shortened trajectories, we assign higher training weights to those that are more consistent with the long generation path. To further enhance gradient stability, we reduce the weights of gradient terms with large magnitude, instead of completely removing them as done in previous works. When fine-tuning the Flux model, LeapAlign consistently outperforms state-of-the-art GRPO-based and direct-gradient methods across various metrics, achieving superior image quality and image-text alignment.

  12. Representations Before Pixels: Semantics-Guided Hierarchical Video Prediction

    Accurate future video prediction requires both high visual fidelity and consistent scene semantics, particularly in complex dynamic environments such as autonomous driving. We present Re2Pix, a hierarchical video prediction framework that decomposes forecasting into two stages: semantic representation prediction and representation-guided visual synthesis. Instead of directly predicting future RGB frames, our approach first forecasts future scene structure in the feature space of a frozen vision foundation model, and then conditions a latent diffusion model on these predicted representations to render photorealistic frames. This decomposition enables the model to focus first on scene dynamics and then on appearance generation. A key challenge arises from the train-test mismatch between ground-truth representations available during training and predicted ones used at inference. To address this, we introduce two conditioning strategies, nested dropout and mixed supervision, that improve robustness to imperfect autoregressive predictions. Experiments on challenging driving benchmarks demonstrate that the proposed semantics-first design significantly improves temporal semantic consistency, perceptual quality, and training efficiency compared to strong diffusion baselines. We provide the implementation code at https://github.com/Sta8is/Re2Pix

  13. KV Packet: Recomputation-Free Context-Independent KV Caching for LLMs

    Large Language Models (LLMs) rely heavily on Key-Value (KV) caching to minimize inference latency. However, standard KV caches are context-dependent: reusing a cached document in a new context requires recomputing KV states to account for shifts in attention distribution. Existing solutions such as CacheBlend, EPIC, and SAM-KV mitigate this issue by selectively recomputing a subset of tokens; however, they still incur non-negligible computational overhead (FLOPs) and increased Time-to-First-Token (TTFT) latency. In this paper, we propose KV Packet, a recomputation-free cache reuse framework that treats cached documents as immutable ``packets'' wrapped in light-weight trainable soft-token adapters, which are trained via self-supervised distillation to bridge context discontinuities. Experiments on Llama-3.1 and Qwen2.5 demonstrate that the proposed KV Packet method achieves near-zero FLOPs and lower TTFT than recomputation-based baselines, while retaining F1 scores comparable to those of the full recomputation baseline.

  14. Boosting Visual Instruction Tuning with Self-Supervised Guidance

    Multimodal large language models (MLLMs) perform well on many vision-language tasks but often struggle with vision-centric problems that require fine-grained visual reasoning. Recent evidence suggests that this limitation arises not from weak visual representations, but from under-utilization of visual information during instruction tuning, where many tasks can be partially solved using language priors alone. We propose a simple and lightweight approach that augments visual instruction tuning with a small number of visually grounded self-supervised tasks expressed as natural language instructions. By reformulating classical self-supervised pretext tasks, such as rotation prediction, color matching, and cross-view correspondence, as image-instruction-response triplets, we introduce supervision that cannot be solved without relying on visual evidence. Our approach requires no human annotations, no architectural modifications, and no additional training stages. Across multiple models, training regimes, and benchmarks, injecting only a small fraction (3-10%) of such visually grounded instructions consistently improves performance on vision-centric evaluations. Our findings highlight instruction tuning with visually grounded SSL tasks as a powerful lever for improving visual reasoning in MLLMs through simple adjustments to the training data distribution. Code available at: https://github.com/sirkosophia/V-GIFT

  15. TRACER: Trace-Based Adaptive Cost-Efficient Routing for LLM Classification

    Every call to an LLM classification endpoint produces a labeled input-output pair already retained in production logs. These pairs constitute a free, growing training set: a lightweight surrogate trained on them can absorb a significant portion of future traffic at near-zero marginal inference cost. The open questions are when the surrogate is reliable enough to deploy, what it handles versus defers, and how that boundary evolves as data accumulates. We introduce TRACER (Trace-based Adaptive Cost-Efficient Routing), an open-source system that trains ML surrogates on an LLM's own production traces and governs deployment through a parity gate: the surrogate is activated only when its agreement with the LLM exceeds a user-specified threshold α. To make the routing boundary transparent, TRACER generates interpretability artifacts describing which input regions the surrogate handles, where it plateaus, and why it defers. On a 77-class intent benchmark with a Sonnet 4.6 teacher, TRACER achieves 83-100% surrogate coverage depending on the quality target α; on a 150-class benchmark, the surrogate fully replaces the teacher. On a natural language inference task, the parity gate correctly refuses deployment because the embedding representation cannot support reliable separation. The system is available as open-source software.

Techmeme(15)

  1. Fermi CEO Toby Neugebauer is leaving the company, as the data center developer faces issues in securing an anchor tenant and construction delays (Amy Harder/Axios)

    Amy Harder / Axios : Fermi CEO Toby Neugebauer is leaving the company, as the data center developer faces issues in securing an anchor tenant and construction delays —  The world's largest data center project — backed by Trump allies and bearing his name — is stalled by delays and logistical hurdles that could stop it before it even starts.

  2. Vercel says its internal systems were accessed via a compromised third-party AI tool, after a user with a ShinyHunters handle claimed a breach on BreachForums (Lawrence Abrams/BleepingComputer)

    Lawrence Abrams / BleepingComputer : Vercel says its internal systems were accessed via a compromised third-party AI tool, after a user with a ShinyHunters handle claimed a breach on BreachForums —  Cloud development platform Vercel has disclosed a security incident after threat actors claimed to have breached its systems and are attempting to sell stolen data.

  3. Sources say NSA is using Mythos Preview, and a source says it is also being used widely within the DoD, despite Anthropic's designation as a supply chain risk (Axios)

    Axios : Sources say NSA is using Mythos Preview, and a source says it is also being used widely within the DoD, despite Anthropic's designation as a supply chain risk —  - The department moved in February to cut off Anthropic and force its vendors to follow suit.

  4. Palantir posts a 22-point summary of Alex Karp's book, promoting hard power, AI weapons and deterrence, while denouncing pluralism and "regressive" cultures (Anthony Ha/TechCrunch)

    Anthony Ha / TechCrunch : Palantir posts a 22-point summary of Alex Karp's book, promoting hard power, AI weapons and deterrence, while denouncing pluralism and “regressive” cultures —  Surveillance and analytics company Palantir recently posted what it called a “brief” 22-point summary of CEO Alexander Karp's book “The Technological Republic.”

  5. Sources: the glowing "26" in Apple's WWDC invite is teasing a revamped Siri, memory shortages may push Mac Studio and touch MacBook Pro launches by a few months (Mark Gurman/Bloomberg)

    Mark Gurman / Bloomberg : Sources: the glowing “26” in Apple's WWDC invite is teasing a revamped Siri, memory shortages may push Mac Studio and touch MacBook Pro launches by a few months —  Also: Memory shortages could push back new Macs.  —  Apple's WWDC 2026 teaser provides a glimpse of the revamped Siri interface coming in iOS 27.

  6. Sources: Google is in talks with Marvell Technology to develop a memory processing unit that works alongside TPUs, and a new TPU for running AI models (Qianer Liu/The Information)

    Qianer Liu / The Information : Sources: Google is in talks with Marvell Technology to develop a memory processing unit that works alongside TPUs, and a new TPU for running AI models —  Google is in talks with Marvell Technology to develop two new chips aimed at running AI models more efficiently, according to two people with direct knowledge of the discussions.

  7. A profile of Quince, an online luxury DTC brand valued at $10B+, which has found success by using data analysis and close manufacturer ties to keep prices low (Amanda Mull/Bloomberg)

    Amanda Mull / Bloomberg : A profile of Quince, an online luxury DTC brand valued at $10B+, which has found success by using data analysis and close manufacturer ties to keep prices low —  The online apparel retailer has become a $10 billion-plus e-commerce giant by mastering the art of the supply chain.

  8. A look at the AI nonprofit METR, whose time-horizon metrics are used by AI researchers and Wall Street investors to track the rapid development of AI systems (Kevin Roose/New York Times)

    Kevin Roose / New York Times : A look at the AI nonprofit METR, whose time-horizon metrics are used by AI researchers and Wall Street investors to track the rapid development of AI systems —  A chart created by METR, a nonprofit A.I. organization, has become an industrywide obsession as it measures the rapid development of big A.I. systems.

  9. Expo, which develops an eponymous React Native framework and provides cloud services for building cross-platform apps, raised a $45M Series B led by Georgian (Maria Deutscher/SiliconANGLE)

    Maria Deutscher / SiliconANGLE : Expo, which develops an eponymous React Native framework and provides cloud services for building cross-platform apps, raised a $45M Series B led by Georgian —  Expo, the developer of a popular open-source tool for building cross-platform applications, today announced that it has raised $45 million in funding.

  10. Binance and Bitget to probe a rally in RaveDAO's RAVE token, which surged 4,500% in a week, after ZachXBT alleged RAVE insiders engineered a large short squeeze (Francisco Rodrigues/CoinDesk)

    Francisco Rodrigues / CoinDesk : Binance and Bitget to probe a rally in RaveDAO's RAVE token, which surged 4,500% in a week, after ZachXBT alleged RAVE insiders engineered a large short squeeze —  Nearly 90% of RAVE's supply was concentrated in just three wallets, and millions of tokens were transferred to exchanges before the price surge.

  11. At the Beijing half-marathon, several humanoid robots beat human winners by 10+ minutes; a robot made by Honor beat the human world record held by Jacob Kiplimo (Reuters)

    Reuters : At the Beijing half-marathon, several humanoid robots beat human winners by 10+ minutes; a robot made by Honor beat the human world record held by Jacob Kiplimo —  Dozens of Chinese-made humanoid robots showed off their fast-improving athleticism and autonomous navigation skills …

  12. Tesla expands its robotaxi service to Dallas and Houston after launching in Austin last year and starting to offer rides without safety drivers in January 2026 (Anthony Ha/TechCrunch)

    Anthony Ha / TechCrunch : Tesla expands its robotaxi service to Dallas and Houston after launching in Austin last year and starting to offer rides without safety drivers in January 2026 —  Tesla is expanding its robotaxi service to Dallas and Houston, according to a social media post from the company.

  13. A look at Dylan Patel's SemiAnalysis, an AI newsletter and research firm that expects $100M+ in 2026 revenue from subscriptions and AI supply chain research (Abram Brown/The Information)

    Abram Brown / The Information : A look at Dylan Patel's SemiAnalysis, an AI newsletter and research firm that expects $100M+ in 2026 revenue from subscriptions and AI supply chain research —  For Dylan Patel, founder of SemiAnalysis, an influential AI industry newsletter and research firm, Nvidia deserves the kind …

  14. Appfigures: app releases across the App Store and Google Play grew 60% YoY in Q1, with App Store releases alone up 80%, possibly driven by AI coding tools (Sarah Perez/TechCrunch)

    Sarah Perez / TechCrunch : Appfigures: app releases across the App Store and Google Play grew 60% YoY in Q1, with App Store releases alone up 80%, possibly driven by AI coding tools —  Everyone said AI would kill apps.  Instead, new app launches are soaring.  —  According to a new analysis from market intelligence …

  15. Mistral, which once aimed for top open models, now leans on being an alternative to Chinese and US labs, says it's on track for $80M in monthly revenue by Dec. (Iain Martin/Forbes)

    Iain Martin / Forbes : Mistral, which once aimed for top open models, now leans on being an alternative to Chinese and US labs, says it's on track for $80M in monthly revenue by Dec. —  Paris-based Mistral wanted to develop a top-tier AI model to rival OpenAI and Anthropic.  That didn't work out.

Solidot(15)

  1. NASA 关闭旅行者 1 号的 LECP 仪器以节省电力维持运行

    NASA JPL 工程师于 4 月 17 日向旅行者 1 号发送指令,关闭低能带电粒子仪(LECP)以节省电力维持探测器的运行。旅行者 1 号与旅行者 2 号都携带了 10 个科学仪器,过去 49 年已有 7 个仪器关闭,剩下 3 个仪器——LECP,三轴磁通门磁强计(MAG)和电浆波系统(PWS)还在运行中。LECP 被用于测量低能带电粒子,包括来自太阳系和银河系的离子、电子和宇宙射线。LECP 提供了星际介质结构的关键数据,探测到日球层外的压力锋面和粒子密度变化区域。旅行者 1 号是距离地球最遥远的人造探测器,也是唯一能探测此类信息的探测器。旅行者号使用放射性同位素热电机作为动力,每年损失约 4 瓦的功率,今年 2 月 27 日工程师检测到功率的意外下降,如果功率再下降将会诱发低电压故障保护系统,为保护探测器而关闭组件。而恢复组件将是一个漫长且充满风险的过程。工程团队因此决定主动关闭仪器以节省电力,他们在三个仍正常工作的仪器中选择了 LECP。工程师估计关闭 LECP 能为旅行者 1 号提供一年的喘息时间,计划在此期间完善名为“Big Bang”的节能方案,以进一步延长旅行者号的运行时间。

  2. 人形机器人打破人类半马世界纪录

    在周日(4 月 19 日)举行的 2026 北京亦庄人形机器人半程马拉松赛上,机器人首次跑赢了人类半马世界纪录,前三名成绩全部优于人类纪录。荣耀旗下的齐天大圣队、雷霆闪电队、星火燎原队分别夺得冠军、亚军、季军,净用时分别为 50 分 26 秒、50 分 56 秒、53 分 01 秒,均优于乌干达名将基普利莫在今年 3 月里斯本半程马拉松赛中创造的 57 分 20 秒人类男子半马世界纪录。这一成绩与去年相比进步显著:一年前在亦庄举行的首届人形机器人半程马拉松上,天工队的人形机器人以 2 小时 40 分 42 秒的成绩夺冠。一年后,参赛队伍数量从 20 支增至百余支,是去年的五倍。整体完赛成绩大幅提升,人形机器人续航更稳、步态更顺、算法更稳。途中仍有部分机器人摔倒,或撞上赛道护栏。技术层面同样有进步:去年工程师们还可以跟着机器人一路小跑“陪跑",今年很多机器人的速度大幅提升,工作人员不得不改坐高尔夫球车跟在后面。去年,天工是赛场上唯一采用"半自主"方式参赛的选手,其余几乎全部依赖遥控。今年,约四成队伍已实现自主导航。为鼓励这一趋势,赛事规则规定遥控操作组成绩须乘以 1.2 加权系数,变相惩罚依赖人工遥控的队伍。

  3. 卫星无人机图像显示美国四成数据中心可能延期

    硅谷科技巨头斥资数千亿美元建造 AI 数据中心,然而大规模的数据中心建设面临巨大的施工和电力挑战。卫星无人机图像显示美国近四成数据中心项目可能无法按计划在今年内完工。分析显示,微软、甲骨文和 OpenAI 等公司的数据中心项目很可能延期三个月以上完工。数据中心项目延误的原因包括劳动力、电力和设备长期短缺,以及获得必要许可的繁琐流程。参与 OpenAI 项目的建筑公司高管提到,他们缺乏足够的技工如电工和管道工去同时开展多个数据中心项目。因巨大的电力需求,科技巨头甚至在建自己的发电厂,大量使用安装在半挂式卡车上的移动式燃气发电机,以及最初为飞机和军舰设计的涡轮发动机。

  4. 内存芯片短缺可能持续到 2030 年

    SK 集团会长崔泰源表示,全球内存芯片晶圆短缺问题可能会一直持续到 2030 年。SK 集团子公司 SK 海力士是最大的 HBM 芯片供应商,其市场份额高达 57%,它同时也是 DRAM 市场的第二大厂商,占据了 32% 的份额。英伟达的 AI 芯片主要使用的是 HBM 内存。崔泰源称 AI 芯片消耗了大量的 HBM,而 HBM 的生产会消耗大量的晶圆。增加晶圆的生产需要至少四到五年时间,目前的短缺可能会一直持续到 2030 年,该公司预计晶圆缺口将超过 20%。他还表示 SK 海力士将努力制定稳定 DRAM 价格的策略。

  5. 果糖不只是糖,它更像是激素

    根据发表在《Nature Metabolism》期刊上的一项研究,果糖在化学结构上与葡萄糖几乎完全相同,然而在行为上几乎完全不同,它更像是激素。当蔗糖和玉米糖浆在口腔内溶解成葡萄糖和果糖,其比例基本相同。葡萄糖会促使胰岛素水平上升,被细胞吸收,肝脏将不需要的部分储存为糖原,任何多余的糖原都会在严格调控下转化为脂肪。果糖的代谢途径则完全不同。它绕过了糖酵解途径中最重要的调控酶——果糖激酶 1 (PFK1)。结果是果糖代谢几乎没有“关闭”机制。果糖不仅仅是一种热量,它是一种代谢信号,能以与葡萄糖截然不同的方式促进脂肪的生成和储存。果糖会告诉肝脏制造脂肪,为饥荒做准备。这对于一只靠秋季浆果增肥的熊而言有意义,但对于春季饮碳酸饮料的人类而言不是好事。过去二十年富裕国家的含糖饮料消费量一直在下降,但肥胖率仍然在持续上升,直到 GLP-1 减肥药流行后才放缓。研究人员怀疑可能存在饮用果糖的滞后效应:果糖是一种缓慢起效的毒药,而不是一种快速增加的热量来源。水果含有果糖,但也含有纤维维生素等其它成分,这些成分会减缓果糖的吸收,减缓其影响,但含有高浓度果糖的碳酸饮料则是另一回事了。

  6. Grinex 交易所声称遭敌对国家黑客入侵

    注册于吉尔吉斯斯坦的加密货币交易所 Grinex 宣布暂停运营,它声称遭到敌对国家政府黑客的入侵,被盗走逾 1300 万美元加密货币。攻击针对的目标是该交易所的俄罗斯用户。Grinex 称,攻击的数字痕迹和性质表明,攻击者拥有前所未有的资源和技术,此类资源和技术通常只有敌对国家政府机构才拥有。初步数据表明攻击是经过协调的,旨在直接损害俄罗斯的金融主权。Grinex 被广泛视为是 Garantex 的新名字,Grinex 去年遭到了美国财政部的制裁。区块链研究公司 Elliptic 称,Grinex 与俄罗斯关系密切,是俄罗斯卢布兑换加密资产的最大交易所之一,迄今交易总额逾 60 亿美元。

  7. 大白鲨面临过热风险

    大白鲨需要维持自身体温高于周围的海水,但在气候变化导致海洋变暖的时代,它们面临过热风险。鲨鱼属于中温动物(mesothermic),它们能利用代谢产生的热使体温高于周围海水,这具有演化上的优势,它们能拥有更高的游动速度、更强的捕食能力以及能长距离迁徙。然而在温暖的水域它们面临过热的风险,即身体产生热量的速度超过了散热的速度。中温动物在海洋暖化的时代不得不减缓游动速度、改变血液流动方式或潜入更冷的水域,捕食因人类过度捕捞而日益减少的食物。根据研究人员的计算,一吨重的温血鲨鱼难以在水温高于 17 摄氏度的水域生存。

  8. 暗能量巡天绘制出迄今最大的高分辨率 3D 宇宙地图

    暗能量光谱仪(Dark Energy Spectroscopic Instrument, DESI)为期五年的暗物质巡天绘制出迄今最大的高分辨率 3D 宇宙地图。DESI 安装于美国亚利桑那州 Kitt Peak 国家天文台的 Nicholas U Mayall 4 米望远镜上,搭载可同时观测 5000 个天体光谱的系统,透过量测星系与类星体的红移重建 3D 宇宙分布。DESI 早期资料显示暗能量可能随时间演变,而非固定不变的宇宙学常数。该结果若被完整五年资料证实,将意味着现有宇宙学模型需要修正,甚至可能牵动对基本物理定律的重新理解。目前 DESI 已测量的星系与类星体数量达到过去所有观测总和的六倍,最终累积超过 4700 万个星系与类星体,以及约 2000 万颗恒星,提供前所未有的统计精度,使宇宙在不同时期的膨胀速率与星系分布差异得以被量测,进而检验暗能量是否随时间改变。

  9. 微软正式将 FAT32 分区大小从 32GB 增加到 2TB

    微软最近释出了预览版 Windows 11 Insider Preview Build 26300.8170,其中一项变化是将 FAT32 分区大小从 32GB 增加到 2TB。FAT32 分区大小限制在 32GB 是微软开发者随手设置的,几十年来一直没变,导致当存储卡和 U 盘容量超过 32GB 时,用户只能选择 exFAT 或 NTFS。现在微软终于移除了这一限制,但该大小限制仍然只在命令行里移除。

  10. 拼多多美团等被罚 36 亿

    市场监管总局发表公告,对上海寻梦信息技术有限公司(拼多多)、北京三快科技有限公司(美团)、北京京东叁佰陆拾度电子商务有限公司(京东)、上海拉扎斯信息科技有限公司(原饿了么,现淘宝闪购)、北京抖音科技有限公司(抖音)、浙江淘宝网络有限公司(淘宝)、浙江天猫网络有限公司(天猫)等7家电商平台“幽灵外卖”系列案,依据《中华人民共和国食品安全法》第一百三十一条、《中华人民共和国电子商务法》第八十三条的规定作出行政处罚决定,责令7家电商平台改正违法行为,暂停新增蛋糕店铺 3 至 9 个月不等,并处以罚没款共计 35.97 亿元。同时,依据《中华人民共和国食品安全法实施条例》第七十五条的规定,对 7 家平台企业法定代表人和食品安全总监合计处以罚款 1968.74 万元。其中拼多多被罚 1521930222.91元——即 15.2 亿。

  11. 英伟达 CEO 反对进一步限制向中国出口芯片

    英伟达执行长黄仁勋在 Dwarkesh Podcast 节目中反驳美国强化对中国芯片设备出口管制的主张,反对进一步限制对中国出口。他指出,中国具备庞大能源资源,可透过扩大产能弥补制程差距,因此中国无法自主发展 AI 芯片的说法“毫无根据”。黄仁勋强调,美国不该放弃全球第二大的算力市场,若迫使中国加速建立本土 AI 技术系统,将损害美国科技领先地位。他批评,现行政策已间接推动中国芯片产业成长,并指出华为去年营收已创下历史新高。

  12. 美国科技巨头成功在欧盟法律中将数据中心环境影响列为保密信息

    微软以及成员包括亚马逊、Google 和 Meta 的游说组织 DigitalEurope 被发现成功在欧盟法律中争取到一则保密条款,阻止公众获取数据中心环境影响的相关信息。法律学者认为该保密条款可能违反了欧盟的透明度规定。该保密条款是在 2024 年添加到 EU Energy Efficiency Directive 修订版中。欧盟委员会在 2023 年发布了第一版草案,按程序听取利益攸关者的反馈。2024 年初微软和 DigitalEurope 提出了他们的反馈意见:将数据中心的环境足迹信息列为机密和商业敏感信息。2024 年 3 月欧盟委员会发布终稿时逐字逐句的加入了微软和 DigitalEurope 的意见。

  13. 乌克兰军方开始大规模使用地面武装机器人

    当人们还在争论是否应该武装机器人时,乌克兰已经开始将此类地面机器人大规模投入战场。乌克兰总统泽连斯基(Volodymyr Zelenskyy)称该国的地面机器人和无人机成功演示了独自突破俄军阵地并迫使俄军士兵投降。他的说法尚未得到独立验证,但他发布了一则宣传视频,称乌克兰军用机器人过去三个月完成了逾 22000 次任务。他的声明可能指的是去年乌克兰第三独立突击旅的一次任务:无人机配合自杀性地面机器人攻击了俄军阵地,在防御工事被摧毁后,俄军士兵向该部队的机器人投降。乌克兰部署了越来越多的配备机枪和榴弹发射器的地面机器人,有时机器人还被改装成了移动炸弹。乌克兰公司 DevDroid 研发的 Droid TW 12.7 就是一辆配备 M2 勃朗宁机枪的履带式机器人,其最高速度与成人行进速度相当,最远能达到 25 公里,能通过 Starlink 进行卫星通信。

  14. Firefox 加入了对 Web Serial API 的支持

    Firefox Nightly 版加入了对 Web Serial API 的支持,而六年前 Mozilla 以不安全为由反对支持该 API。Web Serial API 允许浏览器与通过串行端口通信的设备交互,此类设备包括 3D 打印机,微控制器如 Arduino 和 ESP32,智能家居面板如 ESPHome,以及通过 USB 或蓝牙模拟串行端口的设备通信。Google Chrome 自 2021 年起加入了对 Web Serial API 的支持,基于 Chromium 的浏览器如 Edge、Opera 和 Vivaldi 也都支持该 API。Mozilla 杰出工程师 Martin Thomso 在 2020 年表示,对于如此强大的功能,无法为用户提供充分的保护,即使用户同意。串行端口是物理连接赋予高度信任的时代的遗物,许多设备允许通过该接口连接的设备在没有任何身份验证的情况下获得管理权限,这一权限甚至超过了 root。两年后 Mozilla 被要求重新考虑其立场,Firefox CTO Bobby Holley 表示 Mozilla 愿意采用和 WebMIDI 相同的附加组件守门机制(add-on-gating mechanism)支持 WebSerial API。Mozilla 目前仍然反对 WebUSB 和 WebHID,而苹果 WebKit 团队仍然对 WebSerial、WebUSB 和 WebHID 持反对态度。

  15. 大自然仍然在铸造人类基因

    一万年对于现代人类的演化历史而言不过是一瞬间,因此科学家认为过去万年就人类演化而言变化甚微。然而根据发表《自然》期刊上的一项研究,科学家分析了 15836 具古代人类遗骸的 DNA,发现了 479 个过去万年受自然选择青睐的基因突变。研究人员认为可能还有数千种基因突变经历了自然选择。研究人员发现,导致麸质过敏腹泻性乳糜泻的突变出现在 4000 年前,意味着它比埃及金字塔建造的时间还要晚。今天全世界可能有 8000 万人患有乳糜泻,这是一种自身免疫性疾病,患者的免疫系统会攻击麸质并损害肠道。由于某种原因,携带这种突变的人比没有这种突变的人有更多的后代。研究人员还发现欧洲居民身上增加吸烟率的基因突变在减少,原因可能与吸烟的危害无关,因为欧洲人吸烟的历史只有 460 年。研究人员承认他们不知道是什么原因导致的。