OrangeBot.AI Digest — 2026-04-03
82 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Oracle Files H-1B Visa Petitions Amid Mass Layoffs (nationaltoday.com)
- Artemis II crew take 'spectacular' image of Earth (www.bbc.com)
- iNaturalist (www.inaturalist.org)
- We replaced RAG with a virtual filesystem for our AI documentation assistant (www.mintlify.com)
- OpenClaw privilege-escalation bug (old.reddit.com)
- F-15E jet shot down over Iran (www.theguardian.com)
- Solar and batteries can power the world (nworbmot.org)
- Marc Andreessen is wrong about introspection (www.joanwestenberg.com)
- SSH certificates: the better SSH experience (jpmens.net)
- Show HN: I built a frontpage for personal blogs (text.blogosphere.app)
- Samsung Magician disk utility takes 18 steps and two reboots to uninstall (chalmovsky.com)
- Critics say EU risks ceding control of its tech laws under U.S. pressure (www.politico.eu)
- What Category Theory Teaches Us About DataFrames (mchav.github.io)
- April 2026 TLDR Setup for Ollama and Gemma 4 26B on a Mac mini (gist.github.com)
- NHS staff refusing to use FDP over Palantir ethical concerns (www.freevacy.com)
GitHub Trending(7)
Product Hunt(15)
- Dashla
Tesla vehicle status, navigation, map + more in a dashboard
- FindThem
Describe ideal lead or investor - get their Linkedin & email
- MAI-Transcribe-1
Production ASR for noisy multilingual audio
- Vxero Neo
SSH-native CLI that manage servers, apps, infrastructure
- Codictate
Free dictation for Any language and any application
- StampCut
Stamp the world around you!
- GeneratePPT
Instantly generated simple, design-forward slides
- EmDash CMS
EmDash is a new open-source CMS from Cloudflare
- ZooClaw
Your proactive team of AI specialists in one place
- Qwen3.6-Plus
Multimodal AI optimized for real-world coding agents
- Package Mate
Master your macOS dev environment from the terminal
- VoiceOS
Say it and it's done. Work 10x faster with your voice.
- Google Gemma 4
Google's most intelligent open models to date
- Otto by Audos.com
Your AI co-founder that builds, launches, and sells for you
- Straude
Strava for Claude Code / Global Tokenmaxxing Leaderboard
Hugging Face(15)
- DataFlex: A Unified Framework for Data-Centric Dynamic Training of Large Language Models
Data-centric training has emerged as a promising direction for improving large language models (LLMs) by optimizing not only model parameters but also the selection, composition, and weighting of training data during optimization. However, existing approaches to data selection, data mixture optimization, and data reweighting are often developed in isolated codebases with inconsistent interfaces, hindering reproducibility, fair comparison, and practical integration. In this paper, we present DataFlex, a unified data-centric dynamic training framework built upon LLaMA-Factory. DataFlex supports three major paradigms of dynamic data optimization: sample selection, domain mixture adjustment, and sample reweighting, while remaining fully compatible with the original training workflow. It provides extensible trainer abstractions and modular components, enabling a drop-in replacement for standard LLM training, and unifies key model-dependent operations such as embedding extraction, inference, and gradient computation, with support for large-scale settings including DeepSpeed ZeRO-3. We conduct comprehensive experiments across multiple data-centric methods. Dynamic data selection consistently outperforms static full-data training on MMLU across both Mistral-7B and Llama-3.2-3B. For data mixture, DoReMi and ODM improve both MMLU accuracy and corpus-level perplexity over default proportions when pretraining Qwen2.5-1.5B on SlimPajama at 6B and 30B token scales. DataFlex also achieves consistent runtime improvements over original implementations. These results demonstrate that DataFlex provides an effective, efficient, and reproducible infrastructure for data-centric dynamic training of LLMs.
- The Latent Space: Foundation, Evolution, Mechanism, Ability, and Outlook
Latent space is rapidly emerging as a native substrate for language-based models. While modern systems are still commonly understood through explicit token-level generation, an increasing body of work shows that many critical internal processes are more naturally carried out in continuous latent space than in human-readable verbal traces. This shift is driven by the structural limitations of explicit-space computation, including linguistic redundancy, discretization bottlenecks, sequential inefficiency, and semantic loss. This survey aims to provide a unified and up-to-date landscape of latent space in language-based models. We organize the survey into five sequential perspectives: Foundation, Evolution, Mechanism, Ability, and Outlook. We begin by delineating the scope of latent space, distinguishing it from explicit or verbal space and from the latent spaces commonly studied in generative visual models. We then trace the field's evolution from early exploratory efforts to the current large-scale expansion. To organize the technical landscape, we examine existing work through the complementary lenses of mechanism and ability. From the perspective of Mechanism, we identify four major lines of development: Architecture, Representation, Computation, and Optimization. From the perspective of Ability, we show how latent space supports a broad capability spectrum spanning Reasoning, Planning, Modeling, Perception, Memory, Collaboration, and Embodiment. Beyond consolidation, we discuss the key open challenges, and outline promising directions for future research. We hope this survey serves not only as a reference for existing work, but also as a foundation for understanding latent space as a general computational and systems paradigm for next-generation intelligence.
- Generative World Renderer
Scaling generative inverse and forward rendering to real-world scenarios is bottlenecked by the limited realism and temporal coherence of existing synthetic datasets. To bridge this persistent domain gap, we introduce a large-scale, dynamic dataset curated from visually complex AAA games. Using a novel dual-screen stitched capture method, we extracted 4M continuous frames (720p/30 FPS) of synchronized RGB and five G-buffer channels across diverse scenes, visual effects, and environments, including adverse weather and motion-blur variants. This dataset uniquely advances bidirectional rendering: enabling robust in-the-wild geometry and material decomposition, and facilitating high-fidelity G-buffer-guided video generation. Furthermore, to evaluate the real-world performance of inverse rendering without ground truth, we propose a novel VLM-based assessment protocol measuring semantic, spatial, and temporal consistency. Experiments demonstrate that inverse renderers fine-tuned on our data achieve superior cross-dataset generalization and controllable generation, while our VLM evaluation strongly correlates with human judgment. Combined with our toolkit, our forward renderer enables users to edit styles of AAA games from G-buffers using text prompts.
- SKILL0: In-Context Agentic Reinforcement Learning for Skill Internalization
Agent skills, structured packages of procedural knowledge and executable resources that agents dynamically load at inference time, have become a reliable mechanism for augmenting LLM agents. Yet inference-time skill augmentation is fundamentally limited: retrieval noise introduces irrelevant guidance, injected skill content imposes substantial token overhead, and the model never truly acquires the knowledge it merely follows. We ask whether skills can instead be internalized into model parameters, enabling zero-shot autonomous behavior without any runtime skill retrieval. We introduce SKILL0, an in-context reinforcement learning framework designed for skill internalization. SKILL0 introduces a training-time curriculum that begins with full skill context and progressively withdraws it. Skills are grouped offline by category and rendered with interaction history into a compact visual context, teaching he model tool invocation and multi-turn task completion. A Dynamic Curriculum then evaluates each skill file's on-policy helpfulness, retaining only those from which the current policy still benefits within a linearly decaying budget, until the agent operates in a fully zero-shot setting. Extensive agentic experiments demonstrate that SKILL0 achieves substantial improvements over the standard RL baseline (+9.7\% for ALFWorld and +6.6\% for Search-QA), while maintaining a highly efficient context of fewer than 0.5k tokens per step. Our code is available at https://github.com/ZJU-REAL/SkillZero.
- EgoSim: Egocentric World Simulator for Embodied Interaction Generation
We introduce EgoSim, a closed-loop egocentric world simulator that generates spatially consistent interaction videos and persistently updates the underlying 3D scene state for continuous simulation. Existing egocentric simulators either lack explicit 3D grounding, causing structural drift under viewpoint changes, or treat the scene as static, failing to update world states across multi-stage interactions. EgoSim addresses both limitations by modeling 3D scenes as updatable world states. We generate embodiment interactions via a Geometry-action-aware Observation Simulation model, with spatial consistency from an Interaction-aware State Updating module. To overcome the critical data bottleneck posed by the difficulty in acquiring densely aligned scene-interaction training pairs, we design a scalable pipeline that extracts static point clouds, camera trajectories, and embodiment actions from in-the-wild large-scale monocular egocentric videos. We further introduce EgoCap, a capture system that enables low-cost real-world data collection with uncalibrated smartphones. Extensive experiments demonstrate that EgoSim significantly outperforms existing methods in terms of visual quality, spatial consistency, and generalization to complex scenes and in-the-wild dexterous interactions, while supporting cross-embodiment transfer to robotic manipulation. Codes and datasets will be open soon. The project page is at egosimulator.github.io.
- Steerable Visual Representations
Pretrained Vision Transformers (ViTs) such as DINOv2 and MAE provide generic image features that can be applied to a variety of downstream tasks such as retrieval, classification, and segmentation. However, such representations tend to focus on the most salient visual cues in the image, with no way to direct them toward less prominent concepts of interest. In contrast, Multimodal LLMs can be guided with textual prompts, but the resulting representations tend to be language-centric and lose their effectiveness for generic visual tasks. To address this, we introduce Steerable Visual Representations, a new class of visual representations, whose global and local features can be steered with natural language. While most vision-language models (e.g., CLIP) fuse text with visual features after encoding (late fusion), we inject text directly into the layers of the visual encoder (early fusion) via lightweight cross-attention. We introduce benchmarks for measuring representational steerability, and demonstrate that our steerable visual features can focus on any desired objects in an image while preserving the underlying representation quality. Our method also matches or outperforms dedicated approaches on anomaly detection and personalized object discrimination, exhibiting zero-shot generalization to out-of-distribution tasks.
- LatentUM: Unleashing the Potential of Interleaved Cross-Modal Reasoning via a Latent-Space Unified Model
Unified models (UMs) hold promise for their ability to understand and generate content across heterogeneous modalities. Compared to merely generating visual content, the use of UMs for interleaved cross-modal reasoning is more promising and valuable, e.g., for solving understanding problems that require dense visual thinking, improving visual generation through self-reflection, or modeling visual dynamics of the physical world guided by stepwise action interventions. However, existing UMs necessitate pixel decoding as a bridge due to their disjoint visual representations for understanding and generation, which is both ineffective and inefficient. In this paper, we introduce LatentUM, a novel unified model that represents all modalities within a shared semantic latent space, eliminating the need for pixel-space mediation between visual understanding and generation. This design naturally enables flexible interleaved cross-modal reasoning and generation. Beyond improved computational efficiency, the shared representation substantially alleviates codec bias and strengthens cross-modal alignment, allowing LatentUM to achieve state-of-the-art performance on the Visual Spatial Planning benchmark, push the limits of visual generation through self-reflection, and support world modeling by predicting future visual states within the shared semantic latent space.
- NearID: Identity Representation Learning via Near-identity Distractors
When evaluating identity-focused tasks such as personalized generation and image editing, existing vision encoders entangle object identity with background context, leading to unreliable representations and metrics. We introduce the first principled framework to address this vulnerability using Near-identity (NearID) distractors, where semantically similar but distinct instances are placed on the exact same background as a reference image, eliminating contextual shortcuts and isolating identity as the sole discriminative signal. Based on this principle, we present the NearID dataset (19K identities, 316K matched-context distractors) together with a strict margin-based evaluation protocol. Under this setting, pre-trained encoders perform poorly, achieving Sample Success Rates (SSR), a strict margin-based identity discrimination metric, as low as 30.7% and often ranking distractors above true cross-view matches. We address this by learning identity-aware representations on a frozen backbone using a two-tier contrastive objective enforcing the hierarchy: same identity > NearID distractor > random negative. This improves SSR to 99.2%, enhances part-level discrimination by 28.0%, and yields stronger alignment with human judgments on DreamBench++, a human-aligned benchmark for personalization. Project page: https://gorluxor.github.io/NearID/
- Omni-SimpleMem: Autoresearch-Guided Discovery of Lifelong Multimodal Agent Memory
AI agents increasingly operate over extended time horizons, yet their ability to retain, organize, and recall multimodal experiences remains a critical bottleneck. Building effective lifelong memory requires navigating a vast design space spanning architecture, retrieval strategies, prompt engineering, and data pipelines; this space is too large and interconnected for manual exploration or traditional AutoML to explore effectively. We deploy an autonomous research pipeline to discover Omni-SimpleMem, a unified multimodal memory framework for lifelong AI agents. Starting from a naïve baseline (F1=0.117 on LoCoMo), the pipeline autonomously executes {sim}50 experiments across two benchmarks, diagnosing failure modes, proposing architectural modifications, and repairing data pipeline bugs, all without human intervention in the inner loop. The resulting system achieves state-of-the-art on both benchmarks, improving F1 by +411% on LoCoMo (0.117to0.598) and +214% on Mem-Gallery (0.254to0.797) relative to the initial configurations. Critically, the most impactful discoveries are not hyperparameter adjustments: bug fixes (+175%), architectural changes (+44%), and prompt engineering (+188% on specific categories) each individually exceed the cumulative contribution of all hyperparameter tuning, demonstrating capabilities fundamentally beyond the reach of traditional AutoML. We provide a taxonomy of six discovery types and identify four properties that make multimodal memory particularly suited for autoresearch, offering guidance for applying autonomous research pipelines to other AI system domains. Code is available at this https://github.com/aiming-lab/SimpleMem.
- Therefore I am. I Think
We consider the question: when a large language reasoning model makes a choice, did it think first and then decide to, or decide first and then think? In this paper, we present evidence that detectable, early-encoded decisions shape chain-of-thought in reasoning models. Specifically, we show that a simple linear probe successfully decodes tool-calling decisions from pre-generation activations with very high confidence, and in some cases, even before a single reasoning token is produced. Activation steering supports this causally: perturbing the decision direction leads to inflated deliberation, and flips behavior in many examples (between 7 - 79% depending on model and benchmark). We also show through behavioral analysis that, when steering changes the decision, the chain-of-thought process often rationalizes the flip rather than resisting it. Together, these results suggest that reasoning models can encode action choices before they begin to deliberate in text.
- VOID: Video Object and Interaction Deletion
Existing video object removal methods excel at inpainting content "behind" the object and correcting appearance-level artifacts such as shadows and reflections. However, when the removed object has more significant interactions, such as collisions with other objects, current models fail to correct them and produce implausible results. We present VOID, a video object removal framework designed to perform physically-plausible inpainting in these complex scenarios. To train the model, we generate a new paired dataset of counterfactual object removals using Kubric and HUMOTO, where removing an object requires altering downstream physical interactions. During inference, a vision-language model identifies regions of the scene affected by the removed object. These regions are then used to guide a video diffusion model that generates physically consistent counterfactual outcomes. Experiments on both synthetic and real data show that our approach better preserves consistent scene dynamics after object removal compared to prior video object removal methods. We hope this framework sheds light on how to make video editing models better simulators of the world through high-level causal reasoning.
- UniDriveVLA: Unifying Understanding, Perception, and Action Planning for Autonomous Driving
Vision-Language-Action (VLA) models have recently emerged in autonomous driving, with the promise of leveraging rich world knowledge to improve the cognitive capabilities of driving systems. However, adapting such models for driving tasks currently faces a critical dilemma between spatial perception and semantic reasoning. Consequently, existing VLA systems are forced into suboptimal compromises: directly adopting 2D Vision-Language Models yields limited spatial perception, whereas enhancing them with 3D spatial representations often impairs the native reasoning capacity of VLMs. We argue that this dilemma largely stems from the coupled optimization of spatial perception and semantic reasoning within shared model parameters. To overcome this, we propose UniDriveVLA, a Unified Driving Vision-Language-Action model based on Mixture-of-Transformers that addresses the perception-reasoning conflict via expert decoupling. Specifically, it comprises three experts for driving understanding, scene perception, and action planning, which are coordinated through masked joint attention. In addition, we combine a sparse perception paradigm with a three-stage progressive training strategy to improve spatial perception while maintaining semantic reasoning capability. Extensive experiments show that UniDriveVLA achieves state-of-the-art performance in open-loop evaluation on nuScenes and closed-loop evaluation on Bench2Drive. Moreover, it demonstrates strong performance across a broad range of perception, prediction, and understanding tasks, including 3D detection, online mapping, motion forecasting, and driving-oriented VQA, highlighting its broad applicability as a unified model for autonomous driving. Code and model have been released at https://github.com/xiaomi-research/unidrivevla
- ASI-Evolve: AI Accelerates AI
Can AI accelerate the development of AI itself? While recent agentic systems have shown strong performance on well-scoped tasks with rapid feedback, it remains unclear whether they can tackle the costly, long-horizon, and weakly supervised research loops that drive real AI progress. We present ASI-Evolve, an agentic framework for AI-for-AI research that closes this loop through a learn-design-experiment-analyze cycle. ASI-Evolve augments standard evolutionary agents with two key components: a cognition base that injects accumulated human priors into each round of exploration, and a dedicated analyzer that distills complex experimental outcomes into reusable insights for future iterations. To our knowledge, ASI-Evolve is the first unified framework to demonstrate AI-driven discovery across three central components of AI development: data, architectures, and learning algorithms. In neural architecture design, it discovered 105 SOTA linear attention architectures, with the best discovered model surpassing DeltaNet by +0.97 points, nearly 3x the gain of recent human-designed improvements. In pretraining data curation, the evolved pipeline improves average benchmark performance by +3.96 points, with gains exceeding 18 points on MMLU. In reinforcement learning algorithm design, discovered algorithms outperform GRPO by up to +12.5 points on AMC32, +11.67 points on AIME24, and +5.04 points on OlympiadBench. We further provide initial evidence that this AI-for-AI paradigm can transfer beyond the AI stack through experiments in mathematics and biomedicine. Together, these results suggest that ASI-Evolve represents a promising step toward enabling AI to accelerate AI across the foundational stages of development, offering early evidence for the feasibility of closed-loop AI research.
- Investigating Autonomous Agent Contributions in the Wild: Activity Patterns and Code Change over Time
The rise of large language models for code has reshaped software development. Autonomous coding agents, able to create branches, open pull requests, and perform code reviews, now actively contribute to real-world projects. Their growing role offers a unique and timely opportunity to investigate AI-driven contributions and their effects on code quality, team dynamics, and software maintainability. In this work, we construct a novel dataset of approximately 110,000 open-source pull requests, including associated commits, comments, reviews, issues, and file changes, collectively representing millions of lines of source code. We compare five popular coding agents, including OpenAI Codex, Claude Code, GitHub Copilot, Google Jules, and Devin, examining how their usage differs in various development aspects such as merge frequency, edited file types, and developer interaction signals, including comments and reviews. Furthermore, we emphasize that code authoring and review are only a small part of the larger software engineering process, as the resulting code must also be maintained and updated over time. Hence, we offer several longitudinal estimates of survival and churn rates for agent-generated versus human-authored code. Ultimately, our findings indicate an increasing agent activity in open-source projects, although their contributions are associated with more churn over time compared to human-authored code.
- CORAL: Towards Autonomous Multi-Agent Evolution for Open-Ended Discovery
Large language model (LLM)-based evolution is a promising approach for open-ended discovery, where progress requires sustained search and knowledge accumulation. Existing methods still rely heavily on fixed heuristics and hard-coded exploration rules, which limit the autonomy of LLM agents. We present CORAL, the first framework for autonomous multi-agent evolution on open-ended problems. CORAL replaces rigid control with long-running agents that explore, reflect, and collaborate through shared persistent memory, asynchronous multi-agent execution, and heartbeat-based interventions. It also provides practical safeguards, including isolated workspaces, evaluator separation, resource management, and agent session and health management. Evaluated on diverse mathematical, algorithmic, and systems optimization tasks, CORAL sets new state-of-the-art results on 10 tasks, achieving 3-10 times higher improvement rates with far fewer evaluations than fixed evolutionary search baselines across tasks. On Anthropic's kernel engineering task, four co-evolving agents improve the best known score from 1363 to 1103 cycles. Mechanistic analyses further show how these gains arise from knowledge reuse and multi-agent exploration and communication. Together, these results suggest that greater agent autonomy and multi-agent evolution can substantially improve open-ended discovery. Code is available at https://github.com/Human-Agent-Society/CORAL.
Techmeme(15)
- Sources: Meta has paused its work with Mercor while it investigates a security breach at the data vendor; OpenAI says it is investigating the security incident (Wired)
Wired : Sources: Meta has paused its work with Mercor while it investigates a security breach at the data vendor; OpenAI says it is investigating the security incident — Major AI labs are investigating a security incident that impacted Mercor, a leading data vendor.
- Utah launches a one-year pilot program allowing Legion Health's AI chatbot to renew prescriptions for 15 low-risk psychiatric maintenance medications (Robert Hart/The Verge)
Robert Hart / The Verge : Utah launches a one-year pilot program allowing Legion Health's AI chatbot to renew prescriptions for 15 low-risk psychiatric maintenance medications — Some psychiatrists are asking what problem, exactly, this is solving. … Utah is allowing an AI system to prescribe psychiatric drugs without a doctor.
- Internal memo: Iranian strikes have rendered two AWS zones "hard down" in Dubai and Bahrain and Amazon expects them to be "unavailable for an extended period" (Alex Kantrowitz/Big Technology)
Alex Kantrowitz / Big Technology : Internal memo: Iranian strikes have rendered two AWS zones “hard down” in Dubai and Bahrain and Amazon expects them to be “unavailable for an extended period” — Amazon tells its employees to deprioritize these regions as the Iran war deals meaningful damage to its infrastructure in the Gulf.
- Interviews with Codex lead Alexander Embiricos, OpenClaw's Peter Steinberger, and others about OpenAI's upcoming superapp that combines ChatGPT with Codex (Alex Heath/Sources)
Alex Heath / Sources : Interviews with Codex lead Alexander Embiricos, OpenClaw's Peter Steinberger, and others about OpenAI's upcoming superapp that combines ChatGPT with Codex — Why Codex is becoming the foundation for everything. Also: Fidji Simo's internal memo about taking a leave of absence. — ∙ Paid
- OpenAI COO Brad Lightcap is transitioning to a special projects role as part of an executive shuffle; Fidji Simo is taking medical leave for several weeks (Shirin Ghaffary/Bloomberg)
Shirin Ghaffary / Bloomberg : OpenAI COO Brad Lightcap is transitioning to a special projects role as part of an executive shuffle; Fidji Simo is taking medical leave for several weeks — OpenAI's chief operating officer is shifting into a new role and two other top executives are going on leave due to health reasons …
- Sources: Mercor asked professionals in fields like entertainment to sell their prior work materials for AI training, even if the IP could belong to ex-employers (Katherine Bindley/Wall Street Journal)
Katherine Bindley / Wall Street Journal : Sources: Mercor asked professionals in fields like entertainment to sell their prior work materials for AI training, even if the IP could belong to ex-employers — AI models from the tech giants constantly need new training data. This $10 billion startup is on the hunt for fresh resources.
- Sources: Musk requires banks seeking roles in SpaceX's IPO to subscribe to Grok and advertise on X; some banks are spending tens of millions integrating Grok (Maureen Farrell/New York Times)
Maureen Farrell / New York Times : Sources: Musk requires banks seeking roles in SpaceX's IPO to subscribe to Grok and advertise on X; some banks are spending tens of millions integrating Grok — Mr. Musk is requiring Wall Street firms to purchase subscriptions to his A.I. chatbot if they want to advise on one of the largest initial public offerings in history.
- Chinese semiconductor companies like SMIC reported record 2025 revenue, driven by AI demand and China's self-sufficiency push as a result of US restrictions (Arjun Kharpal/CNBC)
Arjun Kharpal / CNBC : Chinese semiconductor companies like SMIC reported record 2025 revenue, driven by AI demand and China's self-sufficiency push as a result of US restrictions — Chinese semiconductor firms have reported record revenue last year driven by AI demand, a shortage of memory chips and U.S. export restrictions …
- Tower Semiconductor's market cap tops $20B, four years after a near-sale to Intel for $5B; shares are up ~60% over the past month and ~525% over the past year (CTech)
CTech : Tower Semiconductor's market cap tops $20B, four years after a near-sale to Intel for $5B; shares are up ~60% over the past month and ~525% over the past year — Once overlooked, the company is now central to investor expectations around AI. — Four years ago, Tower Semiconductor was nearly sold to Intel for $5 billion.
- Sources: Huawei's Ascend 950PR chip, set for mass production soon, saw prices rise 20% after Chinese tech giants placed bulk orders to run DeepSeek's V4 model (Qianer Liu/The Information)
Qianer Liu / The Information : Sources: Huawei's Ascend 950PR chip, set for mass production soon, saw prices rise 20% after Chinese tech giants placed bulk orders to run DeepSeek's V4 model — When DeepSeek introduces its next-generation model, likely in the next few weeks, it will mark a milestone in China's yearslong quest for semiconductor self-sufficiency.
- Docs: Israeli AI chip startup Hailo is pursuing an urgent IPO via a SPAC merger at a valuation of less than $500M; it was last valued at $1.2B in 2024 (Meir Orbach/CTech)
Meir Orbach / CTech : Docs: Israeli AI chip startup Hailo is pursuing an urgent IPO via a SPAC merger at a valuation of less than $500M; it was last valued at $1.2B in 2024 — Israeli firm seeks critical funding through a SPAC merger amid mounting market pressures. — Israeli chip company Hailo …
- A US bill seeks to ban exports of DUV lithography tech to China, whose imports of chipmaking equipment reportedly grew from $10.7B in 2016 to ~$51.1B in 2025 (Jared Perlo/NBC News)
Jared Perlo / NBC News : A US bill seeks to ban exports of DUV lithography tech to China, whose imports of chipmaking equipment reportedly grew from $10.7B in 2016 to ~$51.1B in 2025 — The MATCH Act would tighten existing restrictions on a critical choke point for the AI industry, banning exports of certain manufacturing tools across China.
- Some startups and researchers who can't access the most advanced chips are adopting a "frugal AI" approach, building smaller models on open-weight systems (Rina Chandran/Rest of World)
Rina Chandran / Rest of World : Some startups and researchers who can't access the most advanced chips are adopting a “frugal AI” approach, building smaller models on open-weight systems — Amid a widening global divide in AI adoption, low-cost AI models that can deliver sovereignty and efficiency …
- The US NLRB rules that Amazon must negotiate with the Amazon Labor Union, which represents ~5,000 workers at its Staten Island warehouse; Amazon plans to appeal (Greg Bensinger/Reuters)
Greg Bensinger / Reuters : The US NLRB rules that Amazon must negotiate with the Amazon Labor Union, which represents ~5,000 workers at its Staten Island warehouse; Amazon plans to appeal — Amazon (AMZN.O) must negotiate with a labor union representing some 5,000 workers at a company warehouse on Staten Island …
- Automattic CEO Matt Mullenweg says EmDash, while open source, is designed "to sell more Cloudflare services" and lacks WordPress' cross-platform democratization (Matt Mullenweg)
Matt Mullenweg : Automattic CEO Matt Mullenweg says EmDash, while open source, is designed “to sell more Cloudflare services” and lacks WordPress' cross-platform democratization — So, two other Matts at Cloudflare announced EmDash — the spiritual successor to WordPress that solves plugin security.
Solidot(15)
- 微软更新服务条款声明 Copilot 仅供娱乐
微软被发现最近更新了 Copilot 的服务条款,包含了一则免责声明:Copilot 仅供娱乐,会犯错,会没有如预期般的工作,不要依赖 Copilot 提供重要建议,使用 Copilot 风险自负。经常使用 AI 聊天机器人的人可能早就知道它提供的信息并不可靠,需要验证。但由于它们过于方便,偷懒的人类变得不那么愿意花时间验证其输出。微软的免责声明再次强调,AI 聊天机器人既不是伴侣,也不是可靠的建议来源。它们是容易出错的工具,可能前一秒大有裨益,下一秒就可能犯错。
- 可再生能源新增装机容量占全球新增装机容量的八成以上
IRENA 最新报告显示,2025 年可再生能源新增装机容量占全球新增装机容量的 85.6%,其中太阳能新增装机容量占到了四分之三。可再生能源新增装机容量约 700 GW,太阳能就有 511 GW,太阳能总装机容量达到了 2.4 TW,比风能和水力发电高 1 TW 以上,但由于太阳能的特性,2024 年的数据显示太阳能的发电量低于风力发电:太阳能占全球总发电量的 7%,风能 8%,核能占 9%。2025 年的数据还没有公布,但根据其快速增长的装机容量太阳能的发电量可能已经超过风电成为仅次于水电的第二大无碳电力来源。
- 考古学家在北美发现距今至少 1.2 万年的骰子
赌博的历史比你想象的要更悠久。考古学家在《American Antiquity》期刊上报告发现了美洲原居民用于赌博的最古老骰子,距今至少 1.2 万年,比旧大陆上的同类活动要早六千年。从掷骰子到赛马,所有机会游戏都依赖于概率,而概率是一个相对反直觉的概念。骰子、机会游戏和赌博一直是美洲原居民文化的重要组成部分,最早的骰子出现在怀俄明州、科罗拉多州和新墨西哥州晚更新世的福尔松地层(Late Pleistocene Folsom)中。这些结果表明,古代美洲原住民掌握了关于机会、随机性和概率的基本知识,因而在对这些概念的理解和实际应用上走在了世界前列。
- 人们日常说话的单词量比上一年减少 300 个单词
发表在《Perspectives on Psychological Science》期刊上的一项研究分析了逾 2000 名参与者在 2005 年到2019 年之间的音频数据,参与者的年龄从 10-94 岁。结果显示,我们每天说话的单词量比上一年减少了约 300 个单词。这意味着一年说话的单词量比上一年减少逾 12 万个单词。说话更少意味着我们花更少的时间与他人交流,也可能意味着更孤独,而孤独与负面健康影响密切相关。研究显示,年轻一代每日说话的单词量下降更快。
- AO3 结束公测
知名同人作品网站 Archive of Our Own(AO3)宣布结束长达 17 年的公测(Open Beta)。AO3 成立于 2008 年 9 月,因保守团体施压导致同人作品被删除账号被关闭,同人社群决定创造自己的网站掌握自己的命运。AO3 于 2009 年 11 月开始公测,上线伊始它只有 347 位用户和 6,598 件同人作品,17 年后它的用户数突破了 1000 万,同人作品突破了 1700 万。运营者表示,AO3 软件已稳定运行了很长时间,结束公测只是外观上的变化,不意味着所有功能已最终完成或完美运行,也不意味着将停止对 AO3 的改进。
- 最富 0.1% 人口的离岸财富超过最穷半数人口的财富总和
根据乐施会(Oxfam)的最新报告,全球最富有 0.1% 人口的未纳税离岸财富总和超过了全球最穷困半数人口(41 亿人)的财富总和。报告呼吁国际社会采取协调一致的行动,对巨额财富征税,终止避税天堂。2024 年全球有 3.55 万亿美元的未纳税财富藏匿于避税天堂和未申报账户中,超过了法国的 GDP,是全球 44 个最不发达国家 GDP 总和的两倍多。最富有的 0.1% 人口拥有约 80% 的未纳税离岸财富或 2.84 万亿美元,最富有的 0.01% 人口拥有半数的未纳税离岸财富或 1.77万亿美元。
- 雄章鱼交接腕兼具感觉和交配功能
雄章鱼用于交配的特化腕足同时也是一种能够检测卵巢激素孕酮的感觉器官。研究人员在这条被称为“交接腕”(hectocotylus)的腕足上发现了化学感受器。在交配过程中,雄性章鱼会探查雌性章鱼的外套膜,旨在寻找用于受精的输卵管。一旦定位成功,精子就会沿着交接腕输送并得到存放。但雄性章鱼如何得知自己何时找到了输卵管?在一个实验装置中,研究人员在管子内壁涂抹了不同的物质,结果发现,只有当交接腕末端的小吸盘接触到孕酮(一种由卵巢产生的激素)时,精子才会释放。涂有其他物质的管子则会“引发回避行为”。
- Artemis II 的厕所是月球任务的一大里程碑
执行阿波罗月球任务的飞船没有专用厕所,它使用的人体排泄物收集系统饱受宇航员诟病,曾在任务期间发生过臭名昭著的粪便漏出事故,迫使宇航员在飞船内追粪便。执行阿耳忒弥斯(Artemis)月球任务的宇航员则有了专门的更舒适的厕所。被称为 Universal Waste Management System (UWMS)的厕所为如厕的宇航员提供了在微重力条件下维持稳定的把手,能同时处理尿液和粪便的系统、男女通用的尿液收集装置,甚至还有营造出私密感的门。Collins Aerospace 公司在 2015 年与 NASA 签订了合同共同研发 UWMS 系统。
- Google 发布开放权重模型 Gemma 4
Google 发布了最新的开放权重模型 Gemma 4,上个版本 Gemma 3 是在一年前发布的。Gemma 4 有四个版本,设计能在本地设备上运行:参数多的两个版本 26B Mixture of Experts 和 31B Dense 设计能在 80GB Nvidia H100 GPU(售价约 20 万人民币)上以 bfloat16 格式未量化运行,量化后降低精度则能使用消费级 GPU;参数少的两个版本 Effective 2B (E2B) 和 Effective 4B (E4B)设计能在移动设备上运行。Google 称它的 Pixel 团队与高通和联发科密切合作,为智能手机、Raspberry Pi 和 Jetson Nano 等设备对这些小模型进行了优化。Gemma 4 采用了 Apache 2.0 授权,在商业用途限制上更灵活。
- Artemis II 宇航员发现电脑上有两个 Outlook 但没有一个能用
Artemis II 宇航员离开地球之后仍然需要用微软的软件,而且和地球的用户一样,微软的软件经常出故障。美国东部时间 4 月 1 日四名宇航员开始了前往月球的 10 天之旅,整个过程一直进行直播。大约在 2 a.m. ET 左右,任务控制中心确认一个流程控制系统出现问题,提出远程协助。本次任务的指挥官、NASA 宇航员 Reid Wiseman 在通话时他说在电脑上看到了两个 Microsoft Outlook,但没有一个能用。休斯顿的任务控制中心表示会通过远程连接调查下问题。一个小时后,任务控制中心表示 Outlook 已经能使用,它会显示为离线,这是预期行为。
- 亚马逊洽谈收购 Globalstar 以挑战 Starlink
亚马逊正在洽谈收购 Globalstar 以帮助它与 SpaceX 的 Starlink 宽带卫星星座展开竞争。苹果持有 Globalstar 五分之一的股份,因此亚马逊和苹果需要展开谈判,增加了交易的复杂性。双方的磋商可能会破裂,无法达成任何协议。Globalstar 成立于 1991 年,受收购传闻的推动,周三市值达到 90 亿美元。苹果是在 2024 年向 Globalstar 投资 15 亿美元从而拥有 20% 股份。
- 实验室手套可能会释放塑料颗粒影响测量结果
研究人员发现,常用的丁腈橡胶和乳胶实验室手套会释放与微塑料相似的硬脂酸盐颗粒,可能会在研究微塑料污染时高估其含量。实验室手套会无意中将颗粒转移到用于分析空气、水等样本的实验室工具上。研究人员建议使用洁净室手套,这种手套释放的颗粒要少得多。硬脂酸盐是一种类肥皂的盐基物质,被添加到一次性手套中,以帮助其在制造过程中轻松与模具分离。由于其化学性质与部分塑料相似,在实验室分析中会难以区分,增加了研究微塑料污染时出现假阳性的风险。
- Anthropic 以版权侵犯为由要求删除上万份 Claude Code 源代码副本
Claude Code 源代码不小心泄漏之后,Anthropic 正以版权侵犯为由要求删除上万份 Claude Code 源码副本,但覆水难收,新的副本仍然源源不断的出现。开发者对该源码的分析揭示了 Anthropic 采用的一些窍门:定期回顾任务以巩固记忆,该过程被称之为“做梦(dreaming)”;某种隐藏身份的卧底模式;被称为 Buddy 的可互动电子宠物。还有开发者用其它 AI 工具和其它编程语言重写了 Claude Code,认为此举称不上版权侵犯,能免于下架的命运。
- SpaceX 申请 IPO
SpaceX 本周向 SEC 秘密递交了上市申请,标志着史上规模最大的 IPO 拉开帷幕。秘密提交上市申请允许企业在不公开披露财务信息的情况下推进上市计划。SpaceX 此次 IPO 计划融资约 750 亿美元,目标估值约 1.75 万亿美元。在美国,只有英伟达、苹果、Alphabet、微软和亚马逊的市值高于 SpaceX。SpaceX 有望迅速加入纳指,因为纳斯达克证券交易所刚刚修改了几乎是为 SpaceX 量身定做的指数纳入方法:取消了公开发行至少 10% 股份的要求——SpaceX 计划发行不到 5% 的股份;上市交易 15 天后即可加入纳斯达克 100 指数。批评人士认为,此举可能会扭曲 IPO 后的价格发现。
- 比特币的签名算法比预计的更容易破解
比特币的数字签名使用了 256 位椭圆曲线算法 secp256k1,此前估计量子计算机需要百万量子比特才能将其破解。Google 研究院的研究人员发表了一篇白皮书,称他们改进了 Shor 算法,使得在 10 分钟内破解比特币地址中的公钥成为可能。他们编译了两个解决椭圆曲线离散对数问题的量子电路,其一需要不到 1200 个逻辑量子比特和 9000 万个托佛利门,其二需要不到 1450 个逻辑量子比特和 7000 万个托佛利门。 比特币作者中本聪早在 2010 年就提出过,如果量子计算机变得切实可行,那么比特币软件需要升级改用其它算法。Google 研究人员推荐加密货币社区迁移到能抵抗量子计算机破解的后量子密码学(PQC)。