DIGEST · 2026-05-04

OrangeBot.AI Digest — 2026-05-04

85 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Microsoft Edge stores all passwords in memory in clear text, even when unused (twitter.com)
  2. Heat pump sales rise across Europe (www.pv-magazine.com)
  3. US healthcare marketplaces shared citizenship and race data with ad tech giants (techcrunch.com)
  4. Days without GitHub incidents (www.dayswithoutgithubincident.com)
  5. Stop big tech from making users behave in ways they don't want to (economist.com)
  6. I am worried about Bun (wwj.dev)
  7. Does Employment Slow Cognitive Decline? Evidence from Labor Market Shocks (www.nber.org)
  8. Incident with Issues and Webhooks – Resolved (www.githubstatus.com)
  9. Removable batteries in smartphones will be mandatory in the EU starting in 2027 (www.ecopv-eu.com)
  10. Redis array: short story of a long development process (antirez.com)
  11. How Monero’s proof of work works (blog.alcazarsec.com)
  12. PyInfra 3.8.0 (github.com)
  13. Talking to strangers at the gym (thienantran.com)
  14. Trademark violation: Fake Notepad++ for Mac (notepad-plus-plus.org)
  15. GameStop makes $55.5B takeover offer for eBay (www.bbc.co.uk)

GitHub Trending(15)

  1. ruvnet / ruflo
  2. TauricResearch / TradingAgents
  3. browserbase / skills
  4. Hmbown / DeepSeek-TUI
  5. soxoj / maigret
  6. qbittorrent / qBittorrent
  7. czlonkowski / n8n-mcp
  8. 1jehuang / jcode
  9. msitarzewski / agency-agents
  10. virattt / dexter
  11. Flowseal / zapret-discord-youtube
  12. fspecii / ace-step-ui
  13. jellyfin / jellyfin
  14. cocoindex-io / cocoindex
  15. docusealco / docuseal

Product Hunt(15)

  1. Sleek Analytics for iOS

    Your website analytics in your pocket

  2. Flowly

    Your personal AI assistant, native to your desktop

  3. Claude Code & Codex Usage Trading Cards by Rudel

    Get your trading card based on your CC & codex usage

  4. Dropy

    Track prices on stores like Amazon, eBay, & AliExpress

  5. Replyke V7

    Pre-Modeled Infra & Client SDKs for User-Powered Products.

  6. Mindra

    Agent Teams You Can Actually Delegate To

  7. Codex Pets

    Animated companions for your Codex workflow

  8. Panels Store

    Buy DRM-free comics and read them instantly in Panels

  9. Manex

    Preserve useful answers, corrections, and context as memory

  10. Regulus by Cumbuca

    AI chatbot trained on Brazil's Central Bank regulations

  11. Visitor profiles and timeline by Croct

    Uncover the story behind every click to optimize your site

  12. Aaavatar

    Branded team headshots in one drop

  13. PandaProbe

    open source agent engineering platform

  14. Huddle01 VMs

    Virtual Machines for Your Agents

  15. Rosentic

    Catch when coding agents break each other before merge

Hugging Face(15)

  1. UniVidX: A Unified Multimodal Framework for Versatile Video Generation via Diffusion Priors

    Recent progress has shown that video diffusion models (VDMs) can be repurposed for diverse multimodal graphics tasks. However, existing methods often train separate models for each problem setting, which fixes the input-output mapping and limits the modeling of correlations across modalities. We present UniVidX, a unified multimodal framework that leverages VDM priors for versatile video generation. UniVidX formulates pixel-aligned tasks as conditional generation in a shared multimodal space, adapts to modality-specific distributions while preserving the backbone's native priors, and promotes cross-modal consistency during synthesis. It is built on three key designs. Stochastic Condition Masking (SCM) randomly partitions modalities into clean conditions and noisy targets during training, enabling omni-directional conditional generation instead of fixed mappings. Decoupled Gated LoRA (DGL) introduces per-modality LoRAs that are activated when a modality serves as the generation target, preserving the strong priors of the VDM. Cross-Modal Self-Attention (CMSA) shares keys and values across modalities while keeping modality-specific queries, facilitating information exchange and inter-modal alignment. We instantiate UniVidX in two domains: UniVid-Intrinsic, for RGB videos and intrinsic maps including albedo, irradiance, and normal; and UniVid-Alpha, for blended RGB videos and their constituent RGBA layers. Experiments show that both models achieve performance competitive with state-of-the-art methods across distinct tasks and generalize robustly to in-the-wild scenarios, even when trained on fewer than 1,000 videos. Project page: https://houyuanchen111.github.io/UniVidX.github.io/

  2. Web2BigTable: A Bi-Level Multi-Agent LLM System for Internet-Scale Information Search and Extraction

    Agentic web search increasingly faces two distinct demands: deep reasoning over a single target, and structured aggregation across many entities and heterogeneous sources. Current systems struggle on both fronts. Breadth-oriented tasks demand schema-aligned outputs with wide coverage and cross-entity consistency, while depth-oriented tasks require coherent reasoning over long, branching search trajectories. We introduce Web2BigTable, a multi-agent framework for web-to-table search that supports both regimes. Web2BigTable adopts a bi-level architecture in which an upper-level orchestrator decomposes the task into sub-problems and lower-level worker agents solve them in parallel. Through a closed-loop run--verify--reflect process, the framework jointly improves decomposition and execution over time via persistent, human-readable external memory, with self-evolving updates to each single-agent. During execution, workers coordinate through a shared workspace that makes partial findings visible, allowing them to reduce redundant exploration, reconcile conflicting evidence, and adapt to emerging coverage gaps. Web2BigTable sets a new state of the art on WideSearch, reaching an Avg@4 Success Rate of 38.50 (7.5times the second best at 5.10), Row F1 of 63.53 (+25.03 over the second best), and Item F1 of 80.12 (+14.42 over the second best). It also generalises to depth-oriented search on XBench-DeepSearch, achieving 73.0 accuracy. Code is available at https://github.com/web2bigtable/web2bigtable.

  3. Map2World: Segment Map Conditioned Text to 3D World Generation

    3D world generation is essential for applications such as immersive content creation or autonomous driving simulation. Recent advances in 3D world generation have shown promising results; however, these methods are constrained by grid layouts and suffer from inconsistencies in object scale throughout the entire world. In this work, we introduce a novel framework, Map2World, that first enables 3D world generation conditioned on user-defined segment maps of arbitrary shapes and scales, ensuring global-scale consistency and flexibility across expansive environments. To further enhance the quality, we propose a detail enhancer network that generates fine details of the world. The detail enhancer enables the addition of fine-grained details without compromising overall scene coherence by incorporating global structure information. We design the entire pipeline to leverage strong priors from asset generators, achieving robust generalization across diverse domains, even under limited training data for scene generation. Extensive experiments demonstrate that our method significantly outperforms existing approaches in user-controllability, scale consistency, and content coherence, enabling users to generate 3D worlds under more complex conditions.

  4. Prox-E: Fine-Grained 3D Shape Editing via Primitive-Based Abstractions

    Text-based 2D image editing models have recently reached an impressive level of maturity, motivating a growing body of work that heavily depends on these models to drive 3D edits. While effective for appearance-based modifications, such 2D-centric 3D editing pipelines often struggle with fine-grained 3D editing, where localized structural changes must be applied while strictly preserving an object's overall identity. To address this limitation, we propose Prox-E, a training-free framework that enables fine-grained 3D control through an explicit, primitive-based geometric abstraction. Our framework first abstracts an input 3D shape into a compact set of geometric primitives. A pretrained vision-language model (VLM) then edits this abstraction to specify primitive-level changes. These structural edits are subsequently used to guide a 3D generative model, enabling fine-grained, localized modifications while preserving unchanged regions of the original shape. Through extensive experiments, we demonstrate that our method consistently balances identity preservation, shape quality, and instruction fidelity more effectively than various existing approaches, including 2D-based 3D editors and training-based methods.

  5. From Skill Text to Skill Structure: The Scheduling-Structural-Logical Representation for Agent Skills

    LLM agents increasingly rely on reusable skills, capability packages that combine instructions, control flow, constraints, and tool calls. In most current agent systems, however, skills are still represented by text-heavy artifacts, including SKILL.md-style documents and structured records whose machine-usable evidence remains embedded largely in natural-language descriptions. This poses a challenge for skill-centered agent systems: managing skill collections and using skills to support agent both require reasoning over invocation interfaces, execution structure, and concrete side effects that are often entangled in a single textual surface. An explicit representation of skill knowledge may therefore help make these artifacts easier for machines to acquire and leverage. Drawing on Memory Organization Packets, Script Theory, and Conceptual Dependency from Schank and Abelson's classical work on linguistic knowledge representation, we introduce what is, to our knowledge, the first structured representation for agent skill artifacts that disentangles skill-level scheduling signals, scene-level execution structure, and logic-level action and resource-use evidence: the Scheduling-Structural-Logical (SSL) representation. We instantiate SSL with an LLM-based normalizer and evaluate it on a corpus of skills in two tasks, Skill Discovery and Risk Assessment, and superiorly outperform the text-only baselines: in Skill Discovery, SSL improves MRR from 0.573 to 0.707; in Risk Assessment, it improves macro F1 from 0.744 to 0.787. These findings reveal that explicit, source-grounded structure makes agent skills easier to search and review. They also suggest that SSL is best understood as a practical step toward more inspectable, reusable, and operationally actionable skill representations for agent systems, rather than as a finished standard or an end-to-end mechanism for managing and using skills.

  6. Stable-GFlowNet: Toward Diverse and Robust LLM Red-Teaming via Contrastive Trajectory Balance

    Large Language Model (LLM) Red-Teaming, which proactively identifies vulnerabilities of LLMs, is an essential process for ensuring safety. Finding effective and diverse attacks in red-teaming is important, but achieving both is challenging. Generative Flow Networks (GFNs) that perform distribution matching are a promising methods, but they are notorious for training instability and mode collapse. In particular, unstable rewards in red-teaming accelerate mode collapse. We propose Stable-GFN (S-GFN), which eliminates partition function Z estimation in GFN and reduces training instability. S-GFN avoids Z-estimation through pairwise comparisons and employs a robust masking methodology against noisy rewards. Additionally, we propose a fluency stabilizer to prevent the model from getting stuck in local optima that produce gibberish. S-GFN provides more stable training while maintaining the optimal policy of GFN. We demonstrate the overwhelming attack performance and diversity of S-GFN across various settings.

  7. Learning while Deploying: Fleet-Scale Reinforcement Learning for Generalist Robot Policies

    Generalist robot policies increasingly benefit from large-scale pretraining, but offline data alone is insufficient for robust real-world deployment. Deployed robots encounter distribution shifts, long-tail failures, task variations, and human correction opportunities that fixed demonstration datasets cannot fully capture. We present Learning While Deploying (LWD), a fleet-scale offline-to-online reinforcement learning framework for continual post-training of generalist Vision-Language-Action (VLA) policies. Starting from a pretrained VLA policy, LWD closes the loop between deployment, shared physical experience, policy improvement, and redeployment by using autonomous rollouts and human interventions collected across a robot fleet. To stabilize learning from heterogeneous, sparse-reward fleet data, LWD combines Distributional Implicit Value Learning (DIVL) for robust value estimation with Q-learning via Adjoint Matching (QAM) for policy extraction in flow-based VLA action generators. We validate LWD on a fleet of 16 dual-arm robots across eight real-world manipulation tasks, including semantic grocery restocking and 3--5 minute long-horizon tasks. A single generalist policy improves as fleet experience accumulates, reaching an average success rate of 95%, with the largest gains on long-horizon tasks.

  8. Let ViT Speak: Generative Language-Image Pre-training

    In this paper, we present Generative Language-Image Pre-training (GenLIP), a minimalist generative pretraining framework for Vision Transformers (ViTs) designed for multimodal large language models (MLLMs). To better align vision encoders with the autoregressive nature of LLMs, GenLIP trains a ViT to predict language tokens directly from visual tokens using a standard language modeling objective, without contrastive batch construction or an additional text decoder. This design offers three key advantages: (1) Simplicity: a single transformer jointly models visual and textual tokens; (2) Scalability: it scales effectively with both data and model size; and (3) Performance: it achieves competitive or superior results across diverse multimodal benchmarks. Trained on 8B samples from Recap-DataComp-1B, GenLIP matches or surpasses strong baselines despite using substantially less pretraining data. After continued pretraining on multi-resolution images at native aspect ratios, GenLIP further improves on detail-sensitive tasks such as OCR and chart understanding, making it a strong foundation for vision encoders in MLLMs.

  9. When Do Diffusion Models learn to Generate Multiple Objects?

    Text-to-image diffusion models achieve impressive visual fidelity, yet they remain unreliable in multi-object generation. Despite extensive empirical evidence of these failures, the underlying causes remain unclear. We begin by asking how much of this limitation arises from the data itself. To disentangle data effects, we consider two regimes across different dataset sizes: (1) concept generalization, where each individual concept is observed during training under potentially imbalanced data distributions, and (2) compositional generalization, where specific combinations of concepts are systematically held out. To study these regimes, we introduce mosaic (Multi-Object Spatial relations, AttrIbution, Counting), a controlled framework for dataset generation. By training diffusion models on mosaic, we find that scene complexity plays a dominant role rather than concept imbalance, and that counting is uniquely difficult to learn in low-data regimes. Moreover, compositional generalization collapses as more concept combinations are held out during training. These findings highlight fundamental limitations of diffusion models and motivate stronger inductive biases and data design for robust multi-object compositional generation.

  10. Trees to Flows and Back: Unifying Decision Trees and Diffusion Models

    Decision trees and diffusion models are ostensibly disparate model classes, one discrete and hierarchical, the other continuous and dynamic. This work unifies the two by establishing a crisp mathematical correspondence between hierarchical decision trees and diffusion processes in appropriate limiting regimes. Our unification reveals a shared optimization principle: Global Trajectory Score Matching (GTSM), for which gradient boosting (in an idealized version) is asymptotically optimal. We underscore the conceptual value of our work through two key practical instantiations: \treeflow, which achieves competitive generation quality on tabular data with higher fidelity and a 2\times computational speedup, and \dsmtree, a novel distillation method that transfers hierarchical decision logic into neural networks, matching teacher performance within 2\% on many benchmarks.

  11. End-to-End Autoregressive Image Generation with 1D Semantic Tokenizer

    Autoregressive image modeling relies on visual tokenizers to compress images into compact latent representations. We design an end-to-end training pipeline that jointly optimizes reconstruction and generation, enabling direct supervision from generation results to the tokenizer. This contrasts with prior two-stage approaches that train tokenizers and generative models separately. We further investigate leveraging vision foundation models to improve 1D tokenizers for autoregressive modeling. Our autoregressive generative model achieves strong empirical results, including a state-of-the-art FID score of 1.48 without guidance on ImageNet 256x256 generation.

  12. Odysseus: Scaling VLMs to 100+ Turn Decision-Making in Games via Reinforcement Learning

    Given the rapidly growing capabilities of vision-language models (VLMs), extending them to interactive decision-making tasks such as video games has emerged as a promising frontier. However, existing approaches either rely on large-scale supervised fine-tuning (SFT) on human trajectories or apply reinforcement learning (RL) only in relatively short-horizon settings (typically around 20--30 turns). In this work, we study RL-based training of VLMs for long-horizon decision-making in Super Mario Land, a visually grounded environment requiring 100+ turns of interaction with coordinated perception, reasoning, and action. We begin with a systematic investigation of key algorithmic components and propose an adapted variant of PPO with a lightweight turn-level critic, which substantially improves training stability and sample efficiency over critic-free methods such as GRPO and Reinforce++. We further show that pretrained VLMs provide strong action priors, significantly improving sample efficiency during RL training and reducing the need for manual design choices such as action engineering, compared to classical deep RL trained from scratch. Building on these insights, we introduce Odysseus, an open training framework for VLM agents, achieving substantial gains across multiple levels of the game and at least 3 times average game progresses than frontier models. Moreover, the trained models exhibit consistent improvements under both in-game and cross-game generalization settings, while maintaining general-domain capabilities. Overall, our results identify key ingredients for making RL stable and effective in long-horizon, multi-modal settings, and provide practical guidance for developing VLMs as embodied agents.

  13. MASCing: Configurable Mixture-of-Experts Behavior via Activation Steering Masks

    Mixture-of-Experts (MoE) architectures in Large Language Models (LLMs) have significantly reduced inference costs through sparse activation. However, this sparse activation paradigm also introduces new safety challenges. Since only a subset of experts is engaged for each input, model behavior becomes coupled to routing decisions, yielding a difficult-to-control mechanism that can vary across safety-relevant scenarios. At the same time, adapting model behavior through full fine-tuning or retraining is costly, especially when developers need to rapidly configure the same model for different safety objectives. We present MASCing (MoE Activation Steering Configuration), the first framework that enables flexible reconfiguration of MoE behavior across diverse safety scenarios without retraining. MASCing uses an LSTM-based surrogate model to capture cross-layer routing dependencies and map routing logits to downstream behaviors. It then optimizes a steering matrix to identify behavior-relevant expert circuits and, at inference time, applies steering masks to the routing gates to override expert selection. This enables targeted enhancement or suppression of specific behaviors while preserving general language utility. To demonstrate its reconfigurability, we apply MASCing to two different safety-related objectives and observe consistent gains with negligible overhead across seven open-source MoE models. For multi-turn jailbreak defense, it improves the average defense success rate from 52.5% to 83.9%, with gains of up to 89.2%. For adult-content generation, MASCing enables models to comply with such requests that would otherwise be refused, increasing the average generation success rate from 52.6% to 82.0%, with gains of up to 93.0%. These results establish MASCing as a practical, lightweight, and flexible framework for scenario-specific safety reconfiguration in MoE models.

  14. Better Models, Faster Training: Sigmoid Attention for single-cell Foundation Models

    Training stable biological foundation models requires rethinking attention mechanisms: we find that using sigmoid attention as a drop in replacement for softmax attention a) produces better learned representations: on six diverse single-cell datasets, sigmoid achieves 25% higher cell-type separation, better cell-type cohesion metrics, and lower validation loss, b) faster training, models with sigmoid attention train up to 10% faster than their softmax counterparts, and c) more stable training by eliminating inherent sources of instability in softmax attention. We establish that sigmoid attention has globally bounded derivatives (leq 0.25) as opposed to softmax, and a diagonal Jacobian structure in contrast with softmax's dense coupling, which together help alleviate training instabilities. In stress tests on 160M-parameter bidirectional attention models trained without gradient clipping on 8K-token sequences, softmax diverges catastrophically, with gradients exploding by four orders of magnitude, while sigmoid remains stable. Finally, we implement and open-source TritonSigmoid, an efficient GPU kernel that achieves 515 TFLOPS on H100 GPUs, outperforming both FlashAttention-2 and FlashSigmoid, with native padding support, which is essential for biological sequences. Our results establish sigmoid attention as both theoretically grounded and empirically superior for biological foundation models. Code is available at https://github.com/MSDLLCpapers/triton-sigmoid

  15. Talker-T2AV: Joint Talking Audio-Video Generation with Autoregressive Diffusion Modeling

    Joint audio-video generation models have shown that unified generation yields stronger cross-modal coherence than cascaded approaches. However, existing models couple modalities throughout denoising via pervasive attention, treating high-level semantics and low-level details in a fully entangled manner. This is suboptimal for talking head synthesis: while audio and facial motion are semantically correlated, their low-level realizations (acoustic signals and visual textures) follow distinct rendering processes. Enforcing joint modeling across all levels causes unnecessary entanglement and reduces efficiency. We propose Talker-T2AV, an autoregressive diffusion framework where high-level cross-modal modeling occurs in a shared backbone, while low-level refinement uses modality-specific decoders. A shared autoregressive language model jointly reasons over audio and video in a unified patch-level token space. Two lightweight diffusion transformer heads decode the hidden states into frame-level audio and video latents. Experiments on talking portrait benchmarks show Talker-T2AV outperforms dual-branch baselines in lip-sync accuracy, video quality, and audio quality, achieving stronger cross-modal consistency than cascaded pipelines.

Techmeme(15)

  1. Duolingo reports Q1 revenue up 27% YoY to $292M, vs. $288.5M est., bookings up 14% to $308.5M, and expects slower growth in Q2; DUOL drops 12%+ after hours (Akash Sriram/Reuters)

    Akash Sriram / Reuters : Duolingo reports Q1 revenue up 27% YoY to $292M, vs. $288.5M est., bookings up 14% to $308.5M, and expects slower growth in Q2; DUOL drops 12%+ after hours —  Duolingo (DUOL.O) posted strong first-quarter results but signaled a more measured growth trajectory ahead, as the language-learning …

  2. Musk v. Altman: Greg Brockman testifies that his OpenAI stake is now worth ~$30B; a Musk attorney asks why he hasn't donated $29B to OpenAI's nonprofit arm (Bloomberg)

    Bloomberg : Musk v. Altman: Greg Brockman testifies that his OpenAI stake is now worth ~$30B; a Musk attorney asks why he hasn't donated $29B to OpenAI's nonprofit arm —  OpenAI co-founder and President Greg Brockman testified that his stake in the startup is now worth almost $30 billion …

  3. Palantir reports Q1 revenue up 85% YoY to $1.63B, vs. $1.54B est., US government revenue up 84% to $687M, and US commercial revenue up 133% to $595M (Jaspreet Singh/Reuters)

    Jaspreet Singh / Reuters : Palantir reports Q1 revenue up 85% YoY to $1.63B, vs. $1.54B est., US government revenue up 84% to $687M, and US commercial revenue up 133% to $595M —  Palantir Technologies (PLTR.O) beat Wall Street estimates for first-quarter revenue on Monday, driven by rising demand for its data analytics software …

  4. Elon Musk agrees to pay $1.5M to settle SEC allegations that he cheated Twitter shareholders in 2022 by failing to disclose the 5%+ stake he had in the company (Nicola M White/Bloomberg)

    Nicola M White / Bloomberg : Elon Musk agrees to pay $1.5M to settle SEC allegations that he cheated Twitter shareholders in 2022 by failing to disclose the 5%+ stake he had in the company —  Elon Musk agreed to settle Securities and Exchange Commission allegations that he cheated Twitter shareholders out of millions …

  5. Pinterest reports Q1 revenue up 18% YoY to $1B, vs. $966M est., MAUs up 11% YoY to 631M, and forecasts Q2 revenue above estimates; PINS jumps 17%+ after hours (Jonathan Vanian/CNBC)

    Jonathan Vanian / CNBC : Pinterest reports Q1 revenue up 18% YoY to $1B, vs. $966M est., MAUs up 11% YoY to 631M, and forecasts Q2 revenue above estimates; PINS jumps 17%+ after hours —  Pinterest reported first-quarter earnings on Monday that beat on the top and bottom lines.  Shares soared 15% after the report.

  6. Musk v. Altman: Stuart Russell, Musk's only AI expert witness, warned of AI risks but his concerns about AI's existential threats were excluded by the judge (Tim Fernholz/TechCrunch)

    Tim Fernholz / TechCrunch : Musk v. Altman: Stuart Russell, Musk's only AI expert witness, warned of AI risks but his concerns about AI's existential threats were excluded by the judge —  When do we take AI doomers seriously?  —  That's a key subtext of Elon Musk's attempt to shut down OpenAI's for-profit AI business.

  7. Sources: the Trump administration is discussing an EO to create an AI working group to examine AI oversight procedures, including vetting models before release (New York Times)

    New York Times : Sources: the Trump administration is discussing an EO to create an AI working group to examine AI oversight procedures, including vetting models before release —  The Trump administration, which took a noninterventionist approach to artificial intelligence, is now discussing imposing oversight …

  8. Former Trump and Biden AI advisers Dean Ball and Ben Buchanan urge bipartisan action on AI security risks, including tighter export controls and safety audits (New York Times)

    New York Times : Former Trump and Biden AI advisers Dean Ball and Ben Buchanan urge bipartisan action on AI security risks, including tighter export controls and safety audits —  We come from different parties and have guided artificial intelligence policy under very different presidents.

  9. Anthropic co-founder explains why there's a 60%+ chance of AI systems autonomously building their successors by 2029 and the consequences of automated AI R&D (Jack Clark/Import AI)

    Jack Clark / Import AI : Anthropic co-founder explains why there's a 60%+ chance of AI systems autonomously building their successors by 2029 and the consequences of automated AI R&D —  The first step towards recursive self improvement  —  Welcome to Import AI, a newsletter about AI research.

  10. Sources: Apple prepares a "Create a Pass" feature for iOS 27, which lets users take a QR code and generate a custom pass around it for concerts and other venues (Mark Gurman/Bloomberg)

    Mark Gurman / Bloomberg : Sources: Apple prepares a “Create a Pass” feature for iOS 27, which lets users take a QR code and generate a custom pass around it for concerts and other venues —  Apple Inc. is preparing a new “Create a Pass” feature for its next major iPhone software update …

  11. Intel taps Alex Katouzian, an ex-Qualcomm EVP, to lead Client Computing & Physical AI group, and names Pushkar Ranade as CTO, after serving on an interim basis (Dylan Martin/CRN)

    Dylan Martin / CRN : Intel taps Alex Katouzian, an ex-Qualcomm EVP, to lead Client Computing & Physical AI group, and names Pushkar Ranade as CTO, after serving on an interim basis —  With the hiring of Alex Katouzian, who was responsible for mounting a new challenge against Intel's PC dominance …

  12. Ex-iRobot CEO Colin Angle launches Familiar Machines & Magic and unveils Familiar, a dog-like, "emotionally intelligent" robot that reacts to owner's feelings (Christopher Mims/Wall Street Journal)

    Christopher Mims / Wall Street Journal : Ex-iRobot CEO Colin Angle launches Familiar Machines & Magic and unveils Familiar, a dog-like, “emotionally intelligent” robot that reacts to owner's feelings —  The inventor behind the world-famous robot vacuum is now designing robots that form an emotional bond with their owners

  13. Nearly 20 US state-run health insurance exchanges include ad trackers that send user data like race and citizenship info to companies like Meta, TikTok, Google (Bloomberg)

    Bloomberg : Nearly 20 US state-run health insurance exchanges include ad trackers that send user data like race and citizenship info to companies like Meta, TikTok, Google —  Nearly all of the 20 state-run health insurance exchanges in the US have added advertising trackers that transmit user activity …

  14. Sources: more than two dozen prediction-market ETFs have been pushed back as the SEC seeks more information; they were originally expected to launch this week (Suzanne McGee/Reuters)

    Suzanne McGee / Reuters : Sources: more than two dozen prediction-market ETFs have been pushed back as the SEC seeks more information; they were originally expected to launch this week —  More than two dozen exchange-traded funds tied to elections, recessions, tech layoffs and other real-world events …

  15. Filing: Blackstone's data center acquisition vehicle seeks to raise as much as $1.75B in its IPO, and will target newly built data centers valued at $250M-$1.5B (Subrat Patnaik/Bloomberg)

    Subrat Patnaik / Bloomberg : Filing: Blackstone's data center acquisition vehicle seeks to raise as much as $1.75B in its IPO, and will target newly built data centers valued at $250M-$1.5B —  Blackstone Digital Infrastructure Trust Inc. is seeking to raise as much as $1.75 billion in a US initial public offering …

Solidot(10)

  1. 科学家发现咖啡如何影响肠道和大脑

    根据发表在《Nature Communications》期刊上的一项研究,科学家发现常饮用含咖啡因和不含咖啡因的咖啡会影响肠道菌群,从而影响情绪和压力水平。研究人员对比了 31 名常饮用咖啡者和 31 名不喝咖啡者。常饮用咖啡者指的是每天饮用 3-5 杯咖啡的人。实验开始时,咖啡饮用者停止饮用咖啡两周。在此期间,研究人员持续收集生物样本监测心理健康状况。实验期间参与者并不知道自己饮用的是含咖啡因的咖啡还是不含咖啡因的咖啡。一半参与者饮用不含咖啡因的咖啡,另一半饮用普通咖啡。参与者都报告情绪有所改善,这一结果显示即使不含咖啡因咖啡也能改善情绪。研究还发现常饮用咖啡者有更高的埃格特菌属(Eggertella sp.)和短隐杆菌(Cryptobacterium curtum),更多的厚壁菌门(Firmicutes)。只有摄入不含咖啡因的人才表现出学习和记忆力的提升,而只有摄入咖啡因的参与者才体验到焦虑减轻以及注意力和警觉性提高。

  2. 天文学家发现 27 颗围绕双恒星运行的候选行星

    天文学家发现了 27 颗围绕双恒星运行的候选行星,类似星球大战里的沙漠行星塔图因(Tatooine)。天文学家至今发现了 18 颗环双星行星,但类似环绕太阳运行的的单恒星行星则发现了逾 8000 颗。科学家以前通过凌日现象识别环双星行星,但需要在特定条件下才能观测到。现在他们采用了轨道进动(apsidal precession),寻找相互绕行且发生掩食的双星系统中轨道出现的摆动,这种摆动通常只能用存在第三个天体去解释。研究团队利用 NASA Transiting Exoplanet Survey Satellite 卫星收集的数据,从 1590 个恒星系统中识别出 36 个候选天体,其中 27 个天体可能具有行星质量。研究人员表示需要更多研究才能确定它们是否是环双星行星。

  3. VS Code 默认在 commit 中插入 Co-Authored-by Copilot

    微软的编辑器 VS Code 被发现默认在 commit 中插入了 Co-Authored-by Copilot,不管用户有没有使用其 AI 助手 Copilot。此事再次在用户中引发了大量批评。微软开发者回应称他们将会在下个版本中解决默认启用的问题,称如果用户没有使用 AI 助手那么就不应该说代码是 Copilot 合作编写的。

  4. 中国三月绿色技术出口增长七成

    因霍尔木兹海峡封锁引发的新一轮能源危机,世界各国正加速向清洁能源转型,最大的绿色技术出口国中国三月的太阳能、电池和电动汽车的出口总额同比增长 70%,其中出口的太阳能装机容量达到 68GW,电池出口额达到 100 亿美元,电动汽车和混合动力汽车出口同比增长 140%。多达 50 个国家从中国进口的太阳能设备都创历史新高。

  5. Steam 用户中使用 Linux 比例占 4.52%

    2026 年 3 月 Steam 玩家中使用 Linux 比例达到了史无前例的 5.33%,比前一个月增加了一倍多。根据 Valve 公布的 2026 年 4 月 Steam 硬件和软件调查,Steam 用户中使用 Linux 比例回落到了 4.52%,减少 0.81%,但仍然比去年同期翻了一番。Windows 操作系统的比例提高到 93.47%,OSX 占 2.01%。有众多证据表明 Linux 上的游戏表现有了翻天覆地的变化,而 Linux 下游戏的一大特性是需要的资源比 Windows 更少,在今天内存价格飙升的时期显得更有吸引力。其它数据显示:简体中文用户比例占 23.41%,英语用户占 36.77% 。用户使用英特尔 CPU 的比例占 55.81%,AMD 占 44.18%,几乎和前一个月相同。

  6. 英国 NHS 以 AI 为由准备关闭所有开源库

    日程安排平台 Cal.com 上月宣布从开源转为闭源,理由是 AI 工具更容易从开源代码中发现漏洞,而安全性依赖于模糊,因此闭源有助于提高安全。现在英国国家医疗服务体系(NHS)以相同的理由准备关闭它几乎所有的开源库,这一决定引发了广泛争议和批评。批评者指出 NHS 公布的大部分开源库是数据集、内部工具、指南、研究工具、前端设计等,它们不会因为安全扫描技术的进步而受到影响。此外是否开源对于 Anthropic Mythos 之类的 AI 工具并无区别,因为它们也能分析二进制程序并寻找漏洞。批评者发表了公开信,呼吁 NHS 保持其代码公开。

  7. 杭州法院裁定以 AI 代替人类为由裁员是系违法

    杭州市中级人民法院公布了一起有关“AI 接替人类员工”的判例,判决公司因“AI 成本比人工低”而辞退员工系违法行为,涉事企业需要支付赔偿金 26 万元人民币。在本案中,现年 35 岁的小周(化名) 2022 年入职杭州某家科技公司担任 AI 大模型“质检员”,负责对 AI 大模型与用户交互形成的答案进行正确性判定。2025 年,该公司以“AI 大模型技术升级,原来需要人工完成的质检工作,现在 AI 自己就能做了”为由,试图对小周进行调岗降薪:从主管降为普通员工、月薪从 2.5 万元人民币降到 1.5 万元。小周拒绝如此安排,随后就被公司解除劳动合同。小周申请劳动仲裁,仲裁庭判定公司应当支付违法解除劳动合同赔偿金 26 万余元。该公司不服,因此诉诸法庭。杭州市中级人民法院审理后认定,该公司解约非因裁撤业务、经营不善、减少亏损等消极因素,而是以 AI 的成本优势为由,不属于劳动合同无法履行的“客观情况重大变化”。而且该公司之前为小周提供的调岗降薪方案,实际上导致待遇大幅下降,并非合理协商方案。因此法庭认定该公司构成违法解除,支持仲裁结果,判决其按 2N 标准支付小周赔偿金。杭州市中级人民法院民事第五庭庭长丁晔对媒体表示,在企业视角下,应用 AI 提效降本是市场竞争的必然选择;而在劳动者视角下,因技术变革而失去岗位或被降薪,实质是公司将正常的技术迭代风险转嫁给劳动者。

  8. 人可以在睡梦中交流和学习

    很多人都有过在睡梦中获得灵感的经历。这种现象促使科学家研究睡眠学习。在 1954 年 Charles W. Simon 和 William H. Emmons 认为大多数睡眠学习研究的参与者其实都是清醒的,因此此类研究都毫无意义。他们将睡眠研究归类为科幻和伪科学,之后几十年很少有人再对此展开研究。但最近几年,科学家再次尝试展开研究。新研究主要针对清醒梦者,即在睡眠中保持意识清醒,并意识到自己正在做梦的人。根据发表在《Neuroscience of Consciousness》期刊上的一项研究,20 名清醒梦者在实验室里尝试睡梦中解谜。每个谜题都与特定的声音配对,旨在促使他们恢复处理相应的谜题。在实验室里,参与者解开了梦中出现的谜题的 42%,对于没有出现在梦中的谜题,他们只解开了 17%。大多数人不会做清醒梦,所以研究对象并不具有代表性。研究人员认为一种解释是:我们在睡着时,更可能将不相关的刺激联系起来。研究人员并不建议为了睡眠中学习而干扰睡眠,因为睡眠是重要的生理过程,干扰这一过程可能得不偿失。

  9. Ask.com 关闭

    有 30 年历史的搜索引擎 Ask.com 于 2026 年 5 月 1 日关闭。Ask.com 创办于 1996 年 6 月,最初的名字叫 AskJeeves.com,2006 年弃用了名字 Jeeves 变成了 Ask.com,成为了搜索引擎,有自己的爬虫和算法。2010 年面对大型搜索引擎的竞争它将网络搜索技术外包,恢复了问答网站的功能。Ask.com 虽然关闭了,但 AskJeeves.com 仍继续运营。Jeeves 的意思是贴身侍从,名字来自于英国作家 P. G. Wodehouse 作品《Jeeves》系列,Jeeves 是绅士 Bertie Wooster 的贴身男仆。

  10. 为什么 OpenAI 的系统提示词要专门限制 Goblins

    OpenAI Codex CLI 系统提示词专门加入了一条对地精(Goblins)等词的限制:“never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user's query”。官方解释称,从 GPT-5.1 开始该公司的模型在比喻中提及 goblin 等词的频率大增,ChatGPT 中 goblin 的使用量增加了 175%,gremlin 使用量增加了 52%。它为此展开了调查,发现是因为 Nerdy 个性无意中奖励了此类比喻,导致高频使用 goblin 的行为扩散。为解决该问题,OpenAI 淘汰了 Nerdy 个性,移除了对 goblin 友好的奖励信号,从训练数据过滤掉相关示例,防止其再次不恰当的出现。