DIGEST · 2026-05-01

OrangeBot.AI Digest — 2026-05-01

84 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Credit cards are vulnerable to brute force attacks (metin.nextc.org)
  2. Ti-84 Evo (education.ti.com)
  3. New research suggests people can communicate and practice skills while dreaming (www.newyorker.com)
  4. City Learns Flock Accessed Cameras in Children's Gymnastics Room as a Sales Demo (www.404media.co)
  5. Spotify adds 'Verified' badges to distinguish human artists from AI (www.bbc.com)
  6. The Gay Jailbreak Technique (github.com)
  7. AI uses less water than the public thinks (californiawaterblog.com)
  8. Flock cameras keep telling police a man who doesn't have a warrant has a warrant (www.youtube.com)
  9. Uber torches 2026 AI budget on Claude Code in four months (www.briefs.co)
  10. Police Have Used License Plate Readers at Least 14x to Stalk Romantic Interests (ij.org)
  11. Ask HN: Who is hiring? (May 2026)
  12. An open letter asking NHS England to keep its code open (keepthingsopen.com)
  13. Your website is not for you (websmith.studio)
  14. Apple accidentally left Claude.md files Apple Support app (x.com)
  15. Show HN: WhatCable, a tiny menu bar app for inspecting USB-C cables (github.com)

GitHub Trending(9)

  1. TauricResearch / TradingAgents
  2. soxoj / maigret
  3. warpdotdev / warp
  4. 1jehuang / jcode
  5. mattpocock / skills
  6. browserbase / skills
  7. simstudioai / sim
  8. obra / superpowers
  9. Flowseal / zapret-discord-youtube

Product Hunt(15)

  1. Postiz

    Agentic social media scheduler for agents like OpenClaw

  2. Bitgrain

    Design studio lighter than Figma & more flexible than Canva

  3. Buda

    Recruit agents to run your company as a synchronous team

  4. Zed 1.0

    High-performance, open source, multiplayer code editor

  5. Zush

    Updated: docs support, BYOK, Local AI (Ollama), Windows App

  6. Marx Finance

    AI agents debate the markets

  7. CipherLock

    Learn ciphers by breaking them

  8. HiveTerm

    One workspace for Claude, Codex, Gemini and your stack

  9. Montage

    The runtime framework for agentic user interfaces!

  10. ScreenVeil

    Hide what shouldn’t be seen on your computer

  11. nudge

    Drop your tasks. AI auto-schedules your whole week.

  12. TrafficClaw

    Have a conversation with your SEO & analytics data

  13. Ghosted: Smart Presence

    Pause media or lock your screen when you step away

  14. LaunchCut

    Interactive iOS Demo Builder

  15. PeekFocus

    One keystroke blurs everything behind your active window

Hugging Face(15)

  1. Heterogeneous Scientific Foundation Model Collaboration

    Agentic large language model systems have demonstrated strong capabilities. However, their reliance on language as the universal interface fundamentally limits their applicability to many real-world problems, especially in scientific domains where domain-specific foundation models have been developed to address specialized tasks beyond natural language. In this work, we introduce Eywa, a heterogeneous agentic framework designed to extend language-centric systems to a broader class of scientific foundation models. The key idea of Eywa is to augment domain-specific foundation models with a language-model-based reasoning interface, enabling language models to guide inference over non-linguistic data modalities. This design allows predictive foundation models, which are typically optimized for specialized data and tasks, to participate in higher-level reasoning and decision-making processes within agentic systems. Eywa can serve as a drop-in replacement for a single-agent pipeline (EywaAgent) or be integrated into existing multi-agent systems by replacing traditional agents with specialized agents (EywaMAS). We further investigate a planning-based orchestration framework in which a planner dynamically coordinates traditional agents and Eywa agents to solve complex tasks across heterogeneous data modalities (EywaOrchestra). We evaluate Eywa across a diverse set of scientific domains spanning physical, life, and social sciences. Experimental results demonstrate that Eywa improves performance on tasks involving structured and domain-specific data, while reducing reliance on language-based reasoning through effective collaboration with specialized foundation models.

  2. Visual Generation in the New Era: An Evolution from Atomic Mapping to Agentic World Modeling

    Recent visual generation models have made major progress in photorealism, typography, instruction following, and interactive editing, yet they still struggle with spatial reasoning, persistent state, long-horizon consistency, and causal understanding. We argue that the field should move beyond appearance synthesis toward intelligent visual generation: plausible visuals grounded in structure, dynamics, domain knowledge, and causal relations. To frame this shift, we introduce a five-level taxonomy: Atomic Generation, Conditional Generation, In-Context Generation, Agentic Generation, and World-Modeling Generation, progressing from passive renderers to interactive, agentic, world-aware generators. We analyze key technical drivers, including flow matching, unified understanding-and-generation models, improved visual representations, post-training, reward modeling, data curation, synthetic data distillation, and sampling acceleration. We further show that current evaluations often overestimate progress by emphasizing perceptual quality while missing structural, temporal, and causal failures. By combining benchmark review, in-the-wild stress tests, and expert-constrained case studies, this roadmap offers a capability-centered lens for understanding, evaluating, and advancing the next generation of intelligent visual generation systems.

  3. Co-Evolving Policy Distillation

    RLVR and OPD have become standard paradigms for post-training. We provide a unified analysis of these two paradigms in consolidating multiple expert capabilities into a single model, identifying capability loss in different ways: mixed RLVR suffers from inter-capability divergence cost, while the pipeline of first training experts and then performing OPD, though avoiding divergence, fails to fully absorb teacher capabilities due to large behavioral pattern gaps between teacher and student. We propose Co-Evolving Policy Distillation (CoPD), which encourages parallel training of experts and introduces OPD during each expert's ongoing RLVR training rather than after complete expert training, with experts serving as mutual teachers (making OPD bidirectional) to co-evolve. This enables more consistent behavioral patterns among experts while maintaining sufficient complementary knowledge throughout. Experiments validate that CoPD achieves all-in-one integration of text, image, and video reasoning capabilities, significantly outperforming strong baselines such as mixed RLVR and MOPD, and even surpassing domain-specific experts. The model parallel training pattern offered by CoPD may inspire a novel training scaling paradigm.

  4. ExoActor: Exocentric Video Generation as Generalizable Interactive Humanoid Control

    Humanoid control systems have made significant progress in recent years, yet modeling fluent interaction-rich behavior between a robot, its surrounding environment, and task-relevant objects remains a fundamental challenge. This difficulty arises from the need to jointly capture spatial context, temporal dynamics, robot actions, and task intent at scale, which is a poor match to conventional supervision. We propose ExoActor, a novel framework that leverages the generalization capabilities of large-scale video generation models to address this problem. The key insight in ExoActor is to use third-person video generation as a unified interface for modeling interaction dynamics. Given a task instruction and scene context, ExoActor synthesizes plausible execution processes that implicitly encode coordinated interactions between robot, environment, and objects. Such video output is then transformed into executable humanoid behaviors through a pipeline that estimates human motion and executes it via a general motion controller, yielding a task-conditioned behavior sequence. To validate the proposed framework, we implement it as an end-to-end system and demonstrate its generalization to new scenarios without additional real-world data collection. Furthermore, we conclude by discussing limitations of the current implementation and outlining promising directions for future research, illustrating how ExoActor provides a scalable approach to modeling interaction-rich humanoid behaviors, potentially opening a new avenue for generative models to advance general-purpose humanoid intelligence.

  5. Efficient Training on Multiple Consumer GPUs with RoundPipe

    Fine-tuning Large Language Models (LLMs) on consumer-grade GPUs is highly cost-effective, yet constrained by limited GPU memory and slow PCIe interconnects. Pipeline parallelism combined with CPU offloading mitigates these hardware bottlenecks by reducing communication overhead. However, existing PP schedules suffer from an inherent limitation termed the weight binding issue. Binding uneven model stages (e.g., the LM head is large) to GPUs limits the pipeline's throughput to that of the GPU with the heaviest load, leading to severe pipeline bubbles. In this paper, we propose RoundPipe, a novel pipeline schedule that breaks the weight binding constraint on consumer GPU servers. RoundPipe treats GPUs as a pool of stateless execution workers and dynamically dispatches computation stages across devices in a round-robin manner, achieving a near-zero-bubble pipeline. To ensure training correctness and system efficiency, RoundPipe integrates a priority-aware transfer scheduling engine, a fine-grained distributed event-based synchronization protocol, and an automated layer partitioning algorithm. Evaluations on an 8times RTX 4090 server demonstrate that RoundPipe achieves 1.48--2.16times speedups over state-of-the-art baselines when fine-tuning 1.7B to 32B models. Remarkably, RoundPipe enables LoRA fine-tuning of the Qwen3-235B model with 31K sequence length on a single server. RoundPipe is publicly available as an open-source Python library with comprehensive documentation.

  6. Claw-Eval-Live: A Live Agent Benchmark for Evolving Real-World Workflows

    LLM agents are expected to complete end-to-end units of work across software tools, business services, and local workspaces. Yet many agent benchmarks freeze a curated task set at release time and grade mainly the final response, making it difficult to evaluate agents against evolving workflow demand or verify whether a task was executed. We introduce Claw-Eval-Live, a live benchmark for workflow agents that separates a refreshable signal layer, updated across releases from public workflow-demand signals, from a reproducible, time-stamped release snapshot. Each release is constructed from public workflow-demand signals, with ClawHub Top-500 skills used in the current release, and materialized as controlled tasks with fixed fixtures, services, workspaces, and graders. For grading, Claw-Eval-Live records execution traces, audit logs, service state, and post-run workspace artifacts, using deterministic checks when evidence is sufficient and structured LLM judging only for semantic dimensions. The release contains 105 tasks spanning controlled business services and local workspace repair, and evaluates 13 frontier models under a shared public pass rule. Experiments reveal that reliable workflow automation remains far from solved: the leading model passes only 66.7% of tasks and no model reaches 70%. Failures are structured by task family and execution surface, with HR, management, and multi-system business workflows as persistent bottlenecks and local workspace repair comparatively easier but unsaturated. Leaderboard rank alone is insufficient because models with similar pass rates can diverge in overall completion, and task-level discrimination concentrates in a middle band of tasks. Claw-Eval-Live suggests that workflow-agent evaluation should be grounded twice, in fresh external demand and in verifiable agent action.

  7. Length Value Model: Scalable Value Pretraining for Token-Level Length Modeling

    Token serves as the fundamental unit of computation in modern autoregressive models, and generation length directly influences both inference cost and reasoning performance. Despite its importance, existing approaches lack fine-grained length modeling, operating primarily at the coarse-grained sequence level. We introduce the Length Value Model (LenVM), a token-level framework that models the remaining generation length. By formulating length modeling as a value estimation problem and assigning a constant negative reward to each generated token, LenVM predicts a bounded, discounted return that serves as a monotone proxy for the remaining generation horizon. This formulation yields supervision that is annotation-free, dense, unbiased, and scalable. Experiments on LLMs and VLMs demonstrate LenVM provides a highly effective signal at inference time. On the LIFEBench exact length matching task, applying LenVM to a 7B model improves the length score from 30.9 to 64.8, significantly outperforming frontier closed-source models. Furthermore, LenVM enables continuous control over the trade off between performance and efficiency. On GSM8K at a budget of 200 tokens, LenVM maintains 63% accuracy compared to 6 percent for token budget baseline. It also accurately predicts total generation length from the prompt boundary. Finally, LenVM's token-level values offer an interpretable view of generation dynamics, revealing how specific tokens shift reasoning toward shorter or longer regimes. Results demonstrate that LenVM supports a broad range of applications and token length can be effectively modeled as a token-level value signal, highlighting the potential of LenVM as a general framework for length modeling and as a length-specific value signal that could support future RL training. Code is available at https://github.com/eric-ai-lab/Length-Value-Model.

  8. Leveraging Verifier-Based Reinforcement Learning in Image Editing

    While Reinforcement Learning from Human Feedback (RLHF) has become a pivotal paradigm for text-to-image generation, its application to image editing remains largely unexplored. A key bottleneck is the lack of a robust general reward model for all editing tasks. Existing edit reward models usually give overall scores without detailed checks, ignoring different instruction requirements and causing biased rewards. To address this, we argue that the key is to move from a simple scorer to a reasoning verifier. We introduce Edit-R1, a framework that builds a chain-of-thought (CoT) verifier-based reasoning reward model (RRM) and then leverages it for downstream image editing. The Edit-RRM breaks instructions into distinct principles, evaluates the edited image against each principle, and aggregates these checks into an interpretable, fine-grained reward. To build such an RRM, we first apply supervised fine-tuning (SFT) as a ``cold-start'' to generate CoT reward trajectories. Then, we introduce Group Contrastive Preference Optimization (GCPO), a reinforcement learning algorithm that leverages human pairwise preference data to reinforce our pointwise RRM. After building the RRM, we use GRPO to train editing models with this non-differentiable yet powerful reward model. Extensive experiments demonstrate that our Edit-RRM surpasses powerful VLMs such as Seed-1.5-VL and Seed-1.6-VL as an editing-specific reward model, and we observe a clear scaling trend, with performance consistently improving from 3B to 7B parameters. Moreover, Edit-R1 delivers gains to editing models like FLUX.1-kontext, highlighting its effectiveness in enhancing image editing.

  9. Intern-Atlas: A Methodological Evolution Graph as Research Infrastructure for AI Scientists

    Existing research infrastructure is fundamentally document-centric, providing citation links between papers but lacking explicit representations of methodological evolution. In particular, it does not capture the structured relationships that explain how and why research methods emerge, adapt, and build upon one another. With the rise of AI-driven research agents as a new class of consumers of scientific knowledge, this limitation becomes increasingly consequential, as such agents cannot reliably reconstruct method evolution topologies from unstructured text. We introduce Intern-Atlas, a methodological evolution graph that automatically identifies method-level entities, infers lineage relationships among methodologies, and captures the bottlenecks that drive transitions between successive innovations. Built from 1,030,314 papers spanning AI conferences, journals, and arXiv preprints, the resulting graph comprises 9,410,201 semantically typed edges, each grounded in verbatim source evidence, forming a queryable causal network of methodological development. To operationalize this structure, we further propose a self-guided temporal tree search algorithm for constructing evolution chains that trace the progression of methods over time. We evaluate the quality of the resulting graph against expert-curated ground-truth evolution chains and observe strong alignment. In addition, we demonstrate that Intern-Atlas enables downstream applications in idea evaluation and automated idea generation. We position methodological evolution graphs as a foundational data layer for the emerging automated scientific discovery.

  10. Nemotron 3 Nano Omni: Efficient and Open Multimodal Intelligence

    We introduce Nemotron 3 Nano Omni, the latest model in the Nemotron multimodal series and the first to natively support audio inputs alongside text, images, and video. Nemotron 3 Nano Omni delivers consistent accuracy improvements over its predecessor, Nemotron Nano V2 VL, across all modalities, enabled by advances in architecture, training data and recipes. In particular, Nemotron 3 delivers leading results in real-world document understanding, long audio-video comprehension, and agentic computer use. Built on the highly efficient Nemotron 3 Nano 30B-A3B backbone, Nemotron 3 Nano Omni further incorporates innovative multimodal token-reduction techniques to deliver substantially lower inference latency and higher throughput than other models of similar size. We are releasing model checkpoints in BF16, FP8, and FP4 formats, along with portions of the training data and codebase to facilitate further research and development.

  11. InteractWeb-Bench: Can Multimodal Agent Escape Blind Execution in Interactive Website Generation?

    With the advancement of multimodal large language models (MLLMs) and coding agents, the website development has shifted from manual programming to agent-based project-level code synthesis. Existing benchmarks rely on idealized assumptions, especially for well-structured, information-rich inputs and static execution settings. In contrast, real-world development is constrained by a critical bottleneck: the semantic misalignment between ambiguous, low-quality instructions from non-expert users and model understanding, which results in a failure mode that we term blind execution. To address this gap, we introduce InteractWeb-Bench, the first multimodal interactive benchmark for website generation under non-expert low-code user conditions. InteractWeb-Bench introduces four types of user agents and persona-driven instruction perturbations to systematically simulate diverse user behaviors, including ambiguity, redundancy, and contradiction, grounded in requirement engineering defect taxonomies. We develop an interactive execution environment for agents, featuring a unified action space comprising Clarify, Implement, Verify, and Submit, enabling iterative intent refinement, code synthesis, and visual feedback-based validation. Extensive experiments and analysis reveal that frontier MLLM-based agents remain trapped in blind execution, exposing limitations in intent recognition and adaptive interaction.

  12. Synthetic Computers at Scale for Long-Horizon Productivity Simulation

    Realistic long-horizon productivity work is strongly conditioned on user-specific computer environments, where much of the work context is stored and organized through directory structures and content-rich artifacts. To scale synthetic data creation for such productivity scenarios, we introduce Synthetic Computers at Scale, a scalable methodology for creating such environments with realistic folder hierarchies and content-rich artifacts (e.g., documents, spreadsheets, and presentations). Conditioned on each synthetic computer, we run long-horizon simulations: one agent creates productivity objectives that are specific to the computer's user and require multiple professional deliverables and about a month of human work; another agent then acts as that user and keeps working across the computer -- for example, navigating the filesystem for grounding, coordinating with simulated collaborators, and producing professional artifacts -- until these objectives are completed. In preliminary experiments, we create 1,000 synthetic computers and run long-horizon simulations on them; each run requires over 8 hours of agent runtime and spans more than 2,000 turns on average. These simulations produce rich experiential learning signals, whose effectiveness is validated by significant improvements in agent performance on both in-domain and out-of-domain productivity evaluations. Given that personas are abundant at billion scale, this methodology can in principle scale to millions or even billions of synthetic user worlds with sufficient compute, enabling broader coverage of diverse professions, roles, contexts, environments, and productivity needs. We argue that scalable synthetic computer creation, together with at-scale simulations, is highly promising as a foundational substrate for agent self-improvement and agentic reinforcement learning in long-horizon productivity scenarios.

  13. Representation Fréchet Loss for Visual Generation

    We show that Fréchet Distance (FD), long considered impractical as a training objective, can in fact be effectively optimized in the representation space. Our idea is simple: decouple the population size for FD estimation (e.g., 50k) from the batch size for gradient computation (e.g., 1024). We term this approach FD-loss. Optimizing FD-loss reveals several surprising findings. First, post-training a base generator with FD-loss in different representation spaces consistently improves visual quality. Under the Inception feature space, a one-step generator achieves0.72 FID on ImageNet 256x256. Second, the same FD-loss repurposes multi-step generators into strong one-step generators without teacher distillation, adversarial training or per-sample targets. Third, FID can misrank visual quality: modern representations can yield better samples despite worse Inception FID. This motivates FDr^k, a multi-representation metric. We hope this work will encourage further exploration of distributional distances in diverse representation spaces as both training objectives and evaluation metrics for generative models.

  14. Compliance versus Sensibility: On the Reasoning Controllability in Large Language Models

    Large Language Models (LLMs) are known to acquire reasoning capabilities through shared inference patterns in pre-training data, which are further elicited via Chain-of-Thought (CoT) practices. However, whether fundamental reasoning patterns, such as induction, deduction, and abduction, can be decoupled from specific problem instances remains a critical challenge for model controllability, and for shedding light on reasoning controllability. In this paper, we present the first systematic investigation of this problem through the lens of reasoning conflicts: an explicit tension between parametric and contextual information induced by mandating logical schemata that deviate from those expected for a target task. Our evaluation reveals that LLMs consistently prioritize sensibility over compliance, favoring task-appropriate reasoning patterns despite conflicting instructions. Notably, task accuracy is not strictly determined by sensibility, with models often maintaining high performance even when using conflicting patterns, suggesting a reliance on internalized parametric memory that increases with model size. We further demonstrate that reasoning conflicts are internally detectable, as confidence scores significantly drop during conflicting episodes. Probing experiments confirm that reasoning types are linearly encoded from middle-to-late layers, indicating the potential for activation-level controllability. Leveraging these insights, we steer models towards compliance, increasing instruction following by up to 29%. Overall, our findings establish that while LLM reasoning is anchored to concrete instances, active mechanistic interventions can effectively decouple logical schemata from data, offering a path toward improved controllability, faithfulness, and generalizability.

  15. The Last Human-Written Paper: Agent-Native Research Artifacts

    Scientific publication compresses a branching, iterative research process into a linear narrative, discarding the majority of what was discovered along the way. This compilation imposes two structural costs: a Storytelling Tax, where failed experiments, rejected hypotheses, and the branching exploration process are discarded to fit a linear narrative; and an Engineering Tax, where the gap between reviewer-sufficient prose and agent-sufficient specification leaves critical implementation details unwritten. Tolerable for human readers, these costs become critical when AI agents must understand, reproduce, and extend published work. We introduce the Agent-Native Research Artifact (ARA), a protocol that replaces the narrative paper with a machine-executable research package structured around four layers: scientific logic, executable code with full specifications, an exploration graph that preserves the failures compilation discards, and evidence grounding every claim in raw outputs. Three mechanisms support the ecosystem: a Live Research Manager that captures decisions and dead ends during ordinary development; an ARA Compiler that translates legacy PDFs and repos into ARAs; and an ARA-native review system that automates objective checks so human reviewers can focus on significance, novelty, and taste. On PaperBench and RE-Bench, ARA raises question-answering accuracy from 72.4% to 93.7% and reproduction success from 57.4% to 64.4%. On RE-Bench's five open-ended extension tasks, preserved failure traces in ARA accelerate progress, but can also constrain a capable agent from stepping outside the prior-run box depending on the agent's capabilities.

Techmeme(15)

  1. Servers operated by Ubuntu and its parent company Canonical have been down for more than a day, following a "sustained, cross-border attack" (Dan Goodin/Ars Technica)

    Dan Goodin / Ars Technica : Servers operated by Ubuntu and its parent company Canonical have been down for more than a day, following a “sustained, cross-border attack” —  Servers operated by Ubuntu and its parent company Canonical were knocked offline on Thursday morning and have remained down ever since …

  2. Apple has stopped offering a 256GB storage option for the Mac mini globally; Mac mini now starts at 512GB for $799 in the US (Joe Rossignol/MacRumors)

    Joe Rossignol / MacRumors : Apple has stopped offering a 256GB storage option for the Mac mini globally; Mac mini now starts at 512GB for $799 in the US —  Apple this week stopped offering a 256GB storage option for the Mac mini worldwide.  As a result, the desktop computer now has a higher starting price.

  3. The Academy of Motion Picture Arts and Sciences issues new rules saying acting and writing must be performed by humans and not AI to be eligible for Oscars (Lisa Richwine/Reuters)

    Lisa Richwine / Reuters : The Academy of Motion Picture Arts and Sciences issues new rules saying acting and writing must be performed by humans and not AI to be eligible for Oscars —  Academy Awards organizers issued new rules on Friday to clarify that acting and writing must be performed by humans …

  4. Sources: Cerebras is seeking to raise as much as $4B in its IPO and is targeting a valuation of about $40B (Bloomberg)

    Bloomberg : Sources: Cerebras is seeking to raise as much as $4B in its IPO and is targeting a valuation of about $40B —  Cerebras Systems Inc. is seeking to raise as much as $4 billion in its initial public offering, according to people familiar with the matter, as demand for the artificial intelligence chipmaker …

  5. How influencers boosting US-based AI and opposing Chinese AI are paid with money tied to Leading the Future, funded by execs from OpenAI, Palantir, a16z, others (Taylor Lorenz/Wired)

    Taylor Lorenz / Wired : How influencers boosting US-based AI and opposing Chinese AI are paid with money tied to Leading the Future, funded by execs from OpenAI, Palantir, a16z, others —  Build American AI, a nonprofit linked to a super PAC bankrolled by executives at OpenAI and Andreessen Horowitz …

  6. "Podslop" is flooding listening platforms like Spotify; Podcast Index: of ~11K new podcast feeds in a recent 9-day span, ~39% were likely for AI content shows (Ashley Carman/Bloomberg)

    Ashley Carman / Bloomberg : “Podslop” is flooding listening platforms like Spotify; Podcast Index: of ~11K new podcast feeds in a recent 9-day span, ~39% were likely for AI content shows —  Over the past nine days, 39% of new podcasts were likely AI-generated, according to the Podcast Index.  —  Welcome back to Soundbite.

  7. Pushing back on AI job loss fears, AWS CEO Matt Garman says Amazon plans to hire 11,000 software engineering interns in 2026, a figure in line with recent years (Ben Shimkus/Business Insider)

    Ben Shimkus / Business Insider : Pushing back on AI job loss fears, AWS CEO Matt Garman says Amazon plans to hire 11,000 software engineering interns in 2026, a figure in line with recent years —  - AWS's CEO, Matt Garman, said Amazon sees demand for hiring new software engineers “accelerating.”

  8. Some retail traders are training AI agents to buy and sell assets on their behalf, as exchanges like Polymarket and Bybit roll out agent-friendly interfaces (Emily Nicolle/Bloomberg)

    Emily Nicolle / Bloomberg : Some retail traders are training AI agents to buy and sell assets on their behalf, as exchanges like Polymarket and Bybit roll out agent-friendly interfaces —  Jake Nesler's AI trading bot got one big decision right in its first week.  It ignored the chase.

  9. The US, UK, Australia, Canada, and New Zealand publish guidance on orgs' use of agentic AI systems, saying many give AI more access than can be safely monitored (Greg Otto/CyberScoop)

    Greg Otto / CyberScoop : The US, UK, Australia, Canada, and New Zealand publish guidance on orgs' use of agentic AI systems, saying many give AI more access than can be safely monitored —  The guidance warns that agents capable of taking real-world actions on networks are already inside critical infrastructure …

  10. The gender gap in AI use may be a matter of visibility more than usage, as data suggests women face more judgment for using AI and are less likely to admit it (Issie Lapowsky/Bloomberg)

    Issie Lapowsky / Bloomberg : The gender gap in AI use may be a matter of visibility more than usage, as data suggests women face more judgment for using AI and are less likely to admit it —  Women are using artificial intelligence at lower rates than men.  But are they being “left behind”?

  11. The FBI warns of rising cyber cargo theft, where attackers hack freight brokers' accounts and dupe carriers; 2025 cargo theft losses in N. America rose 60% YoY (Pierluigi Paganini/Security Affairs)

    Pierluigi Paganini / Security Affairs : The FBI warns of rising cyber cargo theft, where attackers hack freight brokers' accounts and dupe carriers; 2025 cargo theft losses in N. America rose 60% YoY —  The FBI warns of rising cyber cargo theft, with hackers targeting brokers and carriers.  Experts say digital attacks are replacing traditional cargo theft.

  12. Meta acquires Assured Robot Intelligence, a startup developing AI models for robots; Assured Robot Intelligence's team will join Meta Superintelligence Labs (Mark Gurman/Bloomberg)

    Mark Gurman / Bloomberg : Meta acquires Assured Robot Intelligence, a startup developing AI models for robots; Assured Robot Intelligence's team will join Meta Superintelligence Labs —  Meta Platforms Inc. has acquired Assured Robot Intelligence, a startup developing artificial intelligence models for robots …

  13. A bug in popular cPanel, WHM, and WP Squared software has reportedly been exploited since Feb.; CISA it gives a 9.8 CVSS score, tells agencies to patch by May 3 (Jonathan Greig/The Record)

    Jonathan Greig / The Record : A bug in popular cPanel, WHM, and WP Squared software has reportedly been exploited since Feb.; CISA it gives a 9.8 CVSS score, tells agencies to patch by May 3 —  Federal agencies have until May 3 to resolve a security issue impacting a critical system for server and website management.

  14. Experts say supply chain attacks compromised SAP and Intercom npm packages, plus the PyPI package Lightning, in a campaign that calls itself Mini Shai-Hulud (Jessica Lyons/The Register)

    Jessica Lyons / The Register : Experts say supply chain attacks compromised SAP and Intercom npm packages, plus the PyPI package Lightning, in a campaign that calls itself Mini Shai-Hulud —  The wave of supply chain attacks aimed at security and developer tools has washed up more victims, namely SAP and Intercom npm packages, plus the lightning PyPI package.

  15. Fun, which builds fiat and crypto payment rails for platforms like Polymarket and Aave, raised a $72M Series A in January, co-led by Multicoin and SignalFire (Ben Weiss/Fortune)

    Ben Weiss / Fortune : Fun, which builds fiat and crypto payment rails for platforms like Polymarket and Aave, raised a $72M Series A in January, co-led by Multicoin and SignalFire —  The finer details of back-end payment systems are enough to make most people's eyes glaze over, even those in crypto.

Solidot(15)

  1. 数据中心开发商 Pure Data 暂停中东投资项目

    在其设施遭袭受损之后,数据中心开发商 Pure Data 暂停所有中东项目投资。Pure Data 在欧洲、亚洲和中东运营或开发逾 1GW 的数据中心。数据中心作为基础设施成为了战争中的一个重要目标。亚马逊 AWS 在中东有三座数据中心遭到袭击,导致中东客户的服务出现大规模中断,迫使亚马逊宣布免除其中东云区域客户所有费用,导致其损失了约 1.5 亿美元。Pure Data 位于阿布扎比 Yas Island 的数据中心园区遭到了弹片的袭击。该公司没有披露发生的时间以及受损情况。

  2. 德国 2025 年新生儿数量降至 1946 年以来最低水平

    德国联邦统计局的初步数据显示,2025 年新生儿数量降至 1946 年以来最低水平。2025 年德国新生儿数约 65.5 万,远低于 1964 年婴儿潮高峰时的 136 万,2024 年的新生儿数据是 68 万。与此同时德国死亡人数接近 101 万,使得 2025 年死亡人数与出生人数之差超过 35.2 万,创战后历史新高。德国出生率连续第四年下降,目前每名妇女平均生育 1.35 个孩子,创历史新低,远低于维持人口稳定所需的 2.1 个孩子。汉堡是唯一一个生育率上升的德国州,2025 年增长了 0.5%。

  3. Google 给你贴上的价格标签

    瑞士邮件服务商 Proton 利用 2025 年广告竞价数据,分析了逾 54,000 个人口画像,估算广告商为触达不同类美国人所支付的价格。结果显示不同人之间的价格差距远超想象。美国人平均每年产生的广告价值约 1,605 美元;一名居住在蒙大拿州 Bozeman 市、年龄 35-44 岁之间、无子女、用台式机进行高价值企业搜索的男性,其广告价值估计为 17,929.30 美元;一位居住在阿肯色州 Fort Smith 市、年龄在 18-24 岁之间、用 Android 手机进行低价值搜索的父亲,其广告价值仅为 31.05 美元。1,605 美元的平均值与 760 美元的中位数显示,少数高价值用户拉高了平均值,而此类商业模式依赖于高价值用户。分析显示,无子女用户的广告价值比有子女用户平均高出约 17%,一旦某个用户被标记为有子女,针对他们的广告投放会从每次点击 6 美元的财富管理广告转向每次点击 2 美元的面包车和幼儿园广告。台式机用户的价值是 Android 用户的 4.9 倍,苹果 iPhone 用户的价值是 Android 用户的 2.7 倍。用户年龄在 35-44 岁之间时广告价值最高,65 岁后广告价值下降——虽然老年用户价值下降,但针对他们的广告则属于高消费类别如医保补充保险、药品和金融产品。老年人的总体价值降低,但广告商的投放力度更精准。为什么蒙大拿州 Bozeman 市居民的广告价值高?因为大量远程科技工作者的涌入和户外休闲消费使其成为全美竞争最激烈的本地广告市场之一。

  4. 亚洲多国加大燃煤发电以应对能源危机

    最新的中东能源危机促使亚洲国家加大燃煤发电,而煤炭是高污染排放来源,如果这一趋势继续,全球气候变化问题将会愈发严峻。印度宣布推迟对国内燃煤电厂的维护检查。国际能源署(IEA)的数据显示,截至 2023 年,印度发电中煤炭占 74%。石油和天然气合计约占 3%,来自中东的采购存在制约,印度通过增加煤炭火力来避免停电风险。泰国电力公司重启原计划停用的 2 座燃煤机组。韩国暂时解除了以发电能力 80% 为上限的煤炭火电站的运行限制,推迟原定于 6 月关闭的两座火电站的关闭时间。日本也将提高煤炭火力发电站的开工率。孟加拉国则增加煤炭的供应来源。全球最大的发电用煤炭出口国印尼计划上调原定为 6 亿吨的 2026 年煤炭生产计划。第二大出口国澳大利亚政府也计划扩大煤炭生产。

  5. 活动邀请 | NVIDIA 开发者见面会:从基础设施到智能体,全链路专家深度解析

    从底层的 GPU 开发,到长上下文的大模型推理,再到能够自主规划的 Agentic AI,AI 技术的演进正全方位重塑软件开发的范式。为了帮助开发者更好地应对日益复杂的全栈挑战,NVIDIA 企业开发者社区诚邀您参加即将在苏州举办的 NVIDIA 开发者见面会。本次见面会将汇集来自 NVIDIA 全球及本地的技术专家,为您带来从基础设施优化到前沿智能体应用的全链路干货分享。  查看全文

  6. 水产养殖的温室气体排放

    发表在《Frontiers of Agricultural Science and Engineering》期刊上的一项研究发现,水产养殖的温室气体排放主要来自四个环节:饲料生产、养殖过程中的能源消耗、池塘或水体中的生物化学过程(如甲烷和氧化亚氮的释放),以及土地利用变化和基础设施建设。其中饲料生产是大多数投饵型养殖系统中最大的排放源,在我国的研究中占比达到 52%。而在我国等以淡水池塘养殖为主的地区,甲烷排放尤为突出,贡献了约 90% 的养殖系统温室气体排放。不同水产养殖物种之间的排放差异显著。例如不依赖投饵的双壳贝类(如牡蛎、蛤蜊)和海藻养殖,排放极低甚至为负值,反而能通过碳固定起到“碳汇”作用。而草食性或杂食性鱼类(如鲤鱼、罗非鱼)在适度养殖强度下排放也相对较低。相比之下,集约化养殖的肉食性鱼类(如鲑鱼、鳟鱼)和虾类由于饲料和能源需求高,碳排放强度显著上升,部分甚至与陆地畜牧业相当。

  7. 微软公开 86-DOS 1.00 源代码

    2018 年微软公开了 MS-DOS 1.25 和 2.11 源代码,2024 年公开了 MS-DOS 4.0 源代码,2026 年 4 月在 86-DOS 1.00 发布 45 周年之际它延续传统公开了 86-DOS 1.00 源代码。86-DOS 的作者是 Tim Paterson,它后来成为 MS-DOS 的基础。发布在 GitHub 上的内容包括了 86-DOS 1.00 内核源代码、内核的多个快照,以及知名工具 CHKDSK 等。

  8. 基因组学先驱 Craig Venter 去世,享年 79 岁

    基因组学先驱 Craig Venter 去世,享年 79 岁。他因在 1990 年代末与人类基因组计划竞争建立一个基因组数据库而闻名,但他的数据库是设想要付费才能访问,因此在科学界并不受欢迎,促使其他科研团队加速公开发布基因组测序结果。2000 年由美国总统克林顿牵手,人类基因组计划和 Venter 创办的 Celera Genomics 公司同意所有人类基因组数据为人类共同财富,不允许专利保护,且对所有研究者公开。

  9. GCC 17 加入对海光 C86-4G CPU 的支持

    GCC 编译器项目合并了支持海光 C86-4G CPU 的补丁。海光最早是与 AMD 合作的半导体企业,授权提供 AMD Zen 1 CPU的本地化版本,其产品仅供国内市场使用。海光去年五月宣布与中科曙光合并,但年底宣布合并计划终止。 C86-4G 为 16 核/32 线程处理器,其性能接近英特尔的 Raptor Lake CPU,支持 DDR5 和 PCIe Gen 5。海光声称 C86-4G 利用了自主研发的新微架构,但仅从 GCC 补丁看它仍然与 AMD Zen 有许多相似之处。C86-4G 包括了 C86-4G-M4 / C86-4G-M6 / C86-4G-M7 系列,其中 C86-4G-M7 支持 AVX-512 指令集。

  10. Zed 编辑器发布 1.0 版本

    用 Rust 开发的文本编辑器项目 Zed 宣布发布 1.0 版本。开发者表示 1.0 版本并不意味着“完成”或“完美”,而是意味着到达了一个关键点。开发者还宣称 Zed 编辑器是一个 AI 原生编辑器,能并行运行多个 AI 智能体,包括 Claude Agent、Codex、OpenCode,以及 Cursor。AI 构建在编辑器的基础架构之中,而不是附加组件。

  11. 城里的鸟更怕女性,原因未知

    根据发表在《People and Nature》期刊上的一项研究,欧洲大山雀和其它 36 种鸟更惧怕女性,原因未知。研究显示,在鸟儿飞走前男性能比女性多靠近大约一米距离。无论男女衣着,是否长发,身高如何,以何种方式接近鸟,这种现象始终存在。鸟类或许能辨别人类的性别,但具体机制并不清楚。研究人员观察了生活在五个欧洲国家城市中心的鸟类,其中包括已知一见到人类就会飞走的鸟类如喜鹊,以及倾向于稍迟才飞走的鸟类如鸽子。对女性表现出过度恐惧的反应在不同鸟类中间是一致的。研究人员猜测鸟可能通过气味或步态辨别性别。

  12. .icu 域名被短暂劫持

    中国网民报告 .icu 顶级域名的权威服务器在短时间内解析到了错误的 IP。.icu 最知名的域名可能是 996.ICU。不同于个别域名被错误解析,域名服务器被错误解析可能会导致影响扩大化。4 月 28 日 6:22 a.m. UTC 进行的测试显示,.icu 权威服务器之一的 b.nic.icu 在 58.13% 的情况下解析到正确 IP,其余解析到污染 IP。到 16:00 UTC 问题基本解决,99.38% 解析到正确 IP,仅 0.63% 返回污染 IP。

  13. 荷兰政府上线开源代码托管平台

    荷兰政府上线了它自己的开源代码托管平台 code.overheid.nl。该平台是基于 Forgejo——Forgejo 是 Gitea 的分支,是一个类似 GitHub 的 Git 软件开发和版本控制平台,支持 bug 跟踪、Wiki 和代码审查等功能。该平台托管于荷兰政府基础设施之上,所有政府机构都可以免费使用。荷兰政府表示此举是加强数字主权行动的一部分。

  14. 报告称逾三分之二婴儿使用屏幕,最多花 8 小时在屏幕上

    一份报告发现,逾三分之二两岁以下婴儿使用屏幕,少数婴儿每天最多花 8 小时在屏幕上。近三分之一新生儿每天看屏幕的时间超过三小时,4-11 个月大的婴儿近 20% 每天使用屏幕逾一小时。已有证据表明,屏幕时间与儿童肥胖、近视、睡眠和行为问题风险,以及长大后社交问题等负面影响相关联。新生儿父母允许婴儿使用屏幕的一大理由是分散他们注意力,让自己有时间做其它家务活或工作。

  15. 马斯克称他创办非盈利的 OpenAI 是为了对抗 Google

    2024 年马斯克(Elon Musk)向旧金山高等法院起诉 OpenAI 及其联合创始人 Sam Altman 和 Greg Brockman 违反公司的创始原则,将商业利益置于公共利益之上。OpenAI 则公开了马斯克的邮件,证明作为曾经的联合创始人,马斯克同意 OpenAI 建立一个盈利实体,还表示将提供资金,但之后暂停了资金支持,他的目的是获得多数股权和董事会控制权,双方最终因此终止了合作。本周这起诉讼正式进入审讯阶段,马斯克在法庭上作证,称创办 OpenAI 是将其作为一家非盈利公司去对抗 Google,如果 OpenAI 的目标是盈利他不会支持它。马斯克称他在与 Google 联合创始人 Larry Page 就 AI 安全问题上发生争执后萌生了创办非盈利 AI 公司的想法。他担心 Page 没有认真对待 AI 安全问题,因此希望通过一个非盈利的开源替代方案去对抗 Google。