OrangeBot.AI Digest — 2026-04-27
88 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Is my blue your blue? (ismy.blue)
- China blocks Meta's acquisition of AI startup Manus (www.cnbc.com)
- GitHub is having issues now (www.githubstatus.com)
- GitHub Copilot is moving to usage-based billing (github.blog)
- US Supreme Court reviews police use of cell location data (www.nytimes.com)
- Dutch central bank ditches AWS and chooses Lidl for European Cloud (www.techzine.eu)
- “Why not just use Lean?” (lawrencecpaulson.github.io)
- Microsoft and OpenAI end their exclusive and revenue-sharing deal (www.bloomberg.com)
- 4TB of voice samples just stolen from 40k AI contractors at Mercor (app.oravys.com)
- Men who stare at walls (www.alexselimov.com)
- Show HN: OSS Agent I built topped the TerminalBench on Gemini-3-flash-preview (github.com)
- Quarkdown – Markdown with Superpowers (quarkdown.com)
- Pgbackrest is no longer being maintained (github.com)
- Fully Featured Audio DSP Firmware for the Raspberry Pi Pico (github.com)
- Flipdiscs (flipdisc.io)
GitHub Trending(13)
- mattpocock / skills
- abhigyanpatwari / GitNexus
- ComposioHQ / awesome-codex-skills
- Alishahryar1 / free-claude-code
- gastownhall / beads
- penpot / penpot
- davila7 / claude-code-templates
- microsoft / VibeVoice
- Z4nzu / hackingtool
- TauricResearch / TradingAgents
- CJackHwang / ds2api
- deepseek-ai / DeepSeek-V3
- donnemartin / system-design-primer
Product Hunt(15)
- Brew Finder
Discover the best coffee shops to work at around you
- GitBar
Every pull request, one menubar. GitHub, GitLab & Azure
- Wafaa.io
Create secure digital contracts in minutes
- Jet AI Agents
Build business AI agents in minutes
- Logic
Build and operate fleets of agents
- PlayJoob
turns dead task boards into a shared strategy map
- SNEWPapers
The World's First AI Newspaper Archive
- Vouch API
AI equity research that proves it isn't lying
- Epismo Agent Package
Run agent workflows the community already built
- Atech
Snap-together electronics built from a chat
- VIDEO AI ME
Create videos with AI actors that sound and look real
- Odyssey-2 Max
Physical accuracy takes a leap in world models
- Waitlister
The waitlist software to launch your product
- Orange Slice
Automate any sales task with AI
- Subgrapher
P2P desktop app for building, browsing, & sharing knowledge
Hugging Face(15)
- Agentic World Modeling: Foundations, Capabilities, Laws, and Beyond
As AI systems move from generating text to accomplishing goals through sustained interaction, the ability to model environment dynamics becomes a central bottleneck. Agents that manipulate objects, navigate software, coordinate with others, or design experiments require predictive environment models, yet the term world model carries different meanings across research communities. We introduce a "levels x laws" taxonomy organized along two axes. The first defines three capability levels: L1 Predictor, which learns one-step local transition operators; L2 Simulator, which composes them into multi-step, action-conditioned rollouts that respect domain laws; and L3 Evolver, which autonomously revises its own model when predictions fail against new evidence. The second identifies four governing-law regimes: physical, digital, social, and scientific. These regimes determine what constraints a world model must satisfy and where it is most likely to fail. Using this framework, we synthesize over 400 works and summarize more than 100 representative systems spanning model-based reinforcement learning, video generation, web and GUI agents, multi-agent social simulation, and AI-driven scientific discovery. We analyze methods, failure modes, and evaluation practices across level-regime pairs, propose decision-centric evaluation principles and a minimal reproducible evaluation package, and outline architectural guidance, open problems, and governance challenges. The resulting roadmap connects previously isolated communities and charts a path from passive next-step prediction toward world models that can simulate, and ultimately reshape, the environments in which agents operate.
- Video Analysis and Generation via a Semantic Progress Function
Transformations produced by image and video generation models often evolve in a highly non-linear manner: long stretches where the content barely changes are followed by sudden, abrupt semantic jumps. To analyze and correct this behavior, we introduce a Semantic Progress Function, a one-dimensional representation that captures how the meaning of a given sequence evolves over time. For each frame, we compute distances between semantic embeddings and fit a smooth curve that reflects the cumulative semantic shift across the sequence. Departures of this curve from a straight line reveal uneven semantic pacing. Building on this insight, we propose a semantic linearization procedure that reparameterizes (or retimes) the sequence so that semantic change unfolds at a constant rate, yielding smoother and more coherent transitions. Beyond linearization, our framework provides a model-agnostic foundation for identifying temporal irregularities, comparing semantic pacing across different generators, and steering both generated and real-world video sequences toward arbitrary target pacing.
- DiffNR: Diffusion-Enhanced Neural Representation Optimization for Sparse-View 3D Tomographic Reconstruction
Neural representations (NRs), such as neural fields and 3D Gaussians, effectively model volumetric data in computed tomography (CT) but suffer from severe artifacts under sparse-view settings. To address this, we propose DiffNR, a novel framework that enhances NR optimization with diffusion priors. At its core is SliceFixer, a single-step diffusion model designed to correct artifacts in degraded slices. We integrate specialized conditioning layers into the network and develop tailored data curation strategies to support model finetuning. During reconstruction, SliceFixer periodically generates pseudo-reference volumes, providing auxiliary 3D perceptual supervision to fix underconstrained regions. Compared to prior methods that embed CT solvers into time-consuming iterative denoising, our repair-and-augment strategy avoids frequent diffusion model queries, leading to better runtime performance. Extensive experiments show that DiffNR improves PSNR by 3.99 dB on average, generalizes well across domains, and maintains efficient optimization.
- LLM Safety From Within: Detecting Harmful Content with Internal Representations
Guard models are widely used to detect harmful content in user prompts and LLM responses. However, state-of-the-art guard models rely solely on terminal-layer representations and overlook the rich safety-relevant features distributed across internal layers. We present SIREN, a lightweight guard model that harnesses these internal features. By identifying safety neurons via linear probing and combining them through an adaptive layer-weighted strategy, SIREN builds a harmfulness detector from LLM internals without modifying the underlying model. Our comprehensive evaluation shows that SIREN substantially outperforms state-of-the-art open-source guard models across multiple benchmarks while using 250 times fewer trainable parameters. Moreover, SIREN exhibits superior generalization to unseen benchmarks, naturally enables real-time streaming detection, and significantly improves inference efficiency compared to generative guard models. Overall, our results highlight LLM internal states as a promising foundation for practical, high-performance harmfulness detection.
- FlowAnchor: Stabilizing the Editing Signal for Inversion-Free Video Editing
We propose FlowAnchor, a training-free framework for stable and efficient inversion-free, flow-based video editing. Inversion-free editing methods have recently shown impressive efficiency and structure preservation in images by directly steering the sampling trajectory with an editing signal. However, extending this paradigm to videos remains challenging, often failing in multi-object scenes or with increased frame counts. We identify the root cause as the instability of the editing signal in high-dimensional video latent spaces, which arises from imprecise spatial localization and length-induced magnitude attenuation. To overcome this challenge, FlowAnchor explicitly anchors both where to edit and how strongly to edit. It introduces Spatial-aware Attention Refinement, which enforces consistent alignment between textual guidance and spatial regions, and Adaptive Magnitude Modulation, which adaptively preserves sufficient editing strength. Together, these mechanisms stabilize the editing signal and guide the flow-based evolution toward the desired target distribution. Extensive experiments demonstrate that FlowAnchor achieves more faithful, temporally coherent, and computationally efficient video editing across challenging multi-object and fast-motion scenarios. The project page is available at https://cuc-mipg.github.io/FlowAnchor.github.io/.
- AgentSearchBench: A Benchmark for AI Agent Search in the Wild
The rapid growth of AI agent ecosystems is transforming how complex tasks are delegated and executed, creating a new challenge of identifying suitable agents for a given task. Unlike traditional tools, agent capabilities are often compositional and execution-dependent, making them difficult to assess from textual descriptions alone. However, existing research and benchmarks typically assume well-specified functionalities, controlled candidate pools, or only executable task queries, leaving realistic agent search scenarios insufficiently studied. We introduce AgentSearchBench, a large-scale benchmark for agent search in the wild, built from nearly 10,000 real-world agents across multiple providers. The benchmark formalizes agent search as retrieval and reranking problems under both executable task queries and high-level task descriptions, and evaluates relevance using execution-grounded performance signals. Experiments reveal a consistent gap between semantic similarity and actual agent performance, exposing the limitations of description-based retrieval and reranking methods. We further show that lightweight behavioral signals, including execution-aware probing, can substantially improve ranking quality, highlighting the importance of incorporating execution signals into agent discovery. Our code is available at https://github.com/Bingo-W/AgentSearchBench.
- Contexts are Never Long Enough: Structured Reasoning for Scalable Question Answering over Long Document Sets
Real-world document question answering is challenging. Analysts must synthesize evidence across multiple documents and different parts of each document. However, any fixed LLM context window can be exceeded as document collections grow. A common workaround is to decompose documents into chunks and assemble answers from chunk-level outputs, but this introduces an aggregation bottleneck: as the number of chunks grows, systems must still combine and reason over an increasingly large body of extracted evidence. We present SLIDERS, a framework for question answering over long document collections through structured reasoning. SLIDERS extracts salient information into a relational database, enabling scalable reasoning over persistent structured state via SQL rather than concatenated text. To make this locally extracted representation globally coherent, SLIDERS introduces a data reconciliation stage that leverages provenance, extraction rationales, and metadata to detect and repair duplicated, inconsistent, and incomplete records. SLIDERS outperforms all baselines on three existing long-context benchmarks, despite all of them fitting within the context window of strong base LLMs, exceeding GPT-4.1 by 6.6 points on average. It also improves over the next best baseline by ~19 and ~32 points on two new benchmarks at 3.9M and 36M tokens, respectively.
- Building a Precise Video Language with Human-AI Oversight
Video-language models (VLMs) learn to reason about the dynamic visual world through natural language. We introduce a suite of open datasets, benchmarks, and recipes for scalable oversight that enable precise video captioning. First, we define a structured specification for describing subjects, scenes, motion, spatial, and camera dynamics, grounded by hundreds of carefully defined visual primitives developed with professional video creators such as filmmakers. Next, to curate high-quality captions, we introduce CHAI (Critique-based Human-AI Oversight), a framework where trained experts critique and revise model-generated pre-captions into improved post-captions. This division of labor improves annotation accuracy and efficiency by offloading text generation to models, allowing humans to better focus on verification. Additionally, these critiques and preferences between pre- and post-captions provide rich supervision for improving open-source models (Qwen3-VL) on caption generation, reward modeling, and critique generation through SFT, DPO, and inference-time scaling. Our ablations show that critique quality in precision, recall, and constructiveness, ensured by our oversight framework, directly governs downstream performance. With modest expert supervision, the resulting model outperforms closed-source models such as Gemini-3.1-Pro. Finally, we apply our approach to re-caption large-scale professional videos (e.g., films, commercials, games) and fine-tune video generation models such as Wan to better follow detailed prompts of up to 400 words, achieving finer control over cinematography including camera motion, angle, lens, focus, point of view, and framing. Our results show that precise specification and human-AI oversight are key to professional-level video understanding and generation. Data and code are available on our project page: https://linzhiqiu.github.io/papers/chai/
- Sessa: Selective State Space Attention
Modern sequence modeling is dominated by two families: Transformers, whose self-attention can access arbitrary elements of the visible sequence, and structured state-space models, which propagate information through an explicit recurrent state. These mechanisms face different limitations on long contexts: when attention is diffuse, the influence of individual tokens is diluted across the effective support, while recurrent state propagation can lose long-range sensitivity unless information is actively preserved. As a result, both mechanisms face challenges in preserving and selectively retrieving information over long contexts. We propose Sessa, a decoder that places attention inside a recurrent feedback path. This creates many attention-based paths through which past tokens can influence future states, rather than relying on a single attention read or a single recurrent chain. We prove that, under explicit assumptions and matched regimes, Sessa admits power-law memory tails O(ell^{-β}) for 0 < β< 1, with slower decay than in the corresponding Transformer and Mamba-style baselines. We further give an explicit construction that achieves this power-law rate. Under the same assumptions, Sessa is the only model class among those considered that realizes flexible selective retrieval, including profiles whose influence does not decay with distance. Consistent with this theoretical advantage, across matched experiments, Sessa achieves the strongest performance on long-context benchmarks while remaining competitive with Transformer and Mamba-style baselines on short-context language modeling.
- Memanto: Typed Semantic Memory with Information-Theoretic Retrieval for Long-Horizon Agents
The transition from stateless language model inference to persistent, multi session autonomous agents has revealed memory to be a primary architectural bottleneck in the deployment of production grade agentic systems. Existing methodologies largely depend on hybrid semantic graph architectures, which impose substantial computational overhead during both ingestion and retrieval. These systems typically require large language model mediated entity extraction, explicit graph schema maintenance, and multi query retrieval pipelines. This paper introduces Memanto, a universal memory layer for agentic artificial intelligence that challenges the prevailing assumption that knowledge graph complexity is necessary to achieve high fidelity agent memory. Memanto integrates a typed semantic memory schema comprising thirteen predefined memory categories, an automated conflict resolution mechanism, and temporal versioning. These components are enabled by Moorcheh's Information Theoretic Search engine, a no indexing semantic database that provides deterministic retrieval within sub ninety millisecond latency while eliminating ingestion delay. Through systematic benchmarking on the LongMemEval and LoCoMo evaluation suites, Memanto achieves state of the art accuracy scores of 89.8 percent and 87.1 percent respectively. These results surpass all evaluated hybrid graph and vector based systems while requiring only a single retrieval query, incurring no ingestion cost, and maintaining substantially lower operational complexity. A five stage progressive ablation study is presented to quantify the contribution of each architectural component, followed by a discussion of the implications for scalable deployment of agentic memory systems.
- EmbodiedMidtrain: Bridging the Gap between Vision-Language Models and Vision-Language-Action Models via Mid-training
Vision-Language-Action Models (VLAs) inherit their visual and linguistic capabilities from Vision-Language Models (VLMs), yet most VLAs are built from off-the-shelf VLMs that are not adapted to the embodied domain, limiting their downstream performance. In this work, we propose EmbodiedMidtrain to bridge the gap between VLMs and VLAs. We first characterize the data distribution gap between them, showing that VLA data occupy compact regions that are largely separated from the broader VLM distribution, while the degree of alignment varies substantially both across and within VLM data sources. Then, we build a mid-training data engine that leverages a lightweight learnable proximity estimator to select the most VLA-aligned candidates from a large VLM pool, and mid-trains the VLM on this curated mixture before downstream VLA fine-tuning. Experiments on three robot manipulation benchmarks show that mid-training consistently improves performance across different VLM backbones, achieving results competitive with expert VLAs and off-the-shelf VLMs trained with larger model scale and training budgets. Further analysis reveals that mid-training provides a stronger initialization for VLA fine-tuning, with gains emerging from the earliest steps and widening throughout training. Moreover, the data engine captures both dataset-level and sample-level alignment signals, favoring spatial reasoning over text-centric tasks while preserving the diversity of the VLM data. We will release all code, data and models for future research.
- dWorldEval: Scalable Robotic Policy Evaluation via Discrete Diffusion World Model
Evaluating robotics policies across thousands of environments and thousands of tasks is infeasible with existing approaches. This motivates the need for a new methodology for scalable robotics policy evaluation. In this paper, we propose dWorldEval, which uses a discrete diffusion world model as a scalable evaluation proxy for robotics policies. Specifically, dWorldEval maps all modalities - including vision, language, and robotic actions - into a unified token space, modeling them via a single transformer-based denoising network. In this paper, we propose dWorldEval, using a discrete diffusion world model as a scalable evaluation proxy for robotics policy. Specifically, it maps all modalities, including vision, language, and robotics action into a unified token space, then denoises them with a single transformer network. Building on this architecture, we employ a sparse keyframe memory to maintain spatiotemporal consistency. We also introduce a progress token that indicates the degree of task completion. At inference, the model jointly predicts future observations and progress token, allowing automatically determine success when the progress reaches 1. Extensive experiments demonstrate that dWorldEval significantly outperforms previous approaches, i.e., WorldEval, Ctrl-World, and WorldGym, on LIBERO, RoboTwin, and multiple real-robot tasks. It paves the way for a new architectural paradigm in building world simulators for robotics evaluation at scale.
- DiagramBank: A Large-scale Dataset of Diagram Design Exemplars with Paper Metadata for Retrieval-Augmented Generation
Recent advances in autonomous ``AI scientist'' systems have demonstrated the ability to automatically write scientific manuscripts and codes with execution. However, producing a publication-grade scientific diagram (e.g., teaser figure) is still a major bottleneck in the ``end-to-end'' paper generation process. For example, a teaser figure acts as a strategic visual interface and serves a different purpose than derivative data plots. It demands conceptual synthesis and planning to translate complex logic workflow into a compelling graphic that guides intuition and sparks curiosity. Existing AI scientist systems usually omit this component or fall back to an inferior alternative. To bridge this gap, we present DiagramBank, a large-scale dataset consisting of 89,422 schematic diagrams curated from existing top-tier scientific publications, designed for multimodal retrieval and exemplar-driven scientific figure generation. DiagramBank is developed through our automated curation pipeline that extracts figures and corresponding in-text references, and uses a CLIP-based filter to differentiate schematic diagrams from standard plots or natural images. Each instance is paired with rich context from abstract, caption, to figure-reference pairs, enabling information retrieval under different query granularities. We release DiagramBank in a ready-to-index format and provide a retrieval-augmented generation codebase to demonstrate exemplar-conditioned synthesis of teaser figures. DiagramBank is publicly available at https://huggingface.co/datasets/zhangt20/DiagramBank with code at https://github.com/csml-rpi/DiagramBank.
- AgriIR: A Scalable Framework for Domain-Specific Knowledge Retrieval
This paper introduces AgriIR, a configurable retrieval augmented generation (RAG) framework designed to deliver grounded, domain-specific answers while maintaining flexibility and low computational cost. Instead of relying on large, monolithic models, AgriIR decomposes the information access process into declarative modular stages -- query refinement, sub-query planning, retrieval, synthesis, and evaluation. This design allows practitioners to adapt the framework to new knowledge verticals without modifying the architecture. Our reference implementation targets Indian agricultural information access, integrating 1B-parameter language models with adaptive retrievers and domain-aware agent catalogues. The system enforces deterministic citation, integrates telemetry for transparency, and includes automated deployment assets to ensure auditable, reproducible operation. By emphasizing architectural design and modular control, AgriIR demonstrates that well-engineered pipelines can achieve domain-accurate, trustworthy retrieval even under constrained resources. We argue that this approach exemplifies ``AI for Agriculture'' by promoting accessibility, sustainability, and accountability in retrieval-augmented generation systems.
- Learning Evidence Highlighting for Frozen LLMs
Large Language Models (LLMs) can reason well, yet often miss decisive evidence when it is buried in long, noisy contexts. We introduce HiLight, an Evidence Emphasis framework that decouples evidence selection from reasoning for frozen LLM solvers. HiLight avoids compressing or rewriting the input, which can discard or distort evidence, by training a lightweight Emphasis Actor to insert minimal highlight tags around pivotal spans in the unaltered context. A frozen Solver then performs downstream reasoning on the emphasized input. We cast highlighting as a weakly supervised decision-making problem and optimize the Actor with reinforcement learning using only the Solver's task reward, requiring no evidence labels and no access to or modification of the Solver. Across sequential recommendation and long-context question answering, HiLight consistently improves performance over strong prompt-based and automated prompt-optimization baselines. The learned emphasis policy transfers zero-shot to both smaller and larger unseen Solver families, including an API-based Solver, suggesting that the Actor captures genuine, reusable evidence structure rather than overfitting to a single backbone.
Techmeme(15)
- Renders based on photos of Samsung's upcoming smart glasses, expected to launch later this year, show a design nearly identical to Ray-Ban Meta glasses (Alexander Maxham/Android Headlines)
Alexander Maxham / Android Headlines : Renders based on photos of Samsung's upcoming smart glasses, expected to launch later this year, show a design nearly identical to Ray-Ban Meta glasses — Add Android Headlines as a preferred source on Google — Samsung's next Android XR product is reportedly a pair of smart glasses codenamed …
- Study: only ~3% of Polymarket accounts drove most price discovery in 2023-2025, suggesting market accuracy comes from an informed minority, not crowd wisdom (Sam Reynolds/CoinDesk)
Sam Reynolds / CoinDesk : Study: only ~3% of Polymarket accounts drove most price discovery in 2023-2025, suggesting market accuracy comes from an informed minority, not crowd wisdom — Researchers show market accuracy comes from a tiny group of informed traders, not broad participation. … What to know:
- Elon Musk boosts an X post by Ronan Farrow promoting his New Yorker article on Sam Altman's alleged deceptions, as Musk's lawsuit against OpenAI heads to trial (Wired)
Wired : Elon Musk boosts an X post by Ronan Farrow promoting his New Yorker article on Sam Altman's alleged deceptions, as Musk's lawsuit against OpenAI heads to trial — The move comes as the trial for Elon Musk's lawsuit against OpenAI kicks off in federal court in Oakland.
- Jury selection begins in Musk v. Altman trial at a federal courthouse in California, with Sam Altman and Greg Brockman in attendance (Ashley Capoot/CNBC)
Ashley Capoot / CNBC : Jury selection begins in Musk v. Altman trial at a federal courthouse in California, with Sam Altman and Greg Brockman in attendance — The nine-person jury was seated on Monday in the high-stakes legal battle between longtime friends turned rivals Elon Musk and Sam Altman at a federal courthouse in Oakland, California.
- Have I Been Pwned: ShinyHunters' breach of ADT exposed the personal data of 5.5M people; ADT previously disclosed data breaches in August 2024 and October 2024 (Sergiu Gatlan/BleepingComputer)
Sergiu Gatlan / BleepingComputer : Have I Been Pwned: ShinyHunters' breach of ADT exposed the personal data of 5.5M people; ADT previously disclosed data breaches in August 2024 and October 2024 — The ShinyHunters extortion group stole the personal information of 5.5 million individuals after breaching the systems …
- GitHub says all Copilot plans will move to usage-based billing on June 1, replacing premium requests with monthly GitHub AI Credits (Mario Rodriguez/The GitHub Blog)
Mario Rodriguez / The GitHub Blog : GitHub says all Copilot plans will move to usage-based billing on June 1, replacing premium requests with monthly GitHub AI Credits — Starting June 1, your Copilot usage will consume GitHub AI Credits. — TL;DR: Today, we are announcing that all GitHub Copilot plans will transition to usage-based billing on June 1, 2026.
- The EU unveils new proposals under the DMA aimed at opening up Android to rivals' AI services; Google says the measures are "unwarranted intervention" (Samuel Stolton/Bloomberg)
Samuel Stolton / Bloomberg : The EU unveils new proposals under the DMA aimed at opening up Android to rivals' AI services; Google says the measures are “unwarranted intervention” — Google was targeted by European Union watchdogs who unveiled a slate of proposals aimed at prising open its Android ecosystem to rivals' AI services.
- More than 600 Google employees, including many from DeepMind, sign a letter to Sundar Pichai demanding he bar the DOD from using Google's AI for classified work (Gerrit De Vynck/Washington Post)
Gerrit De Vynck / Washington Post : More than 600 Google employees, including many from DeepMind, sign a letter to Sundar Pichai demanding he bar the DOD from using Google's AI for classified work — “We want to see AI benefit humanity; not to see it being used in inhumane or extremely harmful ways,” Google employees wrote. — Summary
- Kashable, which lets companies offer "socially responsible" credit and financial wellness programs for employees as a voluntary benefit, raised a $60M Series C (Mary Ann Azevedo/Crunchbase News)
Mary Ann Azevedo / Crunchbase News : Kashable, which lets companies offer “socially responsible” credit and financial wellness programs for employees as a voluntary benefit, raised a $60M Series C — Kashable, a fintech that provides access to “socially responsible” credit and financial wellness programs for employees …
- The founder of car rental platform PocketOS says a Cursor agent using Claude Opus 4.6 accidentally deleted a production database while in a staging environment (Jer/@lifeof_jer)
Jer / @lifeof_jer : The founder of car rental platform PocketOS says a Cursor agent using Claude Opus 4.6 accidentally deleted a production database while in a staging environment — An AI Agent Just Destroyed Our Production Data. It Confessed in Writing.
- Canva says it "moved quickly to investigate and fix" an issue with its Magic Layers feature that replaced the word "Palestine" in designs, after a viral X post (Jess Weatherbed/The Verge)
Jess Weatherbed / The Verge : Canva says it “moved quickly to investigate and fix” an issue with its Magic Layers feature that replaced the word “Palestine” in designs, after a viral X post — The Magic Layers feature is off to a good start. … One of Canva's new AI features …
- Ineffable Intelligence, founded by ex-Google DeepMind Principal Scientist David Silver, raised a $1.1B seed at a $5.1B valuation to build AI "superlearners" (Will Knight/Wired)
Will Knight / Wired : Ineffable Intelligence, founded by ex-Google DeepMind Principal Scientist David Silver, raised a $1.1B seed at a $5.1B valuation to build AI “superlearners” — David Silver has a new billion-dollar company that aims to build AI “superlearners.” — David Silver gave the world its very first glimpse of superintelligence.
- Microsoft and OpenAI remove a clause that would have given Microsoft IP rights until OpenAI achieved "AGI"; Microsoft retains use of OpenAI's models until 2032 (Aaron Holmes/The Information)
Aaron Holmes / The Information : Microsoft and OpenAI remove a clause that would have given Microsoft IP rights until OpenAI achieved “AGI”; Microsoft retains use of OpenAI's models until 2032 — Microsoft and OpenAI amended the terms of their arrangement, allowing OpenAI to sell its models on competing cloud providers, the companies said on Monday.
- Quantum Art, a quantum computing startup focused on enhancing computational throughput using its unique "multicore" architecture, extends its Series A to $140M (Mike Wheatley/SiliconANGLE)
Mike Wheatley / SiliconANGLE : Quantum Art, a quantum computing startup focused on enhancing computational throughput using its unique “multicore” architecture, extends its Series A to $140M — Quantum Art Ltd., a quantum computing startup focused on enhancing computational throughput, said today it has extended …
- Microsoft and OpenAI amend their deal to let OpenAI serve all its products across any cloud provider; Microsoft will no longer pay a revenue share to OpenAI (OpenAI)
OpenAI : Microsoft and OpenAI amend their deal to let OpenAI serve all its products across any cloud provider; Microsoft will no longer pay a revenue share to OpenAI — Amended agreement provides long-term clarity. — The rapid pace of innovation requires us to continue to evolve our partnership to benefit our customers and both companies.
Solidot(15)
- 老房子闹鬼可能源于陈旧设施产生的次声波
觉得老房子闹鬼?你可能是受到了陈旧设施如旧管道和旧锅炉产生的次声波的影响。根据发表在《Frontiers in Behavioural Neuroscience》期刊上的一项研究,研究人员让 36 名志愿者听轻音乐或鬼屋景点播放的那种令人心神不宁的音乐。在参与者不知情下,研究人员悄悄在半数情况下播放了次声波。结果显示,次声波让志愿者感到更烦躁和恼怒,觉得音乐更悲伤,且唾液中的皮质醇水平更高。研究人员称,人耳听不到次声波,但身体和情绪仍然能做出反应,且通常是不愉快的反应。《The Science of Weird Shit: Why Our Minds Conjure the Paranormal》一书的作者 Chris French 教授认为用次声波解释闹鬼有点牵强。
- 欧洲批准了 Moderna 的流感和 COVID-19 联合疫苗
欧洲批准了 Moderna 研发的基于 mRNA 技术的流感和 COVID-19 联合疫苗。被称为 mRNA-1083 或 mCOMBRIAX 的疫苗成为全球首个获得批准的针对这两种呼吸道病毒的联合疫苗。疫苗获批是基于一项 4000 名成年人参与的 III 期临床试验结果。试验分为两组,一组为 50-64 岁的受试者,与标准流感疫苗进行比较;另一组为 65 岁及以上的受试者,与高剂量流感疫苗进行比较。两组受试者中,相对于对照组 mCOMBRIAX 疫苗都能诱导对常见流感病毒株(A/H1N1、A/H3N2 和 B/Victoria)以及 SARS-CoV-2 病毒产生统计上显著更高的免疫反应。试验未发现安全性或不良反应方面的问题。
- 杀虫剂导致北美蝴蝶数量大减
2025 年 3 月科学家在《科学》期刊上发表研究,Xerces Society for Invertebrate Conservation 保护协会随后发表了蝴蝶现状报告。研究发现,从 2000 年到 2020 年全美蝴蝶总数下降了 22%,有 24 种蝴蝶数量下降 90% 或以上。杀虫剂被认为是导致这一结果的主要原因。1960 年代化学公司研制出了强效杀虫剂滴滴涕(DDT),公众对滴滴涕的反对促使企业研制出弱化对人类伤害但强化对昆虫杀伤力的新杀虫剂。多种混合型杀虫剂的使用导致蝴蝶等昆虫在 21 世纪加速减少。生态学家 Matt Forister 等人在《Environmental Toxicology and Chemistry》期刊上报告,他们分析了 336 株植物只有 22 株植物没有检测到农药残留。这些植物至少含有三种化学物质,其中 71 株植物的农药浓度对蝴蝶而言是致命或接近致命。在 2022 年的一项类似研究中,Forister 团队分析了 33 家苗圃出售的 235 株乳草(对帝王蝶至关重要的植物),发现每株植物平均含有 12.2 种杀虫剂。
- 发改委要求撤销对 Manus 的收购
国家发展改革委周一发布通报,外商投资安全审查工作机制办公室(国家发改委)依法依规对外资收购 Manus 项目作出禁止投资决定,要求当事人撤销该收购交易。《外商投资安全审查办法》于 2020 年 12 月 19 日由国家发展改革委、商务部联合发布,自 2021 年 1 月 18 日开始施行,对适用审查的外商投资类型、审查机构、审查范围、审查程序、审查决定监督执行和违规处理等进行规定。跟据该文件,国家建立外商投资安全审查工作机制,工作机制办公室设在国家发展改革委,由国家发展改革委、商务部牵头,承担外商投资安全审查的日常工作。
- Greg K-H 使用基于 AMD Ryzen AI Max 的 AI 工具发现内核 Bug
稳定版内核维护者 Greg Kroah-Hartman 正在使用名为 gkh_clanker_t1000 的 AI 工具发现内核 Bug。他在 Mastodon 上分享了 gkh_clanker_t1000 的硬件图片。gkh_clanker_t1000 运行在搭载 AMD Ryzen AI Max+“Strix Halo”APU 的 Framework Desktop 之上,Ryzen AI Max+ 提供了最高 128GB 的统一内存,可分配 96GB 内存给 GPU 使用,其性能足以运行本地大模型以及其它基于开源软件栈的 AI 工具。Greg K-H 尚未透露 gkh_clanker_t1000 软件方面的信息。
- AI 成本可能高于人工成本
多家企业在 AI 上的支出已经超过了员工薪资,IT 预算严重超支。Uber CTO 的 2026 年 AI 预算因 token 费用超支。根据 Gartner 预测,2026 年全球 IT 支出预计将达到 6.31 万亿美元,比 2025 年增长 13.5%。这一增长是由 AI 基础设施、软件和云服务的“持续发展势头”推动的。即使是 IT 预算最充足的公司也需要证明 AI 投入的长期回报。当 AI 实验室提高价格时,对 AI 的大量投入可能会从一种炫耀的资本变成一种负担。
- 台积电泄密案 Tokyo Electron 子公司被判有罪
台智慧财产及商业法院 4 月 27 日就台积电(TSMC)机密外泄案做出判决,对日本 Tokyo Electron 台子公司判处缓刑 3 年,罚款 1亿 5000 万新台币(约合人民币 3257.8万元),须赔偿台积电 1 亿新台币。任职于该子公司的台积电前员工陈力铭被判处 10 年有期徒刑。台高等检察署于 2025 年 8 月起诉台积电的三名前员工陈力铭、吴秉骏和戈一平,指控他们非法获取和利用台积电的机密信息。陈力铭表示其目的是改善该公司的生产设备。吴秉骏获刑 3 年,戈一平 2 年、陈韦杰 6 年,卢怡尹 10 个月缓刑 3 年。检方指出,陈力铭为求协助新东家争取订单与技术突破,铤而走险,找上昔日具有师徒关系的台积电工程师吴秉骏与戈一平协助。2 人涉嫌利用远端连线方式进入内部资料库,再偷拍 12 张涉及 2 纳米制程的机密照片,传送给陈力铭,供其整理后写入 Tokyo Electron 内部文件,企图协助量产。
- Linux 7.1-rc1 释出
Linus Torvalds 在内核邮件列表上宣布释出 Linux 7.1-rc1。主要变化包括:移除对部分旧硬件的支持,其中包括 i486 和俄罗斯 Baikal CPU,清理部分旧 PCMCIA 驱动;新 NTFS 驱动,支持 12 种新 SoC,新联想 Legion Go 驱动,Intel QAT Zstd 支持,AMD Zen 6 支持,Intel Linear Address Space Separation (LASS),等等。
- 切尔诺贝利灾难 40 年后
1986 年 4 月 26 日切尔诺贝利核电站 4 号反应堆发生燃料棒破裂堆芯熔毁事故,这起事故是历史上最严重的核电事故。40 年后这场灾难对在核电站周围的野生动物产生了什么影响?人们担心核事故会对附近的动植物造成毁灭性影响,因为人类还能疏散,但动植物无法。40 年后动植物们仍然在隔离区/禁区内繁衍生息,但影响也是显著的。研究人员报告,禁区内的树蛙比外面的同类颜色更深,其体内的黑色素可能起到了屏障作用,能减缓辐射的影响。田鼠线粒体遗传多样性比非污染区的同类更高,可能是辐射暴露引起的基因突变,也可能是其它因素。辐射敏感的松树大批死亡,取而代之的是桦树。曾经人类生活的地区如今游荡着狼、熊、野牛、鹿、野猪和麋鹿,禁区内的狼群数量甚至七倍于自然保护区内的狼群。2014 年一台禁区内的红外相机拍摄到了一只棕熊,而棕熊已经一百多年没有在该地区出现过了。研究人员认为反应堆厂房内生长的黑色真菌提供了黑色素能部分抵抗电离辐射影响的证据。
- 调查称半数澳大利亚青少年仍然能访问社交媒体
澳大利亚禁止 16 岁以下青少年使用社媒的禁令于 2025 年 12 月 10 日生效,社媒公司如 Meta、Snap 和 TikTok 必须删除和停用逾百万未成年人账户,违反者将面临最高 3250 万美元的罚款。但根据英国预防自杀组织 Molly Rose Foundation 对 1050 名 12 至 15 岁澳大利亚青少年的调查,禁令生效前拥有社媒账号的青少年逾六成仍然能访问至少一个平台。TikTok、YouTube 和 Instagram 仍然保留了逾半数 16 岁以下用户账号。三分之二年轻用户表示这些平台未采取任何行动删除禁令前已存在的账户。澳大利亚互联网监管机构正呼吁对五家最大的社媒平台展开调查,以查明其是否存在违反禁令的行为。
- Peter Thiel 移居阿根廷
最近热衷于研究敌基督的 Palantir 联合创始人 Peter Thiel 移居阿根廷。他在布宜诺斯艾利斯富人区购置了一座价值 1200 万美元的新豪宅,据报道计划长期定居,他与其丈夫 Matt Danzeisen 本周在玫瑰宫(Casa Rosada)会见了阿根廷总统 Javier Milei,当时在场的还有阿外交部长 Pablo Quirno。Thiel 等科技亿万富翁被认为是在寻找避险之地,因为他们相信随着科技进步造成广泛的经济和政治混乱,美国等国将经历一场崩溃,而新西兰以及阿根廷都被视为是极佳的候选避险之地,这些地方比较偏远,但有着较高的公共卫生水平,以及低成本的食品产地。他已经在新西兰获得了公民身份,拥有欧洲地区的马耳他护照。
- 特朗普解雇了美国国家科学委员会所有 24 名成员
特朗普解雇了美国国家科学委员会所有 24 名成员。美国国家科学委员会成立于 1950 年,其成员由总统任命,他们通常是杰出学者和行业领袖,任期六年,每两年改选八名成员。他们的工作包括就国家科学政策向政府和国会提供建议,监督美国国家科学基金会(NSF)的运作。他们本身不负责拨款,而是就国家科研优先事项提供建议。他们被解雇被认为是在 2025 年 5 月发表联署公开信,反对削减国家科学基金会的预算,从而激怒了特朗普政府。
- AI 是一种压迫性技术
Ali Alkhatib 认为 AI 是一个政治项目,其意图是将权力和能动性从个人和组织转移至中心化的权力结构。这些权力结构目前主要集中在少数科技巨头以及它们投入巨资的 AI 实验室手中。现代 AI 系统脱胎于暴力也是基于暴力:其第一种暴力形式是数据获取的暴力,AI 公司的爬虫无视任何规则抓取数据;第二种暴力形式发生在数据标注和清洗过程中;第三种暴力形式则是其数据集带有殖民主义的西方视角;第四种是利用 AI 工具对边缘群体施加暴力,比如 grok 的比基尼改图。AI 被认为能提高生产力,但实际过程中它却是被用于压榨员工。企业通过指出 AI 可取代员工迫使员工更努力工作,而不是要求加薪或改善工作条件,这种威胁削弱了员工的个人力量,让员工认为自己没有价值。自 ChatGPT 流行以来,制作虚假信息和操纵叙事比以往任何时候都认为,破坏了人类建立起来的可靠生产、验证和传播可信信息的基础设施,如果没有对 AI 输出的验证,使用者将是用信仰取代可信度。
- 座头鲸数量恢复,形成大规模超群
20 世纪的大规模工业捕鲸几乎导致座头鲸灭绝,海洋中剩下的鲸鱼不足工业捕鲸前数量的 5%。40 年前一项全球捕鲸禁令生效,座头鲸种群开始恢复。尽管部分座头鲸种群仍然濒临灭绝,但总体数量正在上升。越来越多地方报告目击到座头鲸的“超群(super-groups)”——指的是 20 头或以上的鲸鱼紧密聚集在一起。座头鲸生活在世界各大洋中,每年都会进行地球哺乳动物中最壮观的迁徙,行程可达 8000 公里,从温暖的繁殖地迁徙到寒冷的觅食地。在此过程中它们将大量的营养物质输送到世界各地,对海洋生态系统的健康至关重要。从2015 年到 2020 年,南非西海岸座头鲸超群目击次数从每年 10 次飙升至 65 次。2025 年 12 月 29 日,两位摄影师在南非西海岸一天内拍摄到了 208 头座头鲸,第二天更是达到了 304 头,这是历史上单日观测到的大型鲸鱼数量最多的一次。
- 三星移动业务可能首次出现亏损
由于内存成本上涨、竞争加剧以及折叠屏手机和智能手表等产品面临的压力,三星移动部门可能在 2026 年出现首次年度亏损。移动业务一直是三星的重要支柱,该业务可能亏损对公司的整体业绩构成了严重威胁。如果预测成真,这将是三星移动业务成立至今首次出现年度亏损。