Weekly Digest — 2026-W18
163 unique stories (2026-04-27 → 2026-05-03), aggregated across 8 sources.
Hacker News(30)
- Is my blue your blue? (ismy.blue)
- China blocks Meta's acquisition of AI startup Manus (www.cnbc.com)
- GitHub is having issues now (www.githubstatus.com)
- GitHub Copilot is moving to usage-based billing (github.blog)
- US Supreme Court reviews police use of cell location data (www.nytimes.com)
- Dutch central bank ditches AWS and chooses Lidl for European Cloud (www.techzine.eu)
- Ghostty is leaving GitHub (mitchellh.com)
- Waymo in Portland (waymo.com)
- Claude.ai unavailable and elevated errors on the API (status.claude.com)
- Anthropic Joins the Blender Development Fund as Corporate Patron (www.blender.org)
- Google and Pentagon reportedly agree on deal for 'any lawful' use of AI (www.theverge.com)
- Your phone is about to stop being yours (keepandroidopen.org)
GitHub Trending(17)
Product Hunt(30)
- Brew Finder
Discover the best coffee shops to work at around you
- GitBar
Every pull request, one menubar. GitHub, GitLab & Azure
- Wafaa.io
Create secure digital contracts in minutes
- Jet AI Agents
Build business AI agents in minutes
- Logic
Build and operate fleets of agents
- PlayJoob
turns dead task boards into a shared strategy map
- Kinhub
Scalable coaching that drives real business impact
- Flitch
Turn your data into insights
- Doza Assist
Open-source local AI that learns how you edit video
- Clera
An AI agent matching candidates to the right roles.
- SureThing.io
Autonomous agent that communicates results like a human
- Happy Horse
Top-tier AI video generation and editing from Alibaba ATH
Hugging Face(30)
- Agentic World Modeling: Foundations, Capabilities, Laws, and Beyond
As AI systems move from generating text to accomplishing goals through sustained interaction, the ability to model environment dynamics becomes a central bottleneck. Agents that manipulate objects, navigate software, coordinate with others, or design experiments require predictive environment models, yet the term world model carries different meanings across research communities. We introduce a "levels x laws" taxonomy organized along two axes. The first defines three capability levels: L1 Predictor, which learns one-step local transition operators; L2 Simulator, which composes them into multi-step, action-conditioned rollouts that respect domain laws; and L3 Evolver, which autonomously revises its own model when predictions fail against new evidence. The second identifies four governing-law regimes: physical, digital, social, and scientific. These regimes determine what constraints a world model must satisfy and where it is most likely to fail. Using this framework, we synthesize over 400 works and summarize more than 100 representative systems spanning model-based reinforcement learning, video generation, web and GUI agents, multi-agent social simulation, and AI-driven scientific discovery. We analyze methods, failure modes, and evaluation practices across level-regime pairs, propose decision-centric evaluation principles and a minimal reproducible evaluation package, and outline architectural guidance, open problems, and governance challenges. The resulting roadmap connects previously isolated communities and charts a path from passive next-step prediction toward world models that can simulate, and ultimately reshape, the environments in which agents operate.
- Video Analysis and Generation via a Semantic Progress Function
Transformations produced by image and video generation models often evolve in a highly non-linear manner: long stretches where the content barely changes are followed by sudden, abrupt semantic jumps. To analyze and correct this behavior, we introduce a Semantic Progress Function, a one-dimensional representation that captures how the meaning of a given sequence evolves over time. For each frame, we compute distances between semantic embeddings and fit a smooth curve that reflects the cumulative semantic shift across the sequence. Departures of this curve from a straight line reveal uneven semantic pacing. Building on this insight, we propose a semantic linearization procedure that reparameterizes (or retimes) the sequence so that semantic change unfolds at a constant rate, yielding smoother and more coherent transitions. Beyond linearization, our framework provides a model-agnostic foundation for identifying temporal irregularities, comparing semantic pacing across different generators, and steering both generated and real-world video sequences toward arbitrary target pacing.
- DiffNR: Diffusion-Enhanced Neural Representation Optimization for Sparse-View 3D Tomographic Reconstruction
Neural representations (NRs), such as neural fields and 3D Gaussians, effectively model volumetric data in computed tomography (CT) but suffer from severe artifacts under sparse-view settings. To address this, we propose DiffNR, a novel framework that enhances NR optimization with diffusion priors. At its core is SliceFixer, a single-step diffusion model designed to correct artifacts in degraded slices. We integrate specialized conditioning layers into the network and develop tailored data curation strategies to support model finetuning. During reconstruction, SliceFixer periodically generates pseudo-reference volumes, providing auxiliary 3D perceptual supervision to fix underconstrained regions. Compared to prior methods that embed CT solvers into time-consuming iterative denoising, our repair-and-augment strategy avoids frequent diffusion model queries, leading to better runtime performance. Extensive experiments show that DiffNR improves PSNR by 3.99 dB on average, generalizes well across domains, and maintains efficient optimization.
- LLM Safety From Within: Detecting Harmful Content with Internal Representations
Guard models are widely used to detect harmful content in user prompts and LLM responses. However, state-of-the-art guard models rely solely on terminal-layer representations and overlook the rich safety-relevant features distributed across internal layers. We present SIREN, a lightweight guard model that harnesses these internal features. By identifying safety neurons via linear probing and combining them through an adaptive layer-weighted strategy, SIREN builds a harmfulness detector from LLM internals without modifying the underlying model. Our comprehensive evaluation shows that SIREN substantially outperforms state-of-the-art open-source guard models across multiple benchmarks while using 250 times fewer trainable parameters. Moreover, SIREN exhibits superior generalization to unseen benchmarks, naturally enables real-time streaming detection, and significantly improves inference efficiency compared to generative guard models. Overall, our results highlight LLM internal states as a promising foundation for practical, high-performance harmfulness detection.
- FlowAnchor: Stabilizing the Editing Signal for Inversion-Free Video Editing
We propose FlowAnchor, a training-free framework for stable and efficient inversion-free, flow-based video editing. Inversion-free editing methods have recently shown impressive efficiency and structure preservation in images by directly steering the sampling trajectory with an editing signal. However, extending this paradigm to videos remains challenging, often failing in multi-object scenes or with increased frame counts. We identify the root cause as the instability of the editing signal in high-dimensional video latent spaces, which arises from imprecise spatial localization and length-induced magnitude attenuation. To overcome this challenge, FlowAnchor explicitly anchors both where to edit and how strongly to edit. It introduces Spatial-aware Attention Refinement, which enforces consistent alignment between textual guidance and spatial regions, and Adaptive Magnitude Modulation, which adaptively preserves sufficient editing strength. Together, these mechanisms stabilize the editing signal and guide the flow-based evolution toward the desired target distribution. Extensive experiments demonstrate that FlowAnchor achieves more faithful, temporally coherent, and computationally efficient video editing across challenging multi-object and fast-motion scenarios. The project page is available at https://cuc-mipg.github.io/FlowAnchor.github.io/.
- AgentSearchBench: A Benchmark for AI Agent Search in the Wild
The rapid growth of AI agent ecosystems is transforming how complex tasks are delegated and executed, creating a new challenge of identifying suitable agents for a given task. Unlike traditional tools, agent capabilities are often compositional and execution-dependent, making them difficult to assess from textual descriptions alone. However, existing research and benchmarks typically assume well-specified functionalities, controlled candidate pools, or only executable task queries, leaving realistic agent search scenarios insufficiently studied. We introduce AgentSearchBench, a large-scale benchmark for agent search in the wild, built from nearly 10,000 real-world agents across multiple providers. The benchmark formalizes agent search as retrieval and reranking problems under both executable task queries and high-level task descriptions, and evaluates relevance using execution-grounded performance signals. Experiments reveal a consistent gap between semantic similarity and actual agent performance, exposing the limitations of description-based retrieval and reranking methods. We further show that lightweight behavioral signals, including execution-aware probing, can substantially improve ranking quality, highlighting the importance of incorporating execution signals into agent discovery. Our code is available at https://github.com/Bingo-W/AgentSearchBench.
- From Skills to Talent: Organising Heterogeneous Agents as a Real-World Company
Individual agent capabilities have advanced rapidly through modular skills and tool integrations, yet multi-agent systems remain constrained by fixed team structures, tightly coupled coordination logic, and session-bound learning. We argue that this reflects a deeper absence: a principled organisational layer that governs how a workforce of agents is assembled, governed, and improved over time, decoupled from what individual agents know. To fill this gap, we introduce OneManCompany (OMC), a framework that elevates multi-agent systems to the organisational level. OMC encapsulates skills, tools, and runtime configurations into portable agent identities called Talents, orchestrated through typed organisational interfaces that abstract over heterogeneous backends. A community-driven Talent Market enables on-demand recruitment, allowing the organisation to close capability gaps and reconfigure itself dynamically during execution. Organisational decision-making is operationalised through an Explore-Execute-Review (E^2R) tree search, which unifies planning, execution, and evaluation in a single hierarchical loop: tasks are decomposed top-down into accountable units and execution outcomes are aggregated bottom-up to drive systematic review and refinement. This loop provides formal guarantees on termination and deadlock freedom while mirroring the feedback mechanisms of human enterprises. Together, these contributions transform multi-agent systems from static, pre-configured pipelines into self-organising and self-improving AI organisations capable of adapting to open-ended tasks across diverse domains. Empirical evaluation on PRDBench shows that OMC achieves an 84.67% success rate, surpassing the state of the art by 15.48 percentage points, with cross-domain case studies further demonstrating its generality.
- World-R1: Reinforcing 3D Constraints for Text-to-Video Generation
Recent video foundation models demonstrate impressive visual synthesis but frequently suffer from geometric inconsistencies. While existing methods attempt to inject 3D priors via architectural modifications, they often incur high computational costs and limit scalability. We propose World-R1, a framework that aligns video generation with 3D constraints through reinforcement learning. To facilitate this alignment, we introduce a specialized pure text dataset tailored for world simulation. Utilizing Flow-GRPO, we optimize the model using feedback from pre-trained 3D foundation models and vision-language models to enforce structural coherence without altering the underlying architecture. We further employ a periodic decoupled training strategy to balance rigid geometric consistency with dynamic scene fluidity. Extensive evaluations reveal that our approach significantly enhances 3D consistency while preserving the original visual quality of the foundation model, effectively bridging the gap between video generation and scalable world simulation.
- ReVSI: Rebuilding Visual Spatial Intelligence Evaluation for Accurate Assessment of VLM 3D Reasoning
Current evaluations of spatial intelligence can be systematically invalid under modern vision-language model (VLM) settings. First, many benchmarks derive question-answer (QA) pairs from point-cloud-based 3D annotations originally curated for traditional 3D perception. When such annotations are treated as ground truth for video-based evaluation, reconstruction and annotation artifacts can miss objects that are clearly visible in the video, mislabel object identities, or corrupt geometry-dependent answers (e.g., size), yielding incorrect or ambiguous QA pairs. Second, evaluations often assume full-scene access, while many VLMs operate on sparsely sampled frames (e.g., 16-64), making many questions effectively unanswerable under the actual model inputs. We improve evaluation validity by introducing ReVSI, a benchmark and protocol that ensures each QA pair is answerable and correct under the model's actual inputs. To this end, we re-annotate objects and geometry across 381 scenes from 5 datasets to improve data quality, and regenerate all QA pairs with rigorous bias mitigation and human verification using professional 3D annotation tools. We further enhance evaluation controllability by providing variants across multiple frame budgets (16/32/64/all) and fine-grained object visibility metadata, enabling controlled diagnostic analyses. Evaluations of general and domain-specific VLMs on ReVSI reveal systematic failure modes that are obscured by prior benchmarks, yielding a more reliable and diagnostic assessment of spatial intelligence.
- Vision-Language-Action Safety: Threats, Challenges, Evaluations, and Mechanisms
Vision-Language-Action (VLA) models are emerging as a unified substrate for embodied intelligence. This shift raises a new class of safety challenges, stemming from the embodied nature of VLA systems, including irreversible physical consequences, a multimodal attack surface across vision, language, and state, real-time latency constraints on defense, error propagation over long-horizon trajectories, and vulnerabilities in the data supply chain. Yet the literature remains fragmented across robotic learning, adversarial machine learning, AI alignment, and autonomous systems safety. This survey provides a unified and up-to-date overview of safety in Vision-Language-Action models. We organize the field along two parallel timing axes, attack timing (training-time vs. inference-time and defense timing (training-time vs. inference-time, linking each class of threat to the stage at which it can be mitigated. We first define the scope of VLA safety, distinguishing it from text-only LLM safety and classical robotic safety, and review the foundations of VLA models, including architectures, training paradigms, and inference mechanisms. We then examine the literature through four lenses: Attacks, Defenses, Evaluation, and Deployment. We survey training-time threats such as data poisoning and backdoors, as well as inference-time attacks including adversarial patches, cross-modal perturbations, semantic jailbreaks, and freezing attacks. We review training-time and runtime defenses, analyze existing benchmarks and metrics, and discuss safety challenges across six deployment domains. Finally, we highlight key open problems, including certified robustness for embodied trajectories, physically realizable defenses, safety-aware training, unified runtime safety architectures, and standardized evaluation.
- Tuna-2: Pixel Embeddings Beat Vision Encoders for Multimodal Understanding and Generation
Unified multimodal models typically rely on pretrained vision encoders and use separate visual representations for understanding and generation, creating misalignment between the two tasks and preventing fully end-to-end optimization from raw pixels. We introduce Tuna-2, a native unified multimodal model that performs visual understanding and generation directly based on pixel embeddings. Tuna-2 drastically simplifies the model architecture by employing simple patch embedding layers to encode visual input, completely discarding the modular vision encoder designs such as the VAE or the representation encoder. Experiments show that Tuna-2 achieves state-of-the-art performance in multimodal benchmarks, demonstrating that unified pixel-space modelling can fully compete with latent-space approaches for high-quality image generation. Moreover, while the encoder-based variant converges faster in early pretraining, Tuna-2's encoder-free design achieves stronger multimodal understanding at scale, particularly on tasks requiring fine-grained visual perception. These results show that pretrained vision encoders are not necessary for multimodal modelling, and end-to-end pixel-space learning offers a scalable path toward stronger visual representations for both generation and perception.
- ClawMark: A Living-World Benchmark for Multi-Turn, Multi-Day, Multimodal Coworker Agents
Language-model agents are increasingly used as persistent coworkers that assist users across multiple working days. During such workflows, the surrounding environment may change independently of the agent: new emails arrive, calendar entries shift, knowledge-base records are updated, and evidence appears across images, scanned PDFs, audio, video, and spreadsheets. Existing benchmarks do not adequately evaluate this setting because they typically run within a single static episode and remain largely text-centric. We introduce , a benchmark for coworker agents built around multi-turn multi-day tasks, a stateful sandboxed service environment whose state evolves between turns, and rule-based verification. The current release contains 100 tasks across 13 professional scenarios, executed against five stateful sandboxed services (filesystem, email, calendar, knowledge base, spreadsheet) and scored by 1537 deterministic Python checkers over post-execution service state; no LLM-as-judge is invoked during scoring. We benchmark seven frontier agent systems. The strongest model reaches 75.8 weighted score, but the best strict Task Success is only 20.0\%, indicating that partial progress is common while complete end-to-end workflow completion remains rare. Turn-level analysis shows that performance drops after the first exogenous environment update, highlighting adaptation to changing state as a key open challenge. We release the benchmark, evaluation harness, and construction pipeline to support reproducible coworker-agent evaluation.
Techmeme(30)
- Renders based on photos of Samsung's upcoming smart glasses, expected to launch later this year, show a design nearly identical to Ray-Ban Meta glasses (Alexander Maxham/Android Headlines)
Alexander Maxham / Android Headlines : Renders based on photos of Samsung's upcoming smart glasses, expected to launch later this year, show a design nearly identical to Ray-Ban Meta glasses — Add Android Headlines as a preferred source on Google — Samsung's next Android XR product is reportedly a pair of smart glasses codenamed …
- Study: only ~3% of Polymarket accounts drove most price discovery in 2023-2025, suggesting market accuracy comes from an informed minority, not crowd wisdom (Sam Reynolds/CoinDesk)
Sam Reynolds / CoinDesk : Study: only ~3% of Polymarket accounts drove most price discovery in 2023-2025, suggesting market accuracy comes from an informed minority, not crowd wisdom — Researchers show market accuracy comes from a tiny group of informed traders, not broad participation. … What to know:
- Elon Musk boosts an X post by Ronan Farrow promoting his New Yorker article on Sam Altman's alleged deceptions, as Musk's lawsuit against OpenAI heads to trial (Wired)
Wired : Elon Musk boosts an X post by Ronan Farrow promoting his New Yorker article on Sam Altman's alleged deceptions, as Musk's lawsuit against OpenAI heads to trial — The move comes as the trial for Elon Musk's lawsuit against OpenAI kicks off in federal court in Oakland.
- Jury selection begins in Musk v. Altman trial at a federal courthouse in California, with Sam Altman and Greg Brockman in attendance (Ashley Capoot/CNBC)
Ashley Capoot / CNBC : Jury selection begins in Musk v. Altman trial at a federal courthouse in California, with Sam Altman and Greg Brockman in attendance — The nine-person jury was seated on Monday in the high-stakes legal battle between longtime friends turned rivals Elon Musk and Sam Altman at a federal courthouse in Oakland, California.
- Have I Been Pwned: ShinyHunters' breach of ADT exposed the personal data of 5.5M people; ADT previously disclosed data breaches in August 2024 and October 2024 (Sergiu Gatlan/BleepingComputer)
Sergiu Gatlan / BleepingComputer : Have I Been Pwned: ShinyHunters' breach of ADT exposed the personal data of 5.5M people; ADT previously disclosed data breaches in August 2024 and October 2024 — The ShinyHunters extortion group stole the personal information of 5.5 million individuals after breaching the systems …
- GitHub says all Copilot plans will move to usage-based billing on June 1, replacing premium requests with monthly GitHub AI Credits (Mario Rodriguez/The GitHub Blog)
Mario Rodriguez / The GitHub Blog : GitHub says all Copilot plans will move to usage-based billing on June 1, replacing premium requests with monthly GitHub AI Credits — Starting June 1, your Copilot usage will consume GitHub AI Credits. — TL;DR: Today, we are announcing that all GitHub Copilot plans will transition to usage-based billing on June 1, 2026.
- AWS launches a desktop app for its Amazon Quick AI assistant, letting users connect their tools and local files to build custom apps, live dashboards, and more (Jigar Thakkar/About Amazon)
Jigar Thakkar / About Amazon : AWS launches a desktop app for its Amazon Quick AI assistant, letting users connect their tools and local files to build custom apps, live dashboards, and more — Amazon Quick brings a personal AI assistant to your desktop. Build presentations, intelligent dashboards, and more.
- Musk v. Altman: Musk testifies he's suing OpenAI because "it is not okay to steal a charity" and its pivot sets a concerning precedent for philanthropic efforts (Bloomberg)
Bloomberg : Musk v. Altman: Musk testifies he's suing OpenAI because “it is not okay to steal a charity” and its pivot sets a concerning precedent for philanthropic efforts — Arguments Begin in Musk, Altman Showdown — Video Player is loading. — Unmute — Current Time 0:00 Loaded: 17.06% Playback Rate
- NXP reports Q1 revenue up 12% YoY to $3.18B, vs. $3.15B est., and forecasts Q2 revenue above estimates; NXPI jumps 13%+ after hours (Christina Kyriasoglou/Bloomberg)
Christina Kyriasoglou / Bloomberg : NXP reports Q1 revenue up 12% YoY to $3.18B, vs. $3.15B est., and forecasts Q2 revenue above estimates; NXPI jumps 13%+ after hours — NXP Semiconductors NV jumped in late trading after giving an upbeat revenue forecast, a sign the chipmaker is bouncing back from a prolonged auto industry slump and tariff uncertainties.
- Sources: the US Commerce Department last week ordered multiple chip equipment companies to halt some shipments to China's second-largest chipmaker, Hua Hong (Karen Freifeld/Reuters)
Karen Freifeld / Reuters : Sources: the US Commerce Department last week ordered multiple chip equipment companies to halt some shipments to China's second-largest chipmaker, Hua Hong — The U.S. Department of Commerce last week ordered multiple chip equipment companies to halt certain tool shipments …
- Robinhood reports Q1 revenue up 15% YoY to $1.07B, vs. $1.14B est., and crypto revenue down 47% to $134M, vs. $147.6M est.; HOOD drops 6%+ after hours (Luke Kawa/Sherwood News)
Luke Kawa / Sherwood News : Robinhood reports Q1 revenue up 15% YoY to $1.07B, vs. $1.14B est., and crypto revenue down 47% to $134M, vs. $147.6M est.; HOOD drops 6%+ after hours — The brokerage just reported quarterly results. — Robinhood MarketsHOOD $77.11 (-2.29%) is sharply lower in postmarket trading after reporting underwhelming Q1 results:
- Seagate reports Q3 revenue up 44% YoY to $3.11B, vs. $2.96B est., and forecasts Q4 revenue and adjusted EPS above estimates; STX jumps 13%+ after hours (Zaheer Kachwala/Reuters)
Zaheer Kachwala / Reuters : Seagate reports Q3 revenue up 44% YoY to $3.11B, vs. $2.96B est., and forecasts Q4 revenue and adjusted EPS above estimates; STX jumps 13%+ after hours — Seagate Technology (STX.O) forecast fourth-quarter revenue and profit above Wall Street expectations on Tuesday, betting on strong demand …
Solidot(26)
- 老房子闹鬼可能源于陈旧设施产生的次声波
觉得老房子闹鬼?你可能是受到了陈旧设施如旧管道和旧锅炉产生的次声波的影响。根据发表在《Frontiers in Behavioural Neuroscience》期刊上的一项研究,研究人员让 36 名志愿者听轻音乐或鬼屋景点播放的那种令人心神不宁的音乐。在参与者不知情下,研究人员悄悄在半数情况下播放了次声波。结果显示,次声波让志愿者感到更烦躁和恼怒,觉得音乐更悲伤,且唾液中的皮质醇水平更高。研究人员称,人耳听不到次声波,但身体和情绪仍然能做出反应,且通常是不愉快的反应。《The Science of Weird Shit: Why Our Minds Conjure the Paranormal》一书的作者 Chris French 教授认为用次声波解释闹鬼有点牵强。
- 欧洲批准了 Moderna 的流感和 COVID-19 联合疫苗
欧洲批准了 Moderna 研发的基于 mRNA 技术的流感和 COVID-19 联合疫苗。被称为 mRNA-1083 或 mCOMBRIAX 的疫苗成为全球首个获得批准的针对这两种呼吸道病毒的联合疫苗。疫苗获批是基于一项 4000 名成年人参与的 III 期临床试验结果。试验分为两组,一组为 50-64 岁的受试者,与标准流感疫苗进行比较;另一组为 65 岁及以上的受试者,与高剂量流感疫苗进行比较。两组受试者中,相对于对照组 mCOMBRIAX 疫苗都能诱导对常见流感病毒株(A/H1N1、A/H3N2 和 B/Victoria)以及 SARS-CoV-2 病毒产生统计上显著更高的免疫反应。试验未发现安全性或不良反应方面的问题。
- 杀虫剂导致北美蝴蝶数量大减
2025 年 3 月科学家在《科学》期刊上发表研究,Xerces Society for Invertebrate Conservation 保护协会随后发表了蝴蝶现状报告。研究发现,从 2000 年到 2020 年全美蝴蝶总数下降了 22%,有 24 种蝴蝶数量下降 90% 或以上。杀虫剂被认为是导致这一结果的主要原因。1960 年代化学公司研制出了强效杀虫剂滴滴涕(DDT),公众对滴滴涕的反对促使企业研制出弱化对人类伤害但强化对昆虫杀伤力的新杀虫剂。多种混合型杀虫剂的使用导致蝴蝶等昆虫在 21 世纪加速减少。生态学家 Matt Forister 等人在《Environmental Toxicology and Chemistry》期刊上报告,他们分析了 336 株植物只有 22 株植物没有检测到农药残留。这些植物至少含有三种化学物质,其中 71 株植物的农药浓度对蝴蝶而言是致命或接近致命。在 2022 年的一项类似研究中,Forister 团队分析了 33 家苗圃出售的 235 株乳草(对帝王蝶至关重要的植物),发现每株植物平均含有 12.2 种杀虫剂。
- 发改委要求撤销对 Manus 的收购
国家发展改革委周一发布通报,外商投资安全审查工作机制办公室(国家发改委)依法依规对外资收购 Manus 项目作出禁止投资决定,要求当事人撤销该收购交易。《外商投资安全审查办法》于 2020 年 12 月 19 日由国家发展改革委、商务部联合发布,自 2021 年 1 月 18 日开始施行,对适用审查的外商投资类型、审查机构、审查范围、审查程序、审查决定监督执行和违规处理等进行规定。跟据该文件,国家建立外商投资安全审查工作机制,工作机制办公室设在国家发展改革委,由国家发展改革委、商务部牵头,承担外商投资安全审查的日常工作。
- Greg K-H 使用基于 AMD Ryzen AI Max 的 AI 工具发现内核 Bug
稳定版内核维护者 Greg Kroah-Hartman 正在使用名为 gkh_clanker_t1000 的 AI 工具发现内核 Bug。他在 Mastodon 上分享了 gkh_clanker_t1000 的硬件图片。gkh_clanker_t1000 运行在搭载 AMD Ryzen AI Max+“Strix Halo”APU 的 Framework Desktop 之上,Ryzen AI Max+ 提供了最高 128GB 的统一内存,可分配 96GB 内存给 GPU 使用,其性能足以运行本地大模型以及其它基于开源软件栈的 AI 工具。Greg K-H 尚未透露 gkh_clanker_t1000 软件方面的信息。
- AI 成本可能高于人工成本
多家企业在 AI 上的支出已经超过了员工薪资,IT 预算严重超支。Uber CTO 的 2026 年 AI 预算因 token 费用超支。根据 Gartner 预测,2026 年全球 IT 支出预计将达到 6.31 万亿美元,比 2025 年增长 13.5%。这一增长是由 AI 基础设施、软件和云服务的“持续发展势头”推动的。即使是 IT 预算最充足的公司也需要证明 AI 投入的长期回报。当 AI 实验室提高价格时,对 AI 的大量投入可能会从一种炫耀的资本变成一种负担。
- 调查显示对接种疫苗犹豫的人更可能阅读新右派新闻
2025 年美国有 43 个州报告了逾 2000 例麻疹病例,几乎所有病例都发生在未接种疫苗人群中。2026 年的麻疹病例数仍在增加。美国学龄儿童的麻风腮三联疫苗(MMR)的接种率持续下降,徘徊在 93% 左右,低于 95% 的群体免疫阈值。研究人员调查了 2970 名成年人,虽然大多数美国人(83%)表示 MMR 疫苗的好处大于风险,但大约六分之一受访者表示对接种疫苗犹豫。犹豫的成年人总体更年轻,62% 的人年龄在 44 岁以下,且更有可能为人父母。他们更有可能是少数族裔、低收入和受教育程度较低的人。他们表达了更保守的政治信念,且更可能认同共和党(39%)或独立派(33%)。犹豫的成年人还更有可能认同“让美国再次健康”运动(MAHA),其比例为 43%,而非犹豫成年人占 27%。疫苗接种犹豫和非犹豫者之间的最大差异是前者偏爱阅读新右派新闻如 Breitbart、Newsmax 和 Zero Hedge。
- 尼安德特人和现代人类大脑之间主要是外观上的差异
尼安德特人和现代智人大脑的头骨在外形上存在明显的差异:尼安德特人的头骨更扁更长,而现代人类更圆。根据发表在 PNAS 期刊上的一项研究《Neanderthal brain and cognition reconsidered》,外形上的差异并不能说明什么。对比 400 名(200 欧裔 200 汉族)现代人类大脑的 MRI 扫描和尼安德特人头骨内模,现代人类大脑之间的差异比更新世智人和和尼安德特人之间的差异更大。鉴于大脑大小并不能准确预测认知能力,尼安德特人的认知能力可能更接近现代人类,这意味着人类可能不是凭借更聪明或更强的适应能力而战胜尼安德特人。研究人员认为,尼安德特人大脑和认知能力差异,完全可纳入现代人类内部的差异。
- 肥胖的记忆会长时间留在免疫系统中
根据发表在《EMBO Reports》期刊上一项历时十年的研究,被称为辅助 T 细胞(helper T cells)的免疫细胞会长期携带肥胖记忆。记忆通过 DNA 甲基化过程标记在免疫细胞的 DNA 上,在成功减肥之后仍然会持续 5-10 年。这意味着减肥者仍然会长期面临肥胖相关疾病风险。研究人员称,短期减肥可能无法立即降低肥胖相关的疾病风险,需要在几年时间里维持减肥后的体重才可能逆转肥胖对 T 细胞的影响。
- Mercor 4TB 语音样本被盗
Mercor 是美国一家 AI 初创公司,主要业务是为其他 AI 公司提供专家帮助训练模型和聊天机器人。它招聘专家/合同工时要求对方提供护照或驾照扫描件、自拍和录制一段语音。本月初勒索组织 Lapsus$ 披露它从 Mercor 窃取了 4TB 语音样本。语音样本加上身份证件,引发了合同工们身份被盗用的担忧。已有至少五名合同工对 Mercor 提起了诉讼,指控该公司以训练数据的名义收集语音特征,但并未明确说明这些特征是永久性的生物识别标识符。现有的语音克隆技术只需要 15 秒钟的清晰参照音频,而 Mercor 要求合同工录制的语音长度达到了 2-5 分钟,足够实现语音克隆。
- 三星家电业务计划退出中国
韩国三星电子计划 2026 年内撤出在中国的家电和电视销售。由于中国企业的崛起,该业务的收益持续低迷。将把该业务的重心转向表现强劲的美国。三星电子最快将在 4月 底最终决定停止在中国国内的家电和电视销售,将开始向当地员工和合作伙伴展开说明,将逐步处理中国国内的库存,2026 年内完全结束销售。该公司将把经营资源转向半导体和智能手机的销售。三星在中国生产冰箱和洗衣机等产品,将维持这些产品的生产体制。在结束向中国国内出货的同时,转为面向海外的供应基地。
- 网易运营的《暗黑4》国服限时免费
网易运营的《暗黑破坏神 IV》国服限时一个月免费,领取需要实名——网易邮箱账号或手机号码。《暗黑破坏神 IV》于 2023 年 6 月 5 日发布,故事时间为《暗黑破坏神 III》的 50 年后,玩家可以自由选择游玩每个区域的顺序,在开放世界中体验旅途。本作中共有五张地图:黑暗森林索格伦、冰天山地裂峰山、无边沙漠干燥平原、毒雾沼泽哈维萨和堡垒废墟凯基斯坦。国服是在 2025 年 12 月 12 日正式上线,免费的是基础版本,扩展内容需要付费购买。