DIGEST · 2026-04-24

OrangeBot.AI Digest — 2026-04-24

87 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Google to invest up to $40B in Anthropic in cash and compute (techcrunch.com)
  2. OpenAI releases GPT-5.5 and GPT-5.5 Pro in the API (developers.openai.com)
  3. SDL Now Supports DOS (github.com)
  4. I cancelled Claude: Token issues, declining quality, and poor support (nickyreinert.de)
  5. Norway set to become latest country to ban social media for under 16s (www.bloomberg.com)
  6. Sabotaging projects by overthinking, scope creep, and structural diffing (kevinlynagh.com)
  7. Hear your agent suffer through your code (github.com)
  8. How to be anti-social – a guide to incoherent and isolating social experiences (nate.leaflet.pub)
  9. UK Biobank leak: Health details of 500k people offered for sale on Alibaba (www.bmj.com)
  10. South Korea police arrest man for posting AI photo of runaway wolf (www.bbc.com)
  11. Spinel: Ruby AOT Native Compiler (github.com)
  12. Show HN: How LLMs Work – Interactive visual guide based on Karpathy's lecture (ynarwal.github.io)
  13. Ubuntu 26.04 (lwn.net)
  14. Habitual coffee intake shapes the microbiome, modifies physiology and cognition (www.nature.com)
  15. DeepSeek v4 (api-docs.deepseek.com)

GitHub Trending(12)

  1. Alishahryar1 / free-claude-code
  2. huggingface / ml-intern
  3. google / osv-scanner
  4. Z4nzu / hackingtool
  5. zilliztech / claude-context
  6. open-metadata / OpenMetadata
  7. PostHog / posthog
  8. dani-garcia / vaultwarden
  9. Anil-matcha / Open-Generative-AI
  10. codecrafters-io / build-your-own-x
  11. deepseek-ai / DeepEP
  12. microsoft / typescript-go

Product Hunt(15)

  1. Beezi AI

    Make AI development structured, secure, and cost-efficient.

  2. Spira AI

    AI Influencer that always on trend, create & grow your brand

  3. Bansi AI by Writesonic

    AI video editor for long-form talking head videos

  4. Emotional intelligence AI for live calls

    Emotional intelligence AI for live sales calls

  5. NotchNest AI

    AI powered by Apple Intelligence now in your Notch

  6. BAND

    Coordinate and govern multi-agent work in a single chat

  7. Codex 3.0 by OpenAI

    Codex can now build, test & debug on autopilot

  8. Ask Product Hunt AI

    Find the right product, just ask

  9. LifeOS

    Turn your AI chats/memory to intros with real humans

  10. MailCue

    Run as a fully hardened production email server.

  11. Haiker

    Hacker News App for non-native english speaker

  12. Universal Gas Framework (UGF)

    Route Actions, Not Liquidity

  13. TuneJourney.com

    AI learns your listening habits and curates your live radio

  14. boots.list

    From Rekordbox collection to set-ready playlist

  15. Tyndale

    Translate your app with the AI you already pay for

Hugging Face(15)

  1. LLaTiSA: Towards Difficulty-Stratified Time Series Reasoning from Visual Perception to Semantics

    Comprehensive understanding of time series remains a significant challenge for Large Language Models (LLMs). Current research is hindered by fragmented task definitions and benchmarks with inherent ambiguities, precluding rigorous evaluation and the development of unified Time Series Reasoning Models(TSRMs). To bridge this gap, we formalize Time Series Reasoning (TSR) via a four-level taxonomy of increasing cognitive complexity. We introduce HiTSR, a hierarchical time series reasoning dataset comprising 83k samples with diverse task combinations and verified Chain-of-Thought (CoT) trajectories. Leveraging HiTSR, we propose LLaTiSA, a strong TSRM that integrates visualized patterns with precision-calibrated numerical tables to enhance the temporal perception of Vision-Language Models (VLMs). Through a multi-stage curriculum fine-tuning strategy, LLaTiSA achieves superior performance and exhibits robust out-of-distribution generalization across diverse TSR tasks and real-world scenarios. Our code is available at https://github.com/RainingNovember/LLaTiSA.

  2. WorldMark: A Unified Benchmark Suite for Interactive Video World Models

    Interactive video generation models such as Genie, YUME, HY-World, and Matrix-Game are advancing rapidly, yet every model is evaluated on its own benchmark with private scenes and trajectories, making fair cross-model comparison impossible. Existing public benchmarks offer useful metrics such as trajectory error, aesthetic scores, and VLM-based judgments, but none supplies the standardized test conditions -- identical scenes, identical action sequences, and a unified control interface -- needed to make those metrics comparable across models with heterogeneous inputs. We introduce WorldMark, the first benchmark that provides such a common playing field for interactive Image-to-Video world models. WorldMark contributes: (1) a unified action-mapping layer that translates a shared WASD-style action vocabulary into each model's native control format, enabling apples-to-apples comparison across six major models on identical scenes and trajectories; (2) a hierarchical test suite of 500 evaluation cases covering first- and third-person viewpoints, photorealistic and stylized scenes, and three difficulty tiers from Easy to Hard spanning 20-60s; and (3) a modular evaluation toolkit for Visual Quality, Control Alignment, and World Consistency, designed so that researchers can reuse our standardized inputs while plugging in their own metrics as the field evolves. We will release all data, evaluation code, and model outputs to facilitate future research. Beyond offline metrics, we launch World Model Arena (warena.ai), an online platform where anyone can pit leading world models against each other in side-by-side battles and watch the live leaderboard.

  3. UniT: Toward a Unified Physical Language for Human-to-Humanoid Policy Learning and World Modeling

    Scaling humanoid foundation models is bottlenecked by the scarcity of robotic data. While massive egocentric human data offers a scalable alternative, bridging the cross-embodiment chasm remains a fundamental challenge due to kinematic mismatches. We introduce UniT (Unified Latent Action Tokenizer via Visual Anchoring), a framework that establishes a unified physical language for human-to-humanoid transfer. Grounded in the philosophy that heterogeneous kinematics share universal visual consequences, UniT employs a tri-branch cross-reconstruction mechanism: actions predict vision to anchor kinematics to physical outcomes, while vision reconstructs actions to filter out irrelevant visual confounders. Concurrently, a fusion branch synergies these purified modalities into a shared discrete latent space of embodiment-agnostic physical intents. We validate UniT across two paradigms: 1) Policy Learning (VLA-UniT): By predicting these unified tokens, it effectively leverages diverse human data to achieve state-of-the-art data efficiency and robust out-of-distribution (OOD) generalization on both humanoid simulation benchmark and real-world deployments, notably demonstrating zero-shot task transfer. 2) World Modeling (WM-UniT): By aligning cross-embodiment dynamics via unified tokens as conditions, it realizes direct human-to-humanoid action transfer. This alignment ensures that human data seamlessly translates into enhanced action controllability for humanoid video generation. Ultimately, by inducing a highly aligned cross-embodiment representation (empirically verified by t-SNE visualizations revealing the convergence of human and humanoid features into a shared manifold), UniT offers a scalable path to distill vast human knowledge into general-purpose humanoid capabilities.

  4. StyleID: A Perception-Aware Dataset and Metric for Stylization-Agnostic Facial Identity Recognition

    Creative face stylization aims to render portraits in diverse visual idioms such as cartoons, sketches, and paintings while retaining recognizable identity. However, current identity encoders, which are typically trained and calibrated on natural photographs, exhibit severe brittleness under stylization. They often mistake changes in texture or color palette for identity drift or fail to detect geometric exaggerations. This reveals the lack of a style-agnostic framework to evaluate and supervise identity consistency across varying styles and strengths. To address this gap, we introduce StyleID, a human perception-aware dataset and evaluation framework for facial identity under stylization. StyleID comprises two datasets: (i) StyleBench-H, a benchmark that captures human same-different verification judgments across diffusion- and flow-matching-based stylization at multiple style strengths, and (ii) StyleBench-S, a supervision set derived from psychometric recognition-strength curves obtained through controlled two-alternative forced-choice (2AFC) experiments. Leveraging StyleBench-S, we fine-tune existing semantic encoders to align their similarity orderings with human perception across styles and strengths. Experiments demonstrate that our calibrated models yield significantly higher correlation with human judgments and enhanced robustness for out-of-domain, artist drawn portraits. All of our datasets, code, and pretrained models are publicly available at https://kwanyun.github.io/StyleID_page/

  5. Co-Evolving LLM Decision and Skill Bank Agents for Long-Horizon Tasks

    Long horizon interactive environments are a testbed for evaluating agents skill usage abilities. These environments demand multi step reasoning, the chaining of multiple skills over many timesteps, and robust decision making under delayed rewards and partial observability. Games are a good testbed for evaluating agent skill usage in environments. Large Language Models (LLMs) offer a promising alternative as game playing agents, but they often struggle with consistent long horizon decision making because they lack a mechanism to discover, retain, and reuse structured skills across episodes. We present COSPLAY, a co evolution framework in which an LLM decision agent retrieves skills from a learnable skill bank to guide action taking, while an agent managed skill pipeline discovers reusable skills from the agents unlabeled rollouts to form a skill bank. Our framework improves both the decision agent to learn better skill retrieval and action generation, while the skill bank agent continually extracts, refines, and updates skills together with their contracts. Experiments across six game environments show that COSPLAY with an 8B base model achieves over 25.1 percent average reward improvement against four frontier LLM baselines on single player game benchmarks while remaining competitive on multi player social reasoning games.

  6. Seeing Fast and Slow: Learning the Flow of Time in Videos

    How can we tell whether a video has been sped up or slowed down? How can we generate videos at different speeds? Although videos have been central to modern computer vision research, little attention has been paid to perceiving and controlling the passage of time. In this paper, we study time as a learnable visual concept and develop models for reasoning about and manipulating the flow of time in videos. We first exploit the multimodal cues and temporal structure naturally present in videos to learn, in a self-supervised manner, to detect speed changes and estimate playback speed. We then show that these learned temporal reasoning models enable us to curate the largest slow-motion video dataset to date from noisy in-the-wild sources. Such slow-motion footage, typically filmed by high-speed cameras, contains substantially richer temporal detail than standard videos. Using this data, we further develop models capable of temporal control, including speed-conditioned video generation, which produces motion at specified playback speed, and temporal super-resolution, which tranforms low-FPS, blurry videos into high-FPS sequences with fine-grained temporal details. Our findings highlight time as a manipulable, perceptual dimension in video learning, opening doors to temporally controllable video generation, temporal forensics detection, and potentially richer world-models that understand how events unfold over time.

  7. VLAA-GUI: Knowing When to Stop, Recover, and Search, A Modular Framework for GUI Automation

    Autonomous GUI agents face two fundamental challenges: early stopping, where agents prematurely declare success without verifiable evidence, and repetitive loops, where agents cycle through the same failing actions without recovery. We present VLAA-GUI, a modular GUI agentic framework built around three integrated components that guide the system on when to Stop, Recover, and Search. First, a mandatory Completeness Verifier enforces UI-observable success criteria and verification at every finish step -- with an agent-level verifier that cross-examines completion claims with decision rules, rejecting those lacking direct visual evidence. Second, a mandatory Loop Breaker provides multi-tier filtering: switching interaction mode after repeated failures, forcing strategy changes after persistent screen-state recurrence, and binding reflection signals to strategy shifts. Third, an on-demand Search Agent searches online for unfamiliar workflows by directly querying a capable LLM with search ability, returning results as plain text. We additionally integrate a Coding Agent for code-intensive actions and a Grounding Agent for precise action grounding, both invoked on demand when required. We evaluate VLAA-GUI across five top-tier backbones, including Opus 4.5, 4.6 and Gemini 3.1 Pro, on two benchmarks with Linux and Windows tasks, achieving top performance on both (77.5% on OSWorld and 61.0% on WindowsAgentArena). Notably, three of the five backbones surpass human performance (72.4%) on OSWorld in a single pass. Ablation studies show that all three proposed components consistently improve a strong backbone, while a weaker backbone benefits more from these tools when the step budget is sufficient. Further analysis also shows that the Loop Breaker nearly halves wasted steps for loop-prone models.

  8. Hybrid Policy Distillation for LLMs

    Knowledge distillation (KD) is a powerful paradigm for compressing large language models (LLMs), whose effectiveness depends on intertwined choices of divergence direction, optimization strategy, and data regime. We break down the design of existing KD methods and present a unified view that establishes connections between them, reformulating KD as a reweighted log-likelihood objective at the token level. We further propose Hybrid Policy Distillation (HPD), which integrates the complementary advantages of forward and reverse KL to balance mode coverage and mode-seeking, and combines off-policy data with lightweight, approximate on-policy sampling. We validate HPD on long-generation math reasoning as well as short-generation dialogue and code tasks, demonstrating improved optimization stability, computational efficiency, and final performance across diverse model families and scales. The code related to this work is available at https://github.com/zwhong714/Hybrid-Policy-Distillation.

  9. TingIS: Real-time Risk Event Discovery from Noisy Customer Incidents at Enterprise Scale

    Real-time detection and mitigation of technical anomalies are critical for large-scale cloud-native services, where even minutes of downtime can result in massive financial losses and diminished user trust. While customer incidents serve as a vital signal for discovering risks missed by monitoring, extracting actionable intelligence from this data remains challenging due to extreme noise, high throughput, and semantic complexity of diverse business lines. In this paper, we present TingIS, an end-to-end system designed for enterprise-grade incident discovery. At the core of TingIS is a multi-stage event linking engine that synergizes efficient indexing techniques with Large Language Models (LLMs) to make informed decisions on event merging, enabling the stable extraction of actionable incidents from just a handful of diverse user descriptions. This engine is complemented by a cascaded routing mechanism for precise business attribution and a multi-dimensional noise reduction pipeline that integrates domain knowledge, statistical patterns, and behavioral filtering. Deployed in a production environment handling a peak throughput of over 2,000 messages per minute and 300,000 messages per day, TingIS achieves a P90 alert latency of 3.5 minutes and a 95\% discovery rate for high-priority incidents. Benchmarks constructed from real-world data demonstrate that TingIS significantly outperforms baseline methods in routing accuracy, clustering quality, and Signal-to-Noise Ratio.

  10. EditCrafter: Tuning-free High-Resolution Image Editing via Pretrained Diffusion Model

    We propose EditCrafter, a high-resolution image editing method that operates without tuning, leveraging pretrained text-to-image (T2I) diffusion models to process images at resolutions significantly exceeding those used during training. Leveraging the generative priors of large-scale T2I diffusion models enables the development of a wide array of novel generation and editing applications. Although numerous image editing methods have been proposed based on diffusion models and exhibit high-quality editing results, they are difficult to apply to images with arbitrary aspect ratios or higher resolutions since they only work at the training resolutions (512x512 or 1024x1024). Naively applying patch-wise editing fails with unrealistic object structures and repetition. To address these challenges, we introduce EditCrafter, a simple yet effective editing pipeline. EditCrafter operates by first performing tiled inversion, which preserves the original identity of the input high-resolution image. We further propose a noise-damped manifold-constrained classifier-free guidance (NDCFG++) that is tailored for high resolution image editing from the inverted latent. Our experiments show that the our EditCrafter can achieve impressive editing results across various resolutions without fine-tuning and optimization.

  11. Context Unrolling in Omni Models

    We present Omni, a unified multimodal model natively trained on diverse modalities, including text, images, videos, 3D geometry, and hidden representations. We find that such training enables Context Unrolling, where the model explicitly reasons across multiple modal representations before producing predictions. This process enables the model to aggregate complementary information across heterogeneous modalities, facilitating a more faithful approximation of the shared multimodal knowledge manifold and improving downstream reasoning fidelity. As a result, Omni achieves strong performance on both multimodal generation and understanding benchmarks, while demonstrating advanced multimodal reasoning capabilities, including in-context generation of text, image, video, and 3D geometry.

  12. Vista4D: Video Reshooting with 4D Point Clouds

    We present Vista4D, a robust and flexible video reshooting framework that grounds the input video and target cameras in a 4D point cloud. Specifically, given an input video, our method re-synthesizes the scene with the same dynamics from a different camera trajectory and viewpoint. Existing video reshooting methods often struggle with depth estimation artifacts of real-world dynamic videos, while also failing to preserve content appearance and failing to maintain precise camera control for challenging new trajectories. We build a 4D-grounded point cloud representation with static pixel segmentation and 4D reconstruction to explicitly preserve seen content and provide rich camera signals, and we train with reconstructed multiview dynamic data for robustness against point cloud artifacts during real-world inference. Our results demonstrate improved 4D consistency, camera control, and visual quality compared to state-of-the-art baselines under a variety of videos and camera paths. Moreover, our method generalizes to real-world applications such as dynamic scene expansion and 4D scene recomposition. See our project page for results, code, and models: https://eyeline-labs.github.io/Vista4D

  13. WebGen-R1: Incentivizing Large Language Models to Generate Functional and Aesthetic Websites with Reinforcement Learning

    While Large Language Models (LLMs) excel at function-level code generation, project-level tasks such as generating functional and visually aesthetic multi-page websites remain highly challenging. Existing works are often limited to single-page static websites, while agentic frameworks typically rely on multi-turn execution with proprietary models, leading to substantial token costs, high latency, and brittle integration. Training a small LLM end-to-end with reinforcement learning (RL) is a promising alternative, yet it faces a critical bottleneck in designing reliable and computationally feasible rewards for website generation. Unlike single-file coding tasks that can be verified by unit tests, website generation requires evaluating inherently subjective aesthetics, cross-page interactions, and functional correctness. To this end, we propose WebGen-R1, an end-to-end RL framework tailored for project-level website generation. We first introduce a scaffold-driven structured generation paradigm that constrains the large open-ended action space and preserves architectural integrity. We then design a novel cascaded multimodal reward that seamlessly couples structural guarantees with execution-grounded functional feedback and vision-based aesthetic supervision. Extensive experiments demonstrate that our WebGen-R1 substantially transforms a 7B base model from generating nearly nonfunctional websites into producing deployable, aesthetically aligned multi-page websites. Remarkably, our WebGen-R1 not only consistently outperforms heavily scaled open-source models (up to 72B), but also rivals the state-of-the-art DeepSeek-R1 (671B) in functional success, while substantially exceeding it in valid rendering and aesthetic alignment. These results position WebGen-R1 as a viable path for scaling small open models from function-level code generation to project-level web application generation.

  14. UniGenDet: A Unified Generative-Discriminative Framework for Co-Evolutionary Image Generation and Generated Image Detection

    In recent years, significant progress has been made in both image generation and generated image detection. Despite their rapid, yet largely independent, development, these two fields have evolved distinct architectural paradigms: the former predominantly relies on generative networks, while the latter favors discriminative frameworks. A recent trend in both domains is the use of adversarial information to enhance performance, revealing potential for synergy. However, the significant architectural divergence between them presents considerable challenges. Departing from previous approaches, we propose UniGenDet: a Unified generative-discriminative framework for co-evolutionary image Generation and generated image Detection. To bridge the task gap, we design a symbiotic multimodal self-attention mechanism and a unified fine-tuning algorithm. This synergy allows the generation task to improve the interpretability of authenticity identification, while authenticity criteria guide the creation of higher-fidelity images. Furthermore, we introduce a detector-informed generative alignment mechanism to facilitate seamless information exchange. Extensive experiments on multiple datasets demonstrate that our method achieves state-of-the-art performance. Code: https://github.com/Zhangyr2022/UniGenDet{https://github.com/Zhangyr2022/UniGenDet}.

  15. Coevolving Representations in Joint Image-Feature Diffusion

    Joint image-feature generative modeling has recently emerged as an effective strategy for improving diffusion training by coupling low-level VAE latents with high-level semantic features extracted from pre-trained visual encoders. However, existing approaches rely on a fixed representation space, constructed independently of the generative objective and kept unchanged during training. We argue that the representation space guiding diffusion should itself adapt to the generative task. To this end, we propose Coevolving Representation Diffusion (CoReDi), a framework in which the semantic representation space evolves during training by learning a lightweight linear projection jointly with the diffusion model. While naively optimizing this projection leads to degenerate solutions, we show that stable coevolution can be achieved through a combination of stop-gradient targets, normalization, and targeted regularization that prevents feature collapse. This formulation enables the semantic space to progressively specialize to the needs of image synthesis, improving its complementarity with image latents. We apply CoReDi to both VAE latent diffusion and pixel-space diffusion, demonstrating that adaptive semantic representations improve generative modeling across both settings. Experiments show that CoReDi achieves faster convergence and higher sample quality compared to joint diffusion models operating in fixed representation spaces.

Techmeme(15)

  1. Series, founded by two seniors at Yale to build an AI-powered social network on iMessage, raised a $5.1M pre-seed from Reddit CEO Steve Huffman and others (Dominic-Madori Davis/TechCrunch)

    Dominic-Madori Davis / TechCrunch : Series, founded by two seniors at Yale to build an AI-powered social network on iMessage, raised a $5.1M pre-seed from Reddit CEO Steve Huffman and others —  Series, a social networking app, announced that it raised a $5.1 million pre-seed round, with investors including Venmo co-founder …

  2. The Wire by Acutus, an AI-generated site with articles attacking AI industry critics, appears to be funded by the OpenAI-backed super PAC Leading The Future (@themidasproj)

    @themidasproj : The Wire by Acutus, an AI-generated site with articles attacking AI industry critics, appears to be funded by the OpenAI-backed super PAC Leading The Future —  The reporters at this news site are AI bots. OpenAI's super PAC appears to be funding it.

  3. Nvidia stock jumps 4.3% to close at a record for the first time since Oct., pushing Nvidia's market cap past $5T, as a rally in Intel pushed chipmakers higher (Jordan Novet/CNBC)

    Jordan Novet / CNBC : Nvidia stock jumps 4.3% to close at a record for the first time since Oct., pushing Nvidia's market cap past $5T, as a rally in Intel pushed chipmakers higher —  Nvidia shares closed at a record on Friday for the first time since October, pushing the company's market cap past $5 trillion …

  4. Sam Altman apologizes to a Canadian town for not alerting police to the activity of a mass shooting suspect when her ChatGPT account was suspended (Wall Street Journal)

    Wall Street Journal : Sam Altman apologizes to a Canadian town for not alerting police to the activity of a mass shooting suspect when her ChatGPT account was suspended —  Sam Altman said AI company will work more closely with governments to prevent future tragedies  —  OpenAI Chief Executive Sam Altman apologized …

  5. Intel stock closes up 23.6%, its best performance since October 1987, as Intel shows signs of renewed growth amid the AI boom; the stock is up 124% YTD (CNBC)

    CNBC : Intel stock closes up 23.6%, its best performance since October 1987, as Intel shows signs of renewed growth amid the AI boom; the stock is up 124% YTD —  Intel shares soared 24% on Friday, their best performance since October 1987, as investors cheered signs of renewed growth due to mounting artificial intelligence demand.

  6. ComfyUI, which gives creators granular control over image, video, and audio outputs from diffusion models, raised $30M at a $500M valuation (Marina Temkin/TechCrunch)

    Marina Temkin / TechCrunch : ComfyUI, which gives creators granular control over image, video, and audio outputs from diffusion models, raised $30M at a $500M valuation —  ComfyUI, a startup that helps creators control image, video, and audio outputs from diffusion models with a node-based workflow …

  7. Sources: Stanford University professor James Zou aims to raise ~$100M at a ~$1B valuation for Human Intelligence, which aims to use AI to study physiology (Rebecca Torrence/Bloomberg)

    Rebecca Torrence / Bloomberg : Sources: Stanford University professor James Zou aims to raise ~$100M at a ~$1B valuation for Human Intelligence, which aims to use AI to study physiology —  Stanford University professor James Zou is raising money at a roughly $1 billion valuation to build artificial intelligence models …

  8. Maine's governor vetoes a bill that would have led to US' first state pause on data centers, citing its failure to exempt a project in a distressed mill town (Jenna Russell/New York Times)

    Jenna Russell / New York Times : Maine's governor vetoes a bill that would have led to US' first state pause on data centers, citing its failure to exempt a project in a distressed mill town —  Gov. Janet Mills said she rejected what would have been the nation's first moratorium on data centers because it failed to exempt a project in a distressed mill town.

  9. Source: Meta has a system in India to "automatically restrict content, at scale" to meet local law, massively expanding the country's censorship powers (Aroon Deep/The Hindu)

    Aroon Deep / The Hindu : Source: Meta has a system in India to “automatically restrict content, at scale” to meet local law, massively expanding the country's censorship powers —  India is now part of a small group of countries where content takedown notices are complied with immediately …

  10. X launches its standalone messaging app XChat on the App Store, saying it supports end-to-end encryption and has no ads (Zac Hall/9to5Mac)

    Zac Hall / 9to5Mac : X launches its standalone messaging app XChat on the App Store, saying it supports end-to-end encryption and has no ads —  XChat, the standalone messaging app from X, is now available.  X's new iPhone and iPad app has officially arrived on the App Store.  —  XChat has officially launched on iPhone and iPad

  11. Sources: AI startups are struggling to access Nvidia GPUs as Microsoft and other cloud providers divert supply to internal teams and large customers like OpenAI (The Information)

    The Information : Sources: AI startups are struggling to access Nvidia GPUs as Microsoft and other cloud providers divert supply to internal teams and large customers like OpenAI —  AI startups are struggling to access Nvidia graphics processing units as Microsoft and other cloud providers divert GPU stockpiles …

  12. Sources: JPMorgan and other banks struggled to spread the risk of billions in loans they made to build data centers leased to Oracle in Texas and Wisconsin (Wall Street Journal)

    Wall Street Journal : Sources: JPMorgan and other banks struggled to spread the risk of billions in loans they made to build data centers leased to Oracle in Texas and Wisconsin —  The AI boom has hit a funding snag, compounding power constraints and a growing public backlash against data centers

  13. The DOJ joins xAI in its legal challenge to a new Colorado law that seeks to prevent discrimination by AI tools in employment and other areas (Madlin Mekelburg/Bloomberg)

    Madlin Mekelburg / Bloomberg : The DOJ joins xAI in its legal challenge to a new Colorado law that seeks to prevent discrimination by AI tools in employment and other areas —  The Trump administration is joining Elon Musk's artificial intelligence company xAI in its legal challenge to Colorado's new state law that seeks …

  14. Anthropic says Google is committing to invest $10B now in cash at a $350B valuation and will invest another $30B if Anthropic hits performance targets (Bloomberg)

    Bloomberg : Anthropic says Google is committing to invest $10B now in cash at a $350B valuation and will invest another $30B if Anthropic hits performance targets —  Google will invest $10 billion in Anthropic PBC, with another $30 billion potentially to follow, strengthening the relationship between …

  15. France's forecasting office refers suspected weather sensor tampering at Paris airport to police, after detecting unusual readings alongside Polymarket betting (Joe Wertz/Bloomberg)

    Joe Wertz / Bloomberg : France's forecasting office refers suspected weather sensor tampering at Paris airport to police, after detecting unusual readings alongside Polymarket betting —  France's forecasting office flagged suspected tampering with weather sensors at the country's largest airport and referred the case to police …

Solidot(15)

  1. Firefox 悄悄集成 Brave 的 Adblock 引擎

    Mozilla 上个月释出的 Firefox 149 悄悄集成了 Brave 的开源 Adblock 引擎 adblock-rust。adblock-rust 在默认情况下没有启用,也没有 UI 或内容过滤列表。dblock-rust 是 Brave 内置广告屏蔽器使用的引擎,使用 Rust 开发,采用 MPL-2.0 许可授权,能处理网络请求拦截、过滤特定元素样式(cosmetic filtering),兼容 uBlock Origin 的过滤列表语法。Firefox 分支 Waterfox 也采用了 adblock-rust。

  2. 减少接触塑料会在短期内大幅减少体内的塑料物质含量

    根据一项研究,减少接触塑料会在短期内减少体内的塑料物质含量最多五成。研究主要针对两种来自塑料制品的化学物质——邻苯二甲酸酯(phthalates)和双酚(bisphenols)。211 名试验参与者的尿检结果显示他们体内都有较高含量的塑料化学物质。其中 60 名参与者参与了一项随机对照试验。研究人员与农民和生产商合作,提供在生产和包装过程中完全不用塑料的食品。一周后相比对照组,减少接触塑料的参与者体内邻苯二甲酸酯减少了 44%,双酚减少了逾 50%。

  3. 英国生物银行 50 万参与者健康数据泄漏

    英国生物银行(UK Biobank)有 50 万参与者的健康数据泄漏,泄漏数据集在阿里巴巴上出售。英国科技大臣 Ian Murray 称英国生物银行运营机构已向政府通报了这起事件,泄露的信息不包含姓名、地址、联系方式或电话号码,但可能包括性别、年龄、出生年月、社会经济地位、生活习惯以及生物样本的测量数据。英国生物银行收集志愿者提供的健康数据,被用于帮助改进痴呆症、癌症和帕金森病的检测和治疗。有逾 1.8 万篇论文使用了英国生物银行的数据。阿里巴巴已经删除了数据,但相关数据又被上传到了 GitHub 上,英国生物银行向 GitHub 递交了大量 DMCA 删除通知。

  4. 天文学家发现银河系的边缘

    银河系是圆盘状恒星,也就是恒星的形成是由里向外,这意味着越远的恒星越年轻。天文学家分析了银河系逾 10 万颗巨型恒星,发现恒星年龄分布由内向外模式在距离银河系中心 3.5-4 万光年间发生逆转,超过该距离恒星年龄变大了。这种突变形成了 U 型年龄分布,曲线最低点对应的恒星形成率急剧下降,代表着银河系恒星形成盘的边界。天文学家认为他们识别了银河系圆盘的边缘。

  5. NASA Roman 太空望远镜最早九月发射

    NASA 宣布 Nancy Grace Roman 太空望远镜最早今年九月发射最晚 2027 年 5 月发射。望远镜以 NASA 首任天文学部门女主任的名字命名,它属于红外望远镜,核心任务包括探测暗能量、发现系外行星及验证广义相对论宇宙时空曲率。望远镜使用美国国家侦察局捐赠的 2.4 米口径主镜(当年 NRO 捐赠了两台主镜),配备了两台科学仪器:3 亿像素多波段红外相机大视场仪表(WFI),能直接观测邻近恒星周围的类木行星的日冕仪(CGI)。

  6. 深度求索发布 DeepSeek-V4 预览版

    深度求索发布了 DeepSeek-V4 预览版。DeepSeek-V4 有两个版本,其中 Pro 版本有 1.6 万亿参数其中 490 亿活跃参数;Flash 版本有 2840 亿参数其中活跃参数 130 亿。两个版本都支持百万上下文。DeepSeek V4 除了支持英伟达 GPU 还支持华为昇腾 NPU。深度求索称,在 Agentic Coding 评测中,V4-Pro 已达到当前开源模型最佳水平,并在其他 Agent 相关评测中同样表现优异;Pro 在世界知识测评中,大幅领先其他开源模型,仅稍逊于顶尖闭源模型 Gemini-Pro-3.1;在数学、STEM、竞赛型代码的测评中,V4-Pro 超越当前所有已公开评测的开源模型。

  7. Tim Cook 称 2012 年发布 Apple Maps 是其任内第一个重大失误

    在最近举行的员工大会上,即将卸任的苹果 CEO 库克(Tim Cook)称 2012 年上线 Apple Maps 是其任内第一个真正的重大失误。当时的 Apple Maps 还比较原始,尚未准备好。它的失败迫使库克向用户道歉,并建议用户使用竞争对手的地图应用。库克表示这次重大失败还是富有价值的,“这体现了我们始终将用户放置在决策中心。现在我们拥有全世界最好的地图应用。我们学会了坚持,犯了错误但做了正确的事。”

  8. 心脏跳动能抑制心脏癌生长

    根据发表在《科学》期刊上的一项研究,心脏持续不断的跳动可能会主动抑制心脏组织中的肿瘤生长。这些组织中的细胞通路会改变癌细胞的基因调控方式,从而阻止其增殖。这些发现揭示了机械力在保护心脏免受癌症侵害方面的作用,并可能为基于机械刺激的新型癌症疗法做好铺垫。在哺乳动物中,心脏癌极为罕见。更重要的是,成年人的心脏自我更新能力有限,心肌细胞的再生率每年约为 1%。人们对这些特征所提出的一种解释是:心脏组织承受着巨大的机械负荷,它必须克服很大的阻力而持续泵血。这种持续的压力似乎会抑制心脏细胞的增殖能力。通过利用基因工程小鼠模型,研究人员发现,即使引入了强力致癌突变,心脏对其也具有显著的抗性。研究发现,组织内的机械力会重塑癌细胞基因组的调节格局,从而影响癌细胞是否能够增殖。Nesprin-2 是这一过程的核心,这是一种将细胞表面的机械信号传递至细胞核的蛋白。作为 LINC 复合物的一个组成部分,Nesprin-2 可感知心脏的机械微环境,并可在功能上改变染色质结构和组蛋白甲基化,从而降低与肿瘤细胞增殖相关的基因活性。当癌细胞中的 Nesprin-2 被沉默后,这些细胞在机械活动的心脏中会重新获得生长能力并形成肿瘤。

  9. Ubuntu 26.04 LTS 释出

    Canonical 释出了代号为 Resolute Raccoon 的 Ubuntu 26.04 LTS。同时释出的还有衍生版本 Edubuntu、Kubuntu、Lubuntu、Ubuntu Budgie、Ubuntu Cinnamon、Ubuntu Kylin、Ubuntu Studio、Ubuntu Unity 和 Xubuntu。Ubuntu Desktop、Ubuntu Server、Ubuntu Cloud、Ubuntu WSL 和 Ubuntu Core 将获得五年的支持,其余版本获得三年的支持,付费扩展支持 ESM (Expanded Security Maintenance)为十年 。Ubuntu 26.04 采用最新的 Linux 7.0 kernel,GNOME 50 桌面环境,引入了基于 TPM 的全盘加密,GStreamer 1.28,沙盒图形加载,Chrony 4.8,等等。

  10. AI 漏洞报告大增促使内核移除缺乏维护的代码

    内核开发者宣布,为了保持自己的理智不被大量涌入的 AI 生成漏洞报告搞疯,他们正在 Linux 内核移除缺乏维护的代码。计划移除的代码包括 ISA 和 PCMCIA Ethernet 驱动、一对 PCI 驱动、ax25 和业余无线电子系统、ATM 协议和驱动,以及 ISDN 子系统。

  11. 猕猴吃土帮助消化游客的高热量垃圾食品

    剑桥大学的一项研究发现,欧洲唯一野猴群、生活在直布罗陀的猕猴群会定期吃土。对猴群的监测显示,与游客接触频繁的猕猴摄入的泥土量更多,而在旅游旺季食土率也更高。科学家认为游客提供或从游客偷取的高热量食物如巧克力、薯片和冰淇淋正在扰乱它们的肠道微生物组成,导致了其习惯的改变。吃土可能有助于猕猴恢复肠胃平衡,土壤可以提供垃圾食品缺乏的细菌和矿物质,可能有助于保护肠道内壁,缓解或预防因摄入过多糖分和脂肪引起的刺激。猕猴群平均每周有 12 次吃土行为。30% 的吃土行为发生在群体中,89% 发生在其它猕猴在场的情况下,其它猕猴通常会在旁观察,表明这种行为是“社会习得的”。大多数猕猴偏爱食用红土(terra rossa),占到了所有事件的 83%。

  12. 特朗普模因币导致投资者损失数十亿美元

    2025 年 1 月总统就职典礼前特朗普推出了官方模因币,成为首位发行个人加密货币的总统。分析认为特朗普家族从中获利逾 2.8 亿美元,但投资该模因币的散户则损失了逾 43 亿美元。大部分特朗普模因币掌握在内部人士手中,通过关键时刻抛售和收集手续费,他们没有损失,反而获利逾 6 亿美元。特朗普代币最初定价 28.73 美元,如今的价格不到 3 美元,币值跌了 93%。第一夫人 Melania Trump 的官方模因币则更惨,币值相比峰值跌了 99%。特朗普家族的加密货币公司 World Liberty Financial 则为其带来了约 50 亿美元的财富。

  13. 53 国齐聚哥伦比亚商讨淘汰化石燃料

    53 国(不包括主要排放国中国、美国、印度和俄罗斯)将于下周齐聚哥伦比亚商讨如何逐步淘汰化石燃料。霍尔木兹海峡的关闭让亚太国家面临能源危机,迫使这些国家采取紧急措施如推行远程办公和关闭学校。与会国并非传统盟友,而是在短时间内组建的松散联盟,这也反映了形势的紧迫性。最新的能源危机可能成为可再生清洁能源的一个转折点。化石燃料行业多年来一直宣称石油、天然气和煤炭是可靠的能源,但危机显示它们并不可靠,可再生能源则廉价、可靠且安全。相比 1970 年代的中东石油危机,今天可作为化石燃料提到的清洁能源已经成熟。自 1970 年代以来,太阳能电池板的价格下降了 99.9%,自 1984 年以来风能成本下降了 91%,自 1991 年以来电池价格下降了 99%。此前韩国七成原油经过霍尔木兹海峡,现在韩国计划四年内将可再生能源装机容量翻一番。过去 30 年的全球气候谈判几乎从未提及化石燃料,部分原因是化石燃料主要出口国和游说团体的阻挠。现在一个松散的国家联盟正绕过全球气候谈判,讨论如何真正逐步淘汰化石燃料。

  14. 加密货币骗子瞄准滞留在霍尔木兹海峡附近的船只

    由于伊朗关闭了霍尔木兹海峡,目前有约 2000 艘船只和 2 万名海员滞留在附近。伊朗几周前表示通过海峡的油轮需要以加密货币形式支付过境费,但美国随后也宣布关闭海峡搜查过往船只。形势相当混乱,而在混乱之中骗子开始浑水摸鱼。据报道,加密货币骗子正冒充伊朗当局,向航运公司发送信息,要求以比特币或泰达币支付过境费。未经证实的消息称,有船只认为已经支付了过境费而试图通过海峡,在遭到伊朗炮击后不得不返回。目前已发生了两次类似事件,一次是 4 月 18 日,一次是 4 月 22 日。

  15. 古代人类曾三次迁徙到南美洲

    南美洲是人类最后定居的一块大陆。科学家以前认为迁徙到南美很简单:大约 1.5 万前来自大约同一群体的古代人类祖先进入这块大陆,逐渐适应从丛林到高原等不同环境。但发表在《自然》期刊上的古代和现代基因组分析研究显示:古代人类至少分三次迁徙到南美洲。第一波迁徙大约发生在 1.27 万年前,第二波是 9000 年前,第三波则是约 1300 年前。第三波迁徙者与中美洲人有关联。