DIGEST · 2026-03-23

OrangeBot.AI Digest — 2026-03-23

90 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Autoresearch on an old research idea (ykumar.me)
  2. US and TotalEnergies reach 'nearly $1B' deal to end offshore wind projects (www.lemonde.fr)
  3. Is it a pint? (isitapint.com)
  4. Cyber.mil serving file downloads using TLS certificate which expired 3 days ago (www.cyber.mil)
  5. If DSPy is so great, why isn't anyone using it? (skylarbpayne.com)
  6. iPhone 17 Pro Demonstrated Running a 400B LLM (twitter.com)
  7. An unsolicited guide to being a researcher [pdf] (emerge-lab.github.io)
  8. Bombadil: Property-based testing for web UIs (github.com)
  9. Student beauty and grades under in-person and remote teaching (www.sciencedirect.com)
  10. GitHub appears to be struggling with measly three nines availability (www.theregister.com)
  11. Migrating to the EU (rz01.org)
  12. I built an AI receptionist for a mechanic shop (www.itsthatlady.dev)
  13. POSSE – Publish on your Own Site, Syndicate Elsewhere (indieweb.org)
  14. Walmart: ChatGPT checkout converted 3x worse than website (searchengineland.com)
  15. Two pilots dead after plane and ground vehicle collide at LaGuardia (www.bbc.com)

GitHub Trending(15)

  1. FujiwaraChoki / MoneyPrinterV2
  2. bytedance / deer-flow
  3. Crosstalk-Solutions / project-nomad
  4. vxcontrol / pentagi
  5. browser-use / browser-use
  6. TauricResearch / TradingAgents
  7. tinygrad / tinygrad
  8. affaan-m / everything-claude-code
  9. NousResearch / hermes-agent
  10. jingyaogong / minimind
  11. hsliuping / TradingAgents-CN
  12. kepano / obsidian-skills
  13. czlonkowski / n8n-mcp
  14. iptv-org / iptv
  15. hesreallyhim / awesome-claude-code

Product Hunt(15)

  1. Claude Usage Tracker

    See exactly how much you spend on Claude, across every tool

  2. Zoer.ai

    Build full-stack webapps from the database up

  3. Honestly

    Real reviews from Reddit & YouTube when shopping online

  4. Pause.do

    Interrupt scrolling, tab overload, and AI autopilot

  5. Tobira.ai

    A network where AI agents find deals for their humans

  6. Fastlane

    Tinder for Marketing

  7. Iris

    Send work beautifully, pinned feedback, see what they viewed

  8. DataSieve 2.0

    Extract structured data from text, files and archives.

  9. AlphaClaw Apex

    OpenClaw harness and fleet manager for Mac

  10. Nomie

    AI wellness app that turns doomscrolling into self‑care

  11. WeixinClawBot

    The official WeChat pipeline for OpenClaw

  12. Embedful

    Easy data visualizations. Embed and share anywhere.

  13. Claude Code Scheduled Tasks

    Schedule recurring tasks locally and in the cloud easily

  14. Edgee Claude Code Compressor

    Extend Claude Pro's limit by 26.2%

  15. Bench for Claude Code

    Store, review, and share your Claude Code sessions

Hugging Face(15)

  1. HopChain: Multi-Hop Data Synthesis for Generalizable Vision-Language Reasoning

    VLMs show strong multimodal capabilities, but they still struggle with fine-grained vision-language reasoning. We find that long CoT reasoning exposes diverse failure modes, including perception, reasoning, knowledge, and hallucination errors, which can compound across intermediate steps. However, most existing vision-language data used for RLVR does not involve complex reasoning chains that rely on visual evidence throughout, leaving these weaknesses largely unexposed. We therefore propose HopChain, a scalable framework for synthesizing multi-hop vision-language reasoning data specifically for RLVR training of VLMs. Each synthesized multi-hop query forms a logically dependent chain of instance-grounded hops, where earlier hops establish the instances, sets, or conditions needed for later hops, while the final answer remains a specific, unambiguous number suitable for verifiable rewards. We add the multi-hop data synthesized by HopChain to the original RLVR data used to train Qwen3.5-35B-A3B and Qwen3.5-397B-A17B, and compare against RLVR on the original RLVR data alone across 24 benchmarks spanning STEM and Puzzle, General VQA, Text Recognition and Document Understanding, and Video Understanding. Although this multi-hop data is not synthesized to target any specific benchmark, adding it improves 20 out of 24 benchmarks on both models, indicating broad and generalizable gains. To demonstrate that full chained queries are important, we replace them with half-multi-hop or single-hop variants, reducing the 24-benchmark average accuracy by 5.3 and 7.0 points, respectively. Multi-hop training also strengthens long-CoT vision-language reasoning, with gains peaking at more than 50 accuracy points in the ultra-long-CoT regime. These experiments establish HopChain as an effective, scalable framework for synthesizing multi-hop data that improves generalizable vision-language reasoning.

  2. Astrolabe: Steering Forward-Process Reinforcement Learning for Distilled Autoregressive Video Models

    Distilled autoregressive (AR) video models enable efficient streaming generation but frequently misalign with human visual preferences. Existing reinforcement learning (RL) frameworks are not naturally suited to these architectures, typically requiring either expensive re-distillation or solver-coupled reverse-process optimization that introduces considerable memory and computational overhead. We present Astrolabe, an efficient online RL framework tailored for distilled AR models. To overcome existing bottlenecks, we introduce a forward-process RL formulation based on negative-aware fine-tuning. By contrasting positive and negative samples directly at inference endpoints, this approach establishes an implicit policy improvement direction without requiring reverse-process unrolling. To scale this alignment to long videos, we propose a streaming training scheme that generates sequences progressively via a rolling KV-cache, applying RL updates exclusively to local clip windows while conditioning on prior context to ensure long-range coherence. Finally, to mitigate reward hacking, we integrate a multi-reward objective stabilized by uncertainty-aware selective regularization and dynamic reference updates. Extensive experiments demonstrate that our method consistently enhances generation quality across multiple distilled AR video models, serving as a robust and scalable alignment solution.

  3. TerraScope: Pixel-Grounded Visual Reasoning for Earth Observation

    Vision-language models (VLMs) have shown promise in earth observation (EO), yet they struggle with tasks that require grounding complex spatial reasoning in precise pixel-level visual representations. To address this problem, we introduce TerraScope, a unified VLM that delivers pixel-grounded geospatial reasoning with two key capabilities: (1) modality-flexible reasoning: it handles single-modality inputs (optical or SAR) and adaptively fuses different modalities into the reasoning process when both are available; (2) multi-temporal reasoning: it integrates temporal sequences for change analysis across multiple time points. In addition, we curate Terra-CoT, a large-scale dataset containing 1 million samples with pixel-level masks embedded in reasoning chains across multiple sources. We also propose TerraScope-Bench, the first benchmark for pixel-grounded geospatial reasoning with six sub-tasks that evaluates both answer accuracy and mask quality to ensure authentic pixel-grounded reasoning. Experiments show that TerraScope significantly outperforms existing VLMs on pixel-grounded geospatial reasoning while providing interpretable visual evidence.

  4. ProactiveBench: Benchmarking Proactiveness in Multimodal Large Language Models

    Effective collaboration begins with knowing when to ask for help. For example, when trying to identify an occluded object, a human would ask someone to remove the obstruction. Can MLLMs exhibit a similar "proactive" behavior by requesting simple user interventions? To investigate this, we introduce ProactiveBench, a benchmark built from seven repurposed datasets that tests proactiveness across different tasks such as recognizing occluded objects, enhancing image quality, and interpreting coarse sketches. We evaluate 22 MLLMs on ProactiveBench, showing that (i) they generally lack proactiveness; (ii) proactiveness does not correlate with model capacity; (iii) "hinting" at proactiveness yields only marginal gains. Surprisingly, we found that conversation histories and in-context learning introduce negative biases, hindering performance. Finally, we explore a simple fine-tuning strategy based on reinforcement learning: its results suggest that proactiveness can be learned, even generalizing to unseen scenarios. We publicly release ProactiveBench as a first step toward building proactive multimodal models.

  5. FlowScene: Style-Consistent Indoor Scene Generation with Multimodal Graph Rectified Flow

    Scene generation has extensive industrial applications, demanding both high realism and precise control over geometry and appearance. Language-driven retrieval methods compose plausible scenes from a large object database, but overlook object-level control and often fail to enforce scene-level style coherence. Graph-based formulations offer higher controllability over objects and inform holistic consistency by explicitly modeling relations, yet existing methods struggle to produce high-fidelity textured results, thereby limiting their practical utility. We present FlowScene, a tri-branch scene generative model conditioned on multimodal graphs that collaboratively generates scene layouts, object shapes, and object textures. At its core lies a tight-coupled rectified flow model that exchanges object information during generation, enabling collaborative reasoning across the graph. This enables fine-grained control of objects' shapes, textures, and relations while enforcing scene-level style coherence across structure and appearance. Extensive experiments show that FlowScene outperforms both language-conditioned and graph-conditioned baselines in terms of generation realism, style consistency, and alignment with human preferences.

  6. The Y-Combinator for LLMs: Solving Long-Context Rot with λ-Calculus

    LLMs are increasingly used as general-purpose reasoners, but long inputs remain bottlenecked by a fixed context window. Recursive Language Models (RLMs) address this by externalising the prompt and recursively solving subproblems. Yet existing RLMs depend on an open-ended read-eval-print loop (REPL) in which the model generates arbitrary control code, making execution difficult to verify, predict, and analyse. We introduce λ-RLM, a framework for long-context reasoning that replaces free-form recursive code generation with a typed functional runtime grounded in λ-calculus. It executes a compact library of pre-verified combinators and uses neural inference only on bounded leaf subproblems, turning recursive reasoning into a structured functional program with explicit control flow. We show that λ-RLM admits formal guarantees absent from standard RLMs, including termination, closed-form cost bounds, controlled accuracy scaling with recursion depth, and an optimal partition rule under a simple cost model. Empirically, across four long-context reasoning tasks and nine base models, λ-RLM outperforms standard RLM in 29 of 36 model-task comparisons, improves average accuracy by up to +21.9 points across model tiers, and reduces latency by up to 4.1x. These results show that typed symbolic control yields a more reliable and efficient foundation for long-context reasoning than open-ended recursive code generation. The complete implementation of λ-RLM, is open-sourced for the community at: https://github.com/lambda-calculus-LLM/lambda-RLM.

  7. LumosX: Relate Any Identities with Their Attributes for Personalized Video Generation

    Recent advances in diffusion models have significantly improved text-to-video generation, enabling personalized content creation with fine-grained control over both foreground and background elements. However, precise face-attribute alignment across subjects remains challenging, as existing methods lack explicit mechanisms to ensure intra-group consistency. Addressing this gap requires both explicit modeling strategies and face-attribute-aware data resources. We therefore propose LumosX, a framework that advances both data and model design. On the data side, a tailored collection pipeline orchestrates captions and visual cues from independent videos, while multimodal large language models (MLLMs) infer and assign subject-specific dependencies. These extracted relational priors impose a finer-grained structure that amplifies the expressive control of personalized video generation and enables the construction of a comprehensive benchmark. On the modeling side, Relational Self-Attention and Relational Cross-Attention intertwine position-aware embeddings with refined attention dynamics to inscribe explicit subject-attribute dependencies, enforcing disciplined intra-group cohesion and amplifying the separation between distinct subject clusters. Comprehensive evaluations on our benchmark demonstrate that LumosX achieves state-of-the-art performance in fine-grained, identity-consistent, and semantically aligned personalized multi-subject video generation. Code and models are available at https://jiazheng-xing.github.io/lumosx-home/.

  8. Hyperagents

    Self-improving AI systems aim to reduce reliance on human engineering by learning to improve their own learning and problem-solving processes. Existing approaches to self-improvement rely on fixed, handcrafted meta-level mechanisms, fundamentally limiting how fast such systems can improve. The Darwin Gödel Machine (DGM) demonstrates open-ended self-improvement in coding by repeatedly generating and evaluating self-modified variants. Because both evaluation and self-modification are coding tasks, gains in coding ability can translate into gains in self-improvement ability. However, this alignment does not generally hold beyond coding domains. We introduce hyperagents, self-referential agents that integrate a task agent (which solves the target task) and a meta agent (which modifies itself and the task agent) into a single editable program. Crucially, the meta-level modification procedure is itself editable, enabling metacognitive self-modification, improving not only the task-solving behavior, but also the mechanism that generates future improvements. We instantiate this framework by extending DGM to create DGM-Hyperagents (DGM-H), eliminating the assumption of domain-specific alignment between task performance and self-modification skill to potentially support self-accelerating progress on any computable task. Across diverse domains, the DGM-H improves performance over time and outperforms baselines without self-improvement or open-ended exploration, as well as prior self-improving systems. Furthermore, the DGM-H improves the process by which it generates new agents (e.g., persistent memory, performance tracking), and these meta-level improvements transfer across domains and accumulate across runs. DGM-Hyperagents offer a glimpse of open-ended AI systems that do not merely search for better solutions, but continually improve their search for how to improve.

  9. Reasoning as Compression: Unifying Budget Forcing via the Conditional Information Bottleneck

    Chain-of-Thought (CoT) prompting improves LLM accuracy on complex tasks but often increases token usage and inference cost. Existing "Budget Forcing" methods reducing cost via fine-tuning with heuristic length penalties, suppress both essential reasoning and redundant filler. We recast efficient reasoning as a lossy compression problem under the Information Bottleneck (IB) principle, and identify a key theoretical gap when applying naive IB to transformers: attention violates the Markov property between prompt, reasoning trace, and response. To resolve this issue, we model CoT generation under the Conditional Information Bottleneck (CIB) principle, where the reasoning trace Z acts as a computational bridge that contains only the information about the response Y that is not directly accessible from the prompt X. This yields a general Reinforcement Learning objective: maximize task reward while compressing completions under a prior over reasoning traces, subsuming common heuristics (e.g., length penalties) as special cases (e.g., uniform priors). In contrast to naive token-counting-based approaches, we introduce a semantic prior that measures token cost by surprisal under a language model prior. Empirically, our CIB objective prunes cognitive bloat while preserving fluency and logic, improving accuracy at moderate compression and enabling aggressive compression with minimal accuracy drop.

  10. A Subgoal-driven Framework for Improving Long-Horizon LLM Agents

    Large language model (LLM)-based agents have emerged as powerful autonomous controllers for digital environments, including mobile interfaces, operating systems, and web browsers. Web navigation, for example, requires handling dynamic content and long sequences of actions, making it particularly challenging. Existing LLM-based agents struggle with long-horizon planning in two main ways. During online execution, they often lose track as new information arrives, lacking a clear and adaptive path toward the final goal. This issue is further exacerbated during reinforcement learning (RL) fine-tuning, where sparse and delayed rewards make it difficult for agents to identify which actions lead to success, preventing them from maintaining coherent reasoning over extended tasks. To address these challenges, we propose two contributions. First, we introduce an agent framework that leverages proprietary models for online planning through subgoal decomposition. Second, we present MiRA (Milestoning your Reinforcement Learning Enhanced Agent), an RL training framework that uses dense, milestone-based reward signals. The real-time planning mechanism improves proprietary models such as Gemini by approximately a 10% absolute increase in success rate (SR) on the WebArena-Lite benchmark. Meanwhile, applying MiRA to the open Gemma3-12B model increases its success rate from 6.4% to 43.0%. This performance surpasses proprietary systems such as GPT-4-Turbo (17.6%) and GPT-4o (13.9%), as well as the previous open-model state of the art, WebRL (38.4%). Overall, our findings demonstrate that combining explicit inference-time planning with milestone-based rewards significantly improves an agent's long-horizon capabilities, paving the way for more robust and general-purpose autonomous systems.

  11. Deep Tabular Research via Continual Experience-Driven Execution

    Large language models often struggle with complex long-horizon analytical tasks over unstructured tables, which typically feature hierarchical and bidirectional headers and non-canonical layouts. We formalize this challenge as Deep Tabular Research (DTR), requiring multi-step reasoning over interdependent table regions. To address DTR, we propose a novel agentic framework that treats tabular reasoning as a closed-loop decision-making process. We carefully design a coupled query and table comprehension for path decision making and operational execution. Specifically, (i) DTR first constructs a hierarchical meta graph to capture bidirectional semantics, mapping natural language queries into an operation-level search space; (ii) To navigate this space, we introduce an expectation-aware selection policy that prioritizes high-utility execution paths; (iii) Crucially, historical execution outcomes are synthesized into a siamese structured memory, i.e., parameterized updates and abstracted texts, enabling continual refinement. Extensive experiments on challenging unstructured tabular benchmarks verify the effectiveness and highlight the necessity of separating strategic planning from low-level execution for long-horizon tabular reasoning.

  12. Versatile Editing of Video Content, Actions, and Dynamics without Training

    Controlled video generation has seen drastic improvements in recent years. However, editing actions and dynamic events, or inserting contents that should affect the behaviors of other objects in real-world videos, remains a major challenge. Existing trained models struggle with complex edits, likely due to the difficulty of collecting relevant training data. Similarly, existing training-free methods are inherently restricted to structure- and motion-preserving edits and do not support modification of motion or interactions. Here, we introduce DynaEdit, a training-free editing method that unlocks versatile video editing capabilities with pretrained text-to-video flow models. Our method relies on the recently introduced inversion-free approach, which does not intervene in the model internals, and is thus model-agnostic. We show that naively attempting to adapt this approach to general unconstrained editing results in severe low-frequency misalignment and high-frequency jitter. We explain the sources for these phenomena and introduce novel mechanisms for overcoming them. Through extensive experiments, we show that DynaEdit achieves state-of-the-art results on complex text-based video editing tasks, including modifying actions, inserting objects that interact with the scene, and introducing global effects.

  13. BEAVER: A Training-Free Hierarchical Prompt Compression Method via Structure-Aware Page Selection

    The exponential expansion of context windows in LLMs has unlocked capabilities for long-document understanding but introduced severe bottlenecks in inference latency and information utilization. Existing compression methods often suffer from high training costs or semantic fragmentation due to aggressive token pruning. In this paper, we propose BEAVER, a novel training-free framework that shifts compression from linear token removal to structure-aware hierarchical selection. BEAVER maximizes hardware parallelism by mapping variable-length contexts into dense page-level tensors via dual-path pooling, and preserves discourse integrity through a hybrid planner combining semantic and lexical dual-branch selection with sentence smoothing. Extensive evaluations on four long-context benchmarks demonstrate that BEAVER achieves comparable performance to state-of-the-art (SOTA) methods like LongLLMLingua. Notably, on the RULER benchmark, BEAVER maintains high fidelity in multi-needle retrieval where baselines deteriorate. Regarding efficiency, BEAVER reduces latency by 26.4x on 128k contexts, offering a scalable solution for high-throughput applications. Our code is available at https://cslikai.cn/BEAVER/.

  14. HiMu: Hierarchical Multimodal Frame Selection for Long Video Question Answering

    Long-form video question answering requires reasoning over extended temporal contexts, making frame selection critical for large vision-language models (LVLMs) bound by finite context windows. Existing methods face a sharp trade-off: similarity-based selectors are fast but collapse compositional queries into a single dense vector, losing sub-event ordering and cross-modal bindings; agent-based methods recover this structure through iterative LVLM inference, but at prohibitive cost. We introduce HiMu, a training-free framework that bridges this gap. A single text-only LLM call decomposes the query into a hierarchical logic tree whose leaves are atomic predicates, each routed to a lightweight expert spanning vision (CLIP, open-vocabulary detection, OCR) and audio (ASR, CLAP). The resulting signals are normalized, temporally smoothed to align different modalities, and composed bottom-up through fuzzy-logic operators that enforce temporal sequencing and adjacency, producing a continuous satisfaction curve. Evaluations on Video-MME, LongVideoBench and HERBench-Lite show that HiMu advances the efficiency-accuracy Pareto front: at 16 frames with Qwen3-VL 8B it outperforms all competing selectors, and with GPT-4o it surpasses agentic systems operating at 32-512 frames while requiring roughly 10x fewer FLOPs.

  15. How Well Does Generative Recommendation Generalize?

    A widely held hypothesis for why generative recommendation (GR) models outperform conventional item ID-based models is that they generalize better. However, there is few systematic way to verify this hypothesis beyond a superficial comparison of overall performance. To address this gap, we categorize each data instance based on the specific capability required for a correct prediction: either memorization (reusing item transition patterns observed during training) or generalization (composing known patterns to predict unseen item transitions). Extensive experiments show that GR models perform better on instances that require generalization, whereas item ID-based models perform better when memorization is more important. To explain this divergence, we shift the analysis from the item level to the token level and show that what appears to be item-level generalization often reduces to token-level memorization for GR models. Finally, we show that the two paradigms are complementary. We propose a simple memorization-aware indicator that adaptively combines them on a per-instance basis, leading to improved overall recommendation performance.

Techmeme(15)

  1. Crunchyroll is investigating a breach after hackers claimed they accessed a support agent's account and stole the personal information of ~6.8M users (Lawrence Abrams/BleepingComputer)

    Lawrence Abrams / BleepingComputer : Crunchyroll is investigating a breach after hackers claimed they accessed a support agent's account and stole the personal information of ~6.8M users —  Popular anime streaming platform Crunchyroll is investigating a breach after hackers claimed to have stolen personal information for approximately 6.8 million people.

  2. Drone delivery startup Zipline raised an additional $200M, including from Paradigm, bringing Zipline's Series H, originally announced in January, to $800M (Kirsten Korosec/TechCrunch)

    Kirsten Korosec / TechCrunch : Drone delivery startup Zipline raised an additional $200M, including from Paradigm, bringing Zipline's Series H, originally announced in January, to $800M —  U.S. autonomous drone delivery and logistics startup Zipline has raised another $200 million, adding to a recent funding round originally announced in January.

  3. Kalshi announces new guardrails to preemptively block politicians, athletes, and others from trading in their relevant markets (Nathan Bomey/Axios)

    Nathan Bomey / Axios : Kalshi announces new guardrails to preemptively block politicians, athletes, and others from trading in their relevant markets —  Prediction market Kalshi plans to block athletes, coaches and officials from betting on their sports and to block political candidates from trading on their campaigns, Axios has learned.

  4. Meta hires the team behind Dreamer, which lets users create AI agents, including Hugo Barra, former Stripe CTO David Singleton, and designer Nicholas Jitkoff (Kurt Wagner/Bloomberg)

    Kurt Wagner / Bloomberg : Meta hires the team behind Dreamer, which lets users create AI agents, including Hugo Barra, former Stripe CTO David Singleton, and designer Nicholas Jitkoff —  Meta Platforms Inc. has hired the founders and team behind the artificial intelligence startup Dreamer, which launched earlier …

  5. The US plans to create a voluntary consortium of countries to invest $4T to secure supply chains for chips, energy, and minerals; the US will contribute $250M (New York Times)

    New York Times : The US plans to create a voluntary consortium of countries to invest $4T to secure supply chains for chips, energy, and minerals; the US will contribute $250M —  Trump officials said on Monday that the war in Iran had emphasized the need to reduce vulnerabilities for energy and technology.

  6. Doc: Kalshi and Polymarket CEOs are investing in a VC fund, led by two early Kalshi employees, that is raising up to $35M to back prediction market startups (Ben Weiss/Fortune)

    Ben Weiss / Fortune : Doc: Kalshi and Polymarket CEOs are investing in a VC fund, led by two early Kalshi employees, that is raising up to $35M to back prediction market startups —  The CEOs of Kalshi and Polymarket are locked in a brutal fight to dominate the white-hot prediction market sector.

  7. Q&A with Jensen Huang, who says "we've achieved AGI", on running Nvidia, AI scaling laws, OpenClaw, future of coding, data centers in space, China, and more (Lex Fridman)

    Lex Fridman : Q&A with Jensen Huang, who says “we've achieved AGI”, on running Nvidia, AI scaling laws, OpenClaw, future of coding, data centers in space, China, and more —  - Watch the full YouTube version of the podcast  —  Table of Contents  —  Here are the loose “chapters” in the conversation.

  8. Polymarket unveils insider trading rules: no bets on stolen confidential info, illegal tips, or events whose outcomes the user can influence as an insider (Emily Nicolle/Bloomberg)

    Emily Nicolle / Bloomberg : Polymarket unveils insider trading rules: no bets on stolen confidential info, illegal tips, or events whose outcomes the user can influence as an insider —  Polymarket has moved to squash some types of insider trading after the prediction markets platform came under scrutiny for suspected manipulation.

  9. Apple announces WWDC 2026 for June 8-12, which will be an online event free for developers, with an in-person event at Apple Park on June 8 (Juli Clover/MacRumors)

    Juli Clover / MacRumors : Apple announces WWDC 2026 for June 8-12, which will be an online event free for developers, with an in-person event at Apple Park on June 8 —  Apple today said that its 37th annual Worldwide Developers Conference is set to begin on Monday, June 8 and end on Friday, June 12.

  10. Sources: Apple is preparing to introduce ads in its Maps app, allowing retailers and brands to bid for ad slots against search queries (Mark Gurman/Bloomberg)

    Mark Gurman / Bloomberg : Sources: Apple is preparing to introduce ads in its Maps app, allowing retailers and brands to bid for ad slots against search queries —  Apple Inc. is preparing to introduce advertising in its Maps app, part of a broader push to generate more money from services.

  11. Norwegian startup Lace, which is building a form of lithography that uses a helium atom beam instead of light to create chip designs, raised a $40M Series A (Reuters)

    Reuters : Norwegian startup Lace, which is building a form of lithography that uses a helium atom beam instead of light to create chip designs, raised a $40M Series A —  Lace, a Norway-headquartered chipmaking equipment startup which is backed by Microsoft, has raised $40 million in funding …

  12. Doctronic, which became the first company to use AI to write prescription refills through a pilot launched in Utah, raised $40M led by Abstract and Lightspeed (Brian Gormley/Wall Street Journal)

    Brian Gormley / Wall Street Journal : Doctronic, which became the first company to use AI to write prescription refills through a pilot launched in Utah, raised $40M led by Abstract and Lightspeed —  Doctronic, which just raised $40 million, links patients to human doctors in virtual visits and has a pilot program that refills prescriptions.

  13. Gimlet Labs, which says it is the first "multi-silicon inference cloud" for running AI workloads across diverse types of hardware, raised an $80M Series A (Julie Bort/TechCrunch)

    Julie Bort / TechCrunch : Gimlet Labs, which says it is the first “multi-silicon inference cloud” for running AI workloads across diverse types of hardware, raised an $80M Series A —  Stanford adjunct professor and successfully exited founder Zain Asgar just raised an $80 million Series A for a startup …

  14. Interviews with Sundar Pichai and other Google executives on being blindsided by ChatGPT's launch, Gemini, Pichai's vision of useful AI everywhere, and more (Harry McCracken/Fast Company)

    Harry McCracken / Fast Company : Interviews with Sundar Pichai and other Google executives on being blindsided by ChatGPT's launch, Gemini, Pichai's vision of useful AI everywhere, and more —  Sundar Pichai was blindsided by ChatGPT.  Soon after being named Google CEO in 2015, he'd declared that the world was entering an AI-first era.

  15. Source: OpenAI is in talks to buy 5 GW of electricity by 2030 from Sam Altman-backed fusion startup Helion; Altman has stepped down as Helion's board chair (Ina Fried/Axios)

    Ina Fried / Axios : Source: OpenAI is in talks to buy 5 GW of electricity by 2030 from Sam Altman-backed fusion startup Helion; Altman has stepped down as Helion's board chair —  OpenAI is in advanced talks to buy electricity from Sam Altman-backed fusion startup Helion Energy, according to a person familiar with the situation.

Solidot(15)

  1. 石油能源危机推动向可再生能源的转型

    霍尔木兹海峡的封锁导致世界再次面临严重的石油能源危机。全球约五分之一石油和液化天然气的运输是经过霍尔木兹海峡,此次危机受影响最大的是亚洲地区。与之前的石油危机不同的是,在很多国家可再生能源已能与化石燃料展开竞争。两大人口大国中国和印度都扩大了可再生能源规模,中国仍然依赖燃煤发电,其可再生能源的规模远超印度。国际能源署的数据显示,中国约十分之一的汽车是电动汽车。中国仍然是世界最大的原油进口国,也是伊朗石油的最大买家。但通过可再生能源实现部分经济领域的电气化,中国已降低了对进口石油的依赖。如果没有这种转变,中国受到影响会更显著。印度目前正面临烹饪用燃气短缺问题,燃气短缺促使居民去抢购电磁炉。太阳能和风能只占日本能源产出的 11%,与印度持平,低于中国的 18%。巴基斯坦加速发展太阳能使该国自 2020 年以来减少进口化石燃料逾 120 亿美元。孟加拉国能源储备有限,该国已关闭大学以节省用电,政府开始实行燃料配给制。

  2. 烟头会在环境中停留十年以上

    根据发表在《Environmental Pollution》期刊上的一项研究,烟头不会完全从环境中消失。由于分解缓慢而且会释放出有毒物质,烟头构成了长期的环境危害。研究人员调查了烟头在十年中的分解过程,发现在富氮条件下烟头的质量会在十年里减少 84%。烟头的分解分为四个过程,初始阶段出现一个峰值,然后在中期再次出现一个峰值,显示旧烟头会带来持续的生态风险。

  3. 三星 Galaxy S26 支持 AirDrop

    在 Google Pixel 10 系列手机之后,三星宣布其 Galaxy S26 系列手机正式支持 AirDrop。Google Android 平台工程副总裁 Eric Kay 此前表示今年会有更多 Android 设备支持 AirDrop。苹果 iPhone 的 AirDrop 以及 Android 的 Quick Share 被用于快速在设备之间分享文件,Galaxy S26 系列包括三个型号 Galaxy S26、Galaxy S26 Plus 和 Galaxy S26 Ultra,对 AirDrop 的支持将首先在韩国推出,然后扩大到美国等其它地区。三星其它型号的智能手机未来也可能支持 AirDrop。

  4. 微软释出紧急更新修复微软账号登录问题

    微软释出了紧急更新 KB5085516,修复三月例行安全更新 KB5079473 释出后出现的微软账号登录问题,该问题影响使用微软账号登录的应用如 Teams、OneDrive、Microsoft Edge、Microsoft 365 Copilot 以及 Excel 和 Word 等 Office 应用,当用户通过微软账号登录这些应用会返回错误信息,声称用户未连接到互联网。微软表示,使用 Microsoft Entra ID 登录的企业客户未受影响。

  5. 龙芯工程师将维护其 DRM 驱动

    因无人维护,Linux 内核处理 LS7A/LS2K SoC 显示控制器的龙芯 DRM(Direct Rendering Manager)驱动被标记为“孤儿状态”。现在龙芯工程师在邮件列表上宣布接手维护工作。龙芯工程师 Jianmin Lv 和 Qianhai 都是龙芯的 GPU 研发工程师,负责内核驱动开发,他们有能力也有责任持续维护龙芯的 GPU 驱动,最小化对用户的影响,在内部讨论之后,团队推荐两人接手维护工作,推荐 Huacai、Mingcong 和 Ruoyao 三人协助。龙芯将根据芯片支持计划推出更新。

  6. 一篇推荐 RSS 阅读器的文章下载了 500 MB 的广告

    在被算法控制和广告轰炸的时代,RSS 阅读器能让我们控制自己阅读的内容而再次受到青睐。PC Gamer 网站发表了一篇推荐 RSS 阅读器的文章,然而讽刺的是网页本身充斥着广告。除了多则弹出式窗口,网页初步完成加载后其大小达到了 37MB,但此后网页还会在后台持续加载广告,在五分钟内它加载了 500MB 的广告。这就是我们需要 RSS 阅读器摆脱这一切的原因。

  7. 联合国警告地球气候愈发失衡

    根据世界气象组织(WMO)报告,随着温室气体浓度驱动的大气和海洋持续变暖以及冰层融化,地球气候失衡比观测史上任何时候都更加严重。WMO《2025年全球气候状况》报告确认:2015-2025 年是有记录以来最热的 11 年,2025 年是有记录以来第二热或第三热年份,比 1850-1900 年的平均水平约高出 1.43°C。世界各地的极端事件,包括酷热、暴雨和热带气旋,造成了干扰和破坏,凸显了互为关联的经济与社会的脆弱性。海洋在持续变暖并吸收二氧化碳。在过去的二十年里,海洋每年吸收的能量相当于人类每年所用能量的十八倍。报告称,北极的年度海冰范围处于或接近历史最低水平,南极海冰范围是有记录以来的第三低,冰川融化持续有增无减。联合国秘书长安东尼奥·古特雷斯表示,“人类刚刚经历了有记录以来最热的十一年。历史重演了十一次,这已不再是巧合。这是行动呼吁。”

  8. 报告称我国居民平均入睡时间 00:10

    中国睡眠研究会发布了《2026中国睡眠健康研究白皮书》。根据最新调研数据,与 2024 年相比,平均入睡时间在 00:10,提前 8 分钟;觉醒时间为 7:27,提前 4 分钟,夜间睡眠时长 6.97 小时,增加 7 分钟。报告还显示,年龄越大睡眠节律越规律,其中 66 岁及以上人群入睡规律占比 45%。睡眠规律的老年人,夜间清醒时长越短。学生群体睡眠最不规律,占比高达 69.6%。睡眠呼吸暂停的占比则随年龄增长而升高,66 岁及以上人群占比高达 26%,男性风险显著高于女性。而夜间睡眠时长不足 5 小时的人群,肥胖比例高达 41.4%。入睡时间越晚,肥胖风险越高,凌晨 2 点后入睡的人群,肥胖比例高达 18.4%。中年人尤为突出,睡眠不足 5 小时的人群中,中年人肥胖占比高达 46.2%。

  9. SE 将在《勇者斗恶龙X》中集成 Google Gemini

    史克威尔艾尼克斯(Square Enix)宣布将在其热门网游《勇者斗恶龙X》中集成 Google 的 Gemini 模型,允许玩家与其对话,为新手玩家提供导引方面的帮助。《勇者斗恶龙X》还将加入一个集成生成式 AI 的新伙伴角色 Chatty Slimey,当玩家使用聊天功能与其对话时,它会自动生成语音进行互动。如果玩家击败强敌或获得稀有道具,Slimey 还可能主动发起对话。《勇者斗恶龙X》于 2012 年上线,中国版于 2016 年上线,由盛大游戏运营,但在 2019 年就终止了服务。《勇者斗恶龙X》运营已超过十年,每月仍然有数十万活跃玩家,但如何吸引新玩家是其面临的一大挑战。

  10. 视觉小说数据库创始人去世

    视觉小说数据库(Visual Novel Database,简称 VNDB)创始人 Yoran Heling 于 3 月 17 日去世。2007 年网名 Yorhel 的 Yoran Heling 在游玩了《时空轮回》之后,惊讶于网络上没有一个供玩家交流视觉小说、寻找视觉小说的地方,故花费了三个星期的时间创立了 VNDB,作为玩家集中讨论、收集视觉小说的场所。在创立后的一年内,视觉小说数据库便已收录 1000 部视觉小说。目前 VNDB 收录了 61125 部视觉小说。网站背景图中的角色是《AS~天使小夜曲》的女主角“菈司蒂·珐姒”。

  11. 猪脑成功冷冻

    在保存大脑精微结构方面,时机至关重要。血液循环停止后数分钟内,酶就会分解神经元,细胞开始自我消化。 人体冷冻技术通常涉及将人体保存在零度以下,希望未来一旦出现针对其病情的治疗方法或疗法,能够使其复活。传统上,该技术旨在通过冷却和添加固定剂在自然死亡后快速保存大脑,但除非冷冻团队守候在患者床边,否则在此之前退化过程早已开始。为规避这一问题,专注于记忆保存的美国科技公 司Nectome 团队制定了一套与医生协助死亡相兼容的方案,即绝症患者可选择自己的离世时间。其理念在于,通过立即干预,科学家可能拥有最佳机会使大脑的保存状态尽可能接近活体状态。该团队在猪身上测试了这一方案,猪的大脑和心血管解剖结构与人类相当。首先,他们在心脏骤停约 1 分钟后将插管插入心脏,随后冲洗血液并将保存液引入大脑。这些液体含有醛类化学物质,可在细胞间形成分子桥,实质上将细胞活动锁定在原地。随后他们引入冷冻保护剂,置换组织中的水分,防止冷却过程中形成冰晶(否则会损害细胞)。接下来,大脑被冷却至约零下 32 摄氏度,在此温度下冷冻保护剂会形成玻璃态。此时大脑结构可近乎永久保存。

  12. 全球暖化加剧极端干旱热浪事件的发生频率

    21 世纪以来,干旱与热浪同时发生的极端天气事件频率显著增加。这些事件造成了严重的社会经济损失。例如 2010 年俄罗斯的干旱、热浪与野火同时发生,导致了 5.5 万人死亡;2019 年至 2020 年澳大利亚的森林大火与干旱与热浪同时发生导致了“黑夏”;2021 年 6 月太平洋西北地区的极端天气事件导致加拿大大不列颠哥伦比亚省和阿尔伯塔省春小麦产量下降 31%。根据发表在《Science Advances》的一项研究,自 2000 年以来,极端干旱热浪事件的发生频率增加了近 8 倍。全球气温每上升 1°C,极端天气事件发生频率从 1.6% 上升至 13.1%。

  13. Cloudflare 将 archive.today 归类为 C&C/Botnet

    在维基百科之后,提供 DDoS 保护、网页应用防火墙、公共 DNS 解析器、反向代理和 CDN 等服务的 Cloudflare 将 Archive.today 以及相关域名 archive.is、archive.ph 等加入到了 Command and Control & Botnet 类别,意味着该域名被 Cloudflare 的 1.1.1.2 停止解析。原因是 Archive.today 被发现劫持用户的浏览器对博主 Jani Patokallio 的个人博客 Gyrovague 发动了 DDoS 攻击。Patokallio 在澳大利亚工作,他是芬兰人,Archive.today 的运营者针对芬兰 IP 的 CAPTCHA 验证页面嵌入脚本对其博客发动攻击,至今该脚本仍然存在,对 Patokallio 的 DDoS 攻击仍然在进行之中。

  14. SystemD 加入可选的 birthDate 字段

    SystemD 项目合并了一个 pull request,为 userdb 管理的 JSON 用户记录添加了一个新的 birthDate 字段,此举旨在响应美国加州和科罗拉多州,以及巴西的年龄验证法律。systemd 作者 Lennart Poettering 强调该字段是完全可选的,只是定义了一个字段,在用户需要存储出生日期时将其标准化。他表示 systemd 本身不会处理出生日期相关的数据,也不会强制要求提供数据,systemd 并不执行年龄验证政策,而是由系统的其它部分决定。

  15. 慢性肾脏病可能会唤醒破坏大脑的病毒

    全世界可能多达九成的人感染了人类多瘤病毒 2——aka JC 病毒,以首位分离出该病毒的患者 John Cunningham 的名字首字母命名。除非被激活,对大部分人而言这种病毒不会有症状。如果激活,病毒会破坏大脑,导致“进行性多灶性白质脑病(缩写 PML)”。PML 型的 JC 病毒会破坏特定的脑细胞,导致神经细胞功能障碍和死亡。PML 被认为非常罕见,但在艾滋病出现之后,在流行早期有 2% 至 5% 的 HIV 感染者会同时出现 PML。这意味着艾滋病相关联的严重免疫抑制可能是 PML 的一种激活条件。现在,根据发表在《Annals of Internal Medicine Case》期刊上的一项研究,研究人员报告了 PML 的另一种可能的激活条件——慢性肾脏病。相比 HIV,慢性肾脏病影响全世界十分之一的人口,慢性肾脏病激活 PML 型 JC 病毒可能会带来严重后果,需要警惕。