DIGEST · 2026-04-14

OrangeBot.AI Digest — 2026-04-14

88 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Tell HN: Fiverr left customer files public and searchable
  2. YouTube now world's largest media company, topping Disney (www.hollywoodreporter.com)
  3. OpenSSL 4.0.0 (github.com)
  4. I wrote to Flock's privacy contact to opt out of their domestic spying program (honeypot.net)
  5. Claude Code Routines (code.claude.com)
  6. Spain to expand internet blocks to tennis, golf, movies broadcasting times (bandaancha.eu)
  7. Google, Microsoft, Meta All Tracking You Even When You Opt Out (www.404media.co)
  8. The future of everything is lies, I guess: Work (aphyr.com)
  9. Rare concert recordings are landing on the Internet Archive (techcrunch.com)
  10. jj – the CLI for Jujutsu (steveklabnik.github.io)
  11. Backblaze has stopped backing up OneDrive and Dropbox folders and maybe others (rareese.com)
  12. An AI Vibe Coding Horror Story (www.tobru.ch)
  13. Introspective Diffusion Language Models (introspective-diffusion.github.io)
  14. Sometimes powerful people just do dumb shit (www.joanwestenberg.com)
  15. Hacker compromises A16Z-backed phone farm, calling them the 'antichrist' (www.404media.co)

GitHub Trending(13)

  1. forrestchang / andrej-karpathy-skills
  2. thedotmack / claude-mem
  3. jamiepine / voicebox
  4. pascalorg / editor
  5. microsoft / markitdown
  6. obra / superpowers
  7. chrislgarry / Apollo-11
  8. virattt / ai-hedge-fund
  9. shiyu-coder / Kronos
  10. NousResearch / hermes-agent
  11. anthropics / claude-cookbooks
  12. tw93 / Mole
  13. shanraisshan / claude-code-best-practice

Product Hunt(15)

  1. ElevenAgents Guardrails 2.0

    Configurable safety control for enterprise agent deployment.

  2. Caveman

    Why use so many token when few do trick?

  3. Open Agents

    Agents that ship real code

  4. Hapax

    Watches your workflows. Builds your Agents. Automatically.

  5. Softr AI Co-Builder

    Build business apps with AI - that actually work

  6. CatDoes v4

    An AI agent with its own computer builds your apps

  7. Recall 2.0

    Curate an AI that knows what you know.

  8. FuseAI

    Close 10x more revenue with AI agents.

  9. Figma for Agents

    Design with AI agents, connected to your design system

  10. Ghost Pepper 🌶️

    100% local private AI for text-to-speech & meeting notes

  11. FlutterAIDev

    Generate flutter app using AI for Android and iOS

  12. AS Notes

    Turn your IDE into a personal knowledge management system

  13. Amadeus

    Learn Any Piano Song

  14. Ovren

    Your AI engineering department that ships your backlog

  15. PixDone

    Earn random pixel rewards every time you smash a task

Hugging Face(15)

  1. QuanBench+: A Unified Multi-Framework Benchmark for LLM-Based Quantum Code Generation

    Large Language Models (LLMs) are increasingly used for code generation, yet quantum code generation is still evaluated mostly within single frameworks, making it difficult to separate quantum reasoning from framework familiarity. We introduce QuanBench+, a unified benchmark spanning Qiskit, PennyLane, and Cirq, with 42 aligned tasks covering quantum algorithms, gate decomposition, and state preparation. We evaluate models with executable functional tests, report Pass@1 and Pass@5, and use KL-divergence-based acceptance for probabilistic outputs. We additionally study Pass@1 after feedback-based repair, where a model may revise code after a runtime error or wrong answer. Across frameworks, the strongest one-shot scores reach 59.5% in Qiskit, 54.8% in Cirq, and 42.9% in PennyLane; with feedback-based repair, the best scores rise to 83.3%, 76.2%, and 66.7%, respectively. These results show clear progress, but also that reliable multi-framework quantum code generation remains unsolved and still depends strongly on framework-specific knowledge.

  2. The Past Is Not Past: Memory-Enhanced Dynamic Reward Shaping

    Despite the success of reinforcement learning for large language models, a common failure mode is reduced sampling diversity, where the policy repeatedly generates similar erroneous behaviors. Classical entropy regularization encourages randomness under the current policy, but does not explicitly discourage recurrent failure patterns across rollouts. We propose MEDS, a Memory-Enhanced Dynamic reward Shaping framework that incorporates historical behavioral signals into reward design. By storing and leveraging intermediate model representations, we capture features of past rollouts and use density-based clustering to identify frequently recurring error patterns. Rollouts assigned to more prevalent error clusters are penalized more heavily, encouraging broader exploration while reducing repeated mistakes. Across five datasets and three base models, MEDS consistently improves average performance over existing baselines, achieving gains of up to 4.13 pass@1 points and 4.37 pass@128 points. Additional analyses using both LLM-based annotations and quantitative diversity metrics show that MEDS increases behavioral diversity during sampling.

  3. Attention Sink in Transformers: A Survey on Utilization, Interpretation, and Mitigation

    As the foundational architecture of modern machine learning, Transformers have driven remarkable progress across diverse AI domains. Despite their transformative impact, a persistent challenge across various Transformers is Attention Sink (AS), in which a disproportionate amount of attention is focused on a small subset of specific yet uninformative tokens. AS complicates interpretability, significantly affecting the training and inference dynamics, and exacerbates issues such as hallucinations. In recent years, substantial research has been dedicated to understanding and harnessing AS. However, a comprehensive survey that systematically consolidates AS-related research and offers guidance for future advancements remains lacking. To address this gap, we present the first survey on AS, structured around three key dimensions that define the current research landscape: Fundamental Utilization, Mechanistic Interpretation, and Strategic Mitigation. Our work provides a pivotal contribution by clarifying key concepts and guiding researchers through the evolution and trends of the field. We envision this survey as a definitive resource, empowering researchers and practitioners to effectively manage AS within the current Transformer paradigm, while simultaneously inspiring innovative advancements for the next generation of Transformers. The paper list of this work is available at https://github.com/ZunhaiSu/Awesome-Attention-Sink.

  4. OmniShow: Unifying Multimodal Conditions for Human-Object Interaction Video Generation

    In this work, we study Human-Object Interaction Video Generation (HOIVG), which aims to synthesize high-quality human-object interaction videos conditioned on text, reference images, audio, and pose. This task holds significant practical value for automating content creation in real-world applications, such as e-commerce demonstrations, short video production, and interactive entertainment. However, existing approaches fail to accommodate all these requisite conditions. We present OmniShow, an end-to-end framework tailored for this practical yet challenging task, capable of harmonizing multimodal conditions and delivering industry-grade performance. To overcome the trade-off between controllability and quality, we introduce Unified Channel-wise Conditioning for efficient image and pose injection, and Gated Local-Context Attention to ensure precise audio-visual synchronization. To effectively address data scarcity, we develop a Decoupled-Then-Joint Training strategy that leverages a multi-stage training process with model merging to efficiently harness heterogeneous sub-task datasets. Furthermore, to fill the evaluation gap in this field, we establish HOIVG-Bench, a dedicated and comprehensive benchmark for HOIVG. Extensive experiments demonstrate that OmniShow achieves overall state-of-the-art performance across various multimodal conditioning settings, setting a solid standard for the emerging HOIVG task.

  5. Strips as Tokens: Artist Mesh Generation with Native UV Segmentation

    Recent advancements in autoregressive transformers have demonstrated remarkable potential for generating artist-quality meshes. However, the token ordering strategies employed by existing methods typically fail to meet professional artist standards, where coordinate-based sorting yields inefficiently long sequences, and patch-based heuristics disrupt the continuous edge flow and structural regularity essential for high-quality modeling. To address these limitations, we propose Strips as Tokens (SATO), a novel framework with a token ordering strategy inspired by triangle strips. By constructing the sequence as a connected chain of faces that explicitly encodes UV boundaries, our method naturally preserves the organized edge flow and semantic layout characteristic of artist-created meshes. A key advantage of this formulation is its unified representation, enabling the same token sequence to be decoded into either a triangle or quadrilateral mesh. This flexibility facilitates joint training on both data types: large-scale triangle data provides fundamental structural priors, while high-quality quad data enhances the geometric regularity of the outputs. Extensive experiments demonstrate that SATO consistently outperforms prior methods in terms of geometric quality, structural coherence, and UV segmentation.

  6. Uni-ViGU: Towards Unified Video Generation and Understanding via A Diffusion-Based Video Generator

    Unified multimodal models integrating visual understanding and generation face a fundamental challenge: visual generation incurs substantially higher computational costs than understanding, particularly for video. This imbalance motivates us to invert the conventional paradigm: rather than extending understanding-centric MLLMs to support generation, we propose Uni-ViGU, a framework that unifies video generation and understanding by extending a video generator as the foundation. We introduce a unified flow method that performs continuous flow matching for video and discrete flow matching for text within a single process, enabling coherent multimodal generation. We further propose a modality-driven MoE-based framework that augments Transformer blocks with lightweight layers for text generation while preserving generative priors. To repurpose generation knowledge for understanding, we design a bidirectional training mechanism with two stages: Knowledge Recall reconstructs input prompts to leverage learned text-video correspondences, while Capability Refinement fine-tunes on detailed captions to establish discriminative shared representations. Experiments demonstrate that Uni-ViGU achieves competitive performance on both video generation and understanding, validating generation-centric architectures as a scalable path toward unified multimodal intelligence. Project Page and Code: https://fr0zencrane.github.io/uni-vigu-page/.

  7. Pseudo-Unification: Entropy Probing Reveals Divergent Information Patterns in Unified Multimodal Models

    Unified multimodal models (UMMs) were designed to combine the reasoning ability of large language models (LLMs) with the generation capability of vision models. In practice, however, this synergy remains elusive: UMMs fail to transfer LLM-like reasoning to image synthesis and exhibit divergent response behaviors. We term this phenomenon pseudo-unification. Diagnosing its internal causes is important, but existing probing methods either lack model-internal insight or ignore prompt-response dependencies. To address these limitations, we propose an information-theoretic probing framework that jointly analyzes how UMMs encode inputs and generate outputs. Applied to ten representative UMMs, our framework reveals that pseudo-unification stems from a dual divergence: (i) Modality-Asymmetric Encoding, where vision and language follow different entropy trajectories, and (ii) Pattern-Split Response, where text generation exhibits high-entropy creativity while image synthesis enforces low-entropy fidelity. Only models that unify both sides (e.g., via contextual prediction) achieve more genuine unification, enabling stronger reasoning-based text-to-image generation even with fewer parameters. Our work provides the first model-internal probing of unification, demonstrating that real multimodal synergy requires consistency in information flow, not just shared parameters.

  8. CodeTracer: Towards Traceable Agent States

    Code agents are advancing rapidly, but debugging them is becoming increasingly difficult. As frameworks orchestrate parallel tool calls and multi-stage workflows over complex tasks, making the agent's state transitions and error propagation hard to observe. In these runs, an early misstep can trap the agent in unproductive loops or even cascade into fundamental errors, forming hidden error chains that make it hard to tell when the agent goes off track and why. Existing agent tracing analyses either focus on simple interaction or rely on small-scale manual inspection, which limits their scalability and usefulness for real coding workflows. We present CodeTracer, a tracing architecture that parses heterogeneous run artifacts through evolving extractors, reconstructs the full state transition history as a hierarchical trace tree with persistent memory, and performs failure onset localization to pinpoint the failure origin and its downstream chain. To enable systematic evaluation, we construct CodeTraceBench from a large collection of executed trajectories generated by four widely used code agent frameworks on diverse code tasks (e.g., bug fixing, refactoring, and terminal interaction), with supervision at both the stage and step levels for failure localization. Experiments show that CodeTracer substantially outperforms direct prompting and lightweight baselines, and that replaying its diagnostic signals consistently recovers originally failed runs under matched budgets. Our code and data are publicly available.

  9. CocoaBench: Evaluating Unified Digital Agents in the Wild

    LLM agents now perform strongly in software engineering, deep research, GUI automation, and various other applications, while recent agent scaffolds and models are increasingly integrating these capabilities into unified systems. Yet, most evaluations still test these capabilities in isolation, which leaves a gap for more diverse use cases that require agents to combine different capabilities. We introduce CocoaBench, a benchmark for unified digital agents built from human-designed, long-horizon tasks that require flexible composition of vision, search, and coding. Tasks are specified only by an instruction and an automatic evaluation function over the final output, enabling reliable and scalable evaluation across diverse agent infrastructures. We also present CocoaAgent, a lightweight shared scaffold for controlled comparison across model backbones. Experiments show that current agents remain far from reliable on CocoaBench, with the best evaluated system achieving only 45.1% success rate. Our analysis further points to substantial room for improvement in reasoning and planning, tool use and execution, and visual grounding.

  10. Tracing the Roots: A Multi-Agent Framework for Uncovering Data Lineage in Post-Training LLMs

    Post-training data plays a pivotal role in shaping the capabilities of Large Language Models (LLMs), yet datasets are often treated as isolated artifacts, overlooking the systemic connections that underlie their evolution. To disentangle these complex relationships, we introduce the concept of data lineage to the LLM ecosystem and propose an automated multi-agent framework to reconstruct the evolutionary graph of dataset development. Through large-scale lineage analysis, we characterize domain-specific structural patterns, such as vertical refinement in math-oriented datasets and horizontal aggregation in general-domain corpora. Moreover, we uncover pervasive systemic issues, including structural redundancy induced by implicit dataset intersections and the propagation of benchmark contamination along lineage paths. To demonstrate the practical value of lineage analysis for data construction, we leverage the reconstructed lineage graph to create a lineage-aware diversity-oriented dataset. By anchoring instruction sampling at upstream root sources, this approach mitigates downstream homogenization and hidden redundancy, yielding a more diverse post-training corpus. We further highlight lineage-centric analysis as an efficient and robust topological alternative to sample-level dataset comparison for large-scale data ecosystems. By grounding data construction in explicit lineage structures, our work advances post-training data curation toward a more systematic and controllable paradigm.

  11. Introspective Diffusion Language Models

    Diffusion language models promise parallel generation, yet still lag behind autoregressive (AR) models in quality. We stem this gap to a failure of introspective consistency: AR models agree with their own generations, while DLMs often do not. We define the introspective acceptance rate, which measures whether a model accepts its previously generated tokens. This reveals why AR training has a structural advantage: causal masking and logit shifting implicitly enforce introspective consistency. Motivated by this observation, we introduce Introspective Diffusion Language Model (I-DLM), a paradigm that retains diffusion-style parallel decoding while inheriting the introspective consistency of AR training. I-DLM uses a novel introspective strided decoding (ISD) algorithm, which enables the model to verify previously generated tokens while advancing new ones in the same forward pass. From a systems standpoint, we build I-DLM inference engine on AR-inherited optimizations and further customize it with a stationary-batch scheduler. To the best of our knowledge, I-DLM is the first DLM to match the quality of its same-scale AR counterpart while outperforming prior DLMs in both model quality and practical serving efficiency across 15 benchmarks. It reaches 69.6 on AIME-24 and 45.7 on LiveCodeBench-v6, exceeding LLaDA-2.1-mini (16B) by more than 26 and 15 points, respectively. Beyond quality, I-DLM is designed for the growing demand of large-concurrency serving, delivering about 3x higher throughput than prior state-of-the-art DLMs.

  12. Audio Flamingo Next: Next-Generation Open Audio-Language Models for Speech, Sound, and Music

    We present Audio Flamingo Next (AF-Next), the next-generation and most capable large audio-language model in the Audio Flamingo series, designed to advance understanding and reasoning over speech, environmental sounds and music. Compared to Audio Flamingo 3, AF-Next introduces: (i) a stronger foundational audio-language model that significantly improves accuracy across diverse audio understanding tasks; (ii) scalable strategies for constructing large-scale audio understanding and reasoning data beyond existing academic benchmarks; (iii) support for long and complex audio inputs up to 30 minutes; and (iv) Temporal Audio Chain-of-Thought, a new reasoning paradigm that explicitly grounds intermediate reasoning steps to timestamps in long audio, enabling fine-grained temporal alignment and improved interpretability. To enable these capabilities, we first conduct a systematic analysis of Audio Flamingo 3 to identify key gaps in audio understanding and reasoning. We then curate and scale new large-scale datasets totaling over 1 million hours to address these limitations and expand the existing AudioSkills-XL, LongAudio-XL, AF-Think and AF-Chat datasets. AF-Next is trained using a curriculum-based strategy spanning pre-training, mid-training and post-training stages. Extensive experiments across 20 audio understanding and reasoning benchmarks, including challenging long-audio tasks, show that AF-Next outperforms similarly sized open models by large margins and remains highly competitive with and sometimes surpasses, much larger open-weight and closed models. Beyond benchmark performance, AF-Next exhibits strong real-world utility and transfers well to unseen tasks, highlighting its robustness and generalization ability. In addition to all data, code and methods, we open-source 3 variants of AF-Next, including AF-Next-Instruct, AF-Next-Think and AF-Next-Captioner.

  13. Solving Physics Olympiad via Reinforcement Learning on Physics Simulators

    We have witnessed remarkable advances in LLM reasoning capabilities with the advent of DeepSeek-R1. However, much of this progress has been fueled by the abundance of internet question-answer (QA) pairs, a major bottleneck going forward, since such data is limited in scale and concentrated mainly in domains like mathematics. In contrast, other sciences such as physics lack large-scale QA datasets to effectively train reasoning-capable models. In this work, we show that physics simulators can serve as a powerful alternative source of supervision for training LLMs for physical reasoning. We generate random scenes in physics engines, create synthetic question-answer pairs from simulated interactions, and train LLMs using reinforcement learning on this synthetic data. Our models exhibit zero-shot sim-to-real transfer to real-world physics benchmarks: for example, training solely on synthetic simulated data improves performance on IPhO (International Physics Olympiad) problems by 5-10 percentage points across model sizes. These results demonstrate that physics simulators can act as scalable data generators, enabling LLMs to acquire deep physical reasoning skills beyond the limitations of internet-scale QA data. Code available at: https://sim2reason.github.io/.

  14. Agentic Aggregation for Parallel Scaling of Long-Horizon Agentic Tasks

    We study parallel test-time scaling for long-horizon agentic tasks such as agentic search and deep research, where multiple rollouts are generated in parallel and aggregated into a final response. While such scaling has proven effective for chain-of-thought reasoning, agentic tasks pose unique challenges: trajectories are long, multi-turn, and tool-augmented, and outputs are often open-ended. Aggregating only final answers discards rich information from trajectories, while concatenating all trajectories exceeds the model's context window. To address this, we propose AggAgent, an aggregation agent that treats parallel trajectories as an environment. We equip it with lightweight tools to inspect candidate solutions and search across trajectories, enabling it to navigate and synthesize information on demand. Across six benchmarks and three model families (GLM-4.7, Qwen3.5, MiniMax-M2.5), AggAgent outperforms all existing aggregation methods-by up to 5.3% absolute on average and 10.3% on two deep research tasks-while adding minimal overhead, as the aggregation cost remains bounded by a single agentic rollout. Our findings establish agentic aggregation as an effective and cost-efficient approach to parallel test-time scaling.

  15. Prompt Relay: Inference-Time Temporal Control for Multi-Event Video Generation

    Video diffusion models have achieved remarkable progress in generating high-quality videos. However, these models struggle to represent the temporal succession of multiple events in real-world videos and lack explicit mechanisms to control when semantic concepts appear, how long they persist, and the order in which multiple events occur. Such control is especially important for movie-grade video synthesis, where coherent storytelling depends on precise timing, duration, and transitions between events. When using a single paragraph-style prompt to describe a sequence of complex events, models often exhibit semantic entanglement, where concepts intended for different moments in the video bleed into one another, resulting in poor text-video alignment. To address these limitations, we propose Prompt Relay, an inference-time, plug-and-play method to enable fine-grained temporal control in multi-event video generation, requiring no architectural modifications and no additional computational overhead. Prompt Relay introduces a penalty into the cross-attention mechanism, so that each temporal segment attends only to its assigned prompt, allowing the model to represent one semantic concept at a time and thereby improving temporal prompt alignment, reducing semantic interference, and enhancing visual quality.

Techmeme(15)

  1. Meta and Broadcom announce an expanded partnership to co-develop multiple generations of Meta's MTIA chips; Broadcom CEO Hock Tan plans to leave Meta's board (CNBC)

    CNBC : Meta and Broadcom announce an expanded partnership to co-develop multiple generations of Meta's MTIA chips; Broadcom CEO Hock Tan plans to leave Meta's board —  Meta and Broadcom on Tuesday announced a sweeping deal that extends an existing partnership between the two companies for the design …

  2. The FCC grants Netgear a conditional approval to import its future consumer routers, cable modems, and cable gateways into the US through October 1, 2027 (Sean Hollister/The Verge)

    Sean Hollister / The Verge : The FCC grants Netgear a conditional approval to import its future consumer routers, cable modems, and cable gateways into the US through October 1, 2027 —  Make it make sense. Make it make sense. … The United States' foreign router ban didn't make a whole lot of sense, and today may not change that.

  3. US-based Credo, which specializes in data center connectivity, agrees to acquire Israeli chip company DustPhotonics in a cash-and-stock deal worth up to $1.3B (CTech)

    CTech : US-based Credo, which specializes in data center connectivity, agrees to acquire Israeli chip company DustPhotonics in a cash-and-stock deal worth up to $1.3B —  The Israeli company's photonic chip technology enables faster, lower-cost data transfer in next-generation AI clusters.

  4. AWS launches Amazon Bio Discovery, a new AI-powered application designed to speed up drug development, giving scientists access to biological foundation models (Reuters)

    Reuters : AWS launches Amazon Bio Discovery, a new AI-powered application designed to speed up drug development, giving scientists access to biological foundation models —  Amazon's (AMZN.O) cloud unit on Tuesday launched Amazon Bio Discovery, an artificial intelligence application designed …

  5. Users accuse Anthropic of degrading the performance of Claude Opus 4.6 and Claude Code; employees publicly deny the company degrades models to manage capacity (Carl Franzen/VentureBeat)

    Carl Franzen / VentureBeat : Users accuse Anthropic of degrading the performance of Claude Opus 4.6 and Claude Code; employees publicly deny the company degrades models to manage capacity —  A growing number of developers and AI power users are taking to social media to accuse Anthropic of degrading the performance …

  6. Kraken co-CEO Arjun Sethi says the crypto exchange has confidentially filed for a US IPO; it was valued at $13.3B this month, down from a $20B peak in late 2025 (Cory Schouten/Semafor)

    Cory Schouten / Semafor : Kraken co-CEO Arjun Sethi says the crypto exchange has confidentially filed for a US IPO; it was valued at $13.3B this month, down from a $20B peak in late 2025 —  The US crypto exchange Kraken has confidentially filed for an initial public offering, co-CEO Arjun Sethi said Tuesday …

  7. OpenAI rolls out GPT-5.4-Cyber, which is fine-tuned for additional cybersecurity use cases, to some participants of its Trusted Access for Cyber program (Rachel Metz/Bloomberg)

    Rachel Metz / Bloomberg : OpenAI rolls out GPT-5.4-Cyber, which is fine-tuned for additional cybersecurity use cases, to some participants of its Trusted Access for Cyber program —  OpenAI is letting a select group of users access a new artificial intelligence model that's meant to be more adept …

  8. Anthropic redesigns Claude Code on desktop, adding a sidebar for managing multiple sessions, a drag-and-drop layout, an integrated terminal, and a file editor (Claude)

    Claude : Anthropic redesigns Claude Code on desktop, adding a sidebar for managing multiple sessions, a drag-and-drop layout, an integrated terminal, and a file editor —  Today, we're releasing a redesign of the Claude Code desktop app, built to help you run more Claude Code tasks at once.

  9. Source: Anthropic is preparing to release Claude Opus 4.7, along with a new AI-powered tool for designing websites and presentations, as soon as this week (Stephanie Palazzolo/The Information)

    Stephanie Palazzolo / The Information : Source: Anthropic is preparing to release Claude Opus 4.7, along with a new AI-powered tool for designing websites and presentations, as soon as this week —  Anthropic is preparing its next flagship model, Claude Opus 4.7, along with a new AI-powered tool for designing websites and presentations …

  10. Nvidia stock rose 18%+ over the past ten days, its longest winning streak since 2023; Jensen Huang said in March that Nvidia has $1T of GPU orders through 2027 (Katie Tarasov/CNBC)

    Katie Tarasov / CNBC : Nvidia stock rose 18%+ over the past ten days, its longest winning streak since 2023; Jensen Huang said in March that Nvidia has $1T of GPU orders through 2027 —  Nvidia stock is on a tear, rising more than 18% over the past ten days.  It's the longest winning streak the artificial intelligence chip giant …

  11. Bluefish, which helps brands manage visibility across AI platforms such as ChatGPT and Claude, raised a $43M Series B, bringing its total funding to $68M (Trishla Ostwal/Adweek)

    Trishla Ostwal / Adweek : Bluefish, which helps brands manage visibility across AI platforms such as ChatGPT and Claude, raised a $43M Series B, bringing its total funding to $68M —  Series B brings total funding to $68 million as brands rethink AI visibility  —  If you want to shape media strategy—not just optimize it—upgrade your expertise.

  12. Microsoft agrees to rent 30,000 Nvidia Vera Rubin chips from Nscale at a site in Norway that was initially intended for OpenAI and marketed as part of Stargate (Bloomberg)

    Bloomberg : Microsoft agrees to rent 30,000 Nvidia Vera Rubin chips from Nscale at a site in Norway that was initially intended for OpenAI and marketed as part of Stargate —  Microsoft Corp. has agreed to rent data center capacity at a site in Norway that was initially intended for OpenAI and marketed …

  13. Anthropic launches a repeatable routines feature for Claude Code as a research preview, allowing developers to schedule and automate software development tasks (Zac Hall/9to5Mac)

    Zac Hall / 9to5Mac : Anthropic launches a repeatable routines feature for Claude Code as a research preview, allowing developers to schedule and automate software development tasks —  Anthropic's Claude Code has a new repeatable routines feature that works even when your Mac is offline.

  14. Google launches Skills, repeatable AI prompts users can run in Chrome with a keyboard shortcut; users can set up their own Skills or choose from 50+ presets (Reece Rogers/Wired)

    Reece Rogers / Wired : Google launches Skills, repeatable AI prompts users can run in Chrome with a keyboard shortcut; users can set up their own Skills or choose from 50+ presets —  The premade Skills available through the Gemini sidebar in Chrome include ways to maximize protein in recipes or summarize YouTube videos.

  15. Glydways, a robocar startup backed by Sam Altman, Khosla, and others, says it is in talks to raise $250M at a $1B+ valuation, following a ~$170M Series C (Min-Jeong Lee/Bloomberg)

    Min-Jeong Lee / Bloomberg : Glydways, a robocar startup backed by Sam Altman, Khosla, and others, says it is in talks to raise $250M at a $1B+ valuation, following a ~$170M Series C —  Glydways Inc., the robocar startup backed by Sam Altman, is in talks to raise an additional $250 million on the heels …

Solidot(15)

  1. Google 将惩罚“后退按钮劫持”行为

    今天的很多网站不让用户“后退”,但到了 6 月 15 日,如果网站还这么做,Google 将会进行惩罚,大幅降低网站的搜索排名。Google 将把这种被称为“后退按钮劫持”的做法定性为恶意行为。“后退按钮劫持”旨在强行将用户留在网站以增加流量,当访客试图通过后退按钮返回上一页,网站会篡改页面浏览历史记录,在用户点击后退按钮时插入其他内容。Google 表示,后退按钮应该始终执行用户预期的功能——返回上一页,任何其他行为都属于一种欺骗性的用户体验。

  2. 德国主权科技基金向 Mastodon 资助 61.4 万欧元

    德国主权科技基金(Sovereign Tech Fund)向联邦宇宙微博客项目 Mastodon 资助 61.4 万欧元,用于支持 Mastodon 及其软件生态系统的改进和更新。这笔资金将投入到改进:黑名单同步;新的 Fediverse Auxiliary Service Provider(FASP)允许服务器之间共享存储和媒体处理资源;自动化内容检测;私信端到端加密;改进文档。相关改进预计在 2026 年底到 2027 年完成。

  3. OpenSSL 4.0 释出

    OpenSSL 项目释出了 v4.0 版本。主要新特性包括:支持 Encrypted Client Hello (ECH),通过加密初始 TLS 握手以及隐藏服务器名称指示(SNI)提供更好的隐私保护;移除 SSLv3 等旧协议/引擎支持;通过支持 RFC 8998 改进后量子加密支持;移除 SSLv2 Client Hello,停止支持 Darwin i386 和 PowerPC/PPC64 等。

  4. Servo 发布首个 crates.io 版本

    Rust 语言开发的浏览器渲染引擎项目 Servo 释出了 servo crate v0.1.0,这是它发布的首个 crates.io 版本,允许 Servo 作为一个库供其他项目使用。Servo 源自 Mozilla,2020 年 8 月 Mozilla 在裁员时砍掉了 Servo 引擎团队的大部分成员。Servo 项目之后脱离 Mozilla 成为一个独立项目,由 Linux 基金会托管,旨在为其它项目提供一个嵌入的高性能的、安全的渲染引擎。Servo 项目于 2025 年 10 月释出了 v0.0.1 版本,之后以每月发布一个新版本的频率发布。开发者表示他们计划以每半年更新一次的频率提供长期支持版本(LTS),因为嵌入开发者可能无法及时更新到最新 Servo 版本,他们更适合使用 LTS 版本。LTS 版本预计将提供九个月的安全更新。

  5. 斯坦福的 AI 报告认为中美差距微乎其微

    斯坦福大学研究院 Institute for Human-Centered Artificial Intelligence(HAI)发布了年度报告 AI Index,报告认为中国顶级 AI 与美国 AI 相差无几。2024 年 1 月美国顶级 AI 的得分比中国顶级 AI 高 10% 左右,到 2026 年 3 月美国 Anthropic 和字节跳动的 AI 得分差距仅为 2.7%。在衡量语言、数学和编程领域难题正确率的基准测试中,差距也在缩小,中美之间的性能差距已基本消除。在开发和运营数据中心数量方面,美国有 5427 个遥遥领先于其他国家,2025 年民间投资额美国也以 2859 亿美元遥遥领先其他国家。中国的民间投资仅为 124 亿美元,但政府投资较大,实际投资额尚不明确。在被引用最多的前 100 篇论文中,中国的论文在 2024 年达到 41 篇,比上年增加 7 篇,缩小了与排名第一的美国(46 篇)的差距。

  6. 人类止痛药对龙虾有效

    根据发表在《Scientific Reports》期刊上的一项研究,龙虾能感受到疼痛,而人类止痛药也能帮助它止痛。这项研究再次表明需要为甲壳类动物开发出更人道的宰杀方法。新研究发现,当挪威龙虾在水中遭受电击时它们会快速摆动尾巴试图逃脱。但如果用止痛药阿司匹林和利多卡因预先为龙虾镇痛,尾巴摆动会减少甚至消失。论文合作者 Lynne Sneddon 称人类止痛药对挪威龙虾也有效,这表明人类生理功能与龙虾十分相似。人类应像对待鸡和牛一样重视对甲壳类动物的饲养和宰杀方式。

  7. 含氟自来水对 IQ 和大脑功能没有影响

    根据发表在 PNAS 期刊上的一项长期研究,儿童时期饮用含氟自来水对直至 80 岁的智商和大脑功能没有影响。美国反疫苗的卫生部长 Robert F. Kennedy Jr.此前宣称饮用水中的氟化物与 IQ 下降相关。美国多个红州——其中包括犹他州和佛罗里达州——已经颁布了禁令禁止在饮用水中添加氟化物,理由是担心氟化物导致 IQ 下降。肯塔基州、路易斯安那州、密苏里州和俄克拉荷马州等也在审议类似的立法。反对饮用水中添加氟化物的人士引用的研究多来自中国等国,这些国家饮用水中的氟化物浓度远高于美国,相关研究认为氟化物可能与儿童 IQ 有关。根据美国 CDC 的数据,预防龋齿的饮用水最优氟化物浓度为每升 0.7 毫克,相当于在 55 加仑的水桶中滴入 3 滴。美国饮用水法定氟化物浓度上限为每升 4.0 毫克。

  8. 31 个 WordPress 插件被收购后植入了后门

    一个印度开发团队以 Essential Plugin 的名义开发了 31 款 WordPress 插件,插件有免费版本也有付费版本。2024 年底由于收入下降了 35-45% 开发者将所有插件出售给了一个有 SEO、加密货币和赌博背景的买家,金额是六位数。2025 年 8 月 8 日买家释出了更新,其中包含了后门,但后门一直处于休眠状态。2026 年 4 月 5-6 日后门激活开始向所有运行相关插件的网站传播恶意载荷。恶意代码从指令控制服务器获取垃圾链接、重定向和虚假页面。这些垃圾信息只会显示给 Google 的机器人 Googlebot,对网站所有者不可见。WordPress.org 插件团队次日关闭了所有插件,但 SEO 垃圾信息注入攻击仍在进行中。这不是 WordPress.org 第一次遭遇模式相似的供应链攻击——收购信任插件然后注入恶意代码,它没有机制标记或审查插件所有权转移,也没有向用户发出插件所有权变更的通知,新所有者也不会触发额外的代码审查。最新攻击有数十万安装这些插件的网站受到影响。

  9. FBI 搜查朝 Sam Altman 住宅扔燃烧瓶的男子家

    OpenAI CEO Sam Altman 过去几天遭遇了两次袭击。第一次是一名男子朝他在旧金山的住宅扔燃烧瓶,并在 OpenAI 总部大楼前发口头威胁;第二次是有人在汽车内朝 Altman 的家开枪。三位嫌疑人都遭到了逮捕。扔燃烧瓶的嫌疑人是 20 岁的 Daniel Moreno-Gama,FBI 搜查了他在德州的家。Moreno-Gama 还被发现在 Substack 上撰写博客表达对 AI 的担忧以及反对 AI 高管。他还是 Discord 服务器 PauseAI 的成员,该组织是一个致力于禁止开发最强大的 AI 模型以保护公众的激进组织。Moreno-Gama 被捕时携带了一份文件,表达对 AI 以及 AI 公司高管的反对立场,他本人此前并没有犯罪记录。

  10. 黑客入侵 a16z 投资的手机农场,试图让手机农场账号发帖称 a16z 是反基督

    一名黑客入侵了硅谷风投 a16z 投资的手机农场 Doublespeed,该公司使用 AI 生成的 TikTok 账号创建虚假网红、生成视频以及发评论。黑客试图控制 Doublespeed 的社交账号发梗图声称 a16z 是“反基督”,图像包含了 a16z 联合创始人、特朗普支持者 Marc Andreessen。Doublespeed 联合创始人 Zuhair Lakhani 称他们已经迅速采取行动阻止了这次未经授权的访问,该公司的社交账号没有发布未经授权的帖子。Doublespeed 从 a16z 获得了 100 万美元的投资。

  11. 扎克伯格可能很快会有他的 AI 克隆

    FT 报道,Meta 正在构建一个 AI 版本的扎克伯格(Mark Zuckerberg),代替真人与员工互动。报道援引知情人士的消息称,这是该公司目前的优先事项,扎克伯格本人亲自参与了 AI 的训练和测试。AI 的训练内容包括他的举止、语气和公开发表的声明,以及近期对公司战略的思考,以便员工能通过与其互动感受到与创始人更紧密的联系。知情人士称,这项工作的重点之一是制作逼真的虚拟 AI 角色,因为需要大量的算力才能实现逼真的效果以及避免在与用户交互时出现延迟,因此扩大规模存在困难之处。如果实验成功的话,未来网红和内容创作者也可以采用这项技术。

  12. 计算机科学的黄金期可能已结束

    2025 年秋季美国四年制大学计算机科学专业的学生入学人数下降了 8.1%。计算机科学专业的本科排名在一年内从第四位跌至第六位,前三则一直是商科、公共卫生和人文科学。从 2008 年到 2024 年,计算机科学一直是美国增长最快的专业,如今它的黄金期可能已经结束。美国主修计算机科学的人数比上一学年少了 54000 人。那么他们选择了什么新专业?数据分析和数据科学招生总人数逾 3.5 万人,而 2020 年它们刚拆分出来时只招了几百人。数据显示,部分有意计算机科学专业的学生转向了相关领域如机器人学。工程专业学生入学人数 2025 年秋季增长了 7.3%,其中增长最快的两个专业是机械工程和电气工程专业,分别增长了 11% 和 14%。大学教授认为由于计算机科学毕业人数供过于求,学生们可能认为机械工程专业更通用,能在 AI 驱动的世界里提供更好的就业机会,如机器人、无人机、航空航天和电动汽车等行业。

  13. 《传送门2社区版》将于 4 月 18 日公测

    《传送门2(Portal 2)》自 2011 年 4 月发布至今已有 15 年,期间模组开发者推出了多个衍生版本,包括《Portal Stories: Mel》、《Aperture Tag: The Paint Gun Testing Initiative》以及《Portal Reloaded》等,现在由 P2:CE Team 开发的最新社区版本《传送门2社区版(Portal 2: Community Edition)》将于 4 月 18 日公测。《社区版》升级了引擎,使用了官方授权、基于 CS:GO 的 Source 引擎重度修改版本 Strata Source,原生支持 64 位改进了性能,新增原生 DirectX 11 支持,移除了旧引擎的很多限制,更新或改进了游戏玩法,提供了允许玩家轻松扩展游戏机制的脚本框架 AngelScript,采用了 Source 2 引擎的 Panorama UI 框架等等。《社区版》将免费提供给现有的《传送门2》玩家。

  14. 长期接触农药可能诱发糖尿病

    2023 年全球农药使用量达 373 万吨,约为 1990 年的两倍。农药相关健康风险研究长期集中在急性中毒、神经毒性和癌症方面。新型基因技术如今已能用于追踪农药对肠道菌群的影响。印度团队对印度南部近 3000 人开展研究后发现,城市地区 23% 的人患有糖尿病,多与肥胖、高胆固醇等典型危险因素相关;但农村地区糖尿病患病率仍高达 16%,且与这些危险因素无关。研究人员怀疑环境化学物质可能发挥了作用研究团队在小鼠身上研究了一种广泛使用的农业杀虫剂——氯氰菊酯的影响。根据印度日常饮食中的农药残留量,研究团队采用了“现实剂量”,持续给药 120 天。研究显示,氯氰菊酯重塑了小鼠肠道菌群,其中乳酸杆菌等有益菌数量下降,幽门螺杆菌等潜在有害菌增多。即便体重没有增加,接触氯氰菊酯的小鼠仍出现了高血糖和糖尿病症状。农药似乎不仅会改变菌群种类,还会影响其代谢活性。在另一项大型研究中,研究人员将 17 种人体肠道代表性细菌暴露于 18 种不同农药,检测到微生物产生的数百种小分子物质发生变化,其中包括短链脂肪酸、胆汁酸和色氨酸相关分子。这些物质能维持肠道黏膜健康、调节炎症反应、调控免疫功能。他们还发现,部分细菌会在细胞内蓄积农药,这可能延长其在人体内的停留时间,增加长期健康风险。

  15. Google Play 下架《心跳文学部》

    Google Play 下架了《心跳文学部(Doki Doki Literature Club)》,理由是游戏内容违反了与敏感主题相关的服务条款。作者 Dan Salvato 在一份声明中表示在致力于让游戏在 Google Play 重新上架。《心跳文学部》描述了一位男高中生加入学校的文学部与四位女性成员交流的故事,看起来是一个简单的恋爱视觉小说,但在完成一个结局之后故事会变得非常古怪,游戏会通过删除文件和存档的方式打破第四面墙。《心跳文学部》的免费版本积累了逾千万下载量,是 Steam 平台排名第一的心理恐怖游戏。关于敏感主题《心跳文学部》会在启动之后发出多次警告。《心跳文学部》有 iOS、Nintendo Switch、PlayStation 等各种版本。