DIGEST · 2026-04-22

OrangeBot.AI Digest — 2026-04-22

87 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. We found a stable Firefox identifier linking all your private Tor identities (fingerprint.com)
  2. Over-editing refers to a model modifying code beyond what is necessary (nrehiew.github.io)
  3. 5x5 Pixel font for tiny screens (maurycyz.com)
  4. Alberta startup sells no-tech tractors for half price (wheelfront.com)
  5. Qwen3.6-27B: Flagship-Level Coding in a 27B Dense Model (qwen.ai)
  6. Scoring Show HN submissions for AI design patterns (www.adriankrebs.ch)
  7. Our eighth generation TPUs: two chips for the agentic era (blog.google)
  8. 3.4M Solar Panels (tech.marksblogg.com)
  9. GitHub CLI now collects pseudoanonymous telemetry (cli.github.com)
  10. Irony as Meta staff unhappy about running surveillance software on work PCs (www.theregister.com)
  11. How does GPS work? (perthirtysix.com)
  12. Windows 9x Subsystem for Linux (social.hails.org)
  13. Making RAM at Home [video] (www.youtube.com)
  14. FBI looks into dead or missing scientists tied to NASA, Blue Origin, SpaceX (fortune.com)
  15. Tell HN: I'm sick of AI everything

GitHub Trending(12)

  1. zilliztech / claude-context
  2. Fincept-Corporation / FinceptTerminal
  3. koala73 / worldmonitor
  4. langfuse / langfuse
  5. KeygraphHQ / shannon
  6. open-metadata / OpenMetadata
  7. ruvnet / RuView
  8. HKUDS / RAG-Anything
  9. sansan0 / TrendRadar
  10. AIDC-AI / Pixelle-Video
  11. Z4nzu / hackingtool
  12. vercel-labs / skills

Product Hunt(15)

  1. Seeknal

    Data & AI/ML CLI for pipelines and NL queries

  2. AdsAgent

    Let Claude run your Google Ads

  3. Cavalry Studio

    Free Motion Design tool by Canva

  4. kimiflare

    kimi k2.6 cli code editor hosted on cloudflare workers AI

  5. Portt

    Transform your photo into any era and any location

  6. Iris Studio

    The Swiss Army Knife for AI & Video Creators

  7. Wrangle

    The markdown editor that understands CLAUDE.md

  8. Kyohansha

    Web-based 60FPS Live2D AI with Lite-RAG long-term memory

  9. ConsoleMini

    Turn a Mac mini into a living-room retro/PS console

  10. Toki 2.0

    Automatically go from ideas to scheduled plan

  11. ChatGPT Images 2.0

    First image model with thinking capabilities

  12. Loomal

    Identity infrastructure for AI agents

  13. Nomie v2

    Replace doomscrolling with a self-care interactive world

  14. Basedash Automations

    Your AI data analyst that works while you sleep.

  15. Cai

    Press ⌥C on anything to run smart actions, locally

Hugging Face(15)

  1. Tstars-Tryon 1.0: Robust and Realistic Virtual Try-On for Diverse Fashion Items

    Recent advances in image generation and editing have opened new opportunities for virtual try-on. However, existing methods still struggle to meet complex real-world demands. We present Tstars-Tryon 1.0, a commercial-scale virtual try-on system that is robust, realistic, versatile, and highly efficient. First, our system maintains a high success rate across challenging cases like extreme poses, severe illumination variations, motion blur, and other in-the-wild conditions. Second, it delivers highly photorealistic results with fine-grained details, faithfully preserving garment texture, material properties, and structural characteristics, while largely avoiding common AI-generated artifacts. Third, beyond apparel try-on, our model supports flexible multi-image composition (up to 6 reference images) across 8 fashion categories, with coordinated control over person identity and background. Fourth, to overcome the latency bottlenecks of commercial deployment, our system is heavily optimized for inference speed, delivering near real-time generation for a seamless user experience. These capabilities are enabled by an integrated system design spanning end-to-end model architecture, a scalable data engine, robust infrastructure, and a multi-stage training paradigm. Extensive evaluation and large-scale product deployment demonstrate that Tstars-Tryon1.0 achieves leading overall performance. To support future research, we also release a comprehensive benchmark. The model has been deployed at an industrial scale on the Taobao App, serving millions of users with tens of millions of requests.

  2. CoInteract: Physically-Consistent Human-Object Interaction Video Synthesis via Spatially-Structured Co-Generation

    Synthesizing human--object interaction (HOI) videos has broad practical value in e-commerce, digital advertising, and virtual marketing. However, current diffusion models, despite their photorealistic rendering capability, still frequently fail on (i) the structural stability of sensitive regions such as hands and faces and (ii) physically plausible contact (e.g., avoiding hand--object interpenetration). We present CoInteract, an end-to-end framework for HOI video synthesis conditioned on a person reference image, a product reference image, text prompts, and speech audio. CoInteract introduces two complementary designs embedded into a Diffusion Transformer (DiT) backbone. First, we propose a Human-Aware Mixture-of-Experts (MoE) that routes tokens to lightweight, region-specialized experts via spatially supervised routing, improving fine-grained structural fidelity with minimal parameter overhead. Second, we propose Spatially-Structured Co-Generation, a dual-stream training paradigm that jointly models an RGB appearance stream and an auxiliary HOI structure stream to inject interaction geometry priors. During training, the HOI stream attends to RGB tokens and its supervision regularizes shared backbone weights; at inference, the HOI branch is removed for zero-overhead RGB generation. Experimental results demonstrate that CoInteract significantly outperforms existing methods in structural stability, logical consistency, and interaction realism.

  3. AgentSPEX: An Agent SPecification and EXecution Language

    Language-model agent systems commonly rely on reactive prompting, in which a single instruction guides the model through an open-ended sequence of reasoning and tool-use steps, leaving control flow and intermediate state implicit and making agent behavior potentially difficult to control. Orchestration frameworks such as LangGraph, DSPy, and CrewAI impose greater structure through explicit workflow definitions, but tightly couple workflow logic with Python, making agents difficult to maintain and modify. In this paper, we introduce AgentSPEX, an Agent SPecification and EXecution Language for specifying LLM-agent workflows with explicit control flow and modular structure, along with a customizable agent harness. AgentSPEX supports typed steps, branching and loops, parallel execution, reusable submodules, and explicit state management, and these workflows execute within an agent harness that provides tool access, a sandboxed virtual environment, and support for checkpointing, verification, and logging. Furthermore, we provide a visual editor with synchronized graph and workflow views for authoring and inspection. We include ready-to-use agents for deep research and scientific research, and we evaluate AgentSPEX on 7 benchmarks. Finally, we show through a user study that AgentSPEX provides a more interpretable and accessible workflow-authoring paradigm than a popular existing agent framework.

  4. AnyRecon: Arbitrary-View 3D Reconstruction with Video Diffusion Model

    Sparse-view 3D reconstruction is essential for modeling scenes from casual captures, but remain challenging for non-generative reconstruction. Existing diffusion-based approaches mitigates this issues by synthesizing novel views, but they often condition on only one or two capture frames, which restricts geometric consistency and limits scalability to large or diverse scenes. We propose AnyRecon, a scalable framework for reconstruction from arbitrary and unordered sparse inputs that preserves explicit geometric control while supporting flexible conditioning cardinality. To support long-range conditioning, our method constructs a persistent global scene memory via a prepended capture view cache, and removes temporal compression to maintain frame-level correspondence under large viewpoint changes. Beyond better generative model, we also find that the interplay between generation and reconstruction is crucial for large-scale 3D scenes. Thus, we introduce a geometry-aware conditioning strategy that couples generation and reconstruction through an explicit 3D geometric memory and geometry-driven capture-view retrieval. To ensure efficiency, we combine 4-step diffusion distillation with context-window sparse attention to reduce quadratic complexity. Extensive experiments demonstrate robust and scalable reconstruction across irregular inputs, large viewpoint gaps, and long trajectories.

  5. TEMPO: Scaling Test-time Training for Large Reasoning Models

    Test-time training (TTT) adapts model parameters on unlabeled test instances during inference time, which continuously extends capabilities beyond the reach of offline training. Despite initial gains, existing TTT methods for LRMs plateau quickly and do not benefit from additional test-time compute. Without external calibration, the self-generated reward signal increasingly drifts as the policy model evolves, leading to both performance plateaus and diversity collapse. We propose TEMPO, a TTT framework that interleaves policy refinement on unlabeled questions with periodic critic recalibration on a labeled dataset. By formalizing this alternating procedure through the Expectation-Maximization (EM) algorithm, we reveal that prior methods can be interpreted as incomplete variants that omit the crucial recalibration step. Reintroducing this step tightens the evidence lower bound (ELBO) and enables sustained improvement. Across diverse model families (Qwen3 and OLMO3) and reasoning tasks, TEMPO improves OLMO3-7B on AIME 2024 from 33.0% to 51.1% and Qwen3-14B from 42.3% to 65.8%, while maintaining high diversity.

  6. PlayCoder: Making LLM-Generated GUI Code Playable

    Large language models (LLMs) have achieved strong results in code generation, but their ability to generate GUI applications, especially games, remains insufficiently studied. Existing benchmarks mainly evaluate correctness through test cases, which are inadequate for GUI applications because these systems are interactive, event-driven, and require correct state transitions across sequences of user actions. Their evaluation therefore should consider interaction flows and UI logic rather than only pass/fail outcomes. To study this problem, we introduce PlayEval, a repository-aware benchmark built from 43 multilingual GUI applications in Python, TypeScript, and JavaScript. Unlike prior GUI benchmarks that are difficult to adapt to desktop environments, PlayEval covers six major GUI application categories and directly supports code-generation evaluation. We further propose Play@k, a metric that measures whether at least one of *k* generated candidates can be played end-to-end without logical errors. To support reliable evaluation, we develop PlayTester, an LLM-based agent that performs task-oriented GUI playthroughs and detects logic violations automatically. Experiments on 10 state-of-the-art code LLMs show that, despite high compilation rates, they achieve near-zero Play@3, revealing major weaknesses in generating logically correct GUI applications. To address this limitation, we present PlayCoder, a multi-agent, repository-aware framework that generates, evaluates, and iteratively repairs GUI application code in a closed loop. PlayCoder substantially improves both functional correctness and semantic alignment for open-source and closed-source models, reaching up to 38.1% Exec@3 and 20.3% Play@3. Case studies further show that it can uncover silent logic bugs missed by traditional metrics and fix them through targeted edits.

  7. ShadowPEFT: Shadow Network for Parameter-Efficient Fine-Tuning

    Parameter-efficient fine-tuning (PEFT) reduces the training cost of full-parameter fine-tuning for large language models (LLMs) by training only a small set of task-specific parameters while freezing the pretrained backbone. However, existing approaches, such as Low-Rank Adaptation (LoRA), achieve adaptation by inserting independent low-rank perturbations directly to individual weights, resulting in a local parameterization of adaptation. We propose ShadowPEFT, a centralized PEFT framework that instead performs layer-level refinement through a depth-shared shadow module. At each transformer layer, ShadowPEFT maintains a parallel shadow state and evolves it repeatedly for progressively richer hidden states. This design shifts adaptation from distributed weight-space perturbations to a shared layer-space refinement process. Since the shadow module is decoupled from the backbone, it can be reused across depth, independently pretrained, and optionally deployed in a detached mode, benefiting edge computing scenarios. Experiments on generation and understanding benchmarks show that ShadowPEFT matches or outperforms LoRA and DoRA under comparable trainable-parameter budgets. Additional analyses on shadow pretraining, cross-dataset transfer, parameter scaling, inference latency, and system-level evaluation suggest that centralized layer-space adaptation is a competitive and flexible alternative to conventional low-rank PEFT.

  8. Chat2Workflow: A Benchmark for Generating Executable Visual Workflows with Natural Language

    At present, executable visual workflows have emerged as a mainstream paradigm in real-world industrial deployments, offering strong reliability and controllability. However, in current practice, such workflows are almost entirely constructed through manual engineering: developers must carefully design workflows, write prompts for each step, and repeatedly revise the logic as requirements evolve-making development costly, time-consuming, and error-prone. To study whether large language models can automate this multi-round interaction process, we introduce Chat2Workflow, a benchmark for generating executable visual workflows directly from natural language, and propose a robust agentic framework to mitigate recurrent execution errors. Chat2Workflow is built from a large collection of real-world business workflows, with each instance designed so that the generated workflow can be transformed and directly deployed to practical workflow platforms such as Dify and Coze. Experimental results show that while state-of-the-art language models can often capture high-level intent, they struggle to generate correct, stable, and executable workflows, especially under complex or changing requirements. Although our agentic framework yields up to 5.34% resolve rate gains, the remaining real-world gap positions Chat2Workflow as a foundation for advancing industrial-grade automation. Code is available at https://github.com/zjunlp/Chat2Workflow.

  9. AJ-Bench: Benchmarking Agent-as-a-Judge for Environment-Aware Evaluation

    As reinforcement learning continues to scale the training of large language model-based agents, reliably verifying agent behaviors in complex environments has become increasingly challenging. Existing approaches rely on rule-based verifiers or LLM-as-a-Judge models, which struggle to generalize beyond narrow domains. Agent-as-a-Judge addresses this limitation by actively interacting with environments and tools to acquire verifiable evidence, yet its capabilities remain underexplored. We introduce a benchmark AJ-Bench to systematically evaluate Agent-as-a-Judge across three domains-search, data systems, and graphical user interfaces-comprising 155 tasks and 516 annotated trajectories. The benchmark comprehensively assesses judge agents' abilities in information acquisition, state verification, and process verification. Experiments demonstrate consistent performance gains over LLM-as-a-Judge baselines, while also revealing substantial open challenges in agent-based verification. Our data and code are available at https://aj-bench.github.io/.

  10. Understanding and Enforcing Weight Disentanglement in Task Arithmetic

    Task arithmetic provides an efficient, training-free way to edit pre-trained models, yet lacks a fundamental theoretical explanation for its success. The existing concept of ``weight disentanglement" describes the ideal outcome of non-interfering task composition but does not reveal its underlying cause. Crucially, what intrinsic properties of the pre-trained model (θ_0) or the task vectors (τ_t) enable this disentanglement remains underexplored. In this paper, we introduce Task-Feature Specialization (TFS), a model's ability to allocate distinct internal features to different tasks, as the fundamental principle. We first prove that TFS is a sufficient condition for weight disentanglement. More importantly, we find that TFS also gives rise to an observable geometric consequence: weight vector orthogonality. This positions TFS as the common cause for both the desired functional outcome (disentanglement) and a measurable geometric property (orthogonality). This relationship provides the key insight for our method: since the abstract TFS property is intractable to enforce directly, we can instead promote weight disentanglement by shaping its concrete geometric consequence, orthogonality. Therefore, we propose OrthoReg, a simple and effective regularization method that actively enforces an internal orthogonal structure on weight updates (ΔW) that constitute τ_t during fine-tuning. And we theoretically prove that OrthoReg promotes disentanglement. Extensive experiments demonstrate that OrthoReg consistently and significantly enhances the performance of various task arithmetic methods. Code is available at https://github.com/RL-MIND/OrthoReg{https://github.com/RL-MIND/OrthoReg}.

  11. Code-Switching Information Retrieval: Benchmarks, Analysis, and the Limits of Current Retrievers

    Code-switching is a pervasive linguistic phenomenon in global communication, yet modern information retrieval systems remain predominantly designed for, and evaluated within, monolingual contexts. To bridge this critical disconnect, we present a holistic study dedicated to code-switching IR. We introduce CSR-L (Code-Switching Retrieval benchmark-Lite), constructing a dataset via human annotation to capture the authentic naturalness of mixed-language queries. Our evaluation across statistical, dense, and late-interaction paradigms reveals that code-switching acts as a fundamental performance bottleneck, degrading the effectiveness of even robust multilingual models. We demonstrate that this failure stems from substantial divergence in the embedding space between pure and code-switched text. Scaling this investigation, we propose CS-MTEB, a comprehensive benchmark covering 11 diverse tasks, where we observe performance declines of up to 27%. Finally, we show that standard multilingual techniques like vocabulary expansion are insufficient to resolve these deficits completely. These findings underscore the fragility of current systems and establish code-switching as a crucial frontier for future IR optimization.

  12. Dual-View Training for Instruction-Following Information Retrieval

    Instruction-following information retrieval (IF-IR) studies retrieval systems that must not only find documents relevant to a query, but also obey explicit user constraints such as required attributes, exclusions, or output preferences. However, most retrievers are trained primarily for semantic relevance and often fail to distinguish documents that match the topic from those that satisfy the instruction. We propose a dual-view data synthesis strategy based on polarity reversal: given a query, a document that is relevant under the instruction, and a hard negative that matches the query but violates the instruction, we prompt an LLM to generate a complementary instruction under which the two documents swap relevance labels. By presenting the same document pair under complementary instructions that invert their relevance labels, the training signal forces the retriever to reconsider the same candidate set through the instruction, rather than relying on fixed topical cues. On a 305M-parameter encoder, our method improves performance on the FollowIR benchmark by 45%, surpassing general-purpose embedding models of comparable or larger scale. Through head-to-head comparisons at matched data budgets, we further show that data diversity and instruction supervision play complementary roles: the former preserves general retrieval quality, while the latter improves instruction sensitivity. These results highlight the value of targeted data synthesis for building retrieval systems that are both broadly capable and instruction-aware.

  13. CityRAG: Stepping Into a City via Spatially-Grounded Video Generation

    We address the problem of generating a 3D-consistent, navigable environment that is spatially grounded: a simulation of a real location. Existing video generative models can produce a plausible sequence that is consistent with a text (T2V) or image (I2V) prompt. However, the capability to reconstruct the real world under arbitrary weather conditions and dynamic object configurations is essential for downstream applications including autonomous driving and robotics simulation. To this end, we present CityRAG, a video generative model that leverages large corpora of geo-registered data as context to ground generation to the physical scene, while maintaining learned priors for complex motion and appearance changes. CityRAG relies on temporally unaligned training data, which teaches the model to semantically disentangle the underlying scene from its transient attributes. Our experiments demonstrate that CityRAG can generate coherent minutes-long, physically grounded video sequences, maintain weather and lighting conditions over thousands of frames, achieve loop closure, and navigate complex trajectories to reconstruct real-world geography.

  14. Speculative Decoding for Autoregressive Video Generation

    Autoregressive video diffusion is emerging as a promising paradigm for streaming video synthesis, with step distillation serving as the primary means of accelerating inference. Whether speculative decoding, the dominant acceleration strategy for large language models, can be effectively adapted to autoregressive video generation remains an open question, because video blocks are continuous spatiotemporal tensors with no token-level distribution for exact rejection sampling. We introduce SDVG, which brings speculative decoding to block-based autoregressive video diffusion by replacing token verification with an image-quality router. A 1.3B drafter proposes candidate blocks via four denoising steps; each block is VAE-decoded and scored by ImageReward using worst-frame aggregation--taking the minimum per-frame reward to catch single-frame artifacts that averaging would mask. Blocks scoring above a fixed threshold tau are accepted into the 14B target's KV cache; the rest are regenerated by the target. Two additional design choices prove critical: the first block is always force-rejected to anchor scene composition, and tau serves as a single knob that traces a smooth quality-speed Pareto frontier. On 1003 MovieGenVideoBench prompts (832x480), SDVG retains 98.1% of target-only VisionReward quality (0.0773 vs. 0.0788) at a 1.59x speedup with tau=-0.7, and reaches 2.09x at 95.7% quality retention--while consistently outperforming draft-only generation by over +17%. The framework is training-free, requires no architectural changes, and can be seamlessly integrated into existing autoregressive video generation pipelines.

  15. Target-Oriented Pretraining Data Selection via Neuron-Activated Graph

    Everyday tasks come with a target, and pretraining models around this target is what turns them into experts. In this paper, we study target-oriented language model (LM) pretraining by introducing Neuron-Activated Graph Ranking (NAG-based Ranking), a training-free and interpretable framework for target pretraining data selection. Rather than using black-box representations, our approach directly characterizes each target input by a sparse set of high-impact neurons in any off-the-shelf LLMs. Concretely, we quantify neuron impact and select the most influential neurons across layers into a compact Neuron-Activated Graph (NAG), and rank candidate data by NAG similarity to target examples. We conduct experiments across six benchmarks, where our NAG-based Ranking improves target-oriented pretraining by 4.9% on average over random sampling, and also outperforms state-of-the-art baselines by 5.3% accuracy on HellaSwag. It also remains effective under a more applicable multi-target setting, where our best setup surpasses two baselines by 1.1% and 4.1%, respectively. Furthermore, we provide a comprehensive analysis on why and how our NAG works, e.g., deactivating NAG-selected neurons (only 0.12% of all) causes a 23.5% performance collapse, and restricting NAG to the final layer incurs a 4.1% average drop, indicating that NAG captures a sparse "functional backbone" for learning target features. We release the code at https://github.com/asillycat/NAG.

Techmeme(15)

  1. Dune Analytics: Polymarket's global trading volumes have fallen behind Kalshi in recent months; sources: product delays are contributing to Polymarket's decline (Bloomberg)

    Bloomberg : Dune Analytics: Polymarket's global trading volumes have fallen behind Kalshi in recent months; sources: product delays are contributing to Polymarket's decline —  Polymarket, the long-time leader in prediction market trading volume, has fallen behind its chief rival as it faces a growing list …

  2. Kalshi suspends and fines congressional candidates Mark Moran of Virginia, Matt Klein of Minnesota, and Ezekiel Enriquez of Texas for political insider trading (Dan Mangan/CNBC)

    Dan Mangan / CNBC : Kalshi suspends and fines congressional candidates Mark Moran of Virginia, Matt Klein of Minnesota, and Ezekiel Enriquez of Texas for political insider trading —  Prediction market platform Kalshi said Wednesday it had suspended and fined three candidates for Congress from Minnesota …

  3. IBM reports Q1 revenue up 9% YoY to $15.92B, vs. $15.62B est., software revenue up 11% to $7.05B, and maintains FY 2026 guidance; IBM drops ~6% after hours (Jordan Novet/CNBC)

    Jordan Novet / CNBC : IBM reports Q1 revenue up 9% YoY to $15.92B, vs. $15.62B est., software revenue up 11% to $7.05B, and maintains FY 2026 guidance; IBM drops ~6% after hours —  IBM shares slipped 6% in extended trading on Wednesday after the hardware, software and consulting provider reported stronger …

  4. ServiceNow reports Q1 subscription revenue up 22% YoY to $3.67B, vs. $3.65B est., says conflict in the Middle East weighed on growth; NOW drops 12%+ after hours (Lola Murti/CNBC)

    Lola Murti / CNBC : ServiceNow reports Q1 subscription revenue up 22% YoY to $3.67B, vs. $3.65B est., says conflict in the Middle East weighed on growth; NOW drops 12%+ after hours —  ServiceNow reported first-quarter results on Wednesday that narrowly beat Wall Street's estimates as the company said the conflict …

  5. Texas Instruments reports Q1 revenue up 19% YoY to $4.83B, vs. $4.52B est., and forecasts Q2 revenue above estimates; TXN jumps 8%+ after hours (Ian King/Bloomberg)

    Ian King / Bloomberg : Texas Instruments reports Q1 revenue up 19% YoY to $4.83B, vs. $4.52B est., and forecasts Q2 revenue above estimates; TXN jumps 8%+ after hours —  Texas Instruments Inc. shares surged in late trading after the chipmaker gave a surprisingly strong forecast, helped by booming spending on data centers and industrial equipment.

  6. Sources: xAI held talks in recent weeks with Mistral and Cursor about a potential three-way partnership; Mistral co-founder Devendra Chaplot joined xAI in March (Grace Kay/Business Insider)

    Grace Kay / Business Insider : Sources: xAI held talks in recent weeks with Mistral and Cursor about a potential three-way partnership; Mistral co-founder Devendra Chaplot joined xAI in March —  - Elon Musk's xAI recently announced a partnership with AI coding startup Cursor.  — The two companies have also explored …

  7. Sean Plankey, President Trump's pick to lead CISA, withdraws from consideration after resistance from Sen. Rick Scott stalled his nomination for over a year (Politico)

    Politico : Sean Plankey, President Trump's pick to lead CISA, withdraws from consideration after resistance from Sen. Rick Scott stalled his nomination for over a year —  Spokespeople for the White House, DHS and the Cybersecurity and Infrastructure Security Agency did not immediately respond to a request for comment.

  8. OpenAI releases Privacy Filter, an open-weight model for masking personally identifiable information in text, with 1.5B total and 50M active parameters (OpenAI)

    OpenAI : OpenAI releases Privacy Filter, an open-weight model for masking personally identifiable information in text, with 1.5B total and 50M active parameters —  Our state of the art model for masking personally identifiable information (PII) in text  —  Today we're releasing OpenAI Privacy Filter …

  9. Meta unveils Live Chats on Threads for real-time conversations during cultural events, launching first within the NBA Threads community during the playoffs (Aisha Malik/TechCrunch)

    Aisha Malik / TechCrunch : Meta unveils Live Chats on Threads for real-time conversations during cultural events, launching first within the NBA Threads community during the playoffs —  Threads is launching “Live Chats” to allow for real-time conversations during cultural events, the Meta-owned platform announced on Wednesday.

  10. Sources: SpaceX isn't acquiring Cursor immediately because the deal could delay its IPO; Cursor is no longer proceeding with its reported $2B funding round (Bloomberg)

    Bloomberg : Sources: SpaceX isn't acquiring Cursor immediately because the deal could delay its IPO; Cursor is no longer proceeding with its reported $2B funding round —  SpaceX said it has an agreement giving it the right to acquire artificial intelligence startup Cursor for $60 billion later this year …

  11. OpenAI announces workspace agents in ChatGPT, letting teams create Codex-powered shared agents for complex tasks, and says they are "an evolution of GPTs" (OpenAI)

    OpenAI : OpenAI announces workspace agents in ChatGPT, letting teams create Codex-powered shared agents for complex tasks, and says they are “an evolution of GPTs” —  Codex-powered agents for teams.  —  Create an agent(opens in a new window)Contact sales(opens in a new window)

  12. Google says 75% of new code created inside the company is now generated by AI and reviewed by human engineers, up from 50% last fall (Hugh Langley/Business Insider)

    Hugh Langley / Business Insider : Google says 75% of new code created inside the company is now generated by AI and reviewed by human engineers, up from 50% last fall —  - Three-quarters of new code at Google is being generated by AI, the company said.  — The number has been steadily increasing as the company pushes staff to adopt AI tools.

  13. Core Automation, co-founded by ex-OpenAI VP Jerry Tworek, launches to build "the world's most automated AI lab" with talent from OpenAI, Anthropic, and DeepMind (Business Insider)

    Business Insider : Core Automation, co-founded by ex-OpenAI VP Jerry Tworek, launches to build “the world's most automated AI lab” with talent from OpenAI, Anthropic, and DeepMind —  - A new AI startup, Core Automation, has entered the chat.  — Started by an ex-OpenAI researcher, it has …

  14. LinkedIn names COO Daniel Shapero as its new CEO, succeeding Ryan Roslansky, who will retain his position as EVP at Microsoft (Jordan Novet/CNBC)

    Jordan Novet / CNBC : LinkedIn names COO Daniel Shapero as its new CEO, succeeding Ryan Roslansky, who will retain his position as EVP at Microsoft —  Microsoft has tapped Dan Shapero to be the new CEO of its LinkedIn division, succeeding Ryan Roslansky, who has run the subsidiary since 2020 and last year took …

  15. Monk, which automates accounts receivable workflows, raised a $25M Series A co-led by Footwork and Acrew Capital, bringing its total funding to $29M (Ryan Lawler/Axios)

    Ryan Lawler / Axios : Monk, which automates accounts receivable workflows, raised a $25M Series A co-led by Footwork and Acrew Capital, bringing its total funding to $29M —  Monk, which automates accounts receivable workflows, raised $25 million in Series A funding led by Footwork, CEO George Kurdin tells Axios exclusively.

Solidot(15)

  1. 阴谋论引发美国总统关注促使 FBI 调查

    美国 UFO 社群将过去几年发生的一系列航空航天和核物理领域无关联的专家死亡或失踪案关联起来,认为其中存在巨大的阴谋,怀疑是黑衣人组织 MIB 在执行清理工作。这一阴谋论最终渗透到美国政府,美国总统特朗普称此事非常严重,白宫表示在与联邦机构合作对此展开调查,共和党领导的众议院监督委员会也表示展开调查,FBI 表示正牵头展开调查,正与能源部、战争部以及州和地方执法部门合作去寻找答案。

  2. Google 发布第八代自研 AI 芯片 TPU 8t 和 TPU 8i

    Google 宣布了第八代自研 AI 芯片 TPU 8t 和 TPU 8i,前者专为大模型训练设计,后者专为大模型推理设计。TPU 8t 拥有更大的计算吞吐量和更多的可扩展带宽去满足计算密集训练工作负载,而 TPU 8i 则拥有更多的内存带宽去满足对延迟最敏感的推理工作负载。Google 称,TPU 8t 设计将前沿模型的开发周期从数月缩短至数周,单个 TPU 8t superpod 可扩展至 9,600 个芯片和 2 PB 共享高带宽内存,芯片间带宽是上一代的两倍,该架构可提供 121 ExaFlops 算力,允许最复杂模型利用单一海量内存池。TPU 8i 芯片则配备了 288 GB 高带宽内存和 384 MB 片上 SRAM,模型活动工作集能完全留在芯片上。

  3. 中国选拔出两名巴基斯坦预备宇航员

    中国载人航天工程首批外籍航天员选拔工作于 2026 年 4 月上旬结束,2 名巴基斯坦籍候选对象 Muhammad Zeeshan Ali(穆罕默德·齐尚·阿里)和 Khurram Daud(胡拉姆·达乌德)最终入选。他们将作为预备航天员来华参加训练,在完成各项训练并通过考核后,其中 1 人将以载荷专家身份参加飞行任务,成为首位进入中国空间站的外籍航天员。2025 年 2 月,中巴在伊斯兰堡签署《关于选拔、训练巴基斯坦航天员并参与中国空间站飞行任务的合作协议》,随即正式启动巴基斯坦航天员选拔工作。经过初选、复选、定选三个阶段的严格筛选和评定,最终选拔出 2 名巴基斯坦预备航天员。

  4. 英国 17 岁以下儿童终身禁止购烟

    英国议会通过了法案 Tobacco and Vapes Bill,17 岁及以下儿童将面临终身禁止购买香烟的禁令。该法案旨在禁止商店向 2009 年 1 月 1 日之后出生的儿童出售烟草,阻止他们吸烟,最终目标是打造无烟一代。吸烟是可预防死因之一,该法案被视为是几十年来英国最大的公共卫生干预措施。

  5. 印度男子用 AI 生成的 MAGA 女孩骗美国保守派男性

    22 岁的 Sam 在印度医学院学习,经济拮据,父母资助的钱大部分花在执业资格考试上,他还在为毕业后移民美国存钱。他尝试了很多赚钱法,制作 YouTube 短视频,向其他医学院学生出售学习笔记等等,效果不一,他在浏览 Instagram 时突然产生了一个想法,为什么不用 Google Gemini 的 Nano Banana Pro 生成一个女孩,然后出售她的比基尼照片?然而他的 AI 女孩基本无人问津,他转而询问 Gemini 背后的原因。AI 指出他创造的是没有特色的性感女孩,需要与网络上数以百万计的类似女孩展开竞争。Gemini 帮助他挑选了一个非常有潜力的细分市场:MAGA/保守派。Gemini 指出美国保守派尤其是老年男性,通常拥有更高的可支配收入,而且更忠诚。Sam 于去年 1 月创建了 Emily Hart(@emily_hart.nurse),一名注册护士,长相类似 Jennifer Lawrence,他在发布照片时配上了很多反对自由派和支持保守派的口号,如“基督是王,堕胎是谋杀,非法移民必须被遣返”之类。Sam 从未去过美国,但却成为 MAGA 思想的忠实拥护者,他说自己每天都会写支持基督教、支持第二修正案、拥护生命权、反堕胎、反觉醒主义和反移民的文章。虽然他觉得自己的骗局过于明显,但他的账号却爆红了。他发布的每则 Reels 视频都能获得 300 万、500 万、1000 万的观看量,算法在推荐此类内容。由于 OnlyFans 限制 AI 模特,因此他在不限制 AI 的平台 Fanvue 发布了 AI 生成的软色情,依靠 Fanvue 订阅收入和 MAGA 主题 T 恤销售,他每月赚到了数千美元。Emily Hart 是众多 AI 生成的 MAGA 女网红之一,此类网红都是基于特定模板生成的:白人金发,从事应急救援工作如警察、消防员或急救员,言论都是常见的右翼观点。MAGA 女网红比较罕见,因为 18-29 岁的女性绝大多数倾向于自由派,年轻 MAGA 女性更能吸引保守派男性眼球。自由派对 AI 的识别度更高,Sam 尝试过生成自由派女网红,根本没有互动。他曾收到一则视频,一名男子对着 Emily 的照片自慰,对方给了 50 美元小费,他的感觉就是“随你便”。@emily_hart.nurse 账号在今年 2 月被 Instagram 以存在欺诈行为为由封禁。Sam 说即使没封他也打算收手了,他需要将重心转移到学业上。

  6. 残疾鹦鹉 Bruce 在保护区里百战百胜

    名叫 Bruce 的啄羊鹦鹉缺少了上半个喙,无法像同类那样用尖利的喙叮咬对手,但 Bruce 学会了用其它方法攻击对手,观察显示它在打斗中从未输过。啄羊鹦鹉专家 Raoul Schwing 于 2013 年发现了幼鸟 Bruce,当时它的上喙已经失去了,它被送到了 Willowbank 野生动物保护区。它在圈养环境下生活的不错,缺少了上喙影响到了它自我清洁的能力。保护区的科学家发现它会使用小石头清洁羽毛上的螨虫和污垢,这是啄羊鹦鹉使用工具自我清洁的第一个案例。除了缺少上喙,它的体型在同伴里也属于最小之列。然而科学家观察到它在与同类的冲突中始终占据了上风,研究期间科学家记录到了它与其它鸟类的 36 次冲突,每次都是它赢。相比对手张开喙咬脖子,Bruce 的打斗技巧是用下喙向前猛扑,动作风格类似于马上挥枪或击剑,攻击对手的背部、头部、翅膀和腿部。科学家以前也记录过残疾动物获得优势地位的少数案例,如著名的珍妮·古道尔(Jane Goodall)写过一只阿尔法雄猩猩因脊髓灰质炎(Poliomyelitis)失去了一只手臂但仍然获得了较高的地位。

  7. 好奇号在火星上发现新有机分子

    根据发表在《Nature Communications》期刊上的一项研究,科学家从好奇号在火星盖尔陨石坑采集的岩石样本中发现了 20 多种有机分子,其中一种含氮分子的结构与形成 DNA 的化合物相似,此前从未在火星上探测到。最新研究成果支持了火星可能曾经拥有适宜生命存在环境的观点,但还无法证明火星曾有过生命。有机分子被认为是生命的关键成分,但它们也可通过非生物过程形成,或通过陨石到达星球。其中一种识别的有机分子苯并噻吩(benzothiophene)通常就与太空陨石相关。科学家是利用好奇号搭载的 Sample Analysis at Mars 仪器,使用化合物 TMAH 分解大有机分子进行研究。

  8. Cal.com 发布开源社区版 Cal.diy

    日程安排平台 Cal.com 在以安全理由从开源转为闭源后,发布了面向个人自托管的开源社区版本 Cal.diy,并且在发布声明中反复强调不要将其用于生产用途。Cal.diy 采用 MIT 许可证,是 Cal.com 的开源分支,移除所有私有的企业版功能如 Teams、Organizations、Insights、Workflows 和 SSO/SAML,不需要 Cal.com 账号或授权。Cal.com 声明,它不推荐将 Cal.diy 用于生产。

  9. 新报告称 2025 年是太阳能占据优势的第一年

    国际能源署最新报告证实,2025 年是太阳能占据优势的第一年。过去两年多数能源的发电量变化不大,但太阳能是一个显著的例外。2025 年太阳能总发电量 2700 太瓦时(terawatt-hours),是三年前的两倍多。太阳能发电量占到了全球总发电量的逾 8%。30 个国家新增太阳能发电装机容量至少 1 GW。太阳能发电离不开电池蓄能,2024 年到 2025 年间新增电池蓄能容量增长了 40%,去年新增容量达到 110 GW。尽管中国在 2025 年投产了大量燃煤电厂,但由于在可再生能源的大规模投资,去年中国的燃煤发电量反而略有下降。中国还在建造大量核电站,如果全部投入运营,核电装机容量将超过美国。去年开工建设的新核电厂装机容量为 12 GW,10 座核电厂有 9 座位于中国。

  10. Scattered Spider 资深成员 Tylerb 认罪

    24 岁的英国公民、勒索组织 Scattered Spider 资深成员 Tyler Robert Buchanan aka Tylerb 认罪。他目前关押在美国,面临逾 20 年监禁。Scattered Spider 以利用社交工程方法入侵公司窃取数据勒索赎金闻名,该组织成员通常采取的方法是冒充员工或合同工欺骗 IT 授予企业的访问权限。作为认罪协议的一部分,Tylerb 承认与其他人合谋,在 2022 年发起数万起基于短信的钓鱼攻击,导致包括 Twilio、LastPass、DoorDash 和 Mailchimp 在内的科技公司系统遭到入侵。攻击者利用窃取的数据对加密货币投资者发起 SIM-swapping 攻击,仅在美国就从受害者手中窃取了至少 800 万美元的虚拟货币。Tylerb 被发现是因为他使用相同的用户名和电子邮件注册了钓鱼域名,域名注册商 NameCheap 提供的注册时使用的 IP 地址暴露了其身份。

  11. 微软降低 Game Pass 订阅费

    在新上任 Xbox CEO Asha Sharma 上周表示 Game Pass 订阅费对玩家而言太贵之后,微软下调了价格,不过代价是《使命召唤》系列新作不会同步加入到 Game Pass 服务,而是会延后一年加入。此前有报道称 Game Pass 订阅服务让《使命召唤》损失了 3 亿美元的收入。根据最新调价:Xbox Game Pass Ultimate 将从每月 29.99 美元降至 22.99 美元,PC Game Pass 将从每月 16.49 美元降至 13.99 美元。

  12. Meta 开始记录员工鼠标移动和按键用于 AI 训练

    公司内部备忘录显示,Meta 正在美国员工电脑上安装追踪软件,捕捉员工鼠标移动、点击和按键数据以用于训练 AI 模型,此举是该公司构建能自动执行工作任务的 AI 智能体的大计划的一部分。被称为 Model Capability Initiative(MCI)的工具将在工作相关应用和网站上运行,会不定时截取屏幕内容的快照。备忘录表示此举旨在改进公司 AI 模型,使其在难以模拟人机交互领域如下拉菜单选择和使用快捷键上表现更出色。耶鲁法学教授Ifeoma Ajunwa 表达了员工被实时监控的担忧,称美国在这方面缺乏监管。加拿大法学教授 Valerio De Stefano 称欧洲法律会禁止此类监控。

  13. Firefox 150 释出

    Mozilla 释出了 Firefox 150,主要新特性包括:对所有用户启用“本地网络访问限制”;通过内置 PDF 编辑器支持对文件进行页面重排序、复制、删除、粘贴和导出;改进分屏浏览;通过 about:translations 体验尊重隐私的即时翻译(或在地址栏直接输入 translate);等等。

  14. SpaceX 达成协议可选以 600 亿美元收购 Cursor

    SpaceX 还在继续为其史上最高估值增加燃料。它宣布获得了一个选择权,可选择在晚些时候以 600 亿美元收购 Cursor,或者支付 100 亿美元建立新的合作伙伴关系。Cursor 提供了 AI 驱动的集成开发环境(IDE),支持代码生成、智能重写和代码库查询等功能。收购或与 Cursor 合作将有助于增强 SpaceX 旗下的 xAI 在 AI 代码生成方面的实力。相比 OpenAI 和 Anthropic,xAI 在 AI 代码生成方面缺乏竞争力。

  15. 创意软件行业向 Adobe 宣战

    帝国终将陨落,创意软件行业如今一致认为 Adobe 结束的时代即将到来,它们纷纷以免费或更低的价格提供与 Adobe 竞争的产品——Adobe 的创意软件过去几十年一直被视为行业标准。Cinema 4D 开发商 Maxon 在收购 Autograph 之后向个人用户提供了免费版本,Autograph 是类似 Adobe After Effects 的 动态图形设计软件,此前的永久授权费用高达 1795 美元;Canva 在收购 Affinity 之后将功能上类似 Adobe Illustrator、Photoshop 和 InDesign 的软件 Affinity Designer 2、Affinity Photo 2 和 Affinity Publisher 2 免费提供给用户,它在收购 Cavalry 之后也将其免费,Cavalry 类似 After Effects;苹果于今年 1 月推出了 Creator Studio 套件,包含了 Final Cut Pro、Logic Pro、Pixelmator Pro、Motion、Compressor 和 MainStage 等创意软件,月订费 12.99 美元,相比下 Adob​​e的 Creative Cloud Pro 每月订阅费用 69.99 美元,苹果没有强制用户订阅,用户仍可购买单个应用的单次授权。