DIGEST · 2026-03-18

OrangeBot.AI Digest — 2026-03-18

81 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. FBI is buying location data to track US citizens, director confirms (techcrunch.com)
  2. Wander – A tiny, decentralised tool to explore the small web (susam.net)
  3. AI coding is gambling (notes.visaint.space)
  4. Show HN: Hacker News archive (47M+ items, 11.6GB) as Parquet, updated every 5m (huggingface.co)
  5. Nvidia NemoClaw (github.com)
  6. OpenRocket (openrocket.info)
  7. Snowflake AI Escapes Sandbox and Executes Malware (www.promptarmor.com)
  8. Death to Scroll Fade (dbushell.com)
  9. Despite Doubts, Federal Cyber Experts Approved Microsoft Cloud Service (www.propublica.org)
  10. Tech hobbyist makes shoulder-mounted guided missile prototype with $96 in parts (www.tomshardware.com)
  11. Rob Pike’s Rules of Programming (1989) (www.cs.unc.edu)
  12. Nightingale – open-source karaoke app that works with any song on your computer (nightingale.cafe)
  13. Write up of my homebrew CPU build (willwarren.com)
  14. JPEG Compression (www.sophielwang.com)
  15. Have a fucking website (www.otherstrangeness.com)

GitHub Trending(6)

  1. jarrodwatts / claude-hud
  2. obra / superpowers
  3. unslothai / unsloth
  4. newton-physics / newton
  5. shadps4-emu / shadPS4
  6. langchain-ai / open-swe

Product Hunt(15)

  1. GPT‑5.4 mini and nano

    Fast and efficient models optimized for coding and subagents

  2. Comet for Enterprise

    Perplexity’s Secure AI browser built for enterprise teams

  3. Banyan AI Lite

    AI detecting & preventing SaaS churn

  4. Permit.io MCP Gateway

    Drop-in MCP Security Developers Love and CISOs Trust

  5. Lightfield

    AI-native CRM that builds itself and does work for you

  6. Forvibe for macOS

    Everything between your build and the App Store

  7. Genie by Databox

    Your AI analyst for business performance

  8. ClipLedger

    Track views & payouts for YouTube Shorts creators

  9. Bookshelf for NotebookLM

    Add folders, search, and sync to Google NotebookLM

  10. Soul 2.0

    Fashion-Grade AI Photos Without the Camera Crew

  11. MetricMap

    Track revenue, ads, web vitals, & user insights in one hub

  12. Lore

    Cursor for your memory. 100% private, open-source & free.

  13. Claude Dispatch

    Text Claude from your phone using “Dispatch”

  14. SharePatch

    Share git patches with clean, review-ready browser diffs

  15. Fantastical MCP for Mac

    Manage your schedule directly with Claude

Hugging Face(15)

  1. InCoder-32B: Code Foundation Model for Industrial Scenarios

    Recent code large language models have achieved remarkable progress on general programming tasks. Nevertheless, their performance degrades significantly in industrial scenarios that require reasoning about hardware semantics, specialized language constructs, and strict resource constraints. To address these challenges, we introduce InCoder-32B (Industrial-Coder-32B), the first 32B-parameter code foundation model unifying code intelligence across chip design, GPU kernel optimization, embedded systems, compiler optimization, and 3D modeling. By adopting an efficient architecture, we train InCoder-32B from scratch with general code pre-training, curated industrial code annealing, mid-training that progressively extends context from 8K to 128K tokens with synthetic industrial reasoning data, and post-training with execution-grounded verification. We conduct extensive evaluation on 14 mainstream general code benchmarks and 9 industrial benchmarks spanning 4 specialized domains. Results show InCoder-32B achieves highly competitive performance on general tasks while establishing strong open-source baselines across industrial domains.

  2. MiroThinker-1.7 & H1: Towards Heavy-Duty Research Agents via Verification

    We present MiroThinker-1.7, a new research agent designed for complex long-horizon reasoning tasks. Building on this foundation, we further introduce MiroThinker-H1, which extends the agent with heavy-duty reasoning capabilities for more reliable multi-step problem solving. In particular, MiroThinker-1.7 improves the reliability of each interaction step through an agentic mid-training stage that emphasizes structured planning, contextual reasoning, and tool interaction. This enables more effective multi-step interaction and sustained reasoning across complex tasks. MiroThinker-H1 further incorporates verification directly into the reasoning process at both local and global levels. Intermediate reasoning decisions can be evaluated and refined during inference, while the overall reasoning trajectory is audited to ensure that final answers are supported by coherent chains of evidence. Across benchmarks covering open-web research, scientific reasoning, and financial analysis, MiroThinker-H1 achieves state-of-the-art performance on deep research tasks while maintaining strong results on specialized domains. We also release MiroThinker-1.7 and MiroThinker-1.7-mini as open-source models, providing competitive research-agent capabilities with significantly improved efficiency.

  3. Qianfan-OCR: A Unified End-to-End Model for Document Intelligence

    We present Qianfan-OCR, a 4B-parameter end-to-end vision-language model that unifies document parsing, layout analysis, and document understanding within a single architecture. It performs direct image-to-Markdown conversion and supports diverse prompt-driven tasks including table extraction, chart understanding, document QA, and key information extraction. To address the loss of explicit layout analysis in end-to-end OCR, we propose Layout-as-Thought, an optional thinking phase triggered by special think tokens that generates structured layout representations -- bounding boxes, element types, and reading order -- before producing final outputs, recovering layout grounding capabilities while improving accuracy on complex layouts. Qianfan-OCR ranks first among end-to-end models on OmniDocBench v1.5 (93.12) and OlmOCR Bench (79.8), achieves competitive results on OCRBench, CCOCR, DocVQA, and ChartQA against general VLMs of comparable scale, and attains the highest average score on public key information extraction benchmarks, surpassing Gemini-3.1-Pro, Seed-2.0, and Qwen3-VL-235B. The model is publicly accessible via the Baidu AI Cloud Qianfan platform.

  4. Kinema4D: Kinematic 4D World Modeling for Spatiotemporal Embodied Simulation

    Simulating robot-world interactions is a cornerstone of Embodied AI. Recently, a few works have shown promise in leveraging video generations to transcend the rigid visual/physical constraints of traditional simulators. However, they primarily operate in 2D space or are guided by static environmental cues, ignoring the fundamental reality that robot-world interactions are inherently 4D spatiotemporal events that require precise interactive modeling. To restore this 4D essence while ensuring the precise robot control, we introduce Kinema4D, a new action-conditioned 4D generative robotic simulator that disentangles the robot-world interaction into: i) Precise 4D representation of robot controls: we drive a URDF-based 3D robot via kinematics, producing a precise 4D robot control trajectory. ii) Generative 4D modeling of environmental reactions: we project the 4D robot trajectory into a pointmap as a spatiotemporal visual signal, controlling the generative model to synthesize complex environments' reactive dynamics into synchronized RGB/pointmap sequences. To facilitate training, we curated a large-scale dataset called Robo4D-200k, comprising 201,426 robot interaction episodes with high-quality 4D annotations. Extensive experiments demonstrate that our method effectively simulates physically-plausible, geometry-consistent, and embodiment-agnostic interactions that faithfully mirror diverse real-world dynamics. For the first time, it shows potential zero-shot transfer capability, providing a high-fidelity foundation for advancing next-generation embodied simulation.

  5. Demystifing Video Reasoning

    Recent advances in video generation have revealed an unexpected phenomenon: diffusion-based video models exhibit non-trivial reasoning capabilities. Prior work attributes this to a Chain-of-Frames (CoF) mechanism, where reasoning is assumed to unfold sequentially across video frames. In this work, we challenge this assumption and uncover a fundamentally different mechanism. We show that reasoning in video models instead primarily emerges along the diffusion denoising steps. Through qualitative analysis and targeted probing experiments, we find that models explore multiple candidate solutions in early denoising steps and progressively converge to a final answer, a process we term Chain-of-Steps (CoS). Beyond this core mechanism, we identify several emergent reasoning behaviors critical to model performance: (1) working memory, enabling persistent reference; (2) self-correction and enhancement, allowing recovery from incorrect intermediate solutions; and (3) perception before action, where early steps establish semantic grounding and later steps perform structured manipulation. During a diffusion step, we further uncover self-evolved functional specialization within Diffusion Transformers, where early layers encode dense perceptual structure, middle layers execute reasoning, and later layers consolidate latent representations. Motivated by these insights, we present a simple training-free strategy as a proof-of-concept, demonstrating how reasoning can be improved by ensembling latent trajectories from identical models with different random seeds. Overall, our work provides a systematic understanding of how reasoning emerges in video generation models, offering a foundation to guide future research in better exploiting the inherent reasoning dynamics of video models as a new substrate for intelligence.

  6. TRUST-SQL: Tool-Integrated Multi-Turn Reinforcement Learning for Text-to-SQL over Unknown Schemas

    Text-to-SQL parsing has achieved remarkable progress under the Full Schema Assumption. However, this premise fails in real-world enterprise environments where databases contain hundreds of tables with massive noisy metadata. Rather than injecting the full schema upfront, an agent must actively identify and verify only the relevant subset, giving rise to the Unknown Schema scenario we study in this work. To address this, we propose TRUST-SQL (Truthful Reasoning with Unknown Schema via Tools). We formulate the task as a Partially Observable Markov Decision Process where our autonomous agent employs a structured four-phase protocol to ground reasoning in verified metadata. Crucially, this protocol provides a structural boundary for our novel Dual-Track GRPO strategy. By applying token-level masked advantages, this strategy isolates exploration rewards from execution outcomes to resolve credit assignment, yielding a 9.9% relative improvement over standard GRPO. Extensive experiments across five benchmarks demonstrate that TRUST-SQL achieves an average absolute improvement of 30.6% and 16.6% for the 4B and 8B variants respectively over their base models. Remarkably, despite operating entirely without pre-loaded metadata, our framework consistently matches or surpasses strong baselines that rely on schema prefilling.

  7. WorldCam: Interactive Autoregressive 3D Gaming Worlds with Camera Pose as a Unifying Geometric Representation

    Recent advances in video diffusion transformers have enabled interactive gaming world models that allow users to explore generated environments over extended horizons. However, existing approaches struggle with precise action control and long-horizon 3D consistency. Most prior works treat user actions as abstract conditioning signals, overlooking the fundamental geometric coupling between actions and the 3D world, whereby actions induce relative camera motions that accumulate into a global camera pose within a 3D world. In this paper, we establish camera pose as a unifying geometric representation to jointly ground immediate action control and long-term 3D consistency. First, we define a physics-based continuous action space and represent user inputs in the Lie algebra to derive precise 6-DoF camera poses, which are injected into the generative model via a camera embedder to ensure accurate action alignment. Second, we use global camera poses as spatial indices to retrieve relevant past observations, enabling geometrically consistent revisiting of locations during long-horizon navigation. To support this research, we introduce a large-scale dataset comprising 3,000 minutes of authentic human gameplay annotated with camera trajectories and textual descriptions. Extensive experiments show that our approach substantially outperforms state-of-the-art interactive gaming world models in action controllability, long-horizon visual quality, and 3D spatial consistency.

  8. Thinking in Uncertainty: Mitigating Hallucinations in MLRMs with Latent Entropy-Aware Decoding

    Recent advancements in multimodal large reasoning models (MLRMs) have significantly improved performance in visual question answering. However, we observe that transition words (e.g., because, however, and wait) are closely associated with hallucinations and tend to exhibit high-entropy states. We argue that adequate contextual reasoning information can be directly extracted from the token probability distribution. Inspired by superposed representation theory, we propose leveraging latent superposed reasoning to integrate multiple candidate semantics and maintain latent reasoning trajectories. The hypothesis is that reliance on discrete textual inputs may drive the model toward sequential explicit reasoning, underutilizing dense contextual cues during high-entropy reasoning stages. Therefore, we propose constructing rich semantic representations from the token probability distributions to enhance in-context reasoning. With this goal, we present Latent Entropy-Aware Decoding (LEAD), an efficient plug-and-play decoding strategy that leverages semantic context to achieve reliable reasoning. The heart of our method lies in entropy-aware reasoning mode switching. The model employs probability-weighted continuous embeddings under high-entropy states and transitions back to discrete token embeddings as entropy decreases. Moreover, we propose a prior-guided visual anchor injection strategy that encourages the model to focus on visual information. Extensive experiments show that LEAD effectively mitigates hallucinations across various MLRMs on multiple benchmarks.

  9. Online Experiential Learning for Language Models

    The prevailing paradigm for improving large language models relies on offline training with human annotations or simulated environments, leaving the rich experience accumulated during real-world deployment entirely unexploited. We propose Online Experiential Learning (OEL), a framework that enables language models to continuously improve from their own deployment experience. OEL operates in two stages: first, transferable experiential knowledge is extracted and accumulated from interaction trajectories collected on the user side; second, this knowledge is consolidated into model parameters via on-policy context distillation, requiring no access to the user-side environment. The two stages are iterated to form an online learning loop, where the improved model collects higher-quality trajectories that yield richer experiential knowledge for subsequent rounds. We evaluate OEL on text-based game environments across multiple model scales and both thinking and non-thinking variants. OEL achieves consistent improvements over successive iterations, enhancing both task accuracy and token efficiency while preserving out-of-distribution performance. Our analysis further shows that extracted experiential knowledge is significantly more effective than raw trajectories, and that on-policy consistency between the knowledge source and the policy model is critical for effective learning.

  10. FinToolBench: Evaluating LLM Agents for Real-World Financial Tool Use

    The integration of Large Language Models (LLMs) into the financial domain is driving a paradigm shift from passive information retrieval to dynamic, agentic interaction. While general-purpose tool learning has witnessed a surge in benchmarks, the financial sector, characterized by high stakes, strict compliance, and rapid data volatility, remains critically underserved. Existing financial evaluations predominantly focus on static textual analysis or document-based QA, ignoring the complex reality of tool execution. Conversely, general tool benchmarks lack the domain-specific rigor required for finance, often relying on toy environments or a negligible number of financial APIs. To bridge this gap, we introduce FinToolBench, the first real-world, runnable benchmark dedicated to evaluating financial tool learning agents. Unlike prior works limited to a handful of mock tools, FinToolBench establishes a realistic ecosystem coupling 760 executable financial tools with 295 rigorous, tool-required queries. We propose a novel evaluation framework that goes beyond binary execution success, assessing agents on finance-critical dimensions: timeliness, intent type, and regulatory domain alignment. Furthermore, we present FATR, a finance-aware tool retrieval and reasoning baseline that enhances stability and compliance. By providing the first testbed for auditable, agentic financial execution, FinToolBench sets a new standard for trustworthy AI in finance. The tool manifest, execution environment, and evaluation code will be open-sourced to facilitate future research.

  11. WiT: Waypoint Diffusion Transformers via Trajectory Conflict Navigation

    While recent Flow Matching models avoid the reconstruction bottlenecks of latent autoencoders by operating directly in pixel space, the lack of semantic continuity in the pixel manifold severely intertwines optimal transport paths. This induces severe trajectory conflicts near intersections, yielding sub-optimal solutions. Rather than bypassing this issue via information-lossy latent representations, we directly untangle the pixel-space trajectories by proposing Waypoint Diffusion Transformers (WiT). WiT factorizes the continuous vector field via intermediate semantic waypoints projected from pre-trained vision models. It effectively disentangles the generation trajectories by breaking the optimal transport into prior-to-waypoint and waypoint-to-pixel segments. Specifically, during the iterative denoising process, a lightweight generator dynamically infers these intermediate waypoints from the current noisy state. They then continuously condition the primary diffusion transformer via the Just-Pixel AdaLN mechanism, steering the evolution towards the next state, ultimately yielding the final RGB pixels. Evaluated on ImageNet 256x256, WiT beats strong pixel-space baselines, accelerating JiT training convergence by 2.2x. Code will be publicly released at https://github.com/hainuo-wang/WiT.git.

  12. Rethinking UMM Visual Generation: Masked Modeling for Efficient Image-Only Pre-training

    Unified Multimodal Models (UMMs) are often constrained by the pre-training of their visual generation components, which typically relies on inefficient paradigms and scarce, high-quality text-image paired data. In this paper, we systematically analyze pre-training recipes for UMM visual generation and identify these two issues as the major bottlenecks. To address them, we propose Image-Only Training for UMMs (IOMM), a data-efficient two-stage training framework. The first stage pre-trains the visual generative component exclusively using abundant unlabeled image-only data, thereby removing the dependency on paired data for this costly phase. The second stage fine-tunes the model using a mixture of unlabeled images and a small curated set of text-image pairs, leading to improved instruction alignment and generative quality. Extensive experiments show that IOMM not only improves training efficiency but also achieves state-of-the-art (SOTA) performance. For example, our IOMM-B (3.6B) model was trained from scratch using only sim 1050 H800 GPU hours (with the vast majority, 1000 hours, dedicated to the efficient image-only pre-training stage). It achieves 0.89 on GenEval and 0.55 on WISE--surpassing strong baselines such as BAGEL-7B (0.82 & 0.55) and BLIP3-o-4B (0.84 & 0.50). Code is available https://github.com/LINs-lab/IOMM{https://github.com/LINs-lab/IOMM}.

  13. MEMO: Memory-Augmented Model Context Optimization for Robust Multi-Turn Multi-Agent LLM Games

    Multi-turn, multi-agent LLM game evaluations often exhibit substantial run-to-run variance. In long-horizon interactions, small early deviations compound across turns and are amplified by multi-agent coupling. This biases win rate estimates and makes rankings unreliable across repeated tournaments. Prompt choice worsens this further by producing different effective policies. We address both instability and underperformance with MEMO (Memory-augmented MOdel context optimization), a self-play framework that optimizes inference-time context by coupling retention and exploration. Retention maintains a persistent memory bank that stores structured insights from self-play trajectories and injects them as priors during later play. Exploration runs tournament-style prompt evolution with uncertainty-aware selection via TrueSkill, and uses prioritized replay to revisit rare and decisive states. Across five text-based games, MEMO raises mean win rate from 25.1% to 49.5% for GPT-4o-mini and from 20.9% to 44.3% for Qwen-2.5-7B-Instruct, using 2,000 self-play games per task. Run-to-run variance also drops, giving more stable rankings across prompt variations. These results suggest that multi-agent LLM game performance and robustness have substantial room for improvement through context optimization. MEMO achieves the largest gains in negotiation and imperfect-information games, while RL remains more effective in perfect-information settings.

  14. GradMem: Learning to Write Context into Memory with Test-Time Gradient Descent

    Many large language model applications require conditioning on long contexts. Transformers typically support this by storing a large per-layer KV-cache of past activations, which incurs substantial memory overhead. A desirable alternative is ompressive memory: read a context once, store it in a compact state, and answer many queries from that state. We study this in a context removal setting, where the model must generate an answer without access to the original context at inference time. We introduce GradMem, which writes context into memory via per-sample test-time optimization. Given a context, GradMem performs a few steps of gradient descent on a small set of prefix memory tokens while keeping model weights frozen. GradMem explicitly optimizes a model-level self-supervised context reconstruction loss, resulting in a loss-driven write operation with iterative error correction, unlike forward-only methods. On associative key--value retrieval, GradMem outperforms forward-only memory writers with the same memory size, and additional gradient steps scale capacity much more effectively than repeated forward writes. We further show that GradMem transfers beyond synthetic benchmarks: with pretrained language models, it attains competitive results on natural language tasks including bAbI and SQuAD variants, relying only on information encoded in memory.

  15. SegviGen: Repurposing 3D Generative Model for Part Segmentation

    We introduce SegviGen, a framework that repurposes native 3D generative models for 3D part segmentation. Existing pipelines either lift strong 2D priors into 3D via distillation or multi-view mask aggregation, often suffering from cross-view inconsistency and blurred boundaries, or explore native 3D discriminative segmentation, which typically requires large-scale annotated 3D data and substantial training resources. In contrast, SegviGen leverages the structured priors encoded in pretrained 3D generative model to induce segmentation through distinctive part colorization, establishing a novel and efficient framework for part segmentation. Specifically, SegviGen encodes a 3D asset and predicts part-indicative colors on active voxels of a geometry-aligned reconstruction. It supports interactive part segmentation, full segmentation, and full segmentation with 2D guidance in a unified framework. Extensive experiments show that SegviGen improves over the prior state of the art by 40% on interactive part segmentation and by 15% on full segmentation, while using only 0.32% of the labeled training data. It demonstrates that pretrained 3D generative priors transfer effectively to 3D part segmentation, enabling strong performance with limited supervision. See our project page at https://fenghora.github.io/SegviGen-Page/.

Techmeme(15)

  1. Democratic voters in Illinois rejected most candidates supported by the crypto and AI industries, marking an unexpected defeat after industry PACs spent $18M+ (Bloomberg)

    Bloomberg : Democratic voters in Illinois rejected most candidates supported by the crypto and AI industries, marking an unexpected defeat after industry PACs spent $18M+ —  Democratic voters in Illinois this week rejected most candidates supported by the crypto and artificial intelligence industries …

  2. Micron reports Q2 revenue up 196% YoY to $23.9B, vs. $19.7B est., expects 2026 capex to exceed $25B, vs. $22.4B est., and forecasts Q3 revenue above estimates (Dina Bass/Bloomberg)

    Dina Bass / Bloomberg : Micron reports Q2 revenue up 196% YoY to $23.9B, vs. $19.7B est., expects 2026 capex to exceed $25B, vs. $22.4B est., and forecasts Q3 revenue above estimates —  Micron Technology Inc. warned that it will need to spend heavily on production to meet burgeoning demand, overshadowing …

  3. Compute startup Andromeda raised new funding from Paradigm at a $1.5B valuation, bringing Paradigm's total to $60M; it passed a $100M revenue run rate in 2025 (Upstarts Media)

    Upstarts Media : Compute startup Andromeda raised new funding from Paradigm at a $1.5B valuation, bringing Paradigm's total to $60M; it passed a $100M revenue run rate in 2025 —  Once a service spun up by AI investors Nat Friedman and Daniel Gross, Andromeda has now raised $60M from Paradigm to stand alone.

  4. During a Senate hearing, FBI Director Kash Patel says FBI does purchase "commercially available information" that can be used to track people's location history (Alfred Ng/Politico)

    Alfred Ng / Politico : During a Senate hearing, FBI Director Kash Patel says FBI does purchase “commercially available information” that can be used to track people's location history —  The U.S. Supreme Court has required law enforcement agencies to obtain a warrant for getting people's location data …

  5. Sources: Polymarket is looking to hire a chief risk officer following a CFTC demand; the company has expanded its legal team in recent months (Bernard Goyder/Bloomberg)

    Bernard Goyder / Bloomberg : Sources: Polymarket is looking to hire a chief risk officer following a CFTC demand; the company has expanded its legal team in recent months —  Prediction markets platform Polymarket is looking to hire a chief risk officer as it works to expand its regulated business in the US, according to people familiar with the matter.

  6. Meta makes its Meta Lab NYC pop-up on Fifth Avenue a permanent flagship location after signing a 10-year lease; it opened its first flagship store in LA in 2025 (Kanika Talwar/WWD)

    Kanika Talwar / WWD : Meta makes its Meta Lab NYC pop-up on Fifth Avenue a permanent flagship location after signing a 10-year lease; it opened its first flagship store in LA in 2025 —  The tech company has signed a decade-long lease for its Meta Lab town house on Fifth Avenue.

  7. Sources: Coinbase and Zerohash are among the companies vying to issue Cloudflare's stablecoin, set to launch this year to support payments for the "agentic web" (Yueqi Yang/The Information)

    Yueqi Yang / The Information : Sources: Coinbase and Zerohash are among the companies vying to issue Cloudflare's stablecoin, set to launch this year to support payments for the “agentic web” —  Crypto exchange Coinbase is moving fast to build the infrastructure that allows AI agents to make payments …

  8. Director Coerte Voorhees says he is using AI to feature Val Kilmer in "a significant part" in an indie film, with the cooperation of the late actor's estate (Brent Lang/Variety)

    Brent Lang / Variety : Director Coerte Voorhees says he is using AI to feature Val Kilmer in “a significant part” in an indie film, with the cooperation of the late actor's estate —  Five years prior to his death in 2025, Val Kilmer was cast as Father Fintan, a Catholic priest and Native American spiritualist, in “As Deep as the Grave.”

  9. Facebook launches Creator Fast Track, offering big Instagram, TikTok and YouTube creators guaranteed monthly pay and boosted reach to post on Facebook (Zach Vallese/CNBC)

    Zach Vallese / CNBC : Facebook launches Creator Fast Track, offering big Instagram, TikTok and YouTube creators guaranteed monthly pay and boosted reach to post on Facebook —  Meta on Wednesday launched a new program aimed at luring top creators from TikTok and YouTube to Facebook, offering guaranteed pay and boosted reach.

  10. Deezer reports a net income of €9M in 2025, its first profit since its 2007 founding, while revenue fell 1.4% YoY to €534M (Financial Times)

    Financial Times : Deezer reports a net income of €9M in 2025, its first profit since its 2007 founding, while revenue fell 1.4% YoY to €534M —  Industry under threat from fraudsters uploading and repeatedly playing tracks created by AI to extract royalties

  11. Ramp data: Anthropic is capturing ~73% of all spending among companies buying AI tools for the first time, up from a 50/50 split with OpenAI in January (Madison Mills/Axios)

    Madison Mills / Axios : Ramp data: Anthropic is capturing ~73% of all spending among companies buying AI tools for the first time, up from a 50/50 split with OpenAI in January —  Anthropic is now capturing over 73% of all spending among companies buying AI tools for the first time, according to customer data from Ramp.

  12. Sources: several tech companies, including OpenAI, are encouraging the DOD behind the scenes to back away from designating Anthropic a "supply chain risk" (Mike Isaac/New York Times)

    Mike Isaac / New York Times : Sources: several tech companies, including OpenAI, are encouraging the DOD behind the scenes to back away from designating Anthropic a “supply chain risk” —  Tech companies have been reluctant to directly confront Trump administration officials over their contract feud with the A.I. start-up.

  13. Israeli startup Raven, which offers runtime monitoring and intervention in attacks as they unfold inside applications, raised a $20M seed (CTech)

    CTech : Israeli startup Raven, which offers runtime monitoring and intervention in attacks as they unfold inside applications, raised a $20M seed —  The Israeli startup aims to secure applications at runtime as threats accelerate.  —  Cybersecurity startup Raven has raised $20 million …

  14. Paraform, which connects tech companies with specialized recruiters, raised a $40M Series B led by Scale VP (Natalie Breymeyer/Axios)

    Natalie Breymeyer / Axios : Paraform, which connects tech companies with specialized recruiters, raised a $40M Series B led by Scale VP —  Paraform, a platform that connects tech companies with specialized recruiters, raised a $40 million Series B, CEO John Kim tells Axios exclusively.

  15. The UK government withdraws a proposal to let AI companies train on copyrighted works unless creators opt out, after backlash from artists like Dua Lipa (Graham Fraser/BBC)

    Graham Fraser / BBC : The UK government withdraws a proposal to let AI companies train on copyrighted works unless creators opt out, after backlash from artists like Dua Lipa —  The UK government has backtracked on its position on copyright and AI, stating it must take time to “get this right”.

Solidot(15)

  1. 瑞士构建 BGP 的安全替代

    边界网关路由(BGP)不是为安全设计的,而是为构成互联网的数以千计的自治系统之间大规模快速路由数据包设计的。过去四十年,BGP 运作良好,但其安全缺陷也日益显现。为堵上漏洞,BGP 引入了一系列补丁和扩展如 Resource Public Key Infrastructure (RPKI)、BGPsec 和 RPKI-based Route Origin Authorization (ROA),但无法从根本上解决问题。瑞士苏黎世联邦理工学院开发的 SCION——代表 Scalability, Control, and Isolation On Next-Generation Networks——尝试从根本上改变互联网的路由架构,提供一种更安全的替代。SCION 的首席架构师 Adrian Perrig 是苏黎世联邦理工的计算机科学教授,一直致力于提升互联网的安全。他发现安全无法拼拼凑凑,必须彻底改变设计。SCION 尝试通过三个关联机制解决 BGP 的安全缺陷:其一是多路径路由,两点之间能同时建立数十条甚至数百条并行路径,一条路径发生故障,系统会在几毫秒内完成重路由;其二是不依赖证书颁发机构的隔离域名 ISD 机制;其三是加密路径验证,路径上的每个路由器都提供一个加密签名。瑞士银行已成功测试了 SCION。

  2. GTC 2026 重磅 AI 会议推荐:注册观看还有机会获得 NVIDIA 定制装备

    重磅一:中国创业生态精彩会议 错过首播也无妨,复制下方链接即可直达回看页面。 GTC 2026 创业企业会议特辑深度聚焦中国创业生态,干货满满,不容错过。 线上演讲:《十载相伴,NVIDIA 赋能创业公司在 AI 时代加速前行》 https://www.nvidia.cn/gtc/session-catalog/sessions/gtc26-s81981/?ncid=so-othe-950414 线上演讲:《基于 NVIDIA 全栈技术打造代理式 AI 与物理 AI 的未来基石》 https://www.nvidia.cn/gtc/session-catalog/sessions/gtc26-s81974/?ncid=so-othe-950414 圆桌论坛:《洞察 2026 中国 AI 市场 — AI 智能体和物理 AI 浪潮下的创业风口》 https://www.nvidia.cn/gtc/session-catalog/sessions/gtc26-s81846/?ncid=so-othe-950414

  3. 韩国游戏发行商 CEO 为避免支付合同承诺的 2.5 亿美元而求助于 ChatGPT

    Unknown Worlds Entertainment 是知名水下生存游戏《Subnautica》的开发商,2021 年韩国游戏发行商 Krafton 以 5 亿美元收购了该游戏工作室,并在合同中承诺如果续作《Subnautica 2》销量足够好将额外支付 2.5 亿美元。Krafton 内部对《Subnautica 2》的销量预测相当乐观,因此 2.5 亿美元看起来不得不兑现了。然而 CEO Changhan Kim 不想支付这笔费用,他因此求助于 AI 聊天机器人 ChatGPT 而不是公司法务讨论如何避免支付 2.5 亿美元。公司法务认为奖金取消不了,但 Kim 在 ChatGPT 的帮助下设计了一个方案以莫须有理由突然解雇了 Unknown Worlds 的主要高管。被解雇的高管提起诉讼曝光了 ChatGPT 的阴谋。法官本周裁决 Unknown Worlds 前 CEO Ted Gill 恢复原职。拖延了很久的《Subnautica 2》预计将于今年 5 月发布抢先体验版本(early access)。

  4. 法官裁决苹果可以以任何理由下架应用

    Musi 是一款免费音乐串流应用,只有 iOS 版本,它本身并不托管音乐,而是利用了 YouTube 上的音乐。它诞生于 2013 年,2024 年 9 月因 YouTube 的投诉而被苹果下架,下架前其下载量超过 6600 万次。Musi 有广告,用户可以一次性支付 5.99 美元移除广告。在下架之后,Musi 起诉了苹果。加州北区联邦地区法院法官 Eumi Lee 本周裁决 Musi 败诉。法官称,根据苹果的开发者协议 Developer Program License Agreement(DPLA),苹果“不论有没有原因”都可以下架应用,它只需要发出终止通知。苹果向 Musi 发出了终止通知,因此下架并没有违反 DPLA。法官还抨击 Musi 的律师在案件中捏造事实,下令由该律所支付诉讼费用。

  5. 社会出身与儿童的努力程度密切相关

    哪些孩子在学校更努力?他们的努力程度与其社会出身有何关系?根据发表在《American Sociological Review》上的一项研究,来自优渥家庭的学龄儿童比来自贫困环境的孩子表现出更高的认知努力,尤其是在没有奖励且仅靠内在动力的情况下。尽管如此,两组之间的差距并不大,且可以通过激励措施得到补偿:当提供小奖品(如玩具或社会认可)时,资源匮乏家庭的孩子所表现出的投入程度与家庭环境较好的同龄人非常接近。研究人员指出,与努力相关的行为受社会环境制约,因为家庭可利用资源以及儿童在日常生活中感受到的安全感起着至关重要的作用。在匮乏的环境中成长(缺乏经济手段或缺乏父母的关注时间)会使个体难以长时间集中精力完成某项任务。

  6. 小行星龙宫含有所有五种核碱基

    日本研究团队在《Nature Astronomy》期刊上报告,探测器隼鸟 2 号从小行星龙宫采集的沙粒样本中含有地球生命所需的五种核碱基,再次为小行星陨石将生命所需成分带到地球的假说提供了佐证。隼鸟 2 号探测器于 2018 年造访了小行星龙宫,向其表面发射了两枚金属弹丸——一枚较小,一枚较大——收集了撞击产生的碎片。隼鸟 2 号于 2020 年携带样本返回地球,此后研究人员一直在对其进行分析。日本北海道大学的 Yasuhiro Oba 和同事分析了两个样本,一个来自小行星表面,另一个由弹丸撞击出的地下物质组成。在两个样本中,研究团队都发现了全部五种主要核碱基,核碱基与糖和磷酸结合构成了核酸 DNA 和 RNA。

  7. 英伟达的 DLSS 5 引发争议和批评

    英伟达演示了计划于今年晚些时候推出的深度学习超级采样技术 DLSS 的新版本,结果在玩家中间引发了广泛争议和批评,因为 DLSS 5 在重构图像过程中戏剧性的改变了游戏画面,为游戏画面加入了一层 AI 滤镜,让游戏中的人物变得面目全非。DLSS 5 在社交媒体上引发了玩家制作大量梗图进行嘲讽,在 reddit 上大量相关讨论被删除(可能是英伟达在公关),就像微软被称为 Microslop,玩家现在开始称英伟达为 Slopvidia。

  8. 太阳可能在几十亿年前从银河中心迁移到外围

    最新证据显示,约在 40-60 亿年前,太阳可能曾参与一次大规模的恒星迁移事件。一群与太阳性质非常相似的太阳类恒星(solar twins)一同离开银河系核心区域并向外迁移。天文学家利用 ESA Gaia 卫星的观测数据进行分析,建立了一份前所未有精确的恒星目录。研究显示,太阳目前位于银河系的位置并非偶然,而可能是这次大规模恒星迁移事件的一部分。太阳约在 46 亿年前诞生,而当时太阳的位置比今天更接近银河中心超过一万光年。恒星化学成分的研究支持这一推论,但这个结果长期让科学家感到困惑。观测显示银河中心存在一个巨大的棒状结构,它会形成所谓的共转屏障,使恒星​​难以从银河中心区域迁移到如此遥远的位置。为了解答这个问题,来自日本的研究团队对银河系中类似太阳的恒星展开了大规模研究。这些恒星在温度、表面重力与化学组成上都与太阳极为相似。研究团队从 Gaia 的资料中挑选出 6594 颗太阳类恒星建立目录。透过这份庞大的资料,研究人员得以重建目前最精确的恒星年龄分布。分析结果显示,在 40-60 亿年前出现一个明显且宽广的年龄峰值,显示一群年龄相近的恒星分布在距离银河中心相似的位置。这代表太阳并非单独迁移,而是属于一次大规模恒星外移事件的一部分。

  9. 微软允许 Windows 11 用户在安装过程中重命名主文件夹名称

    众所周知,Windows 在安装过程中并不允许用户重命名主文件夹名称,而是根据用户账号或邮箱地址自动生成名称。去年微软开始测试允许用户重命名主文件夹名称,但非常繁琐。现在微软终于将重命名主文件夹名称作为安装流程的一部分提供给用户。微软释出了预览版本 Windows 11 Insider Preview Build 26220.8062,在安装流程的“设备名称”页面包含了一个重命名主文件夹名称选项,如果用户跳过这一步骤,那么主文件夹仍然会使用默认名称。命名文件夹名称需要遵循微软的命名规定。

  10. 德国法庭裁决 TCL 的 QLED 不是真的 QLED

    德国的一家法庭裁决 TCL 误导消费者,它的多款宣称“量子点电视(QLED)”的产品并不是真的 QLED,没有提供 QLED 电视应有的色彩还原。法院命令 TCL 停止在德国宣传或销售相关型号的电视机。相关诉讼由韩国公司韩松化学提起,它是 TCL 竞争对手三星的合作伙伴。韩松化学委托进行的测试显示,三款以量子点名义出售的 TV 都未检测出铟和镉,它们是不可或缺的量子点材料。TCL 对测试结果提出异议,称量子点含量因供应商而异,它公布了自己的测试结果。TCL 的测试结果与韩松化学的测试结果相矛盾,但双方采用了不同的测试方法:TCL 的测试侧重于其使用的量子点薄膜,而韩松化学测试的是 TCL 电视机。韩松化学在包括美国在内的多国提起了针对 TCL 的诉讼,另一家中国电视制造商海信也面临类似的诉讼。

  11. Marknote 1.5 释出

    基于 Markdown 的笔记管理应用 Marknote 释出了 v1.5。主要新特性包括:新的 Source Mode 模式,不使用富文本 WYSIWYG 界面直接编辑 Markdown 内容;支持维基风格的笔记文档链接,支持跨笔记查找;简化笔记和笔记本管理,每个笔记本会显示包含的笔记数量,如需要在笔记本之间移动笔记可通过拖放完成;Duplicate Note 操作可创建模板复制现有笔记;KRunner 插件;等等。

  12. kagi 翻译支持翻译到 LinkedIn Speak

    职业社交网络 LinkedIn 的用户已经形成了一套独特的语言风格,这种风格被称为 LinkedIn Speak,其特点是能将任何琐事自我包装成积极向上的宏大叙事。举例来说:你失业了,但用 LinkedIn Speak 写出来变成了“开启了人生的新篇章”,去五百强企业做清洁工变成了荣幸加盟;等等。在中国,阿里巴巴的职场语言套话如“赋能、闭环、沉淀、生态”可能与 LinkedIn Speak 最为相似。现在,kagi 翻译工具加入了对 LinkedIn Speak 输出的支持,让任何人可以通过自然语言输出职场套话。

  13. 2026 年 Debian 项目领导人竞选开始,只有一名候选人

    一年一度的 Debian 项目领导人(DPL)竞选启动,今年只有一位候选人 Sruthi Chandran——她是一位来自印度的图书管理员,2025 年的 DPL Andreas Tille 没有再次参选。竞选期持续到 4 月 3 日,投票期从 4 月 4 日持续到 4 月 17 日。Debian 项目的选民们将要在同意 Sruthi Chandran 担任 DPL 或不同意(以上皆非)两个选项中进行投票。DPL 选举采用的是孔多塞投票法。

  14. GIMP 3.2 释出

    图形编辑器项目 GIMP 释出了 v3.2。此举是该项目自 GIMP 3.0 发布之后加快版本发布计划的一部分。从 GIMP 2.0 到 3.0,项目经历了逾 20 年的时间,开发者不希望让用户等待六七年才等到一个小版本更新,等二十年才有一个大版本。GIMP 3.2 新特性包括:MyPaint Brush 画笔工具新增 20 种新画笔;overwrite 绘画模式;新的和升级的文件格式;UI 改进;新的非破坏性图层,使用 Link Layers 整合外部图像,缩放、旋转和变换图像而不会损失质量或清晰度,源文件修改后 Link Layers 会同步更新,Path 工具能创建 Vector Layers;等等。

  15. GTA Wiki 从 Fandom 迁移到独立维基网站

    在 Minecraft Wiki 之后,另一个大型维基社区 GTA Wiki 正从 Fandom 迁移到独立维基网站 gta.wiki。Fandom 是吉米·威尔士等人联合创办的商业化维基托管平台,因广告和使用体验下降而招致了用户不满。GTA Wiki 称,今年 2 月 Fandom 进行了重组,任命了一名亲 AI 的 CEO,而用户对 Fandom 最大的抱怨包括广告太多和内容政策过于严厉。GTA Wiki 在 Fandom 的内容将全部复制到 gta.wiki,编辑和管理员可以选择留在旧平台或迁移到新平台。