DIGEST · 2026-03-17

OrangeBot.AI Digest — 2026-03-17

81 headlines across 8 sources, aggregated for this day.

Hacker News(15)

  1. Python 3.15's JIT is now back on track (fidget-spinner.github.io)
  2. A Decade of Slug (terathon.com)
  3. If you thought code writing speed was your problem you have bigger problems (andrewmurphy.io)
  4. Illinois Introducing Operating System Account Age Bill (www.ilga.gov)
  5. GPT‑5.4 Mini and Nano (openai.com)
  6. Microsoft's 'unhackable' Xbox One has been hacked by 'Bliss' (www.tomshardware.com)
  7. FFmpeg 8.1 (ffmpeg.org)
  8. Give Django your time and money, not your tokens (www.better-simple.com)
  9. Node.js needs a virtual file system (blog.platformatic.dev)
  10. A proposal to classify happiness as a psychiatric disorder (1992) (pmc.ncbi.nlm.nih.gov)
  11. Silicon Valley's "Pronatalists" Killed WFH. The Strait of Hormuz Brought It Back (www.governance.fyi)
  12. Reddit User Uncovers Who Is Behind Meta's $2B Lobbying for Age Verification Tech (www.gadgetreview.com)
  13. Kagi Small Web (kagi.com)
  14. The unlikely story of Teardown Multiplayer (blog.voxagon.se)
  15. Sci-Fi Short Film “There Is No Antimemetics Division” [video] (www.youtube.com)

GitHub Trending(6)

  1. obra / superpowers
  2. codecrafters-io / build-your-own-x
  3. abhigyanpatwari / GitNexus
  4. langchain-ai / deepagents
  5. jarrodwatts / claude-hud
  6. cloudflare / workerd

Product Hunt(15)

  1. DLSS 5

    The GPT moment for real-time computer graphics

  2. Sokosumi

    Marketing agents that research, plan, and manage for you

  3. Easy App Reports

    Bring your app's data to Looker Studio, BigQuery, or AI

  4. Kipps.AI Campaign

    Lead Qualification, Bulk Outreach and Anniversary’s Reminder

  5. Angy

    Multi‑agent pipelines w/ AI‑driven scheduling + safety check

  6. Ocean Orchestrator

    Run AI jobs from your IDE with a one-click workflow

  7. Lightning Rod

    Turn real-world data into training datasets fast

  8. My Computer by Manus AI

    Automate files, apps, and workflows with Manus Desktop

  9. JusRecruit

    AI ATS that handles phone screens + first-round interviews

  10. Codex Subagents

    Parallel custom agents for complex tasks

  11. discli

    Discord CLI for AI agents and humans

  12. OpenFlags

    Fast, self-hosted, edge-ready feature flags for modern teams

  13. Kira 4.0

    Turn your friends into shareable content

  14. Folderly

    Get revenue from every email campaign with 99.9% inbox rate

  15. mTarsier

    Open-source platform for managing MCP servers and clients

Hugging Face(15)

  1. AI Can Learn Scientific Taste

    Great scientists have strong judgement and foresight, closely tied to what we call scientific taste. Here, we use the term to refer to the capacity to judge and propose research ideas with high potential impact. However, most relative research focuses on improving an AI scientist's executive capability, while enhancing an AI's scientific taste remains underexplored. In this work, we propose Reinforcement Learning from Community Feedback (RLCF), a training paradigm that uses large-scale community signals as supervision, and formulate scientific taste learning as a preference modeling and alignment problem. For preference modeling, we train Scientific Judge on 700K field- and time-matched pairs of high- vs. low-citation papers to judge ideas. For preference alignment, using Scientific Judge as a reward model, we train a policy model, Scientific Thinker, to propose research ideas with high potential impact. Experiments show Scientific Judge outperforms SOTA LLMs (e.g., GPT-5.2, Gemini 3 Pro) and generalizes to future-year test, unseen fields, and peer-review preference. Furthermore, Scientific Thinker proposes research ideas with higher potential impact than baselines. Our findings show that AI can learn scientific taste, marking a key step toward reaching human-level AI scientists.

  2. OpenSeeker: Democratizing Frontier Search Agents by Fully Open-Sourcing Training Data

    Deep search capabilities have become an indispensable competency for frontier Large Language Model (LLM) agents, yet the development of high-performance search agents remains dominated by industrial giants due to a lack of transparent, high-quality training data. This persistent data scarcity has fundamentally hindered the progress of the broader research community in developing and innovating within this domain. To bridge this gap, we introduce OpenSeeker, the first fully open-source search agent (i.e., model and data) that achieves frontier-level performance through two core technical innovations: (1) Fact-grounded scalable controllable QA synthesis, which reverse-engineers the web graph via topological expansion and entity obfuscation to generate complex, multi-hop reasoning tasks with controllable coverage and complexity. (2) Denoised trajectory synthesis, which employs a retrospective summarization mechanism to denoise the trajectory, therefore promoting the teacher LLMs to generate high-quality actions. Experimental results demonstrate that OpenSeeker, trained (a single training run) on only 11.7k synthesized samples, achieves state-of-the-art performance across multiple benchmarks including BrowseComp, BrowseComp-ZH, xbench-DeepSearch, and WideSearch. Notably, trained with simple SFT, OpenSeeker significantly outperforms the second-best fully open-source agent DeepDive (e.g., 29.5% v.s. 15.3% on BrowseComp), and even surpasses industrial competitors such as Tongyi DeepResearch (trained via extensive continual pre-training, SFT, and RL) on BrowseComp-ZH (48.4% v.s. 46.7%). We fully open-source the complete training dataset and the model weights to democratize frontier search agent research and foster a more transparent, collaborative ecosystem.

  3. EnterpriseOps-Gym: Environments and Evaluations for Stateful Agentic Planning and Tool Use in Enterprise Settings

    Large language models are shifting from passive information providers to active agents intended for complex workflows. However, their deployment as reliable AI workers in enterprise is stalled by benchmarks that fail to capture the intricacies of professional environments, specifically, the need for long-horizon planning amidst persistent state changes and strict access protocols. In this work, we introduce EnterpriseOps-Gym, a benchmark designed to evaluate agentic planning in realistic enterprise settings. Specifically, EnterpriseOps-Gym features a containerized sandbox with 164 database tables and 512 functional tools to mimic real-world search friction. Within this environment, agents are evaluated on 1,150 expert-curated tasks across eight mission-critical verticals (including Customer Service, HR, and IT). Our evaluation of 14 frontier models reveals critical limitations in state-of-the-art models: the top-performing Claude Opus 4.5 achieves only 37.4% success. Further analysis shows that providing oracle human plans improves performance by 14-35 percentage points, pinpointing strategic reasoning as the primary bottleneck. Additionally, agents frequently fail to refuse infeasible tasks (best model achieves 53.9%), leading to unintended and potentially harmful side effects. Our findings underscore that current agents are not yet ready for autonomous enterprise deployment. More broadly, EnterpriseOps-Gym provides a concrete testbed to advance the robustness of agentic planning in professional workflows.

  4. Grounding World Simulation Models in a Real-World Metropolis

    What if a world simulation model could render not an imagined environment but a city that actually exists? Prior generative world models synthesize visually plausible yet artificial environments by imagining all content. We present Seoul World Model (SWM), a city-scale world model grounded in the real city of Seoul. SWM anchors autoregressive video generation through retrieval-augmented conditioning on nearby street-view images. However, this design introduces several challenges, including temporal misalignment between retrieved references and the dynamic target scene, limited trajectory diversity and data sparsity from vehicle-mounted captures at sparse intervals. We address these challenges through cross-temporal pairing, a large-scale synthetic dataset enabling diverse camera trajectories, and a view interpolation pipeline that synthesizes coherent training videos from sparse street-view images. We further introduce a Virtual Lookahead Sink to stabilize long-horizon generation by continuously re-grounding each chunk to a retrieved image at a future location. We evaluate SWM against recent video world models across three cities: Seoul, Busan, and Ann Arbor. SWM outperforms existing methods in generating spatially faithful, temporally consistent, long-horizon videos grounded in actual urban environments over trajectories reaching hundreds of meters, while supporting diverse camera movements and text-prompted scenario variations.

  5. HSImul3R: Physics-in-the-Loop Reconstruction of Simulation-Ready Human-Scene Interactions

    We present HSImul3R, a unified framework for simulation-ready 3D reconstruction of human-scene interactions (HSI) from casual captures, including sparse-view images and monocular videos. Existing methods suffer from a perception-simulation gap: visually plausible reconstructions often violate physical constraints, leading to instability in physics engines and failure in embodied AI applications. To bridge this gap, we introduce a physically-grounded bi-directional optimization pipeline that treats the physics simulator as an active supervisor to jointly refine human dynamics and scene geometry. In the forward direction, we employ Scene-targeted Reinforcement Learning to optimize human motion under dual supervision of motion fidelity and contact stability. In the reverse direction, we propose Direct Simulation Reward Optimization, which leverages simulation feedback on gravitational stability and interaction success to refine scene geometry. We further present HSIBench, a new benchmark with diverse objects and interaction scenarios. Extensive experiments demonstrate that HSImul3R produces the first stable, simulation-ready HSI reconstructions and can be directly deployed to real-world humanoid robots.

  6. Attention Residuals

    Residual connections with PreNorm are standard in modern LLMs, yet they accumulate all layer outputs with fixed unit weights. This uniform aggregation causes uncontrolled hidden-state growth with depth, progressively diluting each layer's contribution. We propose Attention Residuals (AttnRes), which replaces this fixed accumulation with softmax attention over preceding layer outputs, allowing each layer to selectively aggregate earlier representations with learned, input-dependent weights. To address the memory and communication overhead of attending over all preceding layer outputs for large-scale model training, we introduce Block AttnRes, which partitions layers into blocks and attends over block-level representations, reducing the memory footprint while preserving most of the gains of full AttnRes. Combined with cache-based pipeline communication and a two-phase computation strategy, Block AttnRes becomes a practical drop-in replacement for standard residual connections with minimal overhead. Scaling law experiments confirm that the improvement is consistent across model sizes, and ablations validate the benefit of content-dependent depth-wise selection. We further integrate AttnRes into the Kimi Linear architecture (48B total / 3B activated parameters) and pre-train on 1.4T tokens, where AttnRes mitigates PreNorm dilution, yielding more uniform output magnitudes and gradient distribution across depth, and improves downstream performance across all evaluated tasks.

  7. Mixture-of-Depths Attention

    Scaling depth is a key driver for large language models (LLMs). Yet, as LLMs become deeper, they often suffer from signal degradation: informative features formed in shallow layers are gradually diluted by repeated residual updates, making them harder to recover in deeper layers. We introduce mixture-of-depths attention (MoDA), a mechanism that allows each attention head to attend to sequence KV pairs at the current layer and depth KV pairs from preceding layers. We further describe a hardware-efficient algorithm for MoDA that resolves non-contiguous memory-access patterns, achieving 97.3% of FlashAttention-2's efficiency at a sequence length of 64K. Experiments on 1.5B-parameter models demonstrate that MoDA consistently outperforms strong baselines. Notably, it improves average perplexity by 0.2 across 10 validation benchmarks and increases average performance by 2.11% on 10 downstream tasks, with a negligible 3.7% FLOPs computational overhead. We also find that combining MoDA with post-norm yields better performance than using it with pre-norm. These results suggest that MoDA is a promising primitive for depth scaling. Code is released at https://github.com/hustvl/MoDA .

  8. Effective Distillation to Hybrid xLSTM Architectures

    There have been numerous attempts to distill quadratic attention-based large language models (LLMs) into sub-quadratic linearized architectures. However, despite extensive research, such distilled models often fail to match the performance of their teacher LLMs on various downstream tasks. We set out the goal of lossless distillation, which we define in terms of tolerance-corrected Win-and-Tie rates between student and teacher on sets of tasks. To this end, we introduce an effective distillation pipeline for xLSTM-based students. We propose an additional merging stage, where individually linearized experts are combined into a single model. We show the effectiveness of this pipeline by distilling base and instruction-tuned models from the Llama, Qwen, and Olmo families. In many settings, our xLSTM-based students recover most of the teacher's performance, and even exceed it on some downstream tasks. Our contributions are an important step towards more energy-efficient and cost-effective replacements for transformer-based LLMs.

  9. Anatomy of a Lie: A Multi-Stage Diagnostic Framework for Tracing Hallucinations in Vision-Language Models

    Vision-Language Models (VLMs) frequently "hallucinate" - generate plausible yet factually incorrect statements - posing a critical barrier to their trustworthy deployment. In this work, we propose a new paradigm for diagnosing hallucinations, recasting them from static output errors into dynamic pathologies of a model's computational cognition. Our framework is grounded in a normative principle of computational rationality, allowing us to model a VLM's generation as a dynamic cognitive trajectory. We design a suite of information-theoretic probes that project this trajectory onto an interpretable, low-dimensional Cognitive State Space. Our central discovery is a governing principle we term the geometric-information duality: a cognitive trajectory's geometric abnormality within this space is fundamentally equivalent to its high information-theoretic surprisal. Hallucination detection is counts as a geometric anomaly detection problem. Evaluated across diverse settings - from rigorous binary QA (POPE) and comprehensive reasoning (MME) to unconstrained open-ended captioning (MS-COCO) - our framework achieves state-of-the-art performance. Crucially, it operates with high efficiency under weak supervision and remains highly robust even when calibration data is heavily contaminated. This approach enables a causal attribution of failures, mapping observable errors to distinct pathological states: perceptual instability (measured by Perceptual Entropy), logical-causal failure (measured by Inferential Conflict), and decisional ambiguity (measured by Decision Entropy). Ultimately, this opens a path toward building AI systems whose reasoning is transparent, auditable, and diagnosable by design.

  10. ViFeEdit: A Video-Free Tuner of Your Video Diffusion Transformer

    Diffusion Transformers (DiTs) have demonstrated remarkable scalability and quality in image and video generation, prompting growing interest in extending them to controllable generation and editing tasks. However, compared to the image counterparts, progress in video control and editing remains limited, mainly due to the scarcity of paired video data and the high computational cost of training video diffusion models. To address this issue, in this paper, we propose a video-free tuning framework termed ViFeEdit for video diffusion transformers. Without requiring any forms of video training data, ViFeEdit achieves versatile video generation and editing, adapted solely with 2D images. At the core of our approach is an architectural reparameterization that decouples spatial independence from the full 3D attention in modern video diffusion transformers, which enables visually faithful editing while maintaining temporal consistency with only minimal additional parameters. Moreover, this design operates in a dual-path pipeline with separate timestep embeddings for noise scheduling, exhibiting strong adaptability to diverse conditioning signals. Extensive experiments demonstrate that our method delivers promising results of controllable video generation and editing with only minimal training on 2D image data. Codes are available https://github.com/Lexie-YU/ViFeEdit.

  11. Safe and Scalable Web Agent Learning via Recreated Websites

    Training autonomous web agents is fundamentally limited by the environments they learn from: real-world websites are unsafe to explore, hard to reset, and rarely provide verifiable feedback. We propose VeriEnv, a framework that treats language models as environment creators, automatically cloning real-world websites into fully executable, verifiable synthetic environments. By exposing controlled internal access via a Python SDK, VeriEnv enables agents to self-generate tasks with deterministic, programmatically verifiable rewards, eliminating reliance on heuristic or LLM-based judges. This design decouples agent learning from unsafe real-world interaction while enabling scalable self-evolution through environment expansion. Through experiments on web agent benchmarks, we show that agents trained with VeriEnv generalize to unseen websites, achieve site-specific mastery through self-evolving training, and benefit from scaling the number of training environments. Code and resources will be released at https://github.com/kyle8581/VeriEnv upon acceptance.

  12. Make it SING: Analyzing Semantic Invariants in Classifiers

    All classifiers, including state-of-the-art vision models, possess invariants, partially rooted in the geometry of their linear mappings. These invariants, which reside in the null-space of the classifier, induce equivalent sets of inputs that map to identical outputs. The semantic content of these invariants remains vague, as existing approaches struggle to provide human-interpretable information. To address this gap, we present Semantic Interpretation of the Null-space Geometry (SING), a method that constructs equivalent images, with respect to the network, and assigns semantic interpretations to the available variations. We use a mapping from network features to multi-modal vision language models. This allows us to obtain natural language descriptions and visual examples of the induced semantic shifts. SING can be applied to a single image, uncovering local invariants, or to sets of images, allowing a breadth of statistical analysis at the class and model levels. For example, our method reveals that ResNet50 leaks relevant semantic attributes to the null space, whereas DinoViT, a ViT pretrained with self-supervised DINO, is superior in maintaining class semantics across the invariant space.

  13. TERMINATOR: Learning Optimal Exit Points for Early Stopping in Chain-of-Thought Reasoning

    Large Reasoning Models (LRMs) achieve impressive performance on complex reasoning tasks via Chain-of-Thought (CoT) reasoning, which enables them to generate intermediate thinking tokens before arriving at the final answer. However, LRMs often suffer from significant overthinking, spending excessive compute time even after the answer is generated early on. Prior work has identified the existence of an optimal reasoning length such that truncating reasoning at this point significantly shortens CoT outputs with virtually no change in performance. However, determining optimal CoT lengths for practical datasets is highly non-trivial as they are fully task and model-dependent. In this paper, we precisely address this and design TERMINATOR, an early-exit strategy for LRMs at inference to mitigate overthinking. The central idea underpinning TERMINATOR is that the first arrival of an LRM's final answer is often predictable, and we leverage these first answer positions to create a novel dataset of optimal reasoning lengths to train TERMINATOR. Powered by this approach, TERMINATOR achieves significant reductions in CoT lengths of 14%-55% on average across four challenging practical datasets: MATH-500, AIME 2025, HumanEval, and GPQA, whilst outperforming current state-of-the-art methods.

  14. WebVR: Benchmarking Multimodal LLMs for WebPage Recreation from Videos via Human-Aligned Visual Rubrics

    Existing web-generation benchmarks rely on text prompts or static screenshots as input. However, videos naturally convey richer signals such as interaction flow, transition timing, and motion continuity, which are essential for faithful webpage recreation. Despite this potential, video-conditioned webpage generation remains largely unexplored, with no dedicated benchmark for this task. To fill this gap, we introduce WebVR, a benchmark that evaluates whether MLLMs can faithfully recreate webpages from demonstration videos. WebVR contains 175 webpages across diverse categories, all constructed through a controlled synthesis pipeline rather than web crawling, ensuring varied and realistic demonstrations without overlap with existing online pages. We also design a fine-grained, human-aligned visual rubric that evaluates the generated webpages across multiple dimensions. Experiments on 19 models reveal substantial gaps in recreating fine-grained style and motion quality, while the rubric-based automatic evaluation achieves 96% agreement with human preferences. We release the dataset, evaluation toolkit, and baseline results to support future research on video-to-webpage generation.

  15. Understanding Reasoning in LLMs through Strategic Information Allocation under Uncertainty

    LLMs often exhibit Aha moments during reasoning, such as apparent self-correction following tokens like "Wait," yet their underlying mechanisms remain unclear. We introduce an information-theoretic framework that decomposes reasoning into procedural information and epistemic verbalization - the explicit externalization of uncertainty that supports downstream control actions. We show that purely procedural reasoning can become informationally stagnant, whereas epistemic verbalization enables continued information acquisition and is critical for achieving information sufficiency. Empirical results demonstrate that strong reasoning performance is driven by uncertainty externalization rather than specific surface tokens. Our framework unifies prior findings on Aha moments and post-training experiments, and offers insights for future reasoning model design.

Techmeme(15)

  1. You.com appoints Saahil Jain as new CTO after co-founder Bryan McCann left to Anthropic; You.com, which began in consumer search, is targeting enterprise more (Kevin McLaughlin/The Information)

    Kevin McLaughlin / The Information : You.com appoints Saahil Jain as new CTO after co-founder Bryan McCann left to Anthropic; You.com, which began in consumer search, is targeting enterprise more —  AI startup You.com, valued at $1.5 billion after a funding round in September, is changing its senior leadership as it focuses on helping businesses adopt AI.

  2. Mistral announces Mistral Forge to help enterprises build custom models actually trained on their own data, using Mistral open-weight models as a starting point (TechCrunch)

    TechCrunch : Mistral announces Mistral Forge to help enterprises build custom models actually trained on their own data, using Mistral open-weight models as a starting point —  Most enterprise AI projects fail not because companies lack the technology, but because the models they're using don't understand their business.

  3. Meta says Quest users will lose access to Meta Horizon Worlds on the headsets on June 15; access will continue on the Meta Horizon mobile app (Riley Griffin/Bloomberg)

    Riley Griffin / Bloomberg : Meta says Quest users will lose access to Meta Horizon Worlds on the headsets on June 15; access will continue on the Meta Horizon mobile app —  Meta Platforms Inc. said that users of its Quest headsets will lose access to Horizon Worlds, a virtual destination where cartoon versions of people …

  4. Jensen Huang says Nvidia is in the process of restarting manufacturing of its H200 chips for shipments to China and it has received orders from "many customers" (Ina Fried/Axios)

    Ina Fried / Axios : Jensen Huang says Nvidia is in the process of restarting manufacturing of its H200 chips for shipments to China and it has received orders from “many customers” —  Nvidia CEO Jensen Huang said the company is in the process of restarting manufacturing of its H200 chips for shipments to China.

  5. Robinhood Ventures Fund I discloses its first investments, buying $14.6M of Stripe shares and $20M of ElevenLabs' preferred stock in March (CoinDesk)

    CoinDesk : Robinhood Ventures Fund I discloses its first investments, buying $14.6M of Stripe shares and $20M of ElevenLabs' preferred stock in March —  The closed-end fund aims to give everyday investors exposure to private firms before they go public.  —  What to know:

  6. Sources: China is penalizing people tied to Meta's $2B Manus acquisition, including by apparently restricting Manus executives from leaving China for Singapore (New York Times)

    New York Times : Sources: China is penalizing people tied to Meta's $2B Manus acquisition, including by apparently restricting Manus executives from leaving China for Singapore —  The country appears to be cracking down on people linked to the acquisition of Manus, a Singapore company with Chinese roots, as President Trump prepares to visit Beijing.

  7. At an all-hands, Andy Jassy said he expects AI to help AWS reach $600B in annual sales by 2036, double his prior estimate; AWS had revenue of $128.7B in 2025 (Greg Bensinger/Reuters)

    Greg Bensinger / Reuters : At an all-hands, Andy Jassy said he expects AI to help AWS reach $600B in annual sales by 2036, double his prior estimate; AWS had revenue of $128.7B in 2025 —  Amazon O> CEO Andy Jassy said during an internal all-hands meeting he expects artificial intelligence could help cloud computing …

  8. Candex, which helps large companies pay small, one-time, or irregular vendors, raised a $40M Series C, bringing its total funding to more than $120M (Mary Ann Azevedo/Crunchbase News)

    Mary Ann Azevedo / Crunchbase News : Candex, which helps large companies pay small, one-time, or irregular vendors, raised a $40M Series C, bringing its total funding to more than $120M —  For companies operating around the world, hiring vendors for one-off purchases in other countries can be a complicated process, eating up time and resources.

  9. NetBlocks and Kentik: Iran further restricted what little internet connectivity remained to curb VPNs, possibly to forestall protests during Festival of Fire (Bloomberg)

    Bloomberg : NetBlocks and Kentik: Iran further restricted what little internet connectivity remained to curb VPNs, possibly to forestall protests during Festival of Fire —  Iran's internet, which has been heavily throttled by the regime since the outbreak of war, has gone even darker over the past 48 hours …

  10. Arizona sues Kalshi for allegedly operating an illegal gambling business and unlawfully facilitating betting on elections; Kalshi faces a similar suit in Nevada (Anna Washenko/Engadget)

    Anna Washenko / Engadget : Arizona sues Kalshi for allegedly operating an illegal gambling business and unlawfully facilitating betting on elections; Kalshi faces a similar suit in Nevada —  It's also facing a similar lawsuit in Nevada.  —  Kalshi has been sued by Arizona's attorney general for operating …

  11. OpenAI launches GPT-5.4 mini and nano, aimed at agents, coding, and multi-modal workflows, and offering near GPT-5.4-level performance at a much lower cost (David Gewirtz/ZDNET)

    David Gewirtz / ZDNET : OpenAI launches GPT-5.4 mini and nano, aimed at agents, coding, and multi-modal workflows, and offering near GPT-5.4-level performance at a much lower cost —  ZDNET's key takeaways  — GPT-5.4 mini runs more than twice as fast as GPT-5 mini.  — New models aim at agents, coding, and multi-modal workflows.

  12. Google expands Personal Intelligence, which lets Gemini tailor its responses by connecting to Gmail and other Google services, from paid users to all US users (Aisha Malik/TechCrunch)

    Aisha Malik / TechCrunch : Google expands Personal Intelligence, which lets Gemini tailor its responses by connecting to Gmail and other Google services, from paid users to all US users —  Google announced on Tuesday that it's expanding Personal Intelligence, its feature that allows its AI assistant to tailor …

  13. Intel unveils the Core Ultra 200HX Plus CPUs for high-end gaming laptops, offering an Intel Binary Optimization Tool to improve native performance in some games (Antonio G. Di Benedetto/The Verge)

    Antonio G. Di Benedetto / The Verge : Intel unveils the Core Ultra 200HX Plus CPUs for high-end gaming laptops, offering an Intel Binary Optimization Tool to improve native performance in some games —  The new flagship Core Ultra 9 290HX Plus is accompanied by a Core Ultra 7 270HX Plus.

  14. Tel Aviv- and Seattle-based Native, which helps companies monitor security across cloud providers, emerges from stealth with an $11M seed and a $31M Series A (Chris Metinko/Axios)

    Chris Metinko / Axios : Tel Aviv- and Seattle-based Native, which helps companies monitor security across cloud providers, emerges from stealth with an $11M seed and a $31M Series A —  What to read next

  15. Memo: Satya Nadella says ex-Snap executive Jacob Andreou will lead Copilot for commercial and consumer clients; Mustafa Suleyman will focus on new AI models (Jordan Novet/CNBC)

    Jordan Novet / CNBC : Memo: Satya Nadella says ex-Snap executive Jacob Andreou will lead Copilot for commercial and consumer clients; Mustafa Suleyman will focus on new AI models —  Microsoft said Tuesday that it's bringing together the engineering groups for its commercial and consumer Copilot assistants, which have yet to gain broad adoption.

Solidot(15)

  1. 英伟达的 DLSS 5 引发争议和批评

    英伟达演示了计划于今年晚些时候推出的深度学习超级采样技术 DLSS 的新版本,结果在玩家中间引发了广泛争议和批评,因为 DLSS 5 在重构图像过程中戏剧性的改变了游戏画面,为游戏画面加入了一层 AI 滤镜,让游戏中的人物变得面目全非。DLSS 5 在社交媒体上引发了玩家制作大量梗图进行嘲讽,在 reddit 上大量相关讨论被删除(可能是英伟达在公关),就像微软被称为 Microslop,玩家现在开始称英伟达为 Slopvidia。

  2. 太阳可能在几十亿年前从银河中心迁移到外围

    最新证据显示,约在 40-60 亿年前,太阳可能曾参与一次大规模的恒星迁移事件。一群与太阳性质非常相似的太阳类恒星(solar twins)一同离开银河系核心区域并向外迁移。天文学家利用 ESA Gaia 卫星的观测数据进行分析,建立了一份前所未有精确的恒星目录。研究显示,太阳目前位于银河系的位置并非偶然,而可能是这次大规模恒星迁移事件的一部分。太阳约在 46 亿年前诞生,而当时太阳的位置比今天更接近银河中心超过一万光年。恒星化学成分的研究支持这一推论,但这个结果长期让科学家感到困惑。观测显示银河中心存在一个巨大的棒状结构,它会形成所谓的共转屏障,使恒星​​难以从银河中心区域迁移到如此遥远的位置。为了解答这个问题,来自日本的研究团队对银河系中类似太阳的恒星展开了大规模研究。这些恒星在温度、表面重力与化学组成上都与太阳极为相似。研究团队从 Gaia 的资料中挑选出 6594 颗太阳类恒星建立目录。透过这份庞大的资料,研究人员得以重建目前最精确的恒星年龄分布。分析结果显示,在 40-60 亿年前出现一个明显且宽广的年龄峰值,显示一群年龄相近的恒星分布在距离银河中心相似的位置。这代表太阳并非单独迁移,而是属于一次大规模恒星外移事件的一部分。

  3. 微软允许 Windows 11 用户在安装过程中重命名主文件夹名称

    众所周知,Windows 在安装过程中并不允许用户重命名主文件夹名称,而是根据用户账号或邮箱地址自动生成名称。去年微软开始测试允许用户重命名主文件夹名称,但非常繁琐。现在微软终于将重命名主文件夹名称作为安装流程的一部分提供给用户。微软释出了预览版本 Windows 11 Insider Preview Build 26220.8062,在安装流程的“设备名称”页面包含了一个重命名主文件夹名称选项,如果用户跳过这一步骤,那么主文件夹仍然会使用默认名称。命名文件夹名称需要遵循微软的命名规定。

  4. 德国法庭裁决 TCL 的 QLED 不是真的 QLED

    德国的一家法庭裁决 TCL 误导消费者,它的多款宣称“量子点电视(QLED)”的产品并不是真的 QLED,没有提供 QLED 电视应有的色彩还原。法院命令 TCL 停止在德国宣传或销售相关型号的电视机。相关诉讼由韩国公司韩松化学提起,它是 TCL 竞争对手三星的合作伙伴。韩松化学委托进行的测试显示,三款以量子点名义出售的 TV 都未检测出铟和镉,它们是不可或缺的量子点材料。TCL 对测试结果提出异议,称量子点含量因供应商而异,它公布了自己的测试结果。TCL 的测试结果与韩松化学的测试结果相矛盾,但双方采用了不同的测试方法:TCL 的测试侧重于其使用的量子点薄膜,而韩松化学测试的是 TCL 电视机。韩松化学在包括美国在内的多国提起了针对 TCL 的诉讼,另一家中国电视制造商海信也面临类似的诉讼。

  5. Marknote 1.5 释出

    基于 Markdown 的笔记管理应用 Marknote 释出了 v1.5。主要新特性包括:新的 Source Mode 模式,不使用富文本 WYSIWYG 界面直接编辑 Markdown 内容;支持维基风格的笔记文档链接,支持跨笔记查找;简化笔记和笔记本管理,每个笔记本会显示包含的笔记数量,如需要在笔记本之间移动笔记可通过拖放完成;Duplicate Note 操作可创建模板复制现有笔记;KRunner 插件;等等。

  6. kagi 翻译支持翻译到 LinkedIn Speak

    职业社交网络 LinkedIn 的用户已经形成了一套独特的语言风格,这种风格被称为 LinkedIn Speak,其特点是能将任何琐事自我包装成积极向上的宏大叙事。举例来说:你失业了,但用 LinkedIn Speak 写出来变成了“开启了人生的新篇章”,去五百强企业做清洁工变成了荣幸加盟;等等。在中国,阿里巴巴的职场语言套话如“赋能、闭环、沉淀、生态”可能与 LinkedIn Speak 最为相似。现在,kagi 翻译工具加入了对 LinkedIn Speak 输出的支持,让任何人可以通过自然语言输出职场套话。

  7. 2026 年 Debian 项目领导人竞选开始,只有一名候选人

    一年一度的 Debian 项目领导人(DPL)竞选启动,今年只有一位候选人 Sruthi Chandran——她是一位来自印度的图书管理员,2025 年的 DPL Andreas Tille 没有再次参选。竞选期持续到 4 月 3 日,投票期从 4 月 4 日持续到 4 月 17 日。Debian 项目的选民们将要在同意 Sruthi Chandran 担任 DPL 或不同意(以上皆非)两个选项中进行投票。DPL 选举采用的是孔多塞投票法。

  8. GIMP 3.2 释出

    图形编辑器项目 GIMP 释出了 v3.2。此举是该项目自 GIMP 3.0 发布之后加快版本发布计划的一部分。从 GIMP 2.0 到 3.0,项目经历了逾 20 年的时间,开发者不希望让用户等待六七年才等到一个小版本更新,等二十年才有一个大版本。GIMP 3.2 新特性包括:MyPaint Brush 画笔工具新增 20 种新画笔;overwrite 绘画模式;新的和升级的文件格式;UI 改进;新的非破坏性图层,使用 Link Layers 整合外部图像,缩放、旋转和变换图像而不会损失质量或清晰度,源文件修改后 Link Layers 会同步更新,Path 工具能创建 Vector Layers;等等。

  9. GTA Wiki 从 Fandom 迁移到独立维基网站

    在 Minecraft Wiki 之后,另一个大型维基社区 GTA Wiki 正从 Fandom 迁移到独立维基网站 gta.wiki。Fandom 是吉米·威尔士等人联合创办的商业化维基托管平台,因广告和使用体验下降而招致了用户不满。GTA Wiki 称,今年 2 月 Fandom 进行了重组,任命了一名亲 AI 的 CEO,而用户对 Fandom 最大的抱怨包括广告太多和内容政策过于严厉。GTA Wiki 在 Fandom 的内容将全部复制到 gta.wiki,编辑和管理员可以选择留在旧平台或迁移到新平台。

  10. GDC 2026 访客减少三成

    2026 年的游戏开发者大会(GDC)访客人数比去年下降了三成,总人数大约 2 万。2022 年的 GDC 大会是疫情之后首度回归,采用了线上线下的混合模式举行,其中线下访客接近 1.2 万,总访客 1.7 万。2023 年访客人数回升到 2.8 万,2024 年突破 3 万创下访客人数纪录,2025 年维持这一水平。但今年的 GDC 大会参观人数因为费用以及国际游客对美国国内情况的担忧而减少。

  11. FSF 希望用户自由是 AI 公司版权诉讼的一个目标

    Anthropic 从 Library Genesis 等影子图书馆下载了逾 700 万本书籍,它与图书作者和解了侵权诉讼,正联系相关图书的作者提供经济补偿。被收录在 Anthropic 图书数据库中的一本书是 Sam Williams 著的《Free as in freedom: Richard Stallman's crusade for free software》,该书由 O'Reilly 和 FSF 根据 GNU Free Documentation License (GNU FDL)许可证出版,GNU FDL 是一种自由许可证,无需付费即可用于任意目的。FSF 表示,它对经济补偿兴趣不大,如果其拥有版权的图书被 AI 公司用于训练大模型,那么它更希望获得的补偿是用户自由:AI 公司与用户共享完整的训练输入,完整的模型、训练配置设置和相应的软件源代码。

  12. 英国涂鸦艺术家 Banksy 是否还应该保持匿名?

    路透发表了一篇调查报告,分析了匿名英国涂鸦艺术家 Banksy 的身份。2022 年 11 月,Banksy 在乌克兰基辅的一处被炸村庄墙壁上制作了涂鸦,当地居民看到了涂鸦者。路透记者对此展开了调查,发现了 Banksy 本人多年前亲笔写下的一份认罪书,承认行为不检的轻罪指控,这份文件揭露了他的真实身份。但 Banksy 的律师督促记者不要公开 Banksy 的身份,称会侵犯艺术家的隐私,干扰他艺术创作,危及他的安全,且会损害公众利益。律师称,“匿名或使用笔名进行创作符合重要的社会利益。它保护了言论自由,使创作者能畅所欲言的向权力说出真话,不必担心遭到报复、审查或迫害——尤其是在涉及政治、宗教或社会正义等敏感问题时。”

  13. 常压超导温度创下新纪录

    根据发表在 PNAS 期刊上的一项研究,休斯顿大学物理系及得克萨斯超导中心的研究团队在环境压力下实现了151开尔文(约零下 122 摄氏度)的超导转变温度,刷新了常压下超导温度的世界纪录。超导体通常需要超高压或超低温,常压室温超导体一直是科学家追求的目标。研究团队采用了一种名为“压力淬火”的新工艺。该方法的原理是,先对预选的材料样本施加极高压力,此过程能改变材料的微观结构,从而显著提升其超导转变温度。在维持高压并降温至特定状态后,迅速将压力完全释放。通过这种快速“淬火”,材料在高压下获得的、更利于超导的亚稳态结构得以“锁定”并保留下来,材料在恢复常压后仍能在比原来高得多的温度下保持超导特性。凭借这一方法,团队将超导材料在常压下的转变温度提升至 151 开尔文。

  14. 逾四成日本人计划工作到 70 岁后

    根据日经的调查,回答“到了 70 岁仍会继续工作”的比例为 42%,自 2018 年调查开始以来首次超过 4 成。回答“70~74 岁”仍会继续工作的比例为 23%,回答“75岁以上”的比例为 19%。工作到多大年龄的平均值为 68.3 岁,高于法定的 65 岁。而日本政府在《老年人雇用稳定法》中规定,确保老年人工作到 70 岁的机会是企业的努力义务。

  15. 波兰核研究机构遭黑客攻击

    波兰核研究机构 National Centre for Nuclear Research(NCBJ)披露其 IT 基础设施遭到网络攻击,但表示安全团队迅速采取行动,挫败了攻击,因此未遭受什么影响。NCBJ 从事核物理、反应堆技术、粒子物理和辐射应用方面的研究,运营着用于科学实验、中子研究和医用同位素生产的核反应堆 MARIA。NCBJ 称 MARIA 反应堆未受影响,仍然全负荷运行。NCBJ 未确定攻击者身份。