OrangeBot.AI Digest — 2026-01-21
53 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Linux from Scratch (www.linuxfromscratch.org)
- Scientists find a way to regrow cartilage in mice and human tissue samples (www.sciencedaily.com)
- Claude's new constitution (www.anthropic.com)
- Waiting for dawn in search: Search index, Google rulings and impact on Kagi (blog.kagi.com)
- Swedish Alecta has sold off an estimated $8B of US Treasury Bonds (www.di.se)
- Skip is now free and open source (skip.dev)
- Tell HN: Bending Spoons laid off almost everybody at Vimeo yesterday
- Show HN: ChartGPU – WebGPU-powered charting library (1M points at 60fps) (github.com)
- Ireland wants to give its cops spyware, ability to crack encrypted messages (www.theregister.com)
- How AI destroys institutions (cyberlaw.stanford.edu)
- Stories removed from the Hacker News Front Page, updated in real time (2024) (github.com)
- SETI@home is in hiberation (setiathome.berkeley.edu)
- EU–INC – A new pan-European legal entity (www.eu-inc.org)
- The percentage of Show HN posts is increasing, but their scores are decreasing (snubi.net)
- EmuDevz: A game about developing emulators (afska.github.io)
GitHub Trending(8)
- tambo-ai / tambo
Generative UI SDK for React
- EveryInc / compound-engineering-plugin
Official Claude Code compound engineering plugin
- twitter / the-algorithm
Source code for the X Recommendation Algorithm
- xai-org / grok-1
Grok open release
- microsoft / agent-lightning
The absolute trainer to light up AI agents.
- VectifyAI / PageIndex
📑 PageIndex: Document Index for Vectorless, Reasoning-based RAG
- microsoft / Data-Science-For-Beginners
10 Weeks, 20 Lessons, Data Science for All!
- tobi / try
fresh directories for every vibe
Hugging Face(15)
- Being-H0.5: Scaling Human-Centric Robot Learning for Cross-Embodiment Generalization
We introduce Being-H0.5, a foundational Vision-Language-Action (VLA) model designed for robust cross-embodiment generalization across diverse robotic platforms. While existing VLAs often struggle with morphological heterogeneity and data scarcity, we propose a human-centric learning paradigm that treats human interaction traces as a universal "mother tongue" for physical interaction. To support this, we present UniHand-2.0, the largest embodied pre-training recipe to date, comprising over 35,000 hours of multimodal data across 30 distinct robotic embodiments. Our approach introduces a Unified Action Space that maps heterogeneous robot controls into semantically aligned slots, enabling low-resource robots to bootstrap skills from human data and high-resource platforms. Built upon this human-centric foundation, we design a unified sequential modeling and multi-task pre-training paradigm to bridge human demonstrations and robotic execution. Architecturally, Being-H0.5 utilizes a Mixture-of-Transformers design featuring a novel Mixture-of-Flow (MoF) framework to decouple shared motor primitives from specialized embodiment-specific experts. Finally, to make cross-embodiment policies stable in the real world, we introduce Manifold-Preserving Gating for robustness under sensory shift and Universal Async Chunking to universalize chunked control across embodiments with different latency and control profiles. We empirically demonstrate that Being-H0.5 achieves state-of-the-art results on simulated benchmarks, such as LIBERO (98.9%) and RoboCasa (53.9%), while also exhibiting strong cross-embodiment capabilities on five robotic platforms.
- Advances and Frontiers of LLM-based Issue Resolution in Software Engineering: A Comprehensive Survey
Issue resolution, a complex Software Engineering (SWE) task integral to real-world development, has emerged as a compelling challenge for artificial intelligence. The establishment of benchmarks like SWE-bench revealed this task as profoundly difficult for large language models, thereby significantly accelerating the evolution of autonomous coding agents. This paper presents a systematic survey of this emerging domain. We begin by examining data construction pipelines, covering automated collection and synthesis approaches. We then provide a comprehensive analysis of methodologies, spanning training-free frameworks with their modular components to training-based techniques, including supervised fine-tuning and reinforcement learning. Subsequently, we discuss critical analyses of data quality and agent behavior, alongside practical applications. Finally, we identify key challenges and outline promising directions for future research. An open-source repository is maintained at https://github.com/DeepSoftwareAnalytics/Awesome-Issue-Resolution to serve as a dynamic resource in this field.
- Toward Efficient Agents: Memory, Tool learning, and Planning
Recent years have witnessed increasing interest in extending large language models into agentic systems. While the effectiveness of agents has continued to improve, efficiency, which is crucial for real-world deployment, has often been overlooked. This paper therefore investigates efficiency from three core components of agents: memory, tool learning, and planning, considering costs such as latency, tokens, steps, etc. Aimed at conducting comprehensive research addressing the efficiency of the agentic system itself, we review a broad range of recent approaches that differ in implementation yet frequently converge on shared high-level principles including but not limited to bounding context via compression and management, designing reinforcement learning rewards to minimize tool invocation, and employing controlled search mechanisms to enhance efficiency, which we discuss in detail. Accordingly, we characterize efficiency in two complementary ways: comparing effectiveness under a fixed cost budget, and comparing cost at a comparable level of effectiveness. This trade-off can also be viewed through the Pareto frontier between effectiveness and cost. From this perspective, we also examine efficiency oriented benchmarks by summarizing evaluation protocols for these components and consolidating commonly reported efficiency metrics from both benchmark and methodological studies. Moreover, we discuss the key challenges and future directions, with the goal of providing promising insights.
- OmniTransfer: All-in-one Framework for Spatio-temporal Video Transfer
Videos convey richer information than images or text, capturing both spatial and temporal dynamics. However, most existing video customization methods rely on reference images or task-specific temporal priors, failing to fully exploit the rich spatio-temporal information inherent in videos, thereby limiting flexibility and generalization in video generation. To address these limitations, we propose OmniTransfer, a unified framework for spatio-temporal video transfer. It leverages multi-view information across frames to enhance appearance consistency and exploits temporal cues to enable fine-grained temporal control. To unify various video transfer tasks, OmniTransfer incorporates three key designs: Task-aware Positional Bias that adaptively leverages reference video information to improve temporal alignment or appearance consistency; Reference-decoupled Causal Learning separating reference and target branches to enable precise reference transfer while improving efficiency; and Task-adaptive Multimodal Alignment using multimodal semantic guidance to dynamically distinguish and tackle different tasks. Extensive experiments show that OmniTransfer outperforms existing methods in appearance (ID and style) and temporal transfer (camera movement and video effects), while matching pose-guided methods in motion transfer without using pose, establishing a new paradigm for flexible, high-fidelity video generation.
- Think3D: Thinking with Space for Spatial Reasoning
Understanding and reasoning about the physical world requires spatial intelligence: the ability to interpret geometry, perspective, and spatial relations beyond 2D perception. While recent vision large models (VLMs) excel at visual understanding, they remain fundamentally 2D perceivers and struggle with genuine 3D reasoning. We introduce Think3D, a framework that enables VLM agents to think with 3D space. By leveraging 3D reconstruction models that recover point clouds and camera poses from images or videos, Think3D allows the agent to actively manipulate space through camera-based operations and ego/global-view switching, transforming spatial reasoning into an interactive 3D chain-of-thought process. Without additional training, Think3D significantly improves the spatial reasoning performance of advanced models such as GPT-4.1 and Gemini 2.5 Pro, yielding average gains of +7.8% on BLINK Multi-view and MindCube, and +4.7% on VSI-Bench. We further show that smaller models, which struggle with spatial exploration, benefit significantly from a reinforcement learning policy that enables the model to select informative viewpoints and operations. With RL, the benefit from tool usage increases from +0.7% to +6.8%. Our findings demonstrate that training-free, tool-augmented spatial exploration is a viable path toward more flexible and human-like 3D reasoning in multimodal agents, establishing a new dimension of multimodal intelligence. Code and weights are released at https://github.com/zhangzaibin/spagent.
- FutureOmni: Evaluating Future Forecasting from Omni-Modal Context for Multimodal LLMs
Although Multimodal Large Language Models (MLLMs) demonstrate strong omni-modal perception, their ability to forecast future events from audio-visual cues remains largely unexplored, as existing benchmarks focus mainly on retrospective understanding. To bridge this gap, we introduce FutureOmni, the first benchmark designed to evaluate omni-modal future forecasting from audio-visual environments. The evaluated models are required to perform cross-modal causal and temporal reasoning, as well as effectively leverage internal knowledge to predict future events. FutureOmni is constructed via a scalable LLM-assisted, human-in-the-loop pipeline and contains 919 videos and 1,034 multiple-choice QA pairs across 8 primary domains. Evaluations on 13 omni-modal and 7 video-only models show that current systems struggle with audio-visual future prediction, particularly in speech-heavy scenarios, with the best accuracy of 64.8% achieved by Gemini 3 Flash. To mitigate this limitation, we curate a 7K-sample instruction-tuning dataset and propose an Omni-Modal Future Forecasting (OFF) training strategy. Evaluations on FutureOmni and popular audio-visual and video-only benchmarks demonstrate that OFF enhances future forecasting and generalization. We publicly release all code (https://github.com/OpenMOSS/FutureOmni) and datasets (https://huggingface.co/datasets/OpenMOSS-Team/FutureOmni).
- MemoryRewardBench: Benchmarking Reward Models for Long-Term Memory Management in Large Language Models
Existing works increasingly adopt memory-centric mechanisms to process long contexts in a segment manner, and effective memory management is one of the key capabilities that enables large language models to effectively propagate information across the entire sequence. Therefore, leveraging reward models (RMs) to automatically and reliably evaluate memory quality is critical. In this work, we introduce MemoryRewardBench, the first benchmark to systematically study the ability of RMs to evaluate long-term memory management processes. MemoryRewardBench covers both long-context comprehension and long-form generation tasks, featuring 10 distinct settings with different memory management patterns, with context length ranging from 8K to 128K tokens. Evaluations on 13 cutting-edge RMs indicate a diminishing performance gap between open-source and proprietary models, with newer-generation models consistently outperforming their predecessors regardless of parameter count. We further expose the capabilities and fundamental limitations of current RMs in evaluating LLM memory management across diverse settings.
- Locate, Steer, and Improve: A Practical Survey of Actionable Mechanistic Interpretability in Large Language Models
Mechanistic Interpretability (MI) has emerged as a vital approach to demystify the opaque decision-making of Large Language Models (LLMs). However, existing reviews primarily treat MI as an observational science, summarizing analytical insights while lacking a systematic framework for actionable intervention. To bridge this gap, we present a practical survey structured around the pipeline: "Locate, Steer, and Improve." We formally categorize Localizing (diagnosis) and Steering (intervention) methods based on specific Interpretable Objects to establish a rigorous intervention protocol. Furthermore, we demonstrate how this framework enables tangible improvements in Alignment, Capability, and Efficiency, effectively operationalizing MI as an actionable methodology for model optimization. The curated paper list of this work is available at https://github.com/rattlesnakey/Awesome-Actionable-MI-Survey.
- UniX: Unifying Autoregression and Diffusion for Chest X-Ray Understanding and Generation
Despite recent progress, medical foundation models still struggle to unify visual understanding and generation, as these tasks have inherently conflicting goals: semantic abstraction versus pixel-level reconstruction. Existing approaches, typically based on parameter-shared autoregressive architectures, frequently lead to compromised performance in one or both tasks. To address this, we present UniX, a next-generation unified medical foundation model for chest X-ray understanding and generation. UniX decouples the two tasks into an autoregressive branch for understanding and a diffusion branch for high-fidelity generation. Crucially, a cross-modal self-attention mechanism is introduced to dynamically guide the generation process with understanding features. Coupled with a rigorous data cleaning pipeline and a multi-stage training strategy, this architecture enables synergistic collaboration between tasks while leveraging the strengths of diffusion models for superior generation. On two representative benchmarks, UniX achieves a 46.1% improvement in understanding performance (Micro-F1) and a 24.2% gain in generation quality (FD-RadDino), using only a quarter of the parameters of LLM-CXR. By achieving performance on par with task-specific models, our work establishes a scalable paradigm for synergistic medical image understanding and generation. Codes and models are available at https://github.com/ZrH42/UniX.
- ToolPRMBench: Evaluating and Advancing Process Reward Models for Tool-using Agents
Reward-guided search methods have demonstrated strong potential in enhancing tool-using agents by effectively guiding sampling and exploration over complex action spaces. As a core design, those search methods utilize process reward models (PRMs) to provide step-level rewards, enabling more fine-grained monitoring. However, there is a lack of systematic and reliable evaluation benchmarks for PRMs in tool-using settings. In this paper, we introduce ToolPRMBench, a large-scale benchmark specifically designed to evaluate PRMs for tool-using agents. ToolPRMBench is built on top of several representative tool-using benchmarks and converts agent trajectories into step-level test cases. Each case contains the interaction history, a correct action, a plausible but incorrect alternative, and relevant tool metadata. We respectively utilize offline sampling to isolate local single-step errors and online sampling to capture realistic multi-step failures from full agent rollouts. A multi-LLM verification pipeline is proposed to reduce label noise and ensure data quality. We conduct extensive experiments across large language models, general PRMs, and tool-specialized PRMs on ToolPRMBench. The results reveal clear differences in PRM effectiveness and highlight the potential of specialized PRMs for tool-using. Code and data will be released at https://github.com/David-Li0406/ToolPRMBench.
- Aligning Agentic World Models via Knowledgeable Experience Learning
Current Large Language Models (LLMs) exhibit a critical modal disconnect: they possess vast semantic knowledge but lack the procedural grounding to respect the immutable laws of the physical world. Consequently, while these agents implicitly function as world models, their simulations often suffer from physical hallucinations-generating plans that are logically sound but physically unexecutable. Existing alignment strategies predominantly rely on resource-intensive training or fine-tuning, which attempt to compress dynamic environmental rules into static model parameters. However, such parametric encapsulation is inherently rigid, struggling to adapt to the open-ended variability of physical dynamics without continuous, costly retraining. To bridge this gap, we introduce WorldMind, a framework that autonomously constructs a symbolic World Knowledge Repository by synthesizing environmental feedback. Specifically, it unifies Process Experience to enforce physical feasibility via prediction errors and Goal Experience to guide task optimality through successful trajectories. Experiments on EB-ALFRED and EB-Habitat demonstrate that WorldMind achieves superior performance compared to baselines with remarkable cross-model and cross-environment transferability.
- Agentic-R: Learning to Retrieve for Agentic Search
Agentic search has recently emerged as a powerful paradigm, where an agent interleaves multi-step reasoning with on-demand retrieval to solve complex questions. Despite its success, how to design a retriever for agentic search remains largely underexplored. Existing search agents typically rely on similarity-based retrievers, while similar passages are not always useful for final answer generation. In this paper, we propose a novel retriever training framework tailored for agentic search. Unlike retrievers designed for single-turn retrieval-augmented generation (RAG) that only rely on local passage utility, we propose to use both local query-passage relevance and global answer correctness to measure passage utility in a multi-turn agentic search. We further introduce an iterative training strategy, where the search agent and the retriever are optimized bidirectionally and iteratively. Different from RAG retrievers that are only trained once with fixed questions, our retriever is continuously improved using evolving and higher-quality queries from the agent. Extensive experiments on seven single-hop and multi-hop QA benchmarks demonstrate that our retriever, termed , consistently outperforms strong baselines across different search agents. Our codes are available at: https://github.com/8421BCD/Agentic-R.
- A BERTology View of LLM Orchestrations: Token- and Layer-Selective Probes for Efficient Single-Pass Classification
Production LLM systems often rely on separate models for safety and other classification-heavy steps, increasing latency, VRAM footprint, and operational complexity. We instead reuse computation already paid for by the serving LLM: we train lightweight probes on its hidden states and predict labels in the same forward pass used for generation. We frame classification as representation selection over the full token-layer hidden-state tensor, rather than committing to a fixed token or fixed layer (e.g., first-token logits or final-layer pooling). To implement this, we introduce a two-stage aggregator that (i) summarizes tokens within each layer and (ii) aggregates across layer summaries to form a single representation for classification. We instantiate this template with direct pooling, a 100K-parameter scoring-attention gate, and a downcast multi-head self-attention (MHA) probe with up to 35M trainable parameters. Across safety and sentiment benchmarks our probes improve over logit-only reuse (e.g., MULI) and are competitive with substantially larger task-specific baselines, while preserving near-serving latency and avoiding the VRAM and latency costs of a separate guard-model pipeline.
- KAGE-Bench: Fast Known-Axis Visual Generalization Evaluation for Reinforcement Learning
Pixel-based reinforcement learning agents often fail under purely visual distribution shift even when latent dynamics and rewards are unchanged, but existing benchmarks entangle multiple sources of shift and hinder systematic analysis. We introduce KAGE-Env, a JAX-native 2D platformer that factorizes the observation process into independently controllable visual axes while keeping the underlying control problem fixed. By construction, varying a visual axis affects performance only through the induced state-conditional action distribution of a pixel policy, providing a clean abstraction for visual generalization. Building on this environment, we define KAGE-Bench, a benchmark of six known-axis suites comprising 34 train-evaluation configuration pairs that isolate individual visual shifts. Using a standard PPO-CNN baseline, we observe strong axis-dependent failures, with background and photometric shifts often collapsing success, while agent-appearance shifts are comparatively benign. Several shifts preserve forward motion while breaking task completion, showing that return alone can obscure generalization failures. Finally, the fully vectorized JAX implementation enables up to 33M environment steps per second on a single GPU, enabling fast and reproducible sweeps over visual factors. Code: https://avanturist322.github.io/KAGEBench/.
- LightOnOCR: A 1B End-to-End Multilingual Vision-Language Model for State-of-the-Art OCR
We present LightOnOCR-2-1B, a 1B-parameter end-to-end multilingual vision--language model that converts document images (e.g., PDFs) into clean, naturally ordered text without brittle OCR pipelines. Trained on a large-scale, high-quality distillation mix with strong coverage of scans, French documents, and scientific PDFs, LightOnOCR-2 achieves state-of-the-art results on OlmOCR-Bench while being 9times smaller and substantially faster than prior best-performing models. We further extend the output format to predict normalized bounding boxes for embedded images, introducing localization during pretraining via a resume strategy and refining it with RLVR using IoU-based rewards. Finally, we improve robustness with checkpoint averaging and task-arithmetic merging. We release model checkpoints under Apache 2.0, and publicly release the dataset and LightOnOCR-bbox-bench evaluation under their respective licenses.
Solidot(15)
- 月球上的射电望远镜
如果一切顺利,2027 年初 SpaceX 将把 LuSEE-Night 发射到月球背面。LuSEE-Night 代表 Lunar Surface Electromagnetics Experiment–Night,它将利用 Firefly Aerospace 的着陆器 Blue Ghost Mission 2 在月之背面登陆。Firefly 的 Blue Ghost 1 去年 3 月完成了私人公司的首次成功月表着陆。月球射电望远镜有助于科学家解开宇宙中最著名的谜团。月球上将能更清晰的观测暗物质、暗能量、中子星和引力波。地球上的观测设备容易受到干扰,而月球可能是内太阳系最安静的地方。月球背面在长达 14 个地球日里可能是内太阳系最黑暗的电磁区域,没有太阳辐射,也没有来自地球的干扰信号。
- cURL 因 AI Slop 将关闭 Bug 悬赏项目
为减少 AI Slop 报告数量,cURL 项目将在一月底终止 Bug 悬赏项目。cURL 维护者 Daniel Stenberg 表示,“为避免被拖下去我们必须努力阻止这股洪流。”cURL 是一个广泛使用的互联网基础工具,几乎被每一个联网的设备和系统使用。今年早些时候,cURL 项目披露收到了大量由 AI 生成的虚假漏洞报告,Stenberg 当时称至今没有看到一份 AI 帮助下完成的漏洞报告是有效的。
- 大多数 CEO 报告 AI 投资零回报
普华永道(PwC)调查了逾 4500 位 CEO ,发现尽管在 AI 上投入了大量资金,但大部分 CEO 表示 AI 投资未带来收入增长或成本降低。接受调查的 4454 位商界领袖中只有 12% 同时实现了成本降低和收入增长,56% 既没有降低成本也没有增加收入,26% 实现了成本降低,但类似比例的人增加了成本。AI 的普及度仍然有限,即使在需求生成(22%)、支持服务(20%)和产品开发(19%)等热门应用场景中,只有少数企业广泛部署 AI。从更宏观的角度,普华永道报告 CEO 们的信心跌至五年以来的最低点,仅 30% 的 CEO 对营收增长乐观(低于去年的 38%),表明地缘政治风险日益加剧,网络威胁升级,同时 AI 的利弊也存在不确定性。
- IPv4 和 IPv6 地址现状
IPv4 的 40 亿地址空间早就枯竭,但有 2^128 地址空间的 IPv6 普及速度并没有预想的快。一大原因是互联网从点对点架构转向了客户端/服务器架构,而 Network Address Translators (NATs)与此完全嵌合。客户端/服务器架构中,客户端共享一个较小的公共地址池,只在与远程服务器建立活动会话后才使用地址。在 NAT 帮助下逾 300 亿联网设备只使用 30 亿个已通告的 IPv4 地址池。但联网设备的持续增长意味着 NAT 无法完全解决问题,增长压力推动 IPv6 的加快部署。2017 年 IPv6 部署激增是受益于印度部署的移动 IPv6 服务,而中国的 IPv6 服务最近几年也在快速普及,IPv6 用户占比从 2024 年初的 32% 增长到 2025 年底的 54%,意味着两年内中国 IPv6 用户数量增加约 9400 万。但非洲、东欧和南欧以及西亚没有出现 IPv6 的大规模部署。
- 华硕为 AI 逐步退出智能手机业务
华硕董事长施崇棠证实该公司将逐步退出智能手机业务,将重心集中到 AI 产品如机器人和智能眼镜上。施崇棠表示,未来不会推出新智能手机机型。但他没有把话完全说死,只是表示公司将采取“无限期观望”态度。华硕的智能手机包括 Zenfone 和 ROG Phone 品牌,前者主打小巧廉价,后者则是主打游戏以高价著称,其价格甚至高于三星的旗舰手机。目前还没有哪家 Android 厂商在停止发布新机型后能重新恢复生产,LG 就是其中的典型代表,该公司在停止推出新机型后最终完全退出手机市场。
- 沥青滴漏实验即将百年
世界持续时间最长的实验即将跨过一百年。沥青滴漏实验始于 1927 年,澳大利亚昆士兰大学物理学家 Thomas Parnell 将已知最粘稠的液体沥青注入到一个封闭的漏斗。1930 年 Parnell 剪开了漏斗封口,标志着沥青滴漏实验正式开始。沥青开始缓慢流动,每一滴高黏度沥青需经近十年时间,方能滴进漏斗下方的烧杯之中,第一滴沥青于 1938 年 12 月滴出。时至今日,已滴出九滴沥青。根据实验结果,沥青的黏度大约是水之千亿倍。上一次沥青滴落是在 2014 年,下一次预计会在 2020 年代的某个时候。尽管有无数双眼睛看着,但至今无人真正目睹沥青滴下。实验一直在直播,但过去的各种故障导致错过了每一个关键时刻。
- TCL 接手索尼电视业务
索尼宣布了剥离电视业务的计划,它的电视业务将合并到一家新的合资企业,中国最大的电视制造商 TCL 在其中持 51% 股份,索尼持 49%。两家公司计划 3 月底达成协议,2027 年 4 月新公司投入运营。这项计划还需要得到监管部门的批准。新公司将保留“索尼”和“Bravia”电视品牌,计划结合索尼的图像和音频技术、品牌价值、供应链管理,以及 TCL 的显示技术、垂直整合的供应链优势、全球市场布局以及端到端的成本效益。这项交易如果达成,将标志着索尼电视时代的结束。
- 地球逾 20 年再次遭遇 S4 级太阳辐射风暴
1 月 18 日暴发的 X1.9 级太阳耀斑朝地球方向释放了一个日冕物质抛射,1 月 19 日日冕物质抛射抵达地球,产生了 G4 级地磁风暴,欧洲各地的夜空,从德国、英国到法国都观察到了明亮的极光。1 月 20 日的地磁风暴等级略低,在 G1 到 G3 之间。地球目前正处于 S4 级太阳辐射风暴中,这是 2003 年以来首次突破 S4 级,也是自测量以来第三强的辐射风暴,预计会对卫星运营商和高频无线电产生影响。
- OzLabs 成员全部脱离 IBM
OzLabs 是一个澳大利亚自由软件开发者组织,该组织成员负责了很多知名的开源项目如 Samba、rsync、Linux PPP、Linux netfilter、Linux Advanced Power Management (APM) 和 OpenBMC。该组织成立于 1999 年,由 Linuxcare 聘请 Andrew Tridgell 负责组建。Linuxcare 在 2001 年的动荡导致大部分成员都离开了,他们加入了 IBM Linux Technology Center 从事 PowerPC Linux 及相关项目。但截至 2026 年 1 月,所有成员都离开了 IBM,结束了 OzLabs 与 IBM 长达 25 年的合作关系。
- 好莱坞单一文化的兴衰
在流媒体和社交媒体的推荐算法时代,人们的注意力很少会被少数几部作品所捕获,曾经引领时代吸引无数人关注的好莱坞作品越来越少见了。1939 年的《乱世佳人》售出了 2 亿张电影票,而当时的美国人口仅为 1.3 亿。《陆军野战医院(M*A*S*H)》在 1983 年播出最后一集时吸引了逾一亿人观看。2025 年只有三部美国电影的票房收入超过 10 亿美元,而 2019 年这一数字为九部。YouTube 之所以能成为电视上最受欢迎的视频平台,不是因为它拥有最热门的节目,而是因为它能满足所有人需求。互联网打破了好莱坞对发行渠道的垄断。
- Threads 移动端日活人数超过 X
Similarweb 最新数据显示,Threads 移动端日活用户数超越了 X。而 X 在 Web 端仍然远超 Threads,日访问量约 1.5 亿,而 Threads 的日访问量为 850 万。Similarweb 的数据显示,截至 2026 年 1 月 7 日 Threads 在 iOS 和 Android 平台上的日活用户数达到 1.415 亿,而 X 在移动设备上的日活用户数为 1.25 亿。Threads 日活增长受益于在 Meta 旗下社交平台如 Facebook 和 Instagram 上的推广,以及对内容创作者的重视和新功能的快速推出。Threads 过去一年新增了兴趣社区、完善筛选功能、私信、长文本、阅后即焚等功能,最近还在测试游戏功能。
- 英伟达被控主动联络安娜的档案以高速下载盗版书库
英伟达除了供应 AI 芯片外,还开发了自己的大模型,如 NeMo、Retro-48B、InstructRetro 和 Megatron。那么这些大模型的训练数据来自何处?图书作者指控英伟达使用盗版书库训练模型。上周五原告修改了诉状,指控英伟达使用了影子图书馆“安娜的档案(Anna’s Archive)”收集的盗版电子书库。诉状援引英伟达内部邮件和文件称,英伟达员工主动联系“安娜的档案”,询问该影子图书馆提供的付费“高速访问”是什么意思。安娜的档案要求英伟达管理层内部批准之后它才会提供该服务。英伟达据报道在一周内批准了这一要求,安娜的档案随后提供了 500 TB 电子书的高速访问。英伟达还被控从 LibGen、Sci-Hub 和 Z-Library 下载书籍。
- Oxfam 报告称全球财富不平等创新高
乐施会(Oxfam)发表年度报告《Resisting the Rule of the Rich》,称全球财富不平等在加速。2025 年亿万富翁的财富激增 2.5 万亿美元,几乎相当于全球半数人口(约 41 亿人)的财富总和。全球亿万富翁人数首次突破 3000 人,世界首富马斯克(Elon Musk)的财富也首次突破 5000 亿美元。报告警告,超级富豪正形成新的寡头政治,他们利用巨额财富收买政治、媒体和司法以维护自身财富,瓦解和摧毁进步政策,剥夺我们的基本公民权利和政治权利。报告举例说,贝佐斯收购《华盛顿邮报》,马斯克收购 Twitter/X,黄馨祥(Patrick Soon-Shion)控制《洛杉矶时报》,法国极右翼亿万富翁 Vincent Bollore 拥有 CNews。乐施会呼吁全世界人民联合起来捍卫自身权利,争取一个取代不平等和寡头政治的替代方案。
- 10 分钟高强度运动有助于降低癌症风险
澳大利亚纽卡斯尔大学的研究人员发现,仅 10 分钟的高强度运动就能提高血液中几种小分子的水平。这些分子能够启动 DNA 修复机制,并关闭癌症生长信号。其中许多分子具有减轻炎症、维护血管健康和促进新陈代谢的作用。这些快速的变化似乎能够抑制肠道癌细胞的生长,同时还能加快受损DNA的修复速度。当科学家在实验室中将含有这些运动驱动分子的血液暴露于肠癌细胞时,他们观察到了广泛的基因变化——超过 1300 个基因的活动发生了改变,包括那些参与DNA修复、能量生产和癌细胞生长的基因。研究表明,运动通过血液发送分子信号,影响控制肿瘤生长和遗传稳定性的基因。这些结果进一步证明,保持身体活动是癌症预防的重要组成部分。研究团队发现,运动增加了支持线粒体能量代谢的基因活性,这有助于细胞更有效地利用氧气。与此同时,与快速细胞分裂相关的基因活性出现下调,可能使癌细胞侵袭性降低。运动后采集的血液还增强了DNA修复能力,激活了一个名为 PNKP 的关键修复基因。
- 近三分之一的社媒研究有着未披露的利益关联
社媒研究员通常需要与平台进行合作,但双方的利益关系很多时候并没有公开。近三分之一的社媒研究有着未披露的利益关联,部分研究员接受了社媒的资助,部分曾与社媒行业的员工合作发表过研究。研究人员认为,此类关联可能会扭曲研究结果。研究人员分析了《科学》、《自然》、PNAS 及其子刊如《Science Advances》和《Nature Communications》上的 295 篇社媒论文,它们的总引用次数有 5 万次,被逾 1.5 万篇新闻报道引用。其中五分之一的署名作者承认获得社媒企业的资助或有过合作。研究人员之后使用 OpenAlex 分析了署名作者和社媒企业之间的关联,发现一半作者存在关联,也就是有 30% 的作者没有披露潜在利益冲突。