OrangeBot.AI Digest — 2026-03-27
87 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- If you don't opt out by Apr 24 GitHub will train on your private repos
- Make macOS consistently bad (unironically) (lr0.org)
- AI got the blame for the Iran school bombing. The truth is more worrying (www.theguardian.com)
- Meow.camera (meow.camera)
- Desk for people who work at home with a cat (soranews24.com)
- People inside Microsoft are fighting to drop mandatory Microsoft Account (www.windowscentral.com)
- Anatomy of the .claude/ folder (blog.dailydoseofds.com)
- Iran-linked hackers claim breach of FBI director's personal email (www.reuters.com)
- Installing a Let's Encrypt TLS certificate on a Brother printer with Certbot (owltec.ca)
- How and why to take a logarithm of an image [video] (www.youtube.com)
- The 'paperwork flood': How I drowned a bureaucrat before dinner (sightlessscribbles.com)
- Hold on to Your Hardware (xn--gckvb8fzb.com)
- ‘Energy independence feels practical’: Europeans building mini solar farms (www.euronews.com)
- Everything old is new again: memory optimization (nibblestew.blogspot.com)
- A Faster Alternative to Jq (micahkepe.com)
GitHub Trending(12)
Product Hunt(15)
- Oli
Scan any product to know it's safe during pregnancy
- Hunna
Profit App for Business Founders
- Stripe Projects
Production-ready dev stack from your terminal
- Focus Flow 6.7 (Jira Plugin)
Plan vs reality and multi-team epic reviews in Jira
- Audos Publishing House
Build an AI business, get up to $100K. No equity taken
- Benchspan
Run agent benchmarks in minutes, not hours
- Noctiluca
A new remote desktop for macOS
- Cockpit AI
Run revenue agents across every channel
- Universal CLI by Composio
Connect AI agents to 1000+ apps directly from your terminal
- Voxtral TTS by Mistral AI
Multilingual TTS model with realistic and expressive speech
- Google Gemini Memory Import
Switch to Gemini without losing your AI memories
- Codex Plugins
Package Codex skills and app integrations as plugins
- AppDesk
Your App Store data, beautifully visualized on your Mac
- Suno v5.5
Create with your voice, tune models to your sound
- Gemini 3.1 Flash Live
Making audio AI more natural and reliable
Hugging Face(15)
- PixelSmile: Toward Fine-Grained Facial Expression Editing
Fine-grained facial expression editing has long been limited by intrinsic semantic overlap. To address this, we construct the Flex Facial Expression (FFE) dataset with continuous affective annotations and establish FFE-Bench to evaluate structural confusion, editing accuracy, linear controllability, and the trade-off between expression editing and identity preservation. We propose PixelSmile, a diffusion framework that disentangles expression semantics via fully symmetric joint training. PixelSmile combines intensity supervision with contrastive learning to produce stronger and more distinguishable expressions, achieving precise and stable linear expression control through textual latent interpolation. Extensive experiments demonstrate that PixelSmile achieves superior disentanglement and robust identity preservation, confirming its effectiveness for continuous, controllable, and fine-grained expression editing, while naturally supporting smooth expression blending.
- Intern-S1-Pro: Scientific Multimodal Foundation Model at Trillion Scale
We introduce Intern-S1-Pro, the first one-trillion-parameter scientific multimodal foundation model. Scaling to this unprecedented size, the model delivers a comprehensive enhancement across both general and scientific domains. Beyond stronger reasoning and image-text understanding capabilities, its intelligence is augmented with advanced agent capabilities. Simultaneously, its scientific expertise has been vastly expanded to master over 100 specialized tasks across critical science fields, including chemistry, materials, life sciences, and earth sciences. Achieving this massive scale is made possible by the robust infrastructure support of XTuner and LMDeploy, which facilitates highly efficient Reinforcement Learning (RL) training at the 1-trillion parameter level while ensuring strict precision consistency between training and inference. By seamlessly integrating these advancements, Intern-S1-Pro further fortifies the fusion of general and specialized intelligence, working as a Specializable Generalist, demonstrating its position in the top tier of open-source models for general capabilities, while outperforming proprietary models in the depth of specialized scientific tasks.
- RealRestorer: Towards Generalizable Real-World Image Restoration with Large-Scale Image Editing Models
Image restoration under real-world degradations is critical for downstream tasks such as autonomous driving and object detection. However, existing restoration models are often limited by the scale and distribution of their training data, resulting in poor generalization to real-world scenarios. Recently, large-scale image editing models have shown strong generalization ability in restoration tasks, especially for closed-source models like Nano Banana Pro, which can restore images while preserving consistency. Nevertheless, achieving such performance with those large universal models requires substantial data and computational costs. To address this issue, we construct a large-scale dataset covering nine common real-world degradation types and train a state-of-the-art open-source model to narrow the gap with closed-source alternatives. Furthermore, we introduce RealIR-Bench, which contains 464 real-world degraded images and tailored evaluation metrics focusing on degradation removal and consistency preservation. Extensive experiments demonstrate our model ranks first among open-source methods, achieving state-of-the-art performance.
- Calibri: Enhancing Diffusion Transformers via Parameter-Efficient Calibration
In this paper, we uncover the hidden potential of Diffusion Transformers (DiTs) to significantly enhance generative tasks. Through an in-depth analysis of the denoising process, we demonstrate that introducing a single learned scaling parameter can significantly improve the performance of DiT blocks. Building on this insight, we propose Calibri, a parameter-efficient approach that optimally calibrates DiT components to elevate generative quality. Calibri frames DiT calibration as a black-box reward optimization problem, which is efficiently solved using an evolutionary algorithm and modifies just ~100 parameters. Experimental results reveal that despite its lightweight design, Calibri consistently improves performance across various text-to-image models. Notably, Calibri also reduces the inference steps required for image generation, all while maintaining high-quality outputs.
- MACRO: Advancing Multi-Reference Image Generation with Structured Long-Context Data
Generating images conditioned on multiple visual references is critical for real-world applications such as multi-subject composition, narrative illustration, and novel view synthesis, yet current models suffer from severe performance degradation as the number of input references grows. We identify the root cause as a fundamental data bottleneck: existing datasets are dominated by single- or few-reference pairs and lack the structured, long-context supervision needed to learn dense inter-reference dependencies. To address this, we introduce MacroData, a large-scale dataset of 400K samples, each containing up to 10 reference images, systematically organized across four complementary dimensions -- Customization, Illustration, Spatial reasoning, and Temporal dynamics -- to provide comprehensive coverage of the multi-reference generation space. Recognizing the concurrent absence of standardized evaluation protocols, we further propose MacroBench, a benchmark of 4,000 samples that assesses generative coherence across graded task dimensions and input scales. Extensive experiments show that fine-tuning on MacroData yields substantial improvements in multi-reference generation, and ablation studies further reveal synergistic benefits of cross-task co-training and effective strategies for handling long-context complexity. The dataset and benchmark will be publicly released.
- Voxtral TTS
We introduce Voxtral TTS, an expressive multilingual text-to-speech model that generates natural speech from as little as 3 seconds of reference audio. Voxtral TTS adopts a hybrid architecture that combines auto-regressive generation of semantic speech tokens with flow-matching for acoustic tokens. These tokens are encoded and decoded with Voxtral Codec, a speech tokenizer trained from scratch with a hybrid VQ-FSQ quantization scheme. In human evaluations conducted by native speakers, Voxtral TTS is preferred for multilingual voice cloning due to its naturalness and expressivity, achieving a 68.4\% win rate over ElevenLabs Flash v2.5. We release the model weights under a CC BY-NC license.
- SlopCodeBench: Benchmarking How Coding Agents Degrade Over Long-Horizon Iterative Tasks
Software development is iterative, yet agentic coding benchmarks overwhelmingly evaluate single-shot solutions against complete specifications. Code can pass the test suite but become progressively harder to extend. Recent iterative benchmarks attempt to close this gap, but constrain the agent's design decisions too tightly to faithfully measure how code quality shapes future extensions. We introduce SlopCodeBench, a language-agnostic benchmark comprising 20 problems and 93 checkpoints, in which agents repeatedly extend their own prior solutions under evolving specifications that force architectural decisions without prescribing internal structure. We track two trajectory-level quality signals: verbosity, the fraction of redundant or duplicated code, and structural erosion, the share of complexity mass concentrated in high-complexity functions. No agent solves any problem end-to-end across 11 models; the highest checkpoint solve rate is 17.2%. Quality degrades steadily: erosion rises in 80% of trajectories and verbosity in 89.8%. Against 48 open-source Python repositories, agent code is 2.2x more verbose and markedly more eroded. Tracking 20 of those repositories over time shows that human code stays flat, while agent code deteriorates with each iteration. A prompt-intervention study shows that initial quality can be improved, but it does not halt degradation. These results demonstrate that pass-rate benchmarks systematically undermeasure extension robustness, and that current agents lack the design discipline iterative software development demands.
- MSA: Memory Sparse Attention for Efficient End-to-End Memory Model Scaling to 100M Tokens
Long-term memory is a cornerstone of human intelligence. Enabling AI to process lifetime-scale information remains a long-standing pursuit in the field. Due to the constraints of full-attention architectures, the effective context length of large language models (LLMs) is typically limited to 1M tokens. Existing approaches, such as hybrid linear attention, fixed-size memory states (e.g., RNNs), and external storage methods like RAG or agent systems, attempt to extend this limit. However, they often suffer from severe precision degradation and rapidly increasing latency as context length grows, an inability to dynamically modify memory content, or a lack of end-to-end optimization. These bottlenecks impede complex scenarios like large-corpus summarization, Digital Twins, and long-history agent reasoning, while limiting memory capacity and slowing inference. We present Memory Sparse Attention (MSA), an end-to-end trainable, efficient, and massively scalable memory model framework. Through core innovations including scalable sparse attention and document-wise RoPE, MSA achieves linear complexity in both training and inference while maintaining exceptional stability, exhibiting less than 9% degradation when scaling from 16K to 100M tokens. Furthermore, KV cache compression, combined with Memory Parallel, enables 100M-token inference on 2xA800 GPUs. We also propose Memory Interleaving to facilitate complex multi-hop reasoning across scattered memory segments. MSA significantly surpasses frontier LLMs, state-of-the-art RAG systems, and leading memory agents in long-context benchmarks. These results demonstrate that by decoupling memory capacity from reasoning, MSA provides a scalable foundation to endow general-purpose models with intrinsic, lifetime-scale memory.
- AVControl: Efficient Framework for Training Audio-Visual Controls
Controlling video and audio generation requires diverse modalities, from depth and pose to camera trajectories and audio transformations, yet existing approaches either train a single monolithic model for a fixed set of controls or introduce costly architectural changes for each new modality. We introduce AVControl, a lightweight, extendable framework built on LTX-2, a joint audio-visual foundation model, where each control modality is trained as a separate LoRA on a parallel canvas that provides the reference signal as additional tokens in the attention layers, requiring no architectural changes beyond the LoRA adapters themselves. We show that simply extending image-based in-context methods to video fails for structural control, and that our parallel canvas approach resolves this. On the VACE Benchmark, we outperform all evaluated baselines on depth- and pose-guided generation, inpainting, and outpainting, and show competitive results on camera control and audio-visual benchmarks. Our framework supports a diverse set of independently trained modalities: spatially-aligned controls such as depth, pose, and edges, camera trajectory with intrinsics, sparse motion control, video editing, and, to our knowledge, the first modular audio-visual controls for a joint generation model. Our method is both compute- and data-efficient: each modality requires only a small dataset and converges within a few hundred to a few thousand training steps, a fraction of the budget of monolithic alternatives. We publicly release our code and trained LoRA checkpoints.
- Representation Alignment for Just Image Transformers is not Easier than You Think
Representation Alignment (REPA) has emerged as a simple way to accelerate Diffusion Transformers training in latent space. At the same time, pixel-space diffusion transformers such as Just image Transformers (JiT) have attracted growing attention because they remove a dependency on a pretrained tokenizer, and then avoid the reconstruction bottleneck of latent diffusion. This paper shows that the REPA can fail for JiT. REPA yields worse FID for JiT as training proceeds and collapses diversity on image subsets that are tightly clustered in the representation space of pretrained semantic encoder on ImageNet. We trace the failure to an information asymmetry: denoising occurs in the high dimensional image space, while the semantic target is strongly compressed, making direct regression a shortcut objective. We propose PixelREPA, which transforms the alignment target and constrains alignment with a Masked Transformer Adapter that combines a shallow transformer adapter with partial token masking. PixelREPA improves both training convergence and final quality. PixelREPA reduces FID from 3.66 to 3.17 for JiT-B/16 and improves Inception Score (IS) from 275.1 to 284.6 on ImageNet 256 times 256, while achieving > 2times faster convergence. Finally, PixelREPA-H/16 achieves FID=1.81 and IS=317.2. Our code is available at https://github.com/kaist-cvml/PixelREPA.
- Less Gaussians, Texture More: 4K Feed-Forward Textured Splatting
Existing feed-forward 3D Gaussian Splatting methods predict pixel-aligned primitives, leading to a quadratic growth in primitive count as resolution increases. This fundamentally limits their scalability, making high-resolution synthesis such as 4K intractable. We introduce LGTM (Less Gaussians, Texture More), a feed-forward framework that overcomes this resolution scaling barrier. By predicting compact Gaussian primitives coupled with per-primitive textures, LGTM decouples geometric complexity from rendering resolution. This approach enables high-fidelity 4K novel view synthesis without per-scene optimization, a capability previously out of reach for feed-forward methods, all while using significantly fewer Gaussian primitives. Project page: https://yxlao.github.io/lgtm/
- S2D2: Fast Decoding for Diffusion LLMs via Training-Free Self-Speculation
Block-diffusion language models offer a promising path toward faster-than-autoregressive generation by combining block-wise autoregressive decoding with within-block parallel denoising. However, in the few-step regime needed for practical acceleration, standard confidence-thresholded decoding is often brittle: aggressive thresholds hurt quality, while conservative thresholds require unnecessary denoising steps. Existing approaches that address this issue either require additional training or incur extra test-time compute. We present S2D2, a training-free self-speculative decoding framework for block-diffusion language models. Our key observation is that a block-diffusion model becomes autoregressive when the block size is reduced to one, allowing the same pretrained model to act as both drafter and verifier. S2D2 inserts a speculative verification step into standard block-diffusion decoding and uses lightweight routing policies to decide when verification is worth its cost. This yields a hybrid decoding trajectory in which diffusion proposes tokens in parallel, while the autoregressive mode acts as a local sequence-level critic. Across three mainstream block-diffusion families, S2D2 consistently improves the accuracy-speed tradeoff over strong confidence-thresholding baselines. On SDAR, we observe up to 4.7times speedup over autoregressive decoding, and up to 1.57times over a tuned dynamic decoding baseline while improving accuracy by up to 4.5 points. On LLaDA2.1-Mini, S2D2 remains complementary to built-in self-correction, including a conservative setting where it is 4.4times faster than the static baseline with slightly higher accuracy.
- MuRF: Unlocking the Multi-Scale Potential of Vision Foundation Models
Vision Foundation Models (VFMs) have become the cornerstone of modern computer vision, offering robust representations across a wide array of tasks. While recent advances allow these models to handle varying input sizes during training, inference typically remains restricted to a single, fixed scale. This prevalent single-scale paradigm overlooks a fundamental property of visual perception: varying resolutions offer complementary inductive biases, where low-resolution views excel at global semantic recognition and high-resolution views are essential for fine-grained refinement. In this work, we propose Multi-Resolution Fusion (MuRF), a simple yet universally effective strategy to harness this synergy at inference time. Instead of relying on a single view, MuRF constructs a unified representation by processing an image at multiple resolutions through a frozen VFM and fusing the resulting features. The universality of MuRF is its most compelling attribute. It is not tied to a specific architecture, serving instead as a fundamental, training-free enhancement to visual representation. We empirically validate this by applying MuRF to a broad spectrum of critical computer vision tasks across multiple distinct VFM families - primarily DINOv2, but also demonstrating successful generalization to contrastive models like SigLIP2.
- FinMCP-Bench: Benchmarking LLM Agents for Real-World Financial Tool Use under the Model Context Protocol
This paper introduces FinMCP-Bench, a novel benchmark for evaluating large language models (LLMs) in solving real-world financial problems through tool invocation of financial model context protocols. FinMCP-Bench contains 613 samples spanning 10 main scenarios and 33 sub-scenarios, featuring both real and synthetic user queries to ensure diversity and authenticity. It incorporates 65 real financial MCPs and three types of samples, single tool, multi-tool, and multi-turn, allowing evaluation of models across different levels of task complexity. Using this benchmark, we systematically assess a range of mainstream LLMs and propose metrics that explicitly measure tool invocation accuracy and reasoning capabilities. FinMCP-Bench provides a standardized, practical, and challenging testbed for advancing research on financial LLM agents.
- BioVITA: Biological Dataset, Model, and Benchmark for Visual-Textual-Acoustic Alignment
Understanding animal species from multimodal data poses an emerging challenge at the intersection of computer vision and ecology. While recent biological models, such as BioCLIP, have demonstrated strong alignment between images and textual taxonomic information for species identification, the integration of the audio modality remains an open problem. We propose BioVITA, a novel visual-textual-acoustic alignment framework for biological applications. BioVITA involves (i) a training dataset, (ii) a representation model, and (iii) a retrieval benchmark. First, we construct a large-scale training dataset comprising 1.3 million audio clips and 2.3 million images, covering 14,133 species annotated with 34 ecological trait labels. Second, building upon BioCLIP2, we introduce a two-stage training framework to effectively align audio representations with visual and textual representations. Third, we develop a cross-modal retrieval benchmark that covers all possible directional retrieval across the three modalities (i.e., image-to-audio, audio-to-text, text-to-image, and their reverse directions), with three taxonomic levels: Family, Genus, and Species. Extensive experiments demonstrate that our model learns a unified representation space that captures species-level semantics beyond taxonomy, advancing multimodal biodiversity understanding. The project page is available at: https://dahlian00.github.io/BioVITA_Page/
Techmeme(15)
- Toronto-based quantum computing company Xanadu's stock closed up 15% in its trading debut on Nasdaq; it also began trading on the Toronto Stock Exchange (Josh Scott/BetaKit)
Josh Scott / BetaKit : Toronto-based quantum computing company Xanadu's stock closed up 15% in its trading debut on Nasdaq; it also began trading on the Toronto Stock Exchange — Toronto-based quantum company the first new Canadian tech firm to list on the TSX since 2021. — Toronto quantum computing firm …
- Sources: Physical Intelligence, which is developing AI models for robotics, is discussing a new funding round of about $1B that would value it at $11B+ (Bloomberg)
Bloomberg : Sources: Physical Intelligence, which is developing AI models for robotics, is discussing a new funding round of about $1B that would value it at $11B+ — Physical Intelligence, a two-year-old robotics startup founded by AI academics and former Google DeepMind researchers …
- Court docs: Zuckerberg texted Musk approvingly about his work with DOGE in February 2025, offering to "take down content doxxing or threatening" DOGE staffers (Karissa Bell/Engadget)
Karissa Bell / Engadget : Court docs: Zuckerberg texted Musk approvingly about his work with DOGE in February 2025, offering to “take down content doxxing or threatening” DOGE staffers — The two have a complicated history, but were apparently on friendlier terms at the start of the Trump Administration.
- Tech stocks suffer their worst week in almost a year, driven by the Iran war and Meta's legal defeats; Meta fell 11%, Alphabet fell ~9%, and Microsoft fell ~7% (Ari Levy/CNBC)
Ari Levy / CNBC : Tech stocks suffer their worst week in almost a year, driven by the Iran war and Meta's legal defeats; Meta fell 11%, Alphabet fell ~9%, and Microsoft fell ~7% — A bad week for stocks was particularly rough for tech investors, as the Nasdaq suffered its worst weekly drop since April 2025.
- US memory chip stocks lost ~$100B in market value this week, led by Micron's 15% drop, after Google Research detailed its TurboQuant compression algorithm (Financial Times)
Financial Times : US memory chip stocks lost ~$100B in market value this week, led by Micron's 15% drop, after Google Research detailed its TurboQuant compression algorithm — New research suggests AI data centres will need much less memory than investors had bargained for — US memory chip stocks …
- Filing: Kalshi has secured a license allowing it to offer margin trading to users, which would make the platform more appealing to institutional investors (Bloomberg)
Bloomberg : Filing: Kalshi has secured a license allowing it to offer margin trading to users, which would make the platform more appealing to institutional investors — Kalshi Inc. has secured a license allowing it to offer margin trading to users, a feature that would make the prediction market platform …
- The Trump administration launches the White House app, promising news "straight from the source, no filter", with news, live, social, and gallery sections (Emily Goodin/New York Post)
Emily Goodin / New York Post : The Trump administration launches the White House app, promising news “straight from the source, no filter”, with news, live, social, and gallery sections — The Trump administration announced the launch of the White House app on Friday, promising news “straight from the source, no filter.”
- The FBI confirms Iran-linked Handala breached Kash Patel's personal email but says the data accessed was "historical in nature" and involves no government info (TechCrunch)
TechCrunch : The FBI confirms Iran-linked Handala breached Kash Patel's personal email but says the data accessed was “historical in nature” and involves no government info — Lorenzo Franceschi-Bicchierai Zack Whittaker — A hacking group backed by the Iranian government dubbed “Handala” …
- Sources: Google nears a deal to help finance Nexus Data Centers' Texas campus that is leased to Anthropic, as Google deepens its partnership with the company (Financial Times)
Financial Times : Sources: Google nears a deal to help finance Nexus Data Centers' Texas campus that is leased to Anthropic, as Google deepens its partnership with the company — Texas site for Nexus Data Centers aims to avoid grid connection delays with direct gas supplies
- China's broadening Manus review raises doubts about the "Singapore-washing" model; sources: Meta moved 100+ Manus employees to Singapore in early March (Anniek Bao/CNBC)
Anniek Bao / CNBC : China's broadening Manus review raises doubts about the “Singapore-washing” model; sources: Meta moved 100+ Manus employees to Singapore in early March — Tech circles from Silicon Valley to Shenzhen buzzed when Meta acquired Manus, a Singaporean AI startup with Chinese roots, for $2 billion late last year.
- Microsoft will lease Crusoe's 900 MW data center in Abilene, Texas, after Oracle and OpenAI reportedly withdrew, with the first building expected by mid-2027 (Matt Day/Bloomberg)
Matt Day / Bloomberg : Microsoft will lease Crusoe's 900 MW data center in Abilene, Texas, after Oracle and OpenAI reportedly withdrew, with the first building expected by mid-2027 — Microsoft Corp. will occupy a data center project previously earmarked for Oracle Corp. and OpenAI, scooping up 900 megawatts …
- Iran-linked hacker group Handala Hack Team claims the breach of FBI Director Kash Patel's personal email and publishes some documents online (Reuters)
Reuters : Iran-linked hacker group Handala Hack Team claims the breach of FBI Director Kash Patel's personal email and publishes some documents online — Iran-linked hackers have publicly claimed the breach of FBI Director Kash Patel's personal inbox, publishing photographs of the director and his purported resume to the internet.
- Meta agrees to fund Entergy Louisiana's new energy infrastructure for its Louisiana data center, including seven natural gas power plants (Nicholas G. Miller/Wall Street Journal)
Nicholas G. Miller / Wall Street Journal : Meta agrees to fund Entergy Louisiana's new energy infrastructure for its Louisiana data center, including seven natural gas power plants — The company will fund new natural gas power plants, transmissions lines and battery energy storage — Meta Platforms agreed to a deal …
- Sources: Aetherflux, founded by Robinhood co-founder Baiju Bhatt to develop orbital data centers, aims to raise $250M to $300M in a Series B at a $2B valuation (Wall Street Journal)
Wall Street Journal : Sources: Aetherflux, founded by Robinhood co-founder Baiju Bhatt to develop orbital data centers, aims to raise $250M to $300M in a Series B at a $2B valuation — Aetherflux, led by Robinhood co-founder Baiju Bhatt, seeks to launch solar-powered satellites for AI computing
- Anthropic says it's testing an AI model that's a "step change" in performance after a draft blog in an unsecured data store revealed the Claude Mythos model (Beatrice Nolan/Fortune)
Beatrice Nolan / Fortune : Anthropic says it's testing an AI model that's a “step change” in performance after a draft blog in an unsecured data store revealed the Claude Mythos model — AI company Anthropic is developing and has begun testing with early access customers a new AI model more capable …
Solidot(15)
- 苹果向 FBI 提供用马甲邮箱发出匿名威胁的用户名字
一位苹果用户在阅读了 FBI 局长 Kash Patel 动用政府资源派遣一整队人马为其女友 Alexis Wilkins 提供安保的新闻之后,使用 iCloud 隐藏真实邮箱的马甲邮箱功能向 Wilkins 发送匿名恐吓信。苹果向 FBI 交出了这名用户的真实名字。这位用户是在 2026 年 2 月 28 日使用马甲邮箱 peaty_terms_1o@icloud.com 发送了邮件,其真实名字是 Alden Ruml。数据显示他的账户生成了 134 个马甲邮箱。执行人员询问了 Ruml,他证实匿名信是其所发送。
- Ubuntu 26.04 LTS Beta 释出
Ubuntu 26.04 LTS Beta 释出,v26.04 是一个长期支持版本,正式发布日期定为 4 月 23 日。Ubuntu 26.04 主要变化包括:Linux 7.0 kernel(还在开发中,即将在一两周内发布)、GNOME 50.0 桌面环境、Mesa 26.0 图形驱动、Python 3.14、GCC 15.2 以及一系列软件更新,等等。
- AI 如何削弱我们的判断力
根据发表在《科学》期刊上的一项研究,为人际关系问题提供建议和支持的 AI 聊天机器人可能会通过明显谄媚的回答而悄然强化有害的信念。研究发现,在各种语境下,聊天机器人肯定人类用户的频率远超真人之间相互肯定的频率;由此产生的有害后果包括:用户更坚信自己正确且更不愿去修复人际关系。研究人员利用 Reddit 社区“AITA”中的帖子评估了 OpenAI、Anthropic、Google 等公司的 11 种先进且广泛使用的 AI 大模型;结果发现,这些系统对用户行为的肯定频率比真人高出 49%,即使是在涉及欺骗、伤害或违法的场景中也是如此。在两项后续的实验中,研究人员探讨了这类结果所导致的行为后果。研究结果显示,在涉及人际交往情境(尤其是冲突)时,与谄媚式 AI 互动的参与者会更坚信自己是正确的,并且即使仅经过一次互动,他们和解或承担责任的意愿也会降低。
- Mozilla 和 Mila 联合推进开源主权 AI
AI 的未来应该属于全人类,不能局限于少数国家或公司。为了实现这一目标,AI 必须开放、值得信赖,且其构建方式应赋予个人、机构和国家真正的选择权。正因如此 Mozilla 宣布与加拿大魁北克人工智能研究所 Mila 建立战略合作伙伴关系,联合推进开源主权 AI。Mila 和 Mozilla 将合作开发相关技术和方法,减少对封闭系统的依赖,为透明度、问责制和共享创新创造更多空间。双方暂时还没有公布更多信息。
- Reddit 开始推出验证用户是否是人类的检查机制
Reddit CEO 兼联合创始人 Steve Huffman 周三宣布开始推出验证用户是否是人类的检查机制,声称会以保护用户隐私作为首要原则,只是为了确认用户是人类而不是具体哪个人。Reddit 表示只有在检测到用户账号可疑时才会要求对该账号进行验证,不会要求网站所有用户进行验证。账号可疑的信息包括了撰写或发布内容的速度。 为了验证账户是否为人类所有,Reddit 将利用第三方工具如苹果、Google、YubiKey 的 passkeys,第三方生物识别服务如 Face ID 甚至 Sam Altman 的 World ID,或者在部分国家要求政府颁发的身份证件。
- 维基百科禁止使用生成式 AI 撰写或改写文章
维基百科禁止使用生成式 AI 撰写或改写文章,但用 AI 翻译或润色仍然可以,前提是需要人工仔细检查内容的准确性,因为基于大模型的 AI 工具可能会以出人意料的方式改变文本含义,与内容引用来源牛头不对马嘴。而使用 AI 工具翻译的志愿编辑需要对翻译的两种语言都有所精通以发现可能的翻译错误。最新的政策如何执行仍然是一大难题。
- 过度耕作如何削弱土壤
耕作是一种通过翻动表层土壤露出新鲜泥土为播种做准备的古老农耕方式,至今仍然流行,但对土壤退化的担忧推动了被称为再生耕作(regenerative)的新农耕方法。华盛顿大学研究人员利用测量地震的方法测量了耕作对土壤水分和保水能力的影响。研究人员在英国的一块实验农场周围铺设了光纤记录了不同耕作以及农机设备造成的地面运动。研究报告发表在《科学》期刊上。结果显示,耕作和农机的压实会破坏土壤中复杂的毛细管网络,而正是毛细管网络赋予了土壤天然的海绵状性质。耕作的目的是在土壤中形成孔隙,使水分能到达植物根部,但实际上它却破坏了土壤中的细小通道,导致雨水积聚在地表,形成一层泥泞的硬壳。随着时间的推移,这会加剧土壤侵蚀和洪水风险。
- 鲸鱼在分娩过程中展开合作
根据发表在《科学》期刊上的一项研究,研究人员捕捉到了一头抹香鲸幼崽的诞生过程,记录了来自两个通常相互独立家族群体的 11 头抹香鲸如何密切协作,旨在幼崽出生后为其提供数小时的支持。这些发现为鲸目动物的直接群体照料行为提供了量化证据;它们同时也表明,在分娩等关键时刻展现的短期、高度协调性合作行为可能对维系抹香鲸群体中复杂的社会结构起着根本性的作用。2023 年 7 月,在多米尼克(Dominica)海岸附近,研究人员观测到一个已知的抹香鲸社会单元中的 11 名成员在距离海面极近的位置聚集。通过无人机拍摄的影像,研究作者记录了长达 34 分钟的幼崽分娩过程;分娩之后是一段紧张但协调的活动,期间多头成年雌性抹香鲸环绕在分娩母鲸的周围。在幼崽出生后的一小时内,该群体展现了惊人的协作行为;来自两个家族的成员会轮流用身体托举推升新生幼崽至水面,此举很可能是为了协助幼崽呼吸。在这一关键时期,整个群体单元始终保持着紧密的组织性。弗氏海豚(Fraser’s dolphins)曾数次近距离掠过该鲸群,并与领航鲸有短暂的互动。在幼崽出生数小时后,这群抹香鲸逐渐分散成规模较小、更为典型的觅食群体。
- 实验室培育食管恢复猪的吞咽能力
英国科学家首次在实验室培养出功能完整的食管,并移植到猪体内,成功恢复了其正常吞咽功能。这项成果被视为器官替代工程领域的重要进步,为未来 5 年内启动人类临床试验奠定了坚实基础,有望彻底改变食管闭锁患儿的命运。食管闭锁是一种严重的先天缺陷,患儿出生时食管与胃部未能连接。全球每年约有 180 名新生儿受此困扰。当前标准治疗方案是通过复杂手术进行重建,但医生常需截取患儿肠道或上提胃部来替代食管。这不仅创伤大,还可能导致终生呼吸、消化问题及远期并发症风险。为应对这一难题,英国大奥蒙德街医院与伦敦大学学院研究团队取得了突破。他们开发了一套创新的生物工程工艺:首先以猪食管为“支架”,经过去细胞化处理,剥离其所有原始细胞,仅保留纯净的天然微观结构框架;随后,团队从受体猪身上获取微小肌肉活检样本,提取肌肉前体细胞,在实验室中大量扩增后,将其注入支架内部。整个结构在生物反应器中培养约两个月,其间持续灌注营养液,模拟体内环境,促使细胞定居、增殖,最终形成一段活的、由受体自身细胞构成的新食管。团队将实验室培育的 8 个食管移植到猪体内,所有受体猪在术后关键 30 天均存活。
- 微软将默认收集 GitHub Copilot 交互数据训练 AI
微软旗下的代码托管平台 GitHub 宣布更新 GitHub Copilot 交互数据使用政策。从 4 月 24 日起,除非用户主动选择退出,否则微软/GitHub 将使用 GitHub Copilot 交互数据训练 AI 模型。微软默认启用了数据收集,此举被认为违反了欧盟的通用数据保护法规 GDPR,GDPR 强制要求默认选择退出。
- 晋江封禁 “老天奶”引争议
晋江文学城因内容审核标准问题陷入舆论漩涡,并因此引发了一场大规模的用户“发票抗议”。风波始于晋江文学城于近日突然锁定热门女性向小说《女主对此感到厌烦》,相关提示为“文章设定有问题,锁文要求清理”。这一操作迅速引发读者的不满与质疑。锁文事件引发舆论后,晋江平台管理员发帖说明《女主对此感到厌烦》临时锁定事件,称该作品因疑似含有挑动对立情绪的内容,网站依规对其进行临时锁定核查。经初步核查确认,未发现该文存在无差别扫射某一群体等明显违规行为,现对该作品恢复上架。然而这份说明中关于内容审核的具体标准,却引发了更大的争议。说明中明确“严禁使用非经汉语权威机构承认的生造字词或短语”,并举例“老天奶(为网络通用口语,类似‘老天爷’的变体)”。但有用户发现,晋江官方账号此前曾使用“老天奶”一词进行内容宣传。在被用户质疑后,相关博文已被删除。3 月 25 日,对晋江文学城不满的用户集体申请历史充值发票,到晚间,排队人数飙升至 19 万以上,规模空前。
- GNOME 基金会宣布面向资深开发者的奖学金计划
GNOME 基金会宣布了面向资深开发者的奖学金(Fellowship)计划,资助有经验的开发者投入 1 年/12 个月时间专注于某个领域的开发。奖学金的申请者必须是 GNOME 项目的资深贡献者,拥有良好的记录。根据其经验和所处地点他们将获得 7 万 - 10 万美元的资助。首批将于 5 月公布。首批重点领域包括构建系统、CI/CD 基础设施、开发者工具、文档、可访问性和解决技术债务。
- Sora 为何失败:每天推理成本最高 1500 万美元总收入仅为 210 万美元
Sora 一度被视为是视频的未来,但却成为 OpenAI 少数关闭的产品之一。很多人为之惋惜,但数据显示这款产品是注定要关闭的,因为其经济模式不可维持。Sora 在鼎盛时期每天的推理成本高达 1500 万美元,一年的服务器总支出可能高达几十亿美元,而该应用至今的总收入为 210 万美元,也就是收入相对于支出几乎为零。Sora 的活跃用户数也远远少于它的聊天机器人 ChatGPT:Sora 在 2025 年 11 月在 iOS 和 Google Play 的下载量为 333 万,但到了 2026 年 2 月下载量 110 万次左右,跌至峰值的三分之一,它的月活用户数在 2025 年 12 月达到峰值,之后开始下降,也就是用户在流失而不是增长。
- 汽车钠离子电池能在 11 分钟内完成充电
在宁德时代和长安发布首款搭载钠离子电池的量产电动汽车后,北汽宣布了它的极光钠离子电池原型。钠离子电池成本比锂离子电池低,对原材料的价格不那么敏感,宁德时代、比亚迪等都在押注钠离子电池以应对不断上涨的锂材料价格。北汽称其钠离子电池能在 -40°C 至 60°C 内稳定工作,-20°C能量保持率超 92%,电芯能量密度 170Wh/kg,充电仅需 11 分钟,搭载该电池的 CLTC 汽车续航里程可达 450 公里。
- 学生说服学校设立 Tor 中继节点
台师大计算机科学和信息工程系学生苏恩立(Su En-Li,NZ)成功说服学校设立了首个校园 Tor 中继节点。Tor 中继节点与出口节点不同,它只是中继加密的流量而不直接向用户传输内容,因此风险较低。苏恩立通过公开正规的行政流程与学校的网络管理员、教授和系主任进行邮件沟通,最终成功在管理严密的学术网络 TANet 内设立了首个 Tor 中继节点。苏恩立还通过组织一系列活动帮助人们理解匿名网络≠犯罪工具。