OrangeBot.AI Digest — 2026-03-09
85 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Bluesky CEO Jay Graber is stepping down (bsky.social)
- Workers report watching Ray-Ban Meta-shot footage of people using the bathroom (arstechnica.com)
- JSLinux Now Supports x86_64 (bellard.org)
- Florida judge rules red light camera tickets are unconstitutional (cbs12.com)
- Building a Procedural Hex Map with Wave Function Collapse (felixturner.github.io)
- Jolla on track to ship new phone with Sailfish OS, user-replaceable battery (liliputing.com)
- Is legal the same as legitimate: AI reimplementation and the erosion of copyleft (writings.hongminhee.org)
- Kuwaiti F/A-18's Triple Friendly Fire Shootdown Gets Stranger by the Day (www.twz.com)
- The engine of Germany's wealth is blocking its future (europeancorrespondent.com)
- New farm bill would condemn pigs to a lifetime in gestation crates (twitter.com)
- The window chrome of our discontent (pxlnv.com)
- FFmpeg at Meta: Media Processing at Scale (engineering.fb.com)
- Ireland shuts last coal plant, becomes 15th coal-free country in Europe (2025) (www.pv-magazine.com)
- Fontcrafter: Turn Your Handwriting into a Real Font (arcade.pirillo.com)
- Show HN: Mcp2cli – One CLI for every API, 96-99% fewer tokens than native MCP (github.com)
GitHub Trending(12)
Product Hunt(15)
- SCRAPR
Turn any website into an API
- Kita
Turn documents into signals for lenders
- Wideframe
AI Coworker for Video Editors
- Flowripple
Easily Trigger workflows from your SaaS
- OpenClix
Agent-driven retention flows for mobile apps.
- simply
ai nutrition app
- Roundtable
Launch your EU investment fund in days, not months
- Hannah & Co
AI coworkers for marketing.
- Dex
Ask your data. Get answers and next steps.
- Timelaps
Know if your marketing is working with real-time insights
- BrandingStudio.ai
Agency-quality branding in 60 minutes, not 6 months
- Reflct
The journaling habit you'll actually keep
- Unite Pro for macOS
Turn websites into Mac apps
- Phi-4-reasoning-vision
Open-weight 15B multimodal model for thinking and GUI agents
- cutefolio
build portfolios that actually look cute.
Hugging Face(15)
- Penguin-VL: Exploring the Efficiency Limits of VLM with LLM-based Vision Encoders
Vision Language Model (VLM) development has largely relied on scaling model size, which hinders deployment on compute-constrained mobile and edge devices such as smartphones and robots. In this work, we explore the performance limits of compact (e.g., 2B and 8B) VLMs. We challenge the prevailing practice that state-of-the-art VLMs must rely on vision encoders initialized via massive contrastive pretraining (e.g., CLIP/SigLIP). We identify an objective mismatch: contrastive learning, optimized for discrimination, enforces coarse and category-level invariances that suppress fine-grained visual cues needed for dense captioning and complex VLM reasoning. To address this issue, we present Penguin-VL, whose vision encoder is initialized from a text-only LLM. Our experiments reveal that Penguin-Encoder serves as a superior alternative to traditional contrastive pretraining, unlocking a higher degree of visual fidelity and data efficiency for multimodal understanding. Across various image and video benchmarks, Penguin-VL achieves performance comparable to leading VLMs (e.g., Qwen3-VL) in mathematical reasoning and surpasses them in tasks such as document understanding, visual knowledge, and multi-perspective video understanding. Notably, these gains are achieved with a lightweight architecture, demonstrating that improved visual representation rather than model scaling is the primary driver of performance. Our ablations show that Penguin-Encoder consistently outperforms contrastive-pretrained encoders, preserving fine-grained spatial and temporal cues that are critical for dense perception and complex reasoning. This makes it a strong drop-in alternative for compute-efficient VLMs and enables high performance in resource-constrained settings. Code: https://github.com/tencent-ailab/Penguin-VL
- BandPO: Bridging Trust Regions and Ratio Clipping via Probability-Aware Bounds for LLM Reinforcement Learning
Proximal constraints are fundamental to the stability of the Large Language Model reinforcement learning. While the canonical clipping mechanism in PPO serves as an efficient surrogate for trust regions, we identify a critical bottleneck: fixed bounds strictly constrain the upward update margin of low-probability actions, disproportionately suppressing high-advantage tail strategies and inducing rapid entropy collapse. To address this, we introduce Band-constrained Policy Optimization (BandPO). BandPO replaces canonical clipping with Band, a unified theoretical operator that projects trust regions defined by f-divergences into dynamic, probability-aware clipping intervals. Theoretical analysis confirms that Band effectively resolves this exploration bottleneck. We formulate this mapping as a convex optimization problem, guaranteeing a globally optimal numerical solution while deriving closed-form solutions for specific divergences. Extensive experiments across diverse models and datasets demonstrate that BandPO consistently outperforms canonical clipping and Clip-Higher, while robustly mitigating entropy collapse.
- Planning in 8 Tokens: A Compact Discrete Tokenizer for Latent World Model
World models provide a powerful framework for simulating environment dynamics conditioned on actions or instructions, enabling downstream tasks such as action planning or policy learning. Recent approaches leverage world models as learned simulators, but its application to decision-time planning remains computationally prohibitive for real-time control. A key bottleneck lies in latent representations: conventional tokenizers encode each observation into hundreds of tokens, making planning both slow and resource-intensive. To address this, we propose CompACT, a discrete tokenizer that compresses each observation into as few as 8 tokens, drastically reducing computational cost while preserving essential information for planning. An action-conditioned world model that occupies CompACT tokenizer achieves competitive planning performance with orders-of-magnitude faster planning, offering a practical step toward real-world deployment of world models.
- WildActor: Unconstrained Identity-Preserving Video Generation
Production-ready human video generation requires digital actors to maintain strictly consistent full-body identities across dynamic shots, viewpoints and motions, a setting that remains challenging for existing methods. Prior methods often suffer from face-centric behavior that neglects body-level consistency, or produce copy-paste artifacts where subjects appear rigid due to pose locking. We present Actor-18M, a large-scale human video dataset designed to capture identity consistency under unconstrained viewpoints and environments. Actor-18M comprises 1.6M videos with 18M corresponding human images, covering both arbitrary views and canonical three-view representations. Leveraging Actor-18M, we propose WildActor, a framework for any-view conditioned human video generation. We introduce an Asymmetric Identity-Preserving Attention mechanism coupled with a Viewpoint-Adaptive Monte Carlo Sampling strategy that iteratively re-weights reference conditions by marginal utility for balanced manifold coverage. Evaluated on the proposed Actor-Bench, WildActor consistently preserves body identity under diverse shot compositions, large viewpoint transitions, and substantial motions, surpassing existing methods in these challenging settings.
- Progressive Residual Warmup for Language Model Pretraining
Transformer architectures serve as the backbone for most modern Large Language Models, therefore their pretraining stability and convergence speed are of central concern. Motivated by the logical dependency of sequentially stacked layers, we propose Progressive Residual Warmup (ProRes) for language model pretraining. ProRes implements an "early layer learns first" philosophy by multiplying each layer's residual with a scalar that gradually warms up from 0 to 1, with deeper layers taking longer warmup steps. In this way, deeper layers wait for early layers to settle into a more stable regime before contributing to learning. We demonstrate the effectiveness of ProRes through pretraining experiments across various model scales, as well as normalization and initialization schemes. Comprehensive analysis shows that ProRes not only stabilizes pretraining but also introduces a unique optimization trajectory, leading to faster convergence, stronger generalization and better downstream performance. Our code is available at https://github.com/dandingsky/ProRes.
- RoboMME: Benchmarking and Understanding Memory for Robotic Generalist Policies
Memory is critical for long-horizon and history-dependent robotic manipulation. Such tasks often involve counting repeated actions or manipulating objects that become temporarily occluded. Recent vision-language-action (VLA) models have begun to incorporate memory mechanisms; however, their evaluations remain confined to narrow, non-standardized settings. This limits their systematic understanding, comparison, and progress measurement. To address these challenges, we introduce RoboMME: a large-scale standardized benchmark for evaluating and advancing VLA models in long-horizon, history-dependent scenarios. Our benchmark comprises 16 manipulation tasks constructed under a carefully designed taxonomy that evaluates temporal, spatial, object, and procedural memory. We further develop a suite of 14 memory-augmented VLA variants built on the π0.5 backbone to systematically explore different memory representations across multiple integration strategies. Experimental results show that the effectiveness of memory representations is highly task-dependent, with each design offering distinct advantages and limitations across different tasks. Videos and code can be found at our website https://robomme.github.io.
- Reasoning Models Struggle to Control their Chains of Thought
Chain-of-thought (CoT) monitoring is a promising tool for detecting misbehaviors and understanding the motivations of modern reasoning models. However, if models can control what they verbalize in their CoT, it could undermine CoT monitorability. To measure this undesirable capability -- CoT controllability -- we introduce the CoT-Control evaluation suite, which includes tasks that require models to solve problems while adhering to CoT instructions, e.g., reasoning about a genetics question without using the word 'chromosome'. We show that reasoning models possess significantly lower CoT controllability than output controllability; for instance, Claude Sonnet 4.5 can control its CoT only 2.7% of the time but 61.9% when controlling its final output. We also find that CoT controllability is higher for larger models and decreases with more RL training, test-time compute, and increased problem difficulty. CoT controllability failures extend even to situations in which models are given incentives (as opposed to direct requests) to evade CoT monitors, although models exhibit slightly higher controllability when they are told they are being monitored. Similarly, eliciting controllability by adversarially optimizing prompts does not meaningfully increase controllability. Our results leave us cautiously optimistic that CoT controllability is currently unlikely to be a failure mode of CoT monitorability. However, the mechanism behind low controllability is not well understood. Given its importance for maintaining CoT monitorability, we recommend that frontier labs track CoT controllability in future models.
- FlashPrefill: Instantaneous Pattern Discovery and Thresholding for Ultra-Fast Long-Context Prefilling
Long-context modeling is a pivotal capability for Large Language Models, yet the quadratic complexity of attention remains a critical bottleneck, particularly during the compute-intensive prefilling phase. While various sparse attention mechanisms have been explored, they typically suffer from either significant search latency or insufficient sparsity. In this paper, we propose FlashPrefill, a framework enabling ultra-fast prefilling via instantaneous pattern discovery and thresholding. FlashPrefill leverages a fast block-searching technique to simultaneously locate dynamic vertical, slash, and block-sparse attention patterns. Crucially, it introduces a dynamic thresholding mechanism that bypasses the prohibitive overhead of sorting or accumulating attention scores while effectively eliminating the long-tail distribution to enhance sparsity. Extensive evaluations demonstrate that FlashPrefill achieves a substantial leap in efficiency, delivering an unprecedented 27.78x speedup on 256K sequences. Notably, unlike existing methods that incur efficiency degradation on shorter contexts, FlashPrefill maintains a 1.71x speedup even at a 4K context length, demonstrating its robustness and practical utility across varying sequence scales.
- HiMAP-Travel: Hierarchical Multi-Agent Planning for Long-Horizon Constrained Travel
Sequential LLM agents fail on long-horizon planning with hard constraints like budgets and diversity requirements. As planning progresses and context grows, these agents drift from global constraints. We propose HiMAP-Travel, a hierarchical multi-agent framework that splits planning into strategic coordination and parallel day-level execution. A Coordinator allocates resources across days, while Day Executors plan independently in parallel. Three key mechanisms enable this: a transactional monitor enforcing budget and uniqueness constraints across parallel agents, a bargaining protocol allowing agents to reject infeasible sub-goals and trigger re-planning, and a single policy trained with GRPO that powers all agents through role conditioning. On TravelPlanner, HiMAP-Travel with Qwen3-8B achieves 52.78% validation and 52.65% test Final Pass Rate (FPR). In a controlled comparison with identical model, training, and tools, it outperforms the sequential DeepTravel baseline by +8.67~pp. It also surpasses ATLAS by +17.65~pp and MTP by +10.0~pp. On FlexTravelBench multi-turn scenarios, it achieves 44.34% (2-turn) and 37.42% (3-turn) FPR while reducing latency 2.5x through parallelization.
- Dynamic Chunking Diffusion Transformer
Diffusion Transformers process images as fixed-length sequences of tokens produced by a static patchify operation. While effective, this design spends uniform compute on low- and high-information regions alike, ignoring that images contain regions of varying detail and that the denoising process progresses from coarse structure at early timesteps to fine detail at late timesteps. We introduce the Dynamic Chunking Diffusion Transformer (DC-DiT), which augments the DiT backbone with a learned encoder-router-decoder scaffold that adaptively compresses the 2D input into a shorter token sequence in a data-dependent manner using a chunking mechanism learned end-to-end with diffusion training. The mechanism learns to compress uniform background regions into fewer tokens and detail-rich regions into more tokens, with meaningful visual segmentations emerging without explicit supervision. Furthermore, it also learns to adapt its compression across diffusion timesteps, using fewer tokens at noisy stages and more tokens as fine details emerge. On class-conditional ImageNet 256{times}256, DC-DiT consistently improves FID and Inception Score over both parameter-matched and FLOP-matched DiT baselines across 4{times} and 16{times} compression, showing this is a promising technique with potential further applications to pixel-space, video and 3D generation. Beyond accuracy, DC-DiT is practical: it can be upcycled from pretrained DiT checkpoints with minimal post-training compute (up to 8{times} fewer training steps) and composes with other dynamic computation methods to further reduce generation FLOPs.
- Mario: Multimodal Graph Reasoning with Large Language Models
Recent advances in large language models (LLMs) have opened new avenues for multimodal reasoning. Yet, most existing methods still rely on pretrained vision-language models (VLMs) to encode image-text pairs in isolation, ignoring the relational structure that real-world multimodal data naturally form. This motivates reasoning on multimodal graphs (MMGs), where each node has textual and visual attributes and edges provide structural cues. Enabling LLM-based reasoning on such heterogeneous multimodal signals while preserving graph topology introduces two key challenges: resolving weak cross-modal consistency and handling heterogeneous modality preference. To address this, we propose Mario, a unified framework that simultaneously resolves the two above challenges and enables effective LLM-based reasoning over MMGs. Mario consists of two innovative stages. Firstly, a graph-conditioned VLM design that jointly refines textual and visual features through fine-grained cross-modal contrastive learning guided by graph topology. Secondly, a modality-adaptive graph instruction tuning mechanism that organizes aligned multimodal features into graph-aware instruction views and employs a learnable router to surface, for each node and its neighborhood, the most informative modality configuration to the LLM. Extensive experiments across diverse MMG benchmarks demonstrate that Mario consistently outperforms state-of-the-art graph models in both supervised and zero-shot scenarios for node classification and link prediction. The code will be made available at https://github.com/sunyuanfu/Mario.
- π-StepNFT: Wider Space Needs Finer Steps in Online RL for Flow-based VLAs
Flow-based vision-language-action (VLA) models excel in embodied control but suffer from intractable likelihoods during multi-step sampling, hindering online reinforcement learning. We propose \textit{boldsymbolπ-StepNFT} (Step-wise Negative-aware Fine-Tuning), a critic-and-likelihood-free framework that requires only a single forward pass per optimization step and eliminates auxiliary value networks. We identify that wider exploration spaces necessitate finer-grained, step-wise guidance for alignment. Empirically, π-StepNFT unlocks latent potential on LIBERO with competitive few-shot robustness. Moreover, it achieves superior generalization on ManiSkill, outperforming value-based baselines in OOD scenarios by preventing overfitting to multimodal features. This property offers a scalable solution promising for complex real-world applications.
- EffectMaker: Unifying Reasoning and Generation for Customized Visual Effect Creation
Visual effects (VFX) are essential for enhancing the expressiveness and creativity of video content, yet producing high-quality effects typically requires expert knowledge and costly production pipelines. Existing AIGC systems face significant challenges in VFX generation due to the scarcity of effect-specific data and the inherent difficulty of modeling supernatural or stylized effects. Moreover, these approaches often require per-effect fine-tuning, which severely limits their scalability and generalization to novel VFX. In this work, we present EffectMaker, a unified reasoning-generation framework that enables reference-based VFX customization. EffectMaker employs a multimodal large language model to interpret high-level effect semantics and reason about how they should adapt to a target subject, while a diffusion transformer leverages in-context learning to capture fine-grained visual cues from reference videos. These two components form a semantic-visual dual-path guidance mechanism that enables accurate, controllable, and effect-consistent synthesis without per-effect fine-tuning. Furthermore, we construct EffectData, the largest high-quality synthetic dataset containing 130k videos across 3k VFX categories, to improve generalization and scalability. Experiments show that EffectMaker achieves superior visual quality and effect consistency over state-of-the-art baselines, offering a scalable and flexible paradigm for customized VFX generation. Project page: https://effectmaker.github.io
- τ-Knowledge: Evaluating Conversational Agents over Unstructured Knowledge
Conversational agents are increasingly deployed in knowledge-intensive settings, where correct behavior depends on retrieving and applying domain-specific knowledge from large, proprietary, and unstructured corpora during live interactions with users. Yet most existing benchmarks evaluate retrieval or tool use independently of each other, creating a gap in realistic, fully agentic evaluation over unstructured data in long-horizon interactions. We introduce τ-Knowledge, an extension of τ-Bench for evaluating agents in environments where success depends on coordinating external, natural-language knowledge with tool outputs to produce verifiable, policy-compliant state changes. Our new domain, τ-Banking, models realistic fintech customer support workflows in which agents must navigate roughly 700 interconnected knowledge documents while executing tool-mediated account updates. Across embedding-based retrieval and terminal-based search, even frontier models with high reasoning budgets achieve only sim25.5% pass^1, with reliability degrading sharply over repeated trials. Agents struggle to retrieve the correct documents from densely interlinked knowledge bases and to reason accurately over complex internal policies. Overall, τ-Knowledge provides a realistic testbed for developing agents that integrate unstructured knowledge in human-facing deployments.
- Dynamic Model Routing and Cascading for Efficient LLM Inference: A Survey
The rapid growth of large language models (LLMs) with diverse capabilities, costs, and domains has created a critical need for intelligent model selection at inference time. While smaller models suffice for routine queries, complex tasks demand more capable models. However, static model deployment does not account for the complexity and domain of incoming queries, leading to suboptimal performance and increased costs. Dynamic routing systems that adaptively select models based on query characteristics have emerged as a solution to this challenge. We provide a systematic analysis of state-of-the-art multi-LLM routing and cascading approaches. In contrast to mixture-of-experts architectures, which route within a single model, we study routing across multiple independently trained LLMs. We cover diverse routing paradigms, including query difficulty, human preferences, clustering, uncertainty quantification, reinforcement learning, multimodality, and cascading. For each paradigm, we analyze representative methods and examine key trade-offs. Beyond taxonomy, we introduce a conceptual framework that characterizes routing systems along three dimensions: when decisions are made, what information is used, and how they are computed. This perspective highlights that practical systems are often compositional, integrating multiple paradigms under operational constraints. Our analysis demonstrates that effective multi-LLM routing requires balancing competing objectives. Choosing the optimal routing strategy depends on deployment and computational constraints. Well-designed routing systems can outperform even the most powerful individual models by strategically leveraging specialized capabilities across models while maximizing efficiency gains. Meanwhile, open challenges remain in developing routing mechanisms that generalize across diverse architectures, modalities, and applications.
Techmeme(15)
- HPE reports Q1 revenue up 18% YoY to $9.3B, vs. $9.37B est., Cloud and AI revenue down 2.7% YoY to $6.3B but reports an AI server backlog of $5B (Dina Bass/Bloomberg)
Dina Bass / Bloomberg : HPE reports Q1 revenue up 18% YoY to $9.3B, vs. $9.37B est., Cloud and AI revenue down 2.7% YoY to $6.3B but reports an AI server backlog of $5B — Hewlett Packard Enterprise Co. gave an outlook for revenue in the current quarter that exceeded analysts' estimates, a sign the company …
- Filing: Anthropic says it had $5B+ in all-time revenue since 2023 and could lose billions after clients paused deal talks due to supply-chain risk designation (Paresh Dave/Wired)
Paresh Dave / Wired : Filing: Anthropic says it had $5B+ in all-time revenue since 2023 and could lose billions after clients paused deal talks due to supply-chain risk designation — Executives at the AI startup say companies paused deal talks after the Trump administration labeled it a supply-chain risk …
- More than 30 staffers from OpenAI and Google, including DeepMind chief scientist Jeff Dean, file an amicus brief in support of Anthropic in its fight with DOD (Maxwell Zeff/Wired)
Maxwell Zeff / Wired : More than 30 staffers from OpenAI and Google, including DeepMind chief scientist Jeff Dean, file an amicus brief in support of Anthropic in its fight with DOD — Google DeepMind chief scientist Jeff Dean is among the AI researchers and engineers rushing to Anthropic's defense.
- MacBook Pro 16" (M5 Max) review: "Super core" architecture and 40-core GPU deliver beastly performance, but still retains a five-year-old design (Brian Westover/PCMag)
Brian Westover / PCMag : MacBook Pro 16" (M5 Max) review: “Super core” architecture and 40-core GPU deliver beastly performance, but still retains a five-year-old design — Outstanding — THE BOTTOM LINE — Pros & Cons — “Super core” architecture delivers incredible performance
- Apple Studio Display XDR review: great reference picture modes, much improved camera, and 120Hz support on newer Macs, but expensive at $3300 and only for Macs (John Higgins/The Verge)
John Higgins / The Verge : Apple Studio Display XDR review: great reference picture modes, much improved camera, and 120Hz support on newer Macs, but expensive at $3300 and only for Macs — It's been almost exactly four years since Apple released the 5K Studio Display that so many wanted, even if it didn't really deliver as a high-end display.
- SoftBank's stock is down ~48% since Nov. 3, as scrutiny into the scale of its OpenAI involvement grows; on Monday, SoftBank fell 9.8% on Stargate delay reports (Financial Times)
Financial Times : SoftBank's stock is down ~48% since Nov. 3, as scrutiny into the scale of its OpenAI involvement grows; on Monday, SoftBank fell 9.8% on Stargate delay reports — Japanese group has suffered from recent share falls and a negative outlook from rating agency S&P
- Anthropic debuts a Code Review feature for Claude Code, which uses agents working in teams to check pull requests for bugs, available in research preview (David Gewirtz/ZDNET)
David Gewirtz / ZDNET : Anthropic debuts a Code Review feature for Claude Code, which uses agents working in teams to check pull requests for bugs, available in research preview — ZDNET's key takeaways — Anthropic launches AI agents to review developer pull requests. — Internal tests tripled meaningful code review feedback.
- Jay Graber is stepping down as Bluesky CEO, saying a "seasoned operator focused on scaling and execution" is needed; VC Toni Schneider is named interim CEO (Kate Knibbs/Wired)
Kate Knibbs / Wired : Jay Graber is stepping down as Bluesky CEO, saying a “seasoned operator focused on scaling and execution” is needed; VC Toni Schneider is named interim CEO — Venture capitalist Toni Schneider will take over as interim CEO as Bluesky's board of directors hunts for a permanent replacement.
- Sources: Apple delayed the release of its smart home display, planned for this month, until later this year to let the company finish work on the new Siri (Mark Gurman/Bloomberg)
Mark Gurman / Bloomberg : Sources: Apple delayed the release of its smart home display, planned for this month, until later this year to let the company finish work on the new Siri — Apple Inc.'s artificial intelligence struggles are rippling through its product plans, forcing the company to delay a long …
- NBC News poll of 1,000 registered voters: just 26% had a positive view of AI, while 46% had a negative view, the third worst net negative score of all topics (Jake Angelo/Fortune)
Jake Angelo / Fortune : NBC News poll of 1,000 registered voters: just 26% had a positive view of AI, while 46% had a negative view, the third worst net negative score of all topics — Artificial intelligence has permeated nearly every part of daily American life. It's being adopted across the professional sphere …
- UK's AI drive, touted since 2024 and featuring splashy NScale and CoreWeave deals, to build datacenters is riddled with phantom investments and shaky accounting (Aisha Down/The Guardian)
Aisha Down / The Guardian : UK's AI drive, touted since 2024 and featuring splashy NScale and CoreWeave deals, to build datacenters is riddled with phantom investments and shaky accounting — Exclusive: Rented datacentres and ‘supercomputer’ site that's still a scaffolding yard raise questions for Starmer's push to ‘mainline AI into veins of economy’
- OpenAI agrees to acquire Promptfoo, which fixes security issues in AI systems being built and is "trusted by 25%+ of Fortune 500", to fold into OpenAI Frontier (OpenAI)
OpenAI : OpenAI agrees to acquire Promptfoo, which fixes security issues in AI systems being built and is “trusted by 25%+ of Fortune 500”, to fold into OpenAI Frontier — Accelerating agentic security testing and evaluation capabilities in OpenAI Frontier
- Source: UK cyber security company Darktrace names Ed Jennings, who previously led work management platform Quickbase, as CEO, its third chief in 18 months (Kieran Smith/Financial Times)
Kieran Smith / Financial Times : Source: UK cyber security company Darktrace names Ed Jennings, who previously led work management platform Quickbase, as CEO, its third chief in 18 months — UK-based cyber security group hires Ed Jennings as private equity owners push for faster growth — Darktrace has appointed …
- Qualcomm unveils the Arduino Ventuno Q, a single-board computer for AI and robotics applications, powered by Dragonwing IQ8 processor and 16GB of RAM (Steve Dent/Engadget)
Steve Dent / Engadget : Qualcomm unveils the Arduino Ventuno Q, a single-board computer for AI and robotics applications, powered by Dragonwing IQ8 processor and 16GB of RAM — It marries a Qualcomm processor with a microcontroller and comes with 16GB of RAM. — Qualcomm, which purchased microcontroller board …
- Australia's online age restrictions take effect, requiring platforms to verify users are 18+ before they can access content including porn and R-rated games (Ahmed Yussuf/ABC)
Ahmed Yussuf / ABC : Australia's online age restrictions take effect, requiring platforms to verify users are 18+ before they can access content including porn and R-rated games — In short: — New online safety codes requiring R-rated games and websites to verify users' ages are now in effect.
Solidot(13)
- 一百万颗卫星会如何影响天空?
SpaceX 计划向太空发射 1 百万颗卫星,理由是建造太空数据中心。暂不讨论太空数据中心的可行性(没什么可行性),这一百万颗卫星会如何影响我们每一个人?SpaceX 已经向 FCC 递交了发射提议,对该计划的公众意见征集于上周五结束,逾千条公众意见绝大多数持反对立场,要求 FCC 停止推进该计划。SpaceX 已经向地球轨道发射了上万颗卫星,一百万颗则是已有数量的一百倍。SpaceX 平均每周发射两次,它的卫星不断上天,也不断坠落。越来越多的证据表明,火箭发射会将污染物排放到空气中,影响大气层,造成潜在的温室效应,可能加剧对臭氧层的威胁。如果 SpaceX 的一百万颗卫星全部脱离轨道,意味着每三分钟就有一颗卫星重返大气层。轨道上的卫星越多,发生碰撞的可能性也越大。卫星也会影响天文观测,该公司一直试图与国际天文学联合会合作减少卫星对天文观测的影响,但一百万颗完全不同的量级,天文学家震惊不已。
- FBI 通过 Proton Mail 识别抗议者身份
FBI 通过瑞士邮件服务商 Proton Mail 提供的信息识别了亚特兰大抗议组织 Defend the Atlanta Forest/Stop Cop City 领导人的身份。Proton Mail 坚称它必须遵守瑞士的法律。Stop Cop City 官方 FB 账号使用的邮箱是 defendtheatlantaforest@protonmail.com,FBI 援引《司法互助条约(Mutual Legal Assistance Treaty 或 MLAT)》,请求瑞士司法部向 Proton Mail 索取信息。瑞士与逾 30 个国家签订了 MLAT。Proton 向瑞士司法机构提供了信息,然后由瑞士转交给了 FBI。Proton AG 通讯主管 Edward Shone 表示,该公司没有直接向 FBI 提供信息,相关信息是 FBI 通过瑞士司法部获取到的。
- 卫星揭示北美和非洲桥梁面临稳定性风险
根据发表在《Nature Communication》期刊上的一项研究,休斯顿大学等机构的研究人员利用卫星分析了全世界 744 座桥梁,评估其状况。研究结果显示,北美桥梁的状况普遍最差,其次是非洲桥梁。研究分析的很多桥梁已接近其设计使用寿命的上限。北美桥梁的建设高峰是在 1960 年代,很多已建成数十年,接近或超过其最初的设计寿命。研究人员利用名为多时相干涉合成孔径雷达(Multi-temporal InSAR, MTInSAR)的卫星遥感方法去监测桥梁结构中的微小位移。
- 耳鸣与睡眠密切相关
幻影知觉(phantom percept)是大脑愚弄我们以为看到、听到、感觉到或闻到了实际上不存在的东西。耳鸣是最常见的幻影知觉,尽管有很多假说,但至今未找到确切的病因或疗法,全世界有 15% 的人口受到耳鸣的困扰。很多耳鸣患者都表示睡眠质量差,睡眠模式紊乱,但耳鸣与睡眠这一重要生理功能之间的潜在关联直到最近才被人所认识。牛津大学的神经学家提出,深度睡眠或非快速眼动睡眠(non-REM)期间出现的大幅度自发性脑电波可能抑制导致耳鸣的脑电活动。雪貂实验显示,耳鸣雪貂在进入 non-REM 睡眠后过度活跃的脑电活动会减弱。研究结果表明,深度睡眠可能有助于缓解耳鸣,有可能揭示了大脑调节异常活动的自然机制。
- 研究人员模拟月球土壤种植收获鹰嘴豆
根据发表在《Scientific Reports》期刊上的一项研究,德州农工的研究人员模拟月球土壤种植收获了鹰嘴豆,但鹰嘴豆的食用安全性尚未确定。月球土壤贫瘠,缺乏营养元素,富含重金属,为克服这些问题,研究人员利用蚯蚓通过废弃物产生的堆肥去提供必需的微生物和营养物质,使用共生真菌 Arbuscular Mycorrhizal Fungi 去促进植物生长和减少对有毒重金属的吸收。结果显示添加堆肥和共生真菌的模拟月壤混合物能像普通地球土壤那样种植和收获鹰嘴豆。研究人员接下来将分析鹰嘴豆的营养成分,检测是否含有重金属,确保人类的食用安全。
- 新法律要求成人网站验证澳大利亚人年龄
在禁止 16 岁以下儿童使用社交媒体三个月后,澳大利亚新法律要求有成人内容的网站验证该国访客的年龄,确保其年满 18 岁,违规者将面临罚款。澳大利亚网络安全监管机构表示,此举旨在保护儿童免受有害内容的侵害。eSafety 专员 Julie Inman Grant 表示,我们不允许儿童进入酒吧、酒专卖店、成人用品店或赌场,但对于网络空间,却没有此类保障措施。平台可能需要使用面部识别技术、数字身份和信用卡信息去验证访客。根据新规,搜索引擎、应用商店、社交媒体和游戏平台、色情网站以及包括聊天机器人在内的 AI 系统必须采取切实有效的措施,防止儿童接触成人内容。
- 印尼和印度卡纳塔克邦将禁止 16 岁以下儿童使用社媒
继澳大利亚之后,印度科技重镇卡纳塔克邦(Karnataka)与印度尼西亚相继宣布将禁止 16 岁以下青少年使用社交媒体及高风险数字平台。印尼通信与数码部长梅蒂雅(Meutya Hafid)星期五发声明说,政府将从 3 月 28 日起,分阶段注销 16 岁以下青少年在“高风险平台”上的账户。首批受影响的平台包括 YouTube、TikTok、Facebook、Instagram、Threads、X、Bigo Live 以及游戏平台 Roblox。梅蒂雅说,颁布禁令是因为青少年面临网络色情、网络欺凌、网络诈骗和网络成瘾的威胁。印度卡纳塔克邦立法议员星期五在邦预算会议上提出禁止16岁以下青少年使用社媒应用的法案。卡纳塔克邦将成为全印度首个实行这项禁令的邦。议员里兹万说:“青少年在未了解后果的情况下,就开始使用社交媒体。我们将与社会人士探讨,如何在社媒落实年龄限制。”
- NASA DART 探测器确认改变了小行星的轨道
NASA 执行双小行星重定向测试(DART)任务的探测器于 2022 年 9 月撞击了小行星 Dimorphos,这是世界首次行星防御技术演示。撞击不仅改变了 Dimorphos 绕较大伴星 Didymos 的运动,也同时使这对双小行星绕太阳的轨道出现可测量的改变。观测结果显示,原本约 770 天的公转周期缩短了 0.15 秒,这是人类制造的物体首次被量测到改变天体绕太阳运行的轨道。先前研究已发现,Dimorphos 绕直径约 805 米的 Didymos 公转周期(约 12 小时)因撞击缩短了 33 分钟。新的研究指出,撞击喷出的碎屑使整个系统的轨道速度改变约每秒 11.7 微米,导致其绕太阳的公转周期改变 0.15 秒。虽然这对轨道来说只是极其微小的变化,但随着时间累积,仍可能造成显著偏移,甚至影响一颗潜在危险小行星是否会撞上地球。
- 在卫星照片披露美国军事基地损失之后 Planet Labs 停止发布卫星照片
过去几天 Planet Labs 公司的卫星图像展示了美国在中东军事基地的战损情况,包括移动雷达 THAAD 遭到攻击的画面。Planet 周五宣布停止发布部分地区的卫星照片。Planet 运营着数百颗地球成像卫星,能每天对地球上的每一块陆地进行一次观测。其客户包括智库、非政府组织、学术机构、新闻媒体以及农业、林业和能源等行业的商业用户。它还与美国军方和情报机构签署了出售卫星照片的合同。Planet 宣布对特定地区的卫星照片强制推迟 96 小时发布。
- AI 翻译工具会将“幻觉”加入到维基百科文章
维基百科编辑实施了新政策,限制使用 AI 翻译工具将英文条目翻译到其它语言的贡献者,原因是他们发现 AI 工具会在译文中加入“幻觉”——即原文不存在的内容。问题与非营利组织 Open Knowledge Association (OKA)有关,该组织主要依赖来自“全球南方(Global South)”的廉价劳工充当翻译合同工,将英文维基百科文章翻译到其他语言。部分译者开始使用 Google Gemini 和 ChatGPT 等工具加快翻译,但编辑在审阅译文时发现了大量错误,包括事实错误、缺少引用以及引用不相关来源。
- 苹果禁止美国用户下载字节跳动的其它应用
拥有中国 App Store 账户的美国 iPhone 用户报告他们无法再下载或更新字节跳动的其它应用。用户会看到警告,“此应用在您所在的国家或地区不可用”。苹果是在遵守 2024 年美国国会通过的 TikTok 法案——《The Protecting Americans from Foreign Adversary Controlled Applications Act》,法案主要针对 TikTok,但也涵盖了字节跳动的其它应用如 TikTok 中国版抖音、AI 智能助手豆包以及阅读平台番茄小说。
- 小行星 2024 YR4 不会撞击月球
去年引发广泛关注的小行星 2024 YR4 一度成为近 20 年发现的最危险的小行星,它撞击地球的可能性已经排除,但有 4% 的概率会在 2032 年 12 月 22 日撞击月球。根据天文学家利用韦伯太空望远镜近红外相机 NIRCam 对其的最新观测,它也不会撞击月球了。2024 YR4 会安全从距离月球 2 万多公里处掠过。
- OpenWrt 25.12.0 释出
面向路由器等嵌入式设备的发行版 OpenWrt 释出了 v25.12.0。OpenWrt 25.12.0 以 2025 年 4 月 1 日去世的 Dave Täht 名字命名,他是 Bufferbloat 项目联合创始人,致力于降低网络延迟,他的工作让无数人的网速更快更稳定可靠。新版本的主要变化包括:包管理器从 opkg 切换到 apk,opkg 不再维护;attended.sysupgrade LuCI 大幅简化了升级流程;保留 Shell 历史;支持大量新设备,总数超过 2200 种,等等。