OrangeBot.AI Digest — 2026-04-09
85 headlines across 8 sources, aggregated for this day.
Hacker News(15)
- Maine is about to become the first state to ban major new data centers (www.gadgetreview.com)
- Native Instant Space Switching on macOS (arhan.sh)
- ChatGPT Pro now starts at $100/month (chatgpt.com)
- EFF is leaving X (www.eff.org)
- The Pentagon Threatened Pope Leo XIV's Ambassador with the Avignon Papacy (www.thelettersfromleo.com)
- The Vercel plugin on Claude Code wants to read your prompts (akshaychugh.xyz)
- Meta removes ads for social media addiction litigation (www.axios.com)
- Am I German or Autistic? (german.millermanschool.com)
- How Pizza Tycoon simulated traffic on a 25 MHz CPU (pizzalegacy.nl)
- Top laptops to use with FreeBSD (freebsdfoundation.github.io)
- Introduction to Nintendo DS Programming (www.patater.com)
- Wit, unker, Git: The lost medieval pronouns of English intimacy (www.bbc.com)
- Reallocating $100/Month Claude Code Spend to Zed and OpenRouter (braw.dev)
- Claude mixes up who said what (dwyer.co.za)
- Help Keep Thunderbird Alive (updates.thunderbird.net)
GitHub Trending(10)
Product Hunt(15)
- Brila
One-page websites from real Google Maps reviews
- GAIA
Proactive personal assistant that handles your day
- Fonic
Turn messy work into interactive, actionable reports
- Bouncer
Filter (and heal) your Twitter feed
- Cyris
Turns every AI decision into audit-ready evidence
- Offsite
Build teams of humans and agents, watch them work.
- Grass
Gives your coding agent a dedicated VM that's ready 24/7
- Plow
Openclaw on your Mac, with permissions you can understand
- Claude Managed Agents
Pre-built agent harness on managed infrastructure
- OpenResource
Discover open-source tools with an AI chat assistant
- ProdShort
Turn meetings into ready-to-post shorts and posts
- AgentMail
Email Inboxes for AI Agents
- Startups.RIP
Rebuild 1,738+ dead YC startups with AI
- Lukan AI Agent, IDE and workstation.
The open-source AI workstation for coding, ops, and life
- ScreenSmooth
Beautiful Screen Recordings in minutes
Hugging Face(15)
- Think in Strokes, Not Pixels: Process-Driven Image Generation via Interleaved Reasoning
Humans paint images incrementally: they plan a global layout, sketch a coarse draft, inspect, and refine details, and most importantly, each step is grounded in the evolving visual states. However, can unified multimodal models trained on text-image interleaved datasets also imagine the chain of intermediate states? In this paper, we introduce process-driven image generation, a multi-step paradigm that decomposes synthesis into an interleaved reasoning trajectory of thoughts and actions. Rather than generating images in a single step, our approach unfolds across multiple iterations, each consisting of 4 stages: textual planning, visual drafting, textual reflection, and visual refinement. The textual reasoning explicitly conditions how the visual state should evolve, while the generated visual intermediate in turn constrains and grounds the next round of textual reasoning. A core challenge of process-driven generation stems from the ambiguity of intermediate states: how can models evaluate each partially-complete image? We address this through dense, step-wise supervision that maintains two complementary constraints: for the visual intermediate states, we enforce the spatial and semantic consistency; for the textual intermediate states, we preserve the prior visual knowledge while enabling the model to identify and correct prompt-violating elements. This makes the generation process explicit, interpretable, and directly supervisable. To validate proposed method, we conduct experiments under various text-to-image generation benchmarks.
- RAGEN-2: Reasoning Collapse in Agentic RL
RL training of multi-turn LLM agents is inherently unstable, and reasoning quality directly determines task performance. Entropy is widely used to track reasoning stability. However, entropy only measures diversity within the same input, and cannot tell whether reasoning actually responds to different inputs. In RAGEN-2, we find that even with stable entropy, models can rely on fixed templates that look diverse but are input-agnostic. We call this template collapse, a failure mode invisible to entropy and all existing metrics. To diagnose this failure, we decompose reasoning quality into within-input diversity (Entropy) and cross-input distinguishability (Mutual Information, MI), and introduce a family of mutual information proxies for online diagnosis. Across diverse tasks, mutual information correlates with final performance much more strongly than entropy, making it a more reliable proxy for reasoning quality. We further explain template collapse with a signal-to-noise ratio (SNR) mechanism. Low reward variance weakens task gradients, letting regularization terms dominate and erase cross-input reasoning differences. To address this, we propose SNR-Aware Filtering to select high-signal prompts per iteration using reward variance as a lightweight proxy. Across planning, math reasoning, web navigation, and code execution, the method consistently improves both input dependence and task performance.
- MARS: Enabling Autoregressive Models Multi-Token Generation
Autoregressive (AR) language models generate text one token at a time, even when consecutive tokens are highly predictable given earlier context. We introduce MARS (Mask AutoRegreSsion), a lightweight fine-tuning method that teaches an instruction-tuned AR model to predict multiple tokens per forward pass. MARS adds no architectural modifications, no extra parameters, and produces a single model that can still be called exactly like the original AR model with no performance degradation. Unlike speculative decoding, which maintains a separate draft model alongside the target, or multi-head approaches such as Medusa, which attach additional prediction heads, MARS requires only continued training on existing instruction data. When generating one token per forward pass, MARS matches or exceeds the AR baseline on six standard benchmarks. When allowed to accept multiple tokens per step, it maintains baseline-level accuracy while achieving 1.5-1.7x throughput. We further develop a block-level KV caching strategy for batch inference, achieving up to 1.71x wall-clock speedup over AR with KV cache on Qwen2.5-7B. Finally, MARS supports real-time speed adjustment via confidence thresholding: under high request load, the serving system can increase throughput on the fly without swapping models or restarting, providing a practical latency-quality knob for deployment.
- Combee: Scaling Prompt Learning for Self-Improving Language Model Agents
Recent advances in prompt learning allow large language model agents to acquire task-relevant knowledge from inference-time context without parameter changes. For example, existing methods (like ACE or GEPA) can learn system prompts to improve accuracy based on previous agent runs. However, these methods primarily focus on single-agent or low-parallelism settings. This fundamentally limits their ability to efficiently learn from a large set of collected agentic traces. It would be efficient and beneficial to run prompt learning in parallel to accommodate the growing trend of learning from many agentic traces or parallel agent executions. Yet without a principled strategy for scaling, current methods suffer from quality degradation with high parallelism. To improve both the efficiency and quality of prompt learning, we propose Combee, a novel framework to scale parallel prompt learning for self-improving agents. Combee speeds up learning and enables running many agents in parallel while learning from their aggregate traces without quality degradation. To achieve this, Combee leverages parallel scans and employs an augmented shuffle mechanism; Combee also introduces a dynamic batch size controller to balance quality and delay. Evaluations on AppWorld, Terminal-Bench, Formula, and FiNER demonstrate that Combee achieves up to 17x speedup over previous methods with comparable or better accuracy and equivalent cost.
- SEVerA: Verified Synthesis of Self-Evolving Agents
Recent advances have shown the effectiveness of self-evolving LLM agents on tasks such as program repair and scientific discovery. In this paradigm, a planner LLM synthesizes an agent program that invokes parametric models, including LLMs, which are then tuned per task to improve performance. However, existing self-evolving agent frameworks provide no formal guarantees of safety or correctness. Because such programs are often executed autonomously on unseen inputs, this lack of guarantees raises reliability and security concerns. We formulate agentic code generation as a constrained learning problem, combining hard formal specifications with soft objectives capturing task utility. We introduce Formally Guarded Generative Models (FGGM), which allow the planner LLM to specify a formal output contract for each generative model call using first-order logic. Each FGGM call wraps the underlying model in a rejection sampler with a verified fallback, ensuring every returned output satisfies the contract for any input and parameter setting. Building on FGGM, we present SEVerA (Self-Evolving Verified Agents), a three-stage framework: Search synthesizes candidate parametric programs containing FGGM calls; Verification proves correctness with respect to hard constraints for all parameter values, reducing the problem to unconstrained learning; and Learning applies scalable gradient-based optimization, including GRPO-style fine-tuning, to improve the soft objective while preserving correctness. We evaluate SEVerA on Dafny program verification, symbolic math synthesis, and policy-compliant agentic tool use (τ^2-bench). Across tasks, SEVerA achieves zero constraint violations while improving performance over unconstrained and SOTA baselines, showing that formal behavioral constraints not only guarantee correctness but also steer synthesis toward higher-quality agents.
- INSPATIO-WORLD: A Real-Time 4D World Simulator via Spatiotemporal Autoregressive Modeling
Building world models with spatial consistency and real-time interactivity remains a fundamental challenge in computer vision. Current video generation paradigms often struggle with a lack of spatial persistence and insufficient visual realism, making it difficult to support seamless navigation in complex environments. To address these challenges, we propose INSPATIO-WORLD, a novel real-time framework capable of recovering and generating high-fidelity, dynamic interactive scenes from a single reference video. At the core of our approach is a Spatiotemporal Autoregressive (STAR) architecture, which enables consistent and controllable scene evolution through two tightly coupled components: Implicit Spatiotemporal Cache aggregates reference and historical observations into a latent world representation, ensuring global consistency during long-horizon navigation; Explicit Spatial Constraint Module enforces geometric structure and translates user interactions into precise and physically plausible camera trajectories. Furthermore, we introduce Joint Distribution Matching Distillation (JDMD). By using real-world data distributions as a regularizing guide, JDMD effectively overcomes the fidelity degradation typically caused by over-reliance on synthetic data. Extensive experiments demonstrate that INSPATIO-WORLD significantly outperforms existing state-of-the-art (SOTA) models in spatial consistency and interaction precision, ranking first among real-time interactive methods on the WorldScore-Dynamic benchmark, and establishing a practical pipeline for navigating 4D environments reconstructed from monocular videos.
- Neural Computers
We propose a new frontier: Neural Computers (NCs) -- an emerging machine form that unifies computation, memory, and I/O in a learned runtime state. Unlike conventional computers, which execute explicit programs, agents, which act over external execution environments, and world models, which learn environment dynamics, NCs aim to make the model itself the running computer. Our long-term goal is the Completely Neural Computer (CNC): the mature, general-purpose realization of this emerging machine form, with stable execution, explicit reprogramming, and durable capability reuse. As an initial step, we study whether early NC primitives can be learned solely from collected I/O traces, without instrumented program state. Concretely, we instantiate NCs as video models that roll out screen frames from instructions, pixels, and user actions (when available) in CLI and GUI settings. These implementations show that learned runtimes can acquire early interface primitives, especially I/O alignment and short-horizon control, while routine reuse, controlled updates, and symbolic stability remain open. We outline a roadmap toward CNCs around these challenges. If overcome, CNCs could establish a new computing paradigm beyond today's agents, world models, and conventional computers.
- TC-AE: Unlocking Token Capacity for Deep Compression Autoencoders
We propose TC-AE, a ViT-based architecture for deep compression autoencoders. Existing methods commonly increase the channel number of latent representations to maintain reconstruction quality under high compression ratios. However, this strategy often leads to latent representation collapse, which degrades generative performance. Instead of relying on increasingly complex architectures or multi-stage training schemes, TC-AE addresses this challenge from the perspective of the token space, the key bridge between pixels and image latents, through two complementary innovations: Firstly, we study token number scaling by adjusting the patch size in ViT under a fixed latent budget, and identify aggressive token-to-latent compression as the key factor that limits effective scaling. To address this issue, we decompose token-to-latent compression into two stages, reducing structural information loss and enabling effective token number scaling for generation. Secondly, to further mitigate latent representation collapse, we enhance the semantic structure of image tokens via joint self-supervised training, leading to more generative-friendly latents. With these designs, TC-AE achieves substantially improved reconstruction and generative performance under deep compression. We hope our research will advance ViT-based tokenizer for visual generation.
- FP4 Explore, BF16 Train: Diffusion Reinforcement Learning via Efficient Rollout Scaling
Reinforcement-Learning-based post-training has recently emerged as a promising paradigm for aligning text-to-image diffusion models with human preferences. In recent studies, increasing the rollout group size yields pronounced performance improvements, indicating substantial room for further alignment gains. However, scaling rollouts on large-scale foundational diffusion models (e.g., FLUX.1-12B) imposes a heavy computational burden. To alleviate this bottleneck, we explore the integration of FP4 quantization into Diffusion RL rollouts. Yet, we identify that naive quantized pipelines inherently introduce risks of performance degradation. To overcome this dilemma between efficiency and training integrity, we propose Sol-RL (Speed-of-light RL), a novel FP4-empowered Two-stage Reinforcement Learning framework. First, we utilize high-throughput NVFP4 rollouts to generate a massive candidate pool and extract a highly contrastive subset. Second, we regenerate these selected samples in BF16 precision and optimize the policy exclusively on them. By decoupling candidate exploration from policy optimization, Sol-RL integrates the algorithmic mechanisms of rollout scaling with the system-level throughput gains of NVFP4. This synergistic algorithm-hardware design effectively accelerates the rollout phase while reserving high-fidelity samples for optimization. We empirically demonstrate that our framework maintains the training integrity of BF16 precision pipeline while fully exploiting the throughput gains enabled by FP4 arithmetic. Extensive experiments across SANA, FLUX.1, and SD3.5-L substantiate that our approach delivers superior alignment performance across multiple metrics while accelerating training convergence by up to 4.64times, unlocking the power of massive rollout scaling at a fraction of the cost.
- Graph-Based Chain-of-Thought Pruning for Reducing Redundant Reflections in Reasoning LLMs
Extending CoT through RL has been widely used to enhance the reasoning capabilities of LLMs. However, due to the sparsity of reward signals, it can also induce undesirable thinking patterns such as overthinking, i.e., generating redundant intermediate reasoning content. In this work, we argue that a major source of such redundancy is inefficient reflection, which often manifests in two problematic patterns: Indiscriminate Reflection, where the model performs broad, low-impact checks throughout reasoning, and Repetitive Reflection, where it repeatedly re-verifies an already established conclusion. To address this, we introduce a graph-based CoT optimization framework. Specifically, we convert each linear CoT into a directed acyclic graph (DAG) with explicit dependency edges, and design a dual pruning strategy: branch-level pruning removes weakly contributing reflection branches, while depth-level pruning eliminates late-stage re-verification. We distill this behavior via a three-stage pipeline: (1) SFT to initialize the policy on pruned concise traces, (2) DPO to prefer correct but less redundant trajectories, and (3) GRPO with length penalty to jointly optimize answer correctness and efficiency. Experiments show that our approach reduces the average reasoning tokens by 42\% while maintaining or improving accuracy.
- FlowInOne:Unifying Multimodal Generation as Image-in, Image-out Flow Matching
Multimodal generation has long been dominated by text-driven pipelines where language dictates vision but cannot reason or create within it. We challenge this paradigm by asking whether all modalities, including textual descriptions, spatial layouts, and editing instructions, can be unified into a single visual representation. We present FlowInOne, a framework that reformulates multimodal generation as a purely visual flow, converting all inputs into visual prompts and enabling a clean image-in, image-out pipeline governed by a single flow matching model. This vision-centric formulation naturally eliminates cross-modal alignment bottlenecks, noise scheduling, and task-specific architectural branches, unifying text-to-image generation, layout-guided editing, and visual instruction following under one coherent paradigm. To support this, we introduce VisPrompt-5M, a large-scale dataset of 5 million visual prompt pairs spanning diverse tasks including physics-aware force dynamics and trajectory prediction, alongside VP-Bench, a rigorously curated benchmark assessing instruction faithfulness, spatial precision, visual realism, and content consistency. Extensive experiments demonstrate that FlowInOne achieves state-of-the-art performance across all unified generation tasks, surpassing both open-source models and competitive commercial systems, establishing a new foundation for fully vision-centric generative modeling where perception and creation coexist within a single continuous visual space.
- Beyond Hard Negatives: The Importance of Score Distribution in Knowledge Distillation for Dense Retrieval
Transferring knowledge from a cross-encoder teacher via Knowledge Distillation (KD) has become a standard paradigm for training retrieval models. While existing studies have largely focused on mining hard negatives to improve discrimination, the systematic composition of training data and the resulting teacher score distribution have received relatively less attention. In this work, we highlight that focusing solely on hard negatives prevents the student from learning the comprehensive preference structure of the teacher, potentially hampering generalization. To effectively emulate the teacher score distribution, we propose a Stratified Sampling strategy that uniformly covers the entire score spectrum. Experiments on in-domain and out-of-domain benchmarks confirm that Stratified Sampling, which preserves the variance and entropy of teacher scores, serves as a robust baseline, significantly outperforming top-K and random sampling in diverse settings. These findings suggest that the essence of distillation lies in preserving the diverse range of relative scores perceived by the teacher.
- Fast Spatial Memory with Elastic Test-Time Training
Large Chunk Test-Time Training (LaCT) has shown strong performance on long-context 3D reconstruction, but its fully plastic inference-time updates remain vulnerable to catastrophic forgetting and overfitting. As a result, LaCT is typically instantiated with a single large chunk spanning the full input sequence, falling short of the broader goal of handling arbitrarily long sequences in a single pass. We propose Elastic Test-Time Training inspired by elastic weight consolidation, that stabilizes LaCT fast-weight updates with a Fisher-weighted elastic prior around a maintained anchor state. The anchor evolves as an exponential moving average of past fast weights to balance stability and plasticity. Based on this updated architecture, we introduce Fast Spatial Memory (FSM), an efficient and scalable model for 4D reconstruction that learns spatiotemporal representations from long observation sequences and renders novel view-time combinations. We pre-trained FSM on large-scale curated 3D/4D data to capture the dynamics and semantics of complex spatial environments. Extensive experiments show that FSM supports fast adaptation over long sequences and delivers high-quality 3D/4D reconstruction with smaller chunks and mitigating the camera-interpolation shortcut. Overall, we hope to advance LaCT beyond the bounded single-chunk setting toward robust multi-chunk adaptation, a necessary step for generalization to genuinely longer sequences, while substantially alleviating the activation-memory bottleneck.
- The Depth Ceiling: On the Limits of Large Language Models in Discovering Latent Planning
The viability of chain-of-thought (CoT) monitoring hinges on models being unable to reason effectively in their latent representations. Yet little is known about the limits of such latent reasoning in LLMs. We test these limits by studying whether models can discover multi-step planning strategies without supervision on intermediate steps and execute them latently, within a single forward pass. Using graph path-finding tasks that precisely control the number of required latent planning steps, we uncover a striking limitation unresolved by massive scaling: tiny transformers trained from scratch discover strategies requiring up to three latent steps, fine-tuned GPT-4o and Qwen3-32B reach five, and GPT-5.4 attains seven under few-shot prompting. Although the maximum latent planning depth models can learn during training is five, the discovered strategy generalizes up to eight latent steps at test-time. This reveals a dissociation between the ability to discover a latent strategy under final-answer supervision alone and the ability to execute it once discovered. If similar limits hold more broadly, strategies requiring multiple coordinated latent planning steps may need to be explicitly taught or externalized, lending credence to CoT monitoring.
- DeonticBench: A Benchmark for Reasoning over Rules
Reasoning with complex, context-specific rules remains challenging for large language models (LLMs). In legal and policy settings, this manifests as deontic reasoning: reasoning about obligations, permissions, and prohibitions under explicit rules. While many recent benchmarks emphasize short-context mathematical reasoning, fewer focus on long-context, high-stakes deontic reasoning. To address this gap, we introduce DEONTICBENCH, a benchmark of 6,232 tasks across U.S. federal taxes, airline baggage policies, U.S. immigration administration, and U.S. state housing law. These tasks can be approached in multiple ways, including direct reasoning in language or with the aid of symbolic computation. Besides free-form chain-of-thought reasoning, DEONTICBENCH enables an optional solver-based workflow in which models translate statutes and case facts into executable Prolog, leading to formal problem interpretations and an explicit program trace. We release reference Prolog programs for all instances. Across frontier LLMs and coding models, best hard-subset performance reaches only 44.4% on SARA Numeric and 46.6 macro-F1 on Housing. We further study training with supervised fine-tuning and reinforcement learning for symbolic program generation. Although training improves Prolog generation quality, current RL methods still fail to solve these tasks reliably. Overall, DEONTICBENCH provides a benchmark for studying context-grounded rule reasoning in real-world domains under both symbolic and non-symbolic settings.
Techmeme(15)
- xAI has filed a lawsuit challenging Colorado's landmark AI anti-discrimination law, set to take effect in the summer, saying it violates free speech protections (Financial Times)
Financial Times : xAI has filed a lawsuit challenging Colorado's landmark AI anti-discrimination law, set to take effect in the summer, saying it violates free speech protections — Elon Musk's AI lab claims the regulations violate free speech protections — Elon Musk's xAI has filed a lawsuit challenging …
- A gap in understanding AI is growing, as casual users cite flaws in old free models while power users point to new models' staggering gains in technical domains (Andrej Karpathy/@karpathy)
Andrej Karpathy / @karpathy : A gap in understanding AI is growing, as casual users cite flaws in old free models while power users point to new models' staggering gains in technical domains — Judging by my tl there is a growing gap in understanding of AI capability. The first issue I think is around recency and tier of use. I think a lot of people tried the free tier of ChatGPT somewhere last year and allowed it to inform their views on AI a little too much. This is
- An OpenAI note to investors after Anthropic announced Mythos says OpenAI's early push to increase computing resources gives it a key advantage over Anthropic (Shirin Ghaffary/Bloomberg)
Shirin Ghaffary / Bloomberg : An OpenAI note to investors after Anthropic announced Mythos says OpenAI's early push to increase computing resources gives it a key advantage over Anthropic — OpenAI told investors this week that its early push to dramatically increase computing resources gives it a key advantage …
- Internal memo: Meta is pulling top engineers into its new Applied AI Engineering division, as part of a push to improve its models and "compete in the AI race" (Jyoti Mann/The Information)
Jyoti Mann / The Information : Internal memo: Meta is pulling top engineers into its new Applied AI Engineering division, as part of a push to improve its models and “compete in the AI race” — Meta Platforms is pulling top engineers from across the company into its new Applied AI Engineering division …
- An Ohio man is the first convicted under the Take It Down Act after pleading guilty to creating and sharing real and AI-generated explicit images of 10+ victims (Ashley Belanger/Ars Technica)
Ashley Belanger / Ars Technica : An Ohio man is the first convicted under the Take It Down Act after pleading guilty to creating and sharing real and AI-generated explicit images of 10+ victims — An Ohio man became the first person convicted under the Take It Down Act after pleading guilty to creating and sharing both real …
- EFF says it is leaving X, as "X is no longer where the fight is happening" and an X post now gets less than 3% of the views a tweet got seven years ago (Kenyatta Thomas/Electronic Frontier ...)
Kenyatta Thomas / Electronic Frontier Foundation : EFF says it is leaving X, as “X is no longer where the fight is happening” and an X post now gets less than 3% of the views a tweet got seven years ago — After almost twenty years on the platform, EFF is logging off of X. This isn't a decision we made lightly, but it might be overdue.
- OpenAI launches a $100/month ChatGPT Pro subscription, which offers 5x more Codex usage than Plus; the $200/month Pro plan offers 20x higher limits than Plus (Zac Hall/9to5Mac)
Zac Hall / 9to5Mac : OpenAI launches a $100/month ChatGPT Pro subscription, which offers 5x more Codex usage than Plus; the $200/month Pro plan offers 20x higher limits than Plus — OpenAI is launching a revamped $100/month ChatGPT subscription aimed at Codex users. Here's how it compares to OpenAI's existing plans and what it includes.
- Google says the Gemini app can now generate interactive 3D models and simulations; users must select the Pro model in the prompt bar (Emma Roth/The Verge)
Emma Roth / The Verge : Google says the Gemini app can now generate interactive 3D models and simulations; users must select the Pro model in the prompt bar — You can interact with the 3D models and adjust variables in real-time. … Google's latest upgrade for Gemini will allow the chatbot to generate interactive 3D models …
- Sources: the White House is pushing back on GOP-led AI bills in Nebraska and Tennessee, putting new pressure on GOP state lawmakers who support AI guardrails (Axios)
Axios : Sources: the White House is pushing back on GOP-led AI bills in Nebraska and Tennessee, putting new pressure on GOP state lawmakers who support AI guardrails — The Trump administration is pushing back on Republican-led AI bills in Nebraska and Tennessee, with sources familiar …
- Anthropic makes Claude Cowork, previously available as a "research preview", generally available to all paid plans, and adds six features for enterprise use (Zac Hall/9to5Mac)
Zac Hall / 9to5Mac : Anthropic makes Claude Cowork, previously available as a “research preview”, generally available to all paid plans, and adds six features for enterprise use — Claude Cowork, available on macOS, is losing the “research preview” label today as Anthropic introduces enterprise capabilities.
- Florida's AG launches a probe into OpenAI and ChatGPT, saying its data could fall "into the hands of America's enemies" and citing a mass shooter's ChatGPT use (Juby Babu/Reuters)
Juby Babu / Reuters : Florida's AG launches a probe into OpenAI and ChatGPT, saying its data could fall “into the hands of America's enemies” and citing a mass shooter's ChatGPT use — Florida Attorney General James Uthmeier on Thursday launched an investigation into OpenAI and its chatbot ChatGPT …
- X brings back Voice Notes to X Chat, supporting one-on-one and group conversations via a new voice input icon, as it tests a standalone X Chat spinoff on iOS (Sarah Perez/TechCrunch)
Sarah Perez / TechCrunch : X brings back Voice Notes to X Chat, supporting one-on-one and group conversations via a new voice input icon, as it tests a standalone X Chat spinoff on iOS — Posting Voice Notes publicly on X may no longer be possible, but you can now share audio messages within X's direct messaging system, X Chat, once again.
- Visa unveils Intelligent Commerce Connect, a platform that facilitates payments for AI agents across multiple card networks, including those of Visa competitors (Kelly Tyko/Axios)
Kelly Tyko / Axios : Visa unveils Intelligent Commerce Connect, a platform that facilitates payments for AI agents across multiple card networks, including those of Visa competitors — Visa is rolling out a new platform for a future where AI agents do the shopping for consumers.
- Chapter, which uses AI to help seniors enroll in Medicare, raised a $100M Series E, doubling its valuation to $3B and bringing its total funding to $285M (Stephanie Palazzolo/The Information)
Stephanie Palazzolo / The Information : Chapter, which uses AI to help seniors enroll in Medicare, raised a $100M Series E, doubling its valuation to $3B and bringing its total funding to $285M — Meta Platforms' Wednesday release of Spark, the first model in its broader Muse model family, doesn't blow other frontier models out of the water …
- RISC-V chip designer SiFive raised a $400M Series G led by Atreides at a $3.65B valuation; CEO Patrick Little says it is the final funding round before an IPO (Stephen Nellis/Reuters)
Stephen Nellis / Reuters : RISC-V chip designer SiFive raised a $400M Series G led by Atreides at a $3.65B valuation; CEO Patrick Little says it is the final funding round before an IPO — Silicon Valley startup SiFive said on Thursday it has raised a $400 million round of funding from Atreides Management, Nvidia …
Solidot(15)
- NASA 宇航员的笔记本电脑运行 VLC
NASA 的视频显示,Artemis II 宇航员的笔记本电脑运行了开源多媒体播放器 VLC,而 VLC 使用了 FFmpeg 多媒体库,这意味着两大开源项目都进入了太空,而且工作正常,不像微软的私有软件 Outlook。Artemis 宇航员除了使用开源软件,还用苹果最新的 iPhone 17 Pro Max 智能手机在飞船内自拍。飞船外的拍摄则使用了 2014 年款的 GoPro Hero 4 Black 相机。
- 欧洲男性过去一万年摄入的肉量一直多于女性
人体骨胶原中碳氮同位素比例能揭示其饮食信息,氮同位素比例反映了摄入的肉量,碳同位素比例则可以用于推断一个人摄入了多少小米以及多少可能比较稀有的海产品。研究人员分析了欧洲 673 个遗址的 12281 名成年人过去一万年间男性和女性消费肉类、小米和/或海产品的比例,发现几乎所有时期男性摄入的肉量都多于女性。新石器时代最平等,但也存在获取动物蛋白质上显著的性别差异。研究人员推测,这种肉食消费的不平等现象可能源自食物禁忌、创世信仰、对女性蛋白质需求的误解,或者将男性需求置于女性需求之上的社会规范。
- 英国科学家量化交通对城市温度的贡献
英国科学家使用来自大曼彻斯特交通局(TfGM)的真实交通数据以及开放数据集,量化了交通产生的热量对城市温度的贡献。研究结果显示,在曼彻斯特交通产生的热量在夏季使模拟气温升高了约 0.16°C,冬季升高了约 0.35°C。虽然升温幅度不大,但在极端高温事件中会产生显著影响。研究还发现,交通产生的热量不仅影响室外温度,还会影响室内温度。街道上释放的热量会传递到建筑物内,增加了夏季对空调的需求。
- FBI 称 2025 年美国因网络犯罪损失 210 亿美元
FBI 称 2025 年美国因网络犯罪损失 210 亿美元,比 2024 年的 166 亿美元增长了 26%。主要网络犯罪类型包括:投资诈骗、商业电邮入侵、技术支持欺骗和数据泄露。美国 Internet Crime Complaint Center (IC3)去年收到的投诉最多的是钓鱼攻击(19.1 万)、勒索(8.9 万)和投资诈骗(7.2 万),商业电邮入侵(24,700 )、数据泄露(3,900)、勒索软件攻击(3,600)和 SIM 卡劫持(971 起)。投资诈骗造成了 86 亿美元的损失。针对加密货币的网络犯罪造成了逾 110 亿美元损失。60 岁以上的美国人报告损失 77 亿美元,比上一年增长了 37%。
- LinkedIn 扫描浏览器扩展面临集体诉讼
LinkedIn 与爱沙尼亚软件公司 Teamfluence 之间的法律纠纷暴露了这家职业社交网络扫描浏览器扩展的行为,并因此面临用户的集体诉讼。Teamfluence 开发了一个抓取 LinkedIn 用户数据的扩展,LinkedIn 以违反用户协议为由关闭了其账号。Teamfluence 随后在德国提起诉讼寻求恢复账号,声称 LinkedIn 违反了多项欧盟法律。Teamfluence 还通过 Fairlinked 的名义发布报告 BrowserGate,指控 LinkedIn 扫描用户的浏览器以收集扩展信息,Fairlinked 称 LinkedIn 使用隐蔽的 JavaScript 程序扫描浏览器,检测是否安装了 6,222 种扩展之一,其中包括微软的竞争对手 Salesforce、HubSpot 和 Pipedrive。因为扫描的扩展还包含 “伊斯兰内容过滤器”、“反犹太复国主义政治标签”等,LinkedIn 被控收集用户的政治观点或宗教立场,根据欧盟的法律收集此类信息需要获得用户的明确同意。LinkedIn 没有否认它扫描扩展的行为,但强调它扫描扩展是为了识别哪些扩展违反了其条款。
- 免费领取价值30/90美金的NVIDIA DLI自学课程并测试获得证书
领取规则:未注册过开发者的用户可以通过如下链接免费选择一门 DLI 在线自主培训的付费课程,配套云端实验环境和可获得 NVIDIA 培训证书。每位用户(每个邮箱账号)仅可选择一门。 目前可选课程包括 7 门英文课,5 门中文课,目前课程列表如下,随时下架,免费名额有限,先到先得: 领取指南: 总共分四步,即可领取课程并学习测试获得证书 一、新注册开发者用户 填写未注册过的邮箱——设置密码——完善资料 二、免费领取一节免费课程 下滑到NVIDIA 培训和认证模块,点击立即领取,并补充个人信息 注意:年龄仅限18岁以上,Location选择China 三、领取一节您感兴趣的中文或者英文课程 右上角可选择语言挑选中/英文课程 四、开始学习并通过测试获得NVIDIA培训证书 学习前,请阅读下图四步进行环境确认 自测链接:http://websocketstest.courses.nvidia.com
- 大电芯降本、AI算力“施压”,谁来替储能系统兜底这笔“物理账”?
领取规则:未注册过开发者的用户可以通过如下链接免费选择一门 DLI 在线自主培训的付费课程,配套云端实验环境和可获得 NVIDIA 培训证书。每位用户(每个邮箱账号)仅可选择一门。 目前可选课程包括 7 门英文课,5 门中文课,目前课程列表如下,随时下架,免费名额有限,先到先得:
- 大电芯降本、AI算力“施压”,谁来替储能系统兜底这笔“物理账”?
领取规则:未注册过开发者的用户可以通过如下链接免费选择一门 DLI 在线自主培训的付费课程,配套云端实验环境和可获得 NVIDIA 培训证书。每位用户(每个邮箱账号)仅可选择一门。 目前可选课程包括 7 门英文课,5 门中文课,目前课程列表如下,随时下架,免费名额有限,先到先得:
- 两个超大质量黑洞可能在百年内合并
马普射电天文研究所研究员领导的国际团队在 Mrk 501 星系中心发现了一对即将合并的黑洞的直接证据。Mrk 501 是一个椭圆星系,位于武仙座,它是一个活跃星系核,其中心黑洞喷射出强大粒子流。研究人员分析了长达 23 年的数据,发现其中心存在两个超大质量黑洞的证据。两个黑洞以约 121 天的周期相互绕转,它们之间的距离相当于地日距离的 250-540 倍,对于质量为太阳 1 亿到 10 亿倍的天体而言,这一距离微不足道。根据它们的质量,两大黑洞有可能在百年内合并。
- 科学家捏造了一种病,AI 告诉人们这是真的
眼睛酸痛发痒?你可能和其它数百万人一样,长时间暴露在屏幕上的蓝光下而眼睛疲劳。你可能会因此多次揉眼睛,眼睑可能会泛起粉红色。如果你将这些症状输入到 AI 聊天工具里,过去一年半 AI 聊天机器人可能会给出一个奇怪的答案:bixonimania。这种疾病没有出现在标准医学文献中,原因是它根本就不存在,是瑞典哥德堡大学 Almira Osmanovic Thunström 团队捏造出来的。研究团队在 2024 年初将两篇关于该虚构皮肤病的论文上传到预印本服务器,测试大模型是否会接受虚假信息并将其作为权威医疗建议发布。结果可能比预想的还要好。上传几周后,大模型就开始鹦鹉学语般重复假消息,仿佛它真的存在。更糟糕的是,该虚构皮肤病论文还被其他研究人员引用,发表在同行评审的期刊上。Osmanovic Thunström 认为这表明部分研究人员依赖于 AI 生成的文献,并没有真的阅读原始论文——论文中包含了大量线索表明它是伪造的,比如作者名字叫 Izgubljenovic,在虚构的加州城市 Nova City 的虚构大学 Asteria Horizon University 工作。
- W玻色子质量测量结果与标准模型一致
当基本粒子的质量测量结果偏离理论预测时,往往可能动摇现有理论体系。但 LHC 的实验结果显示,W 玻色子质量测量结果与标准模型一致。W 玻色子是传递弱相互作用的基本粒子之一,而弱相互作用是自然界四种基本作用力之一。这种作用力使粒子能够发生“身份转换”,例如质子可以转变为中子,反之亦然。正是这种转变驱动了放射性衰变,也使得为太阳提供能量的核聚变反应成为可能。研究团队对 LHC 在 2016 年产生的超过 10 亿次质子碰撞数据进行了分析。研究团队测得 W 玻色子的质量为 80360.2±9.9 MeV,这一结果与标准模型的理论预测相符。
- 微软考虑加固数据中心以抵御战火
微软正重新评估在冲突频发地区设计和建造数据中心的方式,如加固数据中心,或将数据中心建造在地堡内。微软总裁 Brad Smith 接受 Nikkei Asia 采访时呼吁制定国际规则促进对民用基础设施的保护,而数据中心也应该包含在内。微软在中东建造运营了大量数据中心,地点包括了阿联酋、卡塔尔和以色列,计划于今年晚些时候在沙特阿拉伯开展业务。这些地点如今被认为容易受到战争影响。
- 微软终止 VeraCrypt 账户 Windows 版本更新暂停
开源加密工具 VeraCrypt 的开发者 Mounir Idrassi 透露,微软终止了他用于给驱动和引导程序签名的账号,没有给出解释,导致该工具的 Windows 版本无法更新。VeraCrypt 基于已经停止开发的加密工具 TrueCrypt,允许用户在硬盘上创建加密分区,或者创建单独的加密卷存储文件。如果用户被迫交出登陆凭证,该工具也允许用户创建一个隐藏卷。Idrassi 称他的账号是在 1 月被关闭的,微软在账号关闭的信息中表示其组织没有通过微软的验证要求,随即关闭了申述通道。Idrassi 的联系信息都返回了自动回复信息。Idrassi 表示这导致他无法再释出 Windows 更新,Linux 和 macOS 的更新仍然能照常,但 Windows 是用户数最多平台,无法更新 Windows 版本是一次重大打击。VeraCrypt 不是唯一受到影响的项目,很多微软 Windows 驱动开发者报告他们的账号被锁定了。
- 纽时记者称 Adam Back 是中本聪
纽时记者 John Carreyrou 在调查之后宣称 Adam Back 是比特币作者中本聪,Adam Back 本人公开否认。Adam Back 出生于 1970 年,是一位英国密码学家和密码朋克,工作量证明系统 Hashcash 的作者,这不是第一次他被认为是比特币作者。Carreyrou 是知名的调查记者和作家,他的调查认为,Back 的写作风格、意识形态、技术背景以及早年发表的帖子都与中本聪有重合之处。他的调查主要是基于芬兰程序员 Martti Malmi 公开的数百封与中本聪交流的电子邮件。Back 通过其 X 账号回应称,他不是中本聪,但他早年非常关注密码学、网络隐私和电子现金的积极社会影响,因此从 1992 年左右开始积极参与电子现金和隐私技术的应用研究,在密码朋克邮件列表中参与讨论,最终促成了 Hashcash 等创意的诞生。
- 全新世最暴力火山正在重新注满岩浆
大约 7300 年前,位于日本九州萨摩半岛的鬼界破火山口发生了大规模喷发。这次喷发是当前地质期全新世规模最大的火山喷发,火山喷出的物质覆盖 4500 平方公里范围。此后这座火山没有再发生大规模喷发,但仍然活跃,发生过零星的小型喷发。根据发表在《Communications Earth & Environment》期刊上的一项研究,日本科学家报告这座大部分位于海底的火山正在重新注满岩浆,引发了它再次喷发的担忧。鉴于该地区的人口密度,任何规模的喷发都可能造成严重破坏。