SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,557 papers  ·  Page 49 of 52

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Breaks Assumption
Settles the long-standing practitioner debate over whether to use training or holdout data for interpreting black-box models with PD/ALE plots.
Mar 17
New Capability
Enables Bayesian model selection and joint posterior inference over combinatorial spaces of up to billions of simulator model instantiations.
Mar 17
Efficiency Breakthrough
Achieves 1,000x speedups in Bayesian inverse problems by replacing repeated MCMC sampling with one-step preconditioned generative transport.
Mar 17
Practical Magic
Imagine a paper-thin sticker you can slap on a wall to listen to the room next door, and get this—it doesn't even need a battery.
Mar 16
Paradigm Challenge
Future 6G antennas are going to literally slide around on your phone to grab a signal so sharp it shouldn't even be possible.
Mar 16
Efficiency Breakthrough
ActTail achieves 80% activation sparsity in LLMs with significantly lower perplexity degradation than uniform methods by using Heavy-Tailed Self-Regularization theory.
Mar 16
Paradigm Shift
This paper proposes a method to align and personalize LLMs directly from raw user interactions using self-distillation, bypassing the need for explicit human labels or RLHF.
Mar 16
Breaks Assumption
The researchers demonstrate that prompt injection is caused by 'role confusion' in the latent space, where models assign authority based on the style of writing rather than the source of the text.
Mar 16
Breaks Assumption
This theoretical work refutes the 'Garbage In, Garbage Out' mantra for modern ML, proving that high-dimensional model capacity can asymptotically overcome predictor error and structural uncertainty.
Mar 16
Paradigm Shift
Introduces the Budget-Sensitive Discovery Score (BSDS), a formally verified metric machine-checked in Lean 4 for evaluating AI-guided scientific candidate selection.
Mar 16
Efficiency Breakthrough
ReBalance is a training-free framework that dynamically modulates 'thinking' length in reasoning models to prune redundancy during overthinking and promote exploration during underthinking.
Mar 16
Breaks Assumption
This study proves that reasoning traces (Chain-of-Thought) causally shape model behavior and generalization, even when the final answer is held constant.
Mar 16
Breaks Assumption
SpectralGuard identifies a 'memory collapse' vulnerability in State Space Models (like Mamba) where adversarial inputs can drive the transition operator's spectral radius to zero.
Mar 16
Open Release
Surg-R1 is a specialized surgical reasoning model released alongside the largest surgical Chain-of-Thought dataset (320,000 pairs).
Mar 16
Paradigm Shift
This paper establishes a systematic protocol for 'stitching' heterogeneous Vision Foundation Models (e.g., CLIP and DINOv2) to share early layers while retaining specialized capabilities.
Mar 16
Efficiency Breakthrough
Achieves 100x speedup in robotic action generation by distilling iterative flow/diffusion models into a one-step policy without a pre-trained teacher.
Mar 16
Paradigm Shift
Introduces Modal Logical Neural Networks (MLNNs) as a differentiable logic layer that bridges deep learning with symbolic Kripke semantics for regulated AI.
Mar 16
Paradigm Shift
Demonstrates a robot that improves its own locomotion by identifying and physically 'self-destructing' redundant or inhibiting limbs during its lifetime.
Mar 16
New Capability
Enables training-free infinite video generation (hour-scale) by using evolving memory tokens to solve identity drift and motion stagnation.
Mar 16
Breaks Assumption
Reveals that standard global correlation metrics for LLM judges fail to predict success in 'best-of-n' selection tasks due to within-prompt signal loss.
Mar 16
Efficiency Breakthrough
Reduces Chain-of-Thought (CoT) compute costs by 14-55% by learning the optimal 'early-exit' points for Large Reasoning Models.
Mar 16
Scaling Insight
Discovers that as LLMs scale, their complex non-linear depth dynamics converge into accurate, low-order linear surrogates.
Mar 16
Paradigm Shift
Derives an exact, unbiased policy gradient for Reinforcement Learning on Diffusion LLMs, bypassing the need for sequence-level likelihood approximations.
Mar 16
Breaks Assumption
Shows that tool-augmented agents suffer from 'recommendation drift' where they provide unsafe advice under tool corruption while maintaining high ranking scores.
Mar 16
Efficiency Breakthrough
Accelerates Diffusion Transformers (DiTs) by 2x using a training-free framework that selectively reduces computation in non-aesthetic image regions.
Mar 16
Breaks Assumption
Challenges the standard practice of deep PPO training by proving that consensus aggregation of 'wider' parallel runs is 8x more sample efficient than multiple epochs.
Mar 16
Open Release
Releases Feynman, an agentic pipeline and 100k-sample dataset for generating high-quality, knowledge-rich diagrams with grounded captions.
Mar 16
Open Release
Introduces the largest-ever multi-modal CAD dataset with 10 million annotations for 1 million models to enable geometric deep learning on BRep data.
Mar 16
New Capability
Unlocks Maximum Entropy RL for high-dimensional humanoid control, matching or doubling the performance of dominant deterministic baselines.
Mar 16
Efficiency Breakthrough
Introduces a training-free framework that allows LLM agents to dynamically scale their reasoning depth based on a pre-defined token/tool budget.
Mar 16
Efficiency Breakthrough
Achieves a 98x speedup in LLM routing on AMD hardware using Flash Attention and prompt compression, enabling high-context classification without a dedicated GPU.
Mar 16
Paradigm Shift
Proposes modeling the world in the feature space of frozen geometry foundation models instead of pixels, achieving 5x faster depth forecasting.
Mar 16
New Capability
A retrosynthesis model that explicitly learns strategic bond-disconnection reasoning via reinforcement learning with a round-trip accuracy reward.
Mar 16
Scaling Insight
Longitudinal evidence reveals that successive ChatGPT versions are converging in output diversity, suggesting potential model collapse from synthetic data saturation.
Mar 16
New Capability
A new system enables humanoid robots to play competitive tennis rallies with humans by learning from imperfect, fragmented motion data.
Mar 16
Scaling Insight
Adversarial test case evolution improves code reinforcement learning by creating harder, more discriminative verification signals that drive better model performance.
Mar 16
Efficiency Breakthrough
Modality-level disaggregation enables cost-optimal MLLM serving across heterogeneous GPUs over commodity PCIe, bypassing the need for expensive NVLink interconnects.
Mar 16
Breaks Assumption
Probing of Vision-Language-Action (VLA) models reveals that the action decoder largely ignores the reasoning logic in Chain-of-Thought, relying almost exclusively on object names.
Mar 16
New Capability
SciDesignBench provides a massive simulator-grounded environment for scientific inverse design, revealing that current LLMs struggle significantly with iterative refinement.
Mar 16
Efficiency Breakthrough
A hardware-algorithm co-design for Spiking Neural Networks achieves up to 69x energy efficiency gains using an SRAM-based Compute-in-Memory accelerator.
Mar 16
Breaks Assumption
The TaoBench benchmark proves that state-of-the-art math LLMs fail on equivalent logic problems when presented outside of the standard 'MathLib' framework.
Mar 16
New Capability
A self-supervised robotic system detects novel objects by training bespoke detectors on-the-fly from human video demonstrations, bypassing language-based prompts.
Mar 16
New Capability
AIM enables post-training modulation of large models to change utility levels or focus features without any retraining or additional data.
Mar 16
Efficiency Breakthrough
Achieves 4x visual token compression and 80% lower training cost while unifying multimodal comprehension and generation.
Mar 16
New Capability
First training-free method for debiasing reward models using Sparse Autoencoder (SAE) interventions.
Mar 16
Breaks Assumption
Breaks the long-standing accuracy-robustness trade-off in VLMs by localizing adversarial robustness to shallow layers.
Mar 16
New Capability
A flow-based navigation policy that achieves zero-shot sim-to-real transfer across wheeled, quadrupedal, and humanoid platforms.
Mar 16
Paradigm Shift
A small-scale molecular reasoning model that outperforms ultra-large foundation models via structured chain-of-thought and RL.
Mar 16
Efficiency Breakthrough
Adaptive VLM Routing reduces inference costs for Computer Use Agents by up to 78% with negligible accuracy loss.
Mar 16
Efficiency Breakthrough
Distills a 2B Vision-Language Retriever into a 70M text-only encoder for visual document retrieval with 50x lower latency.
Mar 16