SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 45 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

New Capability
Incorporating PDE residuals into fine-tuning allows pre-trained physics foundation models to adapt to new tasks without requiring ground-truth solutions.
Mar 17
Efficiency Breakthrough
PrismMirror is the first monocular human frontal view synthesis model to achieve real-time inference (24 FPS) without external geometric models.
Mar 17
Breaks Assumption
Challenges the 'Flat Minima' hypothesis by showing that grokking is driven by anisotropic noise rectification rather than finding flat regions.
Mar 17
Efficiency Breakthrough
A 4B parameter model matches a 120B parameter model in program verification through a rigorous data curation pipeline.
Mar 17
Efficiency Breakthrough
Bridges the gap between generative (MAE) and predictive (I-JEPA) self-supervised learning, achieving a 10% performance boost.
Mar 17
New Capability
Mamba-3 introduces MIMO formulations and complex-valued updates to solve the state-tracking failures of previous linear models.
Mar 17
Open Release
Democratizes the development of 'Deep Search' agents by open-sourcing the specialized training data and trajectory synthesis methods.
Mar 17
Breaks Assumption
Proves that simple deterministic ranking beats expensive LLM-based structuring for conversational memory retrieval.
Mar 17
Efficiency Breakthrough
Accelerates state-of-the-art 3D human mesh recovery by over 10x, enabling real-time vision-only humanoid teleoperation.
Mar 17
Paradigm Shift
Introduces an adversarial co-evolution framework where Code and Test LLMs optimize against each other to improve code generation.
Mar 17
New Capability
Uses Sparse Autoencoders (SAEs) to mechanisticially repair 'moral indifference' in LLM latent representations.
Mar 17
New Capability
A benchmark for unsolved math problems with automated verification, enabling the measurement of true mathematical discovery.
Mar 17
Efficiency Breakthrough
Introduces Mixture-of-Depths Attention (MoDA) to solve signal degradation in deep LLMs with hardware-efficient implementation.
Mar 17
Breaks Assumption
Proves that standard acquisition functions like UCB are sufficient for asynchronous Bayesian Optimization, debunking the need for complex diversity-enforcing strategies.
Mar 17
Breaks Assumption
Settles the long-standing practitioner debate over whether to use training or holdout data for interpreting black-box models with PD/ALE plots.
Mar 17
New Capability
Enables Bayesian model selection and joint posterior inference over combinatorial spaces of up to billions of simulator model instantiations.
Mar 17
Efficiency Breakthrough
Achieves 1,000x speedups in Bayesian inverse problems by replacing repeated MCMC sampling with one-step preconditioned generative transport.
Mar 17
Practical Magic
Imagine a paper-thin sticker you can slap on a wall to listen to the room next door, and get this—it doesn't even need a battery.
Mar 16
Paradigm Challenge
Future 6G antennas are going to literally slide around on your phone to grab a signal so sharp it shouldn't even be possible.
Mar 16
Efficiency Breakthrough
ActTail achieves 80% activation sparsity in LLMs with significantly lower perplexity degradation than uniform methods by using Heavy-Tailed Self-Regularization theory.
Mar 16
Paradigm Shift
This paper proposes a method to align and personalize LLMs directly from raw user interactions using self-distillation, bypassing the need for explicit human labels or RLHF.
Mar 16
Breaks Assumption
The researchers demonstrate that prompt injection is caused by 'role confusion' in the latent space, where models assign authority based on the style of writing rather than the source of the text.
Mar 16
Breaks Assumption
This theoretical work refutes the 'Garbage In, Garbage Out' mantra for modern ML, proving that high-dimensional model capacity can asymptotically overcome predictor error and structural uncertainty.
Mar 16
Paradigm Shift
Introduces the Budget-Sensitive Discovery Score (BSDS), a formally verified metric machine-checked in Lean 4 for evaluating AI-guided scientific candidate selection.
Mar 16
Efficiency Breakthrough
ReBalance is a training-free framework that dynamically modulates 'thinking' length in reasoning models to prune redundancy during overthinking and promote exploration during underthinking.
Mar 16
Breaks Assumption
This study proves that reasoning traces (Chain-of-Thought) causally shape model behavior and generalization, even when the final answer is held constant.
Mar 16
Breaks Assumption
SpectralGuard identifies a 'memory collapse' vulnerability in State Space Models (like Mamba) where adversarial inputs can drive the transition operator's spectral radius to zero.
Mar 16
Open Release
Surg-R1 is a specialized surgical reasoning model released alongside the largest surgical Chain-of-Thought dataset (320,000 pairs).
Mar 16
Paradigm Shift
This paper establishes a systematic protocol for 'stitching' heterogeneous Vision Foundation Models (e.g., CLIP and DINOv2) to share early layers while retaining specialized capabilities.
Mar 16
Efficiency Breakthrough
Achieves 100x speedup in robotic action generation by distilling iterative flow/diffusion models into a one-step policy without a pre-trained teacher.
Mar 16
Paradigm Shift
Introduces Modal Logical Neural Networks (MLNNs) as a differentiable logic layer that bridges deep learning with symbolic Kripke semantics for regulated AI.
Mar 16
Paradigm Shift
Demonstrates a robot that improves its own locomotion by identifying and physically 'self-destructing' redundant or inhibiting limbs during its lifetime.
Mar 16
New Capability
Enables training-free infinite video generation (hour-scale) by using evolving memory tokens to solve identity drift and motion stagnation.
Mar 16
Breaks Assumption
Reveals that standard global correlation metrics for LLM judges fail to predict success in 'best-of-n' selection tasks due to within-prompt signal loss.
Mar 16
Efficiency Breakthrough
Reduces Chain-of-Thought (CoT) compute costs by 14-55% by learning the optimal 'early-exit' points for Large Reasoning Models.
Mar 16
Scaling Insight
Discovers that as LLMs scale, their complex non-linear depth dynamics converge into accurate, low-order linear surrogates.
Mar 16
Paradigm Shift
Derives an exact, unbiased policy gradient for Reinforcement Learning on Diffusion LLMs, bypassing the need for sequence-level likelihood approximations.
Mar 16
Breaks Assumption
Shows that tool-augmented agents suffer from 'recommendation drift' where they provide unsafe advice under tool corruption while maintaining high ranking scores.
Mar 16
Efficiency Breakthrough
Accelerates Diffusion Transformers (DiTs) by 2x using a training-free framework that selectively reduces computation in non-aesthetic image regions.
Mar 16
Breaks Assumption
Challenges the standard practice of deep PPO training by proving that consensus aggregation of 'wider' parallel runs is 8x more sample efficient than multiple epochs.
Mar 16
Open Release
Releases Feynman, an agentic pipeline and 100k-sample dataset for generating high-quality, knowledge-rich diagrams with grounded captions.
Mar 16
Open Release
Introduces the largest-ever multi-modal CAD dataset with 10 million annotations for 1 million models to enable geometric deep learning on BRep data.
Mar 16
New Capability
Unlocks Maximum Entropy RL for high-dimensional humanoid control, matching or doubling the performance of dominant deterministic baselines.
Mar 16
Efficiency Breakthrough
Introduces a training-free framework that allows LLM agents to dynamically scale their reasoning depth based on a pre-defined token/tool budget.
Mar 16
Efficiency Breakthrough
Achieves a 98x speedup in LLM routing on AMD hardware using Flash Attention and prompt compression, enabling high-context classification without a dedicated GPU.
Mar 16
Paradigm Shift
Proposes modeling the world in the feature space of frozen geometry foundation models instead of pixels, achieving 5x faster depth forecasting.
Mar 16
New Capability
A retrosynthesis model that explicitly learns strategic bond-disconnection reasoning via reinforcement learning with a round-trip accuracy reward.
Mar 16
Scaling Insight
Longitudinal evidence reveals that successive ChatGPT versions are converging in output diversity, suggesting potential model collapse from synthetic data saturation.
Mar 16
New Capability
A new system enables humanoid robots to play competitive tennis rallies with humans by learning from imperfect, fragmented motion data.
Mar 16
Scaling Insight
Adversarial test case evolution improves code reinforcement learning by creating harder, more discriminative verification signals that drive better model performance.
Mar 16