SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 40 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Efficiency Breakthrough
Reduces high-quality 3D head avatar creation time from over 24 hours to 0.5 seconds per frame.
Mar 18
Breaks Assumption
Reveals that models with identical predictive performance produce fundamentally different feature attributions based solely on their hypothesis class.
Mar 18
Paradigm Shift
Introduces a privacy-preserving ML framework that achieves strong non-invertibility without the utility loss of Differential Privacy or the cost of Homomorphic Encryption.
Mar 18
Efficiency Breakthrough
Fuses categorical sampling into the LM-head matmul to eliminate logit materialization and speed up LLM decoding by up to 19%.
Mar 18
Paradigm Shift
Analyses over 10,000 experiments to prove that LLM agents are capable of genuine architectural discovery rather than just hyperparameter tuning.
Mar 18
Breaks Assumption
Provides empirical evidence that structural sparsity in Vision Transformers does not lead to improved semantic interpretability.
Mar 18
New Capability
Demonstrates a complete AI-assisted mathematical research loop where a mathematician wrote zero lines of formal code to verify complex physics equilibria.
Mar 18
New Capability
Integrates LLM agents with the industry-standard Rosetta software to automate physics-based protein design for non-canonical amino acids.
Mar 18
Breaks Assumption
Releases 70B parameter models that operate entirely on bytes, effectively 'liberating' LLMs from static tokenizers.
Mar 18
Scaling Insight
Derives closed-form power-law scaling for hyperparameters like learning rate and batch size using modern optimization theory rather than expensive empirical sweeps.
Mar 18
Paradigm Shift
Introduces per-token adapter routing, allowing a single sequence to dynamically utilize multiple specialized LoRA experts.
Mar 18
Breaks Assumption
Provides the first formal proof that safety is non-compositional, meaning two individually safe AI agents can become hazardous when combined.
Mar 18
New Capability
Enables the prediction of an adapter's task, performance, and attributes directly from its LoRA weights without any inference or data access.
Mar 18
Paradigm Shift
Finds that filtering knowledge at 'write-time' (ingestion) maintains 100% RAG accuracy under noise levels where standard 'read-time' filtering completely collapses.
Mar 18
Paradigm Shift
Proposes a protocol that replaces complex multi-agent coding frameworks with a simple, interpretable filesystem structure.
Mar 18
Paradigm Shift
Establishes a duality between sequence-axis attention and depth-wise residual connections, treating layer depth as an ordered variable.
Mar 18
Efficiency Breakthrough
Achieves microsecond-level kinodynamic motion planning for high-DOF robots by using differential flatness to solve boundary value problems analytically.
Mar 18
New Capability
Introduces ARISE, a hierarchical reinforcement learning framework that allows LLMs to evolve and reuse a tiered library of reasoning skills rather than treating every math problem in isolation.
Mar 18
Breaks Assumption
Challenges the standard use of bilinear/bicubic interpolation for upsampling saliency maps, proving it creates spurious importance regions and proposing a mass-redistribution alternative.
Mar 18
Efficiency Breakthrough
Demonstrates that masked diffusion language models can be 21.8x more compute-efficient than traditional autoregressive models when scaled correctly.
Mar 18
New Capability
Proposes the Vision-Sound-Language-Action (VSLA) paradigm, enabling robots to respond to real-time environmental acoustics during task execution.
Mar 18
Breaks Assumption
Debunks the widely held 'intra-modal misalignment hypothesis' which claimed CLIP embeddings are inherently poor for image-only tasks.
Mar 18
Efficiency Breakthrough
Introduces Helium, a serving framework that treats agentic workflows as data query plans to optimize redundant LLM calls and KV caches.
Mar 18
Efficiency Breakthrough
Presents ZipCal, a model-agnostic calibration data selection strategy for pruning and quantization that is 240x faster than model-based methods.
Mar 18
Paradigm Shift
Proves that compositional generalization failure in neural networks is an architectural issue and provides a category-theoretic framework to fix it.
Mar 18
Breaks Assumption
Discovers that skipping learning rate decay during pre-training, while appearing worse for pre-train loss, significantly improves the model's adaptability during supervised fine-tuning (SFT).
Mar 18
Breaks Assumption
Proves that noisy/incorrect labels are destructive to Reinforcement Learning with Verifiable Rewards (RLVR), contradicting recent high-profile claims that noise doesn't matter.
Mar 18
New Capability
Successfully trains a 0.9B parameter pure Spiking Neural Network (SNN) from scratch for language modeling, achieving performance without Transformer distillation.
Mar 18
Paradigm Shift
Formulates Hierarchical Instruction Following as a Constrained Markov Decision Process to ensure LLMs prioritize system prompts over user instructions.
Mar 18
New Capability
Localizes reinforcement learning updates for code generation by using execution traces to identify the exact point of semantic failure.
Mar 18
Breaks Assumption
Challenges the standard 'pretrain-then-finetune' pipeline by showing that repeating domain-specific data during pretraining is significantly more effective.
Mar 18
Breaks Assumption
A rigorous multi-method audit revealing that frontier LLM performance on MMLU is significantly inflated by data contamination and memorization.
Mar 18
Paradigm Shift
Introduces modular, composable safety alignment via learnable control tokens rather than static parameter-level tuning.
Mar 18
New Capability
Uses an asymmetric Draft-Verify-Recover pipeline to enable high-quality personalized AI assistants without compromising user privacy.
Mar 18
New Capability
A self-supervised RLVR method that escapes the 'spurious majority' trap by using a temporary unlearning process for exploration.
Mar 18
Paradigm Shift
Decouples perceptual failures from logical errors in Vision-Language reward models to enable more reliable test-time scaling.
Mar 18
Paradigm Shift
Researchers identified a 'critique vector' in the latent space of Large Reasoning Models that can be steered to improve self-correction and test-time scaling.
Mar 18
New Capability
Omnilingual MT scales machine translation to over 1,600 languages, an 8x increase in coverage over previous state-of-the-art systems.
Mar 18
New Capability
This paper demonstrates precise behavioral steering of agentic traits in a 35B parameter MoE model using Sparse Autoencoder (SAE) decoded probe vectors.
Mar 18
Paradigm Shift
FederatedFactory solves the 'extreme non-IID' problem in Federated Learning by federating generative priors instead of model weights.
Mar 18
Paradigm Shift
Laya introduces the first EEG foundation model based on Joint Embedding Predictive Architecture (JEPA), outperforming traditional reconstruction-based models.
Mar 18
New Capability
Introduces a method to give frozen LLMs persistent memory in their continuous latent space, bypassing the need for text-level RAG or retraining.
Mar 18
Paradigm Shift
IndexRAG shifts cross-document reasoning from inference-time prompting to offline indexing by generating 'bridging facts' at index time.
Mar 18
Paradigm Shift
Provides a theoretical framework for why training AI on what to avoid (negative constraints) is structurally superior and more stable than training on preferences.
Mar 18
Efficiency Breakthrough
VQKV uses Vector Quantization to achieve over 80% KV cache compression with almost zero loss in model performance.
Mar 18
New Capability
Capability-Guided Compression uses Sparse Autoencoders (SAEs) to prevent 'capability loss' during model pruning and quantization.
Mar 18
Breaks Assumption
A causal analysis reveals that LLMs often ignore their own intermediate reasoning (Chain-of-Thought) when making final decisions.
Mar 18
Efficiency Breakthrough
FEAT is a linear-complexity foundation model designed specifically for extremely large-scale structured (tabular) data.
Mar 18
Open Release
Kamino is a massively parallel GPU physics solver that natively supports complex kinematic loops and multi-body systems.
Mar 18
New Capability
Detects and mitigates Vision-Language Model hallucinations at inference time by analyzing visual attention entropy rather than text outputs.
Mar 18