SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 39 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Efficiency Breakthrough
S-VGGT introduces structure-aware subscene decomposition to break the quadratic scaling bottleneck of 3D foundation models.
Mar 19
New Capability
Introduces a framework to generate complex, non-linear environments with mathematically guaranteed ground-truth optimal policies for RL benchmarking.
Mar 19
Efficiency Breakthrough
DSS-GAN is the first generative adversarial network to use a Mamba (State Space Model) backbone for high-quality image synthesis.
Mar 19
New Capability
VectorWorld enables stable, real-time 1km+ closed-loop world model rollouts for autonomous driving using diffusion flow on vector graphs.
Mar 19
New Capability
REAL achieves extreme quadruped parkour agility that is robust even to a 1-meter visual blind zone.
Mar 19
Breaks Assumption
FINER discovers that MLLMs are highly prone to hallucination when images contain fine-grained mismatches co-occurring with real elements.
Mar 19
Efficiency Breakthrough
Synthetic videos of simple geometric shapes are more effective than massive real-world datasets for teaching video-language models fundamental temporal reasoning.
Mar 19
New Capability
Lifting 2D features into a volumetric representation for robot manipulation policies yields a 14.8% success rate improvement by solving the 2D-3D spatial reasoning mismatch.
Mar 19
Paradigm Shift
A new self-refining surrogate framework enables neural models to simulate complex dynamical systems over arbitrarily long horizons without the standard failure mode of compounding error.
Mar 19
Breaks Assumption
Massive activation outliers in Transformers are an adaptive response to 'gradient sinks' during training, rather than just an inference-time quirk.
Mar 19
Paradigm Shift
The 'consensus trap' in label-free RL—where models reinforce their own systematic errors—can be broken by co-evolving the model in alternating generator and verifier roles.
Mar 19
Breaks Assumption
In-context memory for LLMs is fundamentally unreliable due to compaction loss and goal drift, but structured 'Knowledge Objects' provide a 252x cheaper and 100% accurate alternative.
Mar 19
Efficiency Breakthrough
Anomaly detection can be performed directly using a primary model's internal neuron output ranges, eliminating the need for expensive external AD models.
Mar 19
Efficiency Breakthrough
Truncated backpropagation for video decoding reduces the memory cost of fine-tuning video diffusion models from linear to constant.
Mar 19
Breaks Assumption
Concept erasure in text-to-image models is largely a facade that can be bypassed using text-free inversion attacks.
Mar 19
Paradigm Shift
LLMs compute and cache confidence scores automatically during answer generation, well before they are prompted to verbalize them.
Mar 19
Efficiency Breakthrough
ProbeFlow achieves 14.8x faster action decoding in Vision-Language-Action (VLA) models without any retraining.
Mar 19
New Capability
DebugLM allows developers to trace an LLM's specific behaviors back to individual training data sources.
Mar 19
Paradigm Shift
Measuring the distance between human languages can now be done quantitatively using the attention mechanisms of multilingual transformers.
Mar 19
Breaks Assumption
Large Language Models can maintain performance with only 16-64 unique weight values per matrix, as only the relative rank of weights matters.
Mar 19
Efficiency Breakthrough
Parallel multi-token prediction can be achieved in standard LLMs without training auxiliary models or modifying weights.
Mar 19
Efficiency Breakthrough
CARE provides a recipe for converting standard GQA models into high-efficiency Multi-head Latent Attention (MLA) architectures.
Mar 19
Efficiency Breakthrough
VideoAtlas enables navigation and reasoning over long-form video using compute that scales only logarithmically with video length.
Mar 19
New Capability
Enforce formal safety and Signal Temporal Logic (STL) constraints on robotics foundation models without retraining.
Mar 19
Efficiency Breakthrough
MUD provides a faster, lower-overhead alternative to Muon for transformer training, achieving up to 2.6x higher throughput.
Mar 19
Efficiency Breakthrough
LoST introduces a semantic-first 3D tokenizer that reduces the token count for 3D shape generation by up to 99.9%.
Mar 19
Paradigm Shift
AgentFactory shifts agent evolution from unreliable textual 'reflections' to a library of verifiable, executable Python subagents.
Mar 19
New Capability
SkeletonLLM allows frozen Multimodal LLMs to reason about human motion by rendering skeleton sequences into their native visual modality.
Mar 19
Paradigm Shift
DAPS++ reinterprets diffusion inverse problems as a decoupled EM-style initialization, significantly increasing restoration speed and stability.
Mar 19
New Capability
Motion-MLLM integrates IMU egomotion data into Video-LLMs to solve the fundamental scale and spatial reasoning ambiguities of purely visual models.
Mar 19
Scaling Insight
Provides the first theoretical proof that Graph Transformers structurally prevent the 'oversmoothing' failure mode inherent to deep GCNs.
Mar 19
First Ever
Imagine an AI virus that doesn't just sit there—it copies itself and jumps from one AI to the next all on its own.
Mar 18
Practical Magic
A new VR headset uses mirrors to kill the lag that makes you want to puke.
Mar 18
Nature Is Weird
These tiny sliding antennas are hacking the laws of physics to give you a perfect signal where your phone usually dies.
Mar 18
Practical Magic
New AI can peer into a computer chip's microscopic guts to find "spy tech" hidden by sketchy manufacturers.
Mar 18
Practical Magic
Researchers built a "ghost mode" for robots that calculates the exact path to sneak around without being seen.
Mar 18
Paradigm Challenge
Turns out the long lines at airport security were secretly keeping the whole U.S. flight network from crashing for the last decade.
Mar 18
Efficiency Breakthrough
RSM achieves 20x faster training for recursive reasoning models and enables test-time scaling for up to 20,000 refinement steps.
Mar 18
Scaling Insight
A factorial study on EHR foundation models reveals that joint encoding of code-attribute pairs (local binding) is the primary driver of performance and efficiency.
Mar 18
Paradigm Shift
Alternating Reinforcement Learning with Rubric Rewards (ARL-RR) replaces brittle scalar reward aggregation with a semantic meta-class optimization framework.
Mar 18
Breaks Assumption
Self-reflective program search matches or outperforms recursive language models for long-context tasks, suggesting recursion itself is not the primary driver of performance.
Mar 18
New Capability
Dynamic Representational Circuit Breaking (DRCB) introduces an architectural defense against steganographic collusion in multi-agent RL by monitoring and shuffling latent communication bottlenecks.
Mar 18
Breaks Assumption
Theoretical and empirical evidence suggests that the 'Key' mechanism in Attention may be redundant, proposing a 'QV' paradigm that simplifies Transformer architectures.
Mar 18
Paradigm Shift
Atlas introduces 'Compiled Memory,' which rewrites an agent's system prompt with distilled task experience rather than using RAG or fine-tuning.
Mar 18
New Capability
Latent Posterior Factors (LPF) bridge neural representations with structured probabilistic reasoning by converting VAE posteriors into factors for Sum-Product Networks.
Mar 18
Scaling Insight
Spectral Edge Dynamics (SED) provides an early-warning signal for grokking, predicting generalization up to 1,700 steps before it occurs.
Mar 18
Paradigm Shift
Transition Flow Matching learns a global transition flow rather than local velocity fields, enabling single-step generation and transfer to arbitrary future time points.
Mar 18
Breaks Assumption
Robot policy performance can be improved by up to 60% by identifying a single 'golden ticket' constant noise vector instead of sampling from a Gaussian.
Mar 18
Paradigm Shift
Simulation Distillation (SimDist) enables rapid sim-to-real adaptation by transferring reward and value models directly into a latent world model.
Mar 18
Scaling Insight
Demonstrates that massive scaling of diverse simulator resets can replace manual curriculum engineering for complex dexterous manipulation.
Mar 18