SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 38 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Paradigm Shift
Proposes Modulated Hazard-aware Policy Optimization (MHPO) to solve the instability and mode collapse common in GRPO-based reinforcement learning.
Mar 19
Efficiency Breakthrough
AwaRes enables low-resolution Vision-Language Models to retrieve only the high-resolution image crops needed for a specific query via tool-calling.
Mar 19
New Capability
Minimum-Action Learning achieves a 10,000x reduction in noise variance for symbolic physical law identification from observational data.
Mar 19
New Capability
Learns task-specific dense reward functions directly from images using vision foundation models, without requiring privileged simulator states.
Mar 19
Breaks Assumption
Uses SMT solvers to formally verify the physical consistency of tree-based ML models across their entire input domain.
Mar 19
Efficiency Breakthrough
Provides a systematic profiling of VLM inference bottlenecks and releases 'recipes' that cut time-to-first-token by up to 93%.
Mar 19
Breaks Assumption
Provides a formal proof and empirical evidence that Transformers can learn symbolic rules entirely absent from training, debunking the 'stochastic parrot' interpolation-only hypothesis.
Mar 19
New Capability
Introduces HopChain, a framework for synthesizing multi-hop vision-language reasoning data that yields generalizable gains across 20+ diverse benchmarks.
Mar 19
Breaks Assumption
Identifies a fundamental conflict in Direct Preference Optimization (DPO) for unified models, where image generation quality resists alignment while understanding improves.
Mar 19
Paradigm Shift
Mathematically proves that the Transformer architecture is functionally equivalent to a Bayesian Network performing loopy belief propagation.
Mar 19
Open Release
Democratizes dexterous robot data collection by enabling high-fidelity 21-DoF teleoperation using only a standard smartphone.
Mar 19
Breaks Assumption
Reveals that cross-lingual knowledge failure in large reasoning models is primarily a script-translation barrier rather than a linguistic or reasoning deficit.
Mar 19
Scaling Insight
Shows that 'Mid-Training' on high-quality reasoning data is the primary driver of model capability, whereas RL only succeeds as a sparse refinement step.
Mar 19
New Capability
Leverages cross-lingual inconsistencies to pinpoint exactly which experts in a Mixture-of-Experts (MoE) model store specific factual knowledge.
Mar 19
Breaks Assumption
Exposes 'hidden clones' in VLM ensembles, where models from the same family share correlated errors that naive voting mechanisms fail to detect.
Mar 19
New Capability
Proposes REAL, a Reinforcement Learning framework tailored for regression and ordinal scoring rather than simple binary accuracy.
Mar 19
New Capability
Introduces a framework for LLM agents to autonomously evolve their policies and skill libraries during system idle time without retraining downtime.
Mar 19
Efficiency Breakthrough
A backbone-agnostic denoising objective that allows small GNNs to outperform large models pretrained on much larger supervised datasets in physical sciences.
Mar 19
Paradigm Shift
Achieves high-performance online continual learning without the massive memory overhead of traditional experience replay buffers.
Mar 19
Breaks Assumption
Internal activation probing detects LLM 'rationalization' more reliably than monitoring the model's own Chain-of-Thought (CoT).
Mar 19
Efficiency Breakthrough
A dynamic data pruning framework that cuts dense retriever training time by 50% while actually improving retrieval accuracy.
Mar 19
New Capability
Automates the generation of synthetic machine learning challenges to train agents that can genuinely learn research skills from doing.
Mar 19
Breaks Assumption
Alignment processes induce a 'normative bias' that makes LLMs worse at predicting real human behavior in strategic scenarios.
Mar 19
New Capability
Enables reliable, training-free emotion steering in speech-generative audio models via direct manipulation of specific emotion-sensitive neurons.
Mar 19
Paradigm Shift
A formal, graph-native memory architecture that treats agent memory as a versioned asset, dramatically outperforming Gemini 2.5 Pro on complex recall.
Mar 19
New Capability
A framework to quantify and fix 'task steerability,' the common failure of robots to respond to new instructions while mid-task.
Mar 19
Efficiency Breakthrough
Achieves up to a 1,000x gain in RLHF data efficiency by using information-directed exploration and epistemic neural networks.
Mar 19
Efficiency Breakthrough
Introduces a reward framework that reduces LLM reasoning verbosity by optimizing for 'Information Density' via entropy reduction per step.
Mar 19
Paradigm Shift
Shifts retrieval from static contrastive vector alignment to dynamic reasoning trajectories using a generative model (T1) and GRPO.
Mar 19
Breaks Assumption
Identifies that reasoning-induced safety failures occur *during* Chain-of-Thought and proposes a shift to 'decide-then-reason' architectures.
Mar 19
Efficiency Breakthrough
Generates 9 million grid points of 3D spatiotemporal physical fields in seconds, a 10,000x speedup over traditional physics simulations.
Mar 19
New Capability
Proposes a world model that jointly generates appearance and binocular geometry using an epipolar-aware attention mechanism.
Mar 19
Open Release
Introduces FineViT and a 450M local caption dataset to solve the 'coarse perception' bottleneck in current CLIP-based encoders.
Mar 19
Paradigm Shift
Provides a sheaf-theoretic proof that local causal consistency in generative models does not guarantee global counterfactual coherence.
Mar 19
Efficiency Breakthrough
Replaces quadratic self-attention with $O(N \log N)$ phase-native coupling for time-series, enabling massive context windows.
Mar 19
New Capability
Introduces a paradigm for vision-language navigation that uses ubiquitously available semantic floor plans as global spatial priors.
Mar 19
New Capability
Embeds invisible, agent-specific 'watermarks' into token distributions to enable forensic attribution and topology reconstruction in multi-agent systems.
Mar 19
Efficiency Breakthrough
Achieves an 80% reduction in Chain-of-Thought (CoT) tokens while slightly increasing reasoning accuracy.
Mar 19
Efficiency Breakthrough
Extends LLM context from 32K to 128K by teaching models to selectively skip global attention for ~80% of tokens.
Mar 19
New Capability
Reduces hallucinations by teaching models 'epistemological humility'—the ability to admit they don't know something—using synthetic non-existent terms.
Mar 19
Breaks Assumption
Develops a zero-watermarking framework that survives AI editing by leveraging invariant relations between image patches.
Mar 19
Paradigm Shift
Unifies large-scale search, recommendation, and reasoning into a single self-contained LLM by treating item IDs as a distinct modality.
Mar 19
Scaling Insight
Video fine-tuning consistently degrades static image understanding in multimodal LLMs, revealing a zero-sum trade-off between spatial and temporal capabilities.
Mar 19
New Capability
Introduces a Prompt-Free Universal Region Proposal Network (PF-RPN) that identifies objects in any domain without needing text or image exemplars.
Mar 19
New Capability
FrescoDiffusion enables coherent, 4K image-to-video generation using a training-free, tiled diffusion method with precomputed latent priors.
Mar 19
Efficiency Breakthrough
Knowledge-Aware Active Learning (KA2L) uses latent space probing to identify what an LLM doesn't know and generates targeted synthetic questions.
Mar 19
Breaks Assumption
Dense retrieval architectures are fundamentally flawed at detecting negation and contradictions due to 'Semantic Collapse' in vector space.
Mar 19
Paradigm Shift
Edit-As-Act reframes 3D scene editing as a goal-regressive planning problem using symbolic action languages rather than purely generative pixel manipulation.
Mar 19
Breaks Assumption
ARES demonstrates high-fidelity data reconstruction from large Federated Learning batches without requiring any architectural modifications to the model.
Mar 19
Scaling Insight
Mechanistic probing reveals a directional asymmetry in how LLMs encode hierarchy: hypernymy is redundant and resilient, while hyponymy is fragile and compact.
Mar 19