Best AI papers explained
A podcast by Enoch H. Kang - Fridays
203 Episodes
-
Transformers for In-Context Reinforcement Learning
Published: 5/17/2025 -
Evaluating Large Language Models Across the Lifecycle
Published: 5/17/2025 -
Active Ranking from Human Feedback with DopeWolfe
Published: 5/16/2025 -
Optimal Designs for Preference Elicitation
Published: 5/16/2025 -
Dual Active Learning for Reinforcement Learning from Human Feedback
Published: 5/16/2025 -
Active Learning for Direct Preference Optimization
Published: 5/16/2025 -
Active Preference Optimization for RLHF
Published: 5/16/2025 -
Test-Time Alignment of Diffusion Models without reward over-optimization
Published: 5/16/2025 -
Test-Time Preference Optimization: On-the-Fly Alignment via Iterative Textual Feedback
Published: 5/16/2025 -
GenARM: Reward Guided Generation with Autoregressive Reward Model for Test-time Alignment
Published: 5/16/2025 -
Advantage-Weighted Regression: Simple and Scalable Off-Policy RL
Published: 5/16/2025 -
Can RLHF be More Efficient with Imperfect Reward Models? A Policy Coverage Perspective
Published: 5/16/2025 -
Transformers can be used for in-context linear regression in the presence of endogeneity
Published: 5/15/2025 -
Bayesian Concept Bottlenecks with LLM Priors
Published: 5/15/2025 -
In-Context Parametric Inference: Point or Distribution Estimators?
Published: 5/15/2025 -
Enough Coin Flips Can Make LLMs Act Bayesian
Published: 5/15/2025 -
Bayesian Scaling Laws for In-Context Learning
Published: 5/15/2025 -
Posterior Mean Matching Generative Modeling
Published: 5/15/2025 -
Can Generative AI Solve Your In-Context Learning Problem? A Martingale Perspective
Published: 5/15/2025 -
Dynamic Search for Inference-Time Alignment in Diffusion Models
Published: 5/15/2025
Men know other men best. Women know other women best. And yes, perhaps AIs know other AIs best. AI explains what you should know about this week's AI research progress.