526 Episoade

  1. Trajectory Bellman Residual Minimization: A Simple Value-Based Method for LLM Reasoning

    Publicat: 25.05.2025
  2. Understanding Best-of-N Language Model Alignment

    Publicat: 25.05.2025
  3. Maximizing Acquisition Functions for Bayesian Optimization - and its relation to Gradient Descent

    Publicat: 24.05.2025
  4. Bayesian Prompt Ensembles: Model Uncertainty Estimation for Black-Box Large Language Models

    Publicat: 24.05.2025
  5. Prompting Strategies for Enabling Large Language Models to Infer Causation from Correlation

    Publicat: 24.05.2025
  6. The Parallel Knowledge Gradient Method for Batch Bayesian Optimization

    Publicat: 24.05.2025
  7. FunBO: Discovering Acquisition Functions for Bayesian Optimization with FunSearch

    Publicat: 24.05.2025
  8. Automated Social Science: A Structural Causal Model-Based Approach

    Publicat: 24.05.2025
  9. Causal Interpretation of Transformer Self-Attention

    Publicat: 24.05.2025
  10. A Causal World Model Underlying Next Token Prediction: Exploring GPT in a Controlled Environment

    Publicat: 24.05.2025
  11. Trace is the Next AutoDiff: Generative Optimization with Rich Feedback, Execution Traces, and LLMs

    Publicat: 24.05.2025
  12. Adaptive Inference-Time Compute: LLMs Can Predict if They Can Do Better, Even Mid-Generation

    Publicat: 24.05.2025
  13. Prompts from Reinforcement Learning (PRL)

    Publicat: 24.05.2025
  14. Logits are All We Need to Adapt Closed Models

    Publicat: 24.05.2025
  15. Large Language Models Are (Bayesian) Latent Variable Models: Explaining and Finding Good Demonstrations for In-Context Learning

    Publicat: 23.05.2025
  16. Inference-Time Intervention: Eliciting Truthful Answers from a Language Model

    Publicat: 23.05.2025
  17. From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models

    Publicat: 23.05.2025
  18. LLM In-Context Learning as Kernel Regression

    Publicat: 23.05.2025
  19. Personalizing LLMs via Decode-Time Human Preference Optimization

    Publicat: 23.05.2025
  20. Almost Surely Safe LLM Inference-Time Alignment

    Publicat: 23.05.2025

15 / 27

Cut through the noise. We curate and break down the most important AI papers so you don’t have to.

Visit the podcast's native language site