• 내용으로 건너뛰기

Out of the Box

사용자 도구

  • 로그인

사이트 도구

  • 최근 바뀜
  • 미디어 관리자
  • 사이트맵
추적: • duality_a_new_approach_to_reinforcement_learning • 2021-07_perceiver_io_a_general_architecture_for_structured_inputs_outputs • fast_attention • mamba • plasticity • llm_game • 2017-11_neural_discrete_representation_learning • 2024-02_genie_generative_interactive_environments • 2021-07_vector_quantized_models_for_planning • offline_rl

old_topic:offline_rl

문서의 이전 판입니다!


Offline RL

  • 2021-07 Offline Model-Based Optimization via Normalized Maximum Likelihood Estimation
  • 2021-07 Offline Meta-Reinforcement Learning with Online Self-Supervision
  • 2021-07 Conservative Objective Models for Effective Offline Model-Based Optimization
  • 2021-06 Reinforcement Learning as One Big Sequence Modeling Problem
  • 2021-06 Decision Transformer: Reinforcement Learning via Sequence Modeling
  • 2021-04 Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills
  • 2021-03 Amortized Conditional Normalized Maximum Likelihood: Reliable Out of Distribution Uncertainty Estimation
  • 2021-01 What Can I Do Here? Learning New Skills by Imagining Visual Affordances
  • 2020-10 Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
  • 2020-07 Hyperparameter Selection for Offline Reinforcement Learning
  • 2020-07 Accelerating Online Reinforcement Learning with Offline Datasets
  • 2020-06 Conservative Q-Learning for Offline Reinforcement Learning
/var/www/html/data/pages/old_topic/offline_rl.txt · 마지막으로 수정됨: 2024/03/23 02:42 저자 127.0.0.1

문서 도구

  • 원본 보기
  • 이전 판
  • 역링크
  • Fold/unfold all
  • 맨 위로
별도로 명시하지 않을 경우, 이 위키의 내용은 다음 라이선스에 따라 사용할 수 있습니다: CC Attribution-Noncommercial-Share Alike 4.0 International
CC Attribution-Noncommercial-Share Alike 4.0 International Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki