• 내용으로 건너뛰기

Out of the Box

사용자 도구

  • 로그인

사이트 도구

  • 최근 바뀜
  • 미디어 관리자
  • 사이트맵
추적: • 2021-01_zero-offload_democratizing_billion-scale_model_training • 2020-01_pcgrl_procedural_content_generation_via_reinforcement_learning • 2024-02_craftax_a_lightning-fast_benchmark_for_open-ended_reinforcement_learning • update-rc.d • shared_experience_actor-critic_for_multi-agent_reinforcement_learning • 2020-01_bringing_stories_alive_generating_interactive_fiction_worlds • policy_optimization_by_genetic_distillation • 2021-04_rapid_exploration_for_open-world_navigation_with_latent_goal_models • 2021-07_generative_adversarial_networks_in_time_series_a_survey_and_taxonomy • 2024-01_bridging_state_and_history_representations_understanding_self-predictive_rl

llm학습

TAG: llm학습

  • 2023-01 GPT in 60 Lines of NumPy
2024/01/22 00:18Hyunsoo Park
  • 2023-12 DiLoCo: Distributed Low-Communication Training of Language Models
2024/01/10 23:19Hyunsoo Park
  • 2023-12 Unicron: Economizing Self-Healing LLM Training at Scale
2024/01/10 23:21Hyunsoo Park
  • 2024-01 Asynchronous Local-SGD Training for Language Modeling
2024/01/18 05:07Hyunsoo Park
  • 2024-01 DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
2024/01/18 00:53Hyunsoo Park
  • 2024-01 Mixtral of Experts
2024/01/10 23:51Hyunsoo Park
  • 2024-01 TinyLlama: An Open-Source Small Language Model
2024/01/07 17:40Hyunsoo Park
  • 2024-07 Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients
2024/07/15 00:50Hyunsoo Park
  • 2024-10 Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
2024/10/08 00:21Hyunsoo Park

문서 도구

  • 문서 보기
  • 이전 판
  • 역링크
  • Fold/unfold all
  • 맨 위로
별도로 명시하지 않을 경우, 이 위키의 내용은 다음 라이선스에 따라 사용할 수 있습니다: CC Attribution-Noncommercial-Share Alike 4.0 International
CC Attribution-Noncommercial-Share Alike 4.0 International Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki