review:2024-11_beyond_the_boundaries_of_proximal_policy_optimization

2024-11 Beyond the Boundaries of Proximal Policy Optimization

with torch.no_grad():
    mf = 0.3
    lr = 0.7
    for param_old, param, m in zip(model_old.parameters(), model.parameters(), momentum):
        grad = param - param_old
        m.copy_(mf * m + grad)
        param.copy_(param_old + lr * (m + mf * grad))
        param_old.copy_(param)
review/2024-11_beyond_the_boundaries_of_proximal_policy_optimization.txt · 마지막으로 수정됨: 2025/01/14 10:40 저자 rex8312