Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rationale for training only on 10% of the buffer ? #7

Open
sabeaussan opened this issue Feb 25, 2025 · 0 comments
Open

Rationale for training only on 10% of the buffer ? #7

sabeaussan opened this issue Feb 25, 2025 · 0 comments

Comments

@sabeaussan
Copy link

When defining the batch size in training_go.py, you comment : 'To avoid overfitting, we want to make sure the agent only sees ~10% of samples in the replay over one checkpoint.' 'That is, batch_size * ckpt_interval <= replay_capacity * 0.1'. Can you expand on this choice ? Intuitively training on a small sample of the buffer will foster overfitting rather than prevent it doesn't it ? Can you explain more in details this choice please :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant