Skip to content

[Feature Request] Support for alpha and beta parameters' schedule in torchrl.data.PrioritizedReplayBuffer #1575

Closed
@Mad-Chuck

Description

@Mad-Chuck

Motivation

Since in the original PER paper the parameter beta is changing its value during the training, it would be desirable to be able to perform similar experiments with torch-rl's Prioritized Experience Replay.

Solution

Being able to manually change beta or both alpha and beta by calling a method on a replay buffer or modifying its properties.

Alternatives

OpenAI baseline's implementation does it by providing an additional parameter beta during sampling (while alpha is fixed). It would be also possible to implement this with a scheduler (similar to lr schedules, like in torch.optim)

Checklist

  • I have checked that there is no similar issue in the repo (required)

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions