You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* add orb decov to circleci config.yml
* Add codecov badge to Readme
* Revert "[BugFix] Changing the dm_control import to fail if not installed (#515)"
This reverts commit d194735.
* codecov coverage w/o orb in circleci
* Revert "Revert "[BugFix] Changing the dm_control import to fail if not installed (#515)""
This reverts commit d0dc7de.
* [CI] generation of coverage reports (#534)
* update test scripts to add coverage
* update test scripts to add coverage
Co-authored-by: Silvestre Bahi <silvestrebahi@fb.com>
* [CI] Add xml coverage reports for codecov (#537)
* update test scripts to add coverage
* update test scripts to add coverage
* generate xml file for coverage
* Update run_test.sh
lint end of file
* Update run_test.sh
lint end of file
* Update run_test.sh
lint end of file
Co-authored-by: Silvestre Bahi <silvestrebahi@fb.com>
* permissions
* permissions
Co-authored-by: Silvestre Bahi <silvestrebahi@fb.com>
Co-authored-by: silvestrebahi <silvestre.bahi@gmail.com>
@@ -32,7 +33,7 @@ one object to another without friction.
32
33
algorithms. For instance, here's how to code a rollout in TorchRL:
33
34
<details>
34
35
<summary>Code</summary>
35
-
36
+
36
37
```diff
37
38
- obs, done = env.reset()
38
39
+ tensordict = env.reset()
@@ -57,7 +58,7 @@ algorithms. For instance, here's how to code a rollout in TorchRL:
57
58
TensorDict abstracts away the input / output signatures of the modules, env, collectors, replay buffers and losses of the library, allowing its primitives
58
59
to be easily recycled across settings.
59
60
Here's another example of an off-policy training loop in TorchRL (assuming that a data collector, a replay buffer, a loss and an optimizer have been instantiated):
60
-
61
+
61
62
```diff
62
63
- for i, (obs, next_obs, action, hidden_state, reward, done) in enumerate(collector):
63
64
+ for i, tensordict in enumerate(collector):
@@ -73,7 +74,7 @@ algorithms. For instance, here's how to code a rollout in TorchRL:
73
74
optim.zero_grad()
74
75
```
75
76
Again, this training loop can be re-used across algorithms as it makes a minimal number of assumptions about the structure of the data.
76
-
77
+
77
78
TensorDict supports multiple tensor operations on its device and shape
78
79
(the shape of TensorDict, or its batch size, is the common arbitrary N first dimensions of all its contained tensors):
79
80
```python
@@ -96,11 +97,11 @@ algorithms. For instance, here's how to code a rollout in TorchRL:
96
97
</details>
97
98
98
99
Check our [TensorDict tutorial](tutorials/tensordict.ipynb) for more information.
99
-
100
+
100
101
- An associated [`TensorDictModule` class](torchrl/modules/tensordict_module/common.py) which is [functorch](https://github.com/pytorch/functorch)-compatible!
0 commit comments