@@ -182,7 +182,7 @@ int main(int argc, char *argv[])
182182 storage.get_masks ()[step]);
183183 }
184184 auto actions_tensor = act_result[1 ].cpu ().to (torch::kFloat );
185- float *actions_array = actions_tensor.data <float >();
185+ float *actions_array = actions_tensor.data_ptr <float >();
186186 std::vector<std::vector<float >> actions (num_envs);
187187 for (int i = 0 ; i < num_envs; ++i)
188188 {
@@ -218,7 +218,7 @@ int main(int argc, char *argv[])
218218 returns_rms->update (returns);
219219 reward_tensor = torch::clamp (reward_tensor / torch::sqrt (returns_rms->get_variance () + 1e-8 ),
220220 -reward_clip_value, reward_clip_value);
221- rewards = std::vector<float >(reward_tensor.data <float >(), reward_tensor.data <float >() + reward_tensor.numel ());
221+ rewards = std::vector<float >(reward_tensor.data_ptr <float >(), reward_tensor.data_ptr <float >() + reward_tensor.numel ());
222222 real_rewards = flatten_vector (step_result->real_reward );
223223 dones_vec = step_result->done ;
224224 }
@@ -233,7 +233,7 @@ int main(int argc, char *argv[])
233233 returns_rms->update (returns);
234234 reward_tensor = torch::clamp (reward_tensor / torch::sqrt (returns_rms->get_variance () + 1e-8 ),
235235 -reward_clip_value, reward_clip_value);
236- rewards = std::vector<float >(reward_tensor.data <float >(), reward_tensor.data <float >() + reward_tensor.numel ());
236+ rewards = std::vector<float >(reward_tensor.data_ptr <float >(), reward_tensor.data_ptr <float >() + reward_tensor.numel ());
237237 real_rewards = flatten_vector (step_result->real_reward );
238238 dones_vec = step_result->done ;
239239 }
0 commit comments