Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RLlib] APPO Training iteration fn. #24545

Merged
merged 25 commits into from
May 17, 2022

Conversation

sven1977
Copy link
Contributor

@sven1977 sven1977 commented May 6, 2022

APPO Training iteration fn.

Why are these changes needed?

Related issue number

Checks

  • I've run scripts/format.sh to lint the changes in this PR.
  • I've included any doc changes needed for https://docs.ray.io/en/master/.
  • I've made sure the tests are passing. Note that there might be a few flaky tests, see the recent failures at https://flakey-tests.ray.io/
  • Testing Strategy
    • Unit tests
    • Release tests
    • This PR is not tested :(

Copy link
Member

@avnishn avnishn left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I have 1 question but otherwise lgtm

rllib/agents/ppo/appo.py Show resolved Hide resolved
@avnishn
Copy link
Member

avnishn commented May 10, 2022

This is all looking pretty good, although is it learning?

@sven1977
Copy link
Contributor Author

Hey @avnishn , please give this another go. I confirmed learning Pong is still good, e.g. on only 7 workers:

(base) ray@ip-172-31-83-222:~/riot_games_atari_benchmarks/ray$ tail ~/output.txt 
Resources requested: 0/8 CPUs, 0/1 GPUs, 0.0/35.56 GiB heap, 0.0/17.78 GiB objects (0.0/1.0 accelerator_type:V100)
Result logdir: /home/ray/ray_results/pong-appo
Number of trials: 1/1 (1 TERMINATED)
+-------------------------------------+------------+---------------------+--------+------------------+---------+----------+----------------------+----------------------+--------------------+
| Trial name                          | status     | loc                 |   iter |   total time (s) |      ts |   reward |   episode_reward_max |   episode_reward_min |   episode_len_mean |
|-------------------------------------+------------+---------------------+--------+------------------+---------+----------+----------------------+----------------------+--------------------|
| APPO_PongNoFrameskip-v4_7fb9b_00000 | TERMINATED | 172.31.83.222:15223 |    154 |          1697.41 | 2931200 |    18.07 |                   20 |                    8 |            7679.47 |
+-------------------------------------+------------+---------------------+--------+------------------+---------+----------+----------------------+----------------------+--------------------+

…a_zero_training_itr

# Conflicts:
#	rllib/agents/slateq/slateq.py
#	rllib/algorithms/ars/README.md
#	rllib/algorithms/ars/ars.py
#	rllib/algorithms/ars/ars_tf_policy.py
#	rllib/algorithms/ars/ars_torch_policy.py
#	rllib/algorithms/ars/tests/test_ars.py
#	rllib/algorithms/es/es.py
#	rllib/algorithms/es/es_tf_policy.py
#	rllib/algorithms/es/es_torch_policy.py
#	rllib/algorithms/es/optimizers.py
#	rllib/algorithms/es/tests/test_es.py
#	rllib/algorithms/es/utils.py
@sven1977 sven1977 merged commit 25001f6 into ray-project:master May 17, 2022
@sven1977 sven1977 deleted the appo_training_itr branch June 2, 2023 20:18
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants