Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RLlib] Cleanup examples folder 13: Add READMEs to folder and all sub-folders. #45382

Open
wants to merge 14 commits into
base: master
Choose a base branch
from

Conversation

sven1977
Copy link
Contributor

Cleanup examples folder 13: Add READMEs to folder and all sub-folders.

Why are these changes needed?

Related issue number

Checks

  • I've signed off every commit(by using the -s flag, i.e., git commit -s) in this PR.
  • I've run scripts/format.sh to lint the changes in this PR.
  • I've included any doc changes needed for https://docs.ray.io/en/master/.
    • I've added any new APIs to the API Reference. For example, if I added a
      method in Tune, I've added it in doc/source/tune/api/ under the
      corresponding .rst file.
  • I've made sure the tests are passing. Note that there might be a few flaky tests, see the recent failures at https://flakey-tests.ray.io/
  • Testing Strategy
    • Unit tests
    • Release tests
    • This PR is not tested :(

Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
…p_examples_folder_11_fractional_gpus

# Conflicts:
#	rllib/utils/test_utils.py
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Signed-off-by: sven1977 <svenmika1977@gmail.com>
@sven1977 sven1977 added rllib RLlib related issues rllib-docs-or-examples Issues related to RLlib documentation or rllib/examples rllib-newstack rllib-oldstack-cleanup Issues related to cleaning up classes, utilities on the old API stack labels May 16, 2024
Signed-off-by: sven1977 <svenmika1977@gmail.com>
Copy link
Contributor

@angelinalg angelinalg left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Just some style nits and a few typos.


num_gpus = 0
num_env_runners = 2
In curriculum learning, the environment can be set by the user to different difficulties
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
In curriculum learning, the environment can be set by the user to different difficulties
In curriculum learning, you can set the environment to different difficulties

num_gpus = 0
num_env_runners = 2
In curriculum learning, the environment can be set by the user to different difficulties
throughout the training process. This allows the algorithm to learn how to solve
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
throughout the training process. This allows the algorithm to learn how to solve
throughout the training process. This setting allows the algorithm to learn how to solve

num_env_runners = 2
In curriculum learning, the environment can be set by the user to different difficulties
throughout the training process. This allows the algorithm to learn how to solve
the actual (final) problem incrementally, by interacting with and exploring in more and
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
the actual (final) problem incrementally, by interacting with and exploring in more and
the actual and final problem incrementally, by interacting with and exploring in more and

throughout the training process. This allows the algorithm to learn how to solve
the actual (final) problem incrementally, by interacting with and exploring in more and
more difficult phases.
Normally, such a curriculum starts with setting the environment to an easy difficulty and
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Normally, such a curriculum starts with setting the environment to an easy difficulty and
Normally, such a curriculum starts with setting the environment to an easy level and

the actual (final) problem incrementally, by interacting with and exploring in more and
more difficult phases.
Normally, such a curriculum starts with setting the environment to an easy difficulty and
then - as training progresses - transitions more and more toward a harder-to-solve difficulty.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
then - as training progresses - transitions more and more toward a harder-to-solve difficulty.
then - as training progresses - transitions more toward a harder-to-solve difficulty.

`num_gpus_per_worker` to 0 (they may be set to 1 by default for your
particular RL algorithm)."""
machine does not have any GPUs, you should set the config keys
`num_gpus_per_learner` and `num_gpus_per_env_runner` to 0 (they may be set to
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
`num_gpus_per_learner` and `num_gpus_per_env_runner` to 0 (they may be set to
`num_gpus_per_learner` and `num_gpus_per_env_runner` to 0. They may be set to

particular RL algorithm)."""
machine does not have any GPUs, you should set the config keys
`num_gpus_per_learner` and `num_gpus_per_env_runner` to 0 (they may be set to
1 by default for your particular RL algorithm)."""
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
1 by default for your particular RL algorithm)."""
1 by default for your particular RL algorithm."""

particular RL algorithm)."""
machine does not have any GPUs, you should set the config keys
`num_gpus_per_learner` and `num_gpus_per_env_runner` to 0 (they may be set to
1 by default for your particular RL algorithm)."""

ERR_MSG_INVALID_ENV_DESCRIPTOR = """The env string you provided ('{}') is:
a) Not a supported/installed environment.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
a) Not a supported/installed environment.
a) Not a supported or installed environment.

@@ -1346,6 +1347,11 @@ def run_rllib_example_script_experiment(
tune_callbacks: A list of Tune callbacks to configure with the tune.Tuner.
In case `args.wandb_key` is provided, will append a WandB logger to this
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
In case `args.wandb_key` is provided, will append a WandB logger to this
In case `args.wandb_key` is provided, appends a WandB logger to this

keep_config: Set this to True, if you don't want this utility to change the
given `base_config` in any way and leave it as-is. This is helpful
for example script that want to demonstrate how to set those settings
that are usually taken care of automatically in this function (e.g.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
that are usually taken care of automatically in this function (e.g.
that are usually taken care of automatically in this function (e.g.,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
rllib RLlib related issues rllib-docs-or-examples Issues related to RLlib documentation or rllib/examples rllib-newstack rllib-oldstack-cleanup Issues related to cleaning up classes, utilities on the old API stack
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants