Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Test hook support #263

Merged
merged 66 commits into from
Nov 7, 2024
Merged

Test hook support #263

merged 66 commits into from
Nov 7, 2024

Conversation

TaekyungHeo
Copy link
Member

@TaekyungHeo TaekyungHeo commented Oct 14, 2024

Summary

This PR introduces hooks to CloudAI. Hooks are tests that run either before or after each test in a test scenario. They are defined globally within a test scenario and are automatically executed for each test. There are two types of hooks: pre-tests and post-tests. Pre-tests run before the tests, while post-tests are executed after the tests. Multiple pre-tests and post-tests can be specified in each scenario.

An example of how hooks are defined within a test scenario:

name = "nccl-test"

pre_test = "nccl_test_pre"
post_test = "nccl_test_post"

[[Tests]]
id = "Tests.1"
test_name = "nccl_test_all_reduce"
num_nodes = "2"
time_limit = "00:20:00"

[[Tests]]
id = "Tests.2"
test_name = "nccl_test_all_gather"
num_nodes = "2"
time_limit = "00:20:00"
  [[Tests.dependencies]]
  type = "start_post_comp"
  id = "Tests.1"

You can see the pre_test and post_test fields. These are used to look up the corresponding hook file. A hook file is a separate test scenario file as shown below:

name = "nccl_test_pre"

[[Tests]]
id = "Tests.1"
test_name = "nccl_test_all_reduce"
time_limit = "00:20:00"

If any of the tests in the pre-test fail, the main test or the post-test will not run. In other words, the main test and post-test run conditionally when the pre-test is successful. The tests in hooks have time limits, just as tests in the main scenario do. Output files should be stored in the output directory, in a subdirectory called "pre_test" or "post_test," following a proper directory hierarchy. Hooks are not supported for NeMo 1.0 (NeMo launcher).

Note

  • Idea
    • We may need to generate reports from plugins.
    • We may need to consider the performance impact of plugins.
    • Dependencies are not implemented for now.

Test Plan

  1. CI passes
  2. Manual run
    2.1 Success
$ cloudai run --system-config ~/cloudaix-main/conf/common/system/israel_1.toml --tests-dir conf/common/test --test-scenario conf/common/test_scenario/nccl_tes
t.toml
/.autodirect/mswg2/E2E/theo/venv/lib/python3.10/site-packages/requests/__init__.py:102: RequestsDependencyWarning: urllib3 (1.26.19) or chardet (5.2.0)/charset_normalizer (2.0.12) doesn't match a supported ver
sion!
  warnings.warn("urllib3 ({}) or chardet ({})/charset_normalizer ({}) doesn't match a supported "                                                                                                                
[INFO] System Name: Israel-1                                                                            
[INFO] Scheduler: slurm                                                                                 
[INFO] Test Scenario Name: nccl-test                                                                    
[INFO] Checking if test templates are installed.                                                        
[INFO] Test Scenario: nccl-test                                                                         

Section Name: Tests.1                                                                                   
  Test Name: nccl_test_all_reduce                                                                       
  Description: all_reduce                                                                               
  No dependencies                                                                                       
[INFO] Initializing Runner [RUN] mode                                                                   
[INFO] Creating SlurmRunner                                                                             
[INFO] Starting test scenario execution.                                                                
[INFO] Starting test: Tests.1                                                                           
[INFO] Running test: Tests.1                                                                            
[INFO] Executing command for test Tests.1: sbatch /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-05-46/Tests.1/0/cloudai_sbatch_script.sh
[INFO] Job completed: Tests.1
[INFO] All test scenario results stored at: /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-05-46
[INFO] All test scenario execution attempts are complete. Please review the 'debug.log' file to confirm successful completion or to identify any issues.
$ cd /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-05-46/Tests.1/0
$ ls
cloudai_sbatch_script.sh  epilogue  prologue  stderr.txt  stdout.txt

$ ls prologue/nccl_test_all_reduce/
stderr.txt  stdout.txt

$ ls epilogue/nccl_test_all_gather/
stderr.txt  stdout.txt

2.2 Failure

$ cloudai run --system-config ~/cloudaix-main/conf/common/system/israel_1.toml --tests-dir conf/common/test --test-scenario conf/common/test_scenario/nccl_test.toml
/.autodirect/mswg2/E2E/theo/venv/lib/python3.10/site-packages/requests/__init__.py:102: RequestsDependencyWarning: urllib3 (1.26.19) or chardet (5.2.0)/charset_normalizer (2.0.12) doesn't match a supported version!
  warnings.warn("urllib3 ({}) or chardet ({})/charset_normalizer ({}) doesn't match a supported "
[INFO] System Name: Israel-1
[INFO] Scheduler: slurm
[INFO] Test Scenario Name: nccl-test
[INFO] Checking if test templates are installed.
[INFO] Test Scenario: nccl-test

Section Name: Tests.1
  Test Name: nccl_test_all_reduce
  Description: all_reduce
  No dependencies
[INFO] Initializing Runner [RUN] mode
[INFO] Creating SlurmRunner
[INFO] Starting test scenario execution.
[INFO] Starting test: Tests.1
[INFO] Running test: Tests.1
[INFO] Executing command for test Tests.1: sbatch /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-16-25/Tests.1/0/cloudai_sbatch_script.sh
[ERROR] Job 383928 for test Tests.1 failed: Missing success indicators in /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-16-25/Tests.1/0/stdout.txt: '# Out of bounds values', '# Avg bus bandwidth'. These keywords are expected to be present in stdout.txt, usually towards the end of the file. Please review the NCCL test output and errors in the file. Ensure the NCCL test ran to completion. You can run the generated sbatch script manually and check if /auto/e2e/israel1/workload_results/nccl-test_2024-10-25_22-16-25/Tests.1/0/stdout.txt is created and contains the expected keywords. If the issue persists, contact the system administrator.
[INFO] Terminating all jobs...
[INFO] All jobs have been killed.

Copy link
Contributor

@amaslenn amaslenn left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

For existing prologue we use real NCCL run. In your examples it seems that we are switching to some predefined commands.

  1. How are we going to generate it?
  2. Will that cover our needs? cc @srivatsankrishnan

I do have some code related notes, but let's leave it for later discussion.

@TaekyungHeo
Copy link
Member Author

@amaslenn

How are we going to generate it?

Yes, it is one of the main design choices that we need to make.

@amaslenn
Copy link
Contributor

Yes, it is one of the main design choices that we need to make.

Can we rely on existing mechanisms? Each plugin will be defined as a regular Test TOML, meaning we can generate a CLI for it for a particular system. This is what we do now and it seems to cover all our needs for this feature.

@TaekyungHeo TaekyungHeo force-pushed the plugin-jan branch 15 times, most recently from 7594c19 to 852fee8 Compare October 24, 2024 19:54
@TaekyungHeo
Copy link
Member Author

@amaslenn , I ran verify-configs and got this warning

$ cloudai verify-configs conf
[WARNING] Test configuration directory not provided, using all found test TOMLs in the specified directory.
[INFO] Checked systems: 3, all passed
[INFO] Checked tests: 40, all passed
[WARNING] System configuration not provided, mocking it.
[WARNING] Prologue 'nccl_test_prologue' not found in plugin mapping. Ensure that a proper plugin directory is set under the working directory.
[WARNING] Epilogue 'nccl_test_epilogue' not found in plugin mapping. Ensure that a proper plugin directory is set under the working directory.
[INFO] Checked scenarios: 9, all passed
[INFO] Checked 52 configuration files, all passed

tests/test_parser.py Outdated Show resolved Hide resolved
tests/test_parser.py Show resolved Hide resolved
src/cloudai/parser.py Outdated Show resolved Hide resolved
src/cloudai/_core/test_scenario_parser.py Outdated Show resolved Hide resolved
src/cloudai/parser.py Outdated Show resolved Hide resolved
src/cloudai/parser.py Outdated Show resolved Hide resolved
@TaekyungHeo TaekyungHeo changed the title Plugin support Test hook support Nov 4, 2024
@TaekyungHeo
Copy link
Member Author

TaekyungHeo commented Nov 4, 2024

do you have any suggestions for fixing the CI error in the verification function?

Let's always add hooks into lookup, we always now where it is:

...
err, tomls = expand_file_list(root, glob="**/*.toml")
err, hook_tomls = expand_file_list(HOOKS_DIR, glob="**/*.toml")
tomls += hook_tomls
...

Let's also change if "conf" in toml_file.parts and "hook" in toml_file.parts to == with hooks' dirs constants in load_tomls_by_type.

src/cloudai/_core/test_scenario_parser.py Show resolved Hide resolved
src/cloudai/parser.py Outdated Show resolved Hide resolved
src/cloudai/cli/handlers.py Show resolved Hide resolved
src/cloudai/cli/handlers.py Outdated Show resolved Hide resolved
Copy link
Contributor

@srivatsankrishnan srivatsankrishnan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

As discussed in the call, Taekyung mentioned that there he tested with different nccl test for both pre and post scenarios. This will be a continuing feature to cover for other use cases.

@TaekyungHeo TaekyungHeo merged commit 95b3681 into NVIDIA:main Nov 7, 2024
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature Jan25 Jan'25 release feature
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants