-
Notifications
You must be signed in to change notification settings - Fork 3
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Started on end-to-end integration test.
- Loading branch information
1 parent
175a58b
commit 8a7087d
Showing
1 changed file
with
159 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,159 @@ | ||
"""Tests the full end-to-end extraction process.""" | ||
|
||
import rootutils | ||
|
||
root = rootutils.setup_root(__file__, dotenv=True, pythonpath=True, cwd=True) | ||
|
||
import subprocess | ||
import tempfile | ||
from pathlib import Path | ||
|
||
import polars as pl | ||
from polars.testing import assert_frame_equal | ||
|
||
SUBJECTS_CSV = """ | ||
MRN,dob,eye_color,height | ||
1195293,06/20/1978,BLUE,164.6868838269085 | ||
239684,12/28/1980,BROWN,175.271115221764 | ||
1500733,07/20/1986,BROWN,158.60131573580904 | ||
814703,03/28/1976,HAZEL,156.48559093209357 | ||
754281,12/19/1988,BROWN,166.22261567137025 | ||
68729,03/09/1978,HAZEL,160.3953106166676 | ||
""" | ||
|
||
ADMIT_VITALS_CSV = """ | ||
MRN,admit_date,disch_date,department,vitals_date,HR,temp | ||
239684,"05/11/2010, 17:41:51","05/11/2010, 19:27:19",CARDIAC,"05/11/2010, 18:57:18",112.6,95.5 | ||
754281,"01/03/2010, 06:27:59","01/03/2010, 08:22:13",PULMONARY,"01/03/2010, 06:27:59",142.0,99.8 | ||
814703,"02/05/2010, 05:55:39","02/05/2010, 07:02:30",ORTHOPEDIC,"02/05/2010, 05:55:39",170.2,100.1 | ||
239684,"05/11/2010, 17:41:51","05/11/2010, 19:27:19",CARDIAC,"05/11/2010, 18:25:35",113.4,95.8 | ||
68729,"05/26/2010, 02:30:56","05/26/2010, 04:51:52",PULMONARY,"05/26/2010, 02:30:56",86.0,97.8 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 20:12:31",112.5,99.8 | ||
1500733,"06/03/2010, 14:54:38","06/03/2010, 16:44:26",ORTHOPEDIC,"06/03/2010, 16:20:49",90.1,100.1 | ||
239684,"05/11/2010, 17:41:51","05/11/2010, 19:27:19",CARDIAC,"05/11/2010, 17:48:48",105.1,96.2 | ||
239684,"05/11/2010, 17:41:51","05/11/2010, 19:27:19",CARDIAC,"05/11/2010, 17:41:51",102.6,96.0 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 19:25:32",114.1,100.0 | ||
1500733,"06/03/2010, 14:54:38","06/03/2010, 16:44:26",ORTHOPEDIC,"06/03/2010, 14:54:38",91.4,100.0 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 20:41:33",107.5,100.4 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 20:24:44",107.7,100.0 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 19:45:19",119.8,99.9 | ||
1195293,"06/20/2010, 19:23:52","06/20/2010, 20:50:04",CARDIAC,"06/20/2010, 19:23:52",109.0,100.0 | ||
1500733,"06/03/2010, 14:54:38","06/03/2010, 16:44:26",ORTHOPEDIC,"06/03/2010, 15:39:49",84.4,100.3 | ||
""" | ||
|
||
EVENT_CFGS_YAML = """ | ||
patient_id_col: MRN | ||
subjects: | ||
eye_color: | ||
code: | ||
- EYE_COLOR | ||
- col(eye_color) | ||
timestamp: null | ||
height: | ||
code: HEIGHT | ||
timestamp: null | ||
numerical_value: height | ||
dob: | ||
code: DOB | ||
timestamp: col(dob) | ||
admit_vitals: | ||
admissions: | ||
code: | ||
- ADMISSION | ||
- col(department) | ||
timestamp: col(admit_date) | ||
timestamp_format: "%m/%d/%Y, %H:%M:%S" | ||
discharge: | ||
code: DISCHARGE | ||
timestamp: col(disch_date) | ||
timestamp_format: "%m/%d/%Y, %H:%M:%S" | ||
HR: | ||
code: HR | ||
timestamp: col(vitals_date) | ||
timestamp_format: "%m/%d/%Y, %H:%M:%S" | ||
numerical_value: HR | ||
temp: | ||
code: TEMP | ||
timestamp: col(vitals_date) | ||
timestamp_format: "%m/%d/%Y, %H:%M:%S" | ||
numerical_value: temp | ||
""" | ||
|
||
|
||
def test_command(script: str, hydra_kwargs: dict[str, str], test_name: str): | ||
command_parts = ["python", script] + [f"{k}={v}" for k, v in hydra_kwargs.items()] | ||
command_out = subprocess.run(" ".join(command_parts), shell=True, capture_output=True) | ||
stderr = command_out.stderr.decode() | ||
stdout = command_out.stdout.decode() | ||
if command_out.returncode != 0: | ||
raise AssertionError(f"{test_name} failed!\nstderr:\n{stderr}\nstdout:\n{stdout}") | ||
|
||
|
||
def test_extraction(): | ||
with tempfile.TemporaryDirectory() as d: | ||
raw_cohort_dir = Path(d) / "raw_cohort" | ||
MEDS_cohort_dir = Path(d) / "MEDS_cohort" | ||
|
||
# Create the directories | ||
raw_cohort_dir.mkdir() | ||
MEDS_cohort_dir.mkdir() | ||
|
||
subjects_csv = raw_cohort_dir / "subjects.csv" | ||
admit_vitals_csv = raw_cohort_dir / "admit_vitals.csv" | ||
event_cfgs_yaml = raw_cohort_dir / "event_cfgs.yaml" | ||
|
||
# Write the CSV files | ||
subjects_csv.write_text(SUBJECTS_CSV) | ||
admit_vitals_csv.write_text(ADMIT_VITALS_CSV) | ||
|
||
# Mix things up -- have one CSV be also in parquet format. | ||
admit_vitals_parquet = raw_cohort_dir / "admit_vitals.parquet" | ||
pl.read_csv(admit_vitals_csv).write_parquet(admit_vitals_parquet, use_pyarrow=True) | ||
|
||
# Write the event config YAML | ||
event_cfgs_yaml.write_text(EVENT_CFGS_YAML) | ||
|
||
# Run the extraction script | ||
# 1. Sub-shard the data (this will be a null operation in this case, but it is worth doing just in | ||
# case. | ||
# 2. Collect the patient splits. | ||
# 3. Extract the events and sub-shard by patient. | ||
# 4. Merge to the final output. | ||
|
||
extraction_config_kwargs = { | ||
"raw_cohort_dir": str(raw_cohort_dir.resolve()), | ||
"MEDS_cohort_dir": str(MEDS_cohort_dir.resolve()), | ||
"event_conversion_config_fp": str(event_cfgs_yaml.resolve()), | ||
"split_fracs.train": 4 / 6, | ||
"split_fracs.tuning": 1 / 6, | ||
"split_fracs.held_out": 1 / 6, | ||
"row_chunksize": 10, | ||
"n_patients_per_shard": 2, | ||
"hydra.verbose": True, | ||
} | ||
|
||
# Step 1: Sub-shard the data | ||
test_command("./scripts/extraction/shard_events.py", extraction_config_kwargs, "shard_events") | ||
|
||
subsharded_dir = MEDS_cohort_dir / "sub_sharded" | ||
|
||
out_files = list(subsharded_dir.glob("*/*.parquet")) | ||
assert len(out_files) == 3, f"Expected 3 output files, got {len(out_files)}." | ||
|
||
# Checking specific out files: | ||
# 1. subjects.parquet | ||
subjects_out = subsharded_dir / "subjects" / "[0-7).parquet" | ||
assert subjects_out.is_file(), f"Expected {subjects_out} to exist." | ||
|
||
assert_frame_equal(pl.read_parquet(subjects_out, glob=False), pl.read_csv(subjects_csv)) | ||
|
||
# 2. admit_vitals.parquet | ||
df_chunks = [] | ||
for chunk in ["[0-10)", "[10-16)"]: | ||
admit_vitals_chunk_fp = subsharded_dir / "admit_vitals" / f"{chunk}.parquet" | ||
assert admit_vitals_chunk_fp.is_file(), f"Expected {admit_vitals_chunk_fp} to exist." | ||
|
||
df_chunks.append(pl.read_parquet(admit_vitals_chunk_fp, glob=False)) | ||
assert_frame_equal(pl.concat(df_chunks), pl.read_csv(admit_vitals_csv)) | ||
|
||
raise NotImplementedError("Finish the test_extraction function.") |