Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Risklist module for production #631

Merged
merged 55 commits into from
Aug 27, 2021
Merged
Show file tree
Hide file tree
Changes from 48 commits
Commits
Show all changes
55 commits
Select commit Hold shift + click to select a range
a563694
listmaking WIP
thcrock Feb 20, 2019
9750c3e
forgot migraton
thcrock Feb 20, 2019
360f8f9
WIP
tweddielin Feb 21, 2019
999a46f
alembic add label_value to list_predictions table
tweddielin Feb 26, 2019
372d9c8
add docstrings
tweddielin Feb 28, 2019
16645bc
move risklist a layer above
tweddielin Mar 13, 2019
914ad76
create risklist module
tweddielin Mar 13, 2019
c92bd8b
__init__lpy
tweddielin Mar 13, 2019
d3c3ba9
fix alembic reversion and replace metta.generate_uuid with filename_f…
tweddielin Mar 13, 2019
0e92fb0
Fix down revision of production schema migration
thcrock Apr 12, 2019
dbd4578
Fix alembic revisions
thcrock Jan 6, 2021
f7d49e5
Enable github checks on this branch too
thcrock Jan 6, 2021
dee930f
Closer to getting tests to run
thcrock Jan 6, 2021
1769b00
Add CLI for risklist
thcrock Jan 8, 2021
52c9ff0
Risklist docs stub
thcrock Jan 8, 2021
173167a
Break up data gathering into experiment and matrix, use pytest fixtur…
thcrock Jan 9, 2021
f6b2d02
Modify schema for list prediction metadata
thcrock Jan 9, 2021
acffa67
fix conflicts and add helper functions for getting imputed features
tweddielin Jan 9, 2021
43c1919
Handle other imputation flag cases, fix tracking indentation error
thcrock Jan 10, 2021
7dfb7e1
Add more tests, fill out doc page
thcrock Jan 11, 2021
cc9fe4a
Fix exception name typo
thcrock Jan 11, 2021
5951565
use timechop and planner to create matrix_metadata for production
tweddielin Jan 15, 2021
537f6c8
retrain and predict forward
tweddielin Apr 15, 2021
b429540
rename to retrain_definition
tweddielin Apr 15, 2021
0045aa5
reusing random seeds from existing models
shaycrk May 8, 2021
9dc3697
fix tests (write experiment to test db)
shaycrk May 10, 2021
da870d5
unit test for reusing model random seeds
shaycrk May 10, 2021
6768ee5
add docstring
shaycrk May 10, 2021
7d6a420
only store random seed in experiment runs
shaycrk May 20, 2021
b8fe6d8
DB migration to remove random seed from experiments table
shaycrk May 20, 2021
8207fcd
debugging
shaycrk May 20, 2021
45c9d68
debug model trainer tests
shaycrk May 21, 2021
a665e7e
debug catwalk utils tests
shaycrk May 21, 2021
ead882b
debug catwalk integration test
shaycrk May 21, 2021
de85f10
use public method
tweddielin May 30, 2021
ad860cd
Merge remote-tracking branch 'origin/kit_rand_seed' into list_making
tweddielin May 31, 2021
40466d5
alembic merge
tweddielin May 31, 2021
83c7385
reuse random seed
tweddielin May 31, 2021
f97089b
use timechop for getting retrain information
tweddielin Jun 30, 2021
6f0af1c
create retrain model hash in retrain level instead of model_trainer l…
tweddielin Jun 30, 2021
42bccaa
move util functions to utils
tweddielin Jun 30, 2021
3ec377f
fix cli and docs
tweddielin Jul 1, 2021
1c4da24
update docs
tweddielin Jul 1, 2021
35bd978
use reconstructed feature dict
tweddielin Jul 1, 2021
9f5a099
add RetrainModel and Retrain
tweddielin Jul 29, 2021
ba84822
remove break point
tweddielin Jul 29, 2021
83e0f66
change experiment_runs to triage_runs
tweddielin Aug 21, 2021
d6f14f5
get retrain_config
tweddielin Aug 22, 2021
d76359b
explicitly include run_type in joins to triage_runs
shaycrk Aug 26, 2021
9698500
DB migration updates
shaycrk Aug 26, 2021
a8a29f1
update argument name in docs
shaycrk Aug 26, 2021
694edcc
ensure correct temporal config is used for predicting forward
shaycrk Aug 27, 2021
583e9bd
debug
shaycrk Aug 27, 2021
815a258
debug
shaycrk Aug 27, 2021
5e183fe
Merge branch 'master' into list_making
shaycrk Aug 27, 2021
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions docs/mkdocs.yml
Original file line number Diff line number Diff line change
Expand Up @@ -120,6 +120,7 @@ nav:
- Using Postmodeling: postmodeling/index.md
- Postmodeling & Crosstabs Configuration: postmodeling/postmodeling-config.md
- Model governance: dirtyduck/ml_governance.md
-Predictlist: predictlist/index.md
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should this be - Predictlist: (I can't remember if YAML caries about the whitespace between the - and list item)?

- Scaling up: dirtyduck/aws_batch.md
- API Reference:
- Audition:
Expand Down
87 changes: 87 additions & 0 deletions docs/sources/predictlist/index.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,87 @@
# Retrain and Predict
Use an existing model group to retrain a new model on all the data up to the current date and then predict forward into the future.

## Examples
Both examples assume you have already run a Triage Experiment in the past, and know these two pieces of information:
1. A `model_group_id` from a Triage model group that you want to use to retrain a model and generate prediction
2. A `today` to generate your predictions on.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we changed the argument to prediction_date rather than today, didn't we?


### CLI
`triage retrainpredict <model_group_id> <today>`

Example:
`triage retrainpredict 30 2021-04-04`

The `retrainpredict` will assume the current path to be the 'project path' to train models and write matrices, but this can be overridden by sending the `--project-path` option

### Python
The `Retrainer` class from `triage.predictlist` module can be used to retrain a model and predict forward.

```python
from triage.predictlist import Retrainer
from triage import create_engine

retrainer = Retrainer(
db_engine=create_engine(<your-db-info>),
project_path='/home/you/triage/project2'
model_group_id=36,
)
retrainer.retrain(today='2021-04-04')
retrainer.predict(today='2021-04-04')

```

## Output
The retrained model is sotred similariy to the matrices created during an Experiment:
- Raw Matrix saved to the matrices directory in project storage
- Raw Model saved to the trained_model directory in project storage
- Retrained Model info saved in a table (triage_metadata.models) where model_comment = 'retrain_2021-04-04 21:19:09.975112'
- Predictions saved in a table (triage_production.predictions)
- Prediction metadata (tiebreaking, random seed) saved in a table (triage_produciton.prediction_metadata)


# Predictlist
If you would like to generate a list of predictions on already-trained Triage model with new data, you can use the 'Predictlist' module.

# Predict Foward with Existed Model
Use an existing model object to generate predictions on new data.

## Examples
Both examples assume you have already run a Triage Experiment in the past, and know these two pieces of information:
1. A `model_id` from a Triage model that you want to use to generate predictions
2. An `as_of_date` to generate your predictions on.

### CLI
`triage predictlist <model_id> <as_of_date>`

Example:
`triage predictlist 46 2019-05-06`

The predictlist will assume the current path to be the 'project path' to find models and write matrices, but this can be overridden by sending the `--project-path` option.

### Python

The `predict_forward_with_existed_model` function from the `triage.predictlist` module can be used similarly to the CLI, with the addition of the database engine and project storage as inputs.
```
from triage.predictlist import generate predict_forward_with_existed_model
from triage import create_engine

predict_forward_with_existed_model(
db_engine=create_engine(<your-db-info>),
project_path='/home/you/triage/project2'
model_id=46,
as_of_date='2019-05-06'
)
```

## Output
The Predictlist is stored similarly to the matrices created during an Experiment:
- Raw Matrix saved to the matrices directory in project storage
- Predictions saved in a table (triage_production.predictions)
- Prediction metadata (tiebreaking, random seed) saved in a table (triage_production.prediction_metadata)

## Notes
- The cohort and features for the Predictlist are all inferred from the Experiment that trained the given model_id (as defined by the experiment_models table).
- The feature list ensures that imputation flag columns are present for any columns that either needed to be imputed in the training process, or that needed to be imputed in the predictlist dataset.


11 changes: 10 additions & 1 deletion src/tests/catwalk_tests/test_integration.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@
MatrixStore,
MatrixStorageEngine,
)
from triage.tracking import initialize_tracking_and_get_run_id
from tests.utils import (
get_matrix_store,
matrix_metadata_creator,
Expand All @@ -19,12 +20,20 @@ def test_ModelTrainTester_generate_tasks(db_engine_with_results_schema, project_
model_storage_engine = ModelStorageEngine(project_storage)
matrix_storage_engine = MatrixStorageEngine(project_storage)
sample_matrix_store = get_matrix_store(project_storage)
experiment_hash = save_experiment_and_get_hash({}, 1234, db_engine)
experiment_hash = save_experiment_and_get_hash({}, db_engine)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=5,
experiment_kwargs={},
db_engine=db_engine_with_results_schema
)
# instantiate pipeline objects
trainer = ModelTrainer(
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
db_engine=db_engine,
run_id=run_id,
)
train_tester = ModelTrainTester(
matrix_storage_engine=matrix_storage_engine,
Expand Down
148 changes: 144 additions & 4 deletions src/tests/catwalk_tests/test_model_trainers.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@

from triage.component.catwalk.model_grouping import ModelGrouper
from triage.component.catwalk.model_trainers import ModelTrainer
from triage.component.catwalk.utils import save_experiment_and_get_hash
from triage.tracking import initialize_tracking_and_get_run_id
from tests.utils import get_matrix_store


Expand All @@ -22,11 +24,24 @@ def grid_config():
@pytest.fixture(scope="function")
def default_model_trainer(db_engine_with_results_schema, project_storage):
model_storage_engine = project_storage.model_storage_engine()
experiment_hash = save_experiment_and_get_hash(
config={'foo': 'bar'},
db_engine=db_engine_with_results_schema
)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=5,
experiment_kwargs={},
db_engine=db_engine_with_results_schema
)
# import pdb; pdb.set_trace()
trainer = ModelTrainer(
experiment_hash=None,
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
db_engine=db_engine_with_results_schema,
model_grouper=ModelGrouper(),
run_id=run_id,
)
yield trainer

Expand All @@ -45,7 +60,6 @@ def set_test_seed():
misc_db_parameters=dict(),
matrix_store=get_matrix_store(project_storage),
)

# assert
# 1. that the models and feature importances table entries are present
records = [
Expand Down Expand Up @@ -132,14 +146,26 @@ def set_test_seed():
"select max(batch_run_time) from triage_metadata.models"
)
][0]
experiment_hash = save_experiment_and_get_hash(
config={'foo': 'bar'},
db_engine=db_engine
)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=5,
experiment_kwargs={},
db_engine=db_engine
)
trainer = ModelTrainer(
experiment_hash=None,
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
model_grouper=ModelGrouper(
model_group_keys=["label_name", "label_timespan"]
),
db_engine=db_engine,
replace=True,
run_id=run_id,
)
set_test_seed()
new_model_ids = trainer.train_models(
Expand Down Expand Up @@ -212,11 +238,23 @@ def test_baseline_exception_handling(default_model_trainer):

def test_custom_groups(grid_config, db_engine_with_results_schema, project_storage):
model_storage_engine = project_storage.model_storage_engine()
experiment_hash = save_experiment_and_get_hash(
config={'foo': 'bar'},
db_engine=db_engine_with_results_schema
)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=5,
experiment_kwargs={},
db_engine=db_engine_with_results_schema
)
trainer = ModelTrainer(
experiment_hash=None,
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
model_grouper=ModelGrouper(["class_path"]),
db_engine=db_engine_with_results_schema,
run_id=run_id,
)
# create training set
model_ids = trainer.train_models(
Expand All @@ -235,6 +273,108 @@ def test_custom_groups(grid_config, db_engine_with_results_schema, project_stora
assert records[0] == model_ids[0]


def test_reuse_model_random_seeds(grid_config, default_model_trainer):
trainer = default_model_trainer
db_engine = trainer.db_engine
project_storage = trainer.model_storage_engine.project_storage
model_storage_engine = trainer.model_storage_engine

# re-using the random seeds requires the association between experiments and models
# to exist, which we're not getting in these tests since we aren't using the experiment
# architecture, so back-fill these associations after each train_models() run
def update_experiment_models(db_engine):
sql = """
INSERT INTO triage_metadata.experiment_models(experiment_hash,model_hash)
SELECT er.run_hash, m.model_hash
FROM triage_metadata.models m
LEFT JOIN triage_metadata.triage_runs er
ON m.built_in_triage_run = er.id
LEFT JOIN triage_metadata.experiment_models em
ON m.model_hash = em.model_hash
AND er.run_hash = em.experiment_hash
WHERE em.experiment_hash IS NULL
"""
db_engine.execute(sql)
db_engine.execute('COMMIT;')

random.seed(5)
model_ids = trainer.train_models(
grid_config=grid_config,
misc_db_parameters=dict(),
matrix_store=get_matrix_store(project_storage),
)
update_experiment_models(db_engine)

# simulate running a new experiment where the experiment hash has changed
# (e.g. because the model grid is different), but experiment seed is the
# same, so previously-trained models should not get new seeds
experiment_hash = save_experiment_and_get_hash(
config={'baz': 'qux'},
db_engine=db_engine
)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=5,
experiment_kwargs={},
db_engine=db_engine
)
trainer = ModelTrainer(
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
db_engine=db_engine,
model_grouper=ModelGrouper(),
run_id=run_id,
)
new_grid = grid_config.copy()
new_grid['sklearn.tree.DecisionTreeClassifier']['min_samples_split'] = [3,10,100]
random.seed(5)
new_model_ids = trainer.train_models(
grid_config=new_grid,
misc_db_parameters=dict(),
matrix_store=get_matrix_store(project_storage),
)
update_experiment_models(db_engine)

# should have received 5 models
assert len(new_model_ids) == 6

# all the original model ids should be in the new set
assert len(set(new_model_ids) & set(model_ids)) == len(model_ids)

# however, we should NOT re-use the random seeds (and so get new model_ids)
# if the experiment-level seed is different
experiment_hash = save_experiment_and_get_hash(
config={'lorem': 'ipsum'},
db_engine=db_engine
)
run_id = initialize_tracking_and_get_run_id(
experiment_hash,
experiment_class_path="",
random_seed=42,
experiment_kwargs={},
db_engine=db_engine
)
trainer = ModelTrainer(
experiment_hash=experiment_hash,
model_storage_engine=model_storage_engine,
db_engine=db_engine,
model_grouper=ModelGrouper(),
run_id=run_id,
)
random.seed(42) # different from above
newer_model_ids = trainer.train_models(
grid_config=new_grid,
misc_db_parameters=dict(),
matrix_store=get_matrix_store(project_storage),
)
update_experiment_models(db_engine)

# should get entirely new models now (different IDs)
assert len(newer_model_ids) == 6
assert len(set(new_model_ids) & set(newer_model_ids)) == 0


def test_n_jobs_not_new_model(default_model_trainer):
grid_config = {
"sklearn.ensemble.AdaBoostClassifier": {"n_estimators": [10, 100, 1000]},
Expand Down
8 changes: 4 additions & 4 deletions src/tests/catwalk_tests/test_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,9 +64,9 @@ def test_save_experiment_and_get_hash():
with testing.postgresql.Postgresql() as postgresql:
engine = create_engine(postgresql.url())
ensure_db(engine)
exp_hash = save_experiment_and_get_hash(experiment_config, 1234, engine)
exp_hash = save_experiment_and_get_hash(experiment_config, engine)
assert isinstance(exp_hash, str)
new_hash = save_experiment_and_get_hash(experiment_config, 1234, engine)
new_hash = save_experiment_and_get_hash(experiment_config, engine)
assert new_hash == exp_hash


Expand All @@ -75,7 +75,7 @@ def test_missing_model_hashes():
db_engine = create_engine(postgresql.url())
ensure_db(db_engine)

experiment_hash = save_experiment_and_get_hash({}, 1234, db_engine)
experiment_hash = save_experiment_and_get_hash({}, db_engine)
model_hashes = ['abcd', 'bcde', 'cdef']

# if we associate model hashes with an experiment but don't actually train the models
Expand All @@ -96,7 +96,7 @@ def test_missing_matrix_uuids():
db_engine = create_engine(postgresql.url())
ensure_db(db_engine)

experiment_hash = save_experiment_and_get_hash({}, 1234, db_engine)
experiment_hash = save_experiment_and_get_hash({}, db_engine)
matrix_uuids = ['abcd', 'bcde', 'cdef']

# if we associate matrix uuids with an experiment but don't actually build the matrices
Expand Down
Loading