Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Validation framework for model #17

Open
sgreenbury opened this issue Apr 19, 2024 · 9 comments
Open

Validation framework for model #17

sgreenbury opened this issue Apr 19, 2024 · 9 comments
Assignees
Labels
validation Model validation and consistency

Comments

@sgreenbury
Copy link
Collaborator

Aim: define metrics to be used at different parts of the modelling to validate model against data. E.g. flows from QUANT model. See section in wiki.

@sgreenbury
Copy link
Collaborator Author

sgreenbury commented Apr 26, 2024

Some measures to consider:

  • Modelled travel times compared to reported travel times in NTS
  • Modelled flows between MSOAs compared to observed flows used in QUANT
  • See here for commuting flows census data. We can check the SPC people to matched workplace flows is consistent with this data.

@sgreenbury sgreenbury added the validation Model validation and consistency label Apr 26, 2024
@Hussein-Mahfouz
Copy link
Collaborator

Guidelines from dft on activity and agent based models: TAG unit M5-4

@Hussein-Mahfouz
Copy link
Collaborator

Putting this here for now but could be a separate issue on calibration later:

"We need data sets to be recognized as components on the same level as models. Such data components can then enter the integrated frameworks at various places, not only at the top, as input to drive the whole integrated model, and at the bottom, to compare with the output and to calibrate the model. Data components can be also used between components to test, adjust, and correct the data flows inside the integrated model. This can substantially increase the efficiency and accuracy of the integration process, and reduce the overall complexity of the calibration task for the whole integrated model." - ‘Integronsters’, integral and integrated modeling

A useful exercise could be to identify the datasets that could be used to calibrate at intermediate points of the pipeline

@sgreenbury
Copy link
Collaborator Author

sgreenbury commented May 9, 2024

Notes on tasks:

  • Retrieve data on MSOA to MSOA flows by mode
  • Function to measure difference in NTS (ground truth) and estimated travel time (from travel time matrices)
  • Calculate the aggregate MSOA to MSOA flows for a SPC matched population

@sgreenbury
Copy link
Collaborator Author

We also need to determine a set of metrics for measuring quality of matching between the two datasets as part of task1.

@BZ-BowenZhang
Copy link
Collaborator

We also need to determine a set of metrics for measuring quality of matching between the two datasets as part of task1.

I completely agree with @sgreenbury, and I think calculating the metrics for comparison is easy. But after generating metrics, I have a question: How do we judge the 'goodness'? In other words, how do we set a threshold value as the acceptable standard? We do not have another candidate dataset to compare, but maybe we can compare it with other synthetic populations published in previous papers.

@sgreenbury
Copy link
Collaborator Author

Adding reference with validation methods from @stuartlynn

@BZ-BowenZhang
Copy link
Collaborator

BZ-BowenZhang commented Aug 28, 2024

  • Add more markdown descriptions for the notebook Validation_SPC_with Cencus

@BZ-BowenZhang BZ-BowenZhang linked a pull request Sep 26, 2024 that will close this issue
@BZ-BowenZhang
Copy link
Collaborator

As discussed on 20th Sep:

  • The Census (commuting OD) dataset will be used as the benchmark dataset to validate the performance of different synthetic datasets (SPC and AcBM). The Census and SPC validation is completed and ready to merge with the main branch PR 17 validation framework for model #50
  • For the next step, more validation work will be done after the whole Leeds population with activity chain is available, still comparing with the census data and comparing with the SPC.
  • In the further London case, we may have more option datasets to compare/validate such as mobile phone data from Tao's Group or travel flow from TfL etc.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
validation Model validation and consistency
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants