You must set up model monitoring for each model that you want to monitor.
- 
When you trained the model, you registered or used a Domino Training Set in the same project from which you’re looking to publish a Model API. 
- 
If you will perform Model Quality analysis, set up your data sources for ground truth data. See Connect a Data Source. 
Set up monitoring for Model APIs
The following topics explain the steps:
- 
Set up Prediction Data Capture. Prediction data is required for monitoring both data drift and model quality. 
- 
Publish the Model API. Domino starts to log prediction data (and convert it to Parquet files), but it does not yet produce monitoring data. 
- 
Set up Drift Detection. Register a training set to monitor data drift. - 
Optional: Configure Set up Notifications or change the scheduled checks. 
 
- 
- 
Validate your Setup. Confirm that your predication data is being captured. 
- 
Set up Model Quality Monitoring. Ingest ground truth data to monitor the quality of the model’s predictions. - 
Optional: Configure Set up Notifications or change the scheduled checks. 
 
- 
- 
Set up Cohort Analysis. Get insights into model quality so you can find underperforming data hotspots for model remediation. 
