ai-center
2024.10
true
- Release Notes
- Before you begin
- Getting started
- Installing AI Center
- Migration and upgrade
- Projects
- Datasets
- Data Labeling
- ML packages
- Out of the box packages
- Pipelines
- Training pipelines
- Evaluation pipelines
- Full pipelines
- Managing pipelines
- Closing the loop
- ML Skills
- ML Logs
- Document UnderstandingTM in AI Center
- AI Center API
- How to
- Licensing
- Basic Troubleshooting Guide
Evaluation pipelines
AI Center User Guide
Last updated Nov 11, 2024
Evaluation pipelines
An Evaluation Pipeline is used to evaluate a trained machine learning model. To use this pipeline, the package must contain
code to evaluate a model (the
evaluate()
function in the train.py file). This code, together with a dataset or sub-folder within a dataset, produce a score (the return of the evaluate()
function) and any arbitrary outputs the user would like to persist in addition to the score.
Create a new evaluation pipeline as described here. Make sure to provide the following evaluation pipeline specific information:
- In the Pipeline type field, select Evaluation run.
- In the Choose evaluation dataset field, select a dataset or folder from which you want to import data for evaluation. All files in this dataset/folder should
be available locally during the runtime of the pipeline, being passed to the argument to your
evaluate()
function. - In the Enter parameters section, enter the environment variables defined and used by your pipeline, if any. The environment variables are:
artifacts_directory
, with default value artifacts: This defines the path to a directory that will be persisted as ancillary data related to this pipeline. Most, if not all users, will never have the need to override this through the UI. Anything can be saved during pipeline execution including images, pdfs, and subfolders. Concretely, any data your code writes in the directory specified by the pathos.environ['artifacts_directory']
will be uploaded at the end of the pipeline run and will be viewable from the Pipeline details page.save_test_data
, with default value false: If set to true,data_directory
folder will be uploaded at the end of the pipeline run as an output of the pipeline under directorydata_directory
.Note: The pipeline execution might take some time. Check back to it after a while to see its status.After the pipeline was executed, in the Pipelines page, the pipeline's status changed to Successful. The Pipeline Details page displays the arbitrary files and folders related to the pipeline run. In our example, the run created a file calledmy-evaluate-artifact.txt
.
Here is a conceptually analogous execution of an
evaluation pipeline on some package, for example version 1.1, the output of a training
pipeline on version 1.0.
Note: This is a simplified example. Its
purpose is to illustrate how datasets and packages interact in an evaluation
pipeline. The steps are merely conceptual and do not represent how the platform
works.
The
_results.json
file contains a summary of the pipeline run execution,
exposing all inputs/outputs and execution times for an evaluation pipeline.
{
"parameters": {
"pipeline": "< Pipeline_name >",
"inputs": {
"package": "<Package_name>",
"version": "<version_number>",
"evaluation_data": "<storage_directory>",
"gpu": "True/False"
},
"env": {
"key": "value",
...
}
},
"run_summary": {
"execution_time": <time>, #in seconds
"start_at": <timestamp>, #in seconds
"end_at": <timestamp>, #in seconds
"outputs": {
"score": <score>, #float
"train_data": "<test_storage_directory>",
"evaluation_data": "<test_storage_directory>/None",
"artifacts_data": "<artifacts_storage_directory>",
}
}
}
{
"parameters": {
"pipeline": "< Pipeline_name >",
"inputs": {
"package": "<Package_name>",
"version": "<version_number>",
"evaluation_data": "<storage_directory>",
"gpu": "True/False"
},
"env": {
"key": "value",
...
}
},
"run_summary": {
"execution_time": <time>, #in seconds
"start_at": <timestamp>, #in seconds
"end_at": <timestamp>, #in seconds
"outputs": {
"score": <score>, #float
"train_data": "<test_storage_directory>",
"evaluation_data": "<test_storage_directory>/None",
"artifacts_data": "<artifacts_storage_directory>",
}
}
}
Artifacts folder, visible only if not empty, is a folder regrouping all the
artifacts generated by the pipeline and saved under the
artifacts_directory
folder.
Dataset folder, existing only if
save_data
was set to the default
true value, is a copy of the evaluation dataset folder.
As in training pipelines, a user can set the
parameter
save_test_data
= true
to snapshot data
passed in for evaluation.