AI engine: TensorFlow 1.8; Environment: Python 3.6; Input and output mode: undefined mode. Select an appropriate input and output mode based on the model function or application scenario. When using the template to import a model, select the model directory containing the model files.
The template input is the TensorFlow-based model package stored on OBS. Ensure that the OBS directory you use and ModelArts are in the same region. For details about model package requirements, see Model Package Example.
Undefined Mode can be overwritten. That is, you can select another input and output mode during model creation.
model/ │ ├── Model file //(Mandatory) The model file format varies according to the engine. For details, see the model package example. ├── Custom Python package //(Optional) User's Python package, which can be directly referenced in the model inference code ├── customize_service.py //(Optional) Model inference code file. The file name must be customize_service.py. Otherwise, the code is not considered as inference code.
Structure of the TensorFlow-based model package
When publishing the model, you only need to specify the model directory.
OBS bucket/directory name |── model (Mandatory) The folder must be named model and is used to store model-related files. ├── <<Custom Python package>> (Optional) User's Python package, which can be directly referenced in the model inference code ├── saved_model.pb (Mandatory) Protocol buffer file, which contains the diagram description of the model ├── variables Mandatory for the main file of the *.pb model. The folder must be named variables and contains the weight deviation of the model. ├── variables.index Mandatory ├── variables.data-00000-of-00001 Mandatory ├──customize_service.py (Optional) Model inference code file. The file must be named customize_service.py. Only one inference code file exists. The .py file on which customize_service.py depends can be directly put in the model directory.