Skip to content

Latest commit

 

History

History

serving

CXR Foundation serving

This folder contains the source code and configuration necessary to serve the model on Vertex AI. The implementation follows this container architecture.

The serving container can be used in both online and batch prediction workflows:

Description of select files and folders

  • serving_framework/: A library for implementing Vertex AI-compatible HTTP servers.

  • vertex_schemata/: Folder containing YAML files that define the PredictSchemata for Vertex AI endpoints.

  • Dockerfile: Defines the Docker image for serving the model.

  • entrypoint.sh: A bash script that is used as the Docker entrypoint. It sets up the necessary environment variables, copies the TensorFlow SavedModel(s) locally and launches the TensorFlow server and the frontend HTTP server.

  • model_config.txtpb: The protocol buffer message used in entrypoint.sh to configure the TensorFlow Model server to run multiple models.

  • predictor.py: Prepares model input, calls the models, and post-processes the output into the final response.

  • requirements.txt: Lists the required Python packages.

  • server_gunicorn.py: Creates the HTTP server that launches the prediction executor.

Dependencies