This project demonstrates a complete workflow for deploying a machine learning model on Google Cloud Platform. It covers the following steps:
- Model Development: Train and build your machine learning model using any framework (e.g., TensorFlow, PyTorch), I used Google Cloud Shell, a free browser-based command-line environment on GCP.
- Dockerization: Create a Dockerfile and containerize your model, including its dependencies and environment. This ensures consistent and portable deployment across different environments.
- Artifact Registry(AR): Push the built Docker image to Google Artifact Registry, a secure and managed repository for storing container images(Docker Images).
- Kubernetes Engine (GKE) Deployment: Deploy the containerized model as a service on Google Kubernetes Engine, a managed container orchestration platform. This allows for scalable and automated deployments.
- Frontend Integration: Create a basic frontend application (e.g., using Flask or Streamlit) to interact with the exposed endpoint of your deployed model.
- Containerization: Enables consistent and portable deployments.
- Artifact Registry: Provides a secure and centralized location for storing container images.
- Kubernetes Engine: Offers an automated, scalable, and flexible platform for container orchestration.