Step-by-Step Guide to Deploying ML Solutions with FastAPI, Docker, and GCP


Deploying machine learning (ML) solutions can be a complex process, but with the right tools and knowledge, it can be simplified. This guide will walk you through deploying an ML model using FastAPI, Docker, and Google Cloud Platform (GCP). We’ll cover everything from basic concepts to advanced deployment techniques with example Python code.

Why Choose FastAPI, Docker, and GCP for ML Deployment?

  • FastAPI: A modern, fast (high-performance), web framework for building APIs with Python 3.6+ based on standard Python type hints.
  • Docker: An open platform for developing, shipping, and running applications. Docker enables you to separate your applications from your infrastructure so you can deliver software quickly.
  • Google Cloud Platform (GCP): Provides a suite of cloud computing services that runs on the same infrastructure that Google uses internally for its end-user products.


Before we begin, make sure you have the following installed and set up:

  • Python 3.6+
  • DockerDocker
  • GCP account

Step 1: Building the ML Model with FastAPI

First, let's create a simple ML model and expose it using FastAPI. Here’s a basic example of a FastAPI application:

from fastapi import FastAPI import numpy as np from sklearn.linear_model import LinearRegression app = FastAPI() # Dummy data for the example X = np.array([[1, 1], [1, 2], [2, 2], [2, 3]]) y =, np.array([1, 2])) + 3 model = LinearRegression().fit(X, y) @app.get("/predict") def predict(x1: float, x2: float): prediction = model.predict([[x1, x2]]) return {"prediction": prediction[0]}

Step 2: Dockerizing the Application

To deploy our application, we need to containerize it using Docker. Create a Dockerfile:

FROM tiangolo/uvicorn-gunicorn-fastapi:python3.7 COPY ./app /app RUN pip install -r /app/requirements.txt CMD ["uvicorn", "main:app", "--host", "", "--port", "80"]

And a requirements.txt file:

fastapi scikit-learn

Build and run the Docker container:

docker build -t fastapi-ml-app . docker run -d --name fastapi-ml-app -p 80:80 fastapi-ml-app

Step 3: Deploying to GCP

  1. Setting up GCP:

    • Create a new project in the GCP Console.
    • Enable the Google Kubernetes Engine (GKE) API.
    • Set up billing for your project.
  2. Creating a Kubernetes Cluster:

    • Use the GCP Console or the gcloud command-line tool to create a GKE cluster.
  3. Deploying Docker Container to GKE:

    • Push your Docker image to Google Container Registry (GCR):

      docker tag fastapi-ml-app[PROJECT-ID]/fastapi-ml-app docker push[PROJECT-ID]/fastapi-ml-app
    • Deploy to your Kubernetes cluster:

      apiVersion: apps/v1 kind: Deployment metadata: name: fastapi-ml-app spec: replicas: 1 selector: matchLabels: app: fastapi-ml-app template: metadata: labels: app: fastapi-ml-app spec: containers: - name: fastapi-ml-app image:[PROJECT-ID]/fastapi-ml-app ports: - containerPort: 80
    • Expose your application:

      kubectl expose deployment fastapi-ml-app --type=LoadBalancer --port 80


Deploying ML models with FastAPI, Docker, and GCP can streamline the process and make your applications scalable and maintainable. By following this guide, you can deploy your ML solutions efficiently, leveraging the power of cloud infrastructure. thank you for reading the page!


Popular posts from this blog

zimbra some services are not running [Solve problem]

How to install php7 on centos 6

Bash script list all IP addresses connected to Server