Build Your Own AI Chat App with Google Gemini and Cloud Run

Large Language Models (LLMs) like Google’s Gemini offer powerful capabilities for creating interactive applications. This guide focuses on building a simple yet functional web-based chat application powered by the Gemini API and deploying it seamlessly using Google Cloud Run. This tutorial is ideal for developers exploring LLM integration or looking for practical ways to deploy AI features in their web projects.

What This Guide Covers

Follow this tutorial to construct a web chat interface that interacts directly with the Gemini API to provide conversational responses. The application will be containerized using Docker and deployed onto Google Cloud Run via Cloud Build and Artifact Registry.

Upon completion, you will have learned how to:

  • Set up a Python web application interacting with the Gemini API.
  • Containerize the application using Docker.
  • Deploy the container to Cloud Run using Google Cloud tools.
  • Gain insights into integrating LLMs into web applications.

Think of it as building your own helpful chat assistant, ready to deploy to the cloud.

Prerequisites

Before starting, ensure the following prerequisites are met:

  • A Google Cloud project.
  • The gcloud command-line interface installed and authenticated.
  • Docker installed on your machine.
  • The Vertex AI API enabled within your Google Cloud project.
  • (Optional: Google Cloud Shell provides a pre-configured environment with most tools ready.)

Step 1: Obtain the Starter Code

Begin by cloning the pre-built Python Flask application template from its repository. Open your terminal and run:

git clone https://github.com/ChloeCodesThings/chat-app-demo
cd chat-app-demo

Inside the cloned directory, you’ll find these key files:

  • app.py: Contains the Flask application routes and the logic for interacting with the Gemini API.
  • index.html: Provides a basic user interface for the chat application.
  • Dockerfile: Includes instructions for building the application’s container image.

Step 2: Build and Store the Docker Image using Cloud Build

Containerizing the application makes deployment consistent and reliable. First, define necessary environment variables in your terminal. Replace us-central1 with your preferred Google Cloud region if needed.

export PROJECT_ID=$(gcloud config get-value project)
export REGION=us-central1
export AR_REPO=chat-app-repo
export SERVICE_NAME=chat-gemini-app

Next, create a repository in Google Artifact Registry to store your Docker image:

gcloud artifacts repositories create $AR_REPO \
--repository-format=docker \
--location=$REGION

Finally, use Cloud Build to build the Docker image based on your Dockerfile and push it to the Artifact Registry repository you just created:

gcloud builds submit --tag $REGION-docker.pkg.dev/$PROJECT_ID/$AR_REPO/$SERVICE_NAME

Cloud Build automates the build process using the instructions in your Dockerfile.

Step 3: Deploy the Application to Cloud Run

With the container image built and stored in Artifact Registry, you can deploy it to Cloud Run, Google’s serverless platform for containerized applications:

gcloud run deploy $SERVICE_NAME \
  --image=$REGION-docker.pkg.dev/$PROJECT_ID/$AR_REPO/$SERVICE_NAME \
  --platform=managed \
  --region=$REGION \
  --allow-unauthenticated \
  --set-env-vars=GCP_PROJECT=$PROJECT_ID,GCP_REGION=$REGION

This command deploys your container image as a new Cloud Run service. The --allow-unauthenticated flag makes the service publicly accessible (for production, you’d typically implement authentication). Cloud Run will provide a URL for your deployed service, similar to `https://chat-gemini-app-xxxxxx.run.app`.

Step 4: Understanding the Gemini API Interaction

The core AI logic resides in app.py. Here’s a snippet showcasing the interaction with the Vertex AI SDK for Python, which facilitates communication with the Gemini API:

from vertexai.preview.language_models import ChatModel

def create_session():
    chat_model = ChatModel.from_pretrained("gemini-pro") # Or other available models like gemini-1.0-pro
    chat = chat_model.start_chat()
    return chat

def response(chat, message):
    result = chat.send_message(message)
    return result.text

This code initializes a chat session using a specified Gemini model (like gemini-pro). When a user submits a message through the web interface, the response function sends this message to the Gemini model via the established chat session and returns the model’s text response back to the user interface.

Test Your AI Chat App

Open the deployment URL provided by Cloud Run in your browser. Enter a prompt or question, for example:

“Explain Google Cloud Run in simple terms.”

You should receive a relevant, conversational response generated by the Gemini model running via your deployed application.

Expanding Your Project

This basic application serves as a great starting point. Consider these potential enhancements for future development:

  • Improving the chat interface for a better user experience.
  • Implementing session persistence to maintain chat history across interactions.
  • Using system prompts to guide Gemini’s personality and response style.
  • Securing the application endpoint using Identity Platform, API Gateway, or other authentication methods.

Summary

Congratulations! You have successfully built and deployed a chat application powered by Google Gemini on Cloud Run. This involved setting up a Flask app, containerizing it with Docker, utilizing Cloud Build and Artifact Registry, and deploying to a scalable serverless platform. This project provides a solid foundation for integrating powerful AI capabilities into your web applications.


Leverage AI Power with Innovative Software Technology

Integrating advanced AI like Google Gemini into business applications unlocks significant value, enhancing customer engagement and automating workflows. At Innovative Software Technology, we specialize in developing and deploying cutting-edge AI solutions tailored to your needs. Our expert team excels at seamless Gemini API integration, building scalable web applications, and managing robust Google Cloud Run deployments. Partner with Innovative Software Technology for expert AI application development and transform your operations with intelligent, high-performing custom software solutions on the cloud. We turn complex AI possibilities into tangible business results.

Leave a Reply

Your email address will not be published. Required fields are marked *

Fill out this field
Fill out this field
Please enter a valid email address.
You need to agree with the terms to proceed