The Google Cloud Vertex AI conceptual model for Generative AI workflow is shown below (image from Google documentation at link.)
Vertex AI includes a model garden with a large number of models which are categorized as foundational, fine-tunable or task specific. These include models built by Google as well as many other industry and open source models. Shown below is screenshot of some of the the currently available models and associated services in the VertexAI Model Garden.
Google Gemini is a large language model that can be used for a variety of tasks, including text generation, text classification, text summarization, text translation, text-to-speech, speech-to-text, image captioning, image classification, image generation, image-to-text, video captioning, video classification, video generation, video-to-text, and multimodal tasks.
Google Gemini is available in multiple (multi-modal input) versions. Flash is the smallest version of Gemini, and is designed for use on mobile devices. Pro is the most powerful version of Gemini, and is designed for use on desktop computers. Ultra is the most powerful version of Gemini, and is designed for use on servers. Gemini has multiple editions and versions.
- Gemini 2.5 Pro announced March 2025 - most advanced model, several versions available.
- Gemini 2.0 several versions (i.e. Flash, etc...) available.
- Gemini model lifecycle information is summarized in this page
- Google Gemma is a family of lightweight, state-of-the art open models built from the same research and technology that we used to create the Gemini models.
Google has created tools and also published a number of resources to get you started working with Gemini. I've linked highlights below.
- Google AI Studio IDE --> https://aistudio.google.com/
- Google AI Studio quickstart --> https://ai.google.dev/tutorials/ai-studio_quickstart
- Google's setup instuctions for Gemini on GCP --> https://github.com/GoogleCloudPlatform/generative-ai/tree/main/setup-env
- Google's Guidance "Build w/Gemini on Google Cloud" --> https://ai.google.dev/gemini-api/docs/migrate-to-cloud
- Prompt guidance --> https://ai.google.dev/docs/prompt_best_practices
- System Instructions in prompts --> https://ai.google.dev/docs/system_instructions
- Multi-modal prompt guidance --> https://ai.google.dev/docs/multimodal_concepts
The following Jupyter Notebook is an example of how to use Google Gemini LLM for text generation. It uses the Pro version of Gemini, and shows how to call the API using Python. It also shows how to use the API to generate text from a prompt. The notebook can be run in Google Colab, or on a local machine with Python 3 installed. The notebook can be found here: Google Gemini LLM Example Notebook
- 📺 Video "How to build Multimodal Retrieval-Augmented Generation (RAG) with Gemini" --> link
- 📚 Vertex AI GenAI example notebooks explained and linked --> link
Example notebooks on GitHub --> link- 📓 Jumpstart Document Summary Application --> link
Gemini Cookbook (notebook examples) --> link- 📓 📓 📓 MANY Model evaluation notebooks --> link
- Source code for GKE demo on Github - end-to-end application with 11 microservices, includes use of LLM Google Gemini
- Google's big list:
Learn More About GenAI--> link - Google's Architecture Center (JumpStarts for GenAI - deployable patterns) --> links
- Article:
GenAI Apps w/VertexAI Models and Langchain--> link - Article:
Generate Text Embeddings with BigQuery--> link - Example code:
ChromaDB w/ Gemini--> link - Article:
Exploring the API of Gemini--> link - Big link list:
Awesome Gemini for Google Cloud--> link

