Cloud Computing

Vertex AI Search and Generative AI (with Gemini) achieve FedRAMP High

In the rapidly evolving AI landscape, security remains paramount. Today, we reinforce that commitment with another significant achievement: FedRAMP High authorization for Google Vertex AI Search and Generative AI on Vertex AI. This follows our announcement earlier this week where we shared that Gemini in Workspace apps and the Gemini app are the first generative […]

Vertex AI Search and Generative AI (with Gemini) achieve FedRAMP High Read More »

Build richer gen AI experiences using model endpoint management

Model endpoint management is available on AlloyDB, AlloyDB Omni and Cloud SQL for PostgreSQL. Model endpoint management helps developers to build new experiences using SQL and provides a flexible interface to call gen AI models running anywhere — right from the database. You can generate embeddings inside the database, perform quality control on your vector

Build richer gen AI experiences using model endpoint management Read More »

Vector similarity search for Cloud SQL for MySQL is now GA

If you used the internet today, you’ve probably already benefited from generative AI. Whether it helped you get your work done faster, research home repairs, or find the perfect gift, gen AI is transforming how we get things done. These generative AI experiences use searches against vector embeddings — multi-dimensional representations of data’s meaning —

Vector similarity search for Cloud SQL for MySQL is now GA Read More »

Announcing BigQuery repositories: Git-based collaboration in BigQuery Studio

Modern data teams want to use Git to collaborate effectively and adopt software engineering best practices for managing their data pipelines and analytics code. But most tools used by data teams don’t offer integration with Git version control systems, making a Git workflow feel out of reach. This forces users to copy and paste code

Announcing BigQuery repositories: Git-based collaboration in BigQuery Studio Read More »

Harvesting hardware: Our approach to carbon-aware fleet deployment

When it comes to managing the infrastructure and AI that powers Google’s products and platforms – from Search to YouTube to Google Cloud – every decision we make has an impact. Traditionally, meeting growing demands for machine capacity means deploying new machines and that has an associated embodied carbon impact. That’s why we’re working to

Harvesting hardware: Our approach to carbon-aware fleet deployment Read More »

Master architecture decision records (ADRs): Best practices for effective decision-making

Architecture decision records (ADRs) help you document and communicate important process and architecture decisions in your engineering projects. Based on our experience implementing over 200 ADRs across multiple projects, we’ve developed best practices that can help you streamline your decision-making processes and improve team collaboration. In this post, you’ll learn: How to implement ADRs in

Master architecture decision records (ADRs): Best practices for effective decision-making Read More »

Pilot light with reserved capacity: How to optimize DR cost using On-Demand Capacity Reservations

For digital enterprises to remain competitive, resilience is essential for maintaining reliability and building customer trust. End users expect applications to be available 24 hours a day, leading companies to develop increasingly sophisticated methods to provide continuous operation of critical services. Some companies, such as financial services companies, have to meet regulatory requirements such as

Pilot light with reserved capacity: How to optimize DR cost using On-Demand Capacity Reservations Read More »

Using RDMA over Converged Ethernet networking for AI on Google Cloud

All workloads are not the same. This is especially the case for AI, ML, and scientific workloads. In this blog we show how Google Cloud makes the RDMA over converged ethernet version 2 (RoCE v2) protocol available for high performance workloads. Traditional workloads Network communication in traditional workloads involves a well-known flow. This includes: Movement

Using RDMA over Converged Ethernet networking for AI on Google Cloud Read More »

Gen AI Toolbox for Databases announces LlamaIndex integration

We are excited to announce LlamaIndex integration for Gen AI Toolbox for Databases (Toolbox). We launched Toolbox in beta last month and are thrilled to continue building on that momentum. Gen AI Toolbox for Databases is an open-source server that streamlines the development and management of sophisticated generative AI tools that can connect to databases.

Gen AI Toolbox for Databases announces LlamaIndex integration Read More »

AlloyDB for PostgreSQL: Two years of innovation and industry leadership

Two years ago, on a mission to redefine enterprise-grade databases we released AlloyDB for PostgreSQL in production. We saw the immense popularity and flexibility of PostgreSQL — a database developers love for being open-source — and we knew we could build upon that strong foundation to create a next generation fully managed database with a

AlloyDB for PostgreSQL: Two years of innovation and industry leadership Read More »