Cloud Computing

Use AI to build AI: Save time on prompt design with AI-powered prompt writing

Crafting the perfect prompt for generative AI models can be an art in itself. The difference between a useful and a generic AI response can sometimes be a well-crafted prompt. But, getting there often requires time-consuming tweaking, iteration, and a learning curve. That’s why we’re thrilled to announce new updates to the AI-powered prompt writing […]

Use AI to build AI: Save time on prompt design with AI-powered prompt writing Read More »

Transforming DoD’s data utilization with generative AI

Generative AI presents both immense opportunities and challenges for the Department of Defense (DoD). The potential to enhance situational awareness, streamline tasks, and improve decision-making is significant. However, the DoD’s unique requirements, especially their stringent security standards for cloud services (IL5), necessitate carefully crafted AI solutions that balance innovation with security. The DoD’s 2023 Data,

Transforming DoD’s data utilization with generative AI Read More »

Secure your data ecosystem: a multi-layered approach with Google Cloud

It’s an exciting time in the world of data and analytics, with more organizations harnessing the power of data and AI to help transform and grow their businesses. But in a threat landscape with increasingly sophisticated attacks around every corner, ensuring the security and integrity of that data is critical. Google Cloud offers a comprehensive

Secure your data ecosystem: a multi-layered approach with Google Cloud Read More »

Cloud CISO Perspectives: The high security cost of legacy tech

Welcome to the first Cloud CISO Perspectives for November 2024. Today I’m joined by Andy Wen, Google Cloud’s senior director of product management for Google Workspace, to discuss a new Google survey into the high security costs of legacy tech. As with all Cloud CISO Perspectives, the contents of this newsletter are posted to the

Cloud CISO Perspectives: The high security cost of legacy tech Read More »

How to deploy Llama 3.2-1B-Instruct model with Google Cloud Run GPU

As open-source large language models (LLMs) become increasingly popular, developers are looking for better ways to access new models and deploy them on Cloud Run GPU. That’s why Cloud Run now offers fully managed NVIDIA GPUs, which removes the complexity of driver installations and library configurations. This means you’ll benefit from the same on-demand availability

How to deploy Llama 3.2-1B-Instruct model with Google Cloud Run GPU Read More »

Pirates in the Data Sea: AI Enhancing Your Adversarial Emulation

Matthijs Gielen, Jay Christiansen Background New solutions, old problems. Artificial intelligence (AI) and large language models (LLMs) are here to signal a new day in the cybersecurity world, but what does that mean for us—the attackers and defenders—and our battle to improve security through all the noise? Data is everywhere. For most organizations, the access

Pirates in the Data Sea: AI Enhancing Your Adversarial Emulation Read More »

Empower your teams with self-service Kubernetes using GKE fleets and Argo CD

Managing applications across multiple Kubernetes clusters is complex, especially when those clusters span different environments or even cloud providers. One powerful and secure solution combines Google Kubernetes Engine (GKE) fleets and, Argo CD, a declarative, GitOps continuous delivery tool for Kubernetes. The solution is further enhanced with Connect Gateway and Workload Identity. This blog post

Empower your teams with self-service Kubernetes using GKE fleets and Argo CD Read More »

Data loading best practices for AI/ML inference on GKE

As AI models increase in sophistication, there’s increasingly large model data needed to serve them. Loading the models and weights along with necessary frameworks to serve them for inference can add seconds or even minutes of scaling delay, impacting both costs and the end-user’s experience.  For example, inference servers such as Triton, Text Generation Inference

Data loading best practices for AI/ML inference on GKE Read More »

65,000 nodes and counting: Google Kubernetes Engine is ready for trillion-parameter AI models

As generative AI evolves, we’re beginning to see the transformative potential it is having across industries and our lives. And as large language models (LLMs) increase in size — current models are reaching hundreds of billions of parameters, and the most advanced ones are approaching 2 trillion — the need for computational power will only

65,000 nodes and counting: Google Kubernetes Engine is ready for trillion-parameter AI models Read More »