Manage consistency groups


This document describes how to manage consistency groups. Consistency groups are resource policies that align replication across multiple disks in the same region or zone.

To learn more about consistency groups, see About Asynchronous Replication.

Limitations

  • Consistency groups aren't supported for disks in sole-tenant nodes.
  • Consistency groups can have a maximum of 128 disks.
  • All disks in a consistency group must be in the same project as the consistency group resource policy.
  • All disks in a consistency group must be in the same zone, for zonal disks, or in the same pair of zones, for regional disks.
  • A consistency group can contain primary disks or secondary disks, but not both.
  • You can't add or remove a primary disk to or from a consistency group while the disk is replicating. If you want to add or remove a primary disk to or from a consistency group, you must first stop replication. You can add or remove secondary disks to or from consistency groups at any time.
  • You can attach a maximum of 16 disks that are in different consistency groups, or disks that aren't in a consistency group to a VM. Disks that are in the same consistency group count as one disk towards the 16 disk limit.

Before you begin

  • If you haven't already, then set up authentication. Authentication is the process by which your identity is verified for access to Google Cloud services and APIs. To run code or samples from a local development environment, you can authenticate to Compute Engine by selecting one of the following options:

    Select the tab for how you plan to use the samples on this page:

    Console

    When you use the Google Cloud console to access Google Cloud services and APIs, you don't need to set up authentication.

    gcloud

    1. After installing the Google Cloud CLI, initialize it by running the following command:

      gcloudinit

      If you're using an external identity provider (IdP), you must first sign in to the gcloud CLI with your federated identity.

    2. Set a default region and zone.

    Terraform

    To use the Terraform samples on this page in a local development environment, install and initialize the gcloud CLI, and then set up Application Default Credentials with your user credentials.

    1. Install the Google Cloud CLI.
    2. If you're using an external identity provider (IdP), you must first sign in to the gcloud CLI with your federated identity.

    3. To initialize the gcloud CLI, run the following command:

      gcloudinit
    4. If you're using a local shell, then create local authentication credentials for your user account:

      gcloudauthapplication-defaultlogin

      You don't need to do this if you're using Cloud Shell.

      If an authentication error is returned, and you are using an external identity provider (IdP), confirm that you have signed in to the gcloud CLI with your federated identity.

    For more information, see Set up authentication for a local development environment.

    REST

    To use the REST API samples on this page in a local development environment, you use the credentials you provide to the gcloud CLI.

      After installing the Google Cloud CLI, initialize it by running the following command:

      gcloudinit

      If you're using an external identity provider (IdP), you must first sign in to the gcloud CLI with your federated identity.

    For more information, see Authenticate for using REST in the Google Cloud authentication documentation.

Create a consistency group

If you need to align replication across multiple disks, create a consistency group in the same region as the primary disks. If you need to align disk clones, create a consistency group in the same region as the secondary disks.

Create a consistency group using the Google Cloud console, the Google Cloud CLI, REST, or Terraform.

Console

Create a consistency group by doing the following:

  1. In the Google Cloud console, go to the Asynchronous replication page.

    Go to Asynchronous replication

  2. Click the Consistency groups tab.

  3. Click Create consistency group.

  4. In the Name field, enter a name for the consistency group.

  5. In the Region field, select the region where your disks are located. If you want to add primary disks to consistency group, select the primary region. If you want to add secondary disks to the consistency group, select the secondary region.

  6. Click Create.

gcloud

Create a consistency group using the gcloud compute resource-policies create disk-consistency-group command:

 gcloud compute resource-policies create disk-consistency-group CONSISTENCY_GROUP_NAME \ --region=REGION

Replace the following:

  • CONSISTENCY_GROUP_NAME: the name for the consistency group.
  • REGION: the region for the consistency group. If you want to add primary disks to consistency group, use the primary region. If you want to add secondary disks to the consistency group, use the secondary region.

Go

import("context""fmt""io"compute"cloud.google.com/go/compute/apiv1"computepb"cloud.google.com/go/compute/apiv1/computepb""google.golang.org/protobuf/proto")// createConsistencyGroup creates a new consistency group for a project in a given region.funccreateConsistencyGroup(wio.Writer,projectID,region,groupNamestring)error{// projectID := "your_project_id"// region := "europe-west4"// groupName := "your_group_name"ctx:=context.Background()disksClient,err:=compute.NewResourcePoliciesRESTClient(ctx)iferr!=nil{returnfmt.Errorf("NewResourcePoliciesRESTClient: %w",err)}deferdisksClient.Close()req:=&computepb.InsertResourcePolicyRequest{Project:projectID,ResourcePolicyResource:&computepb.ResourcePolicy{Name:proto.String(groupName),DiskConsistencyGroupPolicy:&computepb.ResourcePolicyDiskConsistencyGroupPolicy{},},Region:region,}op,err:=disksClient.Insert(ctx,req)iferr!=nil{returnfmt.Errorf("unable to create group: %w",err)}iferr=op.Wait(ctx);err!=nil{returnfmt.Errorf("unable to wait for the operation: %w",err)}fmt.Fprintf(w,"Group created\n")returnnil}

Java

importcom.google.cloud.compute.v1.InsertResourcePolicyRequest;importcom.google.cloud.compute.v1.Operation;importcom.google.cloud.compute.v1.Operation.Status;importcom.google.cloud.compute.v1.ResourcePoliciesClient;importcom.google.cloud.compute.v1.ResourcePolicy;importjava.io.IOException;importjava.util.concurrent.ExecutionException;importjava.util.concurrent.TimeUnit;importjava.util.concurrent.TimeoutException;publicclassCreateConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project you want to use.Stringproject="YOUR_PROJECT_ID";// Name of the region in which you want to create the consistency group.Stringregion="us-central1";// Name of the consistency group you want to create.StringconsistencyGroupName="YOUR_CONSISTENCY_GROUP_NAME";createConsistencyGroup(project,region,consistencyGroupName);}// Creates a new consistency group resource policy in the specified project and region.publicstaticStatuscreateConsistencyGroup(Stringproject,Stringregion,StringconsistencyGroupName)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(ResourcePoliciesClientregionResourcePoliciesClient=ResourcePoliciesClient.create()){ResourcePolicyresourcePolicy=ResourcePolicy.newBuilder().setName(consistencyGroupName).setRegion(region).setDiskConsistencyGroupPolicy(ResourcePolicy.newBuilder().getDiskConsistencyGroupPolicy()).build();InsertResourcePolicyRequestrequest=InsertResourcePolicyRequest.newBuilder().setProject(project).setRegion(region).setResourcePolicyResource(resourcePolicy).build();Operationresponse=regionResourcePoliciesClient.insertAsync(request).get(1,TimeUnit.MINUTES);if(response.hasError()){thrownewError("Error creating consistency group! "+response.getError());}returnresponse.getStatus();}}}

Node.js

// Import the Compute libraryconstcomputeLib=require('@google-cloud/compute');constcompute=computeLib.protos.google.cloud.compute.v1;// Instantiate a resourcePoliciesClientconstresourcePoliciesClient=newcomputeLib.ResourcePoliciesClient();// Instantiate a regionOperationsClientconstregionOperationsClient=newcomputeLib.RegionOperationsClient();/** * TODO(developer): Update/uncomment these variables before running the sample. */// The project that contains the consistency group.constprojectId=awaitresourcePoliciesClient.getProjectId();// The region for the consistency group.// If you want to add primary disks to consistency group, use the same region as the primary disks.// If you want to add secondary disks to the consistency group, use the same region as the secondary disks.// region = 'europe-central2';// The name for consistency group.// consistencyGroupName = 'consistency-group-name';asyncfunctioncallCreateConsistencyGroup(){// Create a resourcePolicyResourceconstresourcePolicyResource=newcompute.ResourcePolicy({diskConsistencyGroupPolicy:newcompute.ResourcePolicyDiskConsistencyGroupPolicy({}),name:consistencyGroupName,});const[response]=awaitresourcePoliciesClient.insert({project:projectId,region,resourcePolicyResource,});letoperation=response.latestResponse;// Wait for the create group operation to complete.while(operation.status!=='DONE'){[operation]=awaitregionOperationsClient.wait({operation:operation.name,project:projectId,region,});}console.log(`Consistency group: ${consistencyGroupName} created.`);}awaitcallCreateConsistencyGroup();

Python

from__future__importannotationsimportsysfromtypingimportAnyfromgoogle.api_core.extended_operationimportExtendedOperationfromgoogle.cloudimportcompute_v1defwait_for_extended_operation(operation:ExtendedOperation,verbose_name:str="operation",timeout:int=300)-> Any:""" Waits for the extended (long-running) operation to complete. If the operation is successful, it will return its result. If the operation ends with an error, an exception will be raised. If there were any warnings during the execution of the operation they will be printed to sys.stderr. Args: operation: a long-running operation you want to wait on. verbose_name: (optional) a more verbose name of the operation, used only during error and warning reporting. timeout: how long (in seconds) to wait for operation to finish. If None, wait indefinitely. Returns: Whatever the operation.result() returns. Raises: This method will raise the exception received from `operation.exception()` or RuntimeError if there is no exception set, but there is an `error_code` set for the `operation`. In case of an operation taking longer than `timeout` seconds to complete, a `concurrent.futures.TimeoutError` will be raised. """result=operation.result(timeout=timeout)ifoperation.error_code:print(f"Error during {verbose_name}: [Code: {operation.error_code}]: {operation.error_message}",file=sys.stderr,flush=True,)print(f"Operation ID: {operation.name}",file=sys.stderr,flush=True)raiseoperation.exception()orRuntimeError(operation.error_message)ifoperation.warnings:print(f"Warnings during {verbose_name}:\n",file=sys.stderr,flush=True)forwarninginoperation.warnings:print(f" - {warning.code}: {warning.message}",file=sys.stderr,flush=True)returnresultdefcreate_consistency_group(project_id:str,region:str,group_name:str,group_description:str)-> compute_v1.ResourcePolicy:""" Creates a consistency group in Google Cloud Compute Engine. Args: project_id (str): The ID of the Google Cloud project. region (str): The region where the consistency group will be created. group_name (str): The name of the consistency group. group_description (str): The description of the consistency group. Returns: compute_v1.ResourcePolicy: The consistency group object """# Initialize the ResourcePoliciesClientclient=compute_v1.ResourcePoliciesClient()# Create the ResourcePolicy object with the provided name, description, and policyresource_policy_resource=compute_v1.ResourcePolicy(name=group_name,description=group_description,disk_consistency_group_policy=compute_v1.ResourcePolicyDiskConsistencyGroupPolicy(),)operation=client.insert(project=project_id,region=region,resource_policy_resource=resource_policy_resource,)wait_for_extended_operation(operation,"Consistency group creation")returnclient.get(project=project_id,region=region,resource_policy=group_name)

REST

Create a consistency group using the resourcePolicies.insert method:

 POST https://compute.googleapis.com/compute/v1/projects/PROJECT/regions/REGION/resourcePolicies { "name": "CONSISTENCY_GROUP_NAME", "diskConsistencyGroupPolicy": { } } 

Replace the following:

  • PROJECT: the project that contains the consistency group.
  • REGION: the region for the consistency group. If you want to add primary disks to consistency group, use the same region as the primary disks. If you want to add secondary disks to the consistency group, use the same region as the secondary disks.
  • CONSISTENCY_GROUP_NAME: the name for the consistency group.

Terraform

To create a consistency group, use the compute_resource_policy resource.

resource "google_compute_resource_policy" "default" { name = "test-consistency-group" region = "us-central1" disk_consistency_group_policy { enabled = true } }

To learn how to apply or remove a Terraform configuration, see Basic Terraform commands.

View disks in a consistency group

View disks in a consistency group using the Google Cloud console, the Google Cloud CLI, or REST.

Console

View the disks included in a consistency group by doing the following:

  1. In the Google Cloud console, go to the Asynchronous replication page.

    Go to Asynchronous replication

  2. Click the Consistency groups tab.

  3. Click the name of the consistency group that you want to view the disks for. The Manage consistency group page opens.

  4. View the Consistency group members section to see all disks included in the consistency group.

gcloud

View the disks included in a consistency group using the gcloud compute disks list command:

 gcloud compute disks list \ --LOCATION_FLAG=LOCATION \ --filter=resourcePolicies=CONSISTENCY_GROUP_NAME

Replace the following:

  • LOCATION_FLAG: the location flag for the disks in the consistency group. If the disks in the consistency group are regional, use --region. If the disks in the consistency group are zonal, use --zone.
  • LOCATION: the region or zone of the disks in the consistency group. For regional disks, use the region. For zonal disks, use the zone.
  • CONSISTENCY_GROUP_NAME: the name of the consistency group.

Go

import("context""fmt""io""strings"compute"cloud.google.com/go/compute/apiv1"computepb"cloud.google.com/go/compute/apiv1/computepb""google.golang.org/api/iterator")// listRegionalConsistencyGroup get list of disks in consistency group for a project in a given region.funclistRegionalConsistencyGroup(wio.Writer,projectID,region,groupNamestring)error{// projectID := "your_project_id"// region := "europe-west4"// groupName := "your_group_name"ifgroupName==""{returnfmt.Errorf("group name cannot be empty")}ctx:=context.Background()// To check for zonal disks in consistency group use compute.NewDisksRESTClientdisksClient,err:=compute.NewRegionDisksRESTClient(ctx)iferr!=nil{returnfmt.Errorf("NewRegionDisksRESTClient: %w",err)}deferdisksClient.Close()// If using zonal disk client, use computepb.ListDisksRequestreq:=&computepb.ListRegionDisksRequest{Project:projectID,Region:region,}it:=disksClient.List(ctx,req)for{disk,err:=it.Next()iferr==iterator.Done{break}iferr!=nil{returnerr}for_,diskPolicy:=rangedisk.GetResourcePolicies(){ifstrings.Contains(diskPolicy,groupName){fmt.Fprintf(w,"- %s\n",disk.GetName())}}}returnnil}

Java

List zonal disks in a consistency group

importcom.google.cloud.compute.v1.Disk;importcom.google.cloud.compute.v1.DisksClient;importcom.google.cloud.compute.v1.ListDisksRequest;importjava.io.IOException;importjava.util.ArrayList;importjava.util.List;importjava.util.concurrent.ExecutionException;publicclassListZonalDisksInConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,InterruptedException,ExecutionException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project you want to use.Stringproject="YOUR_PROJECT_ID";// Name of the consistency group.StringconsistencyGroupName="CONSISTENCY_GROUP_ID";// Zone of the disk.StringdisksLocation="us-central1-a";// Region of the consistency group.StringconsistencyGroupLocation="us-central1";listZonalDisksInConsistencyGroup(project,consistencyGroupName,consistencyGroupLocation,disksLocation);}// Lists disks in a consistency group.publicstaticList<Disk>listZonalDisksInConsistencyGroup(Stringproject,StringconsistencyGroupName,StringconsistencyGroupLocation,StringdisksLocation)throwsIOException{Stringfilter=String.format("https://www.googleapis.com/compute/v1/projects/%s/regions/%s/resourcePolicies/%s",project,consistencyGroupLocation,consistencyGroupName);List<Disk>disksList=newArrayList<>();// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(DisksClientdisksClient=DisksClient.create()){ListDisksRequestrequest=ListDisksRequest.newBuilder().setProject(project).setZone(disksLocation).build();DisksClient.ListPagedResponseresponse=disksClient.list(request);for(Diskdisk:response.iterateAll()){if(disk.getResourcePoliciesList().contains(filter)){disksList.add(disk);}}}System.out.println(disksList.size());returndisksList;}}

List regional disks in a consistency group

importcom.google.cloud.compute.v1.Disk;importcom.google.cloud.compute.v1.ListRegionDisksRequest;importcom.google.cloud.compute.v1.RegionDisksClient;importjava.io.IOException;importjava.util.ArrayList;importjava.util.List;importjava.util.concurrent.ExecutionException;publicclassListRegionalDisksInConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,InterruptedException,ExecutionException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project you want to use.Stringproject="YOUR_PROJECT_ID";// Name of the consistency group.StringconsistencyGroupName="CONSISTENCY_GROUP_ID";// Region of the disk.StringdisksLocation="us-central1";// Region of the consistency group.StringconsistencyGroupLocation="us-central1";listRegionalDisksInConsistencyGroup(project,consistencyGroupName,consistencyGroupLocation,disksLocation);}// Lists disks in a consistency group.publicstaticList<Disk>listRegionalDisksInConsistencyGroup(Stringproject,StringconsistencyGroupName,StringconsistencyGroupLocation,StringdisksLocation)throwsIOException{Stringfilter=String.format("https://www.googleapis.com/compute/v1/projects/%s/regions/%s/resourcePolicies/%s",project,consistencyGroupLocation,consistencyGroupName);List<Disk>disksList=newArrayList<>();// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(RegionDisksClientdisksClient=RegionDisksClient.create()){ListRegionDisksRequestrequest=ListRegionDisksRequest.newBuilder().setProject(project).setRegion(disksLocation).build();RegionDisksClient.ListPagedResponseresponse=disksClient.list(request);for(Diskdisk:response.iterateAll()){if(disk.getResourcePoliciesList().contains(filter)){disksList.add(disk);}}}System.out.println(disksList.size());returndisksList;}}

Node.js

// Import the Compute libraryconstcomputeLib=require('@google-cloud/compute');// If you want to get regional disks, you should use: RegionDisksClient.// Instantiate a disksClientconstdisksClient=newcomputeLib.DisksClient();/** * TODO(developer): Update/uncomment these variables before running the sample. */// The project that contains the disks.constprojectId=awaitdisksClient.getProjectId();// If you use RegionDisksClient- define region, if DisksClient- define zone.// The zone or region of the disks.// disksLocation = 'europe-central2-a';// The name of the consistency group.// consistencyGroupName = 'consistency-group-name';// The region of the consistency group.// consistencyGroupLocation = 'europe-central2';asyncfunctioncallConsistencyGroupDisksList(){constfilter=`https://www.googleapis.com/compute/v1/projects/${projectId}/regions/${consistencyGroupLocation}/resourcePolicies/${consistencyGroupName}`;const[response]=awaitdisksClient.list({project:projectId,// If you use RegionDisksClient, pass region as an argument instead of zone.zone:disksLocation,});// Filtering must be done manually for now, since list filtering inside disksClient.list is not supported yet.constfilteredDisks=response.filter(disk=> disk.resourcePolicies.includes(filter));console.log(JSON.stringify(filteredDisks));}awaitcallConsistencyGroupDisksList();

Python

fromgoogle.cloudimportcompute_v1deflist_disks_consistency_group(project_id:str,disk_location:str,consistency_group_name:str,consistency_group_region:str,)-> list:""" Lists disks that are part of a specified consistency group. Args: project_id (str): The ID of the Google Cloud project. disk_location (str): The region or zone of the disk disk_region_flag (bool): Flag indicating if the disk is regional. consistency_group_name (str): The name of the consistency group. consistency_group_region (str): The region of the consistency group. Returns: list: A list of disks that are part of the specified consistency group. """consistency_group_link=(f"https://www.googleapis.com/compute/v1/projects/{project_id}/regions/"f"{consistency_group_region}/resourcePolicies/{consistency_group_name}")# If the final character of the disk_location is a digit, it is a regional diskifdisk_location[-1].isdigit():region_client=compute_v1.RegionDisksClient()disks=region_client.list(project=project_id,region=disk_location)# For zonal disks we use DisksClientelse:client=compute_v1.DisksClient()disks=client.list(project=project_id,zone=disk_location)return[diskfordiskindisksifconsistency_group_linkindisk.resource_policies]

REST

View the disks in a consistency group by using a query filter with one of the following methods:

  • View zonal disks in a consistency group using the disks.get method:

     GET https://compute.googleapis.com/compute/v1/projects/PROJECT/zones/ZONE/disks?filter=resourcePolicies%3DCONSISTENCY_GROUP_NAME
  • View regional disks in a consistency group using the regionDisks.get method:

     GET https://compute.googleapis.com/compute/v1/projects/PROJECT/regions/REGION/disks?filter=resourcePolicies%3DCONSISTENCY_GROUP_NAME

Replace the following:

  • PROJECT: the project that contains the consistency group
  • ZONE: the zone of the disks in the consistency group
  • REGION: the region of the disks in the consistency group
  • CONSISTENCY_GROUP_NAME: the name of the consistency group

Add a disk to a consistency group

If you want to add primary disks to a consistency group, you must add disks to the consistency group before you start replication. You can add secondary disks to a consistency group at any time. All disks in a consistency group must be in the same zone, for zonal disks, or in the same pair of zones, for regional disks.

Add a disk to a consistency group using the Google Cloud console, the Google Cloud CLI, REST, or Terraform.

Console

Add disks to a consistency group by doing the following:

  1. In the Google Cloud console, go to the Asynchronous replication page.

    Go to Asynchronous replication

  2. Click the Consistency groups tab.

  3. Click the name of the consistency group that you want to add disks to. The Manage consistency group page opens.

  4. Click Assign disks. The Assign disks page opens.

  5. Select the disks that you want to add to the consistency group.

  6. Click Assign disks. When prompted, click Add.

gcloud

Add a disk to a consistency group using the gcloud compute disks add-resource-policiescommand:

 gcloud compute disks add-resource-policies DISK_NAME \ --LOCATION_FLAG=LOCATION \ --resource-policies=CONSISTENCY_GROUP

Replace the following:

  • DISK_NAME: the name of the disk to add to the consistency group.
  • LOCATION_FLAG: the location flag for the disk. For a regional disk, use --region. For a zonal disk, use --zone.
  • LOCATION: the region or zone of the disk. For regional disks, use the region. For zonal disks, use the zone.
  • CONSISTENCY_GROUP: the URL of the consistency group. For example, projects/PROJECT/regions/REGION/resourcePolicies/CONSISTENCY_GROUP_NAME.

Go

import("context""fmt""io"compute"cloud.google.com/go/compute/apiv1"computepb"cloud.google.com/go/compute/apiv1/computepb")// addDiskConsistencyGroup adds a disk to a consistency group for a project in a given region.funcaddDiskConsistencyGroup(wio.Writer,projectID,region,groupName,diskNamestring)error{// projectID := "your_project_id"// region := "europe-west4"// diskName := "your_disk_name"// groupName := "your_group_name"ctx:=context.Background()disksClient,err:=compute.NewRegionDisksRESTClient(ctx)iferr!=nil{returnfmt.Errorf("NewResourcePoliciesRESTClient: %w",err)}deferdisksClient.Close()consistencyGroupUrl:=fmt.Sprintf("projects/%s/regions/%s/resourcePolicies/%s",projectID,region,groupName)req:=&computepb.AddResourcePoliciesRegionDiskRequest{Project:projectID,Disk:diskName,RegionDisksAddResourcePoliciesRequestResource:&computepb.RegionDisksAddResourcePoliciesRequest{ResourcePolicies:[]string{consistencyGroupUrl},},Region:region,}op,err:=disksClient.AddResourcePolicies(ctx,req)iferr!=nil{returnfmt.Errorf("unable to add disk: %w",err)}iferr=op.Wait(ctx);err!=nil{returnfmt.Errorf("unable to wait for the operation: %w",err)}fmt.Fprintf(w,"Disk added\n")returnnil}

Java

importcom.google.cloud.compute.v1.AddResourcePoliciesDiskRequest;importcom.google.cloud.compute.v1.AddResourcePoliciesRegionDiskRequest;importcom.google.cloud.compute.v1.DisksAddResourcePoliciesRequest;importcom.google.cloud.compute.v1.DisksClient;importcom.google.cloud.compute.v1.Operation;importcom.google.cloud.compute.v1.Operation.Status;importcom.google.cloud.compute.v1.RegionDisksAddResourcePoliciesRequest;importcom.google.cloud.compute.v1.RegionDisksClient;importjava.io.IOException;importjava.util.Arrays;importjava.util.concurrent.ExecutionException;importjava.util.concurrent.TimeUnit;importjava.util.concurrent.TimeoutException;publicclassAddDiskToConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project that contains the disk.Stringproject="YOUR_PROJECT_ID";// Zone or region of the disk.Stringlocation="us-central1";// Name of the disk.StringdiskName="DISK_NAME";// Name of the consistency group.StringconsistencyGroupName="CONSISTENCY_GROUP";// Region of the consistency group.StringconsistencyGroupLocation="us-central1";addDiskToConsistencyGroup(project,location,diskName,consistencyGroupName,consistencyGroupLocation);}// Adds a disk to a consistency group.publicstaticStatusaddDiskToConsistencyGroup(Stringproject,Stringlocation,StringdiskName,StringconsistencyGroupName,StringconsistencyGroupLocation)throwsIOException,ExecutionException,InterruptedException,TimeoutException{StringconsistencyGroupUrl=String.format("https://www.googleapis.com/compute/v1/projects/%s/regions/%s/resourcePolicies/%s",project,consistencyGroupLocation,consistencyGroupName);Operationresponse;if(Character.isDigit(location.charAt(location.length()-1))){// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(RegionDisksClientdisksClient=RegionDisksClient.create()){AddResourcePoliciesRegionDiskRequestrequest=AddResourcePoliciesRegionDiskRequest.newBuilder().setDisk(diskName).setRegion(location).setProject(project).setRegionDisksAddResourcePoliciesRequestResource(RegionDisksAddResourcePoliciesRequest.newBuilder().addAllResourcePolicies(Arrays.asList(consistencyGroupUrl)).build()).build();response=disksClient.addResourcePoliciesAsync(request).get(1,TimeUnit.MINUTES);}}else{try(DisksClientdisksClient=DisksClient.create()){AddResourcePoliciesDiskRequestrequest=AddResourcePoliciesDiskRequest.newBuilder().setDisk(diskName).setZone(location).setProject(project).setDisksAddResourcePoliciesRequestResource(DisksAddResourcePoliciesRequest.newBuilder().addAllResourcePolicies(Arrays.asList(consistencyGroupUrl)).build()).build();response=disksClient.addResourcePoliciesAsync(request).get(1,TimeUnit.MINUTES);}}if(response.hasError()){thrownewError("Error adding disk to consistency group! "+response.getError());}returnresponse.getStatus();}}

Node.js

// Import the Compute libraryconstcomputeLib=require('@google-cloud/compute');constcompute=computeLib.protos.google.cloud.compute.v1;// If you want to add regional disk,// you should use: RegionDisksClient and RegionOperationsClient.// Instantiate a disksClientconstdisksClient=newcomputeLib.DisksClient();// Instantiate a zoneconstzoneOperationsClient=newcomputeLib.ZoneOperationsClient();/** * TODO(developer): Update/uncomment these variables before running the sample. */// The project that contains the disk.constprojectId=awaitdisksClient.getProjectId();// The name of the disk.// diskName = 'disk-name';// If you use RegionDisksClient- define region, if DisksClient- define zone.// The zone or region of the disk.// diskLocation = 'europe-central2-a';// The name of the consistency group.// consistencyGroupName = 'consistency-group-name';// The region of the consistency group.// consistencyGroupLocation = 'europe-central2';asyncfunctioncallAddDiskToConsistencyGroup(){const[response]=awaitdisksClient.addResourcePolicies({disk:diskName,project:projectId,// If you use RegionDisksClient, pass region as an argument instead of zone.zone:diskLocation,disksAddResourcePoliciesRequestResource:newcompute.DisksAddResourcePoliciesRequest({resourcePolicies:[`https://www.googleapis.com/compute/v1/projects/${projectId}/regions/${consistencyGroupLocation}/resourcePolicies/${consistencyGroupName}`,],}),});letoperation=response.latestResponse;// Wait for the add disk operation to complete.while(operation.status!=='DONE'){[operation]=awaitzoneOperationsClient.wait({operation:operation.name,project:projectId,// If you use RegionDisksClient, pass region as an argument instead of zone.zone:operation.zone.split('/').pop(),});}console.log(`Disk: ${diskName} added to consistency group: ${consistencyGroupName}.`);}awaitcallAddDiskToConsistencyGroup();

Python

fromgoogle.cloudimportcompute_v1defadd_disk_consistency_group(project_id:str,disk_name:str,disk_location:str,consistency_group_name:str,consistency_group_region:str,)-> None:"""Adds a disk to a specified consistency group. Args: project_id (str): The ID of the Google Cloud project. disk_name (str): The name of the disk to be added. disk_location (str): The region or zone of the disk consistency_group_name (str): The name of the consistency group. consistency_group_region (str): The region of the consistency group. Returns: None """consistency_group_link=(f"regions/{consistency_group_region}/resourcePolicies/{consistency_group_name}")# Checking if the disk is zonal or regional# If the final character of the disk_location is a digit, it is a regional diskifdisk_location[-1].isdigit():policy=compute_v1.RegionDisksAddResourcePoliciesRequest(resource_policies=[consistency_group_link])disk_client=compute_v1.RegionDisksClient()disk_client.add_resource_policies(project=project_id,region=disk_location,disk=disk_name,region_disks_add_resource_policies_request_resource=policy,)# For zonal disks we use DisksClientelse:print("Using DisksClient")policy=compute_v1.DisksAddResourcePoliciesRequest(resource_policies=[consistency_group_link])disk_client=compute_v1.DisksClient()disk_client.add_resource_policies(project=project_id,zone=disk_location,disk=disk_name,disks_add_resource_policies_request_resource=policy,)print(f"Disk {disk_name} added to consistency group {consistency_group_name}")

REST

Add disks to a consistency group using one of the following methods:

  • Add zonal disks to a consistency group using the disks.addResourcePolicies method:

     POST https://compute.googleapis.com/compute/v1/projects/PROJECT/zones/LOCATION/disks/DISK_NAME/addResourcePolicies { "resourcePolicies": "CONSISTENCY_GROUP" } 
  • Add regional disks to a consistency group using the regionDisks.addResourcePolicies method:

     POST https://compute.googleapis.com/compute/v1/projects/PROJECT/regions/LOCATION/disks/DISK_NAME/addResourcePolicies { "resourcePolicies": "CONSISTENCY_GROUP" } 

Replace the following:

  • PROJECT: the project that contains the disk.
  • LOCATION: the zone or region of the disk. For zonal disks, use the zone. For regional disks, use the region.
  • DISK_NAME: the name of the disk to add to the consistency group.
  • CONSISTENCY_GROUP: the URL of the consistency group. For example, projects/PROJECT/regions/REGION/resourcePolicies/CONSISTENCY_GROUP_NAME.

Terraform

To add the disk to the consistency group, use the compute_disk_resource_policy_attachment resource.

  • In case of Regional disk specify region in place of zone.

    resource "google_compute_disk_resource_policy_attachment" "default" { name = google_compute_resource_policy.default.name disk = google_compute_disk.default.name zone = "us-central1-a" }

    To learn how to apply or remove a Terraform configuration, see Basic Terraform commands.

Remove a disk from a consistency group

Before you can remove a disk from a consistency group, you must stop replication for the disk.

Remove a disk from a consistency group using the Google Cloud console, the Google Cloud CLI, or REST.

Console

Remove primary disks from a consistency group by doing the following:

  1. In the Google Cloud console, go to the Asynchronous replication page.

    Go to Asynchronous replication

  2. Click the Consistency groups tab.

  3. Click the name of the consistency group that you want to add disks to. The Manage consistency group page opens.

  4. Select the disks that you want to remove from the consistency group.

  5. Click Remove disks. When prompted, click Remove.

gcloud

Remove a disk from a consistency group using the gcloud compute disks remove-resource-policies command:

 gcloud compute disks remove-resource-policies DISK_NAME \ --LOCATION_FLAG=LOCATION \ --resource-policies=CONSISTENCY_GROUP

Replace the following:

  • DISK_NAME: the name of the disk to remove from the consistency group.
  • LOCATION_FLAG: the location flag for the disk. For a regional disk, use --region. For a zonal disk, use --zone.
  • LOCATION: the region or zone of the disk. For regional disks, use the region. For zonal disks, use the zone.
  • CONSISTENCY_GROUP: the URL of the consistency group. For example, projects/PROJECT/regions/REGION/resourcePolicies/CONSISTENCY_GROUP_NAME.

Go

import("context""fmt""io"compute"cloud.google.com/go/compute/apiv1"computepb"cloud.google.com/go/compute/apiv1/computepb")// removeDiskConsistencyGroup removes a disk from consistency group for a project in a given region.funcremoveDiskConsistencyGroup(wio.Writer,projectID,region,groupName,diskNamestring)error{// projectID := "your_project_id"// region := "europe-west4"// diskName := "your_disk_name"// groupName := "your_group_name"ctx:=context.Background()disksClient,err:=compute.NewRegionDisksRESTClient(ctx)iferr!=nil{returnfmt.Errorf("NewResourcePoliciesRESTClient: %w",err)}deferdisksClient.Close()consistencyGroupUrl:=fmt.Sprintf("projects/%s/regions/%s/resourcePolicies/%s",projectID,region,groupName)req:=&computepb.RemoveResourcePoliciesRegionDiskRequest{Project:projectID,Disk:diskName,RegionDisksRemoveResourcePoliciesRequestResource:&computepb.RegionDisksRemoveResourcePoliciesRequest{ResourcePolicies:[]string{consistencyGroupUrl},},Region:region,}op,err:=disksClient.RemoveResourcePolicies(ctx,req)iferr!=nil{returnfmt.Errorf("unable to remove disk: %w",err)}iferr=op.Wait(ctx);err!=nil{returnfmt.Errorf("unable to wait for the operation: %w",err)}fmt.Fprintf(w,"Disk removed\n")returnnil}

Java

importcom.google.cloud.compute.v1.DisksClient;importcom.google.cloud.compute.v1.DisksRemoveResourcePoliciesRequest;importcom.google.cloud.compute.v1.Operation;importcom.google.cloud.compute.v1.Operation.Status;importcom.google.cloud.compute.v1.RegionDisksClient;importcom.google.cloud.compute.v1.RegionDisksRemoveResourcePoliciesRequest;importcom.google.cloud.compute.v1.RemoveResourcePoliciesDiskRequest;importcom.google.cloud.compute.v1.RemoveResourcePoliciesRegionDiskRequest;importjava.io.IOException;importjava.util.Arrays;importjava.util.concurrent.ExecutionException;importjava.util.concurrent.TimeUnit;importjava.util.concurrent.TimeoutException;publicclassRemoveDiskFromConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project that contains the disk.Stringproject="YOUR_PROJECT_ID";// Zone or region of the disk.Stringlocation="us-central1";// Name of the disk.StringdiskName="DISK_NAME";// Name of the consistency group.StringconsistencyGroupName="CONSISTENCY_GROUP";// Region of the consistency group.StringconsistencyGroupLocation="us-central1";removeDiskFromConsistencyGroup(project,location,diskName,consistencyGroupName,consistencyGroupLocation);}// Removes a disk from a consistency group.publicstaticStatusremoveDiskFromConsistencyGroup(Stringproject,Stringlocation,StringdiskName,StringconsistencyGroupName,StringconsistencyGroupLocation)throwsIOException,ExecutionException,InterruptedException,TimeoutException{StringconsistencyGroupUrl=String.format("https://www.googleapis.com/compute/v1/projects/%s/regions/%s/resourcePolicies/%s",project,consistencyGroupLocation,consistencyGroupName);Operationresponse;if(Character.isDigit(location.charAt(location.length()-1))){// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(RegionDisksClientdisksClient=RegionDisksClient.create()){RemoveResourcePoliciesRegionDiskRequestrequest=RemoveResourcePoliciesRegionDiskRequest.newBuilder().setDisk(diskName).setRegion(location).setProject(project).setRegionDisksRemoveResourcePoliciesRequestResource(RegionDisksRemoveResourcePoliciesRequest.newBuilder().addAllResourcePolicies(Arrays.asList(consistencyGroupUrl)).build()).build();response=disksClient.removeResourcePoliciesAsync(request).get(1,TimeUnit.MINUTES);}}else{try(DisksClientdisksClient=DisksClient.create()){RemoveResourcePoliciesDiskRequestrequest=RemoveResourcePoliciesDiskRequest.newBuilder().setDisk(diskName).setZone(location).setProject(project).setDisksRemoveResourcePoliciesRequestResource(DisksRemoveResourcePoliciesRequest.newBuilder().addAllResourcePolicies(Arrays.asList(consistencyGroupUrl)).build()).build();response=disksClient.removeResourcePoliciesAsync(request).get(1,TimeUnit.MINUTES);}}if(response.hasError()){thrownewError("Error removing disk from consistency group! "+response.getError());}returnresponse.getStatus();}}

Node.js

// Import the Compute libraryconstcomputeLib=require('@google-cloud/compute');constcompute=computeLib.protos.google.cloud.compute.v1;// If you want to remove regional disk,// you should use: RegionDisksClient and RegionOperationsClient.// Instantiate a disksClientconstdisksClient=newcomputeLib.DisksClient();// Instantiate a zoneconstzoneOperationsClient=newcomputeLib.ZoneOperationsClient();/** * TODO(developer): Update/uncomment these variables before running the sample. */// The project that contains the disk.constprojectId=awaitdisksClient.getProjectId();// The name of the disk.// diskName = 'disk-name';// If you use RegionDisksClient- define region, if DisksClient- define zone.// The zone or region of the disk.// diskLocation = 'europe-central2-a';// The name of the consistency group.// consistencyGroupName = 'consistency-group-name';// The region of the consistency group.// consistencyGroupLocation = 'europe-central2';asyncfunctioncallDeleteDiskFromConsistencyGroup(){const[response]=awaitdisksClient.removeResourcePolicies({disk:diskName,project:projectId,// If you use RegionDisksClient, pass region as an argument instead of zone.zone:diskLocation,disksRemoveResourcePoliciesRequestResource:newcompute.DisksRemoveResourcePoliciesRequest({resourcePolicies:[`https://www.googleapis.com/compute/v1/projects/${projectId}/regions/${consistencyGroupLocation}/resourcePolicies/${consistencyGroupName}`,],}),});letoperation=response.latestResponse;// Wait for the delete disk operation to complete.while(operation.status!=='DONE'){[operation]=awaitzoneOperationsClient.wait({operation:operation.name,project:projectId,// If you use RegionDisksClient, pass region as an argument instead of zone.zone:operation.zone.split('/').pop(),});}console.log(`Disk: ${diskName} deleted from consistency group: ${consistencyGroupName}.`);}awaitcallDeleteDiskFromConsistencyGroup();

Python

fromgoogle.cloudimportcompute_v1defremove_disk_consistency_group(project_id:str,disk_name:str,disk_location:str,consistency_group_name:str,consistency_group_region:str,)-> None:"""Removes a disk from a specified consistency group. Args: project_id (str): The ID of the Google Cloud project. disk_name (str): The name of the disk to be deleted. disk_location (str): The region or zone of the disk consistency_group_name (str): The name of the consistency group. consistency_group_region (str): The region of the consistency group. Returns: None """consistency_group_link=(f"regions/{consistency_group_region}/resourcePolicies/{consistency_group_name}")# Checking if the disk is zonal or regional# If the final character of the disk_location is a digit, it is a regional diskifdisk_location[-1].isdigit():policy=compute_v1.RegionDisksRemoveResourcePoliciesRequest(resource_policies=[consistency_group_link])disk_client=compute_v1.RegionDisksClient()disk_client.remove_resource_policies(project=project_id,region=disk_location,disk=disk_name,region_disks_remove_resource_policies_request_resource=policy,)# For zonal disks we use DisksClientelse:policy=compute_v1.DisksRemoveResourcePoliciesRequest(resource_policies=[consistency_group_link])disk_client=compute_v1.DisksClient()disk_client.remove_resource_policies(project=project_id,zone=disk_location,disk=disk_name,disks_remove_resource_policies_request_resource=policy,)print(f"Disk {disk_name} removed from consistency group {consistency_group_name}")

REST

Remove a disk from a consistency group using the disks.removeResourcePolicies method for zonal disks, or the regionDisks.removeResourcePolicies method for regional disks.

  • Remove a zonal disk from a consistency group:

     POST https://compute.googleapis.com/compute/v1/projects/PROJECT/zones/LOCATION/disks/DISK_NAME/removeResourcePolicies { "resourcePolicies": "CONSISTENCY_GROUP" } 
  • Remove a regional disk from a consistency group:

     POST https://compute.googleapis.com/compute/v1/projects/PROJECT/regions/LOCATION/disks/DISK_NAME/removeResourcePolicies { "resourcePolicies": "CONSISTENCY_GROUP" } 

Replace the following:

  • PROJECT: the project that contains the disk.
  • LOCATION: the zone or region of the disk. For zonal disks, use the zone. For regional disks, use the region.
  • DISK_NAME: the name of the disk to remove from the consistency group.
  • CONSISTENCY_GROUP: the URL of the consistency group. For example, projects/PROJECT/regions/REGION/resourcePolicies/CONSISTENCY_GROUP_NAME.

Delete a consistency group

Delete a consistency group using the Google Cloud console, the Google Cloud CLI, or REST.

Console

Delete a consistency by doing the following:

  1. In the Google Cloud console, go to the Asynchronous replication page.

    Go to Asynchronous replication

  2. Click the Consistency groups tab.

  3. Select the consistency group that you want to delete.

  4. Click Delete. The Delete consistency group window opens.

  5. Click Delete.

gcloud

Delete the resource policy using the gcloud compute resource-policies delete command:

 gcloud compute resource-policies delete CONSISTENCY_GROUP \ --region=REGION

Replace the following:

  • CONSISTENCY_GROUP: the name of the consistency group
  • REGION: the region of the consistency group

Go

import("context""fmt""io"compute"cloud.google.com/go/compute/apiv1"computepb"cloud.google.com/go/compute/apiv1/computepb")// deleteConsistencyGroup deletes consistency group for a project in a given region.funcdeleteConsistencyGroup(wio.Writer,projectID,region,groupNamestring)error{// projectID := "your_project_id"// region := "europe-west4"// groupName := "your_group_name"ctx:=context.Background()disksClient,err:=compute.NewResourcePoliciesRESTClient(ctx)iferr!=nil{returnfmt.Errorf("NewResourcePoliciesRESTClient: %w",err)}deferdisksClient.Close()req:=&computepb.DeleteResourcePolicyRequest{Project:projectID,ResourcePolicy:groupName,Region:region,}op,err:=disksClient.Delete(ctx,req)iferr!=nil{returnfmt.Errorf("unable to delete group: %w",err)}iferr=op.Wait(ctx);err!=nil{returnfmt.Errorf("unable to wait for the operation: %w",err)}fmt.Fprintf(w,"Group deleted\n")returnnil}

Java

importcom.google.cloud.compute.v1.Operation;importcom.google.cloud.compute.v1.Operation.Status;importcom.google.cloud.compute.v1.ResourcePoliciesClient;importjava.io.IOException;importjava.util.concurrent.ExecutionException;importjava.util.concurrent.TimeUnit;importjava.util.concurrent.TimeoutException;publicclassDeleteConsistencyGroup{publicstaticvoidmain(String[]args)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// TODO(developer): Replace these variables before running the sample.// Project ID or project number of the Cloud project you want to use.Stringproject="YOUR_PROJECT_ID";// Region in which your consistency group is located.Stringregion="us-central1";// Name of the consistency group you want to delete.StringconsistencyGroupName="YOUR_CONSISTENCY_GROUP_NAME";deleteConsistencyGroup(project,region,consistencyGroupName);}// Deletes a consistency group resource policy in the specified project and region.publicstaticStatusdeleteConsistencyGroup(Stringproject,Stringregion,StringconsistencyGroupName)throwsIOException,ExecutionException,InterruptedException,TimeoutException{// Initialize client that will be used to send requests. This client only needs to be created// once, and can be reused for multiple requests.try(ResourcePoliciesClientresourcePoliciesClient=ResourcePoliciesClient.create()){Operationresponse=resourcePoliciesClient.deleteAsync(project,region,consistencyGroupName).get(1,TimeUnit.MINUTES);if(response.hasError()){thrownewError("Error deleting disk! "+response.getError());}returnresponse.getStatus();}}}

Node.js

// Import the Compute libraryconstcomputeLib=require('@google-cloud/compute');// Instantiate a resourcePoliciesClientconstresourcePoliciesClient=newcomputeLib.ResourcePoliciesClient();// Instantiate a regionOperationsClientconstregionOperationsClient=newcomputeLib.RegionOperationsClient();/** * TODO(developer): Update/uncomment these variables before running the sample. */// The project that contains the consistency group.constprojectId=awaitresourcePoliciesClient.getProjectId();// The region of the consistency group.// region = 'europe-central2';// The name of the consistency group.// consistencyGroupName = 'consistency-group-name';asyncfunctioncallDeleteConsistencyGroup(){// Delete a resourcePolicyResourceconst[response]=awaitresourcePoliciesClient.delete({project:projectId,region,resourcePolicy:consistencyGroupName,});letoperation=response.latestResponse;// Wait for the delete group operation to complete.while(operation.status!=='DONE'){[operation]=awaitregionOperationsClient.wait({operation:operation.name,project:projectId,region,});}console.log(`Consistency group: ${consistencyGroupName} deleted.`);}awaitcallDeleteConsistencyGroup();

Python

from__future__importannotationsimportsysfromtypingimportAnyfromgoogle.api_core.extended_operationimportExtendedOperationfromgoogle.cloudimportcompute_v1defwait_for_extended_operation(operation:ExtendedOperation,verbose_name:str="operation",timeout:int=300)-> Any:""" Waits for the extended (long-running) operation to complete. If the operation is successful, it will return its result. If the operation ends with an error, an exception will be raised. If there were any warnings during the execution of the operation they will be printed to sys.stderr. Args: operation: a long-running operation you want to wait on. verbose_name: (optional) a more verbose name of the operation, used only during error and warning reporting. timeout: how long (in seconds) to wait for operation to finish. If None, wait indefinitely. Returns: Whatever the operation.result() returns. Raises: This method will raise the exception received from `operation.exception()` or RuntimeError if there is no exception set, but there is an `error_code` set for the `operation`. In case of an operation taking longer than `timeout` seconds to complete, a `concurrent.futures.TimeoutError` will be raised. """result=operation.result(timeout=timeout)ifoperation.error_code:print(f"Error during {verbose_name}: [Code: {operation.error_code}]: {operation.error_message}",file=sys.stderr,flush=True,)print(f"Operation ID: {operation.name}",file=sys.stderr,flush=True)raiseoperation.exception()orRuntimeError(operation.error_message)ifoperation.warnings:print(f"Warnings during {verbose_name}:\n",file=sys.stderr,flush=True)forwarninginoperation.warnings:print(f" - {warning.code}: {warning.message}",file=sys.stderr,flush=True)returnresultdefdelete_consistency_group(project_id:str,region:str,group_name:str)-> None:""" Deletes a consistency group in Google Cloud Compute Engine. Args: project_id (str): The ID of the Google Cloud project. region (str): The region where the consistency group is located. group_name (str): The name of the consistency group to delete. Returns: None """# Initialize the ResourcePoliciesClientclient=compute_v1.ResourcePoliciesClient()# Delete the (consistency group) from the specified project and regionoperation=client.delete(project=project_id,region=region,resource_policy=group_name,)wait_for_extended_operation(operation,"Consistency group deletion")

REST

Delete a consistency using the resourcePolicies.delete method:

 DELETE https://compute.googleapis.com/compute/v1/projects/PROJECT/regions/REGION/resourcePolicies/CONSISTENCY_GROUP_NAME

Replace the following:

  • PROJECT: the project that contains the consistency group
  • REGION: the region of the consistency group
  • CONSISTENCY_GROUP: the name of the consistency group

What's next