mirror of
https://github.com/letic/terraform-provider-google.git
synced 2024-07-03 08:42:39 +00:00
changes to (r)igm so that resources can be converted from beta back to ga (#2260)
<!-- This change is generated by MagicModules. --> /cc @danawillow
This commit is contained in:
parent
b76bc0331b
commit
bd348c5797
|
@ -3,6 +3,7 @@ package google
|
|||
import (
|
||||
"fmt"
|
||||
"log"
|
||||
"regexp"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
|
@ -14,6 +15,11 @@ import (
|
|||
"google.golang.org/api/compute/v1"
|
||||
)
|
||||
|
||||
var (
|
||||
instanceGroupManagerIdRegex = regexp.MustCompile("^" + ProjectRegex + "/[a-z0-9-]+/[a-z0-9-]+$")
|
||||
instanceGroupManagerIdNameRegex = regexp.MustCompile("^[a-z0-9-]+$")
|
||||
)
|
||||
|
||||
func resourceComputeInstanceGroupManager() *schema.Resource {
|
||||
return &schema.Resource{
|
||||
Create: resourceComputeInstanceGroupManagerCreate,
|
||||
|
@ -322,7 +328,7 @@ func resourceComputeInstanceGroupManagerCreate(d *schema.ResourceData, meta inte
|
|||
}
|
||||
|
||||
// It probably maybe worked, so store the ID now
|
||||
d.SetId(manager.Name)
|
||||
d.SetId(instanceGroupManagerId{Project: project, Zone: zone, Name: manager.Name}.terraformId())
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Creating InstanceGroupManager")
|
||||
|
@ -372,39 +378,45 @@ func flattenFixedOrPercent(fixedOrPercent *computeBeta.FixedOrPercent) []map[str
|
|||
|
||||
func getManager(d *schema.ResourceData, meta interface{}) (*computeBeta.InstanceGroupManager, error) {
|
||||
config := meta.(*Config)
|
||||
|
||||
project, err := getProject(d, config)
|
||||
zonalID, err := parseInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
region, err := getRegion(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
if zonalID.Project == "" {
|
||||
project, err := getProject(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
zonalID.Project = project
|
||||
}
|
||||
if zonalID.Zone == "" {
|
||||
zonalID.Zone, _ = getZone(d, config)
|
||||
}
|
||||
|
||||
getInstanceGroupManager := func(zone string) (interface{}, error) {
|
||||
return config.clientComputeBeta.InstanceGroupManagers.Get(project, zone, d.Id()).Do()
|
||||
return config.clientComputeBeta.InstanceGroupManagers.Get(zonalID.Project, zone, zonalID.Name).Do()
|
||||
}
|
||||
|
||||
var manager *computeBeta.InstanceGroupManager
|
||||
var e error
|
||||
if zone, _ := getZone(d, config); zone != "" {
|
||||
manager, e = config.clientComputeBeta.InstanceGroupManagers.Get(project, zone, d.Id()).Do()
|
||||
|
||||
if e != nil {
|
||||
return nil, handleNotFoundError(e, d, fmt.Sprintf("Instance Group Manager %q", d.Get("name").(string)))
|
||||
}
|
||||
} else {
|
||||
if zonalID.Zone == "" {
|
||||
// If the resource was imported, the only info we have is the ID. Try to find the resource
|
||||
// by searching in the region of the project.
|
||||
var resource interface{}
|
||||
resource, e = getZonalBetaResourceFromRegion(getInstanceGroupManager, region, config.clientComputeBeta, project)
|
||||
if e != nil {
|
||||
return nil, e
|
||||
region, err := getRegion(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
resource, err := getZonalBetaResourceFromRegion(getInstanceGroupManager, region, config.clientComputeBeta, zonalID.Project)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if resource != nil {
|
||||
manager = resource.(*computeBeta.InstanceGroupManager)
|
||||
}
|
||||
} else {
|
||||
manager, err = config.clientComputeBeta.InstanceGroupManagers.Get(zonalID.Project, zonalID.Zone, zonalID.Name).Do()
|
||||
if err != nil {
|
||||
return nil, handleNotFoundError(err, d, fmt.Sprintf("Instance Group Manager %q", zonalID.Name))
|
||||
}
|
||||
|
||||
manager = resource.(*computeBeta.InstanceGroupManager)
|
||||
}
|
||||
|
||||
if manager == nil {
|
||||
|
@ -528,14 +540,21 @@ func performZoneUpdate(config *Config, id string, updateStrategy string, rolling
|
|||
func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
project, err := getProject(d, config)
|
||||
zonalID, err := parseInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
zone, err := getZone(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
if zonalID.Project == "" {
|
||||
zonalID.Project, err = getProject(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
if zonalID.Zone == "" {
|
||||
zonalID.Zone, err = getZone(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
d.Partial(true)
|
||||
|
@ -555,14 +574,14 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
}
|
||||
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.SetTargetPools(
|
||||
project, zone, d.Id(), setTargetPools).Do()
|
||||
zonalID.Project, zonalID.Zone, zonalID.Name, setTargetPools).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating InstanceGroupManager: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Updating InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Updating InstanceGroupManager")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -581,14 +600,14 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
|
||||
// Make the request:
|
||||
op, err := config.clientComputeBeta.InstanceGroups.SetNamedPorts(
|
||||
project, zone, d.Id(), setNamedPorts).Do()
|
||||
zonalID.Project, zonalID.Zone, zonalID.Name, setNamedPorts).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating InstanceGroupManager: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete:
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Updating InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Updating InstanceGroupManager")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -599,14 +618,14 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
if d.HasChange("target_size") {
|
||||
targetSize := int64(d.Get("target_size").(int))
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.Resize(
|
||||
project, zone, d.Id(), targetSize).Do()
|
||||
zonalID.Project, zonalID.Zone, zonalID.Name, targetSize).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating InstanceGroupManager: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Updating InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Updating InstanceGroupManager")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -622,14 +641,14 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
}
|
||||
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.SetAutoHealingPolicies(
|
||||
project, zone, d.Id(), setAutoHealingPoliciesRequest).Do()
|
||||
zonalID.Project, zonalID.Zone, zonalID.Name, setAutoHealingPoliciesRequest).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating AutoHealingPolicies: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Updating AutoHealingPolicies")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Updating AutoHealingPolicies")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -644,21 +663,21 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
InstanceTemplate: d.Get("instance_template").(string),
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.SetInstanceTemplate(project, zone, d.Id(), setInstanceTemplate).Do()
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.SetInstanceTemplate(zonalID.Project, zonalID.Zone, zonalID.Name, setInstanceTemplate).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating InstanceGroupManager: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Updating InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Updating InstanceGroupManager")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
updateStrategy := d.Get("update_strategy").(string)
|
||||
rollingUpdatePolicy := expandUpdatePolicy(d.Get("rolling_update_policy").([]interface{}))
|
||||
err = performZoneUpdate(config, d.Id(), updateStrategy, rollingUpdatePolicy, nil, project, zone)
|
||||
err = performZoneUpdate(config, zonalID.Name, updateStrategy, rollingUpdatePolicy, nil, zonalID.Project, zonalID.Zone)
|
||||
d.SetPartial("instance_template")
|
||||
}
|
||||
|
||||
|
@ -667,7 +686,7 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
updateStrategy := d.Get("update_strategy").(string)
|
||||
rollingUpdatePolicy := expandUpdatePolicy(d.Get("rolling_update_policy").([]interface{}))
|
||||
versions := expandVersions(d.Get("version").([]interface{}))
|
||||
err = performZoneUpdate(config, d.Id(), updateStrategy, rollingUpdatePolicy, versions, project, zone)
|
||||
err = performZoneUpdate(config, zonalID.Name, updateStrategy, rollingUpdatePolicy, versions, zonalID.Project, zonalID.Zone)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -683,22 +702,31 @@ func resourceComputeInstanceGroupManagerUpdate(d *schema.ResourceData, meta inte
|
|||
func resourceComputeInstanceGroupManagerDelete(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
project, err := getProject(d, config)
|
||||
zonalID, err := parseInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
zone, err := getZone(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
if zonalID.Project == "" {
|
||||
zonalID.Project, err = getProject(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.Delete(project, zone, d.Id()).Do()
|
||||
if zonalID.Zone == "" {
|
||||
zonalID.Zone, err = getZone(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.InstanceGroupManagers.Delete(zonalID.Project, zonalID.Zone, zonalID.Name).Do()
|
||||
attempt := 0
|
||||
for err != nil && attempt < 20 {
|
||||
attempt++
|
||||
time.Sleep(2000 * time.Millisecond)
|
||||
op, err = config.clientComputeBeta.InstanceGroupManagers.Delete(project, zone, d.Id()).Do()
|
||||
op, err = config.clientComputeBeta.InstanceGroupManagers.Delete(zonalID.Project, zonalID.Zone, zonalID.Name).Do()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
|
@ -708,7 +736,7 @@ func resourceComputeInstanceGroupManagerDelete(d *schema.ResourceData, meta inte
|
|||
currentSize := int64(d.Get("target_size").(int))
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Deleting InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Deleting InstanceGroupManager")
|
||||
|
||||
for err != nil && currentSize > 0 {
|
||||
if !strings.Contains(err.Error(), "timeout") {
|
||||
|
@ -716,7 +744,7 @@ func resourceComputeInstanceGroupManagerDelete(d *schema.ResourceData, meta inte
|
|||
}
|
||||
|
||||
instanceGroup, err := config.clientComputeBeta.InstanceGroups.Get(
|
||||
project, zone, d.Id()).Do()
|
||||
zonalID.Project, zonalID.Zone, zonalID.Name).Do()
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error getting instance group size: %s", err)
|
||||
}
|
||||
|
@ -729,7 +757,7 @@ func resourceComputeInstanceGroupManagerDelete(d *schema.ResourceData, meta inte
|
|||
|
||||
log.Printf("[INFO] timeout occured, but instance group is shrinking (%d < %d)", instanceGroupSize, currentSize)
|
||||
currentSize = instanceGroupSize
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Deleting InstanceGroupManager")
|
||||
err = computeSharedOperationWait(config.clientCompute, op, zonalID.Project, "Deleting InstanceGroupManager")
|
||||
}
|
||||
|
||||
d.SetId("")
|
||||
|
@ -838,5 +866,43 @@ func flattenAutoHealingPolicies(autoHealingPolicies []*computeBeta.InstanceGroup
|
|||
|
||||
func resourceInstanceGroupManagerStateImporter(d *schema.ResourceData, meta interface{}) ([]*schema.ResourceData, error) {
|
||||
d.Set("wait_for_instances", false)
|
||||
zonalID, err := parseInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if zonalID.Zone == "" || zonalID.Project == "" {
|
||||
return nil, fmt.Errorf("Invalid instance group manager import ID. Expecting {projectId}/{zone}/{name}.")
|
||||
}
|
||||
d.Set("project", zonalID.Project)
|
||||
d.Set("zone", zonalID.Zone)
|
||||
d.Set("name", zonalID.Name)
|
||||
return []*schema.ResourceData{d}, nil
|
||||
}
|
||||
|
||||
type instanceGroupManagerId struct {
|
||||
Project string
|
||||
Zone string
|
||||
Name string
|
||||
}
|
||||
|
||||
func (i instanceGroupManagerId) terraformId() string {
|
||||
return fmt.Sprintf("%s/%s/%s", i.Project, i.Zone, i.Name)
|
||||
}
|
||||
|
||||
func parseInstanceGroupManagerId(id string) (*instanceGroupManagerId, error) {
|
||||
switch {
|
||||
case instanceGroupManagerIdRegex.MatchString(id):
|
||||
parts := strings.Split(id, "/")
|
||||
return &instanceGroupManagerId{
|
||||
Project: parts[0],
|
||||
Zone: parts[1],
|
||||
Name: parts[2],
|
||||
}, nil
|
||||
case instanceGroupManagerIdNameRegex.MatchString(id):
|
||||
return &instanceGroupManagerId{
|
||||
Name: id,
|
||||
}, nil
|
||||
default:
|
||||
return nil, fmt.Errorf("Invalid instance group manager specifier. Expecting either {projectId}/{zone}/{name} or {name}, where {projectId} and {zone} will be derived from the provider.")
|
||||
}
|
||||
}
|
||||
|
|
|
@ -2,16 +2,8 @@ package google
|
|||
|
||||
import (
|
||||
"fmt"
|
||||
"reflect"
|
||||
"strconv"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
computeBeta "google.golang.org/api/compute/v0.beta"
|
||||
"google.golang.org/api/compute/v1"
|
||||
|
||||
"sort"
|
||||
|
||||
"github.com/hashicorp/terraform/helper/acctest"
|
||||
"github.com/hashicorp/terraform/helper/resource"
|
||||
"github.com/hashicorp/terraform/terraform"
|
||||
|
@ -20,8 +12,6 @@ import (
|
|||
func TestAccInstanceGroupManager_basic(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -32,21 +22,15 @@ func TestAccInstanceGroupManager_basic(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_basic(template, target, igm1, igm2),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-no-tp", &manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-no-tp",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
|
@ -58,8 +42,6 @@ func TestAccInstanceGroupManager_basic(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_targetSizeZero(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
templateName := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igmName := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
|
@ -68,26 +50,21 @@ func TestAccInstanceGroupManager_targetSizeZero(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_targetSizeZero(templateName, igmName),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-basic", &manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
if manager.TargetSize != 0 {
|
||||
t.Errorf("Expected target_size to be 0, got %d", manager.TargetSize)
|
||||
}
|
||||
}
|
||||
|
||||
func TestAccInstanceGroupManager_update(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
template1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target2 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -99,31 +76,21 @@ func TestAccInstanceGroupManager_update(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_update(template1, target1, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckInstanceGroupManagerUpdated("google_compute_instance_group_manager.igm-update", 2, []string{target1}, template1),
|
||||
testAccCheckInstanceGroupManagerNamedPorts(
|
||||
"google_compute_instance_group_manager.igm-update",
|
||||
map[string]int64{"customhttp": 8080},
|
||||
&manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
Config: testAccInstanceGroupManager_update2(template1, target1, target2, template2, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckInstanceGroupManagerUpdated(
|
||||
"google_compute_instance_group_manager.igm-update", 3,
|
||||
[]string{target1, target2}, template2),
|
||||
testAccCheckInstanceGroupManagerNamedPorts(
|
||||
"google_compute_instance_group_manager.igm-update",
|
||||
map[string]int64{"customhttp": 8080, "customhttps": 8443},
|
||||
&manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -132,8 +99,6 @@ func TestAccInstanceGroupManager_update(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_updateLifecycle(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
tag1 := "tag1"
|
||||
tag2 := "tag2"
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -143,21 +108,21 @@ func TestAccInstanceGroupManager_updateLifecycle(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_updateLifecycle(tag1, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-update", &manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
Config: testAccInstanceGroupManager_updateLifecycle(tag2, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckInstanceGroupManagerTemplateTags(
|
||||
"google_compute_instance_group_manager.igm-update", []string{tag2}),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -166,7 +131,6 @@ func TestAccInstanceGroupManager_updateLifecycle(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_updateStrategy(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
resource.Test(t, resource.TestCase{
|
||||
|
@ -174,14 +138,13 @@ func TestAccInstanceGroupManager_updateStrategy(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_updateStrategy(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-update-strategy", &manager),
|
||||
testAccCheckInstanceGroupManagerUpdateStrategy(
|
||||
"google_compute_instance_group_manager.igm-update-strategy", "NONE"),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-update-strategy",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -190,8 +153,6 @@ func TestAccInstanceGroupManager_updateStrategy(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_rollingUpdatePolicy(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
resource.Test(t, resource.TestCase{
|
||||
|
@ -201,43 +162,11 @@ func TestAccInstanceGroupManager_rollingUpdatePolicy(t *testing.T) {
|
|||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
Config: testAccInstanceGroupManager_rollingUpdatePolicy(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerBetaExists(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", &manager),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "update_strategy", "ROLLING_UPDATE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.type", "PROACTIVE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.minimal_action", "REPLACE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_surge_percent", "50"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_unavailable_percent", "50"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.min_ready_sec", "20"),
|
||||
),
|
||||
},
|
||||
// No import step because rolling updates are broken and the field will be removed in 2.0.0.
|
||||
// TODO(danawillow): Remove this test once we've removed the field.
|
||||
resource.TestStep{
|
||||
Config: testAccInstanceGroupManager_rollingUpdatePolicy2(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerBetaExists(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", &manager),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "update_strategy", "ROLLING_UPDATE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.type", "PROACTIVE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.minimal_action", "REPLACE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_surge_fixed", "2"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_unavailable_fixed", "2"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.min_ready_sec", "20"),
|
||||
testAccCheckInstanceGroupManagerRollingUpdatePolicy(
|
||||
&manager, "google_compute_instance_group_manager.igm-rolling-update-policy"),
|
||||
),
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -246,8 +175,6 @@ func TestAccInstanceGroupManager_rollingUpdatePolicy(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_separateRegions(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
igm1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm2 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
|
@ -256,14 +183,18 @@ func TestAccInstanceGroupManager_separateRegions(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_separateRegions(igm1, igm2),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckInstanceGroupManagerExists(
|
||||
"google_compute_instance_group_manager.igm-basic-2", &manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic-2",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -272,8 +203,6 @@ func TestAccInstanceGroupManager_separateRegions(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_versions(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
primaryTemplate := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
canaryTemplate := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -283,14 +212,10 @@ func TestAccInstanceGroupManager_versions(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_versions(primaryTemplate, canaryTemplate, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerBetaExists("google_compute_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckInstanceGroupManagerVersions("google_compute_instance_group_manager.igm-basic", primaryTemplate, canaryTemplate),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
|
@ -302,8 +227,6 @@ func TestAccInstanceGroupManager_versions(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -314,15 +237,10 @@ func TestAccInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_autoHealingPolicies(template, target, igm, hck),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckInstanceGroupManagerBetaExists(
|
||||
"google_compute_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckInstanceGroupManagerAutoHealingPolicies("google_compute_instance_group_manager.igm-basic", hck, 10),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
|
@ -338,8 +256,6 @@ func TestAccInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
|||
func TestAccInstanceGroupManager_selfLinkStability(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -351,10 +267,13 @@ func TestAccInstanceGroupManager_selfLinkStability(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccInstanceGroupManager_selfLinkStability(template, target, igm, hck, autoscaler),
|
||||
Check: testAccCheckInstanceGroupManagerBetaExists(
|
||||
"google_compute_instance_group_manager.igm-basic", &manager),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -367,8 +286,18 @@ func testAccCheckInstanceGroupManagerDestroy(s *terraform.State) error {
|
|||
if rs.Type != "google_compute_instance_group_manager" {
|
||||
continue
|
||||
}
|
||||
_, err := config.clientCompute.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
id, err := parseInstanceGroupManagerId(rs.Primary.ID)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if id.Project == "" {
|
||||
id.Project = config.Project
|
||||
}
|
||||
if id.Zone == "" {
|
||||
id.Zone = rs.Primary.Attributes["zone"]
|
||||
}
|
||||
_, err = config.clientCompute.InstanceGroupManagers.Get(
|
||||
id.Project, id.Zone, id.Name).Do()
|
||||
if err == nil {
|
||||
return fmt.Errorf("InstanceGroupManager still exists")
|
||||
}
|
||||
|
@ -377,319 +306,6 @@ func testAccCheckInstanceGroupManagerDestroy(s *terraform.State) error {
|
|||
return nil
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerExists(n string, manager *compute.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
found, err := config.clientCompute.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if found.Name != rs.Primary.ID {
|
||||
return fmt.Errorf("InstanceGroupManager not found")
|
||||
}
|
||||
|
||||
*manager = *found
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerBetaExists(n string, manager *computeBeta.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
found, err := config.clientComputeBeta.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if found.Name != rs.Primary.ID {
|
||||
return fmt.Errorf("InstanceGroupManager not found")
|
||||
}
|
||||
|
||||
*manager = *found
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerUpdated(n string, size int64, targetPools []string, template string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// Cannot check the target pool as the instance creation is asynchronous. However, can
|
||||
// check the target_size.
|
||||
if manager.TargetSize != size {
|
||||
return fmt.Errorf("instance count incorrect")
|
||||
}
|
||||
|
||||
tpNames := make([]string, 0, len(manager.TargetPools))
|
||||
for _, targetPool := range manager.TargetPools {
|
||||
tpNames = append(tpNames, GetResourceNameFromSelfLink(targetPool))
|
||||
}
|
||||
|
||||
sort.Strings(tpNames)
|
||||
sort.Strings(targetPools)
|
||||
if !reflect.DeepEqual(tpNames, targetPools) {
|
||||
return fmt.Errorf("target pools incorrect. Expected %s, got %s", targetPools, tpNames)
|
||||
}
|
||||
|
||||
// check that the instance template updated
|
||||
instanceTemplate, err := config.clientCompute.InstanceTemplates.Get(
|
||||
config.Project, template).Do()
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error reading instance template: %s", err)
|
||||
}
|
||||
|
||||
if instanceTemplate.Name != template {
|
||||
return fmt.Errorf("instance template not updated")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerNamedPorts(n string, np map[string]int64, instanceGroupManager *compute.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
var found bool
|
||||
for _, namedPort := range manager.NamedPorts {
|
||||
found = false
|
||||
for name, port := range np {
|
||||
if namedPort.Name == name && namedPort.Port == port {
|
||||
found = true
|
||||
}
|
||||
}
|
||||
if !found {
|
||||
return fmt.Errorf("named port incorrect")
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerVersions(n string, primaryTemplate string, canaryTemplate string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientComputeBeta.InstanceGroupManagers.Get(config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if len(manager.Versions) != 2 {
|
||||
return fmt.Errorf("Expected # of versions to be 2, got %d", len(manager.Versions))
|
||||
}
|
||||
|
||||
primaryVersion := manager.Versions[0]
|
||||
if !strings.Contains(primaryVersion.InstanceTemplate, primaryTemplate) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", primaryTemplate, primaryVersion.InstanceTemplate)
|
||||
}
|
||||
|
||||
canaryVersion := manager.Versions[1]
|
||||
if !strings.Contains(canaryVersion.InstanceTemplate, canaryTemplate) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", canaryTemplate, canaryVersion.InstanceTemplate)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerAutoHealingPolicies(n, hck string, initialDelaySec int64) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientComputeBeta.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if len(manager.AutoHealingPolicies) != 1 {
|
||||
return fmt.Errorf("Expected # of auto healing policies to be 1, got %d", len(manager.AutoHealingPolicies))
|
||||
}
|
||||
autoHealingPolicy := manager.AutoHealingPolicies[0]
|
||||
|
||||
if !strings.Contains(autoHealingPolicy.HealthCheck, hck) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", hck, autoHealingPolicy.HealthCheck)
|
||||
}
|
||||
|
||||
if autoHealingPolicy.InitialDelaySec != initialDelaySec {
|
||||
return fmt.Errorf("Expected auto healing policy inital delay to be %d, got %d", initialDelaySec, autoHealingPolicy.InitialDelaySec)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerTemplateTags(n string, tags []string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.InstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["zone"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// check that the instance template updated
|
||||
instanceTemplate, err := config.clientCompute.InstanceTemplates.Get(
|
||||
config.Project, GetResourceNameFromSelfLink(manager.InstanceTemplate)).Do()
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error reading instance template: %s", err)
|
||||
}
|
||||
|
||||
if !reflect.DeepEqual(instanceTemplate.Properties.Tags.Items, tags) {
|
||||
return fmt.Errorf("instance template not updated")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerUpdateStrategy(n, strategy string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
if rs.Primary.Attributes["update_strategy"] != strategy {
|
||||
return fmt.Errorf("Expected strategy to be %s, got %s",
|
||||
strategy, rs.Primary.Attributes["update_strategy"])
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckInstanceGroupManagerRollingUpdatePolicy(manager *computeBeta.InstanceGroupManager, resource string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs := s.RootModule().Resources[resource]
|
||||
|
||||
updatePolicy := manager.UpdatePolicy
|
||||
|
||||
surgeFixed, _ := strconv.ParseInt(rs.Primary.Attributes["rolling_update_policy.0.max_surge_fixed"], 10, 64)
|
||||
if updatePolicy.MaxSurge.Fixed != surgeFixed {
|
||||
return fmt.Errorf("Expected update policy MaxSurge to be %d, got %d", surgeFixed, updatePolicy.MaxSurge.Fixed)
|
||||
}
|
||||
|
||||
surgePercent, _ := strconv.ParseInt(rs.Primary.Attributes["rolling_update_policy.0.max_surge_percent"], 10, 64)
|
||||
if updatePolicy.MaxSurge.Percent != surgePercent {
|
||||
return fmt.Errorf("Expected update policy MaxSurge to be %d, got %d", surgePercent, updatePolicy.MaxSurge.Percent)
|
||||
}
|
||||
|
||||
unavailableFixed, _ := strconv.ParseInt(rs.Primary.Attributes["rolling_update_policy.0.max_unavailable_fixed"], 10, 64)
|
||||
if updatePolicy.MaxUnavailable.Fixed != unavailableFixed {
|
||||
return fmt.Errorf("Expected update policy MaxUnavailable to be %d, got %d", unavailableFixed, updatePolicy.MaxUnavailable.Fixed)
|
||||
}
|
||||
|
||||
unavailablePercent, _ := strconv.ParseInt(rs.Primary.Attributes["rolling_update_policy.0.max_unavailable_percent"], 10, 64)
|
||||
if updatePolicy.MaxUnavailable.Percent != unavailablePercent {
|
||||
return fmt.Errorf("Expected update policy MaxUnavailable to be %d, got %d", unavailablePercent, updatePolicy.MaxUnavailable.Percent)
|
||||
}
|
||||
|
||||
policyType := rs.Primary.Attributes["rolling_update_policy.0.type"]
|
||||
if updatePolicy.Type != policyType {
|
||||
return fmt.Errorf("Expected update policy Type to be \"%s\", got \"%s\"", policyType, updatePolicy.Type)
|
||||
}
|
||||
|
||||
policyAction := rs.Primary.Attributes["rolling_update_policy.0.minimal_action"]
|
||||
if updatePolicy.MinimalAction != policyAction {
|
||||
return fmt.Errorf("Expected update policy MinimalAction to be \"%s\", got \"%s\"", policyAction, updatePolicy.MinimalAction)
|
||||
}
|
||||
|
||||
minReadySec, _ := strconv.ParseInt(rs.Primary.Attributes["rolling_update_policy.0.min_ready_sec"], 10, 64)
|
||||
if updatePolicy.MinReadySec != minReadySec {
|
||||
return fmt.Errorf("Expected update policy MinReadySec to be %d, got %d", minReadySec, updatePolicy.MinReadySec)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccInstanceGroupManager_basic(template, target, igm1, igm2 string) string {
|
||||
return fmt.Sprintf(`
|
||||
data "google_compute_image" "my_image" {
|
||||
|
@ -1019,7 +635,7 @@ func testAccInstanceGroupManager_updateStrategy(igm string) string {
|
|||
base_instance_name = "igm-update-strategy"
|
||||
zone = "us-central1-c"
|
||||
target_size = 2
|
||||
update_strategy = "NONE"
|
||||
update_strategy = "REPLACE"
|
||||
named_port {
|
||||
name = "customhttp"
|
||||
port = 8080
|
||||
|
|
|
@ -3,6 +3,7 @@ package google
|
|||
import (
|
||||
"fmt"
|
||||
"log"
|
||||
"regexp"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
|
@ -14,6 +15,11 @@ import (
|
|||
computeBeta "google.golang.org/api/compute/v0.beta"
|
||||
)
|
||||
|
||||
var (
|
||||
regionInstanceGroupManagerIdRegex = regexp.MustCompile("^" + ProjectRegex + "/[a-z0-9-]+/[a-z0-9-]+$")
|
||||
regionInstanceGroupManagerIdNameRegex = regexp.MustCompile("^[a-z0-9-]+$")
|
||||
)
|
||||
|
||||
func resourceComputeRegionInstanceGroupManager() *schema.Resource {
|
||||
return &schema.Resource{
|
||||
Create: resourceComputeRegionInstanceGroupManagerCreate,
|
||||
|
@ -272,6 +278,11 @@ func resourceComputeRegionInstanceGroupManagerCreate(d *schema.ResourceData, met
|
|||
return err
|
||||
}
|
||||
|
||||
region, err := getRegion(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if _, ok := d.GetOk("rolling_update_policy"); d.Get("update_strategy") == "ROLLING_UPDATE" && !ok {
|
||||
return fmt.Errorf("[rolling_update_policy] must be set when 'update_strategy' is set to 'ROLLING_UPDATE'")
|
||||
}
|
||||
|
@ -291,13 +302,13 @@ func resourceComputeRegionInstanceGroupManagerCreate(d *schema.ResourceData, met
|
|||
ForceSendFields: []string{"TargetSize"},
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.Insert(project, d.Get("region").(string), manager).Do()
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.Insert(project, region, manager).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error creating RegionInstanceGroupManager: %s", err)
|
||||
}
|
||||
|
||||
d.SetId(manager.Name)
|
||||
d.SetId(regionInstanceGroupManagerId{Project: project, Region: region, Name: manager.Name}.terraformId())
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWait(config.clientCompute, op, project, "Creating InstanceGroupManager")
|
||||
|
@ -312,19 +323,28 @@ type getInstanceManagerFunc func(*schema.ResourceData, interface{}) (*computeBet
|
|||
func getRegionalManager(d *schema.ResourceData, meta interface{}) (*computeBeta.InstanceGroupManager, error) {
|
||||
config := meta.(*Config)
|
||||
|
||||
project, err := getProject(d, config)
|
||||
regionalID, err := parseRegionInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
region, err := getRegion(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
if regionalID.Project == "" {
|
||||
regionalID.Project, err = getProject(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
}
|
||||
|
||||
manager, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(project, region, d.Id()).Do()
|
||||
if regionalID.Region == "" {
|
||||
regionalID.Region, err = getRegion(d, config)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
}
|
||||
|
||||
manager, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(regionalID.Project, regionalID.Region, regionalID.Name).Do()
|
||||
if err != nil {
|
||||
return nil, handleNotFoundError(err, d, fmt.Sprintf("Region Instance Manager %q", d.Get("name").(string)))
|
||||
return nil, handleNotFoundError(err, d, fmt.Sprintf("Region Instance Manager %q", regionalID.Name))
|
||||
}
|
||||
|
||||
return manager, nil
|
||||
|
@ -352,10 +372,16 @@ func resourceComputeRegionInstanceGroupManagerRead(d *schema.ResourceData, meta
|
|||
return err
|
||||
}
|
||||
|
||||
project, err := getProject(d, config)
|
||||
regionalID, err := parseRegionInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if regionalID.Project == "" {
|
||||
regionalID.Project, err = getProject(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
d.Set("base_instance_name", manager.BaseInstanceName)
|
||||
d.Set("instance_template", manager.InstanceTemplate)
|
||||
|
@ -365,7 +391,7 @@ func resourceComputeRegionInstanceGroupManagerRead(d *schema.ResourceData, meta
|
|||
d.Set("name", manager.Name)
|
||||
d.Set("region", GetResourceNameFromSelfLink(manager.Region))
|
||||
d.Set("description", manager.Description)
|
||||
d.Set("project", project)
|
||||
d.Set("project", regionalID.Project)
|
||||
d.Set("target_size", manager.TargetSize)
|
||||
d.Set("target_pools", manager.TargetPools)
|
||||
d.Set("named_port", flattenNamedPortsBeta(manager.NamedPorts))
|
||||
|
@ -463,7 +489,10 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
return err
|
||||
}
|
||||
|
||||
region := d.Get("region").(string)
|
||||
region, err := getRegion(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
d.Partial(true)
|
||||
|
||||
|
@ -481,7 +510,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.SetTargetPools(
|
||||
project, region, d.Id(), setTargetPools).Do()
|
||||
project, region, d.Get("name").(string), setTargetPools).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating RegionInstanceGroupManager: %s", err)
|
||||
|
@ -503,7 +532,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.SetInstanceTemplate(
|
||||
project, region, d.Id(), setInstanceTemplate).Do()
|
||||
project, region, d.Get("name").(string), setInstanceTemplate).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating RegionInstanceGroupManager: %s", err)
|
||||
|
@ -526,7 +555,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
updateStrategy := d.Get("update_strategy").(string)
|
||||
rollingUpdatePolicy := expandUpdatePolicy(d.Get("rolling_update_policy").([]interface{}))
|
||||
versions := expandVersions(d.Get("version").([]interface{}))
|
||||
err = performRegionUpdate(config, d.Id(), updateStrategy, rollingUpdatePolicy, versions, project, region)
|
||||
err = performRegionUpdate(config, d.Get("name").(string), updateStrategy, rollingUpdatePolicy, versions, project, region)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -543,7 +572,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
|
||||
// Make the request:
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroups.SetNamedPorts(
|
||||
project, region, d.Id(), setNamedPorts).Do()
|
||||
project, region, d.Get("name").(string), setNamedPorts).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating RegionInstanceGroupManager: %s", err)
|
||||
|
@ -561,7 +590,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
if d.HasChange("target_size") {
|
||||
targetSize := int64(d.Get("target_size").(int))
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.Resize(
|
||||
project, region, d.Id(), targetSize).Do()
|
||||
project, region, d.Get("name").(string), targetSize).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error resizing RegionInstanceGroupManager: %s", err)
|
||||
|
@ -583,7 +612,7 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.SetAutoHealingPolicies(
|
||||
project, region, d.Id(), setAutoHealingPoliciesRequest).Do()
|
||||
project, region, d.Get("name").(string), setAutoHealingPoliciesRequest).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error updating AutoHealingPolicies: %s", err)
|
||||
|
@ -606,21 +635,33 @@ func resourceComputeRegionInstanceGroupManagerUpdate(d *schema.ResourceData, met
|
|||
func resourceComputeRegionInstanceGroupManagerDelete(d *schema.ResourceData, meta interface{}) error {
|
||||
config := meta.(*Config)
|
||||
|
||||
project, err := getProject(d, config)
|
||||
regionalID, err := parseRegionInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
region := d.Get("region").(string)
|
||||
if regionalID.Project == "" {
|
||||
regionalID.Project, err = getProject(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.Delete(project, region, d.Id()).Do()
|
||||
if regionalID.Region == "" {
|
||||
regionalID.Region, err = getRegion(d, config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
op, err := config.clientComputeBeta.RegionInstanceGroupManagers.Delete(regionalID.Project, regionalID.Region, regionalID.Name).Do()
|
||||
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error deleting region instance group manager: %s", err)
|
||||
}
|
||||
|
||||
// Wait for the operation to complete
|
||||
err = computeSharedOperationWaitTime(config.clientCompute, op, project, int(d.Timeout(schema.TimeoutDelete).Minutes()), "Deleting RegionInstanceGroupManager")
|
||||
err = computeSharedOperationWaitTime(config.clientCompute, op, regionalID.Project, int(d.Timeout(schema.TimeoutDelete).Minutes()), "Deleting RegionInstanceGroupManager")
|
||||
|
||||
d.SetId("")
|
||||
return nil
|
||||
|
@ -664,5 +705,40 @@ func hashZoneFromSelfLinkOrResourceName(value interface{}) int {
|
|||
|
||||
func resourceRegionInstanceGroupManagerStateImporter(d *schema.ResourceData, meta interface{}) ([]*schema.ResourceData, error) {
|
||||
d.Set("wait_for_instances", false)
|
||||
regionalID, err := parseRegionInstanceGroupManagerId(d.Id())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
d.Set("project", regionalID.Project)
|
||||
d.Set("region", regionalID.Region)
|
||||
d.Set("name", regionalID.Name)
|
||||
return []*schema.ResourceData{d}, nil
|
||||
}
|
||||
|
||||
type regionInstanceGroupManagerId struct {
|
||||
Project string
|
||||
Region string
|
||||
Name string
|
||||
}
|
||||
|
||||
func (r regionInstanceGroupManagerId) terraformId() string {
|
||||
return fmt.Sprintf("%s/%s/%s", r.Project, r.Region, r.Name)
|
||||
}
|
||||
|
||||
func parseRegionInstanceGroupManagerId(id string) (*regionInstanceGroupManagerId, error) {
|
||||
switch {
|
||||
case regionInstanceGroupManagerIdRegex.MatchString(id):
|
||||
parts := strings.Split(id, "/")
|
||||
return ®ionInstanceGroupManagerId{
|
||||
Project: parts[0],
|
||||
Region: parts[1],
|
||||
Name: parts[2],
|
||||
}, nil
|
||||
case regionInstanceGroupManagerIdNameRegex.MatchString(id):
|
||||
return ®ionInstanceGroupManagerId{
|
||||
Name: id,
|
||||
}, nil
|
||||
default:
|
||||
return nil, fmt.Errorf("Invalid region instance group manager specifier. Expecting either {projectId}/{region}/{name} or {name}, where {projectId} and {region} will be derived from the provider.")
|
||||
}
|
||||
}
|
||||
|
|
|
@ -2,15 +2,9 @@ package google
|
|||
|
||||
import (
|
||||
"fmt"
|
||||
"reflect"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
computeBeta "google.golang.org/api/compute/v0.beta"
|
||||
"google.golang.org/api/compute/v1"
|
||||
|
||||
"sort"
|
||||
|
||||
"github.com/hashicorp/terraform/helper/acctest"
|
||||
"github.com/hashicorp/terraform/helper/resource"
|
||||
"github.com/hashicorp/terraform/terraform"
|
||||
|
@ -19,8 +13,6 @@ import (
|
|||
func TestAccRegionInstanceGroupManager_basic(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -31,14 +23,18 @@ func TestAccRegionInstanceGroupManager_basic(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_basic(template, target, igm1, igm2),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-no-tp", &manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-no-tp",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -47,8 +43,6 @@ func TestAccRegionInstanceGroupManager_basic(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_targetSizeZero(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
templateName := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igmName := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
|
@ -57,26 +51,21 @@ func TestAccRegionInstanceGroupManager_targetSizeZero(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_targetSizeZero(templateName, igmName),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
if manager.TargetSize != 0 {
|
||||
t.Errorf("Expected target_size to be 0, got %d", manager.TargetSize)
|
||||
}
|
||||
}
|
||||
|
||||
func TestAccRegionInstanceGroupManager_update(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
template1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target2 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -88,36 +77,21 @@ func TestAccRegionInstanceGroupManager_update(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_update(template1, target1, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerNamedPorts(
|
||||
"google_compute_region_instance_group_manager.igm-update",
|
||||
map[string]int64{"customhttp": 8080},
|
||||
&manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_update2(template1, target1, target2, template2, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerUpdated(
|
||||
"google_compute_region_instance_group_manager.igm-update", 3,
|
||||
[]string{target1, target2}, template2),
|
||||
testAccCheckRegionInstanceGroupManagerNamedPorts(
|
||||
"google_compute_region_instance_group_manager.igm-update",
|
||||
map[string]int64{"customhttp": 8080, "customhttps": 8443},
|
||||
&manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
ImportStateVerifyIgnore: []string{"update_strategy"},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -126,8 +100,6 @@ func TestAccRegionInstanceGroupManager_update(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_updateLifecycle(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
tag1 := "tag1"
|
||||
tag2 := "tag2"
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -137,21 +109,21 @@ func TestAccRegionInstanceGroupManager_updateLifecycle(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_updateLifecycle(tag1, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-update", &manager),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_updateLifecycle(tag2, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-update", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerTemplateTags(
|
||||
"google_compute_region_instance_group_manager.igm-update", []string{tag2}),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -160,7 +132,6 @@ func TestAccRegionInstanceGroupManager_updateLifecycle(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_updateStrategy(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
resource.Test(t, resource.TestCase{
|
||||
|
@ -170,12 +141,11 @@ func TestAccRegionInstanceGroupManager_updateStrategy(t *testing.T) {
|
|||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
Config: testAccRegionInstanceGroupManager_updateStrategy(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-update-strategy", &manager),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-update-strategy", "update_strategy", "NONE"),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-update-strategy",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -184,8 +154,6 @@ func TestAccRegionInstanceGroupManager_updateStrategy(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_rollingUpdatePolicy(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
resource.Test(t, resource.TestCase{
|
||||
|
@ -195,53 +163,18 @@ func TestAccRegionInstanceGroupManager_rollingUpdatePolicy(t *testing.T) {
|
|||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
Config: testAccRegionInstanceGroupManager_rollingUpdatePolicy(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerBetaExists(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", &manager),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "update_strategy", "ROLLING_UPDATE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.type", "PROACTIVE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.minimal_action", "REPLACE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_surge_fixed", "2"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_unavailable_fixed", "2"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.min_ready_sec", "20"),
|
||||
),
|
||||
},
|
||||
// No import step because rolling updates are broken and the field will be removed in 2.0.0.
|
||||
// TODO(danawillow): Remove this test once we've removed the field.
|
||||
resource.TestStep{
|
||||
Config: testAccRegionInstanceGroupManager_rollingUpdatePolicy2(igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerBetaExists(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", &manager),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "update_strategy", "ROLLING_UPDATE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.type", "PROACTIVE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.minimal_action", "REPLACE"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_surge_fixed", "2"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.max_unavailable_fixed", "0"),
|
||||
resource.TestCheckResourceAttr(
|
||||
"google_compute_region_instance_group_manager.igm-rolling-update-policy", "rolling_update_policy.0.min_ready_sec", "10"),
|
||||
testAccCheckInstanceGroupManagerRollingUpdatePolicy(
|
||||
&manager, "google_compute_region_instance_group_manager.igm-rolling-update-policy"),
|
||||
),
|
||||
},
|
||||
},
|
||||
})
|
||||
}
|
||||
|
||||
func TestAccRegionInstanceGroupManager_separateRegions(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager compute.InstanceGroupManager
|
||||
|
||||
igm1 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm2 := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
||||
|
@ -250,14 +183,18 @@ func TestAccRegionInstanceGroupManager_separateRegions(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_separateRegions(igm1, igm2),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic-2", &manager),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic-2",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -266,8 +203,6 @@ func TestAccRegionInstanceGroupManager_separateRegions(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_versions(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
primaryTemplate := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
canaryTemplate := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -277,14 +212,10 @@ func TestAccRegionInstanceGroupManager_versions(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_versions(primaryTemplate, canaryTemplate, igm),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerBetaExists("google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerVersions("google_compute_region_instance_group_manager.igm-basic", primaryTemplate, canaryTemplate),
|
||||
),
|
||||
},
|
||||
resource.TestStep{
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
|
@ -296,8 +227,6 @@ func TestAccRegionInstanceGroupManager_versions(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
target := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
|
@ -308,13 +237,13 @@ func TestAccRegionInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_autoHealingPolicies(template, target, igm, hck),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerBetaExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerAutoHealingPolicies("google_compute_region_instance_group_manager.igm-basic", hck, 10),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -323,8 +252,6 @@ func TestAccRegionInstanceGroupManager_autoHealingPolicies(t *testing.T) {
|
|||
func TestAccRegionInstanceGroupManager_distributionPolicy(t *testing.T) {
|
||||
t.Parallel()
|
||||
|
||||
var manager computeBeta.InstanceGroupManager
|
||||
|
||||
template := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
igm := fmt.Sprintf("igm-test-%s", acctest.RandString(10))
|
||||
zones := []string{"us-central1-a", "us-central1-b"}
|
||||
|
@ -334,13 +261,13 @@ func TestAccRegionInstanceGroupManager_distributionPolicy(t *testing.T) {
|
|||
Providers: testAccProviders,
|
||||
CheckDestroy: testAccCheckRegionInstanceGroupManagerDestroy,
|
||||
Steps: []resource.TestStep{
|
||||
resource.TestStep{
|
||||
{
|
||||
Config: testAccRegionInstanceGroupManager_distributionPolicy(template, igm, zones),
|
||||
Check: resource.ComposeTestCheckFunc(
|
||||
testAccCheckRegionInstanceGroupManagerBetaExists(
|
||||
"google_compute_region_instance_group_manager.igm-basic", &manager),
|
||||
testAccCheckRegionInstanceGroupManagerDistributionPolicy("google_compute_region_instance_group_manager.igm-basic", zones),
|
||||
),
|
||||
},
|
||||
{
|
||||
ResourceName: "google_compute_region_instance_group_manager.igm-basic",
|
||||
ImportState: true,
|
||||
ImportStateVerify: true,
|
||||
},
|
||||
},
|
||||
})
|
||||
|
@ -353,8 +280,18 @@ func testAccCheckRegionInstanceGroupManagerDestroy(s *terraform.State) error {
|
|||
if rs.Type != "google_compute_region_instance_group_manager" {
|
||||
continue
|
||||
}
|
||||
_, err := config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
id, err := parseRegionInstanceGroupManagerId(rs.Primary.ID)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if id.Project == "" {
|
||||
id.Project = config.Project
|
||||
}
|
||||
if id.Region == "" {
|
||||
id.Region = rs.Primary.Attributes["region"]
|
||||
}
|
||||
_, err = config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
id.Project, id.Region, id.Name).Do()
|
||||
if err == nil {
|
||||
return fmt.Errorf("RegionInstanceGroupManager still exists")
|
||||
}
|
||||
|
@ -363,301 +300,6 @@ func testAccCheckRegionInstanceGroupManagerDestroy(s *terraform.State) error {
|
|||
return nil
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerExists(n string, manager *compute.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
found, err := config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if found.Name != rs.Primary.ID {
|
||||
return fmt.Errorf("RegionInstanceGroupManager not found")
|
||||
}
|
||||
|
||||
*manager = *found
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerBetaExists(n string, manager *computeBeta.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
found, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if found.Name != rs.Primary.ID {
|
||||
return fmt.Errorf("RegionInstanceGroupManager not found")
|
||||
}
|
||||
|
||||
*manager = *found
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerUpdated(n string, size int64, targetPools []string, template string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// Cannot check the target pool as the instance creation is asynchronous. However, can
|
||||
// check the target_size.
|
||||
if manager.TargetSize != size {
|
||||
return fmt.Errorf("instance count incorrect")
|
||||
}
|
||||
|
||||
tpNames := make([]string, 0, len(manager.TargetPools))
|
||||
for _, targetPool := range manager.TargetPools {
|
||||
tpNames = append(tpNames, GetResourceNameFromSelfLink(targetPool))
|
||||
}
|
||||
|
||||
sort.Strings(tpNames)
|
||||
sort.Strings(targetPools)
|
||||
if !reflect.DeepEqual(tpNames, targetPools) {
|
||||
return fmt.Errorf("target pools incorrect. Expected %s, got %s", targetPools, tpNames)
|
||||
}
|
||||
|
||||
// check that the instance template updated
|
||||
instanceTemplate, err := config.clientCompute.InstanceTemplates.Get(
|
||||
config.Project, template).Do()
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error reading instance template: %s", err)
|
||||
}
|
||||
|
||||
if instanceTemplate.Name != template {
|
||||
return fmt.Errorf("instance template not updated")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerNamedPorts(n string, np map[string]int64, instanceGroupManager *compute.InstanceGroupManager) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
var found bool
|
||||
for _, namedPort := range manager.NamedPorts {
|
||||
found = false
|
||||
for name, port := range np {
|
||||
if namedPort.Name == name && namedPort.Port == port {
|
||||
found = true
|
||||
}
|
||||
}
|
||||
if !found {
|
||||
return fmt.Errorf("named port incorrect")
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerVersions(n string, primaryTemplate string, canaryTemplate string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if len(manager.Versions) != 2 {
|
||||
return fmt.Errorf("Expected # of versions to be 2, got %d", len(manager.Versions))
|
||||
}
|
||||
|
||||
primaryVersion := manager.Versions[0]
|
||||
if !strings.Contains(primaryVersion.InstanceTemplate, primaryTemplate) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", primaryTemplate, primaryVersion.InstanceTemplate)
|
||||
}
|
||||
|
||||
canaryVersion := manager.Versions[1]
|
||||
if !strings.Contains(canaryVersion.InstanceTemplate, canaryTemplate) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", canaryTemplate, canaryVersion.InstanceTemplate)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerAutoHealingPolicies(n, hck string, initialDelaySec int64) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if len(manager.AutoHealingPolicies) != 1 {
|
||||
return fmt.Errorf("Expected # of auto healing policies to be 1, got %d", len(manager.AutoHealingPolicies))
|
||||
}
|
||||
autoHealingPolicy := manager.AutoHealingPolicies[0]
|
||||
|
||||
if !strings.Contains(autoHealingPolicy.HealthCheck, hck) {
|
||||
return fmt.Errorf("Expected string \"%s\" to appear in \"%s\"", hck, autoHealingPolicy.HealthCheck)
|
||||
}
|
||||
|
||||
if autoHealingPolicy.InitialDelaySec != initialDelaySec {
|
||||
return fmt.Errorf("Expected auto healing policy inital delay to be %d, got %d", initialDelaySec, autoHealingPolicy.InitialDelaySec)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerDistributionPolicy(n string, zones []string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientComputeBeta.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if manager.DistributionPolicy == nil {
|
||||
return fmt.Errorf("Expected distribution policy to exist")
|
||||
}
|
||||
|
||||
zoneConfigs := manager.DistributionPolicy.Zones
|
||||
if len(zoneConfigs) != len(zones) {
|
||||
return fmt.Errorf("Expected number of zones in distribution policy to match; had %d, expected %d", len(zoneConfigs), len(zones))
|
||||
}
|
||||
|
||||
sort.Strings(zones)
|
||||
sortedExisting := make([]string, 0)
|
||||
for _, zone := range zoneConfigs {
|
||||
sortedExisting = append(sortedExisting, zone.Zone)
|
||||
}
|
||||
sort.Strings(sortedExisting)
|
||||
|
||||
for i := 0; i < len(zones); i++ {
|
||||
if !strings.HasSuffix(sortedExisting[i], zones[i]) {
|
||||
return fmt.Errorf("found mismatched zone configuration: expected entry #%d as '%s', got %s", i, zones[i], sortedExisting[i])
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccCheckRegionInstanceGroupManagerTemplateTags(n string, tags []string) resource.TestCheckFunc {
|
||||
return func(s *terraform.State) error {
|
||||
rs, ok := s.RootModule().Resources[n]
|
||||
if !ok {
|
||||
return fmt.Errorf("Not found: %s", n)
|
||||
}
|
||||
|
||||
if rs.Primary.ID == "" {
|
||||
return fmt.Errorf("No ID is set")
|
||||
}
|
||||
|
||||
config := testAccProvider.Meta().(*Config)
|
||||
|
||||
manager, err := config.clientCompute.RegionInstanceGroupManagers.Get(
|
||||
config.Project, rs.Primary.Attributes["region"], rs.Primary.ID).Do()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// check that the instance template updated
|
||||
instanceTemplate, err := config.clientCompute.InstanceTemplates.Get(
|
||||
config.Project, GetResourceNameFromSelfLink(manager.InstanceTemplate)).Do()
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error reading instance template: %s", err)
|
||||
}
|
||||
|
||||
if !reflect.DeepEqual(instanceTemplate.Properties.Tags.Items, tags) {
|
||||
return fmt.Errorf("instance template not updated")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func testAccRegionInstanceGroupManager_basic(template, target, igm1, igm2 string) string {
|
||||
return fmt.Sprintf(`
|
||||
data "google_compute_image" "my_image" {
|
||||
|
|
Loading…
Reference in New Issue
Block a user