diff options
author | Yngve Aasheim <yngveaasheim@users.noreply.github.com> | 2018-01-09 09:15:36 +0100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2018-01-09 09:15:36 +0100 |
commit | c9422014ac0fb0e4173fe21b4485ac18bbbc398a (patch) | |
tree | 721439f3ca38dbabb897c7fef4b372cb00ec4efa | |
parent | 08a484cbdeb14130e39eb61c315b7a5da4171e9b (diff) | |
parent | c856fe16156f0fe9d4d42e90e3af487a1a710f10 (diff) |
Merge pull request #4569 from vespa-engine/mpolden/reintroduce-redundancy-requirement
Reintroduce redundancy requirement
5 files changed, 22 insertions, 20 deletions
diff --git a/node-repository/src/main/java/com/yahoo/vespa/hosted/provision/provisioning/CapacityPolicies.java b/node-repository/src/main/java/com/yahoo/vespa/hosted/provision/provisioning/CapacityPolicies.java index e2ff17ba782..3f392674b20 100644 --- a/node-repository/src/main/java/com/yahoo/vespa/hosted/provision/provisioning/CapacityPolicies.java +++ b/node-repository/src/main/java/com/yahoo/vespa/hosted/provision/provisioning/CapacityPolicies.java @@ -3,6 +3,7 @@ package com.yahoo.vespa.hosted.provision.provisioning; import com.yahoo.config.provision.Capacity; import com.yahoo.config.provision.ClusterSpec; +import com.yahoo.config.provision.SystemName; import com.yahoo.config.provision.Zone; import com.yahoo.config.provision.Flavor; import com.yahoo.config.provision.NodeFlavors; @@ -60,9 +61,9 @@ public class CapacityPolicies { * @throws IllegalArgumentException if only one node is requested */ private int ensureRedundancy(int nodeCount) { - // TODO: Reactivate this check when we have sufficient capacity in ap-northeast - // if (nodeCount == 1) - // throw new IllegalArgumentException("Deployments to prod require at least 2 nodes per cluster for redundancy"); + if (nodeCount == 1 && zone.system() != SystemName.cd) { + throw new IllegalArgumentException("Deployments to prod require at least 2 nodes per cluster for redundancy"); + } return nodeCount; } diff --git a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/InactiveAndFailedExpirerTest.java b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/InactiveAndFailedExpirerTest.java index 048856bc698..9c5c86317aa 100644 --- a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/InactiveAndFailedExpirerTest.java +++ b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/InactiveAndFailedExpirerTest.java @@ -80,18 +80,18 @@ public class InactiveAndFailedExpirerTest { @Test public void reboot_generation_is_increased_when_node_moves_to_dirty() { ProvisioningTester tester = new ProvisioningTester(new Zone(Environment.prod, RegionName.from("us-east"))); - List<Node> nodes = tester.makeReadyNodes(1, "default"); + List<Node> nodes = tester.makeReadyNodes(2, "default"); // Allocate and deallocate a single node - ClusterSpec cluster = ClusterSpec.request(ClusterSpec.Type.content, + ClusterSpec cluster = ClusterSpec.request(ClusterSpec.Type.content, ClusterSpec.Id.from("test"), Version.fromString("6.42")); - tester.prepare(applicationId, cluster, Capacity.fromNodeCount(1), 1); + tester.prepare(applicationId, cluster, Capacity.fromNodeCount(2), 1); tester.activate(applicationId, ProvisioningTester.toHostSpecs(nodes)); - assertEquals(1, tester.getNodes(applicationId, Node.State.active).size()); + assertEquals(2, tester.getNodes(applicationId, Node.State.active).size()); tester.deactivate(applicationId); List<Node> inactiveNodes = tester.getNodes(applicationId, Node.State.inactive).asList(); - assertEquals(1, inactiveNodes.size()); + assertEquals(2, inactiveNodes.size()); // Check reboot generation before node is moved. New nodes transition from provisioned to dirty, so their // wanted reboot generation will always be 1. @@ -102,7 +102,7 @@ public class InactiveAndFailedExpirerTest { tester.advanceTime(Duration.ofMinutes(14)); new InactiveExpirer(tester.nodeRepository(), tester.clock(), Duration.ofMinutes(10), new JobControl(tester.nodeRepository().database())).run(); List<Node> dirty = tester.nodeRepository().getNodes(Node.State.dirty); - assertEquals(1, dirty.size()); + assertEquals(2, dirty.size()); // Reboot generation is increased assertEquals(wantedRebootGeneration + 1, dirty.get(0).status().reboot().wanted()); diff --git a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/RetiredExpirerTest.java b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/RetiredExpirerTest.java index 12e2eb3f323..931f2a8f275 100644 --- a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/RetiredExpirerTest.java +++ b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/maintenance/RetiredExpirerTest.java @@ -108,7 +108,7 @@ public class RetiredExpirerTest { ClusterSpec cluster = ClusterSpec.request(ClusterSpec.Type.content, ClusterSpec.Id.from("test"), Version.fromString("6.42")); activate(applicationId, cluster, 8, 8, provisioner); - activate(applicationId, cluster, 1, 1, provisioner); + activate(applicationId, cluster, 2, 2, provisioner); assertEquals(8, nodeRepository.getNodes(applicationId, Node.State.active).size()); assertEquals(0, nodeRepository.getNodes(applicationId, Node.State.inactive).size()); @@ -116,10 +116,10 @@ public class RetiredExpirerTest { clock.advance(Duration.ofHours(30)); // Retire period spent MockDeployer deployer = new MockDeployer(provisioner, - Collections.singletonMap(applicationId, new MockDeployer.ApplicationContext(applicationId, cluster, Capacity.fromNodeCount(1, Optional.of("default")), 1))); + Collections.singletonMap(applicationId, new MockDeployer.ApplicationContext(applicationId, cluster, Capacity.fromNodeCount(2, Optional.of("default")), 1))); new RetiredExpirer(nodeRepository, deployer, clock, Duration.ofHours(12), new JobControl(nodeRepository.database())).run(); - assertEquals(1, nodeRepository.getNodes(applicationId, Node.State.active).size()); - assertEquals(7, nodeRepository.getNodes(applicationId, Node.State.inactive).size()); + assertEquals(2, nodeRepository.getNodes(applicationId, Node.State.active).size()); + assertEquals(6, nodeRepository.getNodes(applicationId, Node.State.inactive).size()); assertEquals(1, deployer.redeployments); // inactivated nodes are not retired diff --git a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/DynamicDockerProvisioningTest.java b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/DynamicDockerProvisioningTest.java index 1dce5830540..14c353e68f8 100644 --- a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/DynamicDockerProvisioningTest.java +++ b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/DynamicDockerProvisioningTest.java @@ -464,7 +464,7 @@ public class DynamicDockerProvisioningTest { ApplicationId application = tester.makeApplicationId(); Flavor flavor = tester.nodeRepository().getAvailableFlavors().getFlavorOrThrow("d-3"); tester.prepare(application, ClusterSpec.request(ClusterSpec.Type.content, ClusterSpec.Id.from("myContent"), Version.fromString("6.100")), - 1, 1, flavor.canonicalName()); + 2, 1, flavor.canonicalName()); } private ApplicationId makeApplicationId(String tenant, String appName) { diff --git a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/ProvisioningTest.java b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/ProvisioningTest.java index a7ea77618bb..26e5a9b49b4 100644 --- a/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/ProvisioningTest.java +++ b/node-repository/src/test/java/com/yahoo/vespa/hosted/provision/provisioning/ProvisioningTest.java @@ -382,7 +382,6 @@ public class ProvisioningTest { tester.activate(application, state.allHosts); } - @Ignore // TODO: Re-activate when the check is reactivate in CapacityPolicies @Test(expected = IllegalArgumentException.class) public void prod_deployment_requires_redundancy() { ProvisioningTester tester = new ProvisioningTester(new Zone(Environment.prod, RegionName.from("us-east"))); @@ -541,25 +540,27 @@ public class ProvisioningTest { ApplicationId application1 = tester.makeApplicationId(); - tester.makeReadyNodes(10, "default"); + tester.makeReadyNodes(14, "default"); // deploy - SystemState state1 = prepare(application1, 2, 2, 3, 3, "default", tester); + SystemState state1 = prepare(application1, 3, 3, 4, 4, "default", tester); tester.activate(application1, state1.allHosts); // decrease cluster sizes - SystemState state2 = prepare(application1, 1, 1, 1, 1, "default", tester); + SystemState state2 = prepare(application1, 2, 2, 2, 2, "default", tester); tester.activate(application1, state2.allHosts); // content0 assertFalse(state2.hostByMembership("content0", 0, 0).membership().get().retired()); - assertTrue( state2.hostByMembership("content0", 0, 1).membership().get().retired()); + assertFalse( state2.hostByMembership("content0", 0, 1).membership().get().retired()); assertTrue( state2.hostByMembership("content0", 0, 2).membership().get().retired()); + assertTrue( state2.hostByMembership("content0", 0, 3).membership().get().retired()); // content1 assertFalse(state2.hostByMembership("content1", 0, 0).membership().get().retired()); - assertTrue( state2.hostByMembership("content1", 0, 1).membership().get().retired()); + assertFalse(state2.hostByMembership("content1", 0, 1).membership().get().retired()); assertTrue( state2.hostByMembership("content1", 0, 2).membership().get().retired()); + assertTrue( state2.hostByMembership("content1", 0, 3).membership().get().retired()); } @Test |