diff options
author | Dave Urschatz <dave.urschatz@cengn.ca> | 2016-03-08 14:42:23 -0500 |
---|---|---|
committer | Dave Urschatz <dave.urschatz@cengn.ca> | 2016-03-08 14:42:23 -0500 |
commit | 099ec3f8168c4b4343cce7e2f83792c31b59edda (patch) | |
tree | f05c16f813102cdf86cf9ebf0fa0d8aeb6be3afe /ci/maas | |
parent | fd4353888d71acf7174bfd85d4d6ef2e33e9dd0e (diff) |
Tune CENGN Lynx Lab Pod1 DHCP Address Ranges for HA
The intent of this commit is to DHCP address ranges after
Orange pod2 to support the higher volumes of deployed services
in HA scenarios where services are deployed in LXC containers
requiring IP address.
JIRA: JOID-52
Change-Id: Ib24c9e2de40cacf9ab76fbafa46254c52aa08cc4
Signed-off-by: Dave Urschatz <dave.urschatz@cengn.ca>
Diffstat (limited to 'ci/maas')
-rwxr-xr-x | ci/maas/cengn_lynx/pod1/deployment.yaml | 24 |
1 files changed, 12 insertions, 12 deletions
diff --git a/ci/maas/cengn_lynx/pod1/deployment.yaml b/ci/maas/cengn_lynx/pod1/deployment.yaml index 6a46f7ae..de3ee85e 100755 --- a/ci/maas/cengn_lynx/pod1/deployment.yaml +++ b/ci/maas/cengn_lynx/pod1/deployment.yaml @@ -97,33 +97,33 @@ cengn-lynx-pod1: broadcast_ip: 10.120.0.255 router_ip: 10.120.0.254 static_range: - low: 10.120.0.20 - high: 10.120.0.100 + low: 10.120.0.50 + high: 10.120.0.80 dynamic_range: - low: 10.120.0.101 - high: 10.120.0.200 + low: 10.120.0.81 + high: 10.120.0.249 - device: eth1 ip: 172.16.121.3 subnet_mask: 255.255.255.0 broadcast_ip: 172.16.121.255 management: 1 static_range: - low: 172.16.121.20 - high: 172.16.121.100 + low: 172.16.121.50 + high: 172.16.121.80 dynamic_range: - low: 172.16.121.101 - high: 172.16.121.200 + low: 172.16.121.81 + high: 172.16.121.249 - device: eth2 ip: 172.16.122.3 subnet_mask: 255.255.255.0 broadcast_ip: 172.16.122.255 management: 1 static_range: - low: 172.16.122.20 - high: 172.16.122.100 + low: 172.16.122.50 + high: 172.16.122.80 dynamic_range: - low: 172.16.122.101 - high: 172.16.122.200 + low: 172.16.122.81 + high: 172.16.122.249 # Defines the physical nodes which are added to the MAAS cluster # controller upon startup of the node. |