This Terraform module allows users to dynamically create and update Cisco ACI Service Redirection Policy and Destinations by leveraging Consul catalog information.
The Cisco ACI fabric can act as a distributed stateless load-balancer sitting in front of any pool of workloads, regardless of their form-factor. For this module to work, the user should have deployed a Service-Graph template with PBR and service redirection enabled. For more information on how to deploy Cisco ACI Service Graph and PBR, please refer to the configuration guide and this white paper.
Using this Terraform module in conjunction with consul-terraform-sync enables administrators to automatically scale out or scale in backend server pools without having to manually reconfigure Cisco ACI policies.
This module supports the following:
- Create, Update and Delete Redirection Destination Policies (vnsRedirectDest)
- Create and Update Service Redirection Policies (vnsSvcRedirectPol).
If there is a missing feature or a bug - open an issue
- Currently Consul nodes MAC address must be specified as
meta
in the Consul service definition - If the redirection policy becomes empty (no remaining healthy service), it is not deleted
- All Consul services part of the same "task" must be defined in the same ACI VRF
~>Note This module only manages Redirection Policy lifecycle, it does not automate the creation of the Service-Graph template and deployment.
The consul-terraform-sync runs as a daemon that enables a publisher-subscriber paradigm between Consul and Cisco ACI to support Network Infrastructure Automation (NIA).
-
consul-terraform-sync subscribes to updates from the Consul catalog and executes one or more automation "tasks" with appropriate value of service variables based on those updates. consul-terraform-sync leverages Terraform as the underlying automation tool and utilizes the Terraform provider ecosystem to drive relevant change to the network infrastructure.
-
Each task consists of a runbook automation written as a compatible Terraform module using resources and data sources for the underlying network infrastructure provider.
Please refer to this link (to be updated) for getting started with consul-terraform-sync
Name | Version |
---|---|
terraform | >= 0.13 |
consul-terraform-sync | >= 0.1.0 |
consul | >= 1.7 |
Name | Version |
---|---|
aci | >= 0.4.1 |
This module is meant for use with consul-terraform-sync >= 0.1.0, Terraform >= 0.13 and Cisco ACI >= 4.2
In order to use this module, you will need to install consul-terraform-sync, create a "task" with this Terraform module as a source within the task, and run consul-terraform-sync.
The users can subscribe to the services in the consul catalog and define the Terraform module which will be executed when there are any updates to the subscribed services using a "task".
~> Note: It is recommended to have the consul-terraform-sync config guide for reference.
- Download the consul-terraform-sync on a node which is highly available (preferably, a node running a consul client)
- Add consul-terraform-sync to the PATH on that node
- Check the installation
$ consul-terraform-sync --version 0.1.0 Compatible with Terraform ~>0.13.0
- Create a config file
tasks.hcl
for consul-terraform-sync. Please note that this just an example.
log_level = "debug"
consul {
address = "172.23.189.57:8500"
}
driver "terraform" {
log = true
required_providers {
aci = {
source = "CiscoDevNet/aci"
version = "0.4.1"
}
}
}
provider "aci" {
alias = "aci1"
username = "admin"
url = "https://172.31.186.4"
private_key = "./terraform.key"
cert_name = "terraform"
}
buffer_period {
min = "5s"
max = "20s"
}
service {
name = "frontend"
datacenter = "dc1"
}
task {
name = "aci-svc-scale"
description = "Automatically Scale ACI Service Redirection Destinations"
source = "CiscoDevNet/terraform-aci-autoscaling-nia"
version = "0.0.1"
providers = ["aci.aci1"]
services = ["frontend"]
variable_files = [ "/Users/nvermand/Documents/Dev/terraform/consul-terraform-sync/inputs.tf"]
}
- Fill the
inputs.tf
file with the required module input and place it in the same directory astasks.hcl
. Currently the user must specify the Cisco ACI Tenant where the policy must be deployed, as well as the Service Redirection Policy name. You can use the example below.
tenant_name = "common"
service_redirection_policy_prefix = "nia"
- Start consul-terraform-sync
$ consul-terraform-sync -config-dir <path_to_configuration_directory>
consul-terraform-sync will create the appropriate policies in accordance to the Consul catalog.
consul-terraform-sync is now subscribed to the Consul catalog. Any updates to the services identified in the task will result in updating the Cisco ACI Redirection Destination Policies.
~> Note: If you are interested in how consul-terraform-sync works, please refer to this section.
Name | Description | Type | Default | Required |
---|---|---|---|---|
aci_tenant | Cisco ACI Tenant name, e.g., prod_tenant. | string |
common | yes |
service_redirection_policy_prefix | Prefix for the service redirection policy that is created when the first service instance is declared in the Consul catalog. The format is <prefix>-<service-name>-svc |
string |
no | |
services | Consul services monitored by consul-terraform-sync | map( |
n/a | yes |
Name | Description |
---|---|
workload_pool | Map that includes information about the created redirection destination policies, namely redirection destination names, service names, IP and MAC addresses |
service_redirection_artifact | Map that includes name and Dn of the created ACI Redirection Policies |
Example of module output:
module_output1 = {
"app" = [
{
"id" = "app1"
"ip" = "172.33.2.8"
"mac" = "F4:14:83:E9:BE:85"
},
]
"web" = [
{
"id" = "web0"
"ip" = "172.31.43.78"
"mac" = "CE:65:9E:1D:6A:DF"
},
{
"id" = "web1"
"ip" = "172.31.51.85"
"mac" = "F1:CC:A2:25:FE:07"
},
{
"id" = "web3"
"ip" = "192.168.128.17"
"mac" = "2E:10:45:0D:FA:EB"
},
]
}
module_output2 = {
"consul-nia-app-svc" = "uni/tn-common/svcCont/svcRedirectPol-consul-nia-app-svc"
"consul-nia-web-svc" = "uni/tn-common/svcCont/svcRedirectPol-consul-nia-web-svc"
}
There are 2 aspects of consul-terraform-sync.
-
Updates from Consul catalog: In the backend, consul-terraform-sync creates a blocking API query session with the Consul agent indentified in the config to get updates from the Consul catalog. consul-terraform-sync. consul-terraform-sync will get an update for the services in the consul catalog when any of the following service attributes are created, updated or deleted. These updates include service creation and deletion as well.
- service id
- service name
- service address
- service port
- service meta
- service tags
- service namespace
- service health status
- node id
- node address
- node datacenter
- node tagged addresses
- node meta
-
Managing the entire Terraform workflow: If a task and is defined, one or more services are associated with the task, provider is declared in the task and a Terraform module is specified using the source field of the task, the following sequence of events will occur:
- consul-terraform-sync will install the required version of Terraform.
- consul-terraform-sync will install the required version of the Terraform provider defined in the config file and declared in the "task".
- A new direstory "nia-tasks" with a sub-directory corresponding to each "task" will be created. This is the reason for having strict guidelines around naming.
- Each sub-directory corresponds to a separate Terraform workspace.
- Within each sub-directory corresponding a task, consul-terraform-sync will template a main.tf, variables.tf, terraform.tfvars and terraform.tfvars.tmpl.
-
main.tf:
- This files contains declaration for the required terraform and provider versions based on the task definition.
- In addition, this file has the module (identified by the 'source' field in the task) declaration with the input variables
- Consul K/V is used as the backend state for fo this Terraform workspace.
Example of generated main.tf:
# This file is generated by Consul Terraform Sync. # # The HCL blocks, arguments, variables, and values are derived from the # operator configuration for Sync. Any manual changes to this file # may not be preserved and could be clobbered by a subsequent update. terraform { required_version = "~>0.13.0" required_providers { aci = { source = "CiscoDevNet/aci" version = "0.4.1" } } backend "consul" { address = "172.23.189.57" gzip = true path = "consul-terraform-sync/terraform" } } provider "aci" { cert_name = var.aci.cert_name private_key = var.aci.private_key url = var.aci.url username = var.aci.username } # Automatically Scale ACI Service Redirection Destinations module "aci-svc-scale" { source = "CiscoDevNet/terraform-aci-autoscaling-nia" version = "0.0.1" services = var.services service_redirection_policy_prefix = var.service_redirection_policy_prefix tenant_name = var.tenant_name }
-
variables.tf:
- This is variables.tf file defined in the module
example generated variables.tf
# # The HCL blocks, arguments, variables, and values are derived from the # operator configuration for Sync. Any manual changes to this file # may not be preserved and could be clobbered by a subsequent update. # Service definition protocol v0 variable "services" { description = "Consul services monitored by Consul Terraform Sync" type = map( object({ id = string name = string address = string port = number meta = map(string) tags = list(string) namespace = string status = string node = string node_id = string node_address = string node_datacenter = string node_tagged_addresses = map(string) node_meta = map(string) }) ) } variable "aci" { default = null description = "Configuration object for aci" type = object({ alias = string cert_name = string private_key = string url = string username = string }) }
-
terraform.tfvars:
- This is the most important file generated by consul-terraform-sync.
- This variables file is generated with the most updated values from Consul catalog for all the services identified in the task.
- consul-terraform-sync updates this file with the latest values when the corresponding service gets updated in Consul catalog.
example terraform.tfvars
services = { "web0" : { id = "web1" name = "web" address = "172.31.51.85" port = 80 meta = { mac_address = "DE:AD:BE:EF:FA:CE" } tags = ["dc1", "nginx", "test", "web"] namespace = null status = "passing" node = "i-0f92f7eb4b6fb460a" node_id = "778506df-a1b2-65e0-fe1e-eafd2d1162a8" node_address = "172.31.51.85" node_datacenter = "us-east-1" node_tagged_addresses = { lan = "172.31.51.85" lan_ipv4 = "172.31.51.85" wan = "172.31.51.85" wan_ipv4 = "172.31.51.85" } node_meta = { consul-network-segment = "" } } }
-
Network Infrastructure Automation (NIA) compatible modules are built to utilize the above service variables
-
- consul-terraform-sync manages the entire Terraform workflow of plan, apply and destroy for all the individual workspaces corrresponding to the defined "tasks" based on the updates to the services to those tasks.
In summary, consul-terraform-sync triggers a Terraform workflow (plan, apply, destroy) based on updates it detects from Consul catalog.