»Terraform Cloud Driver
This feature requires Consul-Terraform-Sync Enterprise which is available with Consul Enterprise.
Consul-Terraform-Sync (CTS) is more powerful when you integrate it with Terraform Cloud. Integrating with Terraform Cloud provides features, such as enhanced workspaces and insight into Terraform operations as CTS dynamically updates your network infrastructure. CTS is compatible with both the self-hosted and managed service versions of Terraform Cloud. It also supports all tiers of the Terraform Cloud managed service.
This page describes how the Terraform Cloud driver operates within CTS.
»Terraform Workspace Automation
CTS manages Terraform runs following the API-driven run workflow for workspaces in Terraform Cloud.
On startup, CTS:
- Creates or discovers Terraform Cloud workspaces corresponding to the configured tasks.
- Prepares the local environment and generates Terraform configuration files that make up the root module for each task.
- Packages the generated files and uploads them as a configuration version for the task's workspace on Terraform Cloud.
Once all workspaces are set up, CTS monitors the Consul catalog for service changes. When relevant changes are detected, the Terraform Cloud driver dynamically updates input variables for that task directly as workspace variables using the Terraform Cloud API. The driver then queues a run on the workspace, with auto-apply enabled, to update your network infrastructure.
Note: Although workspaces for tasks are executed in isolated environments, this does not guarantee the infrastructure changes from concurrent task executions are independent. Ensure that modules across all tasks are not modifying the same resource objects or have overlapping changes that may result in race conditions during automation.
CTS will discover or create a new workspaces based on your configured tasks. The task configuration
description are used to set the workspace name and description. The task configuration
terraform_cloud_workspace is used to set options like Terraform version, execution mode, and agent pool if relevant. CTS will also use any globally set workspace configurations, specified in the driver configuration
Workspace automation requirements for CTS are in place to avoid overriding other workspaces unintentionally.
- Must be set to remote or agent execution mode
- Cannot be connected to a VCS
- Cannot have an existing configuration version uploaded by another application
- Must satisfy workspace tag requirements and tag restrictions set by the CTS operator
Workspaces created by CTS will be configured with the following settings:
|Workspace name||CTS task name|
|Description||CTS task description|
|Apply method||Auto apply|
Other workspace settings can be pre-configured or updated, such as setting the workspace to manual apply or adding a run notification to send messages to a Slack channel when CTS updates your network infrastructure.
CTS can automate remote workspaces with either auto apply or manual apply configured. Having CTS manage workspaces with manual apply is useful to add an approval stage to CTS automation. Operators can manually inspect and approve or discard runs that CTS had queued based on the task run condition.
When CTS detects new changes for a workspace that already has a run pending on approval, CTS will discard the stale run and queue a new run with the latest values. The new run will go through plan and then again wait on an operator to approve it. Only once the run is approved will the infrastructure be updated with the latest Consul changes.
There are two approaches to setup manual apply for a workspace managed by CTS based on how the workspace is created.
- For CTS created workspaces, update the apply method from auto to manual via the Terraform Cloud web application or API.
- For pre-configured workspaces, create the workspace prior to CTS task automation via the Terraform Cloud web application or API.
- Create a workspace with the same name as the desired task.
- Set the workspace to API-driven run workflow and the execution mode to remote.
- Ensure that the apply method for the workspace is set to manual apply.
- Configure the task for the workspace and run CTS.
Tip: Setup run notifications for workspaces with manual apply to not miss automated runs by CTS. Look into setting the buffer period or a schedule condition to group changes together and reduce runs requiring approval.
An example configuration version for a task named "cts-example" would have the folder structure below when running with the Terraform Cloud driver and using the default working directory.
$ tree sync-tasks/ sync-tasks/ └── cts-example/ ├── main.tf └── variables.tf
main.tf- The main file contains the terraform block, provider blocks, and a module block calling the module configured for the task.
terraformblock - The corresponding provider source and versions for the task from the configuration files are placed into this block for the root module.
providerblocks - The provider blocks generated in the root module resemble the
terraform_providerblocks from the configuration for CTS. They have identical arguments present and are set from the intermediate variable created per provider.
moduleblock - The module block is where the task's module is called as a child module. The child module contains the core logic for automation. Required and optional input variables are passed as arguments to the module.
variables.tf- This file contains three types of variable declarations:
servicesinput variable (required) determines module compatibility with Consul-Terraform Sync (read more on compatible Terraform modules for more details).
- Any additional optional input variables provided by CTS that the module may use.
- Various intermediate variables used to configure providers. Intermediate provider variables are interpolated from the provider blocks and arguments configured in the CTS configuration.
variables.module.tf- This file is created if there are variables configured for the task and contains the interpolated variable declarations that match the variables from configuration. These are then used to proxy the configured variables to the module through explicit assignment in the module block.
CTS uses Terraform input variables to reflect the latest Consul service information. They are used as parameters for your Terraform module. Input variables are dynamic and are updated by the driver throughout the runtime of CTS.
You can view the latest service information in the Terraform UI by navigating to that workspace and clicking the "Variables" tab in the workspace navigation.
Caution: Dynamic variables maintained by CTS are formatted for automation. Unexpected manual changes to these variables may result in automation errors.
»Setting Up Terraform Cloud Driver
Because a CTS instance can only be configured with one driver, an instance can only be associated with either a Terraform driver or a Terraform Cloud driver. If there is a need to run both types of drivers, users will need to deploy a separate CTS instance for each type of driver. Relatedly, if there is a need to run CTS across multiple Terraform Cloud organizations, users will need to deploy a separate instance for each organization.
This section captures requirements for setting up CTS to integrate with your Terraform Cloud solution.
- Hostname of your Terraform Cloud, self-hosted distribution
- Name of your organization
- Team API token used for authentication with Terraform Cloud
Prior to running CTS with a Terraform Cloud driver, you will need an account and organization set up, as well as a dedicated token. We recommend using a team token that is restricted to Manage Workspaces-level permissions. Below are the steps for the recommended setup.
The first step is to create an account with your Terraform Cloud service. After creating an account, create a new organization or select an existing organization. The address of your Terraform Cloud service will be used to configure the
hostname, and the organization name will be used to configure the
organization on the Terraform Cloud driver.
Once you have an account and organization, the next step is to create a team. We recommend using a dedicated team and team token to run and authenticate CTS. Using a team token has the benefits of restricting organization permissions as well as associating CTS automated actions with the team rather than an individual.
After creating a dedicated team, update the team's permissions with "Manage Workspaces" organization access-level. CTS's main work revolves around creating and managing workspaces. Therefore restricting the dedicated team's permission to Manage Workspaces level is sufficient and reduces security risk.
After setting the team's permissions, the final setup step is to generate the associated team token, which can be done on the same team management page. This token will be used by CTS for API authentication and will be used to configure the
token on the Terraform Cloud driver.
We recommend configuring workspaces managed by CTS with run notifications through the Terraform web application. Run notifications notify external systems about the progress of runs and could help notify users of CTS events, particularly errored runs.
In order to configure a run notification, users can manually create a notification configuration for workspaces automated by CTS. A workspace may already exist for a task if the workspace name is identical to the configured task's
name. This may occur if CTS has already already run and created the workspace for the task. This may also occur if the workspace is manually created for the task prior to CTS running.