Managing configuration effectively is crucial for ensuring that deployments remain consistent and adaptable to various environments. Configuration data for every aspect of ARO HCP is stored in a configuration file and used for infrastructure and service deployments alike.
Nested YAML structures, override layers, and region-agnostic templating of config values allow sharing common configuration elements across environments and regions. These mechanisms provide the flexibility to adapt settings for a specific cloud, environment, or region when necessary. The configuration structure is enforced by a schema, ensuring the correctness of the configuration while allowing for elaborate override scenarios.
ARO HCP configuration data is stored in YAML format, allowing for a structured representation of settings. The configuration supports nested structures, enabling hierarchical organization of properties and logical grouping of related settings.
...
frontend:
cosmosDB:
name: arohcp-rp
private: true
zoneRedundantMode: 'Auto'
cert:
name: frontend-cert
issuer: OneCertV2-PublicCA
...
Configuration properties are embedded into a layered structure within the configuration file. This layering approach enables reusability of common configuration settings while allowing targeted overrides for specific clouds, environments, and regions.
- Default: This is the base layer of the configuration. It should contain options that apply to most clouds, environments, and regions.
- Cloud: This layer holds overrides relevant to a specific Azure cloud, such as
public
orfairfax
. - Environments: This layer provides overrides for a specific deployment environment within a cloud. Examples include
dev
,personal-dev
,integration
,stage
, andproduction
. - Region: This layer holds overrides for a specific region within a deployment environment, allowing for fine-tuned configuration adjustments.
Below is an example representation:
defaults:
<configuration goes here>
clouds:
public:
defaults:
<configuration goes here>
environments:
prod:
defaults:
<configuration goes here>
regions:
westus3:
<configuration goes here>
<other-region>:
...
<other-env>:
...
fairfax:
...
Configuration in layers can be partial, meaning not all required fields need to be defined in every level if they do not make sense in that context. As long as all layers together eventually provide the necessary fields, the configuration remains valid.
defaults:
frontend:
cosmosDB:
private: true (1)
zoneRedundantMode: 'Auto' (2)
clouds:
public:
defaults: {}
environments:
dev:
defaults:
frontend:
cosmosDB:
private: false (3)
name: arohcp-rp-dev (4)
prod:
defaults:
frontend:
cosmosDB:
name: arohcp-rp-prod (5)
In this example:
- The global defaults (1) and (2) set
private: true
andzoneRedundantMode: 'Auto'
for all Cosmos DB instances, as these are generally good settings for almost all Cosmos DB instances. - The Cosmos DB name is not set in the global defaults, as each deployment requires a unique name.
- In the
public
clouddev
environment, theprivate
(3) setting is overridden tofalse
to make it easier for developers to access their Cosmos DB instance. The DBname
(4) is introduced at this level, as it is unique to the environment. - In the
public
cloudprod
environment, only thename
(5) is overridden and the inheritedprivate
andzoneRedundantMode
settings from the global defaults remain unchanged.
Certain configuration fields require unique values within a deployment environment or even an Azure cloud. For example, Key Vault names must be unique within an Azure cloud, while management cluster names must be unique within a deployment environment.
To prevent repetitive declarations of such values, templating can be used within the configuration. Templating is supported via Go templates in property values, allowing dynamic value substitution based on contextual variables:
ctx.region
: The full Azure region name.- Length: up to 20 characters long.
- Consists of letters and digits. Starts with a letter.
ctx.regionShort
: A shorter version of the region name.- Length: 2 to 4 characters.
- Consists of letters and digits. Starts with a letter.
ctx.stamp
: The numerical value to enumerate the instances of management clusters.- Relates to the EV2 stamp.
- Usually starts with 1
Using these variables, configuration files can remain mostly region-agnostic, avoiding almost all regional overrides.
In this example the name of the Cosmos DB instance is defined by using the ctx.regionShort
template variable. This ensures that the Cosmos DB name is unique within the region.
$schema: config.schema.json
defaults:
frontend:
cosmosDB:
...
name: arohcp-rp-{{ .ctx.regionShort }}
- Avoid Hardcoding: Use templating to avoid hardcoding values that are unique to a region or environment. Also avoid hardcoding values if they are repeated in various contexts, e.g. a deployment script, Helm chart or Bicep template.
- Check name restrictions: Ensure that the names of resources are compliant with Azure naming restrictions, especially when constructing names with templating using
{{ .ctx }}
. We have plans to improve configuration validation to catch these issues automatically. - Be aware of the scope of name uniqueness: Understand the scope of uniqueness for resource names and use appropriate measures like resourcename prefixes and templating. For example
- Key Vault names must be unique within an Azure cloud - this is an Azure restriction
- management cluster names must be unique within a deployment environment - this is an architectural restriction
With multiple layers of overrides and templating in use, it can be difficult to determine the resulting configuration for a specific cloud/environment/region combination. To address this, tooling is available to materialize the configuration for a given deployment scenario.
-
For a quick way to inspect the public cloud configuration of
config/config.yaml
, use:./templatize.sh $deployenv
-
For more control specify a custom configuration file, cloud, and region as follows:
CONFIG_FILE=path_to_config.yaml ./templatize.sh $deployenv -c $cloud -r $region
-
PR checks require materialized configurations for well-known cloud/environment/region combinations to be present in pull requests. These well-known combinations are specified in the config/Makefile and represent important deployment targets for the project. Failing to run
make -C config materialize
and commit the result will cause the PR checks to fail.
Configuration settings can be used in pipeline files and bicepparam files to customize service and infra deployments.
Individual configuration properties can be referenced in pipeline files for use in shell steps:
steps:
- ...
action: Shell
variables:
- name: FRONTEND_COSMOS_DB_NAME
configRef: frontend.cosmosDB.name
For more details on shell steps, refer to the Shell Step Documentation.
To use configuration values for Bicep templates, bicepparam files are processed as Go templates, allowing configuration lookups using the following syntax:
param rpCosmosDbName = '{{ .frontend.cosmosDB.name }}' // quote strings ...
param rpCosmosDbPrivate = {{ .frontend.cosmosDB.private }} // ... but not boolean or number values
- Only basic fields (string, boolean, or number types) should be referenced from pipeline files and Bicepparam files. Complex data types do not translate well to EV2 configuration settings right now.
- Avoid using arrays in configuration. Instead, represent arrays as a list of comma separated values and parse them in Bicep templates using the
csvToArray
function frommodules/common.bicep
. Arrays do not translate well to EV2 configuration settings right now.
The structure of the configuration is strictly defined by a JSON schema to ensure correctness, enforce required fields, and enable validation. This schema is maintained in config.schema.json and dictates the format of the YAML configuration, including supported properties, nested structures, and allowed values.
By enforcing a schema, configuration files remain predictable and can be automatically validated before deployment, reducing misconfigurations and ensuring consistency across environments.
- config.yaml - Contains the configuration for the Red Hat development environments.
- dev: integrated DEV environment - the first environment where all services are deployed together.
- cspr: CS PR environment - a dedicated environment for testing Cluster Service PRs.
- personal-dev: personal DEV environment - used by developers to create new personal ARO HCP instances.
- personal-perscale: personal perfscale environment - used by the perfscale team to create new ARO HCP instances with production grade management cluster settings
- config.msft.yaml - Contains the configuration for the Microsoft deployment environments.
- int: MSIT INT environment - a dedicated environment for testing EV2 deployments and MISE.
-
Update the respective YAML file and run:
make -C config materialize
-
Inspect the effects of the changes in the materialized configuration files.
-
Commit the materialized files, open a PR, review, and merge.
Check the section about Materializing Configuration and Propagate Configuration Changes for more details.
Propagation of configuration changes varies depending on the environment:
-
- Only the dev and cspr environments are automatically reconciled with new changes for configuration, infrastructure, and service deployments.
- personal-dev are fully controlled by developers. If there are relevant changes, notify developers so they can apply updates manually.
-
- Propagation is not automated.
- Refer to the EV2 deployment documentation for details on how to trigger a deployment.