Using Enhanced Azure Scripts: An Introductory Guide

Okay, here is the article on “Using Enhanced Azure Scripts: An Introductory Guide,” aiming for approximately 5000 words.


Using Enhanced Azure Scripts: An Introductory Guide

Microsoft Azure provides a powerful and flexible cloud platform, offering a vast array of services for computing, storage, networking, databases, analytics, AI, and more. Managing these resources effectively, especially at scale, requires automation. While the Azure portal offers a user-friendly graphical interface, scripting provides the power, speed, consistency, and repeatability necessary for efficient cloud operations.

Basic Azure scripts, using either Azure PowerShell or the Azure Command-Line Interface (CLI), can perform simple tasks like creating a resource group or starting a virtual machine. However, as environments grow in complexity and operational maturity increases, the need for more sophisticated scripting techniques becomes paramount. This is where Enhanced Azure Scripts come into play.

“Enhanced Azure Scripts” isn’t an official Microsoft product name but rather a concept representing scripts that go beyond simple command sequences. They embody best practices in software development, security, and operations applied to cloud automation. These scripts are robust, reusable, maintainable, secure, and efficient. They often integrate seamlessly with broader automation frameworks, Infrastructure as Code (IaC) pipelines, and monitoring systems.

This guide provides an introduction to the principles, techniques, and tools involved in creating and utilizing enhanced Azure scripts. Whether you’re an Azure administrator, a DevOps engineer, or a developer looking to automate cloud tasks, this guide will help you elevate your scripting capabilities.

Table of Contents

  1. Introduction: Why Enhance Your Azure Scripts?
  2. Prerequisites and Setup
    • Azure Account and Permissions
    • Choosing Your Tool: Azure PowerShell vs. Azure CLI
    • Installation and Configuration
    • Development Environment
  3. Fundamentals Revisited: The Building Blocks
    • Basic Script Structure (PowerShell & CLI)
    • Connecting to Azure
    • Working with Subscriptions and Context
    • Executing Basic Commands
  4. Core Principles of Enhanced Scripting
    • Modularity and Reusability: Functions and Modules
    • Parameterization and Configuration: Making Scripts Flexible
    • Error Handling and Logging: Building Robust Scripts
    • Idempotency: Ensuring Consistent Outcomes
    • Security Best Practices: Protecting Credentials and Resources
    • Performance Optimization: Writing Efficient Code
  5. Advanced Scripting Techniques
    • Working with Azure Resource Manager (ARM) and Bicep Templates
    • Complex Resource Management (Querying, Tagging, Dependencies)
    • Integrating with Other Azure Services (Key Vault, Storage, Monitor, Entra ID)
    • Asynchronous Operations and Parallel Processing
    • Leveraging Azure REST APIs Directly
  6. Infrastructure as Code (IaC) Integration
    • Scripts as Part of the IaC Lifecycle
    • Pre- and Post-Deployment Scripting
    • Combining Scripts with ARM/Bicep/Terraform
  7. Development Workflow and Best Practices
    • Version Control with Git
    • Testing Your Scripts (Pester for PowerShell)
    • Continuous Integration and Continuous Deployment (CI/CD)
    • Documentation and Commenting Standards
  8. Security Deep Dive
    • Managed Identities for Azure Resources
    • Service Principals: Creation and Usage
    • Securely Handling Secrets with Azure Key Vault
    • Role-Based Access Control (RBAC) for Scripts
  9. Real-World Examples and Use Cases
    • Automated VM Deployment and Configuration
    • Scheduled Resource Cleanup
    • User Access Management (Azure AD / Entra ID)
    • Automated Monitoring and Alert Responses
  10. Beyond Standalone Scripts: Azure Automation, Functions, and Logic Apps
  11. Conclusion: Embracing Enhanced Automation

1. Introduction: Why Enhance Your Azure Scripts?

Simple scripts often suffice for one-off tasks or small environments. However, relying solely on basic scripts presents several challenges as complexity grows:

  • Lack of Reusability: Copy-pasting code across scripts leads to duplication and maintenance nightmares.
  • Brittleness: Scripts fail unexpectedly without proper error handling, leaving the environment in an inconsistent state.
  • Inflexibility: Hardcoded values (resource names, locations, SKUs) make scripts difficult to adapt to different environments or requirements.
  • Security Risks: Storing credentials directly in scripts is a major security vulnerability.
  • Poor Maintainability: Unstructured, uncommented code is hard to understand, debug, and modify.
  • Scalability Issues: Simple sequential execution can be slow for tasks involving numerous resources.
  • Lack of Auditability: Poor logging makes it difficult to track what a script did and why it failed.

Enhanced Azure scripts address these challenges by incorporating principles that lead to:

  • Reliability: Robust error handling and idempotency ensure scripts run predictably.
  • Maintainability: Modular design, clear structure, and good commenting make scripts easier to manage over time.
  • Reusability: Functions and modules allow code to be shared and reused across different automation tasks.
  • Security: Proper credential management (Managed Identities, Key Vault) protects sensitive information.
  • Efficiency: Optimized code and parallel processing speed up execution.
  • Scalability: Scripts are designed to handle larger and more complex Azure environments.
  • Integration: Scripts work seamlessly within larger automation frameworks like CI/CD pipelines.

Investing time in enhancing your Azure scripts pays dividends in reduced operational overhead, increased reliability, improved security posture, and faster delivery of value.

2. Prerequisites and Setup

Before diving into enhanced scripting, ensure you have the necessary foundation.

Azure Account and Permissions

You need an active Azure subscription. The permissions required depend on the tasks your scripts will perform. For resource creation and management, roles like Contributor or Owner might be necessary at the relevant scope (Subscription, Resource Group). However, always adhere to the principle of least privilege – grant only the permissions needed for the script to function. For read-only tasks, the Reader role is often sufficient.

Choosing Your Tool: Azure PowerShell vs. Azure CLI

Azure offers two primary command-line tools for interacting with the platform:

  • Azure PowerShell: A set of PowerShell modules (Az module) providing cmdlets to manage Azure resources.
    • Pros: Native to Windows environments, powerful object-oriented pipeline processing, strong integration with the PowerShell ecosystem, mature. Preferred by those already comfortable with PowerShell.
    • Cons: Can have a steeper learning curve for non-Windows users, larger installation footprint.
  • Azure CLI (Command-Line Interface): A cross-platform command-line tool (az command) using Python.
    • Pros: Cross-platform (Windows, macOS, Linux), concise syntax often mirroring Azure REST APIs, output often defaults to JSON (easy for parsing), generally faster for simple commands. Preferred by those comfortable with Bash/Linux environments or wanting cross-platform consistency.
    • Cons: Pipeline processing is text-based (requires tools like jq for complex JSON manipulation), less integrated with native PowerShell object handling.

Which to choose?
* If you primarily work on Windows and are proficient in PowerShell, Azure PowerShell is a natural fit.
* If you work across different operating systems or prefer a Bash-like experience, Azure CLI is an excellent choice.
* It’s not strictly an either/or decision. Many professionals use both, selecting the tool best suited for a specific task or script. You can even call Azure CLI commands from within a PowerShell script or vice-versa.

This guide will provide examples using both tools where practical.

Installation and Configuration

  • Azure PowerShell: Follow the official Microsoft documentation to install the Az PowerShell module. The recommended method is often via the PowerShell Gallery:
    powershell
    # Ensure PowerShellGet is up-to-date
    Install-Module -Name PowerShellGet -Force -AllowClobber
    # Install the Az module for the current user
    Install-Module -Name Az -Scope CurrentUser -Repository PSGallery -Force
  • Azure CLI: Follow the official installation instructions for your operating system (Windows, macOS, Linux). Packages are available via MSI, Homebrew, apt, yum, etc.

After installation, you need to authenticate.

Connecting to Azure

  • Azure PowerShell:
    “`powershell
    # Interactive login (opens browser)
    Connect-AzAccount

    Login with a specific tenant

    Connect-AzAccount -TenantId “

    Login using a Service Principal (more secure for automation – see Security section)

    $credential = Get-Credential # Prompts for SP AppID (username) and Secret (password)
    Connect-AzAccount -ServicePrincipal -Credential $credential -TenantId “
    * **Azure CLI:**bash

    Interactive login (opens browser)

    az login

    Login with a specific tenant

    az login –tenant “

    Login using a Service Principal (more secure for automation – see Security section)

    az login –service-principal -u -p –tenant
    “`
    Note: Storing secrets directly in scripts or command history is insecure. Later sections will cover secure alternatives like Managed Identities and Azure Key Vault.

Development Environment

A good development environment enhances productivity:

  • Editor/IDE: Visual Studio Code (VS Code) is highly recommended. It’s free, cross-platform, and has excellent extensions for both PowerShell and Azure CLI, providing syntax highlighting, IntelliSense (code completion), debugging, and Git integration.
  • Source Control: Git is the de facto standard. Use platforms like GitHub, Azure Repos, or GitLab.
  • Terminal: Use a modern terminal like Windows Terminal, PowerShell 7+, or standard terminals on macOS/Linux.

3. Fundamentals Revisited: The Building Blocks

Before enhancing scripts, let’s briefly review the basics.

Basic Script Structure

  • PowerShell (.ps1 files):
    “`powershell
    # Script block comment
    <#
    .SYNOPSIS
    Brief description of the script.
    .DESCRIPTION
    More detailed description.
    .PARAMETER ResourceGroupName
    Name of the resource group.
    .EXAMPLE
    .\MyScript.ps1 -ResourceGroupName “MyRG”
    #>
    param(
    [Parameter(Mandatory=$true)]
    [string]$ResourceGroupName,

    [string]$Location = "WestEurope"
    

    )

    Import required modules (best practice)

    Import-Module Az.Accounts
    Import-Module Az.Resources

    Connect to Azure (handle authentication securely in real scripts)

    Connect-AzAccount …

    Write-Host “Checking if Resource Group ‘$ResourceGroupName’ exists in location ‘$Location’…”

    Example command

    $rg = Get-AzResourceGroup -Name $ResourceGroupName -ErrorAction SilentlyContinue

    if ($rg) {
    Write-Host “Resource Group ‘$ResourceGroupName’ already exists.”
    } else {
    Write-Host “Creating Resource Group ‘$ResourceGroupName’ in ‘$Location’…”
    New-AzResourceGroup -Name $ResourceGroupName -Location $Location
    Write-Host “Resource Group created successfully.”
    }

    Disconnect (optional, depends on session management)

    Disconnect-AzAccount

    * **Azure CLI (Bash/Shell scripts - .sh files, or Batch - .bat/.cmd):**bash

    !/bin/bash

    Bash script example

    — Configuration —

    RESOURCE_GROUP_NAME=””
    LOCATION=”westeurope”

    — Argument Parsing (Basic Example) —

    while [[ “$#” -gt 0 ]]; do
    case $1 in
    -g|–resource-group) RESOURCE_GROUP_NAME=”$2″; shift ;;
    -l|–location) LOCATION=”$2″; shift ;;
    *) echo “Unknown parameter passed: $1”; exit 1 ;;
    esac
    shift
    done

    if [[ -z “$RESOURCE_GROUP_NAME” ]]; then
    echo “Error: Resource group name is required. Use -g or –resource-group.”
    exit 1
    fi

    — Login (handle authentication securely in real scripts) —

    az login …

    echo “Checking if Resource Group ‘$RESOURCE_GROUP_NAME’ exists…”

    Example command – use query to check existence, redirect stderr

    az group show –name “$RESOURCE_GROUP_NAME” > /dev/null 2>&1

    if [[ $? -eq 0 ]]; then
    echo “Resource Group ‘$RESOURCE_GROUP_NAME’ already exists.”
    else
    echo “Creating Resource Group ‘$RESOURCE_GROUP_NAME’ in ‘$LOCATION’…”
    az group create –name “$RESOURCE_GROUP_NAME” –location “$LOCATION” –output table
    if [[ $? -ne 0 ]]; then
    echo “Error creating resource group.”
    exit 1
    fi
    echo “Resource Group created successfully.”
    fi

    Logout (optional)

    az logout

    exit 0
    “`

Working with Subscriptions and Context

You might have access to multiple Azure subscriptions. Scripts need to target the correct one.

  • Azure PowerShell:
    “`powershell
    # List available subscriptions
    Get-AzSubscription

    Get the current context (subscription)

    Get-AzContext

    Select a specific subscription

    $subscriptionId = “
    Set-AzContext -SubscriptionId $subscriptionId

    Or using the subscription name

    $subscriptionName = “MySubscriptionName”
    Get-AzSubscription -SubscriptionName $subscriptionName | Set-AzContext
    * **Azure CLI:**bash

    List available subscriptions

    az account list –output table

    Show the current active subscription

    az account show –output table

    Set a specific subscription

    az account set –subscription “
    “`

Ensure your scripts explicitly set the desired subscription context early on to avoid accidental operations in the wrong environment.

Executing Basic Commands

Both tools follow a pattern: Verb-Noun for PowerShell (e.g., New-AzResourceGroup, Get-AzVM) and az group command --parameters for CLI (e.g., az group create, az vm show). Use the built-in help systems (Get-Help <CmdletName> -Full in PowerShell, az <command> --help in CLI) extensively.

4. Core Principles of Enhanced Scripting

These principles transform simple scripts into robust automation assets.

Modularity and Reusability: Functions and Modules

Avoid monolithic scripts. Break down logic into smaller, reusable units.

  • Functions (within a script): Encapsulate specific tasks.
    • PowerShell:
      “`powershell
      function New-AzStorageAccountIfNotExists {
      param(
      [Parameter(Mandatory=$true)]
      [string]$ResourceGroupName,
      [Parameter(Mandatory=$true)]
      [string]$AccountName,
      [string]$Location = ‘WestEurope’,
      [string]$SkuName = ‘Standard_LRS’
      )

      $storageAccount = Get-AzStorageAccount -ResourceGroupName $ResourceGroupName -Name $AccountName -ErrorAction SilentlyContinue
      if (-not $storageAccount) {
          Write-Host "Creating Storage Account '$AccountName'..."
          New-AzStorageAccount -ResourceGroupName $ResourceGroupName -Name $AccountName -Location $Location -SkuName $SkuName
      } else {
          Write-Host "Storage Account '$AccountName' already exists."
      }
      

      }

      Call the function

      New-AzStorageAccountIfNotExists -ResourceGroupName “MyRG” -AccountName “mystorageacc123”
      * **Bash/CLI:** Functions are defined using `function name { commands }` or `name() { commands }`. Parameter handling is positional (`$1`, `$2`, etc.) or requires parsing.bash
      function create_storage_account_if_not_exists() {
      local resource_group_name=”$1″
      local account_name=”$2″
      local location=”${3:-westeurope}” # Default value if $3 is unset
      local sku_name=”${4:-Standard_LRS}”

      az storage account show --name "$account_name" --resource-group "$resource_group_name" > /dev/null 2>&1
      if [[ $? -ne 0 ]]; then
          echo "Creating Storage Account '$account_name'..."
          az storage account create --name "$account_name" --resource-group "$resource_group_name" --location "$location" --sku "$sku_name" --output none
          if [[ $? -ne 0 ]]; then echo "Error creating storage account."; return 1; fi
      else
          echo "Storage Account '$account_name' already exists."
      fi
      return 0
      

      }

      Call the function

      create_storage_account_if_not_exists “MyRG” “mystorageacc123”
      if [[ $? -ne 0 ]]; then exit 1; fi
      * **Modules (PowerShell) / Script Libraries (Bash):** Group related functions into separate files (`.psm1` for PowerShell modules, `.sh` for Bash libraries) that can be imported or sourced by other scripts. This promotes code sharing across projects.
      * **PowerShell Module (`MyAzureHelpers.psm1`):**
      powershell

      MyAzureHelpers.psm1

      function Get-AzResourceIfExists { # … implementation … }
      function Set-AzResourceTag { # … implementation … }
      Export-ModuleMember -Function Get-AzResourceIfExists, Set-AzResourceTag
      *Usage in another script:*powershell
      Import-Module .\MyAzureHelpers.psm1 -Force
      Get-AzResourceIfExists -Name “MyVM” # …
      * **Bash Library (`azure_helpers.sh`):**bash

      azure_helpers.sh

      check_resource_exists() { # … implementation using az show … }
      tag_resource() { # … implementation using az tag update … }
      *Usage in another script:*bash

      !/bin/bash

      source ./azure_helpers.sh # Or provide full path
      check_resource_exists “MyRG” “Microsoft.Compute/virtualMachines” “MyVM”
      “`

Parameterization and Configuration: Making Scripts Flexible

Avoid hardcoding values. Use parameters to make scripts adaptable to different environments (Dev, Test, Prod), regions, or resource configurations.

  • Script Parameters: As shown in the basic script structures, use param() in PowerShell and argument parsing (getopt, while loop) in Bash. Make parameters mandatory where appropriate. Provide sensible defaults. Use validation attributes (PowerShell) or checks (Bash).
    • PowerShell Validation:
      “`powershell
      param(
      [Parameter(Mandatory=$true)]
      [ValidateNotNullOrEmpty()]
      [string]$ResourceGroupName,

      [Parameter(Mandatory=$true)]
      [ValidateSet("WestEurope", "EastUS", "WestUS2")]
      [string]$Location,
      
      [ValidateRange(1, 5)]
      [int]$VmCount = 1
      

      )
      * **Configuration Files:** For numerous or complex configurations, external files (JSON, YAML, INI, .psd1) are better than command-line arguments.
      * **JSON Example (`config.json`):**
      json
      {
      “resourceGroupName”: “ProdRG”,
      “location”: “WestEurope”,
      “vmSize”: “Standard_D2s_v3”,
      “tags”: {
      “Environment”: “Production”,
      “Project”: “Phoenix”
      }
      }
      * **Reading Config in PowerShell:**powershell
      $configPath = “.\config.json”
      if (-not (Test-Path $configPath)) { throw “Config file not found: $configPath” }
      $config = Get-Content $configPath | ConvertFrom-Json

      $rgName = $config.resourceGroupName
      $location = $config.location

      … use $config.vmSize, $config.tags etc.

      * **Reading Config in Bash (using `jq`):**bash
      CONFIG_FILE=”config.json”
      if [[ ! -f “$CONFIG_FILE” ]]; then echo “Config file not found: $CONFIG_FILE”; exit 1; fi

      Requires jq to be installed (sudo apt install jq / brew install jq)

      RESOURCE_GROUP_NAME=$(jq -r ‘.resourceGroupName’ “$CONFIG_FILE”)
      LOCATION=$(jq -r ‘.location’ “$CONFIG_FILE”)
      VM_SIZE=$(jq -r ‘.vmSize’ “$CONFIG_FILE”)

      … use values …

      ``
      * **Environment Variables:** Useful for CI/CD pipelines or containerized environments. Read them within the script (e.g.,
      $env:VARIABLE_NAMEin PowerShell,$VARIABLE_NAME` in Bash).

Error Handling and Logging: Building Robust Scripts

Scripts must gracefully handle expected and unexpected errors.

  • Error Detection:
    • PowerShell: Use try/catch/finally blocks. Control non-terminating errors with -ErrorAction Stop to make them terminating and catchable. Inspect the $Error automatic variable or the exception object ($_) inside catch.
      powershell
      try {
      # Potentially failing command
      New-AzResourceGroup -Name "ExistingRG" -Location "WestEurope" -ErrorAction Stop
      } catch {
      Write-Error "Failed to create resource group: $($_.Exception.Message)"
      # Add logging, cleanup, or exit logic here
      # exit 1 # Optional: terminate script on failure
      } finally {
      # Code here runs whether try succeeded or failed (e.g., cleanup)
      Write-Host "Resource group creation attempt finished."
      }
    • Azure CLI: Check the exit code ($? in Bash, $LASTEXITCODE in PowerShell) after each az command. An exit code of 0 usually indicates success; non-zero indicates failure. Use set -e in Bash to make the script exit immediately if any command fails. Use command grouping { ... } or subshells ( ... ) with error checks. Redirect standard error (2>) for specific handling.
      “`bash
      az vm create –name “MyVM” # … other params …
      if [[ $? -ne 0 ]]; then
      echo “Error: Failed to create VM ‘MyVM’.”
      # Add logging or cleanup
      exit 1
      fi

      Alternative with set -e (script exits automatically on error)

      set -e
      echo “Creating VM…”
      az vm create –name “MyVM” # …
      echo “VM Created.”
      set +e # Disable exit on error if needed later
      * **Logging:** Provide informative output about the script's progress, actions taken, and any errors encountered.
      * **PowerShell:** Use `Write-Host` (console output), `Write-Verbose` (detailed optional output, enabled with `-Verbose` switch), `Write-Warning`, `Write-Error`. Consider using dedicated logging modules like `PSFramework` or simply writing to log files using `Out-File` or `Add-Content`. Include timestamps.
      powershell
      $logFile = “.\script-log-$(Get-Date -Format ‘yyyyMMdd_HHmmss’).log”
      function Write-Log {
      param([string]$Message)
      $timestamp = Get-Date -Format ‘yyyy-MM-dd HH:mm:ss’
      $logEntry = “[$timestamp] $Message”
      Write-Host $logEntry # Also output to console
      Add-Content -Path $logFile -Value $logEntry
      }

      Write-Log “Script started.”
      try {
      # … operations …
      Write-Log “Resource ‘X’ created successfully.”
      } catch {
      Write-Log “[ERROR] Failed: $($.Exception.Message)”
      Write-Error $
      .Exception.Message # Also write to error stream
      }
      Write-Log “Script finished.”
      * **Bash:** Use `echo` to print messages. Redirect output (`>` or `>>`) to log files. Use `logger` to send messages to syslog. Prepend timestamps using `date`.bash
      LOG_FILE=”script-log-$(date +%Y%m%d_%H%M%S).log”
      log() {
      local message=”$1″
      local timestamp=$(date ‘+%Y-%m-%d %H:%M:%S’)
      echo “[$timestamp] $message” | tee -a “$LOG_FILE” # Output to console and append to file
      }

      log “Script started.”
      az group create –name “MyRG” # … > /dev/null # Suppress normal output if needed
      if [[ $? -ne 0 ]]; then
      log “[ERROR] Failed to create resource group.”
      exit 1
      else
      log “Resource group created.”
      fi
      log “Script finished.”
      “`
      * Azure Monitor Integration: For advanced logging, consider sending script logs and metrics directly to Azure Monitor Logs (Log Analytics) using REST APIs or specific cmdlets/commands.

Idempotency: Ensuring Consistent Outcomes

An idempotent script can be run multiple times with the same input parameters and produce the same end result without unintended side effects. This is crucial for automation reliability.

  • Check Before Creating: Before creating a resource, check if it already exists in the desired state.
    • PowerShell: Use Get-AzResource, Get-AzVM, Get-AzStorageAccount, etc., often with -ErrorAction SilentlyContinue.
    • Azure CLI: Use az resource show, az vm show, az storage account show, etc. Check the exit code or parse the output (e.g., using --query).
  • Update Instead of Failing: If a resource exists but differs from the desired state (e.g., different tags, VM size), update it instead of throwing an error or creating a duplicate (if applicable). Use Set-AzResource, Update-AzVM, az resource update, az vm update, etc.
  • Design for Repeatability: Structure logic to handle both initial creation and subsequent updates or no-ops gracefully.

Example (Conceptual Idempotent Resource Group Creation):

“`powershell
param([string]$ResourceGroupName, [string]$Location, [hashtable]$Tags)

$rg = Get-AzResourceGroup -Name $ResourceGroupName -ErrorAction SilentlyContinue

if ($rg) {
Write-Host “Resource Group ‘$ResourceGroupName’ exists.”
# Optional: Check if location matches (cannot change location)
if ($rg.Location -ne $Location) {
Write-Warning “Resource Group ‘$ResourceGroupName’ exists but in location ‘$($rg.Location)’ instead of desired ‘$Location’.”
}
# Optional: Check and update tags if different
if ($rg.Tags -ne $Tags) { # Simplified comparison; real check is more complex
Write-Host “Updating tags for Resource Group ‘$ResourceGroupName’…”
Set-AzResourceGroup -Name $ResourceGroupName -Tag $Tags
}
} else {
Write-Host “Creating Resource Group ‘$ResourceGroupName’ in ‘$Location’ with tags…”
New-AzResourceGroup -Name $ResourceGroupName -Location $Location -Tag $Tags
}
“`

Security Best Practices: Protecting Credentials and Resources

This is paramount. Never hardcode credentials (passwords, secrets, keys).

  • Authentication:
    • Managed Identities: The preferred method for scripts running on Azure resources (VMs, App Service, Functions, Azure Automation). The script authenticates as the identity of the Azure resource itself, eliminating the need to manage credentials.
      • System-Assigned: Tied to the lifecycle of the Azure resource.
      • User-Assigned: Standalone Azure resource that can be assigned to multiple Azure services.
    • Service Principals: An identity created for applications, hosted services, and automated tools to access Azure resources. Requires creating an App Registration in Azure AD (Entra ID) and granting it RBAC roles. Authentication uses Client ID, Tenant ID, and either a Client Secret or a Certificate. Store these credentials securely (see Key Vault below).
    • User Accounts (Interactive): Suitable for manual execution or attended automation, but not recommended for unattended scripts due to security risks and potential MFA challenges.
  • Secret Management:
    • Azure Key Vault: A secure store for secrets, keys, and certificates. Scripts can authenticate to Key Vault (ideally using a Managed Identity or Service Principal) and retrieve secrets at runtime. This avoids embedding secrets in code or configuration files.
      • PowerShell: Get-AzKeyVaultSecret
      • Azure CLI: az keyvault secret show
  • Least Privilege Principle: Grant the Managed Identity or Service Principal only the minimum RBAC roles required to perform its tasks at the narrowest possible scope (e.g., Resource Group level instead of Subscription level). Regularly review permissions.
  • Code Scanning: Use security scanning tools (built into Azure DevOps, GitHub, or third-party tools) to detect hardcoded secrets or vulnerable code patterns.

(More details in the Security Deep Dive section.)

Performance Optimization: Writing Efficient Code

Slow scripts waste time and resources.

  • Filtering Early: When retrieving Azure resources, use server-side filtering parameters whenever possible instead of retrieving large collections and filtering client-side.
    • Good (PowerShell): Get-AzVM -ResourceGroupName "MyRG" -Name "MyVM*"
    • Less Efficient (PowerShell): Get-AzVM | Where-Object { $_.ResourceGroupName -eq "MyRG" -and $_.Name -like "MyVM*" }
    • Good (CLI): az vm list --resource-group "MyRG" --query "[?starts_with(name,'MyVM')].{Name:name, PowerState:powerState}"
    • Less Efficient (CLI): az vm list --query "[].{Name:name, PowerState:powerState, RG:resourceGroup}" | jq '.[] | select(.RG=="MyRG" and (.Name | startswith("MyVM")))' (though jq is fast, server-side filtering is better)
  • Selecting Specific Properties: If you only need certain properties of an object, retrieve only those.
    • PowerShell: Select-Object (client-side, less optimal than server-side selection if available). Some cmdlets might have specific parameters for this.
    • Azure CLI: Use the --query parameter with JMESPath expressions to select and reshape data server-side.
  • Parallel Processing: For tasks involving many independent resources (e.g., tagging 100 VMs), perform operations in parallel.
    • PowerShell: ForEach-Object -Parallel (PowerShell 7+), Start-Job, Start-ThreadJob, or PowerShell Workflows (older).
    • Azure CLI / Bash: Use background processes (&), wait, xargs -P, or parallel.
  • Caching: If fetching static or slowly changing data repeatedly (e.g., available VM sizes in a region), cache the results temporarily within the script run.
  • API Throttling: Be aware of Azure API rate limits. Implement exponential backoff and retry logic for transient errors or throttling responses (HTTP 429). Most Az cmdlets and az commands have some built-in retry logic, but complex scripts might need explicit handling.

5. Advanced Scripting Techniques

Moving beyond basic CRUD operations.

Working with Azure Resource Manager (ARM) and Bicep Templates

While scripts can create resources directly, using ARM (JSON) or Bicep (DSL) templates is the standard for Infrastructure as Code deployments. Scripts can orchestrate these template deployments.

  • Triggering Deployments:
    • PowerShell: New-AzResourceGroupDeployment, New-AzSubscriptionDeployment
    • Azure CLI: az deployment group create, az deployment sub create
  • Passing Parameters: Templates define parameters; scripts provide the values dynamically or from configuration files.
    • PowerShell: -TemplateParameterFile "params.json", or -TemplateParameterObject (Get-Content params.json | ConvertFrom-Json) or using hashtables for parameters.
    • Azure CLI: --parameters "params.json" or --parameters key1=value1 key2=value2.
  • Handling Outputs: Templates can define outputs (e.g., the public IP address of a deployed VM). Scripts can retrieve these outputs after deployment for subsequent steps.
    • PowerShell: The deployment object returned by New-Az*Deployment contains an Outputs property.
    • Azure CLI: Use az deployment group show -g <RG> -n <DeploymentName> --query properties.outputs.
  • What-If Deployment: Preview the changes a template deployment would make without applying them. A crucial safety feature.
    • PowerShell: Add the -WhatIf switch to New-Az*Deployment.
    • Azure CLI: Use az deployment group what-if, az deployment sub what-if.

“`powershell

PowerShell Example: Deploying an ARM template

$params = @{
vmName = “myVMfromScript”
adminUsername = “azureuser”
adminPassword = (ConvertTo-SecureString “YourComplexP@ssw0rd!” -AsPlainText -Force) # Better: Use Key Vault reference
}
New-AzResourceGroupDeployment -ResourceGroupName “MyRG” -TemplateFile “.\azuredeploy.json” -TemplateParameterObject $params -WhatIf # Run WhatIf first!

Remove -WhatIf to deploy

$deployment = New-AzResourceGroupDeployment -ResourceGroupName “MyRG” -TemplateFile “.\azuredeploy.json” -TemplateParameterObject $params
$publicIp = $deployment.Outputs.publicIPAddress.value
Write-Host “VM Deployed with Public IP: $publicIp”
“`

“`bash

Azure CLI Example: Deploying a Bicep file

Ensure Bicep CLI is installed or use az bicep install

PARAM_FILE=”params.json” # Contains parameter values
RG_NAME=”MyRG”
BICEP_FILE=”main.bicep”
DEPLOYMENT_NAME=”bicepDeploy_$(date +%s)”

Run What-If first

az deployment group what-if –resource-group “$RG_NAME” –template-file “$BICEP_FILE” –parameters “$PARAM_FILE”

Deploy for real

az deployment group create –resource-group “$RG_NAME” –template-file “$BICEP_FILE” –parameters “$PARAM_FILE” –name “$DEPLOYMENT_NAME”

Get outputs

PUBLIC_IP=$(az deployment group show –resource-group “$RG_NAME” –name “$DEPLOYMENT_NAME” –query “properties.outputs.publicIPAddress.value” -o tsv)
echo “VM Deployed with Public IP: $PUBLIC_IP”
“`

Complex Resource Management

  • Advanced Querying:
    • PowerShell: Use -Filter parameters where available (OData syntax). Pipe to Where-Object for complex client-side filtering.
    • Azure CLI: Leverage the powerful --query parameter with JMESPath syntax for server-side filtering and shaping of JSON output.
    • Azure Resource Graph: For fast, large-scale querying across subscriptions, use Azure Resource Graph.
      • PowerShell: Search-AzGraph
      • Azure CLI: az graph query
        “`powershell

      Find all VMs with tag ‘Environment=Production’ across subscriptions

      Search-AzGraph -Query “Resources | where type =~ ‘microsoft.compute/virtualmachines’ and tags.Environment =~ ‘Production’ | project name, location, resourceGroup”
      bash

      Find all VMs with tag ‘Environment=Production’ across subscriptions

      az graph query -q “Resources | where type =~ ‘microsoft.compute/virtualmachines’ and tags.Environment =~ ‘Production’ | project name, location, resourceGroup” –output table
      “`

  • Tagging Strategies: Use scripts to enforce tagging standards, find untagged resources, or update tags in bulk.
  • Managing Dependencies: Scripts often need to create resources in a specific order (e.g., VNet before VM). Implement checks or use dependsOn within ARM/Bicep templates orchestrated by the script.
  • Resource Locking: Use scripts to apply or remove management locks (CanNotDelete, ReadOnly) to protect critical resources.
    • PowerShell: New-AzResourceLock, Remove-AzResourceLock
    • Azure CLI: az lock create, az lock delete

Integrating with Other Azure Services

Enhance scripts by interacting with other relevant Azure services:

  • Azure Key Vault: Retrieve secrets, keys, or certificates securely at runtime.
  • Azure Storage: Upload/download script files, configuration data, logs, or outputs to Blob Storage. Manage file shares.
  • Azure Monitor: Query logs (Log Analytics) using Invoke-AzOperationalInsightsQuery (PowerShell) or az monitor log-analytics query (CLI). Send custom logs or metrics. Trigger alert rules.
  • Azure Active Directory (Entra ID): Manage users, groups, service principals, app registrations.
    • PowerShell: Az.Resources (for Service Principals), Microsoft.Graph module (preferred for comprehensive user/group management).
    • Azure CLI: az ad sp, az ad user, az ad group. az rest for Microsoft Graph API calls.
  • Azure Policy: Trigger policy evaluations or remediation tasks. Assign policies or initiatives.
  • Azure App Configuration: Centralized store for application settings and feature flags, accessible by scripts.

Asynchronous Operations and Parallel Processing

For long-running operations (like VM creation) or bulk tasks, don’t wait sequentially.

  • Azure PowerShell:
    • -AsJob Parameter: Many long-running cmdlets (e.g., New-AzVM) support the -AsJob parameter. This returns a job object immediately, and you can check its status later with Get-Job and retrieve results with Receive-Job.
    • ForEach-Object -Parallel (PS 7+): Efficiently run script blocks in parallel threads. Use $using: scope modifier to pass variables. Control throttle limit.
    • Start-ThreadJob / Start-Job: More general-purpose background job execution.
  • Azure CLI / Bash:
    • --no-wait Parameter: Some az commands support --no-wait, returning immediately while the operation continues in Azure. You typically need to poll the resource state later using az vm show or similar commands.
    • Backgrounding (&): Run commands in the background: az vm create ... &. Capture the process ID ($!) if needed. Use wait to pause the script until background jobs complete.
    • xargs / parallel: Powerful tools for parallel execution of commands based on input lists.
      bash
      # Example: Restart multiple VMs in parallel using xargs
      az vm list -g MyRG --query "[?powerState=='VM running'].name" -o tsv | \
      xargs -P 5 -I {} az vm restart -g MyRG -n {} --no-wait
      # -P 5: Run up to 5 jobs in parallel
      # -I {}: Replace {} with the VM name from input
      # --no-wait: Return immediately after initiating restart

Leveraging Azure REST APIs Directly

When PowerShell cmdlets or CLI commands don’t offer the required functionality, or you need fine-grained control, you can call the Azure Resource Manager REST APIs directly.

  • PowerShell: Invoke-AzRestMethod or Invoke-RestMethod. Need to handle authentication (acquiring an access token) and construct the URI and request body. Invoke-AzRestMethod simplifies authentication when logged in via Connect-AzAccount.
  • Azure CLI: az rest. Automatically handles authentication based on the current login context. You specify the URI, method (GET, POST, PUT, PATCH, DELETE), and optionally a request body.

“`powershell

PowerShell Example: Get resource group details using REST API

$rgName = “MyRG”
$subscriptionId = (Get-AzContext).Subscription.Id
$uri = “https://management.azure.com/subscriptions/$subscriptionId/resourcegroups/$rgName?api-version=2021-04-01”
Invoke-AzRestMethod -Path $uri -Method GET
“`

“`bash

Azure CLI Example: Get resource group details using REST API

RG_NAME=”MyRG”
SUBSCRIPTION_ID=$(az account show –query id -o tsv)
URI=”/subscriptions/$SUBSCRIPTION_ID/resourcegroups/$RG_NAME?api-version=2021-04-01″
az rest –method get –uri $URI
“`

6. Infrastructure as Code (IaC) Integration

Enhanced scripts often play a vital role within a broader IaC strategy.

Scripts as Part of the IaC Lifecycle

While declarative tools like ARM, Bicep, or Terraform define the desired state, scripts are often needed for imperative tasks that these tools don’t handle easily:

  • Running custom configuration inside a VM after provisioning.
  • Seeding a database.
  • Calling external APIs.
  • Performing complex validation logic.
  • Orchestrating multiple template deployments with dependencies.

Pre- and Post-Deployment Scripting

  • Pre-Deployment: Scripts can run before an IaC deployment to:
    • Validate prerequisites (e.g., check resource quotas, existence of dependent resources).
    • Generate dynamic parameter values.
    • Retrieve secrets from Key Vault to pass into the deployment.
  • Post-Deployment: Scripts can run after a successful IaC deployment to:
    • Perform configuration tasks not covered by the template (using VM extensions like Custom Script Extension or dscExtension).
    • Run integration tests.
    • Update DNS records.
    • Send notifications.

Combining Scripts with ARM/Bicep/Terraform

  • ARM/Bicep: Use deploymentScripts resource type (runs PowerShell/CLI scripts as part of the deployment). Alternatively, use Custom Script Extension or dscExtension for VMs. Trigger scripts externally from a CI/CD pipeline before/after the template deployment.
  • Terraform: Use local-exec or remote-exec provisioners to run scripts during resource creation or destruction (use with caution, especially remote-exec). Use external data source to call scripts and consume their output. The recommended approach is often to separate script execution into distinct stages in a CI/CD pipeline rather than embedding it heavily within Terraform apply.

7. Development Workflow and Best Practices

Treat your automation scripts like production code.

Version Control with Git

  • Store all scripts and configuration files in a Git repository (Azure Repos, GitHub, GitLab).
  • Use meaningful commit messages.
  • Utilize branching strategies (e.g., Gitflow, GitHub Flow) for developing new features or fixes without impacting the main/production branch.
  • Implement Pull Requests (PRs) with code reviews for collaboration and quality control.

Testing Your Scripts

Testing ensures scripts work as expected and prevents regressions.

  • Pester (PowerShell): The standard testing framework for PowerShell. Write tests (*.Tests.ps1 files) to:
    • Unit Test: Test individual functions in isolation (using mocking if they interact with Azure).
    • Integration Test: Test script interaction with live Azure resources (use a dedicated test environment!).
    • Infrastructure Test: Validate the state of Azure resources after a script has run.
      “`powershell

    Example Pester Test (MyScript.Tests.ps1)

    Describe ‘New-AzResourceGroupIfNotExists Function’ {
    BeforeAll {
    # Mock the Get-AzResourceGroup and New-AzResourceGroup cmdlets
    Mock Get-AzResourceGroup { return $null } -ParameterFilter { $Name -eq ‘TestRG’ }
    Mock New-AzResourceGroup { return @{ ResourceGroupName = ‘TestRG’; Location = ‘WestEurope’; ProvisioningState = ‘Succeeded’ } } -Verifiable
    }

    It 'Should call New-AzResourceGroup if the group does not exist' {
        New-AzResourceGroupIfNotExists -ResourceGroupName 'TestRG' -Location 'WestEurope'
        Assert-VerifiableMocks
    }
    
    # More tests: case where group exists, validation failures etc.
    

    }
    ``
    * **Azure CLI / Bash:** Testing is less standardized. Use shell testing frameworks like
    ShellSpec,bats, or write simple validation scripts that check exit codes and output. Perform integration tests by running scripts against a test Azure environment and validating the resulting infrastructure state usingaz` commands and queries.

Continuous Integration and Continuous Deployment (CI/CD)

Integrate scripts into automated pipelines (Azure Pipelines, GitHub Actions).

  • CI Pipeline:
    • Triggered on code commits/PRs.
    • Lints the script code (e.g., PSScriptAnalyzer for PowerShell).
    • Runs automated tests (Pester, etc.).
    • Builds artifacts (e.g., PowerShell modules).
  • CD Pipeline:
    • Triggered after successful CI or manually.
    • Securely connects to Azure (using Service Connections with Workload Identity Federation or Service Principals).
    • Deploys/Runs scripts against target environments (Dev, Test, Prod).
    • Includes approval gates for sensitive environments.
    • Logs execution details.

Documentation and Commenting Standards

Make scripts understandable for others (and your future self).

  • Header Blocks: Include synopsis, description, parameters, examples (like PowerShell comment-based help).
  • Inline Comments: Explain complex logic, assumptions, or workarounds.
  • Function/Module Documentation: Document purpose, parameters, and return values.
  • README Files: Provide an overview of the script/module library, setup instructions, usage examples, and configuration details in the Git repository.

8. Security Deep Dive

Revisiting security with more detail.

Managed Identities for Azure Resources

The most secure way for scripts running within Azure to authenticate.

  • Enabling: Enable System-Assigned or assign a User-Assigned Managed Identity to the Azure resource running the script (VM, Automation Account, Function App, etc.).
  • Granting Permissions: Assign the necessary RBAC roles to the Managed Identity (e.g., grant the VM’s identity the Contributor role on a specific Resource Group).
  • Using in Scripts:
    • PowerShell (Az module): Connect-AzAccount -Identity (automatically uses the identity of the environment it’s running in).
    • Azure CLI: az login --identity (for system-assigned) or az login --identity -u <USER_ASSIGNED_MI_CLIENT_ID> (for user-assigned).
    • REST API: Obtain a token from the instance metadata service endpoint (IMDS) available on the Azure resource.

Service Principals: Creation and Usage

For scripts running outside Azure (developer machine, CI/CD agent, on-premises server).

  • Creation:
    • Azure Portal: App registrations -> New registration. Create a client secret or certificate.
    • Azure CLI: az ad sp create-for-rbac --name <SPName> --role <RoleName> --scopes <Scope> (Creates SP and assigns role). Note down the appId, password (secret), and tenant.
    • PowerShell: New-AzADServicePrincipal (more manual steps often required).
  • Secure Storage: Never store the secret or certificate directly in the script. Use:
    • Azure Key Vault.
    • Secure environment variables in CI/CD systems (e.g., Azure Pipelines Secret Variables, GitHub Secrets).
    • Secure configuration mechanisms on the host machine.
  • Authentication: Use the Client ID, Tenant ID, and Secret/Certificate with Connect-AzAccount -ServicePrincipal or az login --service-principal.
  • Workload Identity Federation (CI/CD): A newer, more secure method for GitHub Actions, Azure Pipelines, etc., to authenticate using a Service Principal without needing a long-lived secret. It uses short-lived tokens based on the CI/CD environment’s identity.

Securely Handling Secrets with Azure Key Vault

  • Store Secrets: Place connection strings, API keys, SP secrets, passwords in Key Vault.
  • Access Control: Grant the script’s identity (Managed Identity or Service Principal) specific permissions (Get, List) on the secrets within Key Vault using Access Policies or RBAC for Key Vault.
  • Retrieve at Runtime:
    “`powershell
    # Using Managed Identity (script running on Azure resource with MI)
    Connect-AzAccount -Identity
    $databasePassword = Get-AzKeyVaultSecret -VaultName “MyKV” -Name “SqlPassword” -AsPlainText

    Using Service Principal (ensure SP has Get permission on the secret)

    Connect-AzAccount -ServicePrincipal …

    $databasePassword = Get-AzKeyVaultSecret -VaultName “MyKV” -Name “SqlPassword” -AsPlainText
    bash

    Using Managed Identity

    az login –identity
    DATABASE_PASSWORD=$(az keyvault secret show –vault-name “MyKV” –name “SqlPassword” –query value -o tsv)

    Using Service Principal

    az login –service-principal …

    DATABASE_PASSWORD=$(az keyvault secret show –vault-name “MyKV” –name “SqlPassword” –query value -o tsv)
    “`

Role-Based Access Control (RBAC) for Scripts

Apply the principle of least privilege rigorously.

  • Analyze the exact operations the script needs to perform.
  • Identify the minimum required RBAC permissions for those operations (e.g., Microsoft.Compute/virtualMachines/start/action, Microsoft.Storage/storageAccounts/read).
  • Use built-in roles (Reader, Virtual Machine Contributor, Storage Blob Data Contributor) whenever possible.
  • If necessary, create custom RBAC roles with only the precise permissions needed.
  • Assign the role to the script’s identity (Managed Identity or Service Principal) at the narrowest possible scope (Resource Group or even individual Resource level, rather than Subscription).

9. Real-World Examples and Use Cases

Applying the principles to common scenarios.

  • Automated VM Deployment and Configuration:
    • Script takes parameters (VM name, size, image, network).
    • Retrieves admin password or SSH key from Key Vault.
    • Uses New-AzVM / az vm create or triggers an ARM/Bicep deployment.
    • Waits for VM provisioning.
    • Uses Custom Script Extension or Invoke-AzVMRunCommand / az vm run-command invoke to run configuration scripts inside the VM (install software, configure services).
    • Logs progress and outputs VM details (IP address).
    • Idempotency: Checks if VM already exists.
  • Scheduled Resource Cleanup:
    • Script runs on a schedule (e.g., using Azure Automation, Azure Functions Timer Trigger, cron job).
    • Authenticates using Managed Identity.
    • Uses Azure Resource Graph or Get-AzResource / az resource list to find resources matching specific criteria (e.g., untagged resources, resources older than X days, resources tagged for deletion).
    • Applies appropriate filtering (e.g., exclude production resources).
    • Logs potential resources for deletion (Dry Run mode).
    • If not Dry Run, proceeds to delete resources with proper error handling and logging.
  • User Access Management (Azure AD / Entra ID):
    • Script for onboarding new users: Takes user details, creates user account, assigns licenses, adds to relevant groups based on department/role. Uses Microsoft.Graph PowerShell module or az ad user/az ad group / az rest with MS Graph API.
    • Script for offboarding: Disables user account, removes group memberships, revokes sessions, archives data (interacts with other services like Exchange Online, SharePoint).
    • Uses Service Principal with appropriate Graph API permissions (User.ReadWrite.All, Group.ReadWrite.All).
  • Automated Monitoring and Alert Responses:
    • Azure Monitor alert triggers an Action Group, which runs an Azure Function, Logic App, or Azure Automation runbook containing a script.
    • Script receives alert context (payload).
    • Authenticates using Managed Identity.
    • Performs automated remediation (e.g., restart VM, increase disk size, add firewall rule).
    • Updates alert status or logs actions taken back to Log Analytics or a ticketing system.

10. Beyond Standalone Scripts: Azure Automation, Functions, and Logic Apps

While enhanced standalone scripts are powerful, Azure offers platform services designed for more complex automation scenarios:

  • Azure Automation:
    • Managed service for process automation (runbooks), configuration management (DSC), and update management.
    • Runbooks can be PowerShell or Python scripts.
    • Provides scheduling, credential/variable/certificate management (secure assets), source control integration, and execution logging.
    • Runbooks run in an Azure sandbox or on a Hybrid Runbook Worker (for on-premises/other cloud resources).
    • Authenticates using Managed Identities (System or User-Assigned) or classic Run As accounts (legacy).
    • Ideal for scheduled tasks, operational procedures, and integrating with DSC.
  • Azure Functions:
    • Serverless compute service for running event-driven code.
    • Supports multiple languages (including PowerShell, Python, C#, Node.js).
    • Can be triggered by HTTP requests, timers, Azure service events (Blob Storage, Service Bus, Event Grid), etc.
    • Excellent for event-based automation, lightweight APIs, and integration tasks.
    • Can leverage PowerShell/CLI via managed dependencies or by invoking the binaries.
    • Authenticates using Managed Identities.
    • Pay-per-execution model (Consumption plan) or dedicated App Service Plan.
  • Azure Logic Apps:
    • Serverless workflow automation service with a visual designer.
    • Connects hundreds of services (Azure and third-party) using pre-built connectors.
    • Low-code/no-code approach for integration and orchestration.
    • Can incorporate Azure Functions or inline code snippets for custom logic.
    • Can trigger scripts or be triggered by script actions via HTTP requests.
    • Ideal for complex workflows involving multiple systems and APIs, especially when visual design is preferred.

When to choose what?

  • Standalone Scripts (PowerShell/CLI): Ad-hoc tasks, local execution, simple CI/CD integration, components within larger systems.
  • Azure Automation Runbooks: Scheduled operational tasks, long-running processes, integration with DSC, centralized script management within Azure.
  • Azure Functions: Event-driven automation, short-lived tasks, API integrations, serverless compute needs, multi-language requirements.
  • Logic Apps: Complex workflow orchestration, connecting diverse SaaS/PaaS services, visual design preference, citizen developer scenarios.

Often, these services are used together. A Logic App might orchestrate a process that calls an Azure Function, which in turn executes an enhanced PowerShell script from an Automation Account.

11. Conclusion: Embracing Enhanced Automation

Moving from basic Azure scripting to creating enhanced, robust automation assets is a crucial step in effectively managing modern cloud environments. By embracing principles like modularity, parameterization, robust error handling, idempotency, and strong security practices, you transform simple command sequences into reliable, maintainable, and scalable solutions.

Leveraging tools like Azure PowerShell and Azure CLI effectively, integrating with Azure services like Key Vault and Monitor, adopting Infrastructure as Code practices, and following sound development workflows (version control, testing, CI/CD) are all part of this journey. Understanding when to use standalone scripts versus platform services like Azure Automation, Functions, or Logic Apps allows you to choose the right tool for the job.

The investment in writing enhanced Azure scripts pays off significantly through increased operational efficiency, reduced errors, improved security posture, and faster response times. It empowers individuals and teams to manage complex Azure estates confidently and consistently. Start incorporating these principles into your scripting practices today, and continuously refine your approach as you gain experience and Azure evolves. The future of cloud management is automated, and enhanced scripting is a cornerstone of that future.


Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top