tangyibo / gpupgrade

GPDB major version upgrade utility

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

gpupgrade Concourse Build Status

gpupgrade runs pg_upgrade across all segments to upgrade a Greenplum cluster across major versions. For further details read the Greenplum Database Upgrade documentation and blog post. We warmly welcome any feedback and contributions.

Purpose:

Greenplum has multiple ways of upgrading including backup & restore and gpcopy. These methods usually require additional diskspace for the required copy and significant downtime. gpupgrade can do fast in-place upgrades without the need for additional hardware, disk space, and with less downtime.

Creating an easy upgrade path enables users to quickly and confidently upgrade. This enables Greenplum to have faster release cycles with faster user feedback. Most importantly it allows Greenplum to reduce its reliance on supporting legacy versions.

Supported Versions:

Source Cluster Target Cluster
5 6
6 7 (future work)

Architecture:

gpupgrade consists of three processes that communicate using gRPC and protocol buffers:

  • CLI
    • Runs on the coordinator host
    • Consists of a gRPC client
  • Hub
    • Runs on the coordinator host
    • Upgrades the coordinator
    • Coordinates the agent processes
    • Consists of a gRPC client and server
  • Agents
    • Run on all segment hosts
    • Upgrade the standby, primary, and mirror segments
    • Execute commands received from the hub
    • Consist of a gRPC server
       CLI                     Hub                     Agent
      ------                  ------                  -------
  gRPC client    <-->      gRPC server
                                ^
                                |
                                V
                           gRPC client     <-->      gRPC server

Steps:

Running gpupgrade consists of several steps (ie: commands):

  • gpupgrade initialize
    • The source cluster can still be running. No downtime.
    • Substeps include creating the gpupgrade state directory, starting the hub and agents, creating the target cluster, and running pre-upgrade checks.
  • gpupgrade execute
    • This step will stop the source cluster. Downtime is needed.
    • Substeps include upgrading the coordinator, copying the coordinator catalog to the segments, and upgrading the primaries.
    • The coordinator contains only catalog information and no data, and is used to upgrade the catalog of the primaries. That is, after the target cluster coordinator is upgraded it's copied to each of the primary data directories on the target cluster to upgrade their catalog. Next, the target cluster primaries data is upgraded using pg_upgrade.
  • gpupgrade finalize
    • After finalizing the upgrade cannot be reverted.
    • Substeps include updating the data directories and coordinator catalog, and upgrading the standby and mirrors.

Optional steps (ie: commands):

  • gpupgrade revert
    • To restore the cluster to the state before the upgrade.
    • Can be run after initialize or execute, but not finalize.
    • Substeps include deleting the target cluster, archiving the gpupgrade log directory, and restoring the source cluster.
  start <---- run migration
    |            scripts  |
run migration             |
  scripts                 ^
    |                     |
    V                     |
initialize ---> revert ----
    |                     ^
    V                     |
 execute  ----> revert ----
    |
    V
 finalize
    |
run migration 
  scripts
    |
    V
   done

Each substep within a step implements crash-only idempotence. This means that if an error occurs and is fixed then on rerun the step will succeed. This requires each substep to clean up any side effects it creates, or possibly check if the work has been done.

Link vs. Copy Mode:

gpupgrade inits a fresh target cluster "next to" the source cluster, and upgrades "into it" in-place using pg_upgrade's copy or link mode.

Attribute Copy Mode Link Mode
Description Copy's source files to the target cluster. Uses hard links to modify the source cluster data in place.
Upgrade Time Slow, since it copy's the data before upgrading. Fast, since the data is modified in place.
Disk Space ~60% free disk space needed. ~20% free disk space needed.
Revert Speed Fast, since the source cluster remains untouched. Slow, since the source files have been modified the primaries and mirrors need to be rebuilt.
Risk Less risky since the source cluster is untouched. More risky since the source cluster is modified.

Getting Started

Prerequisites

  • Golang. See the top of go.mod for the current version used.
  • protoc. This is the compiler for the gRPC protobuf system which can be installed on macOS with brew install protobuf.
  • Run make && make depend-dev to install other developer dependencies. Note make needs to be run first.

Setting up your IDE

Vim

Checkout vim-go and go-delve.

IntelliJ

Golang
  • In Preferences > Plugins, install the "Go" plugin from JetBrains.
Imports
  • Preferences > Editor > Code Style > Go > select "Imports" tab
    • uncheck "Use back quotes for imports"
    • uncheck "Add parentheses for a single import"
    • uncheck "Remove redundant import aliases"
    • Sorting type: gofmt
    • check "Move all imports in a single declaration"
    • check "Group stdlib imports"
      • check "Move all stdlib imports in a single group"
    • check "Group"
      • check "Current project packages"
Copyright
  • Preferences > Editor > Copyright > Copyright Profiles
    • Add new profile called "vmware" with the following text:

      Copyright (c) 2017-$originalComment.match("Copyright \(c\) (\d+)", 1, "-")$today.year VMware, Inc. or its affiliates
      SPDX-License-Identifier: Apache-2.0
      
  • Preferences > Editor > Copyright
    • Select "vmware" for default project copyright.
  • Preferences > Editor > Copyright > Formatting
    • Select "Use custom formatting options"
    • For Comment Type: select "use line comment"
    • For Relative Location: select "Before other comments" and check "Add blank line after"
Formatting
  • Preferences > Tools > Actions on Save
    • check "Reformat code" and "Optimize imports"

Build and Test

make         # builds gpupgrade binary locally
make check   # runs tests
make install # installs gpupgrade into $GOBIN

Cross-compile with:

  • make build_linux
  • make build_mac

Running

gpupgrade initialize --file ./gpupgrade_config
OR
gpupgrade initialize --source-gphome "$GPHOME" --target-gphome "$GPHOME" --source-master-port 6000 --disk-free-ratio 0
gpupgrade execute
gpupgrade finalize

Running Tests

Unit tests

make unit

Integration tests

Tests that run against the gpupgrade binary to verify the interaction between components. Before writing a new integration test please review the README.

make integration

Acceptance tests

Tests more end-to-end acceptance-level behavior between components. Tests are located in the test directory and use the BATS (Bash Automated Testing System) framework which can be installed on macOS with brew install bats-core. Please review the integrations/README.

# Some tests require GPDB installed and running
make gpupgrade-tests
make pg-upgrade-tests

To run all tests in a suite:

bats test/acceptance/gpupgrade/finalize.bats

To run a single test or set of tests:

bats -f "gpupgrade finalize should" test/acceptance/gpupgrade/finalize.bats

All local tests

# Runs all local tests
make check

End-to-End tests

Creates a Concourse pipeline that includes various multi-host X-to-Y upgrade and functional tests. These cannot be run locally.

make set-pipeline

Concourse Pipeline

To update the pipeline edit the yaml files in the ci directory and run make set-pipeline.

The yaml files in the ci directory are concatenated to create ci/generated/template.yml. Next, go generate ./ci is executed which runs go run ./parser/parse_template.go generated/template.yml generated/pipeline.yml to create ci/generated/pipeline.yml. None of the generated files template.yml or pipeline.yml are checked in.

To update the production pipeline: PIPELINE_NAME=gpupgrade FLY_TARGET=prod make set-pipeline

To make the pipeline publicly visible run make expose-pipeline. This will allow anyone to see the pipeline and its status. However, the task details will not be visible unless one logs into Concourse.

Note: If your dev pipeline is failing on the build job while verifying the rpm then the most likely cause is needing to sync the latest tags on origin with your remote. This allows the GPDB test rpm to have the correct version number. On your GPDB branch run the following:

$ git fetch --tags origin
$ git push --tags <yourRemoteName>

If you already flew a pipeline before pushing tags you will likely need to delete it, push tags, and re-fly as Concourse has some weird caching issues.

Generating gRPC code

To recompile proto files to generate gRPC client and server code run go generate ./idl

Bash Completion

To enable tab completion of gpupgrade commands source the cli/bash/gpupgrade.bash script from your ~/.bash_completion config, or copy it into your system's completions directory such as /etc/bash_completion.d.

Log Locations

Logs are located on all hosts.

  • gpupgrade logs: $HOME/gpAdminLogs/gpupgrade
    • After finalize the directory is archived with format gpupgrade-<timestamp-upgradeID>.
  • pg_upgrade logs: $HOME/gpAdminLogs/gpupgrade/pg_upgrade
  • greenplum utility logs: $HOME/gpAdminLogs
  • source cluster pg_log: $MASTER_DATA_DIRECTORY/pg_log
  • target cluster pg_log: $(gpupgrade config show --target-datadir)/pg_log
    • The target cluster data directories are located next to the source cluster directories with the format -<upgradeID>.<contentID>

Debugging

  • Identify the High Level Failure
    • What mode was used - copy vs. link?
    • What step failed - initialize, execute, finalize, or revert?
    • What specific substep failed?
  • Identify the Failing Host
    • Did the Hub (coordinator) vs. Agent (segment) fail?
    • What specific host failed?
  • Identify the Failed Utility
    • Did gpupgrade fail, or an underlying utility such as pg_upgrade, gpinitsystem, gpstart, etc.?
  • Identify the Specific Failure
    • Based on the error context and logs what is the specific error?

Debugging Hub and Agent Processes

  • Set a breakpoint in the CLI
    • For example in cli/commands/initialize.go, execute.go, or finalize.go right before the gRPC call to the hub.
  • Run gpupgrade to hit the breakpoint in the CLI and start the hub process.
  • When using intellij "Attach to Process" and select the hub and/or agent processes.
  • Set additional breakpoints in the hub or agent code to aid in debugging.
  • Continue execution on the CLI until the additional breakpoints in the hub or agent code are hit. Step through the code to debug.
  • For faster iterations:
    • Make any local changes in the code
    • Rebuild with make && make install
    • Reload the new code with gpupgrade restart-services or manually stop and restart the hub.
    • Repeat the above breakpoints and attaching to the new processes as their PIDs have changed.

About

GPDB major version upgrade utility

License:Other


Languages

Language:Go 78.5%Language:Shell 17.8%Language:PLpgSQL 2.6%Language:Makefile 0.7%Language:Python 0.4%Language:Ruby 0.1%