1016 lines
		
	
	
		
			33 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
			
		
		
	
	
			1016 lines
		
	
	
		
			33 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
| #!/bin/bash
 | |
| 
 | |
| #
 | |
| # Copyright (C) 2015 The Android Open Source Project
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| #      http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| #
 | |
| 
 | |
| # Script to generate a Brillo update for use by the update engine.
 | |
| #
 | |
| # usage: brillo_update_payload COMMAND [ARGS]
 | |
| # The following commands are supported:
 | |
| #  generate    generate an unsigned payload
 | |
| #  hash        generate a payload or metadata hash
 | |
| #  sign        generate a signed payload
 | |
| #  properties  generate a properties file from a payload
 | |
| #  verify      verify a payload by recreating a target image.
 | |
| #  check       verify a payload using paycheck (static testing)
 | |
| #
 | |
| #  Generate command arguments:
 | |
| #  --payload                  generated unsigned payload output file
 | |
| #  --source_image             if defined, generate a delta payload from the
 | |
| #                             specified image to the target_image
 | |
| #  --target_image             the target image that should be sent to clients
 | |
| #  --metadata_size_file       if defined, generate a file containing the size
 | |
| #                             of the ayload metadata in bytes to the specified
 | |
| #                             file
 | |
| #  --disable_fec_computation  Disable the on device fec data computation for
 | |
| #                             incremental update. This feature is enabled by
 | |
| #                             default
 | |
| #  --force_minor_version      Override the minor version used for delta
 | |
| #                             generation.
 | |
| #
 | |
| #  Hash command arguments:
 | |
| #  --unsigned_payload    the input unsigned payload to generate the hash from
 | |
| #  --signature_size      signature sizes in bytes in the following format:
 | |
| #                        "size1:size2[:...]"
 | |
| #  --payload_hash_file   if defined, generate a payload hash and output to the
 | |
| #                        specified file
 | |
| #  --metadata_hash_file  if defined, generate a metadata hash and output to the
 | |
| #                        specified file
 | |
| #
 | |
| #  Sign command arguments:
 | |
| #  --unsigned_payload        the input unsigned payload to insert the signatures
 | |
| #  --payload                 the output signed payload
 | |
| #  --signature_size          signature sizes in bytes in the following format:
 | |
| #                            "size1:size2[:...]"
 | |
| #  --payload_signature_file  the payload signature files in the following
 | |
| #                            format:
 | |
| #                            "payload_signature1:payload_signature2[:...]"
 | |
| #  --metadata_signature_file the metadata signature files in the following
 | |
| #                            format:
 | |
| #                            "metadata_signature1:metadata_signature2[:...]"
 | |
| #  --metadata_size_file      if defined, generate a file containing the size of
 | |
| #                            the signed payload metadata in bytes to the
 | |
| #                            specified file
 | |
| #  Note that the number of signature sizes and payload signatures have to match.
 | |
| #
 | |
| #  Properties command arguments:
 | |
| #  --payload                 the input signed or unsigned payload
 | |
| #  --properties_file         the output path where to write the properties, or
 | |
| #                            '-' for stdout.
 | |
| #  Verify command arguments:
 | |
| #  --payload             payload input file
 | |
| #  --source_image        verify payload to the specified source image.
 | |
| #  --target_image        the target image to verify upon.
 | |
| #
 | |
| #  Check command arguments:
 | |
| #     Symmetrical with the verify command.
 | |
| 
 | |
| 
 | |
| # Exit codes:
 | |
| EX_UNSUPPORTED_DELTA=100
 | |
| 
 | |
| warn() {
 | |
|   echo "brillo_update_payload: warning: $*" >&2
 | |
| }
 | |
| 
 | |
| die() {
 | |
|   echo "brillo_update_payload: error: $*" >&2
 | |
|   exit 1
 | |
| }
 | |
| 
 | |
| # Loads shflags. We first look at the default install location; then our own
 | |
| # directory; finally the parent directory.
 | |
| load_shflags() {
 | |
|   local my_dir="$(dirname "$(readlink -f "$0")")"
 | |
|   local path
 | |
|   for path in /usr/share/misc \
 | |
|     "${my_dir}"/lib/shflags \
 | |
|     "${my_dir}"/../lib/shflags; do
 | |
|     if [[ -r "${path}/shflags" ]]; then
 | |
|       . "${path}/shflags" || die "Could not load ${path}/shflags."
 | |
|       return
 | |
|     fi
 | |
|   done
 | |
|   die "Could not find shflags."
 | |
| }
 | |
| 
 | |
| load_shflags
 | |
| 
 | |
| HELP_GENERATE="generate: Generate an unsigned update payload."
 | |
| HELP_HASH="hash: Generate the hashes of the unsigned payload and metadata used \
 | |
| for signing."
 | |
| HELP_SIGN="sign: Insert the signatures into the unsigned payload."
 | |
| HELP_PROPERTIES="properties: Extract payload properties to a file."
 | |
| HELP_VERIFY="verify: Verify a (signed) update payload using delta_generator."
 | |
| HELP_CHECK="check: Check a (signed) update payload using paycheck (static \
 | |
| testing)."
 | |
| 
 | |
| usage() {
 | |
|   echo "Supported commands:"
 | |
|   echo
 | |
|   echo "${HELP_GENERATE}"
 | |
|   echo "${HELP_HASH}"
 | |
|   echo "${HELP_SIGN}"
 | |
|   echo "${HELP_PROPERTIES}"
 | |
|   echo "${HELP_VERIFY}"
 | |
|   echo "${HELP_CHECK}"
 | |
|   echo
 | |
|   echo "Use: \"$0 <command> --help\" for more options."
 | |
| }
 | |
| 
 | |
| # Check that a command is specified.
 | |
| if [[ $# -lt 1 ]]; then
 | |
|   echo "Please specify a command [generate|hash|sign|properties|verify|check]"
 | |
|   exit 1
 | |
| fi
 | |
| 
 | |
| # Parse command.
 | |
| COMMAND="${1:-}"
 | |
| shift
 | |
| 
 | |
| case "${COMMAND}" in
 | |
|   generate)
 | |
|     FLAGS_HELP="${HELP_GENERATE}"
 | |
|     ;;
 | |
| 
 | |
|   hash)
 | |
|     FLAGS_HELP="${HELP_HASH}"
 | |
|     ;;
 | |
| 
 | |
|   sign)
 | |
|     FLAGS_HELP="${HELP_SIGN}"
 | |
|     ;;
 | |
| 
 | |
|   properties)
 | |
|     FLAGS_HELP="${HELP_PROPERTIES}"
 | |
|     ;;
 | |
| 
 | |
|   verify)
 | |
|     FLAGS_HELP="${HELP_VERIFY}"
 | |
|     ;;
 | |
| 
 | |
|   check)
 | |
|     FLAGS_HELP="${HELP_CHECK}"
 | |
|     ;;
 | |
| 
 | |
|   *)
 | |
|     echo "Unrecognized command: \"${COMMAND}\"" >&2
 | |
|     usage >&2
 | |
|     exit 1
 | |
|     ;;
 | |
| esac
 | |
| 
 | |
| # Flags
 | |
| FLAGS_HELP="Usage: $0 ${COMMAND} [flags]
 | |
| ${FLAGS_HELP}"
 | |
| 
 | |
| if [[ "${COMMAND}" == "generate" ]]; then
 | |
|   DEFINE_string payload "" \
 | |
|     "Path to output the generated unsigned payload file."
 | |
|   DEFINE_string target_image "" \
 | |
|     "Path to the target image that should be sent to clients."
 | |
|   DEFINE_string source_image "" \
 | |
|     "Optional: Path to a source image. If specified, this makes a delta update."
 | |
|   DEFINE_string metadata_size_file "" \
 | |
|     "Optional: Path to output metadata size."
 | |
|   DEFINE_string max_timestamp "" \
 | |
|     "Optional: The maximum unix timestamp of the OS allowed to apply this \
 | |
| payload, should be set to a number higher than the build timestamp of the \
 | |
| system running on the device, 0 if not specified."
 | |
|   DEFINE_string partition_timestamps "" \
 | |
|     "Optional: Per-partition maximum unix timestamp of the OS allowed to \
 | |
| apply this payload. Should be a comma separated key value pairs. e.x.\
 | |
| system:1234,vendor:456"
 | |
|   DEFINE_string disable_fec_computation "" \
 | |
|     "Optional: Disables the on device fec data computation for incremental \
 | |
| update. This feature is enabled by default."
 | |
|   DEFINE_string disable_verity_computation "" \
 | |
|     "Optional: Disables the on device verity computation for incremental \
 | |
| update. This feature is enabled by default."
 | |
|   DEFINE_string is_partial_update "" \
 | |
|     "Optional: True if the payload is for partial update. i.e. it only updates \
 | |
| a subset of partitions on device."
 | |
|   DEFINE_string full_boot "" "Will include full boot image"
 | |
|   DEFINE_string disable_vabc "" \
 | |
|     "Optional: Disables Virtual AB Compression when installing the OTA"
 | |
|   DEFINE_string enable_vabc_xor "" \
 | |
|     "Optional: Enable the use of Virtual AB Compression XOR feature"
 | |
|   DEFINE_string force_minor_version "" \
 | |
|     "Optional: Override the minor version for the delta generation."
 | |
|   DEFINE_string compressor_types "" \
 | |
|     "Optional: allowed compressor types. Colon separated, allowe values are bz2 and brotli"
 | |
|   DEFINE_string enable_zucchini "" \
 | |
|     "Optional: Whether to enable zucchini diffing"
 | |
|   DEFINE_string enable_lz4diff "" \
 | |
|     "Optional: Whether to enable lz4 diffing for EROFS"
 | |
|   DEFINE_string liblz4_path "" \
 | |
|     "Required if --enabled_lz4diff true is passed. Path to liblz4.so. delta_generator will use this copy of liblz4.so for compression. It is important that this copy of liblz4.so is the same as the one on source build."
 | |
|   DEFINE_string erofs_compression_param "" \
 | |
|     "Compression parameter passed to mkfs.erofs's -z option."
 | |
| fi
 | |
| if [[ "${COMMAND}" == "hash" || "${COMMAND}" == "sign" ]]; then
 | |
|   DEFINE_string unsigned_payload "" "Path to the input unsigned payload."
 | |
|   DEFINE_string signature_size "" \
 | |
|     "Signature sizes in bytes in the following format: size1:size2[:...]"
 | |
| fi
 | |
| if [[ "${COMMAND}" == "hash" ]]; then
 | |
|   DEFINE_string metadata_hash_file "" \
 | |
|     "Optional: Path to output metadata hash file."
 | |
|   DEFINE_string payload_hash_file "" \
 | |
|     "Optional: Path to output payload hash file."
 | |
| fi
 | |
| if [[ "${COMMAND}" == "sign" ]]; then
 | |
|   DEFINE_string payload "" \
 | |
|     "Path to output the generated unsigned payload file."
 | |
|   DEFINE_string metadata_signature_file "" \
 | |
|     "The metatada signatures in the following format: \
 | |
| metadata_signature1:metadata_signature2[:...]"
 | |
|   DEFINE_string payload_signature_file "" \
 | |
|     "The payload signatures in the following format: \
 | |
| payload_signature1:payload_signature2[:...]"
 | |
|   DEFINE_string metadata_size_file "" \
 | |
|     "Optional: Path to output metadata size."
 | |
| fi
 | |
| if [[ "${COMMAND}" == "properties" ]]; then
 | |
|   DEFINE_string payload "" \
 | |
|     "Path to the input signed or unsigned payload file."
 | |
|   DEFINE_string properties_file "-" \
 | |
|     "Path to output the extracted property files. If '-' is passed stdout will \
 | |
| be used."
 | |
| fi
 | |
| if [[ "${COMMAND}" == "verify" || "${COMMAND}" == "check" ]]; then
 | |
|   DEFINE_string payload "" \
 | |
|     "Path to the input payload file."
 | |
|   DEFINE_string target_image "" \
 | |
|     "Path to the target image to verify upon."
 | |
|   DEFINE_string source_image "" \
 | |
|     "Optional: Path to a source image. If specified, the delta update is \
 | |
| applied to this."
 | |
| fi
 | |
| 
 | |
| DEFINE_string work_dir "${TMPDIR:-/tmp}" "Where to dump temporary files."
 | |
| 
 | |
| # Parse command line flag arguments
 | |
| FLAGS "$@" || exit 1
 | |
| eval set -- "${FLAGS_ARGV}"
 | |
| set -e
 | |
| 
 | |
| # Override the TMPDIR with the passed work_dir flags, which anyway defaults to
 | |
| # ${TMPDIR}.
 | |
| TMPDIR="${FLAGS_work_dir}"
 | |
| export TMPDIR
 | |
| 
 | |
| # Associative arrays from partition name to file in the source and target
 | |
| # images. The size of the updated area must be the size of the file.
 | |
| declare -A SRC_PARTITIONS
 | |
| declare -A DST_PARTITIONS
 | |
| 
 | |
| # Associative arrays for the .map files associated with each src/dst partition
 | |
| # file in SRC_PARTITIONS and DST_PARTITIONS.
 | |
| declare -A SRC_PARTITIONS_MAP
 | |
| declare -A DST_PARTITIONS_MAP
 | |
| 
 | |
| # List of partition names in order.
 | |
| declare -a PARTITIONS_ORDER
 | |
| 
 | |
| # A list of PIDs of the extract_image workers.
 | |
| EXTRACT_IMAGE_PIDS=()
 | |
| 
 | |
| # A list of temporary files to remove during cleanup.
 | |
| CLEANUP_FILES=()
 | |
| 
 | |
| # Global options to force the version of the payload.
 | |
| FORCE_MAJOR_VERSION=""
 | |
| FORCE_MINOR_VERSION=""
 | |
| 
 | |
| # Path to the postinstall config file in target image if exists.
 | |
| POSTINSTALL_CONFIG_FILE=""
 | |
| 
 | |
| # Path to the dynamic partition info file in target image if exists.
 | |
| DYNAMIC_PARTITION_INFO_FILE=""
 | |
| 
 | |
| # Path to the META/apex_info.pb found in target build
 | |
| APEX_INFO_FILE=""
 | |
| 
 | |
| # read_option_int <file.txt> <option_key> [default_value]
 | |
| #
 | |
| # Reads the unsigned integer value associated with |option_key| in a key=value
 | |
| # file |file.txt|. Prints the read value if found and valid, otherwise prints
 | |
| # the |default_value|.
 | |
| read_option_uint() {
 | |
|   local file_txt="$1"
 | |
|   local option_key="$2"
 | |
|   local default_value="${3:-}"
 | |
|   local value
 | |
|   if value=$(grep "^${option_key}=" "${file_txt}" | tail -n 1); then
 | |
|     if value=$(echo "${value}" | cut -f 2- -d "=" | grep -E "^[0-9]+$"); then
 | |
|       echo "${value}"
 | |
|       return
 | |
|     fi
 | |
|   fi
 | |
|   echo "${default_value}"
 | |
| }
 | |
| 
 | |
| # truncate_file <file_path> <file_size>
 | |
| #
 | |
| # Truncate the given |file_path| to |file_size| using python.
 | |
| # The truncate binary might not be available.
 | |
| truncate_file() {
 | |
|   local file_path="$1"
 | |
|   local file_size="$2"
 | |
|   python -c "open(\"${file_path}\", 'a').truncate(${file_size})"
 | |
| }
 | |
| 
 | |
| # Create a temporary file in the work_dir with an optional pattern name.
 | |
| # Prints the name of the newly created file.
 | |
| create_tempfile() {
 | |
|   local pattern="${1:-tempfile.XXXXXX}"
 | |
|   mktemp --tmpdir="${FLAGS_work_dir}" "${pattern}"
 | |
| }
 | |
| 
 | |
| cleanup() {
 | |
|   local err=""
 | |
|   rm -f "${CLEANUP_FILES[@]}" || err=1
 | |
| 
 | |
|   # If we are cleaning up after an error, or if we got an error during
 | |
|   # cleanup (even if we eventually succeeded) return a non-zero exit
 | |
|   # code. This triggers additional logging in most environments that call
 | |
|   # this script.
 | |
|   if [[ -n "${err}" ]]; then
 | |
|     die "Cleanup encountered an error."
 | |
|   fi
 | |
| }
 | |
| 
 | |
| cleanup_on_error() {
 | |
|   trap - INT TERM ERR EXIT
 | |
|   cleanup
 | |
|   die "Cleanup success after an error."
 | |
| }
 | |
| 
 | |
| cleanup_on_exit() {
 | |
|   trap - INT TERM ERR EXIT
 | |
|   cleanup
 | |
| }
 | |
| 
 | |
| trap cleanup_on_error INT TERM ERR
 | |
| trap cleanup_on_exit EXIT
 | |
| 
 | |
| # extract_file <zip_file> <entry_name> <destination>
 | |
| #
 | |
| # Extracts |entry_name| from |zip_file| to |destination|.
 | |
| extract_file() {
 | |
|   local zip_file="$1"
 | |
|   local entry_name="$2"
 | |
|   local destination="$3"
 | |
| 
 | |
|   # unzip -p won't report error upon ENOSPC. Therefore, create a temp directory
 | |
|   # as the destination of the unzip, and move the file to the intended
 | |
|   # destination.
 | |
|   local output_directory=$(
 | |
|     mktemp --directory --tmpdir="${FLAGS_work_dir}" "TEMP.XXXXXX")
 | |
|   unzip "${zip_file}" "${entry_name}" -d "${output_directory}" ||
 | |
|     { rm -rf "${output_directory}"; die "Failed to extract ${entry_name}"; }
 | |
| 
 | |
|   mv "${output_directory}/${entry_name}" "${destination}"
 | |
|   rm -rf "${output_directory}"
 | |
| }
 | |
| 
 | |
| # extract_image <image> <partitions_array> [partitions_order]
 | |
| #
 | |
| # Detect the format of the |image| file and extract its updatable partitions
 | |
| # into new temporary files. Add the list of partition names and its files to the
 | |
| # associative array passed in |partitions_array|. If |partitions_order| is
 | |
| # passed, set it to list of partition names in order.
 | |
| extract_image() {
 | |
|   local image="$1"
 | |
| 
 | |
|   # Brillo images are zip files. We detect the 4-byte magic header of the zip
 | |
|   # file.
 | |
|   local magic=$(xxd -p -l4 "${image}")
 | |
|   if [[ "${magic}" == "504b0304" ]]; then
 | |
|     echo "Detected .zip file, extracting Brillo image."
 | |
|     extract_image_brillo "$@"
 | |
|     return
 | |
|   fi
 | |
| 
 | |
|   # Chrome OS images are GPT partitioned disks. We should have the cgpt binary
 | |
|   # bundled here and we will use it to extract the partitions, so the GPT
 | |
|   # headers must be valid.
 | |
|   if cgpt show -q -n "${image}" >/dev/null; then
 | |
|     echo "Detected GPT image, extracting Chrome OS image."
 | |
|     extract_image_cros "$@"
 | |
|     return
 | |
|   fi
 | |
| 
 | |
|   die "Couldn't detect the image format of ${image}"
 | |
| }
 | |
| 
 | |
| # extract_image_cros <image.bin> <partitions_array> [partitions_order]
 | |
| #
 | |
| # Extract Chromium OS recovery images into new temporary files.
 | |
| extract_image_cros() {
 | |
|   local image="$1"
 | |
|   local partitions_array="$2"
 | |
|   local partitions_order="${3:-}"
 | |
| 
 | |
|   local kernel root
 | |
|   kernel=$(create_tempfile "kernel.bin.XXXXXX")
 | |
|   CLEANUP_FILES+=("${kernel}")
 | |
|   root=$(create_tempfile "root.bin.XXXXXX")
 | |
|   CLEANUP_FILES+=("${root}")
 | |
| 
 | |
|   cros_generate_update_payload --extract \
 | |
|     --image "${image}" \
 | |
|     --kern_path "${kernel}" --root_path "${root}"
 | |
| 
 | |
|   # Chrome OS now uses major_version 2 payloads for all boards.
 | |
|   # See crbug.com/794404 for more information.
 | |
|   FORCE_MAJOR_VERSION="2"
 | |
| 
 | |
|   eval ${partitions_array}[kernel]=\""${kernel}"\"
 | |
|   eval ${partitions_array}[root]=\""${root}"\"
 | |
| 
 | |
|   if [[ -n "${partitions_order}" ]]; then
 | |
|     eval "${partitions_order}=( \"root\" \"kernel\" )"
 | |
|   fi
 | |
| 
 | |
|   local part varname
 | |
|   for part in kernel root; do
 | |
|     varname="${partitions_array}[${part}]"
 | |
|     printf "md5sum of %s: " "${varname}"
 | |
|     md5sum "${!varname}"
 | |
|   done
 | |
| }
 | |
| 
 | |
| # extract_partition_brillo <target_files.zip> <partitions_array> <partition>
 | |
| #     <part_file> <part_map_file>
 | |
| #
 | |
| # Extract the <partition> from target_files zip file into <part_file> and its
 | |
| # map file into <part_map_file>.
 | |
| extract_partition_brillo() {
 | |
|   local image="$1"
 | |
|   local partitions_array="$2"
 | |
|   local part="$3"
 | |
|   local part_file="$4"
 | |
|   local part_map_file="$5"
 | |
| 
 | |
|   # For each partition, we in turn look for its image file under IMAGES/ and
 | |
|   # RADIO/ in the given target_files zip file.
 | |
|   local path path_in_zip
 | |
|   for path in IMAGES RADIO; do
 | |
|     if unzip -l "${image}" "${path}/${part}.img" >/dev/null; then
 | |
|       path_in_zip="${path}"
 | |
|       break
 | |
|     fi
 | |
|   done
 | |
|   [[ -n "${path_in_zip}" ]] || die "Failed to find ${part}.img"
 | |
|   extract_file "${image}" "${path_in_zip}/${part}.img" "${part_file}"
 | |
| 
 | |
|   # If the partition is stored as an Android sparse image file, we need to
 | |
|   # convert them to a raw image for the update.
 | |
|   local magic=$(xxd -p -l4 "${part_file}")
 | |
|   if [[ "${magic}" == "3aff26ed" ]]; then
 | |
|     local temp_sparse=$(create_tempfile "${part}.sparse.XXXXXX")
 | |
|     echo "Converting Android sparse image ${part}.img to RAW."
 | |
|     mv "${part_file}" "${temp_sparse}"
 | |
|     simg2img "${temp_sparse}" "${part_file}"
 | |
|     rm -f "${temp_sparse}"
 | |
|   fi
 | |
| 
 | |
|   # Extract the .map file (if one is available).
 | |
|   if unzip -l "${image}" "${path_in_zip}/${part}.map" > /dev/null; then
 | |
|     extract_file "${image}" "${path_in_zip}/${part}.map" "${part_map_file}"
 | |
|   fi
 | |
| 
 | |
|   # delta_generator only supports images multiple of 4 KiB. For target images
 | |
|   # we pad the data with zeros if needed, but for source images we truncate
 | |
|   # down the data since the last block of the old image could be padded on
 | |
|   # disk with unknown data.
 | |
|   local filesize=$(stat -c%s "${part_file}")
 | |
|   if [[ $(( filesize % 4096 )) -ne 0 ]]; then
 | |
|     if [[ "${partitions_array}" == "SRC_PARTITIONS" ]]; then
 | |
|       echo "Rounding DOWN partition ${part}.img to a multiple of 4 KiB."
 | |
|       : $(( filesize = filesize & -4096 ))
 | |
|     else
 | |
|       echo "Rounding UP partition ${part}.img to a multiple of 4 KiB."
 | |
|       : $(( filesize = (filesize + 4095) & -4096 ))
 | |
|     fi
 | |
|     truncate_file "${part_file}" "${filesize}"
 | |
|   fi
 | |
| 
 | |
|   echo "Extracted ${partitions_array}[${part}]: ${filesize} bytes"
 | |
| }
 | |
| 
 | |
| # extract_image_brillo <target_files.zip> <partitions_array> [partitions_order]
 | |
| #
 | |
| # Extract the A/B updated partitions from a Brillo target_files zip file into
 | |
| # new temporary files.
 | |
| extract_image_brillo() {
 | |
|   local image="$1"
 | |
|   local partitions_array="$2"
 | |
|   local partitions_order="${3:-}"
 | |
| 
 | |
|   local partitions=( "boot" "system" )
 | |
|   local ab_partitions_list
 | |
|   ab_partitions_list=$(create_tempfile "ab_partitions_list.XXXXXX")
 | |
|   CLEANUP_FILES+=("${ab_partitions_list}")
 | |
|   if unzip -l "${image}" "META/ab_partitions.txt" > /dev/null; then
 | |
|     extract_file "${image}" "META/ab_partitions.txt" "${ab_partitions_list}"
 | |
|     if grep -v -E '^[a-zA-Z0-9_-]*$' "${ab_partitions_list}" >&2; then
 | |
|       die "Invalid partition names found in the partition list."
 | |
|     fi
 | |
|     # Get partition list without duplicates.
 | |
|     partitions=($(awk '!seen[$0]++' "${ab_partitions_list}"))
 | |
|     if [[ ${#partitions[@]} -eq 0 ]]; then
 | |
|       die "The list of partitions is empty. Can't generate a payload."
 | |
|     fi
 | |
|   else
 | |
|     warn "No ab_partitions.txt found. Using default."
 | |
|   fi
 | |
|   echo "List of A/B partitions for ${partitions_array}: ${partitions[@]}"
 | |
| 
 | |
|   if [[ -n "${partitions_order}" ]]; then
 | |
|     eval "${partitions_order}=(${partitions[@]})"
 | |
|   fi
 | |
| 
 | |
|   # All Brillo updaters support major version 2.
 | |
|   FORCE_MAJOR_VERSION="2"
 | |
| 
 | |
|   if [[ "${partitions_array}" == "SRC_PARTITIONS" ]]; then
 | |
|     # Source image
 | |
|     local ue_config=$(create_tempfile "ue_config.XXXXXX")
 | |
|     CLEANUP_FILES+=("${ue_config}")
 | |
|     if unzip -l "${image}" "META/update_engine_config.txt" > /dev/null; then
 | |
|       extract_file "${image}" "META/update_engine_config.txt" "${ue_config}"
 | |
|     else
 | |
|       warn "No update_engine_config.txt found. Assuming pre-release image, \
 | |
| using payload minor version 2"
 | |
|     fi
 | |
|     # For delta payloads, we use the major and minor version supported by the
 | |
|     # old updater.
 | |
|     FORCE_MINOR_VERSION=$(read_option_uint "${ue_config}" \
 | |
|       "PAYLOAD_MINOR_VERSION" 2)
 | |
|     if [[ -n "${FLAGS_force_minor_version}" ]]; then
 | |
|       FORCE_MINOR_VERSION="${FLAGS_force_minor_version}"
 | |
|     fi
 | |
|     FORCE_MAJOR_VERSION=$(read_option_uint "${ue_config}" \
 | |
|       "PAYLOAD_MAJOR_VERSION" 2)
 | |
| 
 | |
|     # Brillo support for deltas started with minor version 3.
 | |
|     if [[ "${FORCE_MINOR_VERSION}" -le 2 ]]; then
 | |
|       warn "No delta support from minor version ${FORCE_MINOR_VERSION}. \
 | |
| Disabling deltas for this source version."
 | |
|       exit ${EX_UNSUPPORTED_DELTA}
 | |
|     fi
 | |
|   else
 | |
|     # Target image
 | |
|     local postinstall_config=$(create_tempfile "postinstall_config.XXXXXX")
 | |
|     CLEANUP_FILES+=("${postinstall_config}")
 | |
|     if unzip -l "${image}" "META/postinstall_config.txt" > /dev/null; then
 | |
|       extract_file "${image}" "META/postinstall_config.txt" \
 | |
|         "${postinstall_config}"
 | |
|       POSTINSTALL_CONFIG_FILE="${postinstall_config}"
 | |
|     fi
 | |
|     local dynamic_partitions_info=$(create_tempfile "dynamic_partitions_info.XXXXXX")
 | |
|     CLEANUP_FILES+=("${dynamic_partitions_info}")
 | |
|     if unzip -l "${image}" "META/dynamic_partitions_info.txt" > /dev/null; then
 | |
|       extract_file "${image}" "META/dynamic_partitions_info.txt" \
 | |
|         "${dynamic_partitions_info}"
 | |
|       DYNAMIC_PARTITION_INFO_FILE="${dynamic_partitions_info}"
 | |
|     fi
 | |
|     local apex_info=$(create_tempfile "apex_info.XXXXXX")
 | |
|     CLEANUP_FILES+=("${apex_info}")
 | |
|     if unzip -l "${image}" "META/apex_info.pb" > /dev/null; then
 | |
|       extract_file "${image}" "META/apex_info.pb" \
 | |
|         "${apex_info}"
 | |
|       APEX_INFO_FILE="${apex_info}"
 | |
|     fi
 | |
|   fi
 | |
| 
 | |
|   local part
 | |
|   for part in "${partitions[@]}"; do
 | |
|     local part_file=$(create_tempfile "${part}.img.XXXXXX")
 | |
|     local part_map_file=$(create_tempfile "${part}.map.XXXXXX")
 | |
|     CLEANUP_FILES+=("${part_file}" "${part_map_file}")
 | |
|     # Extract partitions in background.
 | |
|     extract_partition_brillo "${image}" "${partitions_array}" "${part}" \
 | |
|         "${part_file}" "${part_map_file}" &
 | |
|     EXTRACT_IMAGE_PIDS+=("$!")
 | |
|     eval "${partitions_array}[\"${part}\"]=\"${part_file}\""
 | |
|     eval "${partitions_array}_MAP[\"${part}\"]=\"${part_map_file}\""
 | |
|   done
 | |
| }
 | |
| 
 | |
| # cleanup_partition_array <partitions_array>
 | |
| #
 | |
| # Remove all empty files in <partitions_array>.
 | |
| cleanup_partition_array() {
 | |
|   local partitions_array="$1"
 | |
|   # Have to use eval to iterate over associative array keys with variable array
 | |
|   # names, we should change it to use nameref once bash 4.3 is available
 | |
|   # everywhere.
 | |
|   for part in $(eval "echo \${!${partitions_array}[@]}"); do
 | |
|     local path="${partitions_array}[$part]"
 | |
|     if [[ ! -s "${!path}" ]]; then
 | |
|       eval "unset ${partitions_array}[${part}]"
 | |
|     fi
 | |
|   done
 | |
| }
 | |
| 
 | |
| extract_payload_images() {
 | |
|   local payload_type=$1
 | |
|   echo "Extracting images for ${payload_type} update."
 | |
| 
 | |
|   if [[ "${payload_type}" == "delta" ]]; then
 | |
|     extract_image "${FLAGS_source_image}" SRC_PARTITIONS
 | |
|   fi
 | |
|   extract_image "${FLAGS_target_image}" DST_PARTITIONS PARTITIONS_ORDER
 | |
|   # Wait for all subprocesses to finish. Not using `wait` since it doesn't die
 | |
|   # on non-zero subprocess exit code. Not using `wait ${EXTRACT_IMAGE_PIDS[@]}`
 | |
|   # as it gives the status of the last process it has waited for.
 | |
|   for pid in ${EXTRACT_IMAGE_PIDS[@]}; do
 | |
|     wait ${pid}
 | |
|   done
 | |
|   cleanup_partition_array SRC_PARTITIONS
 | |
|   cleanup_partition_array SRC_PARTITIONS_MAP
 | |
|   cleanup_partition_array DST_PARTITIONS
 | |
|   cleanup_partition_array DST_PARTITIONS_MAP
 | |
| }
 | |
| 
 | |
| get_payload_type() {
 | |
|   if [[ -z "${FLAGS_source_image}" ]]; then
 | |
|     echo "full"
 | |
|   else
 | |
|     echo "delta"
 | |
|   fi
 | |
| }
 | |
| 
 | |
| validate_generate() {
 | |
|   [[ -n "${FLAGS_payload}" ]] ||
 | |
|     die "You must specify an output filename with --payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_target_image}" ]] ||
 | |
|     die "You must specify a target image with --target_image FILENAME"
 | |
| }
 | |
| 
 | |
| cmd_generate() {
 | |
|   local payload_type=$(get_payload_type)
 | |
|   extract_payload_images ${payload_type}
 | |
| 
 | |
|   echo "Generating ${payload_type} update."
 | |
|   # Common payload args:
 | |
|   GENERATOR_ARGS=( --out_file="${FLAGS_payload}" )
 | |
| 
 | |
|   local part old_partitions="" new_partitions="" partition_names=""
 | |
|   local old_mapfiles="" new_mapfiles=""
 | |
|   for part in "${PARTITIONS_ORDER[@]}"; do
 | |
|     if [[ -n "${partition_names}" ]]; then
 | |
|       partition_names+=":"
 | |
|       new_partitions+=":"
 | |
|       old_partitions+=":"
 | |
|       new_mapfiles+=":"
 | |
|       old_mapfiles+=":"
 | |
|     fi
 | |
|     partition_names+="${part}"
 | |
|     new_partitions+="${DST_PARTITIONS[${part}]}"
 | |
|     if [ "${FLAGS_full_boot}" == "true" ] && [ "${part}" == "boot" ]; then
 | |
|       # Skip boot partition.
 | |
|       old_partitions+=""
 | |
|     else
 | |
|       old_partitions+="${SRC_PARTITIONS[${part}]:-}"
 | |
|     fi
 | |
|     new_mapfiles+="${DST_PARTITIONS_MAP[${part}]:-}"
 | |
|     old_mapfiles+="${SRC_PARTITIONS_MAP[${part}]:-}"
 | |
|   done
 | |
| 
 | |
|   # Target image args:
 | |
|   GENERATOR_ARGS+=(
 | |
|     --partition_names="${partition_names}"
 | |
|     --new_partitions="${new_partitions}"
 | |
|     --new_mapfiles="${new_mapfiles}"
 | |
|   )
 | |
| 
 | |
|   if [[ "${FLAGS_is_partial_update}" == "true" ]]; then
 | |
|     GENERATOR_ARGS+=( --is_partial_update="true" )
 | |
|     # Need at least minor version 7 for partial update, so generate with minor
 | |
|     # version 7 if we don't have a source image. Let the delta_generator to fail
 | |
|     # the other incompatiable minor versions.
 | |
|     if [[ -z "${FORCE_MINOR_VERSION}" ]]; then
 | |
|       FORCE_MINOR_VERSION="7"
 | |
|     fi
 | |
|   fi
 | |
| 
 | |
|   if [[ "${payload_type}" == "delta" ]]; then
 | |
|     # Source image args:
 | |
|     GENERATOR_ARGS+=(
 | |
|       --old_partitions="${old_partitions}"
 | |
|       --old_mapfiles="${old_mapfiles}"
 | |
|     )
 | |
|     if [[ -n "${FLAGS_disable_fec_computation}" ]]; then
 | |
|       GENERATOR_ARGS+=(
 | |
|         --disable_fec_computation="${FLAGS_disable_fec_computation}" )
 | |
|     fi
 | |
|     if [[ -n "${FLAGS_disable_verity_computation}" ]]; then
 | |
|       GENERATOR_ARGS+=(
 | |
|         --disable_verity_computation="${FLAGS_disable_verity_computation}" )
 | |
|     fi
 | |
|     if [[ -n "${FLAGS_compressor_types}" ]]; then
 | |
|       GENERATOR_ARGS+=(
 | |
|         --compressor_types="${FLAGS_compressor_types}" )
 | |
|     fi
 | |
|     if [[ -n "${FLAGS_enable_zucchini}" ]]; then
 | |
|       GENERATOR_ARGS+=(
 | |
|         --enable_zucchini="${FLAGS_enable_zucchini}" )
 | |
|     fi
 | |
|     if [[ -n "${FLAGS_enable_lz4diff}" ]]; then
 | |
|       if [[ "${FLAGS_enable_lz4diff}" == "true" && -z "${FLAGS_liblz4_path}" ]]; then
 | |
|         echo "--liblz4_path is required when enable_lz4diff is set to true."
 | |
|         exit 1
 | |
|       fi
 | |
|       GENERATOR_ARGS+=(
 | |
|         --enable_lz4diff="${FLAGS_enable_lz4diff}" )
 | |
|     fi
 | |
|     if [[ -n "${FLAGS_erofs_compression_param}" ]]; then
 | |
|       GENERATOR_ARGS+=(
 | |
|         --erofs_compression_param="${FLAGS_erofs_compression_param}" )
 | |
|     fi
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FLAGS_enable_vabc_xor}" ]]; then
 | |
|     GENERATOR_ARGS+=(
 | |
|       --enable_vabc_xor="${FLAGS_enable_vabc_xor}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FLAGS_disable_vabc}" ]]; then
 | |
|     GENERATOR_ARGS+=(
 | |
|       --disable_vabc="${FLAGS_disable_vabc}" )
 | |
|   fi
 | |
| 
 | |
|   # minor version is set only for delta or partial payload.
 | |
|   if [[ -n "${FORCE_MINOR_VERSION}" ]]; then
 | |
|     GENERATOR_ARGS+=( --minor_version="${FORCE_MINOR_VERSION}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FORCE_MAJOR_VERSION}" ]]; then
 | |
|     GENERATOR_ARGS+=( --major_version="${FORCE_MAJOR_VERSION}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FLAGS_metadata_size_file}" ]]; then
 | |
|     GENERATOR_ARGS+=( --out_metadata_size_file="${FLAGS_metadata_size_file}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FLAGS_max_timestamp}" ]]; then
 | |
|     GENERATOR_ARGS+=( --max_timestamp="${FLAGS_max_timestamp}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FLAGS_partition_timestamps}" ]]; then
 | |
|     GENERATOR_ARGS+=( --partition_timestamps="${FLAGS_partition_timestamps}" )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${POSTINSTALL_CONFIG_FILE}" ]]; then
 | |
|     GENERATOR_ARGS+=(
 | |
|       --new_postinstall_config_file="${POSTINSTALL_CONFIG_FILE}"
 | |
|     )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "{DYNAMIC_PARTITION_INFO_FILE}" ]]; then
 | |
|     GENERATOR_ARGS+=(
 | |
|       --dynamic_partition_info_file="${DYNAMIC_PARTITION_INFO_FILE}"
 | |
|     )
 | |
|   fi
 | |
|   if [[ -n "{APEX_INFO_FILE}" ]]; then
 | |
|     GENERATOR_ARGS+=(
 | |
|       --apex_info_file="${APEX_INFO_FILE}"
 | |
|     )
 | |
|   fi
 | |
| 
 | |
|   echo "Running delta_generator with args: ${GENERATOR_ARGS[@]}"
 | |
|   if [[ -n "${FLAGS_enable_lz4diff}" ]]; then
 | |
|     LD_PRELOAD=${LD_PRELOAD}:${FLAGS_liblz4_path} "${GENERATOR}" "${GENERATOR_ARGS[@]}"
 | |
|   else
 | |
|     "${GENERATOR}" "${GENERATOR_ARGS[@]}"
 | |
|   fi
 | |
| 
 | |
|   echo "Done generating ${payload_type} update."
 | |
| }
 | |
| 
 | |
| validate_hash() {
 | |
|   [[ -n "${FLAGS_signature_size}" ]] ||
 | |
|     die "You must specify signature size with --signature_size SIZES"
 | |
| 
 | |
|   [[ -n "${FLAGS_unsigned_payload}" ]] ||
 | |
|     die "You must specify the input unsigned payload with \
 | |
| --unsigned_payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_payload_hash_file}" ]] ||
 | |
|     die "You must specify --payload_hash_file FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_metadata_hash_file}" ]] ||
 | |
|     die "You must specify --metadata_hash_file FILENAME"
 | |
| }
 | |
| 
 | |
| cmd_hash() {
 | |
|   "${GENERATOR}" \
 | |
|       --in_file="${FLAGS_unsigned_payload}" \
 | |
|       --signature_size="${FLAGS_signature_size}" \
 | |
|       --out_hash_file="${FLAGS_payload_hash_file}" \
 | |
|       --out_metadata_hash_file="${FLAGS_metadata_hash_file}"
 | |
| 
 | |
|   echo "Done generating hash."
 | |
| }
 | |
| 
 | |
| validate_sign() {
 | |
|   [[ -n "${FLAGS_signature_size}" ]] ||
 | |
|     die "You must specify signature size with --signature_size SIZES"
 | |
| 
 | |
|   [[ -n "${FLAGS_unsigned_payload}" ]] ||
 | |
|     die "You must specify the input unsigned payload with \
 | |
| --unsigned_payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_payload}" ]] ||
 | |
|     die "You must specify the output signed payload with --payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_payload_signature_file}" ]] ||
 | |
|     die "You must specify the payload signature file with \
 | |
| --payload_signature_file SIGNATURES"
 | |
| 
 | |
|   [[ -n "${FLAGS_metadata_signature_file}" ]] ||
 | |
|     die "You must specify the metadata signature file with \
 | |
| --metadata_signature_file SIGNATURES"
 | |
| }
 | |
| 
 | |
| cmd_sign() {
 | |
|   GENERATOR_ARGS=(
 | |
|     --in_file="${FLAGS_unsigned_payload}"
 | |
|     --signature_size="${FLAGS_signature_size}"
 | |
|     --payload_signature_file="${FLAGS_payload_signature_file}"
 | |
|     --metadata_signature_file="${FLAGS_metadata_signature_file}"
 | |
|     --out_file="${FLAGS_payload}"
 | |
|   )
 | |
| 
 | |
|   if [[ -n "${FLAGS_metadata_size_file}" ]]; then
 | |
|     GENERATOR_ARGS+=( --out_metadata_size_file="${FLAGS_metadata_size_file}" )
 | |
|   fi
 | |
| 
 | |
|   "${GENERATOR}" "${GENERATOR_ARGS[@]}"
 | |
|   echo "Done signing payload."
 | |
| }
 | |
| 
 | |
| validate_properties() {
 | |
|   [[ -n "${FLAGS_payload}" ]] ||
 | |
|     die "You must specify the payload file with --payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_properties_file}" ]] ||
 | |
|     die "You must specify a non empty --properties_file FILENAME"
 | |
| }
 | |
| 
 | |
| cmd_properties() {
 | |
|   "${GENERATOR}" \
 | |
|       --in_file="${FLAGS_payload}" \
 | |
|       --properties_file="${FLAGS_properties_file}"
 | |
| }
 | |
| 
 | |
| validate_verify_and_check() {
 | |
|   [[ -n "${FLAGS_payload}" ]] ||
 | |
|     die "Error: you must specify an input filename with --payload FILENAME"
 | |
| 
 | |
|   [[ -n "${FLAGS_target_image}" ]] ||
 | |
|     die "Error: you must specify a target image with --target_image FILENAME"
 | |
| }
 | |
| 
 | |
| cmd_verify() {
 | |
|   local payload_type=$(get_payload_type)
 | |
|   extract_payload_images ${payload_type}
 | |
| 
 | |
|   declare -A TMP_PARTITIONS
 | |
|   for part in "${PARTITIONS_ORDER[@]}"; do
 | |
|     local tmp_part=$(create_tempfile "tmp_part.bin.XXXXXX")
 | |
|     echo "Creating temporary target partition ${tmp_part} for ${part}"
 | |
|     CLEANUP_FILES+=("${tmp_part}")
 | |
|     TMP_PARTITIONS[${part}]=${tmp_part}
 | |
|     local FILESIZE=$(stat -c%s "${DST_PARTITIONS[${part}]}")
 | |
|     echo "Truncating ${TMP_PARTITIONS[${part}]} to ${FILESIZE}"
 | |
|     truncate_file "${TMP_PARTITIONS[${part}]}" "${FILESIZE}"
 | |
|   done
 | |
| 
 | |
|   echo "Verifying ${payload_type} update."
 | |
|   # Common payload args:
 | |
|   GENERATOR_ARGS=( --in_file="${FLAGS_payload}" )
 | |
| 
 | |
|   local part old_partitions="" new_partitions="" partition_names=""
 | |
|   for part in "${PARTITIONS_ORDER[@]}"; do
 | |
|     if [[ -n "${partition_names}" ]]; then
 | |
|       partition_names+=":"
 | |
|       new_partitions+=":"
 | |
|       old_partitions+=":"
 | |
|     fi
 | |
|     partition_names+="${part}"
 | |
|     new_partitions+="${TMP_PARTITIONS[${part}]}"
 | |
|     old_partitions+="${SRC_PARTITIONS[${part}]:-}"
 | |
|   done
 | |
| 
 | |
|   # Target image args:
 | |
|   GENERATOR_ARGS+=(
 | |
|     --partition_names="${partition_names}"
 | |
|     --new_partitions="${new_partitions}"
 | |
|   )
 | |
| 
 | |
|   if [[ "${payload_type}" == "delta" ]]; then
 | |
|     # Source image args:
 | |
|     GENERATOR_ARGS+=(
 | |
|       --old_partitions="${old_partitions}"
 | |
|     )
 | |
|   fi
 | |
| 
 | |
|   if [[ -n "${FORCE_MAJOR_VERSION}" ]]; then
 | |
|     GENERATOR_ARGS+=( --major_version="${FORCE_MAJOR_VERSION}" )
 | |
|   fi
 | |
| 
 | |
|   echo "Running delta_generator to verify ${payload_type} payload with args: \
 | |
| ${GENERATOR_ARGS[@]}"
 | |
|   "${GENERATOR}" "${GENERATOR_ARGS[@]}" || true
 | |
| 
 | |
|   echo "Done applying ${payload_type} update."
 | |
|   echo "Checking the newly generated partitions against the target partitions"
 | |
|   local need_pause=false
 | |
|   for part in "${PARTITIONS_ORDER[@]}"; do
 | |
|     local not_str=""
 | |
|     if ! cmp "${TMP_PARTITIONS[${part}]}" "${DST_PARTITIONS[${part}]}"; then
 | |
|       not_str="in"
 | |
|       need_pause=true
 | |
|     fi
 | |
|     echo "The new partition (${part}) is ${not_str}valid."
 | |
|   done
 | |
|   # All images will be cleaned up when script exits, pause here to give a chance
 | |
|   # to inspect the images.
 | |
|   if [[ "$need_pause" == true ]]; then
 | |
|     read -n1 -r -s -p "Paused to investigate invalid partitions, \
 | |
| press any key to exit."
 | |
|   fi
 | |
| }
 | |
| 
 | |
| cmd_check() {
 | |
|   local payload_type=$(get_payload_type)
 | |
|   extract_payload_images ${payload_type}
 | |
| 
 | |
|   local part dst_partitions="" src_partitions=""
 | |
|   for part in "${PARTITIONS_ORDER[@]}"; do
 | |
|     if [[ -n "${dst_partitions}" ]]; then
 | |
|       dst_partitions+=" "
 | |
|       src_partitions+=" "
 | |
|     fi
 | |
|     dst_partitions+="${DST_PARTITIONS[${part}]}"
 | |
|     src_partitions+="${SRC_PARTITIONS[${part}]:-}"
 | |
|   done
 | |
| 
 | |
|   # Common payload args:
 | |
|   PAYCHECK_ARGS=( "${FLAGS_payload}" --type ${payload_type} \
 | |
|     --part_names ${PARTITIONS_ORDER[@]} \
 | |
|     --dst_part_paths ${dst_partitions} )
 | |
| 
 | |
|   if [[ ! -z "${SRC_PARTITIONS[@]}" ]]; then
 | |
|     PAYCHECK_ARGS+=( --src_part_paths ${src_partitions} )
 | |
|   fi
 | |
| 
 | |
|   echo "Checking ${payload_type} update."
 | |
|   check_update_payload ${PAYCHECK_ARGS[@]} --check
 | |
| }
 | |
| 
 | |
| # Check that the real generator exists:
 | |
| [[ -x "${GENERATOR}" ]] || GENERATOR="$(which delta_generator || true)"
 | |
| [[ -x "${GENERATOR}" ]] || die "can't find delta_generator"
 | |
| 
 | |
| case "$COMMAND" in
 | |
|   generate) validate_generate
 | |
|             cmd_generate
 | |
|             ;;
 | |
|   hash) validate_hash
 | |
|         cmd_hash
 | |
|         ;;
 | |
|   sign) validate_sign
 | |
|         cmd_sign
 | |
|         ;;
 | |
|   properties) validate_properties
 | |
|               cmd_properties
 | |
|               ;;
 | |
|   verify) validate_verify_and_check
 | |
|           cmd_verify
 | |
|           ;;
 | |
|   check) validate_verify_and_check
 | |
|          cmd_check
 | |
|          ;;
 | |
| esac
 |