Bump buildroot to 2019.02
This commit is contained in:
@@ -5,28 +5,31 @@ set -e
|
||||
|
||||
# Download helper for bzr, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../bzr [-q] OUT_FILE REPO_URL REV BASENAME
|
||||
# Options:
|
||||
# -q Be quiet
|
||||
# -o FILE Generate archive in FILE.
|
||||
# -u URI Clone from repository at URI.
|
||||
# -c CSET Use changeset (or revision) CSET.
|
||||
# -n NAME Use basename NAME.
|
||||
#
|
||||
# Environment:
|
||||
# BZR : the bzr command to call
|
||||
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q;;
|
||||
o) output="${OPTARG}";;
|
||||
u) uri="${OPTARG}";;
|
||||
c) cset="${OPTARG}";;
|
||||
n) basename="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
repo="${2}"
|
||||
rev="${3}"
|
||||
basename="${4}"
|
||||
|
||||
shift 4 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -51,5 +54,5 @@ if [ ${bzr_version} -ge ${bzr_min_version} ]; then
|
||||
fi
|
||||
|
||||
_bzr export ${verbose} --root="'${basename}/'" --format=tgz \
|
||||
${timestamp_opt} - "${@}" "'${repo}'" -r "'${rev}'" \
|
||||
${timestamp_opt} - "${@}" "'${uri}'" -r "'${cset}'" \
|
||||
>"${output}"
|
||||
|
||||
@@ -5,33 +5,41 @@ set -e
|
||||
|
||||
# Download helper for cvs, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../cvs [-q] OUT_FILE CVS_URL REV PKG_NAME BASENAME
|
||||
# Options:
|
||||
# -q Be quiet
|
||||
# -o FILE Generate archive in FILE.
|
||||
# -u URI Checkout from repository at URI.
|
||||
# -c REV Use revision REV.
|
||||
# -N RAWNAME Use rawname (aka module name) RAWNAME.
|
||||
# -n NAME Use basename NAME.
|
||||
#
|
||||
# Environment:
|
||||
# CVS : the cvs command to call
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-Q;;
|
||||
o) output="${OPTARG}";;
|
||||
u) uri="${OPTARG#*://}";;
|
||||
c) rev="${OPTARG}";;
|
||||
N) rawname="${OPTARG}";;
|
||||
n) basename="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
repo="${2}"
|
||||
rev="${3}"
|
||||
rawname="${4}"
|
||||
basename="${5}"
|
||||
|
||||
shift 5 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
# being expanded a second time (in case there are spaces in them).
|
||||
# If the CVS server is deadlocked, the client will never return (cfr.
|
||||
# http://autobuild.buildroot.net/results/23d/23d1034b33d0354de15de2ec4a8ccd0603e8db78/build-end.log
|
||||
# ). Since nobody sane will put large code bases in CVS, a timeout of
|
||||
# 10 minutes should do the trick.
|
||||
_cvs() {
|
||||
eval ${CVS} "${@}"
|
||||
eval timeout 10m ${CVS} "${@}"
|
||||
}
|
||||
|
||||
if [[ ${rev} =~ ^[0-9] ]]; then
|
||||
@@ -42,14 +50,14 @@ else
|
||||
select="-r"
|
||||
fi
|
||||
|
||||
# The absence of an initial : on ${repo} means access method undefined
|
||||
if [[ ! "${repo}" =~ ^: ]]; then
|
||||
# The absence of an initial : on ${uri} means access method undefined
|
||||
if [[ ! "${uri}" =~ ^: ]]; then
|
||||
# defaults to anonymous pserver
|
||||
repo=":pserver:anonymous@${repo}"
|
||||
uri=":pserver:anonymous@${uri}"
|
||||
fi
|
||||
|
||||
export TZ=UTC
|
||||
_cvs ${verbose} -z3 -d"'${repo}'" \
|
||||
_cvs ${verbose} -z3 -d"'${uri}'" \
|
||||
co "${@}" -d "'${basename}'" ${select} "'${rev}'" -P "'${rawname}'"
|
||||
|
||||
tar czf "${output}" "${basename}"
|
||||
|
||||
@@ -4,8 +4,6 @@
|
||||
# Its role is to ensure atomicity when saving downloaded files
|
||||
# back to BR2_DL_DIR, and not clutter BR2_DL_DIR with partial,
|
||||
# failed downloads.
|
||||
#
|
||||
# Call it with -h to see some help.
|
||||
|
||||
# To avoid cluttering BR2_DL_DIR, we download to a trashable
|
||||
# location, namely in $(BUILD_DIR).
|
||||
@@ -19,33 +17,51 @@
|
||||
# We want to catch any unexpected failure, and exit immediately.
|
||||
set -e
|
||||
|
||||
export BR_BACKEND_DL_GETOPTS=":hc:d:o:n:N:H:ru:qf:e"
|
||||
|
||||
main() {
|
||||
local OPT OPTARG
|
||||
local backend output hfile recurse quiet
|
||||
local backend output hfile recurse quiet rc
|
||||
local -a uris
|
||||
|
||||
# Parse our options; anything after '--' is for the backend
|
||||
while getopts :hb:o:H:rq OPT; do
|
||||
while getopts ":c:d:D:o:n:N:H:rf:u:q" OPT; do
|
||||
case "${OPT}" in
|
||||
h) help; exit 0;;
|
||||
b) backend="${OPTARG}";;
|
||||
c) cset="${OPTARG}";;
|
||||
d) dl_dir="${OPTARG}";;
|
||||
D) old_dl_dir="${OPTARG}";;
|
||||
o) output="${OPTARG}";;
|
||||
n) raw_base_name="${OPTARG}";;
|
||||
N) base_name="${OPTARG}";;
|
||||
H) hfile="${OPTARG}";;
|
||||
r) recurse="-r";;
|
||||
f) filename="${OPTARG}";;
|
||||
u) uris+=( "${OPTARG}" );;
|
||||
q) quiet="-q";;
|
||||
:) error "option '%s' expects a mandatory argument\n" "${OPTARG}";;
|
||||
\?) error "unknown option '%s'\n" "${OPTARG}";;
|
||||
esac
|
||||
done
|
||||
|
||||
# Forget our options, and keep only those for the backend
|
||||
shift $((OPTIND-1))
|
||||
|
||||
if [ -z "${backend}" ]; then
|
||||
error "no backend specified, use -b\n"
|
||||
fi
|
||||
if [ -z "${output}" ]; then
|
||||
error "no output specified, use -o\n"
|
||||
fi
|
||||
|
||||
# Legacy handling: check if the file already exists in the global
|
||||
# download directory. If it does, hard-link it. If it turns out it
|
||||
# was an incorrect download, we'd still check it below anyway.
|
||||
# If we can neither link nor copy, fallback to doing a download.
|
||||
# NOTE! This is not atomic, is subject to TOCTTOU, but the whole
|
||||
# dl-wrapper runs under an flock, so we're safe.
|
||||
if [ ! -e "${output}" -a -e "${old_dl_dir}/${filename}" ]; then
|
||||
ln "${old_dl_dir}/${filename}" "${output}" || \
|
||||
cp "${old_dl_dir}/${filename}" "${output}" || \
|
||||
true
|
||||
fi
|
||||
|
||||
# If the output file already exists and:
|
||||
# - there's no .hash file: do not download it again and exit promptly
|
||||
# - matches all its hashes: do not download it again and exit promptly
|
||||
@@ -64,36 +80,85 @@ main() {
|
||||
warn "Re-downloading '%s'...\n" "${output##*/}"
|
||||
fi
|
||||
|
||||
# tmpd is a temporary directory in which backends may store intermediate
|
||||
# by-products of the download.
|
||||
# tmpf is the file in which the backends should put the downloaded content.
|
||||
# tmpd is located in $(BUILD_DIR), so as not to clutter the (precious)
|
||||
# $(BR2_DL_DIR)
|
||||
# We let the backends create tmpf, so they are able to set whatever
|
||||
# permission bits they want (although we're only really interested in
|
||||
# the executable bit.)
|
||||
tmpd="$(mktemp -d "${BUILD_DIR}/.${output##*/}.XXXXXX")"
|
||||
tmpf="${tmpd}/output"
|
||||
# Look through all the uris that we were given to download the package
|
||||
# source
|
||||
download_and_check=0
|
||||
rc=1
|
||||
for uri in "${uris[@]}"; do
|
||||
backend_urlencode="${uri%%+*}"
|
||||
backend="${backend_urlencode%|*}"
|
||||
case "${backend}" in
|
||||
git|svn|cvs|bzr|file|scp|hg) ;;
|
||||
*) backend="wget" ;;
|
||||
esac
|
||||
uri=${uri#*+}
|
||||
|
||||
# Helpers expect to run in a directory that is *really* trashable, so
|
||||
# they are free to create whatever files and/or sub-dirs they might need.
|
||||
# Doing the 'cd' here rather than in all backends is easier.
|
||||
cd "${tmpd}"
|
||||
urlencode=${backend_urlencode#*|}
|
||||
# urlencode must be "urlencode"
|
||||
[ "${urlencode}" != "urlencode" ] && urlencode=""
|
||||
|
||||
# If the backend fails, we can just remove the temporary directory to
|
||||
# remove all the cruft it may have left behind. Then we just exit in
|
||||
# error too.
|
||||
if ! "${OLDPWD}/support/download/${backend}" ${quiet} ${recurse} "${tmpf}" "${@}"; then
|
||||
rm -rf "${tmpd}"
|
||||
exit 1
|
||||
fi
|
||||
# tmpd is a temporary directory in which backends may store
|
||||
# intermediate by-products of the download.
|
||||
# tmpf is the file in which the backends should put the downloaded
|
||||
# content.
|
||||
# tmpd is located in $(BUILD_DIR), so as not to clutter the (precious)
|
||||
# $(BR2_DL_DIR)
|
||||
# We let the backends create tmpf, so they are able to set whatever
|
||||
# permission bits they want (although we're only really interested in
|
||||
# the executable bit.)
|
||||
tmpd="$(mktemp -d "${BUILD_DIR}/.${output##*/}.XXXXXX")"
|
||||
tmpf="${tmpd}/output"
|
||||
|
||||
# cd back to free the temp-dir, so we can remove it later
|
||||
cd "${OLDPWD}"
|
||||
# Helpers expect to run in a directory that is *really* trashable, so
|
||||
# they are free to create whatever files and/or sub-dirs they might need.
|
||||
# Doing the 'cd' here rather than in all backends is easier.
|
||||
cd "${tmpd}"
|
||||
|
||||
# Check if the downloaded file is sane, and matches the stored hashes
|
||||
# for that file
|
||||
if ! support/download/check-hash ${quiet} "${hfile}" "${tmpf}" "${output##*/}"; then
|
||||
# If the backend fails, we can just remove the content of the temporary
|
||||
# directory to remove all the cruft it may have left behind, and try
|
||||
# the next URI until it succeeds. Once out of URI to try, we need to
|
||||
# cleanup and exit.
|
||||
if ! "${OLDPWD}/support/download/${backend}" \
|
||||
$([ -n "${urlencode}" ] && printf %s '-e') \
|
||||
-c "${cset}" \
|
||||
-d "${dl_dir}" \
|
||||
-n "${raw_base_name}" \
|
||||
-N "${base_name}" \
|
||||
-f "${filename}" \
|
||||
-u "${uri}" \
|
||||
-o "${tmpf}" \
|
||||
${quiet} ${recurse} -- "${@}"
|
||||
then
|
||||
# cd back to keep path coherence
|
||||
cd "${OLDPWD}"
|
||||
rm -rf "${tmpd}"
|
||||
continue
|
||||
fi
|
||||
|
||||
# cd back to free the temp-dir, so we can remove it later
|
||||
cd "${OLDPWD}"
|
||||
|
||||
# Check if the downloaded file is sane, and matches the stored hashes
|
||||
# for that file
|
||||
if support/download/check-hash ${quiet} "${hfile}" "${tmpf}" "${output##*/}"; then
|
||||
rc=0
|
||||
else
|
||||
if [ ${?} -ne 3 ]; then
|
||||
rm -rf "${tmpd}"
|
||||
continue
|
||||
fi
|
||||
|
||||
# the hash file exists and there was no hash to check the file
|
||||
# against
|
||||
rc=1
|
||||
fi
|
||||
download_and_check=1
|
||||
break
|
||||
done
|
||||
|
||||
# We tried every URI possible, none seems to work or to check against the
|
||||
# available hash. *ABORT MISSION*
|
||||
if [ "${download_and_check}" -eq 0 ]; then
|
||||
rm -rf "${tmpd}"
|
||||
exit 1
|
||||
fi
|
||||
@@ -141,51 +206,8 @@ main() {
|
||||
rm -f "${tmp_output}"
|
||||
exit 1
|
||||
fi
|
||||
}
|
||||
|
||||
help() {
|
||||
cat <<_EOF_
|
||||
NAME
|
||||
${my_name} - download wrapper for Buildroot
|
||||
|
||||
SYNOPSIS
|
||||
${my_name} [OPTION]... -- [BACKEND OPTION]...
|
||||
|
||||
DESCRIPTION
|
||||
Wrapper script around different download mechanisms. Ensures that
|
||||
concurrent downloads do not conflict, that partial downloads are
|
||||
properly evicted without leaving temporary files, and that access
|
||||
rights are maintained.
|
||||
|
||||
-h This help text.
|
||||
|
||||
-b BACKEND
|
||||
Wrap the specified BACKEND. Known backends are:
|
||||
bzr Bazaar
|
||||
cp Local files
|
||||
cvs Concurrent Versions System
|
||||
git Git
|
||||
hg Mercurial
|
||||
scp Secure copy
|
||||
svn Subversion
|
||||
wget HTTP download
|
||||
|
||||
-o FILE
|
||||
Store the downloaded archive in FILE.
|
||||
|
||||
-H FILE
|
||||
Use FILE to read hashes from, and check them against the downloaded
|
||||
archive.
|
||||
|
||||
Exit status:
|
||||
0 if OK
|
||||
!0 in case of error
|
||||
|
||||
ENVIRONMENT
|
||||
|
||||
BUILD_DIR
|
||||
The path to Buildroot's build dir
|
||||
_EOF_
|
||||
return ${rc}
|
||||
}
|
||||
|
||||
trace() { local msg="${1}"; shift; printf "%s: ${msg}" "${my_name}" "${@}"; }
|
||||
|
||||
@@ -5,8 +5,11 @@ set -e
|
||||
|
||||
# Download helper for cp, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../cp [-q] OUT_FILE SRC_FILE
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -o FILE Copy to file FILE.
|
||||
# -f FILE Copy from basename file FILE.
|
||||
# -u DIR Copy from FILE in DIR.
|
||||
#
|
||||
# Environment:
|
||||
# LOCALFILES: the cp command to call
|
||||
@@ -17,18 +20,18 @@ set -e
|
||||
# Make 'cp' verbose by default, so it behaves a bit like the others.
|
||||
verbose=-v
|
||||
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=;;
|
||||
o) output="${OPTARG}";;
|
||||
f) file="${OPTARG}";;
|
||||
u) dir="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
source="${2}"
|
||||
|
||||
shift 2 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -36,4 +39,4 @@ _localfiles() {
|
||||
eval ${LOCALFILES} "${@}"
|
||||
}
|
||||
|
||||
_localfiles ${verbose} "${@}""'${source}'" "'${output}'"
|
||||
_localfiles ${verbose} "'${dir##file://}/${file}'" "'${output}'"
|
||||
@@ -1,65 +1,119 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
# We want to catch any unexpected failure, and exit immediately
|
||||
set -e
|
||||
set -E
|
||||
|
||||
# Download helper for git, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../git [-q] [-r] OUT_FILE REPO_URL CSET BASENAME
|
||||
#
|
||||
# -q Be quiet.
|
||||
# -r Clone and archive sub-modules.
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -r Clone and archive sub-modules.
|
||||
# -o FILE Generate archive in FILE.
|
||||
# -u URI Clone from repository at URI.
|
||||
# -c CSET Use changeset CSET.
|
||||
# -n NAME Use basename NAME.
|
||||
#
|
||||
# Environment:
|
||||
# GIT : the git command to call
|
||||
|
||||
# Save our path and options in case we need to call ourselves again
|
||||
myname="${0}"
|
||||
declare -a OPTS=("${@}")
|
||||
|
||||
# This function is called when an error occurs. Its job is to attempt a
|
||||
# clone from scratch (only once!) in case the git tree is borked, or in
|
||||
# case an unexpected and unsupported situation arises with submodules
|
||||
# or uncommitted stuff (e.g. if the user manually mucked around in the
|
||||
# git cache).
|
||||
_on_error() {
|
||||
local ret=${?}
|
||||
|
||||
printf "Detected a corrupted git cache.\n" >&2
|
||||
if ${BR_GIT_BACKEND_FIRST_FAULT:-false}; then
|
||||
printf "This is the second time in a row; bailing out\n" >&2
|
||||
exit ${ret}
|
||||
fi
|
||||
export BR_GIT_BACKEND_FIRST_FAULT=true
|
||||
|
||||
printf "Removing it and starting afresh.\n" >&2
|
||||
|
||||
popd >/dev/null
|
||||
rm -rf "${git_cache}"
|
||||
|
||||
exec "${myname}" "${OPTS[@]}" || exit ${ret}
|
||||
}
|
||||
|
||||
verbose=
|
||||
recurse=0
|
||||
while getopts :qr OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q; exec >/dev/null;;
|
||||
r) recurse=1;;
|
||||
o) output="${OPTARG}";;
|
||||
u) uri="${OPTARG}";;
|
||||
c) cset="${OPTARG}";;
|
||||
d) dl_dir="${OPTARG}";;
|
||||
n) basename="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
repo="${2}"
|
||||
cset="${3}"
|
||||
basename="${4}"
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
shift 4 # Get rid of our options
|
||||
# Create and cd into the directory that will contain the local git cache
|
||||
git_cache="${dl_dir}/git"
|
||||
mkdir -p "${git_cache}"
|
||||
pushd "${git_cache}" >/dev/null
|
||||
|
||||
# Any error now should try to recover
|
||||
trap _on_error ERR
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
_git() {
|
||||
eval ${GIT} "${@}"
|
||||
eval GIT_DIR="${git_cache}/.git" ${GIT} "${@}"
|
||||
}
|
||||
|
||||
# Try a shallow clone, since it is faster than a full clone - but that only
|
||||
# works if the version is a ref (tag or branch). Before trying to do a shallow
|
||||
# clone we check if ${cset} is in the list provided by git ls-remote. If not
|
||||
# we fall back on a full clone.
|
||||
# Create a warning file, that the user should not use the git cache.
|
||||
# It's ours. Our precious.
|
||||
cat <<-_EOF_ >"${dl_dir}/git.readme"
|
||||
IMPORTANT NOTE!
|
||||
|
||||
The git tree located in this directory is for the exclusive use
|
||||
by Buildroot, which uses it as a local cache to reduce bandwidth
|
||||
usage.
|
||||
|
||||
Buildroot *will* trash any changes in that tree whenever it needs
|
||||
to use it. Buildroot may even remove it in case it detects the
|
||||
repository may have been damaged or corrupted.
|
||||
|
||||
Do *not* work in that directory; your changes will eventually get
|
||||
lost. Do *not* even use it as a remote, or as the source for new
|
||||
worktrees; your commits will eventually get lost.
|
||||
_EOF_
|
||||
|
||||
# Initialise a repository in the git cache. If the repository already
|
||||
# existed, this is a noop, unless the repository was broken, in which
|
||||
# case this magically restores it to working conditions. In the latter
|
||||
# case, we might be missing blobs, but that's not a problem: we'll
|
||||
# fetch what we need later anyway.
|
||||
#
|
||||
# Messages for the type of clone used are provided to ease debugging in case of
|
||||
# problems
|
||||
git_done=0
|
||||
if [ -n "$(_git ls-remote "'${repo}'" "'${cset}'" 2>&1)" ]; then
|
||||
printf "Doing shallow clone\n"
|
||||
if _git clone ${verbose} "${@}" --depth 1 -b "'${cset}'" "'${repo}'" "'${basename}'"; then
|
||||
git_done=1
|
||||
else
|
||||
printf "Shallow clone failed, falling back to doing a full clone\n"
|
||||
fi
|
||||
fi
|
||||
if [ ${git_done} -eq 0 ]; then
|
||||
printf "Doing full clone\n"
|
||||
_git clone ${verbose} "${@}" "'${repo}'" "'${basename}'"
|
||||
# We can still go through the wrapper, because 'init' does not use the
|
||||
# path pointed to by GIT_DIR, but really uses the directory passed as
|
||||
# argument.
|
||||
_git init .
|
||||
|
||||
# Ensure the repo has an origin (in case a previous run was killed).
|
||||
if ! _git remote |grep -q -E '^origin$'; then
|
||||
_git remote add origin "'${uri}'"
|
||||
fi
|
||||
|
||||
pushd "${basename}" >/dev/null
|
||||
_git remote set-url origin "'${uri}'"
|
||||
|
||||
printf "Fetching all references\n"
|
||||
_git fetch origin
|
||||
_git fetch origin -t
|
||||
|
||||
# Try to get the special refs exposed by some forges (pull-requests for
|
||||
# github, changes for gerrit...). There is no easy way to know whether
|
||||
@@ -72,9 +126,48 @@ if ! _git fetch origin "'${cset}:${cset}'" >/dev/null 2>&1; then
|
||||
printf "Could not fetch special ref '%s'; assuming it is not special.\n" "${cset}"
|
||||
fi
|
||||
|
||||
# Check that the changeset does exist. If it does not, re-cloning from
|
||||
# scratch won't help, so we don't want to trash the repository for a
|
||||
# missing commit. We just exit without going through the ERR trap.
|
||||
if ! _git rev-parse --quiet --verify "'${cset}^{commit}'" >/dev/null 2>&1; then
|
||||
printf "Commit '%s' does not exist in this repository\n." "${cset}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# The new cset we want to checkout might have different submodules, or
|
||||
# have sub-dirs converted to/from a submodule. So we would need to
|
||||
# deregister _current_ submodules before we checkout.
|
||||
#
|
||||
# Using "git submodule deinit --all" would remove all the files for
|
||||
# all submodules, including the corresponding .git files or directories.
|
||||
# However, it was only introduced with git-1.8.3, which is too recent
|
||||
# for some enterprise-grade distros.
|
||||
#
|
||||
# So, we fall-back to just removing all submodules directories. We do
|
||||
# not need to be recursive, as removing a submodule will de-facto remove
|
||||
# its own submodules.
|
||||
#
|
||||
# For recent git versions, the repository for submodules is stored
|
||||
# inside the repository of the super repository, so the following will
|
||||
# only remove the working copies of submodules, effectively caching the
|
||||
# submodules.
|
||||
#
|
||||
# For older versions however, the repository is stored in the .git/ of
|
||||
# the submodule directory, so the following will effectively remove the
|
||||
# the working copy as well as the repository, which means submodules
|
||||
# will not be cached for older versions.
|
||||
#
|
||||
cmd='printf "Deregistering submodule \"%s\"\n" "${path}" && cd .. && rm -rf "${path##*/}"'
|
||||
_git submodule --quiet foreach "'${cmd}'"
|
||||
|
||||
# Checkout the required changeset, so that we can update the required
|
||||
# submodules.
|
||||
_git checkout -q "'${cset}'"
|
||||
_git checkout -f -q "'${cset}'"
|
||||
|
||||
# Get rid of now-untracked directories (in case a git operation was
|
||||
# interrupted in a previous run, or to get rid of empty directories
|
||||
# that were parents of submodules removed above).
|
||||
_git clean -ffdx
|
||||
|
||||
# Get date of commit to generate a reproducible archive.
|
||||
# %cD is RFC2822, so it's fully qualified, with TZ and all.
|
||||
@@ -85,20 +178,24 @@ if [ ${recurse} -eq 1 ]; then
|
||||
_git submodule update --init --recursive
|
||||
fi
|
||||
|
||||
# We do not want the .git dir; we keep other .git files, in case they
|
||||
# are the only files in their directory.
|
||||
# Generate the archive, sort with the C locale so that it is reproducible.
|
||||
# We do not want the .git dir; we keep other .git files, in case they are the
|
||||
# only files in their directory.
|
||||
# The .git dir would generate non reproducible tarballs as it depends on
|
||||
# the state of the remote server. It also would generate large tarballs
|
||||
# (gigabytes for some linux trees) when a full clone took place.
|
||||
rm -rf .git
|
||||
find . -not -type d \
|
||||
-and -not -path "./.git/*" >"${output}.list"
|
||||
LC_ALL=C sort <"${output}.list" >"${output}.list.sorted"
|
||||
|
||||
popd >/dev/null
|
||||
|
||||
# Generate the archive, sort with the C locale so that it is reproducible
|
||||
find "${basename}" -not -type d >"${basename}.list"
|
||||
LC_ALL=C sort <"${basename}.list" >"${basename}.list.sorted"
|
||||
# Create GNU-format tarballs, since that's the format of the tarballs on
|
||||
# sources.buildroot.org and used in the *.hash files
|
||||
tar cf - --numeric-owner --owner=0 --group=0 --mtime="${date}" --format=gnu \
|
||||
-T "${basename}.list.sorted" >"${output}.tar"
|
||||
tar cf - --transform="s#^\./#${basename}/#" \
|
||||
--numeric-owner --owner=0 --group=0 --mtime="${date}" --format=gnu \
|
||||
-T "${output}.list.sorted" >"${output}.tar"
|
||||
gzip -6 -n <"${output}.tar" >"${output}"
|
||||
|
||||
rm -f "${output}.list"
|
||||
rm -f "${output}.list.sorted"
|
||||
|
||||
popd >/dev/null
|
||||
|
||||
@@ -5,27 +5,30 @@ set -e
|
||||
|
||||
# Download helper for hg, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../hg [-q] OUT_FILE REPO_URL CSET BASENAME
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -o FILE Generate archive in FILE.
|
||||
# -u URI Clone from repository at URI.
|
||||
# -c CSET Use changeset (or revision) CSET.
|
||||
# -n NAME Use basename NAME.
|
||||
#
|
||||
# Environment:
|
||||
# HG : the hg command to call
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q;;
|
||||
o) output="${OPTARG}";;
|
||||
u) uri="${OPTARG}";;
|
||||
c) cset="${OPTARG}";;
|
||||
n) basename="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
repo="${2}"
|
||||
cset="${3}"
|
||||
basename="${4}"
|
||||
|
||||
shift 4 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -33,7 +36,7 @@ _hg() {
|
||||
eval ${HG} "${@}"
|
||||
}
|
||||
|
||||
_hg clone ${verbose} "${@}" --noupdate "'${repo}'" "'${basename}'"
|
||||
_hg clone ${verbose} "${@}" --noupdate "'${uri}'" "'${basename}'"
|
||||
|
||||
_hg archive ${verbose} --repository "'${basename}'" --type tgz \
|
||||
--prefix "'${basename}'" --rev "'${cset}'" \
|
||||
|
||||
@@ -5,25 +5,28 @@ set -e
|
||||
|
||||
# Download helper for scp, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../scp [-q] OUT_FILE SRC_URL
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -o FILE Copy to local file FILE.
|
||||
# -f FILE Copy from remote file FILE.
|
||||
# -u URI Download file at URI.
|
||||
#
|
||||
# Environment:
|
||||
# SCP : the scp command to call
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q;;
|
||||
o) output="${OPTARG}";;
|
||||
f) filename="${OPTARG}";;
|
||||
u) uri="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
url="${2}"
|
||||
|
||||
shift 2 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -31,4 +34,7 @@ _scp() {
|
||||
eval ${SCP} "${@}"
|
||||
}
|
||||
|
||||
_scp ${verbose} "${@}" "'${url}'" "'${output}'"
|
||||
# Remove any scheme prefix
|
||||
uri="${uri##scp://}"
|
||||
|
||||
_scp ${verbose} "${@}" "'${uri}/${filename}'" "'${output}'"
|
||||
|
||||
@@ -5,27 +5,30 @@ set -e
|
||||
|
||||
# Download helper for svn, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../svn [-q] OUT_FILE REPO_URL REV BASNAME
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -o FILE Generate archive in FILE.
|
||||
# -u URI Checkout from repository at URI.
|
||||
# -c REV Use revision REV.
|
||||
# -n NAME Use basename NAME.
|
||||
#
|
||||
# Environment:
|
||||
# SVN : the svn command to call
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q;;
|
||||
o) output="${OPTARG}";;
|
||||
u) uri="${OPTARG}";;
|
||||
c) rev="${OPTARG}";;
|
||||
n) basename="${OPTARG}";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
repo="${2}"
|
||||
rev="${3}"
|
||||
basename="${4}"
|
||||
|
||||
shift 4 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -33,6 +36,6 @@ _svn() {
|
||||
eval ${SVN} "${@}"
|
||||
}
|
||||
|
||||
_svn export ${verbose} "${@}" "'${repo}@${rev}'" "'${basename}'"
|
||||
_svn export ${verbose} "${@}" "'${uri}@${rev}'" "'${basename}'"
|
||||
|
||||
tar czf "${output}" "${basename}"
|
||||
|
||||
@@ -5,25 +5,30 @@ set -e
|
||||
|
||||
# Download helper for wget, to be called from the download wrapper script
|
||||
#
|
||||
# Call it as:
|
||||
# .../wget [-q] OUT_FILE URL
|
||||
# Options:
|
||||
# -q Be quiet.
|
||||
# -o FILE Save into file FILE.
|
||||
# -f FILENAME The filename of the tarball to get at URL
|
||||
# -u URL Download file at URL.
|
||||
# -e ENCODE Tell wget to urlencode the filename passed to it
|
||||
#
|
||||
# Environment:
|
||||
# WGET : the wget command to call
|
||||
|
||||
verbose=
|
||||
while getopts :q OPT; do
|
||||
while getopts "${BR_BACKEND_DL_GETOPTS}" OPT; do
|
||||
case "${OPT}" in
|
||||
q) verbose=-q;;
|
||||
o) output="${OPTARG}";;
|
||||
f) filename="${OPTARG}";;
|
||||
u) url="${OPTARG}";;
|
||||
e) encode="-e";;
|
||||
:) printf "option '%s' expects a mandatory argument\n" "${OPTARG}"; exit 1;;
|
||||
\?) printf "unknown option '%s'\n" "${OPTARG}" >&2; exit 1;;
|
||||
esac
|
||||
done
|
||||
shift $((OPTIND-1))
|
||||
|
||||
output="${1}"
|
||||
url="${2}"
|
||||
|
||||
shift 2 # Get rid of our options
|
||||
shift $((OPTIND-1)) # Get rid of our options
|
||||
|
||||
# Caller needs to single-quote its arguments to prevent them from
|
||||
# being expanded a second time (in case there are spaces in them)
|
||||
@@ -31,4 +36,8 @@ _wget() {
|
||||
eval ${WGET} "${@}"
|
||||
}
|
||||
|
||||
_wget ${verbose} "${@}" -O "'${output}'" "'${url}'"
|
||||
# Replace every '?' with '%3F' in the filename; only for the PRIMARY and BACKUP
|
||||
# mirror
|
||||
[ -n "${encode}" ] && filename=${filename//\?/%3F}
|
||||
|
||||
_wget ${verbose} "${@}" -O "'${output}'" "'${url}/${filename}'"
|
||||
|
||||
Reference in New Issue
Block a user