* ci: add Dockerfile and action to build celery worker image * ci: build celery worker on push to jennifer/celery branch * ci: also build celery worker for main branch * ci: Add comment to celery Dockerfile * chore: first stab at a celery/rabbitmq docker-compose * feat: add celery configuration and test task / endpoint * chore: run mq/celery containers for dev work * chore: point to ghcr.io image for celery worker * refactor: move XML parsing duties into XMLDraft Move some PlaintextDraft methods into the Draft base class and implement for the XMLDraft class. Use xml2rfc code from ietf.submit as a model for the parsing. This leaves some mismatch between the PlaintextDraft and the Draft class spec for the get_author_list() method to be resolved. * feat: add api_upload endpoint and beginnings of async processing This adds an api_upload() that behaves analogously to the api_submit() endpoint. Celery tasks to handle asynchronous processing are added but are not yet functional enough to be useful. * perf: index Submission table on submission_date This substantially speeds up submission rate threshold checks. * feat: remove existing files when accepting a new submission After checking that a submission is not in progress, remove any files in staging that have the same name/rev with any extension. This should guard against stale files confusing the submission process if the usual cleanup fails or is skipped for some reason. * refactor: make clear that deduce_group() uses only the draft name * refactor: extract only draft name/revision in clean() method Minimizing the amount of validation done when accepting a file. The data extraction will be moved to asynchronous processing. * refactor: minimize checks and data extraction in api_upload() view * ci: fix dockerfiles to match sandbox testing * ci: tweak celery container docker-compose settings * refactor: clean up Draft parsing API and usage * remove get_draftname() from Draft api; set filename during init * further XMLDraft work - remember xml_version after parsing - extract filename/revision during init - comment out long broken get_abstract() method * adjust form clean() method to use changed API * feat: flesh out async submission processing First basically working pass! * feat: add state name for submission being validated asynchronously * feat: cancel submissions that async processing can't handle * refactor: simplify/consolidate async tasks and improve error handling * feat: add api_submission_status endpoint * refactor: return JSON from submission api endpoints * refactor: reuse cancel_submission method * refactor: clean up error reporting a bit * feat: guard against cancellation of a submission while validating Not bulletproof but should prevent * feat: indicate that a submission is still being validated * fix: do not delete submission files after creating them * chore: remove debug statement * test: add tests of the api_upload and api_submission_status endpoints * test: add tests and stubs for async side of submission handling * fix: gracefully handle (ignore) invalid IDs in async submit task * test: test process_uploaded_submission method * fix: fix failures of new tests * refactor: fix type checker complaints * test: test submission_status view of submission in "validating" state * fix: fix up migrations * fix: use the streamlined SubmissionBaseUploadForm for api_upload * feat: show submission history event timestamp as mouse-over text * fix: remove 'manual' as next state for 'validating' submission state * refactor: share SubmissionBaseUploadForm code with Deprecated version * fix: validate text submission title, update a couple comments * chore: disable requirements updating when celery dev container starts * feat: log traceback on unexpected error during submission processing * feat: allow secretariat to cancel "validating" submission * feat: indicate time since submission on the status page * perf: check submission rate thresholds earlier when possible No sense parsing details of a draft that is going to be dropped regardless of those details! * fix: create Submission before saving to reduce race condition window * fix: call deduce_group() with filename * refactor: remove code lint * refactor: change the api_upload URL to api/submission * docs: update submission API documentation * test: add tests of api_submission's text draft consistency checks * refactor: rename api_upload to api_submission to agree with new URL * test: test API documentation and submission thresholds * fix: fix a couple api_submission view renames missed in templates * chore: use base image + add arm64 support * ci: try to fix workflow_dispatch for celery worker * ci: another attempt to fix workflow_dispatch * ci: build celery image for submit-async branch * ci: fix typo * ci: publish celery worker to ghcr.io/painless-security * ci: install python requirements in celery image * ci: fix up requirements install on celery image * chore: remove XML_LIBRARY references that crept back in * feat: accept 'replaces' field in api_submission * docs: update api_submission documentation * fix: remove unused import * test: test "replaces" validation for submission API * test: test that "replaces" is set by api_submission * feat: trap TERM to gracefully stop celery container * chore: tweak celery/mq settings * docs: update installation instructions * ci: adjust paths that trigger celery worker image build * ci: fix branches/repo names left over from dev * ci: run manage.py check when initializing celery container Driver here is applying the patches. Starting the celery workers also invokes the check task, but this should cause a clearer failure if something fails. * docs: revise INSTALL instructions * ci: pass filename to pip update in celery container * docs: update INSTALL to include freezing pip versions Will be used to coordinate package versions with the celery container in production. * docs: add explanation of frozen-requirements.txt * ci: build image for sandbox deployment * ci: add additional build trigger path * docs: tweak INSTALL * fix: change INSTALL process to stop datatracker before running migrations * chore: use ietf.settings for manage.py check in celery container * chore: set uid/gid for celery worker * chore: create user/group in celery container if needed * chore: tweak docker compose/init so celery container works in dev * ci: build mq docker image * fix: move rabbitmq.pid to writeable location * fix: clear password when CELERY_PASSWORD is empty Setting to an empty password is really not a good plan! * chore: add shutdown debugging option to celery image * chore: add django-celery-beat package * chore: run "celery beat" in datatracker-celery image * chore: fix docker image name * feat: add task to cancel stale submissions * test: test the cancel_stale_submissions task * chore: make f-string with no interpolation a plain string Co-authored-by: Nicolas Giard <github@ngpixel.com> Co-authored-by: Robert Sparks <rjsparks@nostrum.com>
76 lines
2.3 KiB
Bash
Executable file
76 lines
2.3 KiB
Bash
Executable file
#!/bin/bash
|
|
#
|
|
# Environment parameters:
|
|
#
|
|
# CELERY_APP - name of application to pass to celery (defaults to ietf)
|
|
#
|
|
# CELERY_ROLE - 'worker' or 'beat' (defaults to 'worker')
|
|
#
|
|
# CELERY_UID - numeric uid for the celery worker process
|
|
#
|
|
# CELERY_GID - numeric gid for the celery worker process
|
|
#
|
|
# UPDATES_REQUIREMENTS_FROM - path, relative to /workspace mount, to a pip requirements
|
|
# file that should be installed at container startup. Default is no package install/update.
|
|
#
|
|
# DEBUG_TERM_TIMING - if non-empty, writes debug messages during shutdown after a TERM signal
|
|
#
|
|
WORKSPACEDIR="/workspace"
|
|
CELERY_ROLE="${CELERY_ROLE:-worker}"
|
|
|
|
cd "$WORKSPACEDIR" || exit 255
|
|
|
|
if [[ -n "${UPDATE_REQUIREMENTS_FROM}" ]]; then
|
|
reqs_file="${WORKSPACEDIR}/${UPDATE_REQUIREMENTS_FROM}"
|
|
echo "Updating requirements from ${reqs_file}..."
|
|
pip install --upgrade -r "${reqs_file}"
|
|
fi
|
|
|
|
if [[ "${CELERY_ROLE}" == "worker" ]]; then
|
|
echo "Running initial checks..."
|
|
/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check
|
|
fi
|
|
|
|
CELERY_OPTS=( "${CELERY_ROLE}" )
|
|
if [[ -n "${CELERY_UID}" ]]; then
|
|
# ensure that some group with the necessary GID exists in container
|
|
if ! id "${CELERY_UID}" ; then
|
|
adduser --system --uid "${CELERY_UID}" --no-create-home --disabled-login "celery-user-${CELERY_UID}"
|
|
fi
|
|
CELERY_OPTS+=("--uid=${CELERY_UID}")
|
|
fi
|
|
|
|
if [[ -n "${CELERY_GID}" ]]; then
|
|
# ensure that some group with the necessary GID exists in container
|
|
if ! getent group "${CELERY_GID}" ; then
|
|
addgroup --gid "${CELERY_GID}" "celery-group-${CELERY_GID}"
|
|
fi
|
|
CELERY_OPTS+=("--gid=${CELERY_GID}")
|
|
fi
|
|
|
|
log_term_timing_msgs () {
|
|
# output periodic debug message
|
|
while true; do
|
|
echo "Waiting for celery worker shutdown ($(date --utc --iso-8601=ns))"
|
|
sleep 0.5s
|
|
done
|
|
}
|
|
|
|
cleanup () {
|
|
# Cleanly terminate the celery app by sending it a TERM, then waiting for it to exit.
|
|
if [[ -n "${celery_pid}" ]]; then
|
|
echo "Gracefully terminating celery worker. This may take a few minutes if tasks are in progress..."
|
|
kill -TERM "${celery_pid}"
|
|
if [[ -n "${DEBUG_TERM_TIMING}" ]]; then
|
|
log_term_timing_msgs &
|
|
fi
|
|
wait "${celery_pid}"
|
|
fi
|
|
}
|
|
|
|
trap 'trap "" TERM; cleanup' TERM
|
|
# start celery in the background so we can trap the TERM signal
|
|
celery --app="${CELERY_APP:-ietf}" "${CELERY_OPTS[@]}" "$@" &
|
|
celery_pid=$!
|
|
wait "${celery_pid}"
|