refactor: adopt --app as celery global option (#15053)
This commit is contained in:
parent
42cb5266fa
commit
9967a5f640
|
|
@ -73,7 +73,7 @@ jobs:
|
|||
setup-postgres
|
||||
- name: Run celery
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: celery worker --app=superset.tasks.celery_app:app -Ofair -c 2 &
|
||||
run: celery --app=superset.tasks.celery_app:app worker -Ofair -c 2 &
|
||||
- name: Python unit tests (PostgreSQL)
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: |
|
||||
|
|
@ -148,7 +148,7 @@ jobs:
|
|||
setup-postgres
|
||||
- name: Run celery
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: celery worker --app=superset.tasks.celery_app:app -Ofair -c 2 &
|
||||
run: celery --app=superset.tasks.celery_app:app worker -Ofair -c 2 &
|
||||
- name: Python unit tests (PostgreSQL)
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: |
|
||||
|
|
|
|||
|
|
@ -62,7 +62,7 @@ jobs:
|
|||
setup-mysql
|
||||
- name: Run celery
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: celery worker --app=superset.tasks.celery_app:app -Ofair -c 2 &
|
||||
run: celery --app=superset.tasks.celery_app:app worker -Ofair -c 2 &
|
||||
- name: Python unit tests (MySQL)
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: |
|
||||
|
|
@ -126,7 +126,7 @@ jobs:
|
|||
setup-postgres
|
||||
- name: Run celery
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: celery worker --app=superset.tasks.celery_app:app -Ofair -c 2 &
|
||||
run: celery --app=superset.tasks.celery_app:app worker -Ofair -c 2 &
|
||||
- name: Python unit tests (PostgreSQL)
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: |
|
||||
|
|
@ -182,7 +182,7 @@ jobs:
|
|||
mkdir ${{ github.workspace }}/.temp
|
||||
- name: Run celery
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: celery worker --app=superset.tasks.celery_app:app -Ofair -c 2 &
|
||||
run: celery --app=superset.tasks.celery_app:app worker -Ofair -c 2 &
|
||||
- name: Python unit tests (SQLite)
|
||||
if: steps.check.outcome == 'failure'
|
||||
run: |
|
||||
|
|
|
|||
|
|
@ -1264,7 +1264,7 @@ To do this, you'll need to:
|
|||
- Start up a celery worker
|
||||
|
||||
```shell script
|
||||
celery worker --app=superset.tasks.celery_app:app -Ofair
|
||||
celery --app=superset.tasks.celery_app:app worker -Ofair
|
||||
```
|
||||
|
||||
Note that:
|
||||
|
|
|
|||
|
|
@ -38,10 +38,10 @@ fi
|
|||
|
||||
if [[ "${1}" == "worker" ]]; then
|
||||
echo "Starting Celery worker..."
|
||||
celery worker --app=superset.tasks.celery_app:app -Ofair -l INFO
|
||||
celery --app=superset.tasks.celery_app:app worker -Ofair -l INFO
|
||||
elif [[ "${1}" == "beat" ]]; then
|
||||
echo "Starting Celery beat..."
|
||||
celery beat --app=superset.tasks.celery_app:app --pidfile /tmp/celerybeat.pid -l INFO -s "${SUPERSET_HOME}"/celerybeat-schedule
|
||||
celery --app=superset.tasks.celery_app:app beat --pidfile /tmp/celerybeat.pid -l INFO -s "${SUPERSET_HOME}"/celerybeat-schedule
|
||||
elif [[ "${1}" == "app" ]]; then
|
||||
echo "Starting web app..."
|
||||
flask run -p 8088 --with-threads --reload --debugger --host=0.0.0.0
|
||||
|
|
|
|||
|
|
@ -1080,11 +1080,11 @@ have the same configuration.
|
|||
|
||||
* To start a Celery worker to leverage the configuration run: ::
|
||||
|
||||
celery worker --app=superset.tasks.celery_app:app --pool=prefork -O fair -c 4
|
||||
celery --app=superset.tasks.celery_app:app worker --pool=prefork -O fair -c 4
|
||||
|
||||
* To start a job which schedules periodic background jobs, run ::
|
||||
|
||||
celery beat --app=superset.tasks.celery_app:app
|
||||
celery --app=superset.tasks.celery_app:app beat
|
||||
|
||||
To setup a result backend, you need to pass an instance of a derivative
|
||||
of ``from cachelib.base.BaseCache`` to the ``RESULTS_BACKEND``
|
||||
|
|
|
|||
|
|
@ -246,7 +246,7 @@ services:
|
|||
- superset
|
||||
- postgres
|
||||
- redis
|
||||
command: "celery worker --app=superset.tasks.celery_app:app --pool=gevent --concurrency=500"
|
||||
command: "celery --app=superset.tasks.celery_app:app worker --pool=gevent --concurrency=500"
|
||||
volumes:
|
||||
- ./config/:/app/pythonpath/
|
||||
beat:
|
||||
|
|
@ -258,7 +258,7 @@ services:
|
|||
- superset
|
||||
- postgres
|
||||
- redis
|
||||
command: "celery beat --app=superset.tasks.celery_app:app --pidfile /tmp/celerybeat.pid --schedule /tmp/celerybeat-schedule"
|
||||
command: "celery --app=superset.tasks.celery_app:app beat --pidfile /tmp/celerybeat.pid --schedule /tmp/celerybeat-schedule"
|
||||
volumes:
|
||||
- ./config/:/app/pythonpath/
|
||||
superset:
|
||||
|
|
|
|||
|
|
@ -57,13 +57,13 @@ CELERY_CONFIG = CeleryConfig
|
|||
To start a Celery worker to leverage the configuration, run the following command:
|
||||
|
||||
```
|
||||
celery worker --app=superset.tasks.celery_app:app --pool=prefork -O fair -c 4
|
||||
celery --app=superset.tasks.celery_app:app worker --pool=prefork -O fair -c 4
|
||||
```
|
||||
|
||||
To start a job which schedules periodic background jobs, run the following command:
|
||||
|
||||
```
|
||||
celery beat --app=superset.tasks.celery_app:app
|
||||
celery --app=superset.tasks.celery_app:app beat
|
||||
```
|
||||
|
||||
To setup a result backend, you need to pass an instance of a derivative of from
|
||||
|
|
|
|||
|
|
@ -22,7 +22,7 @@ maintainers:
|
|||
- name: craig-rueda
|
||||
email: craig@craigrueda.com
|
||||
url: https://github.com/craig-rueda
|
||||
version: 0.1.3
|
||||
version: 0.1.4
|
||||
dependencies:
|
||||
- name: postgresql
|
||||
version: 10.2.0
|
||||
|
|
|
|||
|
|
@ -203,7 +203,7 @@ supersetCeleryBeat:
|
|||
command:
|
||||
- "/bin/sh"
|
||||
- "-c"
|
||||
- ". {{ .Values.configMountPath }}/superset_bootstrap.sh; celery beat --app=superset.tasks.celery_app:app --pidfile /tmp/celerybeat.pid --schedule /tmp/celerybeat-schedule"
|
||||
- ". {{ .Values.configMountPath }}/superset_bootstrap.sh; celery --app=superset.tasks.celery_app:app beat --pidfile /tmp/celerybeat.pid --schedule /tmp/celerybeat-schedule"
|
||||
forceReload: false # If true, forces deployment to reload on each upgrade
|
||||
initContainers:
|
||||
- name: wait-for-postgres
|
||||
|
|
|
|||
Loading…
Reference in New Issue