Django APScheduler
APScheduler for Django.
This is a Django app that adds a lightweight wrapper around APScheduler. It enables storing persistent jobs in the database using Django's ORM.
django-apscheduler is a great choice for quickly and easily adding basic scheduling features to your Django applications with minimal dependencies and very little additional configuration. The ideal use case probably involves running a handful of tasks on a fixed execution schedule.
PLEASE NOTE: the trade-off of this simplicity is that you need to be careful to ensure that you have only ONE scheduler actively running at a particular point in time.
This limitation is due to the fact that APScheduler does not currently have any interprocess synchronization and signalling scheme that would enable the scheduler to be notified when a job has been added, modified, or removed from a job store (in other words, different schedulers won't be able to tell if a job has already been run by another scheduler, and changing a job's scheduled run time directly in the database does nothing unless you also restart the scheduler).
Depending on how you are currently doing your
Django deployments, working with this
limitation might require a bit of thought. It is quite common to start up many webserver worker process in production
environments in order to scale and handle large volumes of user traffic. If each of these worker processes end up
running their own scheduler then this can result in jobs being missed or executed multiple times, as well as duplicate
entries being created in the DjangoJobExecution
table.
Support for sharing a persistent job store between multiple schedulers appears to be planned for an upcoming APScheduler 4.0 release.
So for now your options are to either:
-
Use a custom Django management command to start a single scheduler in its own dedicated process (recommended - see the
runapscheduler.py
example below); or -
Implement your own remote processing logic to ensure that a single
DjangoJobStore
can be used by all of the webserver's worker processes in a coordinated and synchronized way (might not be worth the extra effort and increased complexity for most use cases); or -
Select an alternative task processing library that does support inter-process communication using some sort of shared message broker like Redis, RabbitMQ, Amazon SQS or the like (see: https://djangopackages.org/grids/g/workers-queues-tasks/ for popular options).
Features
- A custom APScheduler job store
(
DjangoJobStore
) that persists scheduled jobs to the Django database. You can view the scheduled jobs and monitor the job execution directly via the Django admin interface:
- The job store also maintains a history of all job executions of the currently scheduled jobs, along with status codes and exceptions (if any):
Note: APScheduler will automatically remove jobs from the job store as soon as their last scheduled execution has been triggered. This will also delete the corresponding job execution entries from the database (i.e. job execution logs are only maintained for 'active' jobs.):
- Job executions can also be triggered manually via the
DjangoJob
admin page:
Note: In order to prevent long running jobs from causing the Django HTTP request to time out, the combined maximum
run time for all APScheduler jobs that are started via the Django admin site is 25 seconds. This timeout value can be
configured via the APSCHEDULER_RUN_NOW_TIMEOUT
setting.
Installation
pip install django-apscheduler
Quick start
- Add
django_apscheduler
to yourINSTALLED_APPS
setting like this:
INSTALLED_APPS = (
# ...
"django_apscheduler",
)
- django-apscheduler comes with sensible configuration defaults out of the box. The defaults can be overridden by adding
the following settings to your Django
settings.py
file:
# Format string for displaying run time timestamps in the Django admin site. The default
# just adds seconds to the standard Django format, which is useful for displaying the timestamps
# for jobs that are scheduled to run on intervals of less than one minute.
#
# See https://docs.djangoproject.com/en/dev/ref/settings/#datetime-format for format string
# syntax details.
APSCHEDULER_DATETIME_FORMAT = "N j, Y, f:s a"
# Maximum run time allowed for jobs that are triggered manually via the Django admin site, which
# prevents admin site HTTP requests from timing out.
#
# Longer running jobs should probably be handed over to a background task processing library
# that supports multiple background worker processes instead (e.g. Dramatiq, Celery, Django-RQ,
# etc. See: https://djangopackages.org/grids/g/workers-queues-tasks/ for popular options).
APSCHEDULER_RUN_NOW_TIMEOUT = 25 # Seconds
-
Run
python manage.py migrate
to create the django_apscheduler models. -
Add a custom Django management command to your project that schedules the APScheduler jobs and starts the scheduler:
# runapscheduler.py
import logging
from django.conf import settings
from apscheduler.schedulers.blocking import BlockingScheduler
from apscheduler.triggers.cron import CronTrigger
from django.core.management.base import BaseCommand
from django_apscheduler.jobstores import DjangoJobStore
from django_apscheduler.models import DjangoJobExecution
from django_apscheduler import util
logger = logging.getLogger(__name__)
def my_job():
# Your job processing logic here...
pass
# The `close_old_connections` decorator ensures that database connections, that have become
# unusable or are obsolete, are closed before and after your job has run. You should use it
# to wrap any jobs that you schedule that access the Django database in any way.
@util.close_old_connections
def delete_old_job_executions(max_age=604_800):
"""
This job deletes APScheduler job execution entries older than `max_age` from the database.
It helps to prevent the database from filling up with old historical records that are no
longer useful.
:param max_age: The maximum length of time to retain historical job execution records.
Defaults to 7 days.
"""
DjangoJobExecution.objects.delete_old_job_executions(max_age)
class Command(BaseCommand):
help = "Runs APScheduler."
def handle(self, *args, **options):
scheduler = BlockingScheduler(timezone=settings.TIME_ZONE)
scheduler.add_jobstore(DjangoJobStore(), "default")
scheduler.add_job(
my_job,
trigger=CronTrigger(second="*/10"), # Every 10 seconds
id="my_job", # The `id` assigned to each job MUST be unique
max_instances=1,
replace_existing=True,
)
logger.info("Added job 'my_job'.")
scheduler.add_job(
delete_old_job_executions,
trigger=CronTrigger(
day_of_week="mon", hour="00", minute="00"
), # Midnight on Monday, before start of the next work week.
id="delete_old_job_executions",
max_instances=1,
replace_existing=True,
)
logger.info(
"Added weekly job: 'delete_old_job_executions'."
)
try:
logger.info("Starting scheduler...")
scheduler.start()
except KeyboardInterrupt:
logger.info("Stopping scheduler...")
scheduler.shutdown()
logger.info("Scheduler shut down successfully!")
-
The management command defined above should be invoked via
./manage.py runapscheduler
whenever the webserver serving your Django application is started. The details of how and where this should be done is implementation specific, and depends on which webserver you are using and how you are deploying your application to production. For most people this should involve configuring a supervisor process of sorts. -
Register any APScheduler jobs as you would normally. Note that if you haven't set
DjangoJobStore
as the'default'
job store, then you will need to includejobstore='djangojobstore'
in yourscheduler.add_job()
calls.
Advanced usage
django-apscheduler assumes that you are already familiar with APScheduler and its proper use. If not, then please head over to the project page and have a look through the APScheduler documentation.
It is possible to make use
of different types of schedulers
depending on your environment and use case. If you would prefer running a BackgroundScheduler
instead of using a
BlockingScheduler
, then you should be aware that using APScheduler with uWSGI requires some additional
configuration steps in
order to re-enable threading support.
Supported databases
Please take note of the list of databases that are officially supported by Django. django-apscheduler probably won't work with unsupported databases like Microsoft SQL Server, MongoDB, and the like.
Database connections and timeouts
django-apscheduler is dependent on the standard Django database configuration settings. These settings, in combination with how your database server has been configured, determine how connection management will be performed for your specific deployment.
The close_old_connections
decorator should be applied to APScheduler jobs that require database access. Doing so
ensures that Django's CONN_MAX_AGE
configuration setting is enforced before and after your job is run. This mirrors the standard Django functionality of
doing the same before and after handling each HTTP request.
If you still encounter any kind of 'lost database connection' errors then it probably means that:
- Your database connections timed out in the middle of executing a job. You should probably consider incorporating a connection pooler as part of your deployment for more robust database connection management (e.g. pgbouncer for PostgreSQL, or the equivalent for other DB platforms).
- Your database server has crashed / been restarted. Django will not reconnect automatically and you need to re-start django-apscheduler as well.
Common footguns
Unless you have a very specific set of requirements, and have intimate knowledge of the inner workings of APScheduler,
you really shouldn't be using BackgroundScheduler
. Doing so can lead to all sorts of temptations like:
- Firing up a scheduler inside of a Django view: this will most likely cause more than one scheduler to run concurrently and lead to jobs running multiple times (see the above introduction to this README for a more thorough treatment of the subject).
- Bootstrapping a scheduler somewhere else inside your Django application: it feels like this should solve the problem mentioned above and guarantee that only one scheduler is running. The downside is that you have just delegated the management of all of your background task processing threads to whatever webserver you are using (Gunicorn, uWSGI, etc.). The webserver will probably kill any long-running threads (your jobs) with extreme prejudice (thinking that they are caused by misbehaving HTTP requests).
Relying on BlockingScheduler
forces you to run APScheduler in its own dedicated process that is not handled or
monitored by the webserver. The example code provided in runapscheduler.py
above is a good starting point.