Implement local task locking using Django locmem cache backend
This commit is contained in:
@@ -1,10 +1,9 @@
|
|||||||
from datetime import timedelta, datetime
|
from datetime import timedelta, datetime
|
||||||
import platform
|
import platform
|
||||||
import time
|
|
||||||
import random
|
|
||||||
|
|
||||||
from django.db.models import Q
|
from django.db.models import Q
|
||||||
from django.utils.translation import ugettext as _
|
from django.utils.translation import ugettext as _
|
||||||
|
from django.core.cache.backends.locmem import CacheClass
|
||||||
|
|
||||||
from celery.decorators import task, periodic_task
|
from celery.decorators import task, periodic_task
|
||||||
from celery.task.control import inspect
|
from celery.task.control import inspect
|
||||||
@@ -18,13 +17,12 @@ from ocr.conf.settings import NODE_CONCURRENT_EXECUTION
|
|||||||
from ocr.conf.settings import REPLICATION_DELAY
|
from ocr.conf.settings import REPLICATION_DELAY
|
||||||
from ocr.conf.settings import QUEUE_PROCESSING_INTERVAL
|
from ocr.conf.settings import QUEUE_PROCESSING_INTERVAL
|
||||||
|
|
||||||
|
LOCK_EXPIRE = 60 * 5 # Lock expires in 5 minutes
|
||||||
|
local_cache = CacheClass([], {})
|
||||||
|
|
||||||
@task
|
@task
|
||||||
def task_process_queue_document(queue_document_id):
|
def task_process_queue_document(queue_document_id):
|
||||||
queue_document = QueueDocument.objects.get(id=queue_document_id)
|
queue_document = QueueDocument.objects.get(id=queue_document_id)
|
||||||
if queue_document.state == QUEUEDOCUMENT_STATE_PROCESSING:
|
|
||||||
#Recheck to avoid race condition
|
|
||||||
return
|
|
||||||
queue_document.state = QUEUEDOCUMENT_STATE_PROCESSING
|
queue_document.state = QUEUEDOCUMENT_STATE_PROCESSING
|
||||||
queue_document.node_name = platform.node()
|
queue_document.node_name = platform.node()
|
||||||
queue_document.result = task_process_queue_document.request.id
|
queue_document.result = task_process_queue_document.request.id
|
||||||
@@ -63,10 +61,12 @@ def reset_orphans():
|
|||||||
|
|
||||||
@periodic_task(run_every=timedelta(seconds=QUEUE_PROCESSING_INTERVAL))
|
@periodic_task(run_every=timedelta(seconds=QUEUE_PROCESSING_INTERVAL))
|
||||||
def task_process_document_queues():
|
def task_process_document_queues():
|
||||||
|
lock_id = '%s-lock-%s' % ('task_process_document_queues', platform.node())
|
||||||
|
acquire_lock = lambda: local_cache.add(lock_id, 'true', LOCK_EXPIRE)
|
||||||
|
release_lock = lambda: local_cache.delete(lock_id)
|
||||||
|
|
||||||
|
if acquire_lock():
|
||||||
reset_orphans()
|
reset_orphans()
|
||||||
#Introduce random 0 < t < 1 second delay to further reduce the
|
|
||||||
#chance of a race condition
|
|
||||||
time.sleep(random.random())
|
|
||||||
q_pending = Q(state=QUEUEDOCUMENT_STATE_PENDING)
|
q_pending = Q(state=QUEUEDOCUMENT_STATE_PENDING)
|
||||||
q_delayed = Q(delay=True)
|
q_delayed = Q(delay=True)
|
||||||
q_delay_interval = Q(datetime_submitted__lt=datetime.now() - timedelta(seconds=REPLICATION_DELAY))
|
q_delay_interval = Q(datetime_submitted__lt=datetime.now() - timedelta(seconds=REPLICATION_DELAY))
|
||||||
@@ -83,3 +83,4 @@ def task_process_document_queues():
|
|||||||
task_process_queue_document.delay(oldest_queued_document.id)
|
task_process_queue_document.delay(oldest_queued_document.id)
|
||||||
except Exception, e:
|
except Exception, e:
|
||||||
print 'DocumentQueueWatcher exception: %s' % e
|
print 'DocumentQueueWatcher exception: %s' % e
|
||||||
|
release_lock()
|
||||||
|
|||||||
Reference in New Issue
Block a user