aboutsummaryrefslogtreecommitdiffstats
path: root/mediagoblin/submit/views.py
diff options
context:
space:
mode:
Diffstat (limited to 'mediagoblin/submit/views.py')
-rw-r--r--mediagoblin/submit/views.py42
1 files changed, 34 insertions, 8 deletions
diff --git a/mediagoblin/submit/views.py b/mediagoblin/submit/views.py
index ba13b755..1ba17954 100644
--- a/mediagoblin/submit/views.py
+++ b/mediagoblin/submit/views.py
@@ -14,19 +14,21 @@
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
+import uuid
+
from os.path import splitext
from cgi import FieldStorage
-from string import split
from werkzeug.utils import secure_filename
+from mediagoblin.db.util import ObjectId
from mediagoblin.util import (
render_to_response, redirect, cleaned_markdown_conversion, \
convert_to_tag_list_of_dicts)
from mediagoblin.util import pass_to_ugettext as _
from mediagoblin.decorators import require_active_login
from mediagoblin.submit import forms as submit_forms, security
-from mediagoblin.process_media import process_media_initial
+from mediagoblin.process_media import process_media, mark_entry_failed
from mediagoblin.messages import add_message, SUCCESS
@@ -51,6 +53,7 @@ def submit_start(request):
# create entry and save in database
entry = request.db.MediaEntry()
+ entry['_id'] = ObjectId()
entry['title'] = (
request.POST['title']
or unicode(splitext(filename)[0]))
@@ -66,10 +69,6 @@ def submit_start(request):
entry['tags'] = convert_to_tag_list_of_dicts(
request.POST.get('tags'))
- # Save, just so we can get the entry id for the sake of using
- # it to generate the file path
- entry.save(validate=False)
-
# Generate a slug from the title
entry.generate_slug()
@@ -88,10 +87,37 @@ def submit_start(request):
# Add queued filename to the entry
entry['queued_media_file'] = queue_filepath
+
+ # We generate this ourselves so we know what the taks id is for
+ # retrieval later.
+ # (If we got it off the task's auto-generation, there'd be a risk of
+ # a race condition when we'd save after sending off the task)
+ task_id = unicode(uuid.uuid4())
+ entry['queued_task_id'] = task_id
+
+ # Save now so we have this data before kicking off processing
entry.save(validate=True)
- # queue it for processing
- process_media_initial.delay(unicode(entry['_id']))
+ # Pass off to processing
+ #
+ # (... don't change entry after this point to avoid race
+ # conditions with changes to the document via processing code)
+ try:
+ process_media.apply_async(
+ [unicode(entry['_id'])], {},
+ task_id=task_id)
+ except BaseException as exc:
+ # The purpose of this section is because when running in "lazy"
+ # or always-eager-with-exceptions-propagated celery mode that
+ # the failure handling won't happen on Celery end. Since we
+ # expect a lot of users to run things in this way we have to
+ # capture stuff here.
+ #
+ # ... not completely the diaper pattern because the exception is
+ # re-raised :)
+ mark_entry_failed(entry[u'_id'], exc)
+ # re-raise the exception
+ raise
add_message(request, SUCCESS, _('Woohoo! Submitted!'))