-
-
Notifications
You must be signed in to change notification settings - Fork 83
/
postprocessor.py
446 lines (381 loc) · 21.5 KB
/
postprocessor.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
unmanic.postprocessor.py
Written by: Josh.5 <jsunnex@gmail.com>
Date: 23 Apr 2019, (7:33 PM)
Copyright:
Copyright (C) Josh Sunnex - All Rights Reserved
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE
OR OTHER DEALINGS IN THE SOFTWARE.
"""
import os
import shutil
import threading
import time
from unmanic import config
from unmanic.libs import common, history
from unmanic.libs.library import Library
from unmanic.libs.notifications import Notifications
from unmanic.libs.plugins import PluginsHandler
"""
The post-processor handles all tasks carried out on completion of a workers task.
This may be on either success or failure of the task.
The post-processor runs as a single thread, processing completed jobs one at a time.
This prevents conflicting copy operations or deleting a file that is also being post processed.
"""
class PostProcessError(Exception):
def __init___(self, expected_var, result_var):
Exception.__init__(self, "Errors found during post process checks. Expected {}, but instead found {}".format(
expected_var, result_var))
self.expected_var = expected_var
self.result_var = result_var
class PostProcessor(threading.Thread):
"""
PostProcessor
"""
def __init__(self, data_queues, task_queue, event):
super(PostProcessor, self).__init__(name='PostProcessor')
self.logger = data_queues["logging"].get_logger(self.name)
self.event = event
self.data_queues = data_queues
self.settings = config.Config()
self.task_queue = task_queue
self.abort_flag = threading.Event()
self.current_task = None
self.ffmpeg = None
self.abort_flag.clear()
def _log(self, message, message2='', level="info"):
message = common.format_message(message, message2)
getattr(self.logger, level)(message)
def stop(self):
self.abort_flag.set()
def run(self):
self._log("Starting PostProcessor Monitor loop...")
while not self.abort_flag.is_set():
self.event.wait(1)
if not self.system_configuration_is_valid():
self.event.wait(2)
continue
while not self.abort_flag.is_set() and not self.task_queue.task_list_processed_is_empty():
self.event.wait(.2)
self.current_task = self.task_queue.get_next_processed_tasks()
if self.current_task:
try:
self._log("Post-processing task - {}".format(self.current_task.get_source_abspath()))
except Exception as e:
self._log("Exception in fetching task absolute path", message2=str(e), level="exception")
if self.current_task.get_task_type() == 'local':
try:
# Post processes the converted file (return it to original directory etc.)
self.post_process_file()
except Exception as e:
self._log("Exception in post-processing local task file", message2=str(e), level="exception")
try:
# Write source and destination data to historic log
self.write_history_log()
except Exception as e:
self._log("Exception in writing history log", message2=str(e), level="exception")
try:
# Remove file from task queue
self.current_task.delete()
except Exception as e:
self._log("Exception in removing task from task list", message2=str(e), level="exception")
else:
try:
# Post processes the remote converted file (return it to original directory etc.)
self.post_process_remote_file()
except Exception as e:
self._log("Exception in post-processing remote task file", message2=str(e), level="exception")
try:
# Write source and destination data to historic log
self.dump_history_log()
except Exception as e:
self._log("Exception in dumping history log for remote task", message2=str(e), level="exception")
try:
# Update the task status to 'complete'
self.current_task.set_status('complete')
except Exception as e:
self._log("Exception in marking remote task as complete", message2=str(e), level="exception")
self._log("Leaving PostProcessor Monitor loop...")
def system_configuration_is_valid(self):
"""
Check and ensure the system configuration is correct for running
:return:
"""
valid = True
plugin_handler = PluginsHandler()
if plugin_handler.get_incompatible_enabled_plugins(self.data_queues.get('frontend_messages')):
valid = False
if not Library.within_library_count_limits(self.data_queues.get('frontend_messages')):
valid = False
return valid
def post_process_file(self):
# Init plugins handler
plugin_handler = PluginsHandler()
# Read current task data
# task_data = self.current_task.get_task_data()
library_id = self.current_task.get_task_library_id()
cache_path = self.current_task.get_cache_path()
source_data = self.current_task.get_source_data()
destination_data = self.current_task.get_destination_data()
# Move file back to original folder and remove source
file_move_processes_success = True
# Create a list for filling with destination paths
destination_files = []
if self.current_task.task.success:
# Run a postprocess file movement on the cache file for for each plugin that configures it
# Fetch all 'postprocessor.file_move' plugin modules
plugin_modules = plugin_handler.get_enabled_plugin_modules_by_type('postprocessor.file_move',
library_id=library_id)
# Check if the source file needs to be removed by default (only if it does not match the destination file)
remove_source_file = False
if source_data['abspath'] != destination_data['abspath']:
remove_source_file = True
# Set initial data (some fields will be overwritten further down)
# - 'library_id' - The library ID for this task
# - 'source_data' - Dictionary of data pertaining to the source file
# - 'remove_source_file' - True to remove the original file (default is True if file name has changed)
# - 'copy_file' - True to run a plugin initiated file copy (default is False unless the plugin says otherwise)
# - 'file_in' - Source path to copy from (if 'copy_file' is True)
# - 'file_out' - Destination path to copy to (if 'copy_file' is True)
# - 'run_default_file_copy' - Prevent the final Unmanic post-process file movement (if different from the original file name)
data = {
'library_id': library_id,
'source_data': None,
'remove_source_file': remove_source_file,
'copy_file': None,
'file_in': None,
'file_out': None,
'run_default_file_copy': True,
}
for plugin_module in plugin_modules:
# Always set source_data to the original file's source_data
data["source_data"] = source_data
# Always set copy_file to False
data["copy_file"] = False
# Always set file in to cache path
data["file_in"] = cache_path
# Always set file out to destination data absolute path
data["file_out"] = destination_data.get('abspath')
# Run plugin to update data
if not plugin_handler.exec_plugin_runner(data, plugin_module.get('plugin_id'), 'postprocessor.file_move'):
# Do not continue with this plugin module's loop
continue
if data.get('copy_file'):
# Copy the file
file_in = os.path.abspath(data.get('file_in'))
file_out = os.path.abspath(data.get('file_out'))
if not self.__copy_file(file_in, file_out, destination_files, plugin_module.get('plugin_id')):
file_move_processes_success = False
else:
self._log("Plugin did not request a file copy ({})".format(plugin_module.get('plugin_id')), level='debug')
# Unmanic's default file movement process
# Only carry out final post-processor file moments if all others were successful
if file_move_processes_success and data.get('run_default_file_copy'):
# Run the default post-process file movement.
# This will always move the file back to the original location.
# If that original location is the same file name, it will overwrite the original file.
if destination_data.get('abspath') == source_data.get('abspath'):
# Only run the final file copy to overwrite the source file if the remove_source_file flag was never set
# The remove_source_file flag will remove the source file in later lines after this copy operation,
# so if we did copy the file here, it would be a waste of time
if not data.get('remove_source_file'):
if not self.__copy_file(cache_path, destination_data.get('abspath'), destination_files, 'DEFAULT',
move=True):
file_move_processes_success = False
elif not self.__copy_file(cache_path, destination_data.get('abspath'), destination_files, 'DEFAULT',
move=True):
file_move_processes_success = False
# Source file removal process
# Only run if all final post-processor file moments were successful
if file_move_processes_success:
# Check if the remove source flag is still True after all plugins have run. If so, we will remove the source file
if data.get('remove_source_file'):
# Only carry out a source removal if the file exists and the final copy was also successful
if file_move_processes_success and os.path.exists(source_data.get('abspath')):
self._log("Removing source: {}".format(source_data.get('abspath')))
os.remove(source_data.get('abspath'))
else:
self._log("Keeping source file '{}'. Not all postprocessor file movement functions completed.".format(
source_data.get('abspath')), level="warning")
# Log a final error if not all file moments were successful
if not file_move_processes_success:
self._log(
"Error while running postprocessor file movement on file '{}'. Not all postprocessor file movement functions completed.".format(
cache_path), level="error")
else:
self._log("Skipping file movement post-processor as the task was not successful '{}'".format(cache_path),
level='warning')
# Fetch all 'postprocessor.task_result' plugin modules
plugin_modules = plugin_handler.get_enabled_plugin_modules_by_type('postprocessor.task_result', library_id=library_id)
for plugin_module in plugin_modules:
data = {
'final_cache_path': cache_path,
'library_id': library_id,
'source_data': source_data,
'task_processing_success': self.current_task.task.success,
'file_move_processes_success': file_move_processes_success,
'destination_files': destination_files,
}
# Run plugin to update data
if not plugin_handler.exec_plugin_runner(data, plugin_module.get('plugin_id'), 'postprocessor.task_result'):
# Do not continue with this plugin module's loop
continue
# Cleanup cache files
self.__cleanup_cache_files(cache_path)
def post_process_remote_file(self):
"""
Process remote files.
Remote files are not processed by plugins.
TODO: Should we move remote tasks to a permanent download location within the cache path? Possibly not...
:return:
"""
# Read current task data
cache_path = self.current_task.get_cache_path()
source_data = self.current_task.get_source_data()
destination_data = self.current_task.get_destination_data()
# Remove the source
if os.path.exists(source_data.get('abspath')):
self._log("Removing remote source: {}".format(source_data.get('abspath')))
os.remove(source_data.get('abspath'))
else:
self._log("Remote source file '{}' does not exist!".format(source_data.get('abspath')), level="warning")
# Copy final cache file to original directory
if os.path.exists(cache_path):
self.__copy_file(cache_path, destination_data.get('abspath'), [], 'DEFAULT', move=True)
else:
self._log("Final cache file '{}' does not exist!".format(cache_path), level="warning")
# Cleanup cache files
self.__cleanup_cache_files(cache_path)
# Modify the task abspath - this may be different now
self.current_task.modify_path(destination_data.get('abspath'))
def __cleanup_cache_files(self, cache_path):
"""
Remove cache files and the cache directory
This ensure we are not simply blindly removing a whole directory.
It ensures were are in-fact only deleting this task's cache files.
:param cache_path:
:return:
"""
task_cache_directory = os.path.dirname(cache_path)
if os.path.exists(task_cache_directory) and "unmanic_file_conversion" in task_cache_directory:
self._log("Removing task cache directory '{}'".format(task_cache_directory))
try:
shutil.rmtree(task_cache_directory)
except Exception as e:
self._log("Exception while clearing cache path '{}'".format(str(e)), level='error')
def __copy_file(self, file_in, file_out, destination_files, plugin_id, move=False):
if move:
self._log("Move file triggered by ({}) {} --> {}".format(plugin_id, file_in, file_out))
else:
self._log("Copy file triggered by ({}) {} --> {}".format(plugin_id, file_in, file_out))
try:
# Ensure the src and dst are not the same file
if os.path.exists(file_out) and os.path.samefile(file_in, file_out):
self._log("The file_in and file_out path are the same file. Nothing will be done! '{}'".format(file_in),
level="warning")
return False
# Get a checksum prior to copy
if not os.path.exists(file_in):
self._log("The file_in path does not exist! '{}'".format(file_in), level="warning")
self.event.wait(1)
self._log("Fetching checksum of source file '{}'.".format(file_in), level='debug')
# Use a '.part' suffix for the file movement, then rename it after
part_file_out = os.path.join("{}.unmanic.part".format(file_out))
# Carry out the file movement
if move:
self._log("Moving file '{}' --> '{}'.".format(file_in, part_file_out), level='debug')
if os.path.exists(part_file_out):
os.remove(part_file_out)
shutil.move(file_in, part_file_out, copy_function=shutil.copyfile)
else:
self._log("Copying file '{}' --> '{}'.".format(file_in, part_file_out), level='debug')
shutil.copyfile(file_in, part_file_out)
# Remove dest file if it already exists (required only for moves)
if os.path.exists(file_out):
self._log("The file_out path already exists. Removing file '{}'".format(file_out), level="debug")
os.remove(file_out)
# Move file from part to final destination
self._log("Renaming file '{}' --> '{}'.".format(part_file_out, file_out), level='debug')
shutil.move(part_file_out, file_out, copy_function=shutil.copyfile)
# Write final path to destination_files list
destination_files.append(file_out)
# Mark move process a success
return True
except Exception as e:
self._log("Exception while copying file {} to {}:".format(file_in, file_out),
message2=str(e), level="exception")
file_move_processes_success = False
return file_move_processes_success
def write_history_log(self):
"""
Record task history
:return:
"""
self._log("Writing task history log.", level='debug')
history_logging = history.History()
task_dump = self.current_task.task_dump()
# If task fails, the add a notification that a task has failed
if not self.current_task.task.success:
notifications = Notifications()
notifications.add(
{
'uuid': 'newFailedTask',
'type': 'error',
'icon': 'report',
'label': 'failedTaskLabel',
'message': 'You have a new failed task in your completed tasks list',
'navigation': {
'push': '/ui/dashboard',
'events': [
'completedTasksShowFailed',
],
},
})
history_logging.save_task_history(
{
'task_label': task_dump.get('task_label', ''),
'abspath': task_dump.get('abspath', ''),
'task_success': task_dump.get('task_success', ''),
'start_time': task_dump.get('start_time', ''),
'finish_time': task_dump.get('finish_time', ''),
'processed_by_worker': task_dump.get('processed_by_worker', ''),
'log': task_dump.get('log', ''),
}
)
def dump_history_log(self):
self._log("Dumping remote task history log.", level='debug')
task_dump = self.current_task.task_dump()
destination_data = self.current_task.get_destination_data()
# Dump history log as metadata in the file's path
tasks_data_file = os.path.join(os.path.dirname(destination_data.get('abspath')), 'data.json')
result = common.json_dump_to_file(
{
'task_label': task_dump.get('task_label', ''),
'abspath': task_dump.get('abspath', ''),
'task_success': task_dump.get('task_success', ''),
'start_time': task_dump.get('start_time', ''),
'finish_time': task_dump.get('finish_time', ''),
'processed_by_worker': task_dump.get('processed_by_worker', ''),
'log': task_dump.get('log', ''),
'checksum': 'UNKNOWN',
}
, tasks_data_file)
if not result['success']:
for message in result['errors']:
self._log("Exception:", message2=str(message), level="exception")
raise Exception("Exception in dumping completed task data to file")