-
Notifications
You must be signed in to change notification settings - Fork 28
/
views.py
230 lines (181 loc) · 8.62 KB
/
views.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
import urllib
from celery.result import AsyncResult
from celery_once import AlreadyQueued
from drf_yasg import openapi
from drf_yasg.utils import swagger_auto_schema
from rest_framework import status
from rest_framework.parsers import MultiPartParser
from rest_framework.permissions import IsAuthenticated
from rest_framework.response import Response
from rest_framework.views import APIView
from core.common.permissions import IsSuperuser
from core.common.services import RedisService
from core.common.swagger_parameters import update_if_exists_param, task_param, result_param, username_param, \
file_upload_param, file_url_param, apps_param, parallel_threads_param
from core.common.tasks import rebuild_indexes, populate_indexes
from core.common.utils import parse_bulk_import_task_id, task_exists, flower_get, queue_bulk_import
from core.importers.constants import ALREADY_QUEUED, INVALID_UPDATE_IF_EXISTS, NO_CONTENT_TO_IMPORT
def import_response(request, import_queue, data, threads=None, inline=False):
if not data:
return Response(dict(exception=NO_CONTENT_TO_IMPORT), status=status.HTTP_400_BAD_REQUEST)
user = request.user
username = user.username
update_if_exists = request.GET.get('update_if_exists', 'true')
if update_if_exists not in ['true', 'false']:
return Response(
dict(exception=INVALID_UPDATE_IF_EXISTS),
status=status.HTTP_400_BAD_REQUEST
)
update_if_exists = update_if_exists == 'true'
data = data.decode('utf-8') if isinstance(data, bytes) else data
try:
task = queue_bulk_import(data, import_queue, username, update_if_exists, threads, inline)
except AlreadyQueued:
return Response(dict(exception=ALREADY_QUEUED), status=status.HTTP_409_CONFLICT)
parsed_task = parse_bulk_import_task_id(task.id)
return Response(
dict(task=task.id, state=task.state, username=username, queue=parsed_task['queue']),
status=status.HTTP_202_ACCEPTED
)
class BulkImportFileUploadView(APIView):
permission_classes = (IsAuthenticated, )
parser_classes = (MultiPartParser, )
@swagger_auto_schema(
manual_parameters=[update_if_exists_param, file_upload_param],
)
def post(self, request, import_queue=None):
file = request.data.get('file', None)
if not file:
return Response(dict(exception=NO_CONTENT_TO_IMPORT), status=status.HTTP_400_BAD_REQUEST)
return import_response(self.request, import_queue, file.read())
class BulkImportFileURLView(APIView):
permission_classes = (IsAuthenticated, )
parser_classes = (MultiPartParser, )
@swagger_auto_schema(
manual_parameters=[update_if_exists_param, file_url_param],
)
def post(self, request, import_queue=None):
file = None
try:
file = urllib.request.urlopen(request.data.get('file_url'))
except: # pylint: disable=bare-except
pass
if not file:
return Response(dict(exception=NO_CONTENT_TO_IMPORT), status=status.HTTP_400_BAD_REQUEST)
return import_response(self.request, import_queue, file.read())
class BulkImportView(APIView):
permission_classes = (IsAuthenticated,)
@swagger_auto_schema(
manual_parameters=[update_if_exists_param],
request_body=openapi.Schema(type=openapi.TYPE_OBJECT)
)
def post(self, request, import_queue=None):
return import_response(self.request, import_queue, request.body)
@swagger_auto_schema(manual_parameters=[task_param, result_param, username_param])
def get(
self, request, import_queue=None
): # pylint: disable=too-many-return-statements,too-many-locals,too-many-branches
task_id = request.GET.get('task')
result_format = request.GET.get('result')
username = request.GET.get('username')
user = self.request.user
if task_id:
parsed_task = parse_bulk_import_task_id(task_id)
username = parsed_task['username']
if not user.is_staff and user.username != username:
return Response(status=status.HTTP_403_FORBIDDEN)
task = AsyncResult(task_id)
if task.successful():
result = task.get()
if result and result_format == 'json':
return Response(result.get('json', None), content_type="application/json")
if result and result_format == 'report':
return Response(result.get('report', None))
if result:
return Response(result.get('detailed_summary', None))
if task.failed():
return Response(dict(exception=str(task.result)), status=status.HTTP_400_BAD_REQUEST)
if task.state == 'STARTED':
service = RedisService()
if service.exists(task_id):
return Response(
dict(
details=service.get_formatted(task_id), task=task.id, state=task.state,
username=username, queue=parsed_task['queue']
),
status=status.HTTP_200_OK
)
if task.state == 'PENDING' and not task_exists(task_id):
return Response(dict(exception='task ' + task_id + ' not found'), status=status.HTTP_404_NOT_FOUND)
return Response(
dict(task=task.id, state=task.state, username=username, queue=parsed_task['queue']),
status=status.HTTP_202_ACCEPTED
)
flower_tasks = flower_get('api/tasks').json()
tasks = []
for task_id, value in flower_tasks.items():
if not value['name'].startswith('core.common.tasks.bulk_import'):
continue
task = parse_bulk_import_task_id(task_id)
if user.is_staff or user.username == task['username']:
if (not import_queue or task['queue'] == import_queue) and \
(not username or task['username'] == username):
tasks.append(
dict(task=task_id, state=value['state'], queue=task['queue'], username=task['username'])
)
return Response(tasks)
class BulkImportParallelInlineView(APIView): # pragma: no cover
permission_classes = (IsAuthenticated, )
parser_classes = (MultiPartParser, )
@swagger_auto_schema(
manual_parameters=[update_if_exists_param, file_url_param, file_upload_param, parallel_threads_param],
)
def post(self, request, import_queue=None):
parallel_threads = request.data.get('parallel') or 5
file = None
try:
if 'file' in request.data:
file = request.data['file']
elif 'file_url' in request.data:
file = urllib.request.urlopen(request.data['file_url'])
except: # pylint: disable=bare-except
pass
if not file:
return Response(dict(exception=NO_CONTENT_TO_IMPORT), status=status.HTTP_400_BAD_REQUEST)
return import_response(self.request, import_queue, file.read(), parallel_threads, True)
class BulkImportInlineView(APIView): # pragma: no cover
permission_classes = (IsAuthenticated, )
parser_classes = (MultiPartParser, )
@swagger_auto_schema(
manual_parameters=[update_if_exists_param, file_url_param, file_upload_param],
)
def post(self, request, import_queue=None):
file = None
try:
if 'file' in request.data:
file = request.data['file']
elif 'file_url' in request.data:
file = urllib.request.urlopen(request.data['file_url'])
except: # pylint: disable=bare-except
pass
if not file:
return Response(dict(exception=NO_CONTENT_TO_IMPORT), status=status.HTTP_400_BAD_REQUEST)
return import_response(self.request, import_queue, file.read(), None, True)
class BaseESIndexView(APIView): # pragma: no cover
permission_classes = (IsSuperuser,)
parser_classes = (MultiPartParser,)
task = None
@swagger_auto_schema(manual_parameters=[apps_param])
def post(self, request):
apps = request.data.get('apps', None)
if apps:
apps = apps.split(',')
result = self.task.delay(apps)
return Response(
dict(state=result.state, username=self.request.user.username, task=result.task_id, queue='default'),
status=status.HTTP_202_ACCEPTED
)
class RebuildESIndexView(BaseESIndexView): # pragma: no cover
task = rebuild_indexes
class PopulateESIndexView(BaseESIndexView): # pragma: no cover
task = populate_indexes