perf: dataset exec 缓存优化

This commit is contained in:
caoqianming 2023-05-31 09:19:32 +08:00
parent 8a1d43b7b8
commit ae071d1d9c
5 changed files with 35 additions and 17 deletions

View File

@ -0,0 +1,18 @@
# Generated by Django 3.2.12 on 2023-05-31 01:09
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('bi', '0001_initial'),
]
operations = [
migrations.AddField(
model_name='dataset',
name='cache_seconds',
field=models.PositiveIntegerField(blank=True, default=10, verbose_name='缓存秒数'),
),
]

View File

@ -8,6 +8,7 @@ class Dataset(CommonBDModel):
description = models.TextField('描述说明', default='', blank=True)
sql_query = models.TextField('sql查询语句', default='', blank=True)
echart_options = models.TextField(default='', blank=True)
cache_seconds = models.PositiveIntegerField('缓存秒数', default=10, blank=True)
# class Report(CommonBDModel):

View File

@ -30,5 +30,4 @@ class DatasetSerializer(CustomModelSerializer):
class DataExecSerializer(serializers.Serializer):
query = serializers.JSONField(label="查询字典参数", required=False, allow_null=True)
return_type = serializers.IntegerField(label="返回格式", required=False, default=2)
query = serializers.JSONField(label="查询字典参数", required=False, allow_null=True)

View File

@ -27,39 +27,42 @@ class DatasetViewSet(CustomModelViewSet):
dt = self.get_object()
rdata = DatasetSerializer(instance=dt).data
query = request.data.get('query', {})
return_type = request.data.get('return_type', 2)
query['r_user'] = request.user.id
query['r_dept'] = request.user.belong_dept.id if request.user.belong_dept else ''
results = {}
seconds = 10
results2 = {}
can_cache = True
if dt.sql_query:
sql_f_ = check_sql_safe(dt.sql_query.format(**query))
sql_f_l = sql_f_.strip(';').split(';')
hash_k = hash(sql_f_.strip(';'))
hash_v = cache.get(hash_k, None)
if hash_v:
return Response(hash_v)
with concurrent.futures.ThreadPoolExecutor(max_workers=6) as executor: # 多线程运行并返回字典结果
fun_ps = []
for ind, val in enumerate(sql_f_l):
res = cache.get(val, None)
if isinstance(res, tuple):
results[f'ds{ind}'] = format_sqldata(res[0], res[1], return_type)
else:
fun_ps.append((f'ds{ind}', execute_raw_sql, val))
fun_ps.append((f'ds{ind}', execute_raw_sql, val))
# 生成执行函数
futures = {executor.submit(i[1], i[2]): i for i in fun_ps}
for future in concurrent.futures.as_completed(futures):
name, *_, sql_f = futures[future] # 获取对应的键
try:
res = future.result()
results[name] = format_sqldata(res[0], res[1], return_type)
if seconds:
cache.set(sql_f, res, seconds)
results[name], results2[name]= format_sqldata(res[0], res[1])
except Exception as e:
results[name] = 'error: ' + str(e)
can_cache = False
rdata['data'] = results
rdata['data2'] = results2
if rdata['echart_options']:
for key in results:
if isinstance(results[key], str):
raise ParseError(results[key])
rdata['echart_options'] = format_json_with_placeholders(rdata['echart_options'], **results)
if results and can_cache:
cache.set(hash_k, rdata, dt.cache_seconds)
return Response(rdata)
@action(methods=['get'], detail=False, perms_map={'get': '*'})

View File

@ -17,11 +17,8 @@ def execute_raw_sql(sql: str, params=None):
rows = cursor.fetchall()
return columns, rows
def format_sqldata(columns, rows, return_type=2):
if return_type == 2:
return [columns] + rows
elif return_type == 1:
return [dict(zip(columns, row)) for row in rows]
def format_sqldata(columns, rows):
return [columns] + rows, [dict(zip(columns, row)) for row in rows]
def query_all_dict(sql, params=None):