mirror of
https://github.com/apache/superset.git
synced 2024-09-17 11:09:47 -04:00
Druid support via SQLAlchemy (#4163)
* Use druiddb * Remove auto formatting * Show prequeries * Fix subtle bug with lists * Move arguments to query object * Fix druid run_query
This commit is contained in:
parent
d997a450cf
commit
686023c8dd
@ -1022,7 +1022,10 @@ class DruidDatasource(Model, BaseDatasource):
|
|||||||
orderby=None,
|
orderby=None,
|
||||||
extras=None, # noqa
|
extras=None, # noqa
|
||||||
columns=None, phase=2, client=None, form_data=None,
|
columns=None, phase=2, client=None, form_data=None,
|
||||||
order_desc=True):
|
order_desc=True,
|
||||||
|
prequeries=None,
|
||||||
|
is_prequery=False,
|
||||||
|
):
|
||||||
"""Runs a query against Druid and returns a dataframe.
|
"""Runs a query against Druid and returns a dataframe.
|
||||||
"""
|
"""
|
||||||
# TODO refactor into using a TBD Query object
|
# TODO refactor into using a TBD Query object
|
||||||
|
@ -370,6 +370,8 @@ class SqlaTable(Model, BaseDatasource):
|
|||||||
)
|
)
|
||||||
logging.info(sql)
|
logging.info(sql)
|
||||||
sql = sqlparse.format(sql, reindent=True)
|
sql = sqlparse.format(sql, reindent=True)
|
||||||
|
if query_obj['is_prequery']:
|
||||||
|
query_obj['prequeries'].append(sql)
|
||||||
return sql
|
return sql
|
||||||
|
|
||||||
def get_sqla_table(self):
|
def get_sqla_table(self):
|
||||||
@ -405,7 +407,10 @@ class SqlaTable(Model, BaseDatasource):
|
|||||||
extras=None,
|
extras=None,
|
||||||
columns=None,
|
columns=None,
|
||||||
form_data=None,
|
form_data=None,
|
||||||
order_desc=True):
|
order_desc=True,
|
||||||
|
prequeries=None,
|
||||||
|
is_prequery=False,
|
||||||
|
):
|
||||||
"""Querying any sqla table from this common interface"""
|
"""Querying any sqla table from this common interface"""
|
||||||
template_kwargs = {
|
template_kwargs = {
|
||||||
'from_dttm': from_dttm,
|
'from_dttm': from_dttm,
|
||||||
@ -564,6 +569,7 @@ class SqlaTable(Model, BaseDatasource):
|
|||||||
|
|
||||||
if is_timeseries and \
|
if is_timeseries and \
|
||||||
timeseries_limit and groupby and not time_groupby_inline:
|
timeseries_limit and groupby and not time_groupby_inline:
|
||||||
|
if self.database.db_engine_spec.inner_joins:
|
||||||
# some sql dialects require for order by expressions
|
# some sql dialects require for order by expressions
|
||||||
# to also be in the select clause -- others, e.g. vertica,
|
# to also be in the select clause -- others, e.g. vertica,
|
||||||
# require a unique inner alias
|
# require a unique inner alias
|
||||||
@ -592,9 +598,44 @@ class SqlaTable(Model, BaseDatasource):
|
|||||||
groupby_exprs[i] == column(gb + '__'))
|
groupby_exprs[i] == column(gb + '__'))
|
||||||
|
|
||||||
tbl = tbl.join(subq.alias(), and_(*on_clause))
|
tbl = tbl.join(subq.alias(), and_(*on_clause))
|
||||||
|
else:
|
||||||
|
# run subquery to get top groups
|
||||||
|
subquery_obj = {
|
||||||
|
'prequeries': prequeries,
|
||||||
|
'is_prequery': True,
|
||||||
|
'is_timeseries': False,
|
||||||
|
'row_limit': timeseries_limit,
|
||||||
|
'groupby': groupby,
|
||||||
|
'metrics': metrics,
|
||||||
|
'granularity': granularity,
|
||||||
|
'from_dttm': inner_from_dttm or from_dttm,
|
||||||
|
'to_dttm': inner_to_dttm or to_dttm,
|
||||||
|
'filter': filter,
|
||||||
|
'orderby': orderby,
|
||||||
|
'extras': extras,
|
||||||
|
'columns': columns,
|
||||||
|
'form_data': form_data,
|
||||||
|
'order_desc': True,
|
||||||
|
}
|
||||||
|
result = self.query(subquery_obj)
|
||||||
|
dimensions = [c for c in result.df.columns if c not in metrics]
|
||||||
|
top_groups = self._get_top_groups(result.df, dimensions)
|
||||||
|
qry = qry.where(top_groups)
|
||||||
|
|
||||||
return qry.select_from(tbl)
|
return qry.select_from(tbl)
|
||||||
|
|
||||||
|
def _get_top_groups(self, df, dimensions):
|
||||||
|
cols = {col.column_name: col for col in self.columns}
|
||||||
|
groups = []
|
||||||
|
for unused, row in df.iterrows():
|
||||||
|
group = []
|
||||||
|
for dimension in dimensions:
|
||||||
|
col_obj = cols.get(dimension)
|
||||||
|
group.append(col_obj.sqla_col == row[dimension])
|
||||||
|
groups.append(and_(*group))
|
||||||
|
|
||||||
|
return or_(*groups)
|
||||||
|
|
||||||
def query(self, query_obj):
|
def query(self, query_obj):
|
||||||
qry_start_dttm = datetime.now()
|
qry_start_dttm = datetime.now()
|
||||||
sql = self.get_query_str(query_obj)
|
sql = self.get_query_str(query_obj)
|
||||||
@ -609,6 +650,12 @@ class SqlaTable(Model, BaseDatasource):
|
|||||||
error_message = (
|
error_message = (
|
||||||
self.database.db_engine_spec.extract_error_message(e))
|
self.database.db_engine_spec.extract_error_message(e))
|
||||||
|
|
||||||
|
# if this is a main query with prequeries, combine them together
|
||||||
|
if not query_obj['is_prequery']:
|
||||||
|
query_obj['prequeries'].append(sql)
|
||||||
|
sql = ';\n\n'.join(query_obj['prequeries'])
|
||||||
|
sql += ';'
|
||||||
|
|
||||||
return QueryResult(
|
return QueryResult(
|
||||||
status=status,
|
status=status,
|
||||||
df=df,
|
df=df,
|
||||||
|
@ -62,6 +62,7 @@ class BaseEngineSpec(object):
|
|||||||
time_groupby_inline = False
|
time_groupby_inline = False
|
||||||
limit_method = LimitMethod.FETCH_MANY
|
limit_method = LimitMethod.FETCH_MANY
|
||||||
time_secondary_columns = False
|
time_secondary_columns = False
|
||||||
|
inner_joins = True
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def fetch_data(cls, cursor, limit):
|
def fetch_data(cls, cursor, limit):
|
||||||
@ -1229,6 +1230,7 @@ class DruidEngineSpec(BaseEngineSpec):
|
|||||||
"""Engine spec for Druid.io"""
|
"""Engine spec for Druid.io"""
|
||||||
engine = 'druid'
|
engine = 'druid'
|
||||||
limit_method = LimitMethod.FETCH_MANY
|
limit_method = LimitMethod.FETCH_MANY
|
||||||
|
inner_joins = False
|
||||||
|
|
||||||
|
|
||||||
engines = {
|
engines = {
|
||||||
|
@ -991,6 +991,12 @@ class Superset(BaseSupersetView):
|
|||||||
query = viz_obj.datasource.get_query_str(query_obj)
|
query = viz_obj.datasource.get_query_str(query_obj)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
return json_error_response(e)
|
return json_error_response(e)
|
||||||
|
|
||||||
|
if query_obj['prequeries']:
|
||||||
|
query_obj['prequeries'].append(query)
|
||||||
|
query = ';\n\n'.join(query_obj['prequeries'])
|
||||||
|
query += ';'
|
||||||
|
|
||||||
return Response(
|
return Response(
|
||||||
json.dumps({
|
json.dumps({
|
||||||
'query': query,
|
'query': query,
|
||||||
|
@ -205,6 +205,8 @@ class BaseViz(object):
|
|||||||
'timeseries_limit_metric': timeseries_limit_metric,
|
'timeseries_limit_metric': timeseries_limit_metric,
|
||||||
'form_data': form_data,
|
'form_data': form_data,
|
||||||
'order_desc': order_desc,
|
'order_desc': order_desc,
|
||||||
|
'prequeries': [],
|
||||||
|
'is_prequery': False,
|
||||||
}
|
}
|
||||||
return d
|
return d
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user