superset/tests/unit_tests/db_engine_specs/test_sqlite.py

147 lines
5.9 KiB
Python

# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
# pylint: disable=invalid-name, unused-argument, import-outside-toplevel, redefined-outer-name
from datetime import datetime
from unittest import mock
import pytest
from sqlalchemy.engine import create_engine
from tests.unit_tests.fixtures.common import dttm
def test_convert_dttm(dttm: datetime) -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
assert SqliteEngineSpec.convert_dttm("TEXT", dttm) == "'2019-01-02 03:04:05.678900'"
def test_convert_dttm_lower(dttm: datetime) -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
assert SqliteEngineSpec.convert_dttm("text", dttm) == "'2019-01-02 03:04:05.678900'"
def test_convert_dttm_invalid_type(dttm: datetime) -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
assert SqliteEngineSpec.convert_dttm("other", dttm) is None
def test_get_all_datasource_names_table() -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
database = mock.MagicMock()
database.get_all_schema_names.return_value = ["schema1"]
table_names = [("table1", "schema1"), ("table2", "schema1")]
get_tables = mock.MagicMock(return_value=table_names)
database.get_all_table_names_in_schema = get_tables
result = SqliteEngineSpec.get_all_datasource_names(database, "table")
assert result == table_names
get_tables.assert_called_once_with(
schema="schema1",
force=True,
cache=database.table_cache_enabled,
cache_timeout=database.table_cache_timeout,
)
def test_get_all_datasource_names_view() -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
database = mock.MagicMock()
database.get_all_schema_names.return_value = ["schema1"]
views_names = [("view1", "schema1"), ("view2", "schema1")]
get_views = mock.MagicMock(return_value=views_names)
database.get_all_view_names_in_schema = get_views
result = SqliteEngineSpec.get_all_datasource_names(database, "view")
assert result == views_names
get_views.assert_called_once_with(
schema="schema1",
force=True,
cache=database.table_cache_enabled,
cache_timeout=database.table_cache_timeout,
)
def test_get_all_datasource_names_invalid_type() -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
database = mock.MagicMock()
database.get_all_schema_names.return_value = ["schema1"]
invalid_type = "asdf"
with pytest.raises(Exception):
SqliteEngineSpec.get_all_datasource_names(database, invalid_type)
@pytest.mark.parametrize(
"dttm,grain,expected",
[
("2022-05-04T05:06:07.89Z", "PT1S", "2022-05-04 05:06:07"),
("2022-05-04T05:06:07.89Z", "PT1M", "2022-05-04 05:06:00"),
("2022-05-04T05:06:07.89Z", "PT1H", "2022-05-04 05:00:00"),
("2022-05-04T05:06:07.89Z", "P1D", "2022-05-04 00:00:00"),
("2022-05-04T05:06:07.89Z", "P1W", "2022-05-01 00:00:00"),
("2022-05-04T05:06:07.89Z", "P1M", "2022-05-01 00:00:00"),
("2022-05-04T05:06:07.89Z", "P1Y", "2022-01-01 00:00:00"),
# ___________________________
# | May 2022 |
# |---------------------------|
# | S | M | T | W | T | F | S |
# |---+---+---+---+---+---+---|
# | 1 | 2 | 3 | 4 | 5 | 6 | 7 |
# ---------------------------
# week ending Saturday
("2022-05-04T05:06:07.89Z", "P1W/1970-01-03T00:00:00Z", "2022-05-07 00:00:00"),
# week ending Sunday
("2022-05-04T05:06:07.89Z", "P1W/1970-01-04T00:00:00Z", "2022-05-08 00:00:00"),
# week starting Sunday
("2022-05-04T05:06:07.89Z", "1969-12-28T00:00:00Z/P1W", "2022-05-01 00:00:00"),
# week starting Monday
("2022-05-04T05:06:07.89Z", "1969-12-29T00:00:00Z/P1W", "2022-05-02 00:00:00"),
# tests for quarter
("2022-01-04T05:06:07.89Z", "P3M", "2022-01-01 00:00:00"),
("2022-02-04T05:06:07.89Z", "P3M", "2022-01-01 00:00:00"),
("2022-03-04T05:06:07.89Z", "P3M", "2022-01-01 00:00:00"),
("2022-04-04T05:06:07.89Z", "P3M", "2022-04-01 00:00:00"),
("2022-05-04T05:06:07.89Z", "P3M", "2022-04-01 00:00:00"),
("2022-06-04T05:06:07.89Z", "P3M", "2022-04-01 00:00:00"),
("2022-07-04T05:06:07.89Z", "P3M", "2022-07-01 00:00:00"),
("2022-08-04T05:06:07.89Z", "P3M", "2022-07-01 00:00:00"),
("2022-09-04T05:06:07.89Z", "P3M", "2022-07-01 00:00:00"),
("2022-10-04T05:06:07.89Z", "P3M", "2022-10-01 00:00:00"),
("2022-11-04T05:06:07.89Z", "P3M", "2022-10-01 00:00:00"),
("2022-12-04T05:06:07.89Z", "P3M", "2022-10-01 00:00:00"),
],
)
def test_time_grain_expressions(dttm: str, grain: str, expected: str) -> None:
from superset.db_engine_specs.sqlite import SqliteEngineSpec
engine = create_engine("sqlite://")
connection = engine.connect()
connection.execute("CREATE TABLE t (dttm DATETIME)")
connection.execute("INSERT INTO t VALUES (?)", dttm)
# pylint: disable=protected-access
expression = SqliteEngineSpec._time_grain_expressions[grain].format(col="dttm")
sql = f"SELECT {expression} FROM t"
result = connection.execute(sql).scalar()
assert result == expected