123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115 |
- # Licensed to the Apache Software Foundation (ASF) under one
- # or more contributor license agreements. See the NOTICE file
- # distributed with this work for additional information
- # regarding copyright ownership. The ASF licenses this file
- # to you under the Apache License, Version 2.0 (the
- # "License"); you may not use this file except in compliance
- # with the License. You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing,
- # software distributed under the License is distributed on an
- # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- # KIND, either express or implied. See the License for the
- # specific language governing permissions and limitations
- # under the License.
- import pandas as pd
- from sqlalchemy import BigInteger, Date, DateTime, String
- from superset import db
- from superset.models.slice import Slice
- from superset.utils.core import get_example_database
- from .helpers import (
- config,
- get_example_data,
- get_slice_json,
- merge_slice,
- misc_dash_slices,
- TBL,
- )
- def load_multiformat_time_series(only_metadata=False, force=False):
- """Loading time series data from a zip file in the repo"""
- tbl_name = "multiformat_time_series"
- database = get_example_database()
- table_exists = database.has_table_by_name(tbl_name)
- if not only_metadata and (not table_exists or force):
- data = get_example_data("multiformat_time_series.json.gz")
- pdf = pd.read_json(data)
- pdf.ds = pd.to_datetime(pdf.ds, unit="s")
- pdf.ds2 = pd.to_datetime(pdf.ds2, unit="s")
- pdf.to_sql(
- tbl_name,
- database.get_sqla_engine(),
- if_exists="replace",
- chunksize=500,
- dtype={
- "ds": Date,
- "ds2": DateTime,
- "epoch_s": BigInteger,
- "epoch_ms": BigInteger,
- "string0": String(100),
- "string1": String(100),
- "string2": String(100),
- "string3": String(100),
- },
- index=False,
- )
- print("Done loading table!")
- print("-" * 80)
- print(f"Creating table [{tbl_name}] reference")
- obj = db.session.query(TBL).filter_by(table_name=tbl_name).first()
- if not obj:
- obj = TBL(table_name=tbl_name)
- obj.main_dttm_col = "ds"
- obj.database = database
- dttm_and_expr_dict = {
- "ds": [None, None],
- "ds2": [None, None],
- "epoch_s": ["epoch_s", None],
- "epoch_ms": ["epoch_ms", None],
- "string2": ["%Y%m%d-%H%M%S", None],
- "string1": ["%Y-%m-%d^%H:%M:%S", None],
- "string0": ["%Y-%m-%d %H:%M:%S.%f", None],
- "string3": ["%Y/%m/%d%H:%M:%S.%f", None],
- }
- for col in obj.columns:
- dttm_and_expr = dttm_and_expr_dict[col.column_name]
- col.python_date_format = dttm_and_expr[0]
- col.dbatabase_expr = dttm_and_expr[1]
- col.is_dttm = True
- db.session.merge(obj)
- db.session.commit()
- obj.fetch_metadata()
- tbl = obj
- print("Creating Heatmap charts")
- for i, col in enumerate(tbl.columns):
- slice_data = {
- "metrics": ["count"],
- "granularity_sqla": col.column_name,
- "row_limit": config["ROW_LIMIT"],
- "since": "2015",
- "until": "2016",
- "viz_type": "cal_heatmap",
- "domain_granularity": "month",
- "subdomain_granularity": "day",
- }
- slc = Slice(
- slice_name=f"Calendar Heatmap multiformat {i}",
- viz_type="cal_heatmap",
- datasource_type="table",
- datasource_id=tbl.id,
- params=get_slice_json(slice_data),
- )
- merge_slice(slc)
- misc_dash_slices.add("Calendar Heatmap multiformat 0")
|