diff --git a/requirements/base.txt b/requirements/base.txt
index 3a5ec607fe9f..b41ea6cb6242 100644
--- a/requirements/base.txt
+++ b/requirements/base.txt
@@ -294,8 +294,6 @@ wtforms==2.3.3
# wtforms-json
wtforms-json==0.3.3
# via apache-superset
-xlsxwriter==3.0.7
- # via apache-superset
# The following packages are considered to be unsafe in a requirements file:
# setuptools
diff --git a/setup.py b/setup.py
index c6850070a0a7..53565f1c764f 100644
--- a/setup.py
+++ b/setup.py
@@ -125,7 +125,6 @@ def get_git_sha() -> str:
"waitress; sys_platform == 'win32'",
"wtforms>=2.3.3, <2.4",
"wtforms-json",
- "xlsxwriter>=3.0.7, <3.1",
],
extras_require={
"athena": ["pyathena[pandas]>=2, <3"],
diff --git a/superset-frontend/src/explore/components/useExploreAdditionalActionsMenu/index.jsx b/superset-frontend/src/explore/components/useExploreAdditionalActionsMenu/index.jsx
index 445db6dc4414..62fcdaaf15ac 100644
--- a/superset-frontend/src/explore/components/useExploreAdditionalActionsMenu/index.jsx
+++ b/superset-frontend/src/explore/components/useExploreAdditionalActionsMenu/index.jsx
@@ -41,7 +41,6 @@ const MENU_KEYS = {
EXPORT_TO_CSV: 'export_to_csv',
EXPORT_TO_CSV_PIVOTED: 'export_to_csv_pivoted',
EXPORT_TO_JSON: 'export_to_json',
- EXPORT_TO_XLSX: 'export_to_xlsx',
DOWNLOAD_AS_IMAGE: 'download_as_image',
SHARE_SUBMENU: 'share_submenu',
COPY_PERMALINK: 'copy_permalink',
@@ -166,16 +165,6 @@ export const useExploreAdditionalActionsMenu = (
[latestQueryFormData],
);
- const exportExcel = useCallback(
- () =>
- exportChart({
- formData: latestQueryFormData,
- resultType: 'results',
- resultFormat: 'xlsx',
- }),
- [latestQueryFormData],
- );
-
const copyLink = useCallback(async () => {
try {
if (!latestQueryFormData) {
@@ -210,11 +199,6 @@ export const useExploreAdditionalActionsMenu = (
setIsDropdownVisible(false);
setOpenSubmenus([]);
- break;
- case MENU_KEYS.EXPORT_TO_XLSX:
- exportExcel();
- setIsDropdownVisible(false);
- setOpenSubmenus([]);
break;
case MENU_KEYS.DOWNLOAD_AS_IMAGE:
downloadAsImage(
@@ -328,12 +312,6 @@ export const useExploreAdditionalActionsMenu = (
>
{t('Download as image')}
-
}
- >
- {t('Export to Excel')}
-
diff --git a/superset/charts/data/api.py b/superset/charts/data/api.py
index 0d0758819ed0..152383e0c66d 100644
--- a/superset/charts/data/api.py
+++ b/superset/charts/data/api.py
@@ -46,7 +46,7 @@
from superset.extensions import event_logger
from superset.utils.async_query_manager import AsyncQueryTokenException
from superset.utils.core import create_zip, get_user_id, json_int_dttm_ser
-from superset.views.base import CsvResponse, generate_download_headers, XlsxResponse
+from superset.views.base import CsvResponse, generate_download_headers
from superset.views.base_api import statsd_metrics
if TYPE_CHECKING:
@@ -353,34 +353,24 @@ def _send_chart_response(
if result_type == ChartDataResultType.POST_PROCESSED:
result = apply_post_process(result, form_data, datasource)
- if result_format in ChartDataResultFormat.table_like():
- # Verify user has permission to export file
+ if result_format == ChartDataResultFormat.CSV:
+ # Verify user has permission to export CSV file
if not security_manager.can_access("can_csv", "Superset"):
return self.response_403()
if not result["queries"]:
return self.response_400(_("Empty query result"))
- is_csv_format = result_format == ChartDataResultFormat.CSV
-
if len(result["queries"]) == 1:
- # return single query results
+ # return single query results csv format
data = result["queries"][0]["data"]
- if is_csv_format:
- return CsvResponse(data, headers=generate_download_headers("csv"))
-
- return XlsxResponse(data, headers=generate_download_headers("xlsx"))
-
- # return multi-query results bundled as a zip file
- def _process_data(query_data: Any) -> Any:
- if result_format == ChartDataResultFormat.CSV:
- encoding = current_app.config["CSV_EXPORT"].get("encoding", "utf-8")
- return query_data.encode(encoding)
- return query_data
+ return CsvResponse(data, headers=generate_download_headers("csv"))
+ # return multi-query csv results bundled as a zip file
+ encoding = current_app.config["CSV_EXPORT"].get("encoding", "utf-8")
files = {
- f"query_{idx + 1}.{result_format}": _process_data(query["data"])
- for idx, query in enumerate(result["queries"])
+ f"query_{idx + 1}.csv": result["data"].encode(encoding)
+ for idx, result in enumerate(result["queries"])
}
return Response(
create_zip(files),
diff --git a/superset/common/chart_data.py b/superset/common/chart_data.py
index 659a64015937..ea31d4f13817 100644
--- a/superset/common/chart_data.py
+++ b/superset/common/chart_data.py
@@ -15,7 +15,6 @@
# specific language governing permissions and limitations
# under the License.
from enum import Enum
-from typing import Set
class ChartDataResultFormat(str, Enum):
@@ -25,11 +24,6 @@ class ChartDataResultFormat(str, Enum):
CSV = "csv"
JSON = "json"
- XLSX = "xlsx"
-
- @classmethod
- def table_like(cls) -> Set["ChartDataResultFormat"]:
- return {cls.CSV} | {cls.XLSX}
class ChartDataResultType(str, Enum):
diff --git a/superset/common/query_context_processor.py b/superset/common/query_context_processor.py
index 703e1d71ddea..a2616013c905 100644
--- a/superset/common/query_context_processor.py
+++ b/superset/common/query_context_processor.py
@@ -46,7 +46,7 @@
from superset.extensions import cache_manager, security_manager
from superset.models.helpers import QueryResult
from superset.models.sql_lab import Query
-from superset.utils import csv, excel
+from superset.utils import csv
from superset.utils.cache import generate_cache_key, set_and_log_cache
from superset.utils.core import (
DatasourceType,
@@ -448,20 +448,15 @@ def processing_time_offsets( # pylint: disable=too-many-locals,too-many-stateme
return CachedTimeOffset(df=rv_df, queries=queries, cache_keys=cache_keys)
def get_data(self, df: pd.DataFrame) -> Union[str, List[Dict[str, Any]]]:
- if self._query_context.result_format in ChartDataResultFormat.table_like():
+ if self._query_context.result_format == ChartDataResultFormat.CSV:
include_index = not isinstance(df.index, pd.RangeIndex)
columns = list(df.columns)
verbose_map = self._qc_datasource.data.get("verbose_map", {})
if verbose_map:
df.columns = [verbose_map.get(column, column) for column in columns]
-
- result = None
- if self._query_context.result_format == ChartDataResultFormat.CSV:
- result = csv.df_to_escaped_csv(
- df, index=include_index, **config["CSV_EXPORT"]
- )
- elif self._query_context.result_format == ChartDataResultFormat.XLSX:
- result = excel.df_to_excel(df, **config["EXCEL_EXPORT"])
+ result = csv.df_to_escaped_csv(
+ df, index=include_index, **config["CSV_EXPORT"]
+ )
return result or ""
return df.to_dict(orient="records")
diff --git a/superset/config.py b/superset/config.py
index c9234e521d75..d77dd81507fd 100644
--- a/superset/config.py
+++ b/superset/config.py
@@ -751,11 +751,6 @@ def _try_json_readsha(filepath: str, length: int) -> Optional[str]:
# note: index option should not be overridden
CSV_EXPORT = {"encoding": "utf-8"}
-# Excel Options: key/value pairs that will be passed as argument to DataFrame.to_excel
-# method.
-# note: index option should not be overridden
-EXCEL_EXPORT = {"encoding": "utf-8"}
-
# ---------------------------------------------------
# Time grain configurations
# ---------------------------------------------------
diff --git a/superset/utils/excel.py b/superset/utils/excel.py
deleted file mode 100644
index 1f68031b6497..000000000000
--- a/superset/utils/excel.py
+++ /dev/null
@@ -1,29 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-import io
-from typing import Any
-
-import pandas as pd
-
-
-def df_to_excel(df: pd.DataFrame, **kwargs: Any) -> Any:
- output = io.BytesIO()
- # pylint: disable=abstract-class-instantiated
- with pd.ExcelWriter(output, engine="xlsxwriter") as writer:
- df.to_excel(writer, **kwargs)
-
- return output.getvalue()
diff --git a/superset/views/base.py b/superset/views/base.py
index ec74b8ccdb3a..487c9c14db48 100644
--- a/superset/views/base.py
+++ b/superset/views/base.py
@@ -693,17 +693,6 @@ class CsvResponse(Response):
default_mimetype = "text/csv"
-class XlsxResponse(Response):
- """
- Override Response to use xlsx mimetype
- """
-
- charset = "utf-8"
- default_mimetype = (
- "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
- )
-
-
def bind_field(
_: Any, form: DynamicForm, unbound_field: UnboundField, options: Dict[Any, Any]
) -> Field:
diff --git a/tests/integration_tests/charts/data/api_tests.py b/tests/integration_tests/charts/data/api_tests.py
index 2818793af017..ed8de062d727 100644
--- a/tests/integration_tests/charts/data/api_tests.py
+++ b/tests/integration_tests/charts/data/api_tests.py
@@ -260,16 +260,6 @@ def test_empty_request_with_csv_result_format(self):
rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
assert rv.status_code == 400
- @pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
- def test_empty_request_with_excel_result_format(self):
- """
- Chart data API: Test empty chart data with Excel result format
- """
- self.query_context_payload["result_format"] = "xlsx"
- self.query_context_payload["queries"] = []
- rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
- assert rv.status_code == 400
-
@pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
def test_with_csv_result_format(self):
"""
@@ -280,17 +270,6 @@ def test_with_csv_result_format(self):
assert rv.status_code == 200
assert rv.mimetype == "text/csv"
- @pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
- def test_with_excel_result_format(self):
- """
- Chart data API: Test chart data with Excel result format
- """
- self.query_context_payload["result_format"] = "xlsx"
- mimetype = "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
- rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
- assert rv.status_code == 200
- assert rv.mimetype == mimetype
-
@pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
def test_with_multi_query_csv_result_format(self):
"""
@@ -306,21 +285,6 @@ def test_with_multi_query_csv_result_format(self):
zipfile = ZipFile(BytesIO(rv.data), "r")
assert zipfile.namelist() == ["query_1.csv", "query_2.csv"]
- @pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
- def test_with_multi_query_excel_result_format(self):
- """
- Chart data API: Test chart data with multi-query Excel result format
- """
- self.query_context_payload["result_format"] = "xlsx"
- self.query_context_payload["queries"].append(
- self.query_context_payload["queries"][0]
- )
- rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
- assert rv.status_code == 200
- assert rv.mimetype == "application/zip"
- zipfile = ZipFile(BytesIO(rv.data), "r")
- assert zipfile.namelist() == ["query_1.xlsx", "query_2.xlsx"]
-
@pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
def test_with_csv_result_format_when_actor_not_permitted_for_csv__403(self):
"""
@@ -333,18 +297,6 @@ def test_with_csv_result_format_when_actor_not_permitted_for_csv__403(self):
rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
assert rv.status_code == 403
- @pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
- def test_with_excel_result_format_when_actor_not_permitted_for_excel__403(self):
- """
- Chart data API: Test chart data with Excel result format
- """
- self.logout()
- self.login(username="gamma_no_csv")
- self.query_context_payload["result_format"] = "xlsx"
-
- rv = self.post_assert_metric(CHART_DATA_URI, self.query_context_payload, "data")
- assert rv.status_code == 403
-
@pytest.mark.usefixtures("load_birth_names_dashboard_with_slices")
def test_with_row_limit_and_offset__row_limit_and_offset_were_applied(self):
"""