优化 sql : SELECT /*aggregation_path_type=single_agg*/ `company_code`, `company_name`, `sale_id`, `sale_name`, round(sum(`checkin_uv`), 2) AS peopleSum, round(sum(`cash_pay`), 2) AS cashPay, round(sum(`account_pay`), 2) AS accountPay, round(sum(`salecard_pay`), 2) AS saleCardPay, round(sum(`online_pay`), 2) AS onlinePay, round(sum(`pay_amount_final`), 2) AS payAmount, round(sum(`total_original_price`), 2) AS originalPrice, CASE WHEN sum(`total_original_price`) = 0 THEN 0 ELSE Round( sum(`pay_amount_final`) / sum(`total_original_price`) * 100, 2 ) END AS discount, CASE WHEN sum(`checkin_uv`) = 0 THEN 0 ELSE Round(sum(`pay_amount_final`) / sum(`checkin_uv`), 2) END AS priceAvg, round(SUM(`settlement_pice_exclude_add`), 2) AS saleExcludeAdd, round(SUM(`original_pice_exclude_add`), 2) AS originalExcludeAdd, CASE WHEN sum(`total_original_price`) = 0 THEN 0 ELSE Round( ( SUM(`settlement_pice_exclude_add`) / sum(`total_original_price`) * 100 ), 2 ) END AS excludeAddDis, CASE WHEN sum(`checkin_uv`) = 0 THEN 0 ELSE Round( SUM(`settlement_pice_exclude_add`) / sum(`checkin_uv`), 2 ) END AS excludeAddPriceAvg FROM `salesman_confim_revenue_v2` WHERE sale_region_code = '8' and DATE_FORMAT(checkin_time, '%Y-%m-%d') >= DATE_FORMAT('2023-04-01', '%Y-%m-%d') and DATE_FORMAT(checkin_time, '%Y-%m-%d') <= DATE_FORMAT('2023-05-22', '%Y-%m-%d') GROUP BY `company_code`, `sale_id` ORDER BY `company_code` DESC

时间: 2024-01-19 07:03:09 浏览: 24
可以尝试对该 SQL 进行如下优化: 1. 使用索引:在 `salesman_confim_revenue_v2` 表中创建 `sale_region_code` 和 `checkin_time` 字段的联合索引,可以提高查询效率。 2. 使用 BETWEEN 代替两个 DATE_FORMAT:可以将 `DATE_FORMAT(checkin_time, '%Y-%m-%d') >= DATE_FORMAT('2023-04-01', '%Y-%m-%d')` 和 `DATE_FORMAT(checkin_time, '%Y-%m-%d') <= DATE_FORMAT('2023-05-22', '%Y-%m-%d')` 替换为 `checkin_time BETWEEN '2023-04-01' AND '2023-05-22'`,可以减少函数的使用,提高查询效率。 3. 减少聚合函数的使用:可以将 `sum` 和 `round` 函数的使用减少,避免重复计算,提高查询效率。 优化后的 SQL 如下: ``` SELECT /*aggregation_path_type=single_agg*/ `company_code`, `company_name`, `sale_id`, `sale_name`, ROUND(SUM(`checkin_uv`), 2) AS peopleSum, ROUND(SUM(`cash_pay`), 2) AS cashPay, ROUND(SUM(`account_pay`), 2) AS accountPay, ROUND(SUM(`salecard_pay`), 2) AS saleCardPay, ROUND(SUM(`online_pay`), 2) AS onlinePay, ROUND(SUM(`pay_amount_final`), 2) AS payAmount, ROUND(SUM(`total_original_price`), 2) AS originalPrice, CASE WHEN SUM(`total_original_price`) = 0 THEN 0 ELSE ROUND(SUM(`pay_amount_final`) / SUM(`total_original_price`) * 100, 2) END AS discount, CASE WHEN SUM(`checkin_uv`) = 0 THEN 0 ELSE ROUND(SUM(`pay_amount_final`) / SUM(`checkin_uv`), 2) END AS priceAvg, ROUND(SUM(`settlement_pice_exclude_add`), 2) AS saleExcludeAdd, ROUND(SUM(`original_pice_exclude_add`), 2) AS originalExcludeAdd, CASE WHEN SUM(`total_original_price`) = 0 THEN 0 ELSE ROUND((SUM(`settlement_pice_exclude_add`) / SUM(`total_original_price`) * 100), 2) END AS excludeAddDis, CASE WHEN SUM(`checkin_uv`) = 0 THEN 0 ELSE ROUND(SUM(`settlement_pice_exclude_add`) / SUM(`checkin_uv`), 2) END AS excludeAddPriceAvg FROM `salesman_confim_revenue_v2` WHERE sale_region_code = '8' AND checkin_time BETWEEN '2023-04-01' AND '2023-05-22' GROUP BY `company_code`, `sale_id` ORDER BY `company_code` DESC ```

相关推荐

class TemporalBlock(nn.Module): """ Temporal block with the following layers: - 2x3x3, 1x3x3, spatio-temporal pyramid pooling - dropout - skip connection. """ def __init__(self, in_channels, out_channels=None, use_pyramid_pooling=False, pool_sizes=None): super().__init__() self.in_channels = in_channels self.half_channels = in_channels // 2 self.out_channels = out_channels or self.in_channels self.kernels = [(2, 3, 3), (1, 3, 3)] # Flag for spatio-temporal pyramid pooling self.use_pyramid_pooling = use_pyramid_pooling # 3 convolution paths: 2x3x3, 1x3x3, 1x1x1 self.convolution_paths = [] for kernel_size in self.kernels: self.convolution_paths.append( nn.Sequential( conv_1x1x1_norm_activated(self.in_channels, self.half_channels), CausalConv3d(self.half_channels, self.half_channels, kernel_size=kernel_size), ) ) self.convolution_paths.append(conv_1x1x1_norm_activated(self.in_channels, self.half_channels)) self.convolution_paths = nn.ModuleList(self.convolution_paths) agg_in_channels = len(self.convolution_paths) * self.half_channels if self.use_pyramid_pooling: assert pool_sizes is not None, "setting must contain the list of kernel_size, but is None." reduction_channels = self.in_channels // 3 self.pyramid_pooling = PyramidSpatioTemporalPooling(self.in_channels, reduction_channels, pool_sizes) agg_in_channels += len(pool_sizes) * reduction_channels # Feature aggregation self.aggregation = nn.Sequential( conv_1x1x1_norm_activated(agg_in_channels, self.out_channels),) if self.out_channels != self.in_channels: self.projection = nn.Sequential( nn.Conv3d(self.in_channels, self.out_channels, kernel_size=1, bias=False), nn.BatchNorm3d(self.out_channels), ) else: self.projection = None网络结构是什么?

arser = argparse.ArgumentParser(description="Run GHCN.") parser.add_argument('--data_path', type=str, default='./data/', help='Input data path') parser.add_argument('--model_path', type=str, default='checkpoint.pt', help='Saved model path.') parser.add_argument('--dataset', type=str, default='Cora', help='Choose a dataset from {Cora, CiteSeer, PubMed}') parser.add_argument('--split', type=str, default='full', help='The type of dataset split {public, full, random}') parser.add_argument('--trim_prob', type=float, default=0.2, help='The probability to trim adj, 0 not trim, 1 trim') parser.add_argument('--seed', type=int, default=123, help='Random seed') parser.add_argument('--epoch', type=int, default=1000, help='Number of epochs to train') parser.add_argument('--lr', type=float, default=0.005, help='Initial learning rate') parser.add_argument('--weight_decay', type=float, default=5e-4, help='Weight decay (L2 norm on parameters)') parser.add_argument('--k', type=int, default=10, help='k-hop aggregation') parser.add_argument('--hidden', type=int, default=64, help='Number of hidden units') parser.add_argument('--dropout', type=float, default=0.7, help='Dropout rate') parser.add_argument('--patience', type=int, default=100, help='How long to wait after last time validation improved') args = parser.parse_args() for arg in vars(args): print('{0} = {1}'.format(arg, getattr(args, arg))) 修改代码要求:如果dataset不等于{Cora, CiteSeer, PubMed}中的任何一项则不打印split

java.sql.SQLException: SQLJob: 28ed8f13-d1ad-41a8-977b-baa49f413c04 failed when executing SQL: /********************************************************************/ insert overwrite table ksh_xjr_ssbk -- 一维度 select count(*) as ssbkqy,industry_name,null as area_name,null as qylx,null as ssbk from tmp_xjr_2 group by industry_name union all select count(*) as ssbkqy,null as industry_name,area_name,null as qylx,null as ssbk from tmp_xjr_2 group by area_name union all select count(*) as ssbkqy,null as industry_name,null as area_name,qylx,null as ssbk from tmp_xjr_2 group by qylx union all select count(*) as ssbkqy,null as industry_name,null as area_name,null as qylx,ssbk from tmp_xjr_2 group by ssbk -- 二维度 union all select count(*) as ssbkqy,industry_name,area_name,null as qylx,null as ssbk from tmp_xjr_2 group by industry_name,area_name union all select count(*) as ssbkqy,industry_name,null as area_name,qylx,null as ssbk from tmp_xjr_2 group by industry_name,qylx union all select count(*) as ssbkqy,industry_name,null as area_name,null as qylx,ssbk from tmp_xjr_2 group by industry_name,ssbk union all select count(*) as ssbkqy,null as industry_name,area_name,qylx,null as ssbk from tmp_xjr_2 group by area_name,qylx union all select count(*) as ssbkqy,null as industry_name,area_name,null as qylx,null as ssbk from tmp_xjr_2 group by area_name,ssbk union all select count(*) as ssbkqy,null as industry_name,null as area_name,qylx,ssbk from tmp_xjr_2 group by qylx,ssbk -- 三维度 union all select count(*) as ssbkqy,industry_name,null as area_name,qylx,ssbk from tmp_xjr_2 group by industry_name,qylx,ssbk union all select count(*) as ssbkqy,industry_name,area_name,null as qylx,ssbk from tmp_xjr_2 group by industry_name,area_name,ssbk union all select count(*) as ssbkqy,industry_name,area_name,qylx,null as ssbk from tmp_xjr_2 group by industry_name,area_name,qylx union all select count(*) as ssbkqy,null as industry_name,area_name,qylx,ssbk from tmp_xjr_2 group by area_name,qylx,ssbk -- 四维度 union all select count(*) as ssbkqy,null as industry_name,area_name,qylx,ssbk from tmp_xjr_2 group by industry_name,area_name,qylx,ssbk;

最新推荐

recommend-type

微软内部资料-SQL性能优化5

A clustered index is like a telephone directory in which all of the rows for customers with the same last name are clustered together in the same part of the book. Just as the organization of a ...
recommend-type

课设毕设基于SSM的毕业生就业信息管理系统-LW+PPT+源码可运行

课设毕设基于SSM的毕业生就业信息管理系统--LW+PPT+源码可运行
recommend-type

zigbee-cluster-library-specification

最新的zigbee-cluster-library-specification说明文档。
recommend-type

管理建模和仿真的文件

管理Boualem Benatallah引用此版本:布阿利姆·贝纳塔拉。管理建模和仿真。约瑟夫-傅立叶大学-格勒诺布尔第一大学,1996年。法语。NNT:电话:00345357HAL ID:电话:00345357https://theses.hal.science/tel-003453572008年12月9日提交HAL是一个多学科的开放存取档案馆,用于存放和传播科学研究论文,无论它们是否被公开。论文可以来自法国或国外的教学和研究机构,也可以来自公共或私人研究中心。L’archive ouverte pluridisciplinaire
recommend-type

实现实时数据湖架构:Kafka与Hive集成

![实现实时数据湖架构:Kafka与Hive集成](https://img-blog.csdnimg.cn/img_convert/10eb2e6972b3b6086286fc64c0b3ee41.jpeg) # 1. 实时数据湖架构概述** 实时数据湖是一种现代数据管理架构,它允许企业以低延迟的方式收集、存储和处理大量数据。与传统数据仓库不同,实时数据湖不依赖于预先定义的模式,而是采用灵活的架构,可以处理各种数据类型和格式。这种架构为企业提供了以下优势: - **实时洞察:**实时数据湖允许企业访问最新的数据,从而做出更明智的决策。 - **数据民主化:**实时数据湖使各种利益相关者都可
recommend-type

SPDK_NVMF_DISCOVERY_NQN是什么 有什么作用

SPDK_NVMF_DISCOVERY_NQN 是 SPDK (Storage Performance Development Kit) 中用于查询 NVMf (Non-Volatile Memory express over Fabrics) 存储设备名称的协议。NVMf 是一种基于网络的存储协议,可用于连接远程非易失性内存存储器。 SPDK_NVMF_DISCOVERY_NQN 的作用是让存储应用程序能够通过 SPDK 查询 NVMf 存储设备的名称,以便能够访问这些存储设备。通过查询 NVMf 存储设备名称,存储应用程序可以获取必要的信息,例如存储设备的IP地址、端口号、名称等,以便能
recommend-type

JSBSim Reference Manual

JSBSim参考手册,其中包含JSBSim简介,JSBSim配置文件xml的编写语法,编程手册以及一些应用实例等。其中有部分内容还没有写完,估计有生之年很难看到完整版了,但是内容还是很有参考价值的。
recommend-type

"互动学习:行动中的多样性与论文攻读经历"

多样性她- 事实上SCI NCES你的时间表ECOLEDO C Tora SC和NCESPOUR l’Ingén学习互动,互动学习以行动为中心的强化学习学会互动,互动学习,以行动为中心的强化学习计算机科学博士论文于2021年9月28日在Villeneuve d'Asq公开支持马修·瑟林评审团主席法布里斯·勒菲弗尔阿维尼翁大学教授论文指导奥利维尔·皮耶昆谷歌研究教授:智囊团论文联合主任菲利普·普雷教授,大学。里尔/CRISTAL/因里亚报告员奥利维耶·西格德索邦大学报告员卢多维奇·德诺耶教授,Facebook /索邦大学审查员越南圣迈IMT Atlantic高级讲师邀请弗洛里安·斯特鲁布博士,Deepmind对于那些及时看到自己错误的人...3谢谢你首先,我要感谢我的两位博士生导师Olivier和Philippe。奥利维尔,"站在巨人的肩膀上"这句话对你来说完全有意义了。从科学上讲,你知道在这篇论文的(许多)错误中,你是我可以依
recommend-type

实现实时监控告警系统:Kafka与Grafana整合

![实现实时监控告警系统:Kafka与Grafana整合](https://imgconvert.csdnimg.cn/aHR0cHM6Ly9tbWJpei5xcGljLmNuL21tYml6X2pwZy9BVldpY3ladXVDbEZpY1pLWmw2bUVaWXFUcEdLT1VDdkxRSmQxZXB5R1lxaWNlUjA2c0hFek5Qc3FyRktudFF1VDMxQVl3QTRXV2lhSWFRMEFRc0I1cW1ZOGcvNjQw?x-oss-process=image/format,png) # 1.1 Kafka集群架构 Kafka集群由多个称为代理的服务器组成,这
recommend-type

Windows 运行Python脚本

要在 Windows 上运行 Python 脚本,你需要先安装 Python。可以从官网下载 Python 安装包并按照提示进行安装。安装完成后,就可以在命令行中输入 `python` 命令,进入 Python 解释器环境。 接着,你可以编写 Python 脚本,保存为 `.py` 后缀的文件。在命令行中进入脚本所在的目录,输入 `python script.py` 命令来运行脚本。其中 `script.py` 是你的脚本文件名。 如果你想在 Windows 上运行一个 Python 程序,但不想打开命令行窗口,可以将脚本文件拖动到 Python 可执行文件 `python.exe` 上,