液化石油气数据更新到2.28;聚烯烃数据获取

This commit is contained in:
workpc 2025-03-10 17:00:07 +08:00
parent 29a65d7c70
commit d4187e8c1e
7 changed files with 428 additions and 1521 deletions

File diff suppressed because it is too large Load Diff

View File

@ -159,7 +159,7 @@ table_name = 'v_tbl_crude_oil_warning'
# 开关 # 开关
is_train = False # 是否训练 is_train = False # 是否训练
is_debug = False # 是否调试 is_debug = False # 是否调试
is_eta = False # 是否使用eta接口 is_eta = True # 是否使用eta接口
is_market = True # 是否通过市场信息平台获取特征 ,在is_eta 为true 的情况下生效 is_market = True # 是否通过市场信息平台获取特征 ,在is_eta 为true 的情况下生效
is_timefurture = True # 是否使用时间特征 is_timefurture = True # 是否使用时间特征
is_fivemodels = False # 是否使用之前保存的最佳的5个模型 is_fivemodels = False # 是否使用之前保存的最佳的5个模型

View File

@ -2,7 +2,7 @@ import logging
import os import os
import logging.handlers import logging.handlers
import datetime import datetime
from lib.tools import MySQLDB,SQLiteHandler from lib.tools import MySQLDB, SQLiteHandler
# eta 接口token # eta 接口token
@ -10,35 +10,35 @@ APPID = "XNLDvxZHHugj7wJ7"
SECRET = "iSeU4s6cKKBVbt94htVY1p0sqUMqb2xa" SECRET = "iSeU4s6cKKBVbt94htVY1p0sqUMqb2xa"
# eta 接口url # eta 接口url
sourcelisturl = 'http://10.189.2.78:8108/v1/edb/source/list' sourcelisturl = 'http://10.189.2.78:8108/v1/edb/source/list'
classifylisturl = 'http://10.189.2.78:8108/v1/edb/classify/list?ClassifyType=' classifylisturl = 'http://10.189.2.78:8108/v1/edb/classify/list?ClassifyType='
uniquecodedataurl = 'http://10.189.2.78:8108/v1/edb/data?UniqueCode=4991c37becba464609b409909fe4d992&StartDate=2024-02-01' uniquecodedataurl = 'http://10.189.2.78:8108/v1/edb/data?UniqueCode=4991c37becba464609b409909fe4d992&StartDate=2024-02-01'
classifyidlisturl = 'http://10.189.2.78:8108/v1/edb/list?ClassifyId=' classifyidlisturl = 'http://10.189.2.78:8108/v1/edb/list?ClassifyId='
edbcodedataurl = 'http://10.189.2.78:8108/v1/edb/data?EdbCode=' edbcodedataurl = 'http://10.189.2.78:8108/v1/edb/data?EdbCode='
edbdatapushurl = 'http://10.189.2.78:8108/v1/edb/push' edbdatapushurl = 'http://10.189.2.78:8108/v1/edb/push'
edbdeleteurl = 'http://10.189.2.78:8108/v1/edb/business/edb/del' edbdeleteurl = 'http://10.189.2.78:8108/v1/edb/business/edb/del'
edbbusinessurl = 'http://10.189.2.78:8108/v1/edb/business/data/del' edbbusinessurl = 'http://10.189.2.78:8108/v1/edb/business/data/del'
edbcodelist = ['ID01385938','lmcads03 lme comdty', edbcodelist = ['ID01385938', 'lmcads03 lme comdty',
'GC1 COMB Comdty', 'GC1 COMB Comdty',
'C2404171822', 'C2404171822',
'dxy curncy', 'dxy curncy',
'S5443199 ', 'S5443199 ',
'S5479800', 'S5479800',
'S5443108', 'S5443108',
'H7358586', 'H7358586',
'LC3FM1 INDEX', 'LC3FM1 INDEX',
'CNY REGN Curncy', 'CNY REGN Curncy',
's0105897', 's0105897',
'M0067419', 'M0067419',
'M0066351', 'M0066351',
'S0266372', 'S0266372',
'S0266438', 'S0266438',
'S0266506', 'S0266506',
'ID01384463'] 'ID01384463']
# 临时写死用指定的列,与上面的edbcode对应后面更改 # 临时写死用指定的列,与上面的edbcode对应后面更改
edbnamelist = [ edbnamelist = [
'ds','y', 'ds', 'y',
'LME铜价', 'LME铜价',
'黄金连1合约', '黄金连1合约',
'Brent-WTI', 'Brent-WTI',
@ -52,66 +52,65 @@ edbnamelist = [
'南华工业品指数', '南华工业品指数',
'PVC期货主力', 'PVC期货主力',
'PE期货收盘价', 'PE期货收盘价',
'PP连续-1月', 'PP连续-1月',
'PP连续-5月', 'PP连续-5月',
'PP连续-9月', 'PP连续-9月',
'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料' 'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料'
] ]
edbcodenamedict = { edbcodenamedict = {
'ID01385938':'PP拉丝1102K市场价青州国家能源宁煤', 'ID01385938': 'PP拉丝1102K市场价青州国家能源宁煤',
'ID01384463':'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料', 'ID01384463': 'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料',
'lmcads03 lme comdty':'LME铜价', 'lmcads03 lme comdty': 'LME铜价',
'GC1 COMB Comdty':'黄金连1合约', 'GC1 COMB Comdty': '黄金连1合约',
'C2404171822':'Brent-WTI', 'C2404171822': 'Brent-WTI',
'dxy curncy':'美元指数', 'dxy curncy': '美元指数',
'S5443199 ':'甲醇鲁南价格', 'S5443199 ': '甲醇鲁南价格',
'S5479800':'甲醇太仓港口价格', 'S5479800': '甲醇太仓港口价格',
'S5443108':'山东丙烯主流价', 'S5443108': '山东丙烯主流价',
'H7358586':'丙烷(山东)', 'H7358586': '丙烷(山东)',
'LC3FM1 INDEX':'FEI丙烷 M1', 'LC3FM1 INDEX': 'FEI丙烷 M1',
'CNY REGN Curncy':'在岸人民币汇率', 'CNY REGN Curncy': '在岸人民币汇率',
's0105897':'南华工业品指数', 's0105897': '南华工业品指数',
'M0067419':'PVC期货主力', 'M0067419': 'PVC期货主力',
'M0066351':'PE期货收盘价', 'M0066351': 'PE期货收盘价',
'S0266372':'PP连续-1月', 'S0266372': 'PP连续-1月',
'S0266438':'PP连续-5月', 'S0266438': 'PP连续-5月',
'S0266506':'PP连续-9月', 'S0266506': 'PP连续-9月',
} }
# eta自有数据指标编码 # eta自有数据指标编码
modelsindex = { modelsindex = {
'NHITS': 'SELF0000077', 'NHITS': 'SELF0000077',
'Informer':'SELF0000078', 'Informer': 'SELF0000078',
'LSTM':'SELF0000079', 'LSTM': 'SELF0000079',
'iTransformer':'SELF0000080', 'iTransformer': 'SELF0000080',
'TSMixer':'SELF0000081', 'TSMixer': 'SELF0000081',
'TSMixerx':'SELF0000082', 'TSMixerx': 'SELF0000082',
'PatchTST':'SELF0000083', 'PatchTST': 'SELF0000083',
'RNN':'SELF0000084', 'RNN': 'SELF0000084',
'GRU':'SELF0000085', 'GRU': 'SELF0000085',
'TCN':'SELF0000086', 'TCN': 'SELF0000086',
'BiTCN':'SELF0000087', 'BiTCN': 'SELF0000087',
'DilatedRNN':'SELF0000088', 'DilatedRNN': 'SELF0000088',
'MLP':'SELF0000089', 'MLP': 'SELF0000089',
'DLinear':'SELF0000090', 'DLinear': 'SELF0000090',
'NLinear':'SELF0000091', 'NLinear': 'SELF0000091',
'TFT':'SELF0000092', 'TFT': 'SELF0000092',
'FEDformer':'SELF0000093', 'FEDformer': 'SELF0000093',
'StemGNN':'SELF0000094', 'StemGNN': 'SELF0000094',
'MLPMultivariate':'SELF0000095', 'MLPMultivariate': 'SELF0000095',
'TiDE':'SELF0000096', 'TiDE': 'SELF0000096',
'DeepNPTS':'SELF0000097' 'DeepNPTS': 'SELF0000097'
} }
# eta 上传预测结果的请求体,后面发起请求的时候更改 model datalist 数据 # eta 上传预测结果的请求体,后面发起请求的时候更改 model datalist 数据
data = { data = {
"IndexCode": "", "IndexCode": "",
"IndexName": "价格预测模型", "IndexName": "价格预测模型",
"Unit": "", "Unit": "",
"Frequency": "日度", "Frequency": "日度",
"SourceName": f"价格预测", "SourceName": f"价格预测",
"Remark": 'ddd', "Remark": 'ddd',
@ -121,19 +120,18 @@ data = {
"Value": 333444 "Value": 333444
} }
] ]
} }
# eta 分类 # eta 分类
# level3才可以获取到数据所以需要人工把能源化工下所有的level3级都找到 # level3才可以获取到数据所以需要人工把能源化工下所有的level3级都找到
# url = 'http://10.189.2.78:8108/v1/edb/list?ClassifyId=1214' # url = 'http://10.189.2.78:8108/v1/edb/list?ClassifyId=1214'
#ParentId ":1160, 能源化工 # ParentId ":1160, 能源化工
# ClassifyId ":1214,原油 # ClassifyId ":1214,原油
#ParentId ":1214,",就是原油下所有的数据。 # ParentId ":1214,",就是原油下所有的数据。
ClassifyId = 1161 ClassifyId = 1161
# 变量定义--测试环境
############################################################################################################### 变量定义--测试环境
server_host = '192.168.100.53' server_host = '192.168.100.53'
login_pushreport_url = f"http://{server_host}:8080/jingbo-dev/api/server/login" login_pushreport_url = f"http://{server_host}:8080/jingbo-dev/api/server/login"
@ -145,7 +143,7 @@ login_data = {
"data": { "data": {
"account": "api_test", "account": "api_test",
# "password": "MmVmNzNlOWI0MmY0ZDdjZGUwNzE3ZjFiMDJiZDZjZWU=", # Shihua@123456 # "password": "MmVmNzNlOWI0MmY0ZDdjZGUwNzE3ZjFiMDJiZDZjZWU=", # Shihua@123456
"password": "ZTEwYWRjMzk0OWJhNTlhYmJlNTZlMDU3ZjIwZjg4M2U=", # 123456 "password": "ZTEwYWRjMzk0OWJhNTlhYmJlNTZlMDU3ZjIwZjg4M2U=", # 123456
"tenantHashCode": "8a4577dbd919675758d57999a1e891fe", "tenantHashCode": "8a4577dbd919675758d57999a1e891fe",
"terminal": "API" "terminal": "API"
}, },
@ -154,41 +152,41 @@ login_data = {
} }
upload_data = { upload_data = {
"funcModule":'研究报告信息', "funcModule": '研究报告信息',
"funcOperation":'上传聚烯烃PP价格预测报告', "funcOperation": '上传聚烯烃PP价格预测报告',
"data":{ "data": {
"groupNo": "000127", "groupNo": "000127",
"ownerAccount":'arui', #报告所属用户账号 "ownerAccount": 'arui', # 报告所属用户账号
"reportType":'OIL_PRICE_FORECAST', # 报告类型固定为OIL_PRICE_FORECAST "reportType": 'OIL_PRICE_FORECAST', # 报告类型固定为OIL_PRICE_FORECAST
"fileName": '2000-40-5-50--100-原油指标数据.xlsx-Brent活跃合约--2024-09-06-15-01-29-预测报告.pdf', #文件名称 "fileName": '2000-40-5-50--100-原油指标数据.xlsx-Brent活跃合约--2024-09-06-15-01-29-预测报告.pdf', # 文件名称
"fileBase64": '' ,#文件内容base64 "fileBase64": '', # 文件内容base64
"categoryNo":'yyjgycbg', # 研究报告分类编码 "categoryNo": 'yyjgycbg', # 研究报告分类编码
"smartBusinessClassCode":'JXTJGYCBG', #分析报告分类编码 "smartBusinessClassCode": 'JXTJGYCBG', # 分析报告分类编码
"reportEmployeeCode":"E40116", # 报告人 "reportEmployeeCode": "E40116", # 报告人
"reportDeptCode" :"D0044" ,# 报告部门 "reportDeptCode": "D0044", # 报告部门
"productGroupCode":"RAW_MATERIAL" # 商品分类 "productGroupCode": "RAW_MATERIAL" # 商品分类
} }
} }
warning_data = { warning_data = {
"funcModule":'原油特征停更预警', "funcModule": '原油特征停更预警',
"funcOperation":'原油特征停更预警', "funcOperation": '原油特征停更预警',
"data":{ "data": {
"groupNo": "000127", "groupNo": "000127",
'WARNING_TYPE_NAME':'特征数据停更预警', 'WARNING_TYPE_NAME': '特征数据停更预警',
'WARNING_CONTENT':'', 'WARNING_CONTENT': '',
'WARNING_DATE':'' 'WARNING_DATE': ''
} }
} }
query_data_list_item_nos_data = { query_data_list_item_nos_data = {
"funcModule": "数据项", "funcModule": "数据项",
"funcOperation": "查询", "funcOperation": "查询",
"data": { "data": {
"dateStart":"20200101", "dateStart": "20200101",
"dateEnd":"20241231", "dateEnd": "20241231",
"dataItemNoList":["Brentzdj","Brentzgj"] # 数据项编码,代表 brent最低价和最高价 "dataItemNoList": ["Brentzdj", "Brentzgj"] # 数据项编码,代表 brent最低价和最高价
} }
} }
@ -196,104 +194,104 @@ query_data_list_item_nos_data = {
# 北京环境数据库 # 北京环境数据库
host = '192.168.101.27' host = '192.168.101.27'
port = 3306 port = 3306
dbusername ='root' dbusername = 'root'
password = '123456' password = '123456'
dbname = 'jingbo_test' dbname = 'jingbo_test'
table_name = 'v_tbl_crude_oil_warning' table_name = 'v_tbl_crude_oil_warning'
### 开关 # 开关
is_train = False # 是否训练 is_train = False # 是否训练
is_debug = True # 是否调试 is_debug = True # 是否调试
is_eta = False # 是否使用eta接口 is_eta = True # 是否使用eta接口
is_market = False # 是否通过市场信息平台获取特征 ,在is_eta 为true 的情况下生效 is_market = False # 是否通过市场信息平台获取特征 ,在is_eta 为true 的情况下生效
is_timefurture = True # 是否使用时间特征 is_timefurture = True # 是否使用时间特征
is_fivemodels = False # 是否使用之前保存的最佳的5个模型 is_fivemodels = False # 是否使用之前保存的最佳的5个模型
is_edbcode = False # 特征使用edbcoding列表中的 is_edbcode = False # 特征使用edbcoding列表中的
is_edbnamelist = False # 自定义特征对应上面的edbnamelist is_edbnamelist = False # 自定义特征对应上面的edbnamelist
is_update_eta = False # 预测结果上传到eta is_update_eta = False # 预测结果上传到eta
is_update_report = True # 是否上传报告 is_update_report = True # 是否上传报告
is_update_warning_data = True # 是否上传预警数据 is_update_warning_data = True # 是否上传预警数据
is_del_corr = 0.6 # 是否删除相关性高的特征,取值为 0-1 0 为不删除0.6 表示删除相关性小于0.6的特征 is_del_corr = 0.6 # 是否删除相关性高的特征,取值为 0-1 0 为不删除0.6 表示删除相关性小于0.6的特征
is_del_tow_month = True # 是否删除两个月不更新的特征 is_del_tow_month = True # 是否删除两个月不更新的特征
# 连接到数据库 # 连接到数据库
db_mysql = MySQLDB(host=host, user=dbusername, password=password, database=dbname) db_mysql = MySQLDB(host=host, user=dbusername,
password=password, database=dbname)
db_mysql.connect() db_mysql.connect()
print("数据库连接成功",host,dbname,dbusername) print("数据库连接成功", host, dbname, dbusername)
# 数据截取日期 # 数据截取日期
start_year = 2020 # 数据开始年份 start_year = 2020 # 数据开始年份
end_time = '2025-01-27' # 数据截取日期 end_time = '' # 数据截取日期
freq = 'B' # 时间频率,"D": 天 "W": 周"M": 月"Q": 季度"A": 年 "H": 小时 "T": 分钟 "S": 秒 "B": 工作日 freq = 'B' # 时间频率,"D": 天 "W": 周"M": 月"Q": 季度"A": 年 "H": 小时 "T": 分钟 "S": 秒 "B": 工作日
delweekenday = True if freq == 'B' else False # 是否删除周末数据 delweekenday = True if freq == 'B' else False # 是否删除周末数据
is_corr = False # 特征是否参与滞后领先提升相关系数 is_corr = False # 特征是否参与滞后领先提升相关系数
add_kdj = False # 是否添加kdj指标 add_kdj = False # 是否添加kdj指标
if add_kdj and is_edbnamelist: if add_kdj and is_edbnamelist:
edbnamelist = edbnamelist+['K','D','J'] edbnamelist = edbnamelist+['K', 'D', 'J']
### 模型参数 # 模型参数
y = 'AVG-金能大唐久泰青州' y = 'AVG-金能大唐久泰青州'
avg_cols = [ avg_cols = [
'PP拉丝1102K出厂价青州国家能源宁煤', 'PP拉丝1102K出厂价青州国家能源宁煤',
'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料', 'PP拉丝L5E89出厂价华北第二区域内蒙古久泰新材料',
'PP拉丝L5E89出厂价河北、鲁北大唐内蒙多伦', 'PP拉丝L5E89出厂价河北、鲁北大唐内蒙多伦',
'PP拉丝HP550J市场价青岛金能化学' 'PP拉丝HP550J市场价青岛金能化学'
] ]
offsite = 80 offsite = 80
offsite_col = ['PP拉丝HP550J市场价青岛金能化学'] offsite_col = ['PP拉丝HP550J市场价青岛金能化学']
horizon =5 # 预测的步长 horizon = 5 # 预测的步长
input_size = 40 # 输入序列长度 input_size = 40 # 输入序列长度
train_steps = 50 if is_debug else 1000 # 训练步数,用来限定epoch次数 train_steps = 50 if is_debug else 1000 # 训练步数,用来限定epoch次数
val_check_steps = 30 # 评估频率 val_check_steps = 30 # 评估频率
early_stop_patience_steps = 5 # 早停的耐心步数 early_stop_patience_steps = 5 # 早停的耐心步数
# --- 交叉验证用的参数 # --- 交叉验证用的参数
test_size = 200 # 测试集大小定义100后面使用的时候重新赋值 test_size = 200 # 测试集大小定义100后面使用的时候重新赋值
val_size = test_size # 验证集大小,同测试集大小 val_size = test_size # 验证集大小,同测试集大小
### 特征筛选用到的参数 # 特征筛选用到的参数
k = 100 # 特征筛选数量如果是0或者值比特征数量大代表全部特征 k = 100 # 特征筛选数量如果是0或者值比特征数量大代表全部特征
corr_threshold = 0.6 # 相关性大于0.6的特征 corr_threshold = 0.6 # 相关性大于0.6的特征
rote = 0.06 # 绘图上下界阈值 rote = 0.06 # 绘图上下界阈值
### 计算准确率 # 计算准确率
weight_dict = [0.4,0.15,0.1,0.1,0.25] # 权重 weight_dict = [0.4, 0.15, 0.1, 0.1, 0.25] # 权重
### 文件 # 文件
data_set = 'PP指标数据.xlsx' # 数据集文件 data_set = 'PP指标数据.xlsx' # 数据集文件
dataset = 'juxitingdataset' # 数据集文件夹 dataset = 'juxitingdataset' # 数据集文件夹
# 数据库名称 # 数据库名称
db_name = os.path.join(dataset,'jbsh_juxiting.db') db_name = os.path.join(dataset, 'jbsh_juxiting.db')
sqlitedb = SQLiteHandler(db_name) sqlitedb = SQLiteHandler(db_name)
sqlitedb.connect() sqlitedb.connect()
settings = f'{input_size}-{horizon}-{train_steps}--{k}-{data_set}-{y}' settings = f'{input_size}-{horizon}-{train_steps}--{k}-{data_set}-{y}'
# 获取日期时间 # 获取日期时间
# now = datetime.datetime.now().strftime('%Y%m%d%H%M%S') # 获取当前日期时间 # now = datetime.datetime.now().strftime('%Y%m%d%H%M%S') # 获取当前日期时间
now = datetime.datetime.now().strftime('%Y-%m-%d') # 获取当前日期时间 now = datetime.datetime.now().strftime('%Y-%m-%d') # 获取当前日期时间
reportname = f'PP大模型预测报告--{end_time}.pdf' # 报告文件名 reportname = f'PP大模型预测报告--{end_time}.pdf' # 报告文件名
reportname = reportname.replace(':', '-') # 替换冒号 reportname = reportname.replace(':', '-') # 替换冒号
if end_time == '': if end_time == '':
end_time = now end_time = now
### 邮件配置 # 邮件配置
username='1321340118@qq.com' username = '1321340118@qq.com'
passwd='wgczgyhtyyyyjghi' passwd = 'wgczgyhtyyyyjghi'
# recv=['liurui_test@163.com','52585119@qq.com'] # recv=['liurui_test@163.com','52585119@qq.com']
recv=['liurui_test@163.com'] recv = ['liurui_test@163.com']
# recv=['liurui_test@163.com'] # recv=['liurui_test@163.com']
title='reportname' title = 'reportname'
content=y+'预测报告请看附件' content = y+'预测报告请看附件'
file=os.path.join(dataset,'reportname') file = os.path.join(dataset, 'reportname')
# file=os.path.join(dataset,'14-7-50--100-原油指标数据.xlsx-Brent连1合约价格--20240731175936-预测报告.pdf') # file=os.path.join(dataset,'14-7-50--100-原油指标数据.xlsx-Brent连1合约价格--20240731175936-预测报告.pdf')
ssl=True ssl = True
### 日志配置 # 日志配置
# 创建日志目录(如果不存在) # 创建日志目录(如果不存在)
log_dir = 'logs' log_dir = 'logs'
@ -305,8 +303,10 @@ logger = logging.getLogger('my_logger')
logger.setLevel(logging.INFO) logger.setLevel(logging.INFO)
# 配置文件处理器,将日志记录到文件 # 配置文件处理器,将日志记录到文件
file_handler = logging.handlers.RotatingFileHandler(os.path.join(log_dir, 'pricepredict.log'), maxBytes=1024 * 1024, backupCount=5) file_handler = logging.handlers.RotatingFileHandler(os.path.join(
file_handler.setFormatter(logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')) log_dir, 'pricepredict.log'), maxBytes=1024 * 1024, backupCount=5)
file_handler.setFormatter(logging.Formatter(
'%(asctime)s - %(name)s - %(levelname)s - %(message)s'))
# 配置控制台处理器,将日志打印到控制台 # 配置控制台处理器,将日志打印到控制台
console_handler = logging.StreamHandler() console_handler = logging.StreamHandler()
@ -317,4 +317,3 @@ logger.addHandler(file_handler)
logger.addHandler(console_handler) logger.addHandler(console_handler)
# logger.info('当前配置:'+settings) # logger.info('当前配置:'+settings)

View File

@ -103,6 +103,14 @@ global_config = {
# ETA配置 # ETA配置
'APPID': None, 'APPID': None,
'SECRET': None, 'SECRET': None,
'classifylisturl': None,
'classifyidlisturl': None,
'edbcodedataurl': None,
'edbcodelist': None,
'edbdatapushurl': None,
'edbdeleteurl': None,
'edbbusinessurl': None,
'ClassifyId': None,
# 数据库配置 # 数据库配置
'sqlitedb': None, 'sqlitedb': None,
@ -1352,7 +1360,7 @@ def style_row(row):
class EtaReader(): class EtaReader():
def __init__(self, signature, classifylisturl, classifyidlisturl, edbcodedataurl, edbcodelist, edbdatapushurl, edbdeleteurl, edbbusinessurl): def __init__(self, signature, classifylisturl, classifyidlisturl, edbcodedataurl, edbcodelist, edbdatapushurl, edbdeleteurl, edbbusinessurl, classifyId):
''' '''
初始化 EtaReader 类的实例 初始化 EtaReader 类的实例
@ -1377,6 +1385,7 @@ class EtaReader():
self.edbcodelist = edbcodelist self.edbcodelist = edbcodelist
self.edbdeleteurl = edbdeleteurl self.edbdeleteurl = edbdeleteurl
self.edbbusinessurl = edbbusinessurl self.edbbusinessurl = edbbusinessurl
self.classifyId = classifyId
def filter_yuanyou_data(self, ClassifyName, data): def filter_yuanyou_data(self, ClassifyName, data):
''' '''
@ -1779,7 +1788,6 @@ class EtaReader():
return df_zhibiaoshuju, df_zhibiaoliebiao return df_zhibiaoshuju, df_zhibiaoliebiao
def get_eta_api_pp_data(self, data_set, dataset=''): def get_eta_api_pp_data(self, data_set, dataset=''):
global ClassifyId
today = datetime.date.today().strftime("%Y-%m-%d") today = datetime.date.today().strftime("%Y-%m-%d")
# 定义你的headers这里可以包含多个参数 # 定义你的headers这里可以包含多个参数
@ -1818,7 +1826,7 @@ class EtaReader():
# 请求成功,处理响应内容 # 请求成功,处理响应内容
# config.logger.info(data.get('Data')) # config.logger.info(data.get('Data'))
# 定义你想要保留的固定值 # 定义你想要保留的固定值
fixed_value = ClassifyId fixed_value = self.classifyId
# 遍历列表,只保留那些'category' key的值为固定值的数据项 # 遍历列表,只保留那些'category' key的值为固定值的数据项
filtered_data = [item for item in data.get( filtered_data = [item for item in data.get(

View File

@ -1,12 +1,75 @@
# 读取配置 # 读取配置
from lib.dataread import *
from lib.tools import SendMail,exception_logger
from models.nerulforcastmodels import ex_Model_Juxiting,model_losss,model_losss_juxiting,brent_export_pdf,tansuanli_export_pdf,pp_export_pdf,model_losss_juxiting
import glob from lib.dataread import *
from config_juxiting import *
from lib.tools import SendMail, exception_logger
from models.nerulforcastmodels import ex_Model, model_losss, model_losss_juxiting, brent_export_pdf, tansuanli_export_pdf, pp_export_pdf, model_losss_juxiting
import datetime
import torch import torch
torch.set_float32_matmul_precision("high") torch.set_float32_matmul_precision("high")
global_config.update({
# 核心参数
'logger': logger,
'dataset': dataset,
'y': y,
'is_debug': is_debug,
'is_train': is_train,
'is_fivemodels': is_fivemodels,
'settings': settings,
# 模型参数
'data_set': data_set,
'input_size': input_size,
'horizon': horizon,
'train_steps': train_steps,
'val_check_steps': val_check_steps,
'val_size': val_size,
'test_size': test_size,
'modelsindex': modelsindex,
'rote': rote,
# 特征工程开关
'is_del_corr': is_del_corr,
'is_del_tow_month': is_del_tow_month,
'is_eta': is_eta,
'is_update_eta': is_update_eta,
'early_stop_patience_steps': early_stop_patience_steps,
# 时间参数
'start_year': start_year,
'end_time': end_time or datetime.datetime.now().strftime("%Y-%m-%d"),
'freq': freq, # 保持列表结构
# 接口配置
'login_pushreport_url': login_pushreport_url,
'login_data': login_data,
'upload_url': upload_url,
'upload_warning_url': upload_warning_url,
'warning_data': warning_data,
# 查询接口
'query_data_list_item_nos_url': query_data_list_item_nos_url,
'query_data_list_item_nos_data': query_data_list_item_nos_data,
# eta 配置
'APPID': APPID,
'SECRET': SECRET,
'etadata': data,
'edbcodelist': edbcodelist,
'ClassifyId': ClassifyId,
'edbcodedataurl': edbcodedataurl,
'classifyidlisturl': classifyidlisturl,
'edbdatapushurl': edbdatapushurl,
'edbdeleteurl': edbdeleteurl,
'edbbusinessurl': edbbusinessurl,
'ClassifyId': ClassifyId,
'classifylisturl': classifylisturl,
# 数据库配置
'sqlitedb': sqlitedb,
})
def predict_main(): def predict_main():
@ -48,31 +111,23 @@ def predict_main():
返回: 返回:
None None
""" """
global end_time end_time = global_config['end_time']
signature = BinanceAPI(APPID, SECRET)
etadata = EtaReader(signature=signature,
classifylisturl=classifylisturl,
classifyidlisturl=classifyidlisturl,
edbcodedataurl=edbcodedataurl,
edbcodelist=edbcodelist,
edbdatapushurl=edbdatapushurl,
edbdeleteurl=edbdeleteurl,
edbbusinessurl=edbbusinessurl
)
# 获取数据 # 获取数据
if is_eta: if is_eta:
logger.info('从eta获取数据...') logger.info('从eta获取数据...')
signature = BinanceAPI(APPID, SECRET) signature = BinanceAPI(APPID, SECRET)
etadata = EtaReader(signature=signature, etadata = EtaReader(signature=signature,
classifylisturl=classifylisturl, classifylisturl=global_config['classifylisturl'],
classifyidlisturl=classifyidlisturl, classifyidlisturl=global_config['classifyidlisturl'],
edbcodedataurl=edbcodedataurl, edbcodedataurl=global_config['edbcodedataurl'],
edbcodelist=edbcodelist, edbcodelist=global_config['edbcodelist'],
edbdatapushurl=edbdatapushurl, edbdatapushurl=global_config['edbdatapushurl'],
edbdeleteurl=edbdeleteurl, edbdeleteurl=global_config['edbdeleteurl'],
edbbusinessurl=edbbusinessurl, edbbusinessurl=global_config['edbbusinessurl'],
classifyId=global_config['ClassifyId'],
) )
df_zhibiaoshuju, df_zhibiaoliebiao = etadata.get_eta_api_pp_data(data_set=data_set, dataset=dataset) # 原始数据,未处理 df_zhibiaoshuju, df_zhibiaoliebiao = etadata.get_eta_api_pp_data(
data_set=data_set, dataset=dataset) # 原始数据,未处理
if is_market: if is_market:
logger.info('从市场信息平台获取数据...') logger.info('从市场信息平台获取数据...')
@ -83,26 +138,26 @@ def predict_main():
df_zhibiaoshuju = get_high_low_data(df_zhibiaoshuju) df_zhibiaoshuju = get_high_low_data(df_zhibiaoshuju)
else: else:
logger.info('从市场信息平台获取数据') logger.info('从市场信息平台获取数据')
df_zhibiaoshuju = get_market_data(end_time,df_zhibiaoshuju) df_zhibiaoshuju = get_market_data(
end_time, df_zhibiaoshuju)
except :
except:
logger.info('最高最低价拼接失败') logger.info('最高最低价拼接失败')
# 保存到xlsx文件的sheet表 # 保存到xlsx文件的sheet表
with pd.ExcelWriter(os.path.join(dataset,data_set)) as file: with pd.ExcelWriter(os.path.join(dataset, data_set)) as file:
df_zhibiaoshuju.to_excel(file, sheet_name='指标数据', index=False) df_zhibiaoshuju.to_excel(file, sheet_name='指标数据', index=False)
df_zhibiaoliebiao.to_excel(file, sheet_name='指标列表', index=False) df_zhibiaoliebiao.to_excel(file, sheet_name='指标列表', index=False)
# 数据处理 # 数据处理
df = datachuli_juxiting(df_zhibiaoshuju, df_zhibiaoliebiao, y=y, dataset=dataset, add_kdj=add_kdj, is_timefurture=is_timefurture, df = datachuli(df_zhibiaoshuju, df_zhibiaoliebiao, y=y, dataset=dataset, add_kdj=add_kdj, is_timefurture=is_timefurture,
end_time=end_time) end_time=end_time)
else: else:
# 读取数据 # 读取数据
logger.info('读取本地数据:' + os.path.join(dataset, data_set)) logger.info('读取本地数据:' + os.path.join(dataset, data_set))
df,df_zhibiaoliebiao = getdata_juxiting(filename=os.path.join(dataset, data_set), y=y, dataset=dataset, add_kdj=add_kdj, df, df_zhibiaoliebiao = getdata(filename=os.path.join(dataset, data_set), y=y, dataset=dataset, add_kdj=add_kdj,
is_timefurture=is_timefurture, end_time=end_time) # 原始数据,未处理 is_timefurture=is_timefurture, end_time=end_time) # 原始数据,未处理
# 更改预测列名称 # 更改预测列名称
df.rename(columns={y: 'y'}, inplace=True) df.rename(columns={y: 'y'}, inplace=True)
@ -124,47 +179,65 @@ def predict_main():
else: else:
for row in first_row.itertuples(index=False): for row in first_row.itertuples(index=False):
row_dict = row._asdict() row_dict = row._asdict()
row_dict['ds'] = row_dict['ds'].strftime('%Y-%m-%d %H:%M:%S') config.logger.info(f'要保存的真实值:{row_dict}')
check_query = sqlitedb.select_data('trueandpredict', where_condition=f"ds = '{row.ds}'") # 判断ds是否为字符串类型,如果不是则转换为字符串类型
if isinstance(row_dict['ds'], (pd.Timestamp, datetime.datetime)):
row_dict['ds'] = row_dict['ds'].strftime('%Y-%m-%d')
elif not isinstance(row_dict['ds'], str):
try:
row_dict['ds'] = pd.to_datetime(
row_dict['ds']).strftime('%Y-%m-%d')
except:
logger.warning(f"无法解析的时间格式: {row_dict['ds']}")
# row_dict['ds'] = row_dict['ds'].strftime('%Y-%m-%d')
# row_dict['ds'] = row_dict['ds'].strftime('%Y-%m-%d %H:%M:%S')
check_query = sqlitedb.select_data(
'trueandpredict', where_condition=f"ds = '{row.ds}'")
if len(check_query) > 0: if len(check_query) > 0:
set_clause = ", ".join([f"{key} = '{value}'" for key, value in row_dict.items()]) set_clause = ", ".join(
sqlitedb.update_data('trueandpredict', set_clause, where_condition=f"ds = '{row.ds}'") [f"{key} = '{value}'" for key, value in row_dict.items()])
sqlitedb.update_data(
'trueandpredict', set_clause, where_condition=f"ds = '{row.ds}'")
continue continue
sqlitedb.insert_data('trueandpredict', tuple(row_dict.values()), columns=row_dict.keys()) sqlitedb.insert_data('trueandpredict', tuple(
row_dict.values()), columns=row_dict.keys())
# 更新accuracy表的y值 # 更新accuracy表的y值
if not sqlitedb.check_table_exists('accuracy'): if not sqlitedb.check_table_exists('accuracy'):
pass pass
else: else:
update_y = sqlitedb.select_data('accuracy',where_condition="y is null") update_y = sqlitedb.select_data(
'accuracy', where_condition="y is null")
if len(update_y) > 0: if len(update_y) > 0:
logger.info('更新accuracy表的y值') logger.info('更新accuracy表的y值')
# 找到update_y 中ds且df中的y的行 # 找到update_y 中ds且df中的y的行
update_y = update_y[update_y['ds']<=end_time] update_y = update_y[update_y['ds'] <= end_time]
logger.info(f'要更新y的信息{update_y}') logger.info(f'要更新y的信息{update_y}')
# try: # try:
for row in update_y.itertuples(index=False): for row in update_y.itertuples(index=False):
try: try:
row_dict = row._asdict() row_dict = row._asdict()
yy = df[df['ds']==row_dict['ds']]['y'].values[0] yy = df[df['ds'] == row_dict['ds']]['y'].values[0]
LOW = df[df['ds']==row_dict['ds']]['Brentzdj'].values[0] LOW = df[df['ds'] == row_dict['ds']]['Brentzdj'].values[0]
HIGH = df[df['ds']==row_dict['ds']]['Brentzgj'].values[0] HIGH = df[df['ds'] == row_dict['ds']]['Brentzgj'].values[0]
sqlitedb.update_data('accuracy', f"y = {yy},LOW_PRICE = {LOW},HIGH_PRICE = {HIGH}", where_condition=f"ds = '{row_dict['ds']}'") sqlitedb.update_data(
'accuracy', f"y = {yy},LOW_PRICE = {LOW},HIGH_PRICE = {HIGH}", where_condition=f"ds = '{row_dict['ds']}'")
except: except:
logger.info(f'更新accuracy表的y值失败{row_dict}') logger.info(f'更新accuracy表的y值失败{row_dict}')
# except Exception as e: # except Exception as e:
# logger.info(f'更新accuracy表的y值失败{e}') # logger.info(f'更新accuracy表的y值失败{e}')
import datetime
# 判断当前日期是不是周一 # 判断当前日期是不是周一
is_weekday = datetime.datetime.now().weekday() == 0 is_weekday = datetime.datetime.now().weekday() == 0
if is_weekday: if is_weekday:
logger.info('今天是周一,更新预测模型') logger.info('今天是周一,更新预测模型')
# 计算最近60天预测残差最低的模型名称 # 计算最近60天预测残差最低的模型名称
model_results = sqlitedb.select_data('trueandpredict', order_by="ds DESC", limit="60") model_results = sqlitedb.select_data(
'trueandpredict', order_by="ds DESC", limit="60")
# 删除空值率为90%以上的列 # 删除空值率为90%以上的列
if len(model_results) > 10: if len(model_results) > 10:
model_results = model_results.dropna(thresh=len(model_results)*0.1,axis=1) model_results = model_results.dropna(
thresh=len(model_results)*0.1, axis=1)
# 删除空行 # 删除空行
model_results = model_results.dropna() model_results = model_results.dropna()
modelnames = model_results.columns.to_list()[2:-1] modelnames = model_results.columns.to_list()[2:-1]
@ -172,51 +245,61 @@ def predict_main():
model_results[col] = model_results[col].astype(np.float32) model_results[col] = model_results[col].astype(np.float32)
# 计算每个预测值与真实值之间的偏差率 # 计算每个预测值与真实值之间的偏差率
for model in modelnames: for model in modelnames:
model_results[f'{model}_abs_error_rate'] = abs(model_results['y'] - model_results[model]) / model_results['y'] model_results[f'{model}_abs_error_rate'] = abs(
model_results['y'] - model_results[model]) / model_results['y']
# 获取每行对应的最小偏差率值 # 获取每行对应的最小偏差率值
min_abs_error_rate_values = model_results.apply(lambda row: row[[f'{model}_abs_error_rate' for model in modelnames]].min(), axis=1) min_abs_error_rate_values = model_results.apply(
lambda row: row[[f'{model}_abs_error_rate' for model in modelnames]].min(), axis=1)
# 获取每行对应的最小偏差率值对应的列名 # 获取每行对应的最小偏差率值对应的列名
min_abs_error_rate_column_name = model_results.apply(lambda row: row[[f'{model}_abs_error_rate' for model in modelnames]].idxmin(), axis=1) min_abs_error_rate_column_name = model_results.apply(
lambda row: row[[f'{model}_abs_error_rate' for model in modelnames]].idxmin(), axis=1)
# 将列名索引转换为列名 # 将列名索引转换为列名
min_abs_error_rate_column_name = min_abs_error_rate_column_name.map(lambda x: x.split('_')[0]) min_abs_error_rate_column_name = min_abs_error_rate_column_name.map(
lambda x: x.split('_')[0])
# 取出现次数最多的模型名称 # 取出现次数最多的模型名称
most_common_model = min_abs_error_rate_column_name.value_counts().idxmax() most_common_model = min_abs_error_rate_column_name.value_counts().idxmax()
logger.info(f"最近60天预测残差最低的模型名称{most_common_model}") logger.info(f"最近60天预测残差最低的模型名称{most_common_model}")
# 保存结果到数据库 # 保存结果到数据库
if not sqlitedb.check_table_exists('most_model'): if not sqlitedb.check_table_exists('most_model'):
sqlitedb.create_table('most_model', columns="ds datetime, most_common_model TEXT") sqlitedb.create_table(
sqlitedb.insert_data('most_model', (datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S'), most_common_model,), columns=('ds', 'most_common_model',)) 'most_model', columns="ds datetime, most_common_model TEXT")
sqlitedb.insert_data('most_model', (datetime.datetime.now().strftime(
'%Y-%m-%d %H:%M:%S'), most_common_model,), columns=('ds', 'most_common_model',))
try: try:
# if is_weekday: if is_weekday:
if True: # if True:
# logger.info('今天是周一,发送特征预警') logger.info('今天是周一,发送特征预警')
# # 上传预警信息到数据库 # 上传预警信息到数据库
# warning_data_df = df_zhibiaoliebiao.copy() warning_data_df = df_zhibiaoliebiao.copy()
# warning_data_df = warning_data_df[warning_data_df['停更周期']> 3 ][['指标名称', '指标id', '频度','更新周期','指标来源','最后更新时间','停更周期']] warning_data_df = warning_data_df[warning_data_df['停更周期'] > 3][[
# # 重命名列名 '指标名称', '指标id', '频度', '更新周期', '指标来源', '最后更新时间', '停更周期']]
# warning_data_df = warning_data_df.rename(columns={'指标名称': 'INDICATOR_NAME', '指标id': 'INDICATOR_ID', '频度': 'FREQUENCY', '更新周期': 'UPDATE_FREQUENCY', '指标来源': 'DATA_SOURCE', '最后更新时间': 'LAST_UPDATE_DATE', '停更周期': 'UPDATE_SUSPENSION_CYCLE'}) # 重命名列名
# from sqlalchemy import create_engine warning_data_df = warning_data_df.rename(columns={'指标名称': 'INDICATOR_NAME', '指标id': 'INDICATOR_ID', '频度': 'FREQUENCY',
# import urllib '更新周期': 'UPDATE_FREQUENCY', '指标来源': 'DATA_SOURCE', '最后更新时间': 'LAST_UPDATE_DATE', '停更周期': 'UPDATE_SUSPENSION_CYCLE'})
# global password from sqlalchemy import create_engine
# if '@' in password: import urllib
# password = urllib.parse.quote_plus(password) global password
if '@' in password:
password = urllib.parse.quote_plus(password)
# engine = create_engine(f'mysql+pymysql://{dbusername}:{password}@{host}:{port}/{dbname}') engine = create_engine(
# warning_data_df['WARNING_DATE'] = datetime.date.today().strftime("%Y-%m-%d %H:%M:%S") f'mysql+pymysql://{dbusername}:{password}@{host}:{port}/{dbname}')
# warning_data_df['TENANT_CODE'] = 'T0004' warning_data_df['WARNING_DATE'] = datetime.date.today().strftime(
# # 插入数据之前查询表数据然后新增id列 "%Y-%m-%d %H:%M:%S")
# existing_data = pd.read_sql(f"SELECT * FROM {table_name}", engine) warning_data_df['TENANT_CODE'] = 'T0004'
# if not existing_data.empty: # 插入数据之前查询表数据然后新增id列
# max_id = existing_data['ID'].astype(int).max() existing_data = pd.read_sql(f"SELECT * FROM {table_name}", engine)
# warning_data_df['ID'] = range(max_id + 1, max_id + 1 + len(warning_data_df)) if not existing_data.empty:
# else: max_id = existing_data['ID'].astype(int).max()
# warning_data_df['ID'] = range(1, 1 + len(warning_data_df)) warning_data_df['ID'] = range(
# warning_data_df.to_sql(table_name, con=engine, if_exists='append', index=False) max_id + 1, max_id + 1 + len(warning_data_df))
# if is_update_warning_data: else:
# upload_warning_info(len(warning_data_df)) warning_data_df['ID'] = range(1, 1 + len(warning_data_df))
warning_data_df.to_sql(
table_name, con=engine, if_exists='append', index=False)
if is_update_warning_data: if is_update_warning_data:
upload_warning_info(10) upload_warning_info(len(warning_data_df))
except: except:
logger.info('上传预警信息到数据库失败') logger.info('上传预警信息到数据库失败')
@ -228,76 +311,75 @@ def predict_main():
row, col = df.shape row, col = df.shape
now = datetime.datetime.now().strftime('%Y%m%d%H%M%S') now = datetime.datetime.now().strftime('%Y%m%d%H%M%S')
# ex_Model_Juxiting(df, ex_Model(df,
# horizon=horizon, horizon=global_config['horizon'],
# input_size=input_size, input_size=global_config['input_size'],
# train_steps=train_steps, train_steps=global_config['train_steps'],
# val_check_steps=val_check_steps, val_check_steps=global_config['val_check_steps'],
# early_stop_patience_steps=early_stop_patience_steps, early_stop_patience_steps=global_config['early_stop_patience_steps'],
# is_debug=is_debug, is_debug=global_config['is_debug'],
# dataset=dataset, dataset=global_config['dataset'],
# is_train=is_train, is_train=global_config['is_train'],
# is_fivemodels=is_fivemodels, is_fivemodels=global_config['is_fivemodels'],
# val_size=val_size, val_size=global_config['val_size'],
# test_size=test_size, test_size=global_config['test_size'],
# settings=settings, settings=global_config['settings'],
# now=now, now=now,
# etadata=etadata, etadata=global_config['etadata'],
# modelsindex=modelsindex, modelsindex=global_config['modelsindex'],
# data=data, data=data,
# is_eta=is_eta, is_eta=global_config['is_eta'],
# end_time=end_time, end_time=global_config['end_time'],
# ) )
logger.info('模型训练完成') logger.info('模型训练完成')
logger.info('训练数据绘图ing') logger.info('训练数据绘图ing')
model_results3 = model_losss_juxiting(sqlitedb) model_results3 = model_losss(sqlitedb, end_time=end_time)
logger.info('训练数据绘图end') logger.info('训练数据绘图end')
# 模型报告
logger.info('制作报告ing')
title = f'{settings}--{end_time}-预测报告' # 报告标题
reportname = f'PP大模型预测报告--{end_time}.pdf' # 报告文件名
reportname = reportname.replace(':', '-') # 替换冒号
pp_export_pdf(dataset=dataset,num_models = 5 if is_fivemodels else 22,time=end_time,
reportname=reportname,sqlitedb=sqlitedb),
logger.info('制作报告end') # # 模型报告
logger.info('模型训练完成') # logger.info('制作报告ing')
# title = f'{settings}--{end_time}-预测报告' # 报告标题
# reportname = f'Brent原油大模型月度预测--{end_time}.pdf' # 报告文件名
# reportname = reportname.replace(':', '-') # 替换冒号
# brent_export_pdf(dataset=dataset, num_models=5 if is_fivemodels else 22, time=end_time,
# reportname=reportname, sqlitedb=sqlitedb),
# logger.info('制作报告end')
# logger.info('模型训练完成')
# # LSTM 单变量模型 # # LSTM 单变量模型
# ex_Lstm(df,input_seq_len=input_size,output_seq_len=horizon,is_debug=is_debug,dataset=dataset) # ex_Lstm(df,input_seq_len=input_size,output_seq_len=horizon,is_debug=is_debug,dataset=dataset)
# # lstm 多变量模型 # # lstm 多变量模型
# ex_Lstm_M(df,n_days=input_size,out_days=horizon,is_debug=is_debug,datasetpath=dataset) # ex_Lstm_M(df,n_days=input_size,out_days=horizon,is_debug=is_debug,datasetpath=dataset)
# # GRU 模型 # # GRU 模型
# # ex_GRU(df) # # ex_GRU(df)
# 发送邮件 # 发送邮件
m = SendMail( # m = SendMail(
username=username, # username=username,
passwd=passwd, # passwd=passwd,
recv=recv, # recv=recv,
title=title, # title=title,
content=content, # content=content,
file=max(glob.glob(os.path.join(dataset,'*.pdf')), key=os.path.getctime), # file=max(glob.glob(os.path.join(dataset,'*.pdf')), key=os.path.getctime),
ssl=ssl, # ssl=ssl,
) # )
# m.send_mail() # m.send_mail()
if __name__ == '__main__': if __name__ == '__main__':
# global end_time # global end_time
# is_on = True # 遍历2024-11-25 到 2024-12-3 之间的工作日日期
# # 遍历2024-11-25 到 2024-12-3 之间的工作日日期 # for i_time in pd.date_range('2022-1-1', '2025-3-26', freq='M'):
# for i_time in pd.date_range('2025-1-20', '2025-2-6', freq='B'):
# end_time = i_time.strftime('%Y-%m-%d')
# try: # try:
# global_config['end_time'] = i_time.strftime('%Y-%m-%d')
# predict_main() # predict_main()
# except: # except Exception as e:
# pass # logger.info(f'预测失败:{e}')
# continue
predict_main() predict_main()

View File

@ -374,13 +374,12 @@ def predict_main():
if __name__ == '__main__': if __name__ == '__main__':
# global end_time # global end_time
# 遍历2024-11-25 到 2024-12-3 之间的工作日日期 # 遍历2024-11-25 到 2024-12-3 之间的工作日日期
for i_time in pd.date_range('2022-1-1', '2025-3-26', freq='M'): # for i_time in pd.date_range('2022-1-1', '2025-3-26', freq='M'):
try: # try:
global_config['end_time'] = i_time.strftime('%Y-%m-%d') # global_config['end_time'] = i_time.strftime('%Y-%m-%d')
predict_main() # predict_main()
except Exception as e: # except Exception as e:
logger.info(f'预测失败:{e}') # logger.info(f'预测失败:{e}')
continue # continue
predict_main()
# predict_main()