From fb47d36048e94b9a506d5c153e3dd19a01e37df1 Mon Sep 17 00:00:00 2001 From: Administrator <admin@example.com> Date: 星期一, 30 十月 2023 16:30:27 +0800 Subject: [PATCH] bug修复 --- third_data/data_server.py | 121 ++++++++++++++++++++++++++++++++-------- 1 files changed, 97 insertions(+), 24 deletions(-) diff --git a/third_data/data_server.py b/third_data/data_server.py index 6feb510..34edcdc 100644 --- a/third_data/data_server.py +++ b/third_data/data_server.py @@ -1,19 +1,21 @@ import http import json +import logging import socketserver +import threading import time from http.server import BaseHTTPRequestHandler import dask -from log_module.log import logger_system -from utils import global_util, tool +from log_module.log import logger_system, logger_debug +from utils import global_util, tool, data_export_util from code_attribute import gpcode_manager from log_module import log, log_analyse, log_export -from l2 import code_price_manager, l2_data_util +from l2 import code_price_manager, l2_data_util, l2_data_manager_new, cancel_buy_strategy, transaction_progress from l2.cancel_buy_strategy import HourCancelBigNumComputer from output.limit_up_data_filter import IgnoreCodeManager -from third_data import kpl_util, kpl_data_manager, kpl_api -from third_data.code_plate_key_manager import RealTimeKplMarketData, KPLPlateForbiddenManager +from third_data import kpl_util, kpl_data_manager, kpl_api, block_info +from third_data.code_plate_key_manager import RealTimeKplMarketData, KPLPlateForbiddenManager, CodePlateKeyBuyManager from third_data.history_k_data_util import HistoryKDatasUtils from third_data.kpl_data_manager import KPLDataManager, KPLLimitUpDataRecordManager, \ KPLCodeLimitUpReasonManager @@ -22,8 +24,13 @@ from urllib.parse import parse_qs from output import code_info_output, limit_up_data_filter, output_util, kp_client_msg_manager -from trade import bidding_money_manager, trade_manager +from trade import bidding_money_manager, trade_manager, l2_trade_util from trade.l2_trade_util import BlackListCodeManager +import concurrent.futures + +# 绂佺敤http.server鐨勬棩蹇楄緭鍑� +logger = logging.getLogger("http.server") +logger.setLevel(logging.CRITICAL) class DataServer(BaseHTTPRequestHandler): @@ -40,6 +47,12 @@ # 绮鹃��,琛屼笟鏁版嵁缂撳瓨 __jingxuan_cache_dict = {} __industry_cache_dict = {} + __latest_limit_up_codes_set = set() + __data_process_thread_pool = concurrent.futures.ThreadPoolExecutor(max_workers=10) + + # 绂佺敤鏃ュ織杈撳嚭 + def log_message(self, format, *args): + pass def __get_limit_up_list(self): # 缁熻鐩墠涓烘鐨勪唬鐮佹定鍋滄暟閲忥紙鍒嗘定鍋滃師鍥狅級 @@ -286,20 +299,36 @@ ps_dict = dict([(k, v[0]) for k, v in parse_qs(url.query).items()]) code = ps_dict['code'] name = ps_dict.get('name') + try: + data = code_info_output.get_output_params(code, self.__jingxuan_cache_dict, self.__industry_cache_dict) + if data["code_name"].find("None") > -1 and name: + data["code_name"] = f"{name} {code}" - data = code_info_output.get_output_params(code, self.__jingxuan_cache_dict, self.__industry_cache_dict) - if data["code_name"].find("None") > -1 and name: - data["code_name"] = f"{name} {code}" + self.__history_plates_dict[code] = (time.time(), data["kpl_code_info"]["code_records"]) + if "plate" in data["kpl_code_info"]: + self.__blocks_dict[code] = (time.time(), data["kpl_code_info"]["plate"]) - self.__history_plates_dict[code] = (time.time(), data["kpl_code_info"]["code_records"]) - if "plate" in data["kpl_code_info"]: - self.__blocks_dict[code] = (time.time(), data["kpl_code_info"]["plate"]) - - response_data = json.dumps({"code": 0, "data": data}) - print("get_score_info 鑰楁椂锛�", time.time() - start_time) + response_data = json.dumps({"code": 0, "data": data}) + print("get_score_info 鑰楁椂锛�", time.time() - start_time) + except Exception as e: + logging.exception(e) # 鑾峰彇璇勫垎淇℃伅 pass + elif url.path == "/get_l2_datas": + # 鑾峰彇L2鐨勬暟鎹� + ps_dict = dict([(k, v[0]) for k, v in parse_qs(url.query).items()]) + code = ps_dict['code'] + datas = data_export_util.get_l2_datas(code) + response_data = json.dumps({"code": 0, "data": datas}) + elif url.path == "/get_trade_progress": + # 鑾峰彇浜ゆ槗杩涘害 + ps_dict = dict([(k, v[0]) for k, v in parse_qs(url.query).items()]) + code = ps_dict['code'] + trade_progress, is_default = transaction_progress.TradeBuyQueue().get_traded_index(code) + datas = data_export_util.get_l2_datas(code) + response_data = json.dumps( + {"code": 0, "data": {"trade_progress": trade_progress, "is_default": is_default}}) elif url.path == "/kpl/get_limit_up_list": response_data = self.__get_limit_up_list() @@ -477,16 +506,61 @@ self.__send_response(result_str) def __process_kpl_data(self, data): - def do_limit_up(result_list): - if result_list: + def do_limit_up(result_list_): + if result_list_: # 淇濆瓨娑ㄥ仠鏃堕棿 - for d in result_list: + codes_set = set() + limit_up_reasons = {} + for d in result_list_: code = d[0] + limit_up_reasons[code] = d[5] + codes_set.add(code) if code.find("00") == 0 or code.find("60") == 0: limit_up_time = time.strftime("%H:%M:%S", time.localtime(d[2])) code_price_manager.Buy1PriceManager().set_limit_up_time(code, limit_up_time) - kpl_data_manager.KPLLimitUpDataRecordManager.save_record(tool.get_now_date_str(), result_list) - self.__kplDataManager.save_data(type_, result_list) + add_codes = codes_set - self.__latest_limit_up_codes_set + self.__latest_limit_up_codes_set = codes_set + + if limit_up_reasons: + # 缁熻娑ㄥ仠鍘熷洜鐨勭エ鐨勪釜鏁� + limit_up_reason_code_dict = {} + for code in limit_up_reasons: + b = limit_up_reasons[code] + if b not in limit_up_reason_code_dict: + limit_up_reason_code_dict[b] = set() + limit_up_reason_code_dict[b].add(code) + cancel_buy_strategy.LCancelRateManager.set_block_limit_up_count(limit_up_reason_code_dict) + + if add_codes: + for code in add_codes: + # 鏍规嵁娑ㄥ仠鍘熷洜鍒ゆ柇鏄惁鍙互涔� + if code.find("00") == 0 or code.find("60") == 0: + try: + # 鍒ゆ柇鏄惁涓嬪崟 + trade_state = trade_manager.CodesTradeStateManager().get_trade_state(code) + if trade_state == trade_manager.TRADE_STATE_BUY_PLACE_ORDER or trade_state == trade_manager.TRADE_STATE_BUY_DELEGATED: + # 濮旀墭涓殑璁㈠崟锛屽垽鏂槸鍚﹂渶瑕佹挙鍗� + if not gpcode_manager.WantBuyCodesManager().is_in_cache(code): + yesterday_codes = kpl_data_manager.get_yesterday_limit_up_codes() + current_limit_up_datas, limit_up_record_datas, yesterday_current_limit_up_codes, before_blocks_dict = kpl_data_manager.KPLLimitUpDataRecordManager.latest_origin_datas, kpl_data_manager.KPLLimitUpDataRecordManager.total_datas, yesterday_codes, block_info.get_before_blocks_dict() + if not current_limit_up_datas: + current_limit_up_datas = [] + if not limit_up_record_datas: + limit_up_record_datas = [] + # 涔扮粷瀵硅�佸ぇ + # 涓�斾笉鑳芥挙鍗� + # if CodePlateKeyBuyManager.is_need_cancel(code, limit_up_reasons.get(code), + # current_limit_up_datas, + # limit_up_record_datas, + # yesterday_current_limit_up_codes, + # before_blocks_dict): + # l2_data_manager_new.L2TradeDataProcessor.cancel_buy(code, + # f"娑ㄥ仠鍘熷洜锛坽limit_up_reasons.get(code)}锛変笉鏄�佸ぇ鎾ゅ崟", + # "鏉垮潡鎾�") + except Exception as e: + logger_debug.exception(e) + kpl_data_manager.KPLLimitUpDataRecordManager.save_record(tool.get_now_date_str(), result_list_) + self.__kplDataManager.save_data(type_, result_list_) type_ = data["type"] print("寮�鐩樺暒type:", type_) @@ -505,8 +579,7 @@ self.__kplDataManager.save_data(type_, result_list) elif type_ == KPLDataType.LIMIT_UP.value: result_list = kpl_util.parseDaBanData(data["data"], kpl_util.DABAN_TYPE_LIMIT_UP) - do_limit_up(result_list) - + self.__data_process_thread_pool.submit(lambda: do_limit_up(result_list)) elif type_ == KPLDataType.OPEN_LIMIT_UP.value: result_list = kpl_util.parseDaBanData(data["data"], kpl_util.DABAN_TYPE_OPEN_LIMIT_UP) if result_list: @@ -576,9 +649,9 @@ handler = DataServer # httpd = socketserver.TCPServer((addr, port), handler) - httpd = ThreadedHTTPServer((addr, port), handler) - print("HTTP server is at: http://%s:%d/" % (addr, port)) try: + httpd = ThreadedHTTPServer((addr, port), handler) + print("HTTP server is at: http://%s:%d/" % (addr, port)) httpd.serve_forever() except Exception as e: logger_system.exception(e) -- Gitblit v1.8.0