提交 80e88bfde60d3b99211049197694143b66d62573
Merge branch 'master' of http://gitlab.ctune.cn/weijunh/DMapManager
merge task
正在显示
3 个修改的文件
包含
219 行增加
和
278 行删除
| 1 | 1 | from datetime import datetime, timedelta |
| 2 | - | |
| 2 | +from operator import or_ | |
| 3 | +from sqlalchemy.sql.expression import intersect_all | |
| 4 | +import re | |
| 3 | 5 | from sqlalchemy.sql.functions import func |
| 4 | -from .models import MonitorLog, db | |
| 6 | +from .models import db, MonitorInfo | |
| 5 | 7 | from sqlalchemy import and_ |
| 6 | 8 | from app.util.component.ApiTemplate import ApiTemplate |
| 9 | +import math | |
| 10 | +from functools import reduce | |
| 7 | 11 | |
| 8 | 12 | |
| 9 | 13 | class Api(ApiTemplate): |
| ... | ... | @@ -13,205 +17,62 @@ class Api(ApiTemplate): |
| 13 | 17 | |
| 14 | 18 | # 返回结果 |
| 15 | 19 | res = {} |
| 16 | - res["data"] = {} | |
| 20 | + res["data"] = [] | |
| 17 | 21 | logs = [] |
| 18 | 22 | try: |
| 19 | 23 | server = self.para.get("server") # server |
| 20 | - # metrics_type = self.para.get("metrics_type") | |
| 21 | - grain = self.para.get("grain") | |
| 22 | - count = int(self.para.get("count")) | |
| 24 | + metrics_type = self.para.get("metrics_type") | |
| 25 | + interval = self.para.get("interval") | |
| 23 | 26 | start_time = self.para.get("start") |
| 24 | 27 | to_time = self.para.get("to") |
| 25 | - | |
| 26 | - cur_now = datetime.now() | |
| 27 | 28 | start_datetime = datetime.strptime(start_time, "%Y-%m-%d %H:%M:%S") |
| 28 | 29 | end_datetime = datetime.strptime(to_time, "%Y-%m-%d %H:%M:%S") |
| 29 | - | |
| 30 | - if grain == "day": | |
| 31 | - # 粒度是一天 | |
| 32 | - format = "%Y-%m-%d" | |
| 33 | - logs = db.session.query(MonitorLog.date_stamp.label("key"), func.avg(MonitorLog.cpu_usage).label("cpu_usage"), | |
| 34 | - func.avg(MonitorLog.total_mem).label( | |
| 35 | - "total_mem"), | |
| 36 | - func.avg(MonitorLog.available_mem).label( | |
| 37 | - "available_mem"), | |
| 38 | - func.avg(MonitorLog.used_mem).label( | |
| 39 | - "used_mem"), | |
| 40 | - func.avg(MonitorLog.disk).label( | |
| 41 | - "disk"), | |
| 42 | - func.avg(MonitorLog.disk_usage).label( | |
| 43 | - "disk_usage"), | |
| 44 | - func.avg(MonitorLog.net_recv).label( | |
| 45 | - "net_recv"), | |
| 46 | - func.avg(MonitorLog.net_send).label("net_send")).filter(and_( | |
| 47 | - MonitorLog.date_stamp <= end_datetime.strftime( | |
| 48 | - format), | |
| 49 | - MonitorLog.date_stamp > start_datetime.strftime( | |
| 50 | - format), | |
| 51 | - MonitorLog.server == server)).group_by(MonitorLog.date_stamp).order_by(MonitorLog.date_stamp.asc()).all() | |
| 52 | - | |
| 53 | - keys = [] | |
| 54 | - keys_map = {} | |
| 55 | - while start_datetime <= end_datetime: | |
| 56 | - keys.append(self.get_monitor_log( | |
| 57 | - start_datetime.strftime('%Y-%m-%d'))) | |
| 58 | - keys_map[start_datetime.strftime('%Y-%m-%d')] = len(keys)-1 | |
| 59 | - start_datetime += timedelta(days=1) | |
| 60 | - | |
| 61 | - res["data"] = self.get_result(logs, keys, keys_map) | |
| 62 | - elif grain == "minu_five": | |
| 63 | - # 粒度是5分钟 | |
| 64 | - cur_minu = int(end_datetime.strftime("%M")) % 5 | |
| 65 | - logs = db.session.query(MonitorLog.hour_stamp.label("hour_stamp"), | |
| 66 | - MonitorLog.f_minu_stamp.label( | |
| 67 | - "minu_stamp"), | |
| 68 | - func.avg(MonitorLog.cpu_usage).label( | |
| 69 | - "cpu_usage"), | |
| 70 | - func.avg(MonitorLog.total_mem).label( | |
| 71 | - "total_mem"), | |
| 72 | - func.avg(MonitorLog.available_mem).label( | |
| 73 | - "available_mem"), | |
| 74 | - func.avg(MonitorLog.used_mem).label( | |
| 75 | - "used_mem"), | |
| 76 | - func.avg(MonitorLog.disk).label( | |
| 77 | - "disk"), | |
| 78 | - func.avg(MonitorLog.disk_usage).label( | |
| 79 | - "disk_usage"), | |
| 80 | - func.avg(MonitorLog.net_recv).label( | |
| 81 | - "net_recv"), | |
| 82 | - func.avg(MonitorLog.net_send).label("net_send")).filter(and_( | |
| 83 | - MonitorLog.time_stamp <= "{}:00".format( | |
| 84 | - end_datetime.strftime("%Y-%m-%d %H:%M")), | |
| 85 | - MonitorLog.time_stamp >= "{}:00".format( | |
| 86 | - start_datetime.strftime("%Y-%m-%d %H:%M")), | |
| 87 | - MonitorLog.server == server)).group_by(MonitorLog.date_stamp, MonitorLog.hour_stamp, MonitorLog.f_minu_stamp).order_by( | |
| 88 | - MonitorLog.hour_stamp.asc(), MonitorLog.f_minu_stamp.asc()).all() | |
| 89 | - | |
| 90 | - keys = [] | |
| 91 | - keys_map = {} | |
| 92 | - while start_datetime <= end_datetime: | |
| 93 | - key = start_datetime.strftime('%H:%M') | |
| 94 | - keys.append(self.get_monitor_log(key)) | |
| 95 | - keys_map[key] = len(keys)-1 | |
| 96 | - start_datetime += timedelta(minutes=5) | |
| 97 | - | |
| 98 | - logs = list(map(lambda log: | |
| 99 | - {"key": "%02d:%02d" % (log.hour_stamp, log.minu_stamp*5+cur_minu), | |
| 100 | - "cpu_usage": log.cpu_usage, "total_mem": log.total_mem, | |
| 101 | - "available_mem": log.available_mem, "used_mem": log.used_mem, "disk": log.disk, "disk_usage": log.disk_usage, | |
| 102 | - "net_recv": log.net_recv, "net_send": log.net_send}, | |
| 103 | - logs)) | |
| 104 | - | |
| 105 | - res["data"] = self.get_result_from_list(logs, keys, keys_map) | |
| 106 | - elif grain == "minu_ten": | |
| 107 | - # 粒度是10分钟 | |
| 108 | - cur_minu = int(end_datetime.strftime("%M")) % 10 | |
| 109 | - logs = db.session.query(MonitorLog.hour_stamp.label("hour_stamp"), | |
| 110 | - MonitorLog.d_minu_stamp.label( | |
| 111 | - "minu_stamp"), | |
| 112 | - func.avg(MonitorLog.cpu_usage).label( | |
| 113 | - "cpu_usage"), | |
| 114 | - func.avg(MonitorLog.total_mem).label( | |
| 115 | - "total_mem"), | |
| 116 | - func.avg(MonitorLog.available_mem).label( | |
| 117 | - "available_mem"), | |
| 118 | - func.avg(MonitorLog.used_mem).label( | |
| 119 | - "used_mem"), | |
| 120 | - func.avg(MonitorLog.disk).label( | |
| 121 | - "disk"), | |
| 122 | - func.avg(MonitorLog.disk_usage).label( | |
| 123 | - "disk_usage"), | |
| 124 | - func.avg(MonitorLog.net_recv).label( | |
| 125 | - "net_recv"), | |
| 126 | - func.avg(MonitorLog.net_send).label("net_send")).filter(and_( | |
| 127 | - MonitorLog.time_stamp <= "{}:00".format( | |
| 128 | - end_datetime.strftime("%Y-%m-%d %H:%M")), | |
| 129 | - MonitorLog.time_stamp >= "{}:00".format( | |
| 130 | - start_datetime.strftime("%Y-%m-%d %H:%M")), | |
| 131 | - MonitorLog.server == server)).group_by(MonitorLog.date_stamp, | |
| 132 | - MonitorLog.hour_stamp, | |
| 133 | - MonitorLog.d_minu_stamp).order_by(MonitorLog.hour_stamp.asc(), MonitorLog.d_minu_stamp.asc()).all() | |
| 134 | - keys = [] | |
| 135 | - keys_map = {} | |
| 136 | - while start_datetime <= end_datetime: | |
| 137 | - key = start_datetime.strftime('%H:%M') | |
| 138 | - keys.append(self.get_monitor_log(key)) | |
| 139 | - keys_map[key] = len(keys)-1 | |
| 140 | - start_datetime += timedelta(minutes=10) | |
| 141 | - | |
| 142 | - logs = list(map(lambda log: | |
| 143 | - {"key": "%02d:%02d" % (log.hour_stamp, log.minu_stamp*10+cur_minu), | |
| 144 | - "cpu_usage": log.cpu_usage, "total_mem": log.total_mem, | |
| 145 | - "available_mem": log.available_mem, "used_mem": log.used_mem, "disk": log.disk, "disk_usage": log.disk_usage, | |
| 146 | - "net_recv": log.net_recv, "net_send": log.net_send}, | |
| 147 | - logs)) | |
| 148 | - | |
| 149 | - res["data"] = self.get_result_from_list(logs, keys, keys_map) | |
| 150 | - elif grain == "hour": | |
| 151 | - # 粒度是一小时 | |
| 152 | - logs = db.session.query(MonitorLog.hour_stamp.label("key"), func.avg(MonitorLog.cpu_usage).label("cpu_usage"), | |
| 153 | - func.avg(MonitorLog.total_mem).label( | |
| 154 | - "total_mem"), | |
| 155 | - func.avg(MonitorLog.available_mem).label( | |
| 156 | - "available_mem"), | |
| 157 | - func.avg(MonitorLog.used_mem).label( | |
| 158 | - "used_mem"), | |
| 159 | - func.avg(MonitorLog.disk).label( | |
| 160 | - "disk"), | |
| 161 | - func.avg(MonitorLog.disk_usage).label( | |
| 162 | - "disk_usage"), | |
| 163 | - func.avg(MonitorLog.net_recv).label( | |
| 164 | - "net_recv"), | |
| 165 | - func.avg(MonitorLog.net_send).label("net_send")).filter(and_( | |
| 166 | - MonitorLog.time_stamp <= "{}:59:59".format( | |
| 167 | - end_datetime.strftime("%Y-%m-%d %H")), | |
| 168 | - MonitorLog.time_stamp >= "{}:59:59".format( | |
| 169 | - start_datetime.strftime("%Y-%m-%d %H")), | |
| 170 | - MonitorLog.server == server)).group_by(MonitorLog.hour_stamp).order_by(MonitorLog.hour_stamp.asc()).all() | |
| 171 | - keys = [] | |
| 172 | - keys_map = {} | |
| 173 | - while start_datetime <= end_datetime: | |
| 174 | - key = int(start_datetime.strftime('%H')) | |
| 175 | - keys.append(self.get_monitor_log("%02d时" % key)) | |
| 176 | - keys_map[str(key)] = len(keys)-1 | |
| 177 | - start_datetime += timedelta(hours=1) | |
| 178 | - | |
| 179 | - res["data"] = self.get_result(logs, keys, keys_map) | |
| 30 | + format = "%Y-%m-%d %H:%M:%S" | |
| 31 | + | |
| 32 | + interval_size = int(re.findall("\d+", interval)[0]) | |
| 33 | + interval_unit = re.findall("\D+", interval)[0] | |
| 34 | + | |
| 35 | + if interval_size == 1 and interval_unit == 'm': | |
| 36 | + logs = db.session.query(MonitorInfo.time_stamp.label('key'), MonitorInfo.value).filter( | |
| 37 | + and_(MonitorInfo.time_stamp <= end_datetime.strftime(format), | |
| 38 | + MonitorInfo.time_stamp > start_datetime.strftime( | |
| 39 | + format), | |
| 40 | + MonitorInfo.server == server, | |
| 41 | + MonitorInfo.metrics == metrics_type)).order_by(MonitorInfo.time_stamp).group_by( | |
| 42 | + MonitorInfo.time_stamp, MonitorInfo.value) | |
| 43 | + datas = list(map(lambda log: | |
| 44 | + {"key": datetime.strftime(log.key, format), | |
| 45 | + "value": log.value}, | |
| 46 | + logs)) | |
| 47 | + datas = reduce(lambda y, x: y if (x['key'] in [i['key'] for i in y]) else ( | |
| 48 | + lambda z, u: (z.append(u), z))(y, x)[1], datas, []) | |
| 49 | + for data in datas: | |
| 50 | + res['data'].append([data['key'], data['value']]) | |
| 180 | 51 | else: |
| 181 | - # 按分钟统计 | |
| 182 | - logs = db.session.query(MonitorLog.hour_stamp.label("hour_stamp"), MonitorLog.minu_stamp.label("minu_stamp"), func.avg(MonitorLog.cpu_usage).label("cpu_usage"), | |
| 183 | - func.avg(MonitorLog.total_mem).label( | |
| 184 | - "total_mem"), | |
| 185 | - func.avg(MonitorLog.available_mem).label( | |
| 186 | - "available_mem"), | |
| 187 | - func.avg(MonitorLog.used_mem).label("used_mem"), | |
| 188 | - func.avg(MonitorLog.disk).label( | |
| 189 | - "disk"), | |
| 190 | - func.avg(MonitorLog.disk_usage).label( | |
| 191 | - "disk_usage"), | |
| 192 | - func.avg(MonitorLog.net_recv).label( | |
| 193 | - "net_recv"), | |
| 194 | - func.avg(MonitorLog.net_send).label("net_send")).filter(and_( | |
| 195 | - MonitorLog.time_stamp <= end_datetime.strftime("%Y-%m-%d %H:%M:%S"), | |
| 196 | - MonitorLog.time_stamp >= start_datetime.strftime("%Y-%m-%d %H:%M:%S"), | |
| 197 | - MonitorLog.server == server)).group_by(MonitorLog.hour_stamp, MonitorLog.minu_stamp).order_by( | |
| 198 | - MonitorLog.hour_stamp.asc(), MonitorLog.minu_stamp.asc()).all() | |
| 199 | - keys = [] | |
| 200 | - keys_map = {} | |
| 201 | - while start_datetime <= end_datetime: | |
| 202 | - key = start_datetime.strftime('%H:%M') | |
| 203 | - keys.append(self.get_monitor_log(key)) | |
| 204 | - keys_map[key] = len(keys)-1 | |
| 205 | - start_datetime += timedelta(minutes=1) | |
| 206 | - | |
| 207 | - logs = list(map(lambda log: | |
| 208 | - {"key": "%02d:%02d" % (log.hour_stamp, log.minu_stamp), | |
| 209 | - "cpu_usage": log.cpu_usage, "total_mem": log.total_mem, | |
| 210 | - "available_mem": log.available_mem, "used_mem": log.used_mem, "disk": log.disk, "disk_usage": log.disk_usage, | |
| 211 | - "net_recv": log.net_recv, "net_send": log.net_send}, | |
| 212 | - logs)) | |
| 213 | - | |
| 214 | - res["data"] = self.get_result_from_list(logs, keys, keys_map) | |
| 52 | + interval_start_datetime = start_datetime | |
| 53 | + interval_end_datatime = self.get_end_interval( | |
| 54 | + interval_start_datetime, interval_unit, interval_size) | |
| 55 | + res_format = "%Y-%m-%d %H:%M:%S" | |
| 56 | + | |
| 57 | + while interval_end_datatime <= end_datetime: | |
| 58 | + logs = db.session.query(MonitorInfo.time_stamp.label('key'), MonitorInfo.value).filter( | |
| 59 | + and_(MonitorInfo.time_stamp <= interval_end_datatime.strftime(format), | |
| 60 | + MonitorInfo.time_stamp > interval_start_datetime.strftime( | |
| 61 | + format), MonitorInfo.server == server, MonitorInfo.metrics == metrics_type)).order_by(MonitorInfo.time_stamp).group_by( | |
| 62 | + MonitorInfo.time_stamp, MonitorInfo.value).all() | |
| 63 | + tmp_data = list(map(lambda log: | |
| 64 | + {"key": datetime.strftime(log.key, res_format), | |
| 65 | + "value": log.value}, | |
| 66 | + logs)) | |
| 67 | + datas = self.get_sample_data(tmp_data) | |
| 68 | + datas = reduce(lambda y, x: y if (x['key'] in [i['key'] for i in y]) else ( | |
| 69 | + lambda z, u: (z.append(u), z))(y, x)[1], datas, []) | |
| 70 | + for data in datas: | |
| 71 | + res['data'].append([data['key'], data['value']]) | |
| 72 | + | |
| 73 | + interval_start_datetime = interval_end_datatime | |
| 74 | + interval_end_datatime = self.get_end_interval( | |
| 75 | + interval_start_datetime, interval_unit, interval_size) | |
| 215 | 76 | res["result"] = True |
| 216 | 77 | except Exception as e: |
| 217 | 78 | raise e |
| ... | ... | @@ -224,14 +85,10 @@ class Api(ApiTemplate): |
| 224 | 85 | "in": "query", |
| 225 | 86 | "type": "string", |
| 226 | 87 | "description": "服务器地址"}, |
| 227 | - {"name": "grain", | |
| 228 | - "in": "query", | |
| 229 | - "type": "string", | |
| 230 | - "description": "指标粒度:minu:分钟,minu_five:5分钟,minu_ten:10分钟,hour:1小时,day:每天"}, | |
| 231 | - {"name": "count", | |
| 88 | + {"name": "interval", | |
| 232 | 89 | "in": "query", |
| 233 | 90 | "type": "string", |
| 234 | - "description": "个数"}, | |
| 91 | + "description": "间隔"}, | |
| 235 | 92 | {"name": "to", |
| 236 | 93 | "in": "query", |
| 237 | 94 | "type": "string", |
| ... | ... | @@ -239,7 +96,11 @@ class Api(ApiTemplate): |
| 239 | 96 | {"name": "start", |
| 240 | 97 | "in": "query", |
| 241 | 98 | "type": "string", |
| 242 | - "description": "查询起始时间"} | |
| 99 | + "description": "查询起始时间"}, | |
| 100 | + {"name": "metrics_type", | |
| 101 | + "in": "query", | |
| 102 | + "type": "string", | |
| 103 | + "description": "查询指标"} | |
| 243 | 104 | ], |
| 244 | 105 | "responses": { |
| 245 | 106 | 200: { |
| ... | ... | @@ -251,40 +112,50 @@ class Api(ApiTemplate): |
| 251 | 112 | } |
| 252 | 113 | } |
| 253 | 114 | |
| 254 | - def get_monitor_log(self, key): | |
| 255 | - return {"key": key, | |
| 256 | - "cpu_usage": None, "total_mem": None, | |
| 257 | - "available_mem": None, "used_mem": None, "disk": None, "disk_usage": None, | |
| 258 | - "net_recv": None, "net_send": None} | |
| 259 | - | |
| 260 | - def get_result(self, logs, keys, keys_map): | |
| 261 | - keys_map_key=keys_map.keys() | |
| 262 | - for log in logs: | |
| 263 | - if str(log.key) in keys_map_key: | |
| 264 | - tmp = keys[keys_map[str(log.key)]] | |
| 265 | - if tmp != None: | |
| 266 | - tmp['cpu_usage'] = log.cpu_usage | |
| 267 | - tmp['total_mem'] = log.total_mem | |
| 268 | - tmp['available_mem'] = log.available_mem | |
| 269 | - tmp['used_mem'] = log.used_mem | |
| 270 | - tmp['disk'] = log.disk | |
| 271 | - tmp['disk_usage'] = log.disk_usage | |
| 272 | - tmp['net_recv'] = log.net_recv | |
| 273 | - tmp['net_send'] = log.net_send | |
| 274 | - return keys | |
| 115 | + def get_end_interval(self, start, unit, size): | |
| 116 | + if unit == 'm': | |
| 117 | + return start+timedelta(minutes=size) | |
| 118 | + elif unit == 'h': | |
| 119 | + return start+timedelta(hours=size) | |
| 120 | + elif unit == 'd': | |
| 121 | + return start+timedelta(days=size) | |
| 122 | + else: | |
| 123 | + return None | |
| 124 | + | |
| 125 | + def get_sample_data(self, orginal): | |
| 126 | + res = [] | |
| 127 | + size = len(orginal) | |
| 128 | + orginal_stamp = {'head': 1, 'tail': size} | |
| 129 | + if size > 1: | |
| 130 | + stamp = {'P0': 1, | |
| 131 | + 'P50': math.floor(0.5*size), | |
| 132 | + # 'P90': math.floor(0.9*size), | |
| 133 | + # 'P95': math.floor(0.95*size), | |
| 134 | + # 'P99': math.floor(0.99*size), | |
| 135 | + 'P100': size} | |
| 136 | + elif size == 1: | |
| 137 | + stamp = {'P0': 1, | |
| 138 | + 'P50': size, | |
| 139 | + # 'P90': size, | |
| 140 | + # 'P95': size, | |
| 141 | + # 'P99': size, | |
| 142 | + 'P100': size} | |
| 143 | + else: | |
| 144 | + return res | |
| 145 | + | |
| 146 | + for key in dict.keys(orginal_stamp): | |
| 147 | + cur_data = orginal[orginal_stamp[key]-1] | |
| 148 | + info = {'key': cur_data['key'], 'value': cur_data['value']} | |
| 149 | + res.append(info) | |
| 150 | + | |
| 151 | + data = sorted(orginal, key=lambda x: x['value']) | |
| 152 | + for key in dict.keys(stamp): | |
| 153 | + cur_data = data[stamp[key]-1] | |
| 154 | + info = {'key': cur_data['key'], 'value': cur_data['value']} | |
| 155 | + res.append(info) | |
| 156 | + | |
| 157 | + res.sort(key=self.takeKey) | |
| 158 | + return res | |
| 275 | 159 | |
| 276 | - def get_result_from_list(self, logs, keys, keys_map): | |
| 277 | - keys_map_key=keys_map.keys() | |
| 278 | - for log in logs: | |
| 279 | - if str(log["key"]) in keys_map_key: | |
| 280 | - tmp = keys[keys_map[str(log["key"])]] | |
| 281 | - if tmp != None: | |
| 282 | - tmp['cpu_usage'] = log["cpu_usage"] | |
| 283 | - tmp['total_mem'] = log["total_mem"] | |
| 284 | - tmp['available_mem'] = log["available_mem"] | |
| 285 | - tmp['used_mem'] = log["used_mem"] | |
| 286 | - tmp['disk'] = log["disk"] | |
| 287 | - tmp['disk_usage'] = log["disk_usage"] | |
| 288 | - tmp['net_recv'] = log["net_recv"] | |
| 289 | - tmp['net_send'] = log["net_send"] | |
| 290 | - return keys | |
| 160 | + def takeKey(self, elem): | |
| 161 | + return elem['key'] | ... | ... |
| 1 | 1 | # coding=utf-8 |
| 2 | -# author: 4N | |
| 3 | -#createtime: 2021/6/11 | |
| 4 | -#email: nheweijun@sina.com | |
| 5 | - | |
| 6 | - | |
| 2 | +# author: resioR | |
| 3 | +#createtime: 2021/12/1 | |
| 4 | +#email: qianyingz@chinadci.com | |
| 7 | 5 | from datetime import datetime |
| 8 | 6 | from time import time |
| 9 | -from sqlalchemy import Column, Integer, String, ForeignKey, Text, DateTime, Time, Float, Binary | |
| 7 | +from sqlalchemy import Column, Integer, String, ForeignKey, Text, DateTime, Time, Float, Binary,TIMESTAMP | |
| 10 | 8 | from sqlalchemy.orm import relationship |
| 11 | 9 | from sqlalchemy.sql.expression import column |
| 12 | 10 | from app.models import db |
| 13 | 11 | |
| 14 | - | |
| 15 | -class MonitorLog(db.Model): | |
| 16 | - ''' | |
| 17 | - ''' | |
| 18 | - __tablename__ = "dmap_monitor_log" | |
| 19 | - guid = Column(String(256), primary_key=True) | |
| 20 | - server = Column(String(256)) | |
| 21 | - date_stamp = Column(Text) | |
| 22 | - hour_stamp = Column(Integer) | |
| 23 | - minu_stamp = Column(Integer) | |
| 24 | - d_minu_stamp = Column(Integer) # 10min粒度 1-6 | |
| 25 | - f_minu_stamp = Column(Integer) # 5min粒度 1~12 | |
| 26 | - time_stamp = Column(DateTime) # 创建时间戳 | |
| 27 | - cpu_usage = Column(Float) | |
| 28 | - total_mem = Column(Float) | |
| 29 | - available_mem = Column(Float) | |
| 30 | - used_mem = Column(Float) | |
| 31 | - disk = Column(Float) | |
| 32 | - disk_usage = Column(Float) | |
| 33 | - net_recv = Column(Float) | |
| 34 | - net_send = Column(Float) | |
| 35 | - | |
| 36 | - | |
| 37 | 12 | class MonitorHost(db.Model): |
| 38 | 13 | ''' |
| 39 | 14 | 监控服务器配置 |
| ... | ... | @@ -44,4 +19,16 @@ class MonitorHost(db.Model): |
| 44 | 19 | user = Column(String(256)) |
| 45 | 20 | password = Column(String(256)) |
| 46 | 21 | type = Column(String(256)) |
| 47 | - host_name=Column(String(256)) | |
| 22 | + host_name = Column(String(256)) | |
| 23 | + | |
| 24 | +class MonitorInfo(db.Model): | |
| 25 | + ''' | |
| 26 | + ''' | |
| 27 | + __tablename__ = "dmap_monitor_info" | |
| 28 | + guid = Column(String(256), primary_key=True) | |
| 29 | + metrics=Column(String(256)) | |
| 30 | + server = Column(String(256)) | |
| 31 | + date_stamp = Column(Text) | |
| 32 | + time_stamp = Column(TIMESTAMP) # 创建时间戳 | |
| 33 | + stamp = Column(Text) # P0,P50,P90,P95,P99,P100 | |
| 34 | + value = Column(Float) | |
| \ No newline at end of file | ... | ... |
| 1 | +# coding=utf-8 | |
| 2 | +# author: resioR | |
| 3 | +#createtime: 2021/12/1 | |
| 4 | +#email: qianyingz@chinadci.com | |
| 5 | + | |
| 1 | 6 | # import schedule |
| 2 | 7 | from flask import json |
| 3 | -import paramiko | |
| 4 | -from sqlalchemy.sql.sqltypes import JSON | |
| 5 | -from .models import MonitorHost, MonitorLog | |
| 8 | +from .models import MonitorHost, MonitorInfo | |
| 6 | 9 | import datetime |
| 7 | 10 | import math |
| 8 | 11 | import time |
| ... | ... | @@ -13,6 +16,10 @@ from app.util.component.PGUtil import PGUtil |
| 13 | 16 | import configure |
| 14 | 17 | from app.util.component.StructuredPrint import StructurePrint |
| 15 | 18 | import traceback |
| 19 | +import requests | |
| 20 | + | |
| 21 | +min_size = 60 | |
| 22 | +size = 60 | |
| 16 | 23 | |
| 17 | 24 | |
| 18 | 25 | def pull_metric(): |
| ... | ... | @@ -22,25 +29,50 @@ def pull_metric(): |
| 22 | 29 | configure.SQLALCHEMY_DATABASE_URI) |
| 23 | 30 | sys_ds = PGUtil.open_pg_data_source( |
| 24 | 31 | 0, configure.SQLALCHEMY_DATABASE_URI) |
| 25 | - | |
| 26 | - #拉取服务器信息 | |
| 32 | + | |
| 33 | + # 拉取服务器信息 | |
| 27 | 34 | |
| 28 | 35 | hosts = sys_session.query( |
| 29 | 36 | MonitorHost.host) |
| 30 | 37 | for host in hosts: |
| 31 | - request_uri="http://{}".format(host.host) | |
| 32 | - | |
| 33 | - | |
| 34 | - | |
| 38 | + try: | |
| 39 | + host_name = host.host | |
| 40 | + request_uri = "http://{}".format(host_name) | |
| 41 | + response = requests.request("get", request_uri) | |
| 42 | + if response.status_code == 200: | |
| 43 | + text = response.text | |
| 44 | + data = json.loads(text) | |
| 45 | + len_metric = len(data) | |
| 46 | + if len_metric > min_size: | |
| 47 | + metric_data = data[len_metric-min_size:len_metric] | |
| 48 | + else: | |
| 49 | + continue | |
| 50 | + | |
| 51 | + type_list = ["cpu_per", "memory_per", "disk_per", | |
| 52 | + "disk_read", "disk_write","disk_read_count", | |
| 53 | + "disk_write_count","network_sent","network_recv"] | |
| 54 | + for type in type_list: | |
| 55 | + sample_data = get_sample_data( | |
| 56 | + metric_data, type, host_name) | |
| 57 | + sys_session.add_all(sample_data) | |
| 58 | + sys_session.commit() | |
| 59 | + | |
| 60 | + except Exception as e: | |
| 61 | + StructurePrint().print(e.__str__()+":" + traceback.format_exc(), "error") | |
| 35 | 62 | # 获取数据并汇聚为1min的数据入库 |
| 36 | 63 | # 结束 |
| 37 | - pass | |
| 38 | 64 | except Exception as e: |
| 39 | 65 | StructurePrint().print(e.__str__()+":" + traceback.format_exc(), "error") |
| 40 | - pass | |
| 66 | + finally: | |
| 67 | + sys_session.rollback() | |
| 68 | + if sys_session: | |
| 69 | + sys_session.close() | |
| 70 | + if sys_ds: | |
| 71 | + sys_ds.Destroy() | |
| 72 | + | |
| 41 | 73 | |
| 42 | 74 | def start_schedule(): |
| 43 | - # # 2分钟巡检一次 | |
| 75 | + # # 1分钟巡检一次 | |
| 44 | 76 | try: |
| 45 | 77 | StructurePrint().print("start_schedule") |
| 46 | 78 | schedule.every(1).minutes.do(pull_metric) |
| ... | ... | @@ -52,3 +84,54 @@ def start_schedule(): |
| 52 | 84 | # # Stop the background thread |
| 53 | 85 | # time.sleep(10) |
| 54 | 86 | # stop_run_continuously.set() |
| 87 | + | |
| 88 | + | |
| 89 | +def get_sample_data(orginal, name, host): | |
| 90 | + res = [] | |
| 91 | + size = len(orginal) | |
| 92 | + orginal_stamp = {'head': 1, 'tail': size} | |
| 93 | + if size > 1: | |
| 94 | + stamp = {'P0': 1, | |
| 95 | + 'P50': math.floor(0.5*size), | |
| 96 | + # 'P90': math.floor(0.9*size), | |
| 97 | + # 'P95': math.floor(0.95*size), | |
| 98 | + # 'P99': math.floor(0.99*size), | |
| 99 | + 'P100': size} | |
| 100 | + elif size == 1: | |
| 101 | + stamp = {'P0': 1, | |
| 102 | + 'P50': size, | |
| 103 | + # 'P90': size, | |
| 104 | + # 'P95': size, | |
| 105 | + # 'P99': size, | |
| 106 | + 'P100': size} | |
| 107 | + else: | |
| 108 | + return res | |
| 109 | + | |
| 110 | + for key in dict.keys(orginal_stamp): | |
| 111 | + cur_data = orginal[orginal_stamp[key]-1] | |
| 112 | + info = get_data(key, host, name, cur_data) | |
| 113 | + res.append(info) | |
| 114 | + | |
| 115 | + data = sorted(orginal, key=lambda x: x[name]) | |
| 116 | + for key in dict.keys(stamp): | |
| 117 | + cur_data = data[stamp[key]-1] | |
| 118 | + info = get_data(key, host, name, cur_data) | |
| 119 | + res.append(info) | |
| 120 | + | |
| 121 | + return res | |
| 122 | + | |
| 123 | + | |
| 124 | +def get_data(stamp, host, metrics_name, cur_data): | |
| 125 | + time_stamp = datetime.datetime.strptime( | |
| 126 | + cur_data['timestamp'], "%Y-%m-%d %H:%M:%S") | |
| 127 | + date_stamp = time_stamp.strftime("%Y-%m-%d") | |
| 128 | + guid = uuid.uuid1().__str__() | |
| 129 | + value = cur_data[metrics_name] | |
| 130 | + info = MonitorInfo(guid=guid, | |
| 131 | + server=host, | |
| 132 | + date_stamp=date_stamp, | |
| 133 | + time_stamp=time_stamp, | |
| 134 | + stamp=stamp, | |
| 135 | + value=value, | |
| 136 | + metrics=metrics_name) | |
| 137 | + return info | ... | ... |
请
注册
或
登录
后发表评论