1151 lines
49 KiB
Python
1151 lines
49 KiB
Python
# coding: utf-8
|
||
# -------------------------------------------------------------------
|
||
# YakPanel
|
||
# -------------------------------------------------------------------
|
||
# Copyright (c) 2015-2099 YakPanel(www.yakpanel.com) All rights reserved.
|
||
# -------------------------------------------------------------------
|
||
# Author: hwliang <hwl@yakpanel.com>
|
||
# -------------------------------------------------------------------
|
||
|
||
# ------------------------------
|
||
# 数据备份模块
|
||
# ------------------------------
|
||
import os
|
||
import sys
|
||
import json
|
||
import re
|
||
import time
|
||
import shlex
|
||
|
||
os.chdir('/www/server/panel')
|
||
if not 'class/' in sys.path:
|
||
sys.path.insert(0, 'class/')
|
||
import public
|
||
from public.hook_import import hook_import
|
||
hook_import()
|
||
|
||
_VERSION = 1.5
|
||
|
||
|
||
class backup:
|
||
_path = None
|
||
_exclude = ""
|
||
_err_log = '/tmp/backup_err.log'
|
||
_inode_min = 10
|
||
_db_mysql = None
|
||
_cloud = None
|
||
_is_save_local = os.path.exists('data/is_save_local_backup.pl')
|
||
_error_msg = ""
|
||
_backup_all = False
|
||
|
||
def __init__(self, cloud_object=None, cron_info={}):
|
||
'''
|
||
@name 数据备份对象
|
||
@param cloud_object 远程上传对象,需具备以下几个属性和方法:
|
||
_title = '中文名称,如:阿里云OSS'
|
||
_name = '英文名称,如:alioss'
|
||
|
||
upload_file(filename,data_type = None)
|
||
文件名 , 数据类型 site/database/path
|
||
|
||
delete_file(filename,data_type = None)
|
||
文件名 , 数据类型 site/database/path
|
||
|
||
给_error_msg赋值,传递错误消息:
|
||
_error_msg = "错误消息"
|
||
'''
|
||
self._cloud = cloud_object
|
||
self.cron_info = None
|
||
if cron_info and 'echo' in cron_info.keys():
|
||
self.cron_info = self.get_cron_info(cron_info["echo"])
|
||
|
||
self._path = public.M('config').where("id=?", (1,)).getField('backup_path')
|
||
|
||
if not public.M('sqlite_master').where('type=? AND name=? AND sql LIKE ?',
|
||
('table', 'backup', '%cron_id%')).count():
|
||
public.M('backup').execute("ALTER TABLE 'backup' ADD 'cron_id' INTEGER DEFAULT 0", ())
|
||
|
||
def echo_start(self):
|
||
print("=" * 90)
|
||
print("|-" + public.lang("Start backup") + "[{}]".format(public.format_date()))
|
||
print("=" * 90)
|
||
|
||
def echo_end(self):
|
||
print("=" * 90)
|
||
print("|-" + public.lang("Backup completed") + "[{}]".format(public.format_date()))
|
||
print("=" * 90)
|
||
print("\n")
|
||
|
||
def echo_info(self, msg):
|
||
print("|-{}".format(msg))
|
||
|
||
def echo_error(self, msg):
|
||
print("=" * 90)
|
||
print("|-Error:{}".format(msg))
|
||
if self._error_msg:
|
||
self._error_msg += "\n"
|
||
self._error_msg += msg
|
||
|
||
# 取排除列表用于计算排除目录大小
|
||
def get_exclude_list(self, exclude=[]):
|
||
if not exclude:
|
||
tmp_exclude = os.getenv('BT_EXCLUDE')
|
||
if tmp_exclude:
|
||
exclude = tmp_exclude.split(',')
|
||
if not exclude: return []
|
||
return exclude
|
||
|
||
# 构造排除
|
||
def get_exclude(self, exclude=[]):
|
||
self._exclude = ""
|
||
if not exclude:
|
||
tmp_exclude = os.getenv('BT_EXCLUDE')
|
||
if tmp_exclude:
|
||
exclude = tmp_exclude.split(',')
|
||
if not exclude: return ""
|
||
for ex in exclude:
|
||
if ex[-1] == '/': ex = ex[:-1]
|
||
self._exclude += " --exclude=\"" + ex + "\""
|
||
self._exclude += " "
|
||
return self._exclude
|
||
|
||
def GetDiskInfo2(self):
|
||
# 取磁盘分区信息
|
||
temp = public.ExecShell("df -T -P|grep '/'|grep -v tmpfs|grep -v 'snap/core'|grep -v udev|grep -v overlay")[0]
|
||
tempInodes = \
|
||
public.ExecShell("df -i -P|grep '/'|grep -v tmpfs|grep -v 'snap/core'|grep -v udev|grep -v overlay")[0]
|
||
temp1 = temp.split('\n')
|
||
tempInodes1 = tempInodes.split('\n')
|
||
diskInfo = []
|
||
n = 0
|
||
cuts = []
|
||
for tmp in temp1:
|
||
n += 1
|
||
try:
|
||
inodes = tempInodes1[n - 1].split()
|
||
disk = re.findall(r"^(.+)\s+([\w\.]+)\s+([\w\.]+)\s+([\w\.]+)\s+([\w\.]+)\s+([\d%]{2,4})\s+(/.{0,50})$",
|
||
tmp.strip())
|
||
if disk: disk = disk[0]
|
||
if len(disk) < 6: continue
|
||
if disk[2].find('M') != -1: continue
|
||
if disk[2].find('K') != -1: continue
|
||
if len(disk[6].split('/')) > 10: continue
|
||
if disk[6] in cuts: continue
|
||
if disk[6].find('docker') != -1: continue
|
||
if disk[1].strip() in ['tmpfs']: continue
|
||
arr = {}
|
||
arr['filesystem'] = disk[0].strip()
|
||
arr['type'] = disk[1].strip()
|
||
arr['path'] = disk[6]
|
||
tmp1 = [disk[2], disk[3], disk[4], disk[5]]
|
||
arr['size'] = tmp1
|
||
if int(inodes[1]) == 0 and int(inodes[2]) == 0:
|
||
arr['inodes'] = [inodes[1], 10000, 10000, 0]
|
||
else:
|
||
arr['inodes'] = [inodes[1], inodes[2], inodes[3], inodes[4]]
|
||
diskInfo.append(arr)
|
||
except:
|
||
continue
|
||
return diskInfo
|
||
|
||
# 取磁盘可用空间
|
||
def get_disk_free(self, dfile):
|
||
diskInfo = self.GetDiskInfo2()
|
||
if not diskInfo: return '', 0, 0
|
||
_root = None
|
||
for d in diskInfo:
|
||
if d['path'] == '/':
|
||
_root = d
|
||
continue
|
||
if re.match("^{}/.+".format(d['path']), dfile):
|
||
return d['path'], float(d['size'][2]) * 1024, int(d['inodes'][2])
|
||
if _root:
|
||
return _root['path'], float(_root['size'][2]) * 1024, int(_root['inodes'][2])
|
||
return '', 0, 0
|
||
|
||
# 备份指定目录
|
||
def backup_path(self, spath, dfile=None, exclude=[], save=3, echo_id=None):
|
||
|
||
error_msg = ""
|
||
self.echo_start()
|
||
if not os.path.exists(spath):
|
||
error_msg = public.lang('The specified directory {} does not exist!', spath)
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{spath}|path")
|
||
return False
|
||
|
||
if spath[-1] == '/':
|
||
spath = spath[:-1]
|
||
|
||
dirname = os.path.basename(spath)
|
||
if not dfile:
|
||
fname = 'path_{}_{}.tar.gz'.format(dirname, public.format_date("%Y%m%d_%H%M%S"))
|
||
# fname = 'path_{}_{}_{}.tar.gz'.format(dirname, public.format_date("%Y%m%d_%H%M%S"),
|
||
# public.GetRandomString(6))
|
||
dfile = os.path.join(self._path, 'path', fname)
|
||
|
||
if not self.backup_path_to(spath, dfile, exclude):
|
||
if self._error_msg:
|
||
error_msg = self._error_msg
|
||
self.send_failture_notification(error_msg, target=f"{spath}|path")
|
||
return False
|
||
|
||
if self._cloud:
|
||
self.echo_info(public.lang('Uploading to {}, please wait ...', self._cloud._title))
|
||
if self._cloud.upload_file(dfile, 'path'):
|
||
self.echo_info(public.lang('Successfully uploaded to {}', self._cloud._title))
|
||
else:
|
||
if hasattr(self._cloud, "error_msg"):
|
||
if self._cloud.error_msg:
|
||
error_msg = self._cloud.error_msg
|
||
if not error_msg:
|
||
error_msg = public.lang("File upload failed, skip this backup!")
|
||
self.echo_error(error_msg)
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
|
||
remark = "Backup to " + self._cloud._title
|
||
self.send_failture_notification(error_msg, target=f"{spath}|path", remark=remark)
|
||
return False
|
||
|
||
filename = dfile
|
||
if self._cloud:
|
||
filename = dfile + '|' + self._cloud._name + '|' + fname
|
||
cron_id = 0
|
||
if echo_id:
|
||
cron_id = public.M("crontab").where('echo=?', (echo_id,)).getField('id')
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': '2',
|
||
'name': spath,
|
||
'pid': 0,
|
||
'filename': filename,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
|
||
if self._cloud:
|
||
_not_save_local = True
|
||
save_local = 0
|
||
if self.cron_info:
|
||
save_local = self.cron_info["save_local"]
|
||
if save_local:
|
||
_not_save_local = False
|
||
else:
|
||
if self._is_save_local:
|
||
_not_save_local = False
|
||
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': '2',
|
||
'name': spath,
|
||
'pid': 0,
|
||
'filename': dfile,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
if _not_save_local:
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
self.echo_info(
|
||
public.lang('User settings do not retain local backups, deleted {}', dfile))
|
||
else:
|
||
self.echo_info(public.lang("Local backup has been kept"))
|
||
|
||
if not self._cloud:
|
||
backups = public.M('backup').where("cron_id=? and type=? and pid=? and name=? and filename NOT LIKE '%|%'",
|
||
(cron_id, '2', 0, spath)).field('id,name,filename').select()
|
||
else:
|
||
backups = public.M('backup').where("cron_id=? and type=? and pid=? and name=? and filename LIKE ?",
|
||
(cron_id, '2', 0, spath, '%{}%'.format(self._cloud._name))).field(
|
||
'id,name,filename').select()
|
||
|
||
self.delete_old(backups, save, 'path')
|
||
self.echo_end()
|
||
self.save_backup_status(True, target=f"{spath}|path")
|
||
return dfile
|
||
|
||
# 清理过期备份文件
|
||
def delete_old(self, backups, save, data_type=None, site_name=None):
|
||
if type(backups) == str:
|
||
self.echo_info(public.lang('Failed to clean expired backup, error: {}', backups))
|
||
return
|
||
self.echo_info(public.lang('Keep the latest number of backups: {} copies', str(save),))
|
||
|
||
# 跳过手动备份文件
|
||
new_backups = []
|
||
for i in range(len(backups)):
|
||
if data_type == 'database' and backups[i]['name'][:3] == 'db_': # 数据库备份
|
||
new_backups.append(backups[i])
|
||
elif data_type == 'site' and backups[i]['name'][:4] == 'web_' and backups[i]['name'][
|
||
-7:] == '.tar.gz': # 网站备份
|
||
new_backups.append(backups[i])
|
||
elif data_type == 'path' and backups[i]['name'][:5] == 'path_': # 目录备份
|
||
new_backups.append(backups[i])
|
||
if new_backups:
|
||
backups = new_backups[:]
|
||
num = len(backups) - int(save)
|
||
if num > 0:
|
||
# self._get_local_backdir()
|
||
self.echo_info('-' * 88)
|
||
for backup in backups:
|
||
# 处理目录备份到远程的情况
|
||
if backup['filename'].find('|') != -1:
|
||
tmp = backup['filename'].split('|')
|
||
backup['filename'] = tmp[0]
|
||
backup['name'] = tmp[-1]
|
||
# 尝试删除本地文件
|
||
if os.path.exists(backup['filename']):
|
||
try:
|
||
os.remove(backup['filename'])
|
||
self.echo_info(public.lang('Expired backup files have been cleaned from disk: {}',
|
||
backup['filename']))
|
||
except:
|
||
pass
|
||
# 尝试删除远程文件
|
||
if self._cloud:
|
||
self._cloud.delete_file(backup['name'], data_type)
|
||
self.echo_info(public.lang('Expired backup files have been cleaned from {}: {}',
|
||
self._cloud._title, backup['name']))
|
||
|
||
# 从数据库清理
|
||
public.M('backup').where('id=?', (backup['id'],)).delete()
|
||
num -= 1
|
||
if num < 1: break
|
||
# if data_type == 'site':
|
||
# backup_path = public.get_backup_path() + '/site'.replace('//', '/')
|
||
# site_lists = os.listdir(backup_path)
|
||
# file_info = []
|
||
# del_list = []
|
||
# check_name = 'web_{}_'.format(site_name)
|
||
# for site_v in site_lists:
|
||
# tmp_dict = {}
|
||
# if check_name == 'web__': continue
|
||
# if site_v.find(check_name) == -1: continue
|
||
# filename = os.path.join(backup_path, site_v)
|
||
# if os.path.isfile(filename):
|
||
# tmp_dict['name'] = filename
|
||
# tmp_dict['time'] = int(os.path.getmtime(filename))
|
||
# file_info.append(tmp_dict)
|
||
# if file_info and len(file_info) > int(save):
|
||
# file_info = sorted(file_info, key=lambda keys: keys['time'])
|
||
# del_list = file_info[:-int(save)]
|
||
# for del_file in del_list:
|
||
# if not del_file: continue
|
||
# if os.path.isfile(del_file['name']):
|
||
# os.remove(del_file['name'])
|
||
# self.echo_info(u"Expired backup files cleaned from disk:" + del_file['name'])
|
||
|
||
# 获取本地备份目录
|
||
# def _get_local_backdir(self):
|
||
# self._local_backdir = public.M('config').field('backup_path').find()['backup_path']
|
||
|
||
# 压缩目录
|
||
def backup_path_to(self, spath, dfile, exclude=[], siteName=None):
|
||
if not os.path.exists(spath):
|
||
self.echo_error(public.lang('The specified directory {} does not exist!', spath))
|
||
return False
|
||
|
||
if spath[-1] == '/':
|
||
spath = spath[:-1]
|
||
|
||
dirname = os.path.basename(spath)
|
||
dpath = os.path.dirname(dfile)
|
||
if not os.path.exists(dpath):
|
||
os.makedirs(dpath, 384)
|
||
self.get_exclude(exclude)
|
||
if self._exclude:
|
||
self._exclude = self._exclude.replace(spath + '/', '')
|
||
exclude_config = self._exclude
|
||
exclude_list = self.get_exclude_list(exclude)
|
||
p_size = public.get_path_size(spath, exclude=exclude_list)
|
||
if not self._exclude:
|
||
exclude_config = "Not set"
|
||
|
||
if siteName:
|
||
self.echo_info(public.lang('Backup site: {}', siteName))
|
||
self.echo_info(public.lang('Website document root: {}', spath))
|
||
else:
|
||
self.echo_info(public.lang('Backup directory: {}', spath))
|
||
|
||
self.echo_info(public.lang('Directory size: {}',str(public.to_size(p_size))))
|
||
self.echo_info(public.lang('Exclusion setting: {}', exclude_config))
|
||
disk_path, disk_free, disk_inode = self.get_disk_free(dfile)
|
||
self.echo_info(public.lang('Partition {} available disk space is: {}, available Inode is: {}',disk_path, str(public.to_size(disk_free)), str(disk_inode)))
|
||
if disk_path:
|
||
if disk_free < p_size:
|
||
self.echo_error(public.lang('The available disk space of the target partition is less than {}, and the backup cannot be completed. Please increase the disk capacity or change the default backup directory on the settings page!',str(public.to_size(p_size))))
|
||
return False
|
||
|
||
if disk_inode < self._inode_min:
|
||
self.echo_error(public.lang('The available Inode of the target partition is less than {}, and the backup cannot be completed. Please increase the disk capacity or change the default backup directory on the settings page!',str(self._inode_min) ))
|
||
return False
|
||
|
||
stime = time.time()
|
||
self.echo_info(public.lang('Start compressing files: {}', public.format_date(times=stime)))
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
public.ExecShell("cd " + os.path.dirname(
|
||
spath) + " && tar zcvf '" + dfile + "' " + self._exclude + " '" + dirname + "' 2>{err_log} 1> /dev/null".format(
|
||
err_log=self._err_log))
|
||
tar_size = os.path.getsize(dfile)
|
||
if tar_size < 1:
|
||
self.echo_error(public.lang("Compression failed!"))
|
||
self.echo_info(public.readFile(self._err_log))
|
||
return False
|
||
compression_time = str('{:.2f}'.format(time.time() - stime))
|
||
self.echo_info(public.lang('Compression completed, took {} seconds, compressed package size: {}',compression_time, str(public.to_size(tar_size))))
|
||
if siteName:
|
||
self.echo_info(public.lang('Site backed up to: {}', dfile))
|
||
else:
|
||
self.echo_info(public.lang('Directory has been backed up to: {}', dfile))
|
||
if os.path.exists(self._err_log):
|
||
os.remove(self._err_log)
|
||
return dfile
|
||
|
||
# 备份指定站点
|
||
def backup_site(self, siteName, save=3, exclude=[], echo_id=None):
|
||
try:
|
||
self.echo_start()
|
||
find = public.M('sites').where('name=?', (siteName,)).field('id,path,project_type').find()
|
||
public.print_log(find)
|
||
if not find or not isinstance(find, dict):
|
||
raise Exception(' The directory for does not exist')
|
||
|
||
# Wordpress
|
||
if find['project_type'] == 'WP2':
|
||
try:
|
||
from wp_toolkit import wpbackup
|
||
bak_info = wpbackup(find['id']).backup_full_get_data()
|
||
self.echo_info(public.lang('Backup wordpress [{}] successfully', siteName))
|
||
self.echo_end()
|
||
return bak_info.bak_file
|
||
except Exception as e:
|
||
public.print_error()
|
||
self.echo_info(str(e))
|
||
self.echo_end()
|
||
return None
|
||
|
||
spath = find['path']
|
||
pid = find['id']
|
||
fname = 'web_{}_{}.tar.gz'.format(siteName, public.format_date("%Y%m%d_%H%M%S"))
|
||
|
||
# fname = 'web_{}_{}_{}.tar.gz'.format(siteName, public.format_date("%Y%m%d_%H%M%S"),
|
||
# public.GetRandomString(6))
|
||
|
||
dfile = os.path.join(self._path, 'site', fname)
|
||
error_msg = ""
|
||
if not self.backup_path_to(spath, dfile, exclude, siteName=siteName):
|
||
if self._error_msg:
|
||
error_msg = self._error_msg
|
||
self.send_failture_notification(error_msg, target=f"{siteName}|site")
|
||
return False
|
||
|
||
if self._cloud:
|
||
self.echo_info(public.lang('Uploading to {}, please wait ...', self._cloud._title))
|
||
if self._cloud.upload_file(dfile, 'site'):
|
||
self.echo_info(public.lang('Successfully uploaded to {}', self._cloud._title))
|
||
else:
|
||
if hasattr(self._cloud, "error_msg"):
|
||
if self._cloud.error_msg:
|
||
error_msg = self._cloud.error_msg
|
||
if not error_msg:
|
||
error_msg = public.lang("File upload failed, skip this backup!")
|
||
self.echo_error(error_msg)
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
|
||
remark = "Backup to " + self._cloud._title
|
||
self.send_failture_notification(error_msg, target=f"{siteName}|site", remark=remark)
|
||
return False
|
||
|
||
filename = dfile
|
||
if self._cloud:
|
||
filename = dfile + '|' + self._cloud._name + '|' + fname
|
||
cron_id = 0
|
||
if echo_id:
|
||
cron_id = public.M("crontab").where('echo=?', (echo_id,)).getField('id')
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': 0,
|
||
'name': fname,
|
||
'pid': pid,
|
||
'filename': filename,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
if self._cloud:
|
||
_not_save_local = True
|
||
save_local = 0
|
||
if self.cron_info:
|
||
save_local = self.cron_info["save_local"]
|
||
if save_local:
|
||
_not_save_local = False
|
||
else:
|
||
if self._is_save_local:
|
||
_not_save_local = False
|
||
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': 0,
|
||
'name': fname,
|
||
'pid': pid,
|
||
'filename': dfile,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
|
||
if _not_save_local:
|
||
if os.path.exists(dfile):
|
||
print(dfile)
|
||
os.remove(dfile)
|
||
self.echo_info(
|
||
public.lang('User settings do not retain local backups, deleted {}', dfile))
|
||
else:
|
||
self.echo_info(public.lang("Local backup has been kept"))
|
||
|
||
# 清理多余备份
|
||
if not self._cloud:
|
||
backups = public.M('backup').where("cron_id=? and type=? and pid=? and filename NOT LIKE '%|%'",
|
||
(cron_id, '0', pid)).field('id,name,filename').select()
|
||
else:
|
||
backups = public.M('backup').where('cron_id=? and type=? and pid=? and filename LIKE ?',
|
||
(cron_id, '0', pid, "%{}%".format(self._cloud._name))).field(
|
||
'id,name,filename').select()
|
||
|
||
self.delete_old(backups, save, 'site', siteName)
|
||
self.echo_end()
|
||
return dfile
|
||
except Exception as e:
|
||
self.send_failture_notification('site {} {}'.format(siteName, e), target=siteName)
|
||
self.echo_error('site {} {}'.format(siteName, e))
|
||
|
||
# 备份所有数据库
|
||
def backup_database_all(self, save=3, echo_id=None):
|
||
databases = public.M('databases').where("type=?", "MySQL").field('name').select()
|
||
self._backup_all = True
|
||
failture_count = 0
|
||
results = []
|
||
for database in databases:
|
||
self._error_msg = ""
|
||
result = self.backup_database(database['name'], save=save, echo_id=echo_id)
|
||
if not result:
|
||
failture_count += 1
|
||
results.append((database['name'], result, self._error_msg,))
|
||
self.save_backup_status(result, target=f"{database['name']}|database", msg=self._error_msg)
|
||
|
||
if failture_count > 0:
|
||
self.send_all_failture_notification("database", results)
|
||
self._backup_all = False
|
||
|
||
# 备份所有站点
|
||
def backup_site_all(self, save=3, echo_id=None):
|
||
sites = public.M('sites').field('name').select()
|
||
self._backup_all = True
|
||
failture_count = 0
|
||
results = []
|
||
for site in sites:
|
||
self._error_msg = ""
|
||
result = self.backup_site(site['name'], save, echo_id=echo_id)
|
||
if not result:
|
||
failture_count += 1
|
||
results.append((site['name'], result, self._error_msg,))
|
||
self.save_backup_status(result, target=f"{site['name']}|site", msg=self._error_msg)
|
||
|
||
if failture_count > 0:
|
||
self.send_all_failture_notification("site", results)
|
||
self._backup_all = False
|
||
|
||
# 配置
|
||
def mypass(self, act):
|
||
conf_file = '/etc/my.cnf'
|
||
conf_file_bak = '/etc/my.cnf.bak'
|
||
if os.path.getsize(conf_file) > 2:
|
||
public.writeFile(conf_file_bak, public.readFile(conf_file))
|
||
public.set_mode(conf_file_bak, 600)
|
||
public.set_own(conf_file_bak, 'mysql')
|
||
elif os.path.getsize(conf_file_bak) > 2:
|
||
public.writeFile(conf_file, public.readFile(conf_file_bak))
|
||
public.set_mode(conf_file, 600)
|
||
public.set_own(conf_file, 'mysql')
|
||
|
||
public.ExecShell("sed -i '/user=root/d' {}".format(conf_file))
|
||
public.ExecShell("sed -i '/password=/d' {}".format(conf_file))
|
||
if act:
|
||
password = public.M('config').where('id=?', (1,)).getField('mysql_root')
|
||
mycnf = public.readFile(conf_file)
|
||
if not mycnf: return False
|
||
src_dump_re = r"\[mysqldump\][^.]"
|
||
sub_dump = "[mysqldump]\nuser=root\npassword=\"{}\"\n".format(password)
|
||
mycnf = re.sub(src_dump_re, sub_dump, mycnf)
|
||
if len(mycnf) > 100: public.writeFile(conf_file, mycnf)
|
||
return True
|
||
return True
|
||
|
||
# map to list
|
||
def map_to_list(self, map_obj):
|
||
try:
|
||
if type(map_obj) != list and type(map_obj) != str: map_obj = list(map_obj)
|
||
return map_obj
|
||
except:
|
||
return []
|
||
|
||
# 备份指定数据库
|
||
def backup_database(self, db_name, dfile=None, save=3, echo_id=None):
|
||
try:
|
||
self.echo_start()
|
||
if not dfile:
|
||
fname = 'db_{}_{}.sql.gz'.format(db_name, public.format_date("%Y%m%d_%H%M%S"))
|
||
# fname = 'db_{}_{}_{}.sql.gz'.format(db_name, public.format_date("%Y%m%d_%H%M%S"),
|
||
# public.GetRandomString(6))
|
||
dfile = os.path.join(self._path, 'database', fname)
|
||
else:
|
||
fname = os.path.basename(dfile)
|
||
|
||
dpath = os.path.dirname(dfile)
|
||
if not os.path.exists(dpath):
|
||
os.makedirs(dpath, 384)
|
||
|
||
error_msg = ""
|
||
# ----- 判断是否为远程数据库START @author hwliang<2021-01-08>--------
|
||
db_find = public.M('databases').where("name=?", (db_name,)).find()
|
||
if db_find['type'] != "MySQL":
|
||
# if db_find['type'] in ['SQLServer','Redis']:
|
||
print("|-{}The database does not currently support backup".format(db_find['type']))
|
||
return False
|
||
# if db_find['type'] == "MongoDB":
|
||
# import databaseModel.mongodbModel as mongodbModel
|
||
# args = public.dict_obj()
|
||
# args.id = db_find['id']
|
||
# args.name = db_find['name']
|
||
# backup_res = mongodbModel.panelMongoDB().ToBackup(args)
|
||
# elif db_find['type'] == "PgSQL":
|
||
# import databaseModel.pgsqlModel as pgsqlModel
|
||
# args = public.dict_obj()
|
||
# args.id = db_find['id']
|
||
# args.name = db_find['name']
|
||
# backup_res = pgsqlModel.panelPgsql().ToBackup(args)
|
||
|
||
# if not isinstance(backup_res,dict) or not 'status' in backup_res:
|
||
|
||
# return False
|
||
|
||
# if not backup_res['status']:
|
||
# print("|-{}数据库备份失败".format(db_find['name']))
|
||
# print("|-{}".format(backup_res['msg']))
|
||
# return False
|
||
# dfile = dfile = os.path.join(self._path,'database',,fname)
|
||
# print("|-{}数据库备份成功".format(db_find['name']))
|
||
# print("|-数据库已备份到:{}".format(dfile))
|
||
|
||
# return False
|
||
conn_config = {}
|
||
self._db_mysql = public.get_mysql_obj(db_name)
|
||
is_cloud_db = db_find['db_type'] in ['1', 1, '2', 2]
|
||
if is_cloud_db:
|
||
# 连接远程数据库
|
||
if db_find['sid']:
|
||
conn_config = public.M('database_servers').where('id=?', db_find['sid']).find()
|
||
if not 'db_name' in conn_config: conn_config['db_name'] = None
|
||
else:
|
||
conn_config = json.loads(db_find['conn_config'])
|
||
conn_config['db_port'] = str(int(conn_config['db_port']))
|
||
if not self._db_mysql or not self._db_mysql.set_host(conn_config['db_host'],
|
||
int(conn_config['db_port']),
|
||
conn_config['db_name'], conn_config['db_user'],
|
||
conn_config['db_password']):
|
||
error_msg = "Failed to connect to remote database [{}:{}]".format(conn_config['db_host'],
|
||
conn_config['db_port'])
|
||
print(error_msg)
|
||
return False
|
||
# ----- 判断是否为远程数据库END @author hwliang<2021-01-08>------------
|
||
d_tmp = self._db_mysql.query(
|
||
"select sum(DATA_LENGTH)+sum(INDEX_LENGTH) from information_schema.tables where table_schema='%s'" % db_name)
|
||
try:
|
||
p_size = self.map_to_list(d_tmp)[0][0]
|
||
except:
|
||
error_msg = public.get_msg_gettext(
|
||
'The database connection is abnormal. Please check whether the root user authority or database configuration parameters are correct.')
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database")
|
||
return False
|
||
|
||
if p_size == None:
|
||
error_msg = public.get_msg_gettext('The specified database [ {} ] has no data!', (db_name,))
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database")
|
||
return False
|
||
|
||
character = public.get_database_character(db_name)
|
||
|
||
self.echo_info(public.get_msg_gettext('Backup database:{}', (db_name,)))
|
||
self.echo_info(public.get_msg_gettext('Database size: {}', (public.to_size(p_size),)))
|
||
self.echo_info(public.get_msg_gettext('Database character set: {}', (character,)))
|
||
disk_path, disk_free, disk_inode = self.get_disk_free(dfile)
|
||
self.echo_info(public.get_msg_gettext(
|
||
'Partition {} available disk space is: {}, available Inode is: {}', (
|
||
disk_path, str(public.to_size(disk_free)), str(disk_inode)
|
||
)
|
||
))
|
||
if disk_path:
|
||
if disk_free < p_size:
|
||
error_msg = public.get_msg_gettext(
|
||
'The available disk space of the target partition is less than {}, and the backup cannot be completed. Please increase the disk capacity or change the default backup directory on the settings page!',
|
||
(
|
||
str(public.to_size(p_size), )
|
||
))
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database")
|
||
return False
|
||
|
||
if disk_inode < self._inode_min:
|
||
error_msg = public.get_msg_gettext(
|
||
'The available Inode of the target partition is less than {}, and the backup cannot be completed. Please increase the disk capacity or change the default backup directory on the settings page!',
|
||
(self._inode_min,))
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database")
|
||
return False
|
||
|
||
stime = time.time()
|
||
self.echo_info(public.get_msg_gettext('Start exporting database: {}', (public.format_date(times=stime),)))
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
# self.mypass(True)
|
||
mysqldump_bin = public.get_mysqldump_bin()
|
||
try:
|
||
if not is_cloud_db:
|
||
# 本地数据库 @author hwliang<2021-01-08>
|
||
password = public.M('config').where('id=?', (1,)).getField('mysql_root')
|
||
password = public.shell_quote(str(password))
|
||
os.environ["MYSQL_PWD"] = password
|
||
backup_cmd = mysqldump_bin + " -E -R --default-character-set=" + character + " --force --hex-blob --opt " + db_name + " -u root -p" + password + " 2>" + self._err_log + "| gzip > " + dfile
|
||
else:
|
||
# 远程数据库 @author hwliang<2021-01-08>
|
||
password = public.shell_quote(str(conn_config['db_password']))
|
||
os.environ["MYSQL_PWD"] = password
|
||
backup_cmd = mysqldump_bin + " -h " + conn_config['db_host'] + " -P " + str(conn_config[
|
||
'db_port']) + " -E -R --default-character-set=" + character + " --force --hex-blob --opt " + db_name + " -u " + str(
|
||
conn_config['db_user']) + " -p" + password + " 2>" + self._err_log + "| gzip > " + dfile
|
||
public.ExecShell(backup_cmd)
|
||
except Exception as e:
|
||
raise
|
||
finally:
|
||
os.environ["MYSQL_PWD"] = ""
|
||
# public.ExecShell("/www/server/mysql/bin/mysqldump --default-character-set="+ character +" --force --hex-blob --opt " + db_name + " 2>"+self._err_log+"| gzip > " + dfile)
|
||
# self.mypass(False)
|
||
gz_size = os.path.getsize(dfile)
|
||
if gz_size < 400:
|
||
error_msg = public.lang("Database export failed!")
|
||
self.echo_error(error_msg)
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database")
|
||
self.echo_info(public.readFile(self._err_log))
|
||
return False
|
||
compressed_time = str('{:.2f}'.format(time.time() - stime))
|
||
self.echo_info(
|
||
public.get_msg_gettext('Compression completed, took {} seconds, compressed package size: {}',
|
||
(str(compressed_time),
|
||
str(public.to_size(gz_size))
|
||
))
|
||
)
|
||
if self._cloud:
|
||
self.echo_info(public.get_msg_gettext('Uploading to {}, please wait ...', (self._cloud._title,)))
|
||
if self._cloud.upload_file(dfile, 'database'):
|
||
self.echo_info(public.get_msg_gettext('Successfully uploaded to {}', (self._cloud._title,)))
|
||
else:
|
||
if hasattr(self._cloud, "error_msg"):
|
||
if self._cloud.error_msg:
|
||
error_msg = self._cloud.error_msg
|
||
if not error_msg:
|
||
error_msg = public.lang("File upload failed, skip this backup!")
|
||
self.echo_error(error_msg)
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
|
||
remark = "Backup to " + self._cloud._title
|
||
self.send_failture_notification(error_msg, target=f"{db_name}|database", remark=remark)
|
||
return False
|
||
|
||
filename = dfile
|
||
if self._cloud:
|
||
filename = dfile + '|' + self._cloud._name + '|' + fname
|
||
self.echo_info(public.get_msg_gettext('Database has been backed up to: {}', (dfile,)))
|
||
if os.path.exists(self._err_log):
|
||
os.remove(self._err_log)
|
||
|
||
pid = public.M('databases').where('name=?', (db_name)).getField('id')
|
||
cron_id = 0
|
||
if echo_id:
|
||
cron_id = public.M("crontab").where('echo=?', (echo_id,)).getField('id')
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': '1',
|
||
'name': fname,
|
||
'pid': pid,
|
||
'filename': filename,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
|
||
if self._cloud:
|
||
_not_save_local = True
|
||
save_local = 0
|
||
if self.cron_info:
|
||
save_local = self.cron_info["save_local"]
|
||
if save_local:
|
||
_not_save_local = False
|
||
else:
|
||
if self._is_save_local:
|
||
_not_save_local = False
|
||
|
||
pdata = {
|
||
'cron_id': cron_id,
|
||
'type': '1',
|
||
'name': fname,
|
||
'pid': pid,
|
||
'filename': dfile,
|
||
'addtime': public.format_date(),
|
||
'size': os.path.getsize(dfile)
|
||
}
|
||
public.M('backup').insert(pdata)
|
||
|
||
if _not_save_local:
|
||
if os.path.exists(dfile):
|
||
os.remove(dfile)
|
||
self.echo_info(
|
||
public.get_msg_gettext('User settings do not retain local backups, deleted {}', (dfile,)))
|
||
else:
|
||
self.echo_info(public.lang("Local backup has been kept"))
|
||
|
||
# 清理多余备份
|
||
if not self._cloud:
|
||
backups = public.M('backup').where("cron_id=? and type=? and pid=? and filename NOT LIKE '%|%'",
|
||
(cron_id, '1', pid)).field('id,name,filename').select()
|
||
else:
|
||
backups = public.M('backup').where('cron_id=? and type=? and pid=? and filename LIKE ?',
|
||
(cron_id, '1', pid, "%{}%".format(self._cloud._name))).field(
|
||
'id,name,filename').select()
|
||
|
||
self.delete_old(backups, save, 'database')
|
||
self.echo_end()
|
||
self.save_backup_status(True, target=f"{db_name}|database")
|
||
return dfile
|
||
except:
|
||
self.send_failture_notification(public.get_msg_gettext('Database {} does not exist', (db_name,)), target=db_name)
|
||
self.echo_error(public.get_msg_gettext('Database {} does not exist', (db_name,)))
|
||
|
||
def generate_success_title(self, task_name):
|
||
from send_mail import send_mail
|
||
sm = send_mail()
|
||
now = public.format_date(format="%Y-%m-%d %H:%M")
|
||
server_ip = sm.GetLocalIp()
|
||
title = public.get_msg_gettext('{}-{} The task was executed successfully', (server_ip, task_name))
|
||
return title
|
||
|
||
def generate_failture_title(self, task_name):
|
||
title = "YakPanel backup task failed reminder".format(task_name)
|
||
return title
|
||
|
||
def generate_all_failture_notice(self, task_name, msg, backup_type, remark=""):
|
||
from send_mail import send_mail
|
||
sm = send_mail()
|
||
now = public.format_date(format="%Y-%m-%d %H:%M:%S")
|
||
server_ip = sm.GetLocalIp()
|
||
if remark:
|
||
remark = "\n* Task notes: {}".format(remark)
|
||
|
||
notice_content = """*YakPanel reminds you that the cron failed to execute*
|
||
* Server IP*: {}
|
||
* Time*: {}
|
||
* Task name*: {} {}
|
||
* The following is a list of {} that failed to backup*:
|
||
{}
|
||
--Notification by YakPanel""".format(
|
||
server_ip, now, task_name, remark, backup_type, msg)
|
||
|
||
# tg_content = """📣‼*YakPanel reminds you that the cron failed to execute*‼
|
||
#
|
||
# * Server IP*: {}
|
||
# * Time*: {}
|
||
# * Task name*: {} {}
|
||
# * The following is a list of {} that failed to backup*:
|
||
# {}
|
||
# --Notification by YakPanel""".format(
|
||
# server_ip, now, task_name, remark, backup_type, msg)
|
||
# return {"mail":notice_content,"tg":tg_content}
|
||
return notice_content
|
||
|
||
def generate_failture_notice(self, task_name, msg, remark):
|
||
from send_mail import send_mail
|
||
sm = send_mail()
|
||
now = public.format_date(format="%Y-%m-%d %H:%M:%S")
|
||
server_ip = sm.GetLocalIp()
|
||
if remark:
|
||
remark = "\n* Task notes: {}".format(remark)
|
||
|
||
notice_content = """*YakPanel reminds you that the cron failed to execute*
|
||
* Server IP*: {}
|
||
* Time*: {}
|
||
* Task name*: {} {}
|
||
* error message*:
|
||
{}
|
||
--Notification by YakPanel""".format(
|
||
server_ip, now, task_name, remark, msg)
|
||
|
||
# tg_content = """📣‼*YakPanel reminds you that the cron failed to execute*‼
|
||
#
|
||
# * Server IP*: {}
|
||
# * Time*: {}
|
||
# * Task name*: {} {}
|
||
# * Error messages*:
|
||
# {}
|
||
#
|
||
# -- Notification by YakPanel""".format(
|
||
# server_ip, now, task_name, remark, msg)
|
||
# return {'mail':notice_content,'tg':tg_content}
|
||
return notice_content
|
||
|
||
def get_cron_info(self, cron_name):
|
||
""" 通过计划任务名称查找计划任务配置参数 """
|
||
try:
|
||
cron_info = public.M('crontab').where('echo=?', (cron_name,)) \
|
||
.field('name,save_local,notice,notice_channel,id,sType').find()
|
||
return cron_info
|
||
except Exception as e:
|
||
pass
|
||
return {}
|
||
|
||
def send_success_notification(self, msg, target="", remark=""):
|
||
pass
|
||
|
||
def send_failture_notification(self, error_msg, target="", remark=""):
|
||
|
||
"""发送任务失败消息
|
||
|
||
:error_msg 错误信息
|
||
:remark 备注
|
||
"""
|
||
if self._backup_all:
|
||
|
||
return
|
||
|
||
if not self.cron_info:
|
||
|
||
return
|
||
cron_info = self.cron_info
|
||
cron_title = cron_info["name"]
|
||
save_local = cron_info["save_local"]
|
||
notice = cron_info["notice"]
|
||
notice_channel = cron_info["notice_channel"]
|
||
|
||
|
||
self.save_backup_status(False, target, msg=error_msg)
|
||
if notice == 0 or not notice_channel:
|
||
return
|
||
|
||
if notice == 1 or notice == 2:
|
||
title = self.generate_failture_title(cron_title)
|
||
task_name = cron_title
|
||
msg = self.generate_failture_notice(task_name, error_msg, remark)
|
||
res = self.send_notification(notice_channel, title, msg)
|
||
if res:
|
||
self.echo_info(public.lang("Notification has been sent"))
|
||
|
||
def send_all_failture_notification(self, backup_type, results, remark=""):
|
||
|
||
"""统一发送任务失败消息
|
||
|
||
:results [(备份对象, 备份结果,错误信息),...]
|
||
:remark 备注
|
||
"""
|
||
if not self.cron_info:
|
||
return
|
||
cron_info = self.cron_info
|
||
cron_title = cron_info["name"]
|
||
save_local = cron_info["save_local"]
|
||
notice = cron_info["notice"]
|
||
notice_channel = cron_info["notice_channel"]
|
||
if notice == 0 or not notice_channel:
|
||
return
|
||
|
||
if notice == 1 or notice == 2:
|
||
title = self.generate_failture_title(cron_title)
|
||
type_desc = {
|
||
"site": "site",
|
||
"database": "database"
|
||
}
|
||
backup_type_desc = type_desc[backup_type]
|
||
task_name = cron_title
|
||
failture_count = 0
|
||
total = 0
|
||
content = ""
|
||
|
||
for obj in results:
|
||
total += 1
|
||
obj_name = obj[0]
|
||
result = obj[1]
|
||
if not result:
|
||
failture_count += 1
|
||
content += "{}、".format(obj_name)
|
||
# content += "<tr><td style='color:red'>{}</td><tr>".format(obj_name)
|
||
|
||
if failture_count > 0:
|
||
if self._cloud:
|
||
remark = public.lang("Backup to {}, a total of {} {}, and failures {}."), (
|
||
self._cloud._title, total, backup_type_desc, failture_count)
|
||
else:
|
||
remark = public.lang("Backup failed {}/total {} sites"), (
|
||
failture_count, total, backup_type_desc)
|
||
|
||
msg = self.generate_all_failture_notice(task_name, content, backup_type_desc, remark)
|
||
res = self.send_notification(notice_channel, title, msg=msg, total=total, failture_count=failture_count)
|
||
if res:
|
||
self.echo_info(public.lang("Notification has been sent"))
|
||
else:
|
||
self.echo_error(public.lang("Failed to send notification"))
|
||
|
||
def send_notification(self, channel, title, msg="", total=0, failture_count=0):
|
||
|
||
"""发送通知
|
||
|
||
Args:
|
||
channel (str): 消息通道,多个用英文逗号隔开
|
||
title (str): 通知标题
|
||
msg (str, optional): 消息内容. Defaults to "".
|
||
|
||
Returns:
|
||
bool: 通知是否发送成功
|
||
"""
|
||
try:
|
||
from config import config
|
||
from panelPush import panelPush
|
||
tongdao = []
|
||
if channel.find(",") >= 0:
|
||
tongdao = channel.split(",")
|
||
else:
|
||
tongdao = [channel]
|
||
|
||
error_count = 0
|
||
con_obj = config()
|
||
get = public.dict_obj()
|
||
msg_channels = con_obj.get_msg_configs(get)
|
||
|
||
error_channel = []
|
||
channel_data = {}
|
||
msg_data = {}
|
||
for ch in tongdao:
|
||
# 根据不同的消息通道准备不同的内容
|
||
if ch == "mail":
|
||
msg_data = {
|
||
"msg": msg.replace("\n", "<br/>"),
|
||
"title": title
|
||
}
|
||
if ch in ["dingding", "weixin", "feishu", "wx_account", "tg"]:
|
||
msg_data["msg"] = msg
|
||
if ch in ["sms"]:
|
||
if total > 0 and failture_count > 0:
|
||
msg_data["sm_type"] = "backup_all"
|
||
msg_data["sm_args"] = {
|
||
"panel_name": public.GetConfigValue('title'),
|
||
"task_name": self.cron_info["name"],
|
||
"failed_count": failture_count,
|
||
"total": total
|
||
}
|
||
else:
|
||
msg_data["sm_type"] = "backup"
|
||
msg_data["sm_args"] = {
|
||
"panel_name": public.GetConfigValue('title'),
|
||
"task_name": self.cron_info["name"]
|
||
}
|
||
channel_data[ch] = msg_data
|
||
# print("channel data:")
|
||
# print(channel_data)
|
||
# 即时推送
|
||
pp = panelPush()
|
||
push_res = pp.push_message_immediately(channel_data)
|
||
if push_res["status"]:
|
||
channel_res = push_res["msg"]
|
||
for ch, res in channel_res.items():
|
||
if not res["status"]:
|
||
if ch in msg_channels:
|
||
error_channel.append(msg_channels[ch]["title"])
|
||
error_count += 1
|
||
if not push_res["status"] or error_count:
|
||
self.echo_error("Notification:{} Failed send!".format(",".join(error_channel)))
|
||
else:
|
||
self.echo_info("Message sent successfully.")
|
||
if error_count == len(tongdao):
|
||
return False
|
||
return True
|
||
except Exception as e:
|
||
import traceback
|
||
print(traceback.format_exc())
|
||
return False
|
||
|
||
def send_notification2(self, channel, title, msg={}):
|
||
try:
|
||
from send_mail import send_mail
|
||
from config import config
|
||
tondao = []
|
||
if channel.find(",") >= 0:
|
||
tongdao = channel.split(",")
|
||
else:
|
||
tongdao = [channel]
|
||
|
||
sm = send_mail()
|
||
c = config()
|
||
send_res = []
|
||
error_count = 0
|
||
channel_names = {
|
||
"mail": "email",
|
||
"telegram": "telegram"
|
||
}
|
||
error_channel = []
|
||
# settings = sm.get_settings()
|
||
settings = c.get_settings2()
|
||
for td in tongdao:
|
||
_res = False
|
||
if td == "mail":
|
||
if len(settings["user_mail"]['mail_list']) == 0:
|
||
continue
|
||
mail_list = settings['user_mail']['mail_list']
|
||
if len(mail_list) == 1:
|
||
mail_list = mail_list[0]
|
||
_res = sm.qq_smtp_send(mail_list, title=title, body=msg['mail'].replace("\n", "<br/>"))
|
||
if not _res:
|
||
error_count += 1
|
||
error_channel.append(channel_names[td])
|
||
if td == "telegram":
|
||
import panel_telegram_bot
|
||
if not settings["telegram"]['setup']:
|
||
continue
|
||
_res = panel_telegram_bot.panel_telegram_bot().send_by_tg_bot(msg['tg'])
|
||
send_res.append(_res)
|
||
if not _res:
|
||
error_count += 1
|
||
error_channel.append(channel_names[td])
|
||
if error_count > 0:
|
||
print("Notification:{} failed to send".format(",".join(error_channel)))
|
||
if error_count == len(tongdao):
|
||
return False
|
||
return True
|
||
except Exception as e:
|
||
print(e)
|
||
return False
|
||
|
||
def save_backup_status(self, status, target="", msg=""):
|
||
"""保存备份的状态"""
|
||
try:
|
||
if not self.cron_info:
|
||
return
|
||
cron_id = self.cron_info["id"]
|
||
sql = public.M("system").dbfile("system").table("backup_status")
|
||
sql.add("id,target,status,msg,addtime", (cron_id, target, status, msg, time.time(),))
|
||
except Exception as e:
|
||
print("Backup status saving error :{}.".format(e))
|
||
|