import os import json from grafana_backup.dashboardApi import search_datasource from grafana_backup.commons import print_horizontal_line def main(args, settings): backup_dir = settings.get('BACKUP_DIR') timestamp = settings.get('TIMESTAMP') grafana_url = settings.get('GRAFANA_URL') http_get_headers = settings.get('HTTP_GET_HEADERS') verify_ssl = settings.get('VERIFY_SSL') debug = settings.get('DEBUG') folder_path = '{0}/datasources/{1}'.format(backup_dir, timestamp) log_file = 'datasources_{0}.txt'.format(timestamp) if not os.path.exists(folder_path): os.makedirs(folder_path) datasources = get_all_datasources_and_save(folder_path, grafana_url, http_get_headers, verify_ssl, debug) print_horizontal_line() def save_datasource(file_name, datasource_setting, folder_path): file_path = folder_path + '/' + file_name + '.datasource' with open(file_path, 'w') as f: f.write(json.dumps(datasource_setting)) print("datasource:{0} is saved to {1}".format(file_name, file_path)) def get_all_datasources_and_save(folder_path, grafana_url, http_get_headers, verify_ssl, debug): status_code_and_content = search_datasource(grafana_url, http_get_headers, verify_ssl, debug) if status_code_and_content[0] == 200: datasources = status_code_and_content[1] print("There are {0} datasources:".format(len(datasources))) for datasource in datasources: print(datasource) save_datasource(datasource['name'], datasource, folder_path) else: print("query datasource failed, status: {0}, msg: {1}".format(status_code_and_content[0], status_code_and_content[1]))