Merge branch 'superscript-v1' into event-listener

This commit is contained in:
Arthur Lu 2021-10-10 19:52:47 -07:00 committed by GitHub
commit 3dbdcfbd35
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
5 changed files with 338 additions and 243 deletions

View File

@ -5,15 +5,31 @@ name: Build Superscript Linux
on: on:
release: release:
types: [published, created, edited] types: [published, edited]
jobs: jobs:
generate: generate:
name: Build Linux name: Build Linux
runs-on: ubuntu-latest runs-on: ubuntu-latest
steps: steps:
- name: Checkout master - name: Checkout master
uses: actions/checkout@master uses: actions/checkout@master
- name: Echo test - name: Install Dependencies
run: echo "test" run: pip install -r requirements.txt
working-directory: src/
- name: Give Execute Permission
run: chmod +x build-CLI.sh
working-directory: build/
- name: Build Binary
run: ./build-CLI.sh
working-directory: build/
- name: Copy Binary to Root Dir
run: cp superscript ..
working-directory: dist/
- name: Upload Release Asset
uses: svenstaro/upload-release-action@v2
with:
repo_token: ${{ secrets.GITHUB_TOKEN }}
file: superscript
asset_name: superscript
tag: ${{ github.ref }}

1
.gitignore vendored
View File

@ -13,6 +13,7 @@
**/profile.* **/profile.*
**/*.log
**/errorlog.txt **/errorlog.txt
/dist/superscript.* /dist/superscript.*
/dist/superscript /dist/superscript

View File

@ -114,13 +114,13 @@ def unkeyify_2l(layered_dict):
out[i] = list(map(lambda x: x[1], add)) out[i] = list(map(lambda x: x[1], add))
return out return out
def get_previous_time(apikey): def get_previous_time(client):
previous_time = get_analysis_flags(apikey, "latest_update") previous_time = get_analysis_flags(client, "latest_update")
if previous_time == None: if previous_time == None:
set_analysis_flags(apikey, "latest_update", 0) set_analysis_flags(client, "latest_update", 0)
previous_time = 0 previous_time = 0
else: else:
@ -129,21 +129,30 @@ def get_previous_time(apikey):
return previous_time return previous_time
def set_current_time(apikey, current_time): def set_current_time(client, current_time):
set_analysis_flags(apikey, "latest_update", {"latest_update":current_time}) set_analysis_flags(client, "latest_update", {"latest_update":current_time})
def load_match(apikey, competition): def get_database_config(client):
return get_match_data_formatted(apikey, competition) remote_config = get_analysis_flags(client, "config")
return remote_config["config"] if remote_config != None else None
def load_metric(apikey, competition, match, group_name, metrics): def set_database_config(client, config):
set_analysis_flags(client, "config", {"config": config})
def load_match(client, competition):
return get_match_data_formatted(client, competition)
def load_metric(client, competition, match, group_name, metrics):
group = {} group = {}
for team in match[group_name]: for team in match[group_name]:
db_data = get_team_metrics_data(apikey, competition, team) db_data = get_team_metrics_data(client, competition, team)
if db_data == None: if db_data == None:
@ -165,23 +174,23 @@ def load_metric(apikey, competition, match, group_name, metrics):
return group return group
def load_pit(apikey, competition): def load_pit(client, competition):
return get_pit_data_formatted(apikey, competition) return get_pit_data_formatted(client, competition)
def push_match(apikey, competition, results): def push_match(client, competition, results):
for team in results: for team in results:
push_team_tests_data(apikey, competition, team, results[team]) push_team_tests_data(client, competition, team, results[team])
def push_metric(apikey, competition, metric): def push_metric(client, competition, metric):
for team in metric: for team in metric:
push_team_metrics_data(apikey, competition, team, metric[team]) push_team_metrics_data(client, competition, team, metric[team])
def push_pit(apikey, competition, pit): def push_pit(client, competition, pit):
for variable in pit: for variable in pit:

View File

@ -9,11 +9,11 @@ def simplestats(data_test):
signal.signal(signal.SIGINT, signal.SIG_IGN) signal.signal(signal.SIGINT, signal.SIG_IGN)
data = np.array(data_test[0]) data = np.array(data_test[3])
data = data[np.isfinite(data)] data = data[np.isfinite(data)]
ranges = list(range(len(data))) ranges = list(range(len(data)))
test = data_test[1] test = data_test[2]
if test == "basic_stats": if test == "basic_stats":
return an.basic_stats(data) return an.basic_stats(data)
@ -48,13 +48,7 @@ def matchloop(client, competition, data, tests, exec_threads):
value = self[item] = type(self)() value = self[item] = type(self)()
return value return value
return_vector = {} input_vector = []
team_filtered = []
variable_filtered = []
variable_data = []
test_filtered = []
result_filtered = []
return_vector = AutoVivification() return_vector = AutoVivification()
for team in data: for team in data:
@ -65,35 +59,35 @@ def matchloop(client, competition, data, tests, exec_threads):
for test in tests[variable]: for test in tests[variable]:
team_filtered.append(team) input_vector.append((team, variable, test, data[team][variable]))
variable_filtered.append(variable)
variable_data.append((data[team][variable], test)) result_filtered = exec_threads.map(simplestats, input_vector)
test_filtered.append(test)
result_filtered = exec_threads.map(simplestats, variable_data)
i = 0 i = 0
result_filtered = list(result_filtered) result_filtered = list(result_filtered)
for result in result_filtered: for result in result_filtered:
filtered = test_filtered[i] filtered = input_vector[i][2]
try: try:
short = short_mapping[filtered] short = short_mapping[filtered]
return_vector[team_filtered[i]][variable_filtered[i]][test_filtered[i]] = result[short] return_vector[input_vector[i][0]][input_vector[i][1]][input_vector[i][2]] = result[short]
except KeyError: # not in mapping except KeyError: # not in mapping
return_vector[team_filtered[i]][variable_filtered[i]][test_filtered[i]] = result return_vector[input_vector[i][0]][input_vector[i][1]][input_vector[i][2]] = result
i += 1 i += 1
return return_vector return return_vector
def metricloop(tbakey, client, competition, timestamp, metrics): # listener based metrics update def metricloop(client, competition, data, metrics): # listener based metrics update
elo_N = metrics["elo"]["N"] elo_N = metrics["elo"]["N"]
elo_K = metrics["elo"]["K"] elo_K = metrics["elo"]["K"]
matches = pull_new_tba_matches(tbakey, competition, timestamp) matches = data
#matches = pull_new_tba_matches(tbakey, competition, timestamp)
red = {} red = {}
blu = {} blu = {}

View File

@ -10,13 +10,19 @@ __changelog__ = """changelog:
1.0.0: 1.0.0:
- superscript now runs in PEP 3143 compliant well behaved daemon on Linux systems - superscript now runs in PEP 3143 compliant well behaved daemon on Linux systems
- linux superscript daemon has integrated websocket output to monitor progress/status remotely - linux superscript daemon has integrated websocket output to monitor progress/status remotely
- linux daemon now sends stderr to errorlog.txt - linux daemon now sends stderr to errorlog.log
- added verbose option to linux superscript to allow for interactive output - added verbose option to linux superscript to allow for interactive output
- moved pymongo import to superscript.py - moved pymongo import to superscript.py
- added profile option to linux superscript to profile runtime of script - added profile option to linux superscript to profile runtime of script
- reduced memory usage slightly by consolidating the unwrapped input data
- added debug option, which performs one loop of analysis and dumps results to local files
- added event and time delay options to config - added event and time delay options to config
- event delay pauses loop until even listener recieves an update - event delay pauses loop until even listener recieves an update
- time delay pauses loop until the time specified has elapsed since the BEGINNING of previous loop - time delay pauses loop until the time specified has elapsed since the BEGINNING of previous loop
- added options to pull config information from database (reatins option to use local config file)
- config-preference option selects between prioritizing local config and prioritizing database config
- synchronize-config option selects whether to update the non prioritized config with the prioritized one
- divided config options between persistent ones (keys), and variable ones (everything else)
0.9.3: 0.9.3:
- improved data loading performance by removing redundant PyMongo client creation (120s to 14s) - improved data loading performance by removing redundant PyMongo client creation (120s to 14s)
- passed singular instance of PyMongo client as standin for apikey parameter in all data.py functions - passed singular instance of PyMongo client as standin for apikey parameter in all data.py functions
@ -158,65 +164,77 @@ import warnings
import websockets import websockets
from interface import splash, log, ERR, INF, stdout, stderr from interface import splash, log, ERR, INF, stdout, stderr
from data import get_previous_time, set_current_time, load_match, push_match, load_pit, push_pit, check_new_database_matches from data import get_previous_time, pull_new_tba_matches, set_current_time, load_match, push_match, load_pit, push_pit, get_database_config, set_database_config, check_new_database_matches
from processing import matchloop, metricloop, pitloop from processing import matchloop, metricloop, pitloop
config_path = "config.json" config_path = "config.json"
sample_json = """{ sample_json = """{
"max-threads": 0.5, "persistent":{
"team": "", "key":{
"competition": "2020ilch", "database":"",
"key":{ "tba":""
"database":"", },
"tba":"" "config-preference":"local",
"synchronize-config":false
}, },
"statistics":{ "variable":{
"match":{ "max-threads":0.5,
"balls-blocked":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "team":"",
"balls-collected":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "competition": "2020ilch",
"balls-lower-teleop":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "statistics":{
"balls-lower-auto":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "match":{
"balls-started":["basic_stats","historical_analyss","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "balls-blocked":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-upper-teleop":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"], "balls-collected":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-upper-auto":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"] "balls-lower-teleop":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-lower-auto":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-started":["basic_stats","historical_analyss","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-upper-teleop":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"],
"balls-upper-auto":["basic_stats","historical_analysis","regression_linear","regression_logarithmic","regression_exponential","regression_polynomial","regression_sigmoidal"]
},
"metric":{
"elo":{
"score":1500,
"N":400,
"K":24
}, },
"gl2":{ "metric":{
"score":1500, "elo":{
"rd":250, "score":1500,
"vol":0.06 "N":400,
"K":24
},
"gl2":{
"score":1500,
"rd":250,
"vol":0.06
},
"ts":{
"mu":25,
"sigma":8.33
}
}, },
"ts":{ "pit":{
"mu":25, "wheel-mechanism":true,
"sigma":8.33 "low-balls":true,
} "high-balls":true,
}, "wheel-success":true,
"pit":{ "strategic-focus":true,
"wheel-mechanism":true, "climb-mechanism":true,
"low-balls":true, "attitude":true
"high-balls":true, },
"wheel-success":true, "event-delay":false,
"strategic-focus":true, "loop-delay":60
"climb-mechanism":true,
"attitude":true
} }
},
"event-delay":false,
"loop-delay":60
}""" }"""
def main(send, verbose = False, profile = False): def main(send, verbose = False, profile = False, debug = False):
def close_all():
if "exec_threads" in locals():
exec_threads.terminate()
exec_threads.join()
exec_threads.close()
if "client" in locals():
client.close()
warnings.filterwarnings("ignore") warnings.filterwarnings("ignore")
sys.stderr = open("errorlog.txt", "w") sys.stderr = open("errorlog.log", "w")
loop_exit_code = 0 exit_code = 0
loop_stored_exception = None
if verbose: if verbose:
splash(__version__) splash(__version__)
@ -227,161 +245,74 @@ def main(send, verbose = False, profile = False):
loop_start = time.time() loop_start = time.time()
current_time = time.time() send(stdout, INF, "current time: " + str(loop_start))
send(stdout, INF, "current time: " + str(current_time))
send(stdout, INF, "loading config at <" + config_path + ">", code = 0)
config = {} config = {}
if load_config(config_path, config) == 1:
if load_config(config_path, config):
send(stderr, ERR, "could not find config at <" + config_path + ">, generating blank config and exiting", code = 100) send(stderr, ERR, "could not find config at <" + config_path + ">, generating blank config and exiting", code = 100)
sys.exit(1) exit_code = 1
break
send(stdout, INF, "found and opened config at <" + config_path + ">", code = 0) send(stdout, INF, "found and loaded config at <" + config_path + ">")
error_flag = False
try:
competition = config["competition"]
except:
send(stderr, ERR, "could not find competition field in config", code = 101)
error_flag = True
try:
match_tests = config["statistics"]["match"]
except:
send(stderr, ERR, "could not find match_tests field in config", code = 102)
error_flag = True
try:
metrics_tests = config["statistics"]["metric"]
except:
send(stderr, ERR, "could not find metrics_tests field in config", code = 103)
error_flag = True
try:
pit_tests = config["statistics"]["pit"]
except:
send(stderr, ERR, "could not find pit_tests field in config", code = 104)
error_flag = True
if error_flag:
sys.exit(1)
error_flag = False
if competition == None or competition == "":
send(stderr, ERR, "competition field in config must not be empty", code = 105)
error_flag = True
if match_tests == None:
send(stderr, ERR, "match_tests field in config must not be empty", code = 106)
error_flag = True
if metrics_tests == None:
send(stderr, ERR, "metrics_tests field in config must not be empty", code = 107)
error_flag = True
if pit_tests == None:
send(stderr, ERR, "pit_tests field in config must not be empty", code = 108)
error_flag = True
if error_flag:
sys.exit(1)
send(stdout, INF, "found and loaded competition, match_tests, metrics_tests, pit_tests from config")
sys_max_threads = os.cpu_count()
try:
cfg_max_threads = config["max-threads"]
except:
send(stderr, ERR, "max-threads field in config must not be empty, refer to documentation for configuration options", code = 109)
sys.exit(1)
if cfg_max_threads > -sys_max_threads and cfg_max_threads < 0 :
alloc_processes = sys_max_threads + cfg_max_threads
elif cfg_max_threads > 0 and cfg_max_threads < 1:
alloc_processes = math.floor(cfg_max_threads * sys_max_threads)
elif cfg_max_threads > 1 and cfg_max_threads <= sys_max_threads:
alloc_processes = cfg_max_threads
elif cfg_max_threads == 0:
alloc_processes = sys_max_threads
else:
send(stderr, ERR, "max-threads must be between -" + str(sys_max_threads) + " and " + str(sys_max_threads) + ", but got " + cfg_max_threads, code = 110)
sys.exit(1)
send(stdout, INF, "found and loaded max-threads from config")
send(stdout, INF, "attempting to start " + str(alloc_processes) + " threads")
try:
exec_threads = Pool(processes = alloc_processes)
except Exception as e:
send(stderr, ERR, "unable to start threads", code = 200)
send(stderr, INF, e)
sys.exit(1)
send(stdout, INF, "successfully initialized " + str(alloc_processes) + " threads")
exit_flag = False
try:
apikey = config["key"]["database"]
except:
send(stderr, ERR, "database key field in config must be present", code = 111)
exit_flag = True
try:
tbakey = config["key"]["tba"]
except:
send(stderr, ERR, "tba key field in config must be present", code = 112)
exit_flag = True
if apikey == None or apikey == "":
send(stderr, ERR, "database key field in config must not be empty, please populate the database key")
exit_flag = True
if tbakey == None or tbakey == "":
send(stderr, ERR, "tba key field in config must not be empty, please populate the tba key")
exit_flag = True
if exit_flag:
sys.exit(1)
flag, apikey, tbakey, preference, sync = parse_config_persistent(send, config)
if flag:
exit_code = 1
break
send(stdout, INF, "found and loaded database and tba keys") send(stdout, INF, "found and loaded database and tba keys")
client = pymongo.MongoClient(apikey) client = pymongo.MongoClient(apikey)
previous_time = get_previous_time(client) send(stdout, INF, "established connection to database")
send(stdout, INF, "analysis backtimed to: " + str(previous_time)) send(stdout, INF, "analysis backtimed to: " + str(get_previous_time(client)))
resolve_config_conflicts(send, client, config, preference, sync)
if config == 1:
exit_code = 1
break
flag, exec_threads, competition, match_tests, metrics_tests, pit_tests = parse_config_variable(send, config)
if flag:
exit_code = 1
break
start = time.time() start = time.time()
send(stdout, INF, "loading match data") send(stdout, INF, "loading match, metric, pit data (this may take a few seconds)")
match_data = load_match(client, competition) match_data = load_match(client, competition)
send(stdout, INF, "finished loading match data in " + str(time.time() - start) + " seconds") metrics_data = pull_new_tba_matches(tbakey, competition, loop_start)
start = time.time()
send(stdout, INF, "performing analysis on match data")
results = matchloop(client, competition, match_data, match_tests, exec_threads)
send(stdout, INF, "finished match analysis in " + str(time.time() - start) + " seconds")
start = time.time()
send(stdout, INF, "uploading match results to database")
push_match(client, competition, results)
send(stdout, INF, "finished uploading match results in " + str(time.time() - start) + " seconds")
start = time.time()
send(stdout, INF, "performing analysis on team metrics")
results = metricloop(tbakey, client, competition, current_time, metrics_tests)
send(stdout, INF, "finished metric analysis and pushed to database in " + str(time.time() - start) + " seconds")
start = time.time()
send(stdout, INF, "loading pit data")
pit_data = load_pit(client, competition) pit_data = load_pit(client, competition)
send(stdout, INF, "finished loading pit data in " + str(time.time() - start) + " seconds") send(stdout, INF, "finished loading match, metric, pit data in "+ str(time.time() - start) + " seconds")
start = time.time() start = time.time()
send(stdout, INF, "performing analysis on pit data") send(stdout, INF, "performing analysis on match, metrics, pit data")
results = pitloop(client, competition, pit_data, pit_tests) match_results = matchloop(client, competition, match_data, match_tests, exec_threads)
send(stdout, INF, "finished pit analysis in " + str(time.time() - start) + " seconds") metrics_results = metricloop(client, competition, metrics_data, metrics_tests)
pit_results = pitloop(client, competition, pit_data, pit_tests)
send(stdout, INF, "finished analysis in " + str(time.time() - start) + " seconds")
start = time.time() start = time.time()
send(stdout, INF, "uploading pit results to database") send(stdout, INF, "uploading match, metrics, pit results to database")
push_pit(client, competition, results) push_match(client, competition, match_results)
send(stdout, INF, "finished uploading pit results in " + str(time.time() - start) + " seconds") push_pit(client, competition, pit_results)
send(stdout, INF, "finished uploading results in " + str(time.time() - start) + " seconds")
client.close() if debug:
f = open("matchloop.log", "w+")
json.dump(match_results, f, ensure_ascii=False, indent=4)
f.close()
set_current_time(client, current_time) f = open("pitloop.log", "w+")
send(stdout, INF, "finished all tests in " + str(time.time() - loop_start) + " seconds, looping") json.dump(pit_results, f, ensure_ascii=False, indent=4)
f.close()
set_current_time(client, loop_start)
close_all()
send(stdout, INF, "closed threads and database client")
send(stdout, INF, "finished all tasks in " + str(time.time() - loop_start) + " seconds, looping")
if profile:
return # return instead of break to avoid sys.exit
event_delay = config["event-delay"] event_delay = config["event-delay"]
if event_delay: if event_delay:
@ -400,24 +331,165 @@ def main(send, verbose = False, profile = False):
except KeyboardInterrupt: except KeyboardInterrupt:
send(stdout, INF, "detected KeyboardInterrupt, killing threads") send(stdout, INF, "detected KeyboardInterrupt, killing threads")
if "exec_threads" in locals(): close_all()
exec_threads.terminate()
exec_threads.join()
exec_threads.close()
send(stdout, INF, "terminated threads, exiting") send(stdout, INF, "terminated threads, exiting")
loop_stored_exception = sys.exc_info()
loop_exit_code = 0 loop_exit_code = 0
break break
except Exception as e: except Exception as e:
send(stderr, ERR, "encountered an exception while running") send(stderr, ERR, "encountered an exception while running", code = 1)
print(e, file = stderr) print(e, file = stderr)
loop_exit_code = 1 exit_code = 1
close_all()
break break
if profile: sys.exit(exit_code)
return
sys.exit(loop_exit_code) def parse_config_persistent(send, config):
exit_flag = False
try:
apikey = config["persistent"]["key"]["database"]
except:
send(stderr, ERR, "database key field in config must be present", code = 111)
exit_flag = True
try:
tbakey = config["persistent"]["key"]["tba"]
except:
send(stderr, ERR, "tba key field in config must be present", code = 112)
exit_flag = True
try:
preference = config["persistent"]["config-preference"]
except:
send(stderr, ERR, "config-preference field in config must be present", code = 113)
exit_flag = True
try:
sync = config["persistent"]["synchronize-config"]
except:
send(stderr, ERR, "synchronize-config field in config must be present", code = 114)
exit_flag = True
if apikey == None or apikey == "":
send(stderr, ERR, "database key field in config must not be empty, please populate the database key", code = 115)
exit_flag = True
if tbakey == None or tbakey == "":
send(stderr, ERR, "tba key field in config must not be empty, please populate the tba key", code = 116)
exit_flag = True
if preference == None or preference == "":
send(stderr, ERR, "config-preference field in config must not be empty, please populate config-preference", code = 117)
exit_flag = True
if sync != True and sync != False:
send(stderr, ERR, "synchronize-config field in config must be a boolean, please populate synchronize-config", code = 118)
exit_flag = True
return exit_flag, apikey, tbakey, preference, sync
def parse_config_variable(send, config):
exit_flag = False
sys_max_threads = os.cpu_count()
try:
cfg_max_threads = config["variable"]["max-threads"]
except:
send(stderr, ERR, "max-threads field in config must not be empty, refer to documentation for configuration options", code = 109)
exit_flag = True
if cfg_max_threads > -sys_max_threads and cfg_max_threads < 0 :
alloc_processes = sys_max_threads + cfg_max_threads
elif cfg_max_threads > 0 and cfg_max_threads < 1:
alloc_processes = math.floor(cfg_max_threads * sys_max_threads)
elif cfg_max_threads > 1 and cfg_max_threads <= sys_max_threads:
alloc_processes = cfg_max_threads
elif cfg_max_threads == 0:
alloc_processes = sys_max_threads
else:
send(stderr, ERR, "max-threads must be between -" + str(sys_max_threads) + " and " + str(sys_max_threads) + ", but got " + cfg_max_threads, code = 110)
exit_flag = True
try:
exec_threads = Pool(processes = alloc_processes)
except Exception as e:
send(stderr, ERR, "unable to start threads", code = 200)
send(stderr, INF, e)
exit_flag = True
send(stdout, INF, "successfully initialized " + str(alloc_processes) + " threads")
try:
competition = config["variable"]["competition"]
except:
send(stderr, ERR, "could not find competition field in config", code = 101)
exit_flag = True
try:
match_tests = config["variable"]["statistics"]["match"]
except:
send(stderr, ERR, "could not find match field in config", code = 102)
exit_flag = True
try:
metrics_tests = config["variable"]["statistics"]["metric"]
except:
send(stderr, ERR, "could not find metrics field in config", code = 103)
exit_flag = True
try:
pit_tests = config["variable"]["statistics"]["pit"]
except:
send(stderr, ERR, "could not find pit field in config", code = 104)
exit_flag = True
if competition == None or competition == "":
send(stderr, ERR, "competition field in config must not be empty", code = 105)
exit_flag = True
if match_tests == None:
send(stderr, ERR, "matchfield in config must not be empty", code = 106)
exit_flag = True
if metrics_tests == None:
send(stderr, ERR, "metrics field in config must not be empty", code = 107)
exit_flag = True
if pit_tests == None:
send(stderr, ERR, "pit field in config must not be empty", code = 108)
exit_flag = True
send(stdout, INF, "found and loaded competition, match, metrics, pit from config")
return exit_flag, exec_threads, competition, match_tests, metrics_tests, pit_tests
def resolve_config_conflicts(send, client, config, preference, sync):
if sync:
if preference == "local" or preference == "client":
send(stdout, INF, "config-preference set to local/client, loading local config information")
remote_config = get_database_config(client)
if remote_config != config["variable"]:
set_database_config(client, config["variable"])
send(stdout, INF, "database config was different and was updated")
return
elif preference == "remote" or preference == "database":
send(stdout, INF, "config-preference set to remote/database, loading remote config information")
remote_config= get_database_config(client)
if remote_config != config["variable"]:
config["variable"] = remote_config
if save_config(config_path, config):
send(stderr, ERR, "local config was different but could not be updated")
config = 1
return
send(stdout, INF, "local config was different and was updated")
return
else:
send(stderr, ERR, "config-preference field in config must be \"local\"/\"client\" or \"remote\"/\"database\"")
config = 1
return
else:
if preference == "local" or preference == "client":
send(stdout, INF, "config-preference set to local/client, loading local config information")
return
elif preference == "remote" or preference == "database":
send(stdout, INF, "config-preference set to remote/database, loading database config information")
config["variable"] = get_database_config(client)
return
else:
send(stderr, ERR, "config-preference field in config must be \"local\"/\"client\" or \"remote\"/\"database\"")
config = 1
return
def load_config(path, config_vector): def load_config(path, config_vector):
try: try:
@ -432,15 +504,12 @@ def load_config(path, config_vector):
return 1 return 1
def save_config(path, config_vector): def save_config(path, config_vector):
try: f = open(path, "w+")
f = open(path) json.dump(config_vector, f, ensure_ascii=False, indent=4)
json.dump(config_vector) f.close()
f.close() return 0
return 0
except:
return 1
def start(pid_path, verbose = False, profile = False): def start(pid_path, verbose = False, profile = False, debug = False):
if profile: if profile:
@ -460,6 +529,10 @@ def start(pid_path, verbose = False, profile = False):
main(log, verbose = verbose) main(log, verbose = verbose)
elif debug:
main(log, verbose = True, profile = True, debug = debug)
else: else:
f = open('errorlog.txt', 'w+') f = open('errorlog.txt', 'w+')
@ -547,10 +620,12 @@ if __name__ == "__main__":
start(None, verbose = True) start(None, verbose = True)
elif 'profile' == sys.argv[1]: elif 'profile' == sys.argv[1]:
start(None, profile=True) start(None, profile=True)
elif 'debug' == sys.argv[1]:
start(None, debug = True)
else: else:
print("usage: %s start|stop|restart|verbose|profile" % sys.argv[0]) print("usage: %s start|stop|restart|verbose|profile|debug" % sys.argv[0])
sys.exit(2) sys.exit(2)
sys.exit(0) sys.exit(0)
else: else:
print("usage: %s start|stop|restart|verbose|profile" % sys.argv[0]) print("usage: %s start|stop|restart|verbose|profile|debug" % sys.argv[0])
sys.exit(2) sys.exit(2)