mirror of
https://github.com/titanscouting/tra-analysis.git
synced 2024-11-10 06:54:44 +00:00
superscript.py v 0.0.1.003
This commit is contained in:
parent
c3746539b3
commit
70b2ff1151
Binary file not shown.
BIN
data analysis/__pycache__/superscript.cpython-37.pyc
Normal file
BIN
data analysis/__pycache__/superscript.cpython-37.pyc
Normal file
Binary file not shown.
@ -61,4 +61,4 @@ def push_team_metrics_data(apikey, competition, team_num, data, dbname = "data_p
|
||||
client = pymongo.MongoClient(apikey)
|
||||
db = client[dbname]
|
||||
mdata = db[colname]
|
||||
mdata.replace_one({"competition" : competition, "team": team_num}, {"_id": competition+str(team_num)+"am", "competition" : competition, "team" : team_num, "data" : data}, True)
|
||||
mdata.replace_one({"competition" : competition, "team": team_num}, {"_id": competition+str(team_num)+"am", "competition" : competition, "team" : team_num, "metrics" : data}, True)
|
@ -3,10 +3,12 @@
|
||||
# Notes:
|
||||
# setup:
|
||||
|
||||
__version__ = "0.0.1.002"
|
||||
__version__ = "0.0.1.003"
|
||||
|
||||
# changelog should be viewed using print(analysis.__changelog__)
|
||||
__changelog__ = """changelog:
|
||||
0.0.1.003:
|
||||
- working
|
||||
0.0.1.002:
|
||||
- started implement of metrics
|
||||
0.0.1.001:
|
||||
@ -61,12 +63,19 @@ from analysis import analysis as an
|
||||
import data as d
|
||||
import time
|
||||
|
||||
main()
|
||||
def testing():
|
||||
|
||||
competition, config = load_config("config.csv")
|
||||
|
||||
apikey = an.load_csv("keys.txt")[0][0]
|
||||
tbakey = an.load_csv("keys.txt")[1][0]
|
||||
|
||||
metricsloop(tbakey, apikey, "2020mokc", 1583084980)
|
||||
|
||||
def main():
|
||||
while(True):
|
||||
current_time = time.time()
|
||||
print("time: " + time)
|
||||
print("time: " + str(current_time))
|
||||
|
||||
print("loading config")
|
||||
competition, config = load_config("config.csv")
|
||||
@ -147,8 +156,56 @@ def push_to_database(apikey, competition, results, metrics):
|
||||
|
||||
d.push_team_tests_data(apikey, competition, team, results[team])
|
||||
|
||||
def metricsloop(apikey, competition, timestamp): # listener based metrics update
|
||||
def metricsloop(tbakey, apikey, competition, timestamp): # listener based metrics update
|
||||
|
||||
matches = d.pull_new_tba_matches(apikey, competition, timestamp)
|
||||
matches = d.pull_new_tba_matches(tbakey, competition, timestamp)
|
||||
|
||||
return
|
||||
red = load_metrics(apikey, competition, matches, "red")
|
||||
blu = load_metrics(apikey, competition, matches, "blue")
|
||||
|
||||
return
|
||||
|
||||
def load_metrics(apikey, competition, matches, group_name):
|
||||
|
||||
for match in matches:
|
||||
|
||||
for team in match[group_name]:
|
||||
|
||||
group = {}
|
||||
|
||||
db_data = d.get_team_metrics_data(apikey, competition, team)
|
||||
|
||||
if d.get_team_metrics_data(apikey, competition, team) == None:
|
||||
|
||||
elo = {"score": 1500, "N": 1500, "K": 1500}
|
||||
gl2 = {"score": 1500, "rd": 250, "vol": 0.06}
|
||||
ts = {"mu": 25, "sigma": 25/3}
|
||||
|
||||
d.push_team_metrics_data(apikey, competition, team, {"elo":elo, "gliko2":gl2,"trueskill":ts})
|
||||
|
||||
group[team] = {"elo": elo, "gl2": gl2, "ts": ts}
|
||||
|
||||
else:
|
||||
|
||||
metrics = db_data["metrics"]
|
||||
elo = metrics["elo"]
|
||||
gl2 = metrics["gliko2"]
|
||||
ts = metrics["trueskill"]
|
||||
|
||||
group[team] = {"elo": elo, "gl2": gl2, "ts": ts}
|
||||
|
||||
return group
|
||||
|
||||
testing()
|
||||
|
||||
"""
|
||||
Metrics Defaults:
|
||||
|
||||
elo starting score = 1500
|
||||
elo N = 1500
|
||||
elo K = 32
|
||||
|
||||
gl2 starting score = 1500
|
||||
gl2 starting rd = 350
|
||||
gl2 starting vol = 0.06
|
||||
"""
|
Loading…
Reference in New Issue
Block a user