-
Notifications
You must be signed in to change notification settings - Fork 1
/
model.py
109 lines (94 loc) · 3.87 KB
/
model.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
# own your model or use this for improve your handle.
# see folder directory /models, copied and modify.
# getting point for another model
# own by me: 0xgan/grey
import pandas as pd
import numpy as np
import os
import requests
import joblib
from sklearn.svm import SVR
from sklearn.model_selection import train_test_split, GridSearchCV
from sklearn.metrics import mean_squared_error, mean_absolute_error, r2_score
from sklearn.preprocessing import StandardScaler
from sklearn.pipeline import Pipeline
from config import data_base_path, model_file_path, TOKEN
def get_data(url):
response = requests.get(url)
if response.status_code == 200:
data = response.json()['history']
df = pd.DataFrame(data)
df['t'] = pd.to_datetime(df['t'], unit='s')
df.columns = ['Timestamp', 'Predict']
df['Predict'] = df['Predict'].apply(lambda x: x * 100)
print(df.head())
else:
print(f"Failed to retrieve data. Status code: {response.status_code}")
return df
def download_data(token):
os.makedirs(data_base_path, exist_ok=True)
if token == 'R':
url = "https://clob.polymarket.com/prices-history?interval=all&market=21742633143463906290569050155826241533067272736897614950488156847949938836455&fidelity=1000"
data = get_data(url)
save_path = os.path.join(data_base_path, 'polymarket_R.csv')
data.to_csv(save_path)
elif token == 'D':
url = "https://clob.polymarket.com/prices-history?interval=all&market=69236923620077691027083946871148646972011131466059644796654161903044970987404&fidelity=1000"
data = get_data(url)
save_path = os.path.join(data_base_path, 'polymarket_D.csv')
data.to_csv(save_path)
def train_model(token):
training_price_data_path = os.path.join(data_base_path, f"polymarket_{token}.csv")
df = pd.read_csv(training_price_data_path)
df['Timestamp'] = pd.to_datetime(df['Timestamp'])
df['year'] = df['Timestamp'].dt.year
df['month'] = df['Timestamp'].dt.month
df['day'] = df['Timestamp'].dt.day
df['hour'] = df['Timestamp'].dt.hour
X = df[['year', 'month', 'day', 'hour']]
y = df['Predict']
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)
# Define pipeline with StandardScaler and SVR
pipeline = Pipeline([
('scaler', StandardScaler()),
('svr', SVR())
])
# SVR hyperparameter tuning
param_grid = {
'svr__C': [0.1, 1, 10, 34],
'svr__gamma': ['scale', 'auto'],
'svr__kernel': ['rbf', 'linear', 'poly', 'sigmoid']
}
grid_search = GridSearchCV(pipeline, param_grid, cv=5, scoring='neg_mean_squared_error', n_jobs=3)
grid_search.fit(X_train, y_train)
best_svr = grid_search.best_estimator_
print(f"Best parameters: {grid_search.best_params_}")
# Making predictions
y_pred = best_svr.predict(X_test)
# Evaluate the model
mae = mean_absolute_error(y_test, y_pred)
mse = mean_squared_error(y_test, y_pred)
rmse = mean_squared_error(y_test, y_pred, squared=False)
r2 = r2_score(y_test, y_pred)
print(f"Mean Absolute Error: {mae}")
print(f"Mean Squared Error: {mse}")
print(f"Root Mean Squared Error: {rmse}")
print(f"R^2 Score: {r2}")
# Save the model
os.makedirs(model_file_path, exist_ok=True)
save_path_model = os.path.join(model_file_path, f'svm_model_{token}.pkl')
joblib.dump(best_svr, save_path_model)
print(f"Trained model saved to {save_path_model}")
def get_inference(token):
save_path_model = os.path.join(model_file_path, f'svm_model_{token}.pkl')
loaded_model = joblib.load(save_path_model)
print("Loaded model successfully")
single_input = pd.DataFrame({
'year': [2024],
'month': [10],
'day': [10],
'hour': [12]
})
# Prediction price
predicted_price = loaded_model.predict(single_input)
return predicted_price[0]