Spaces:
Runtime error
Runtime error
cyberosa
commited on
Commit
Β·
2628969
1
Parent(s):
0b5bfca
trying to fix error with loading files
Browse files- app.py +13 -12
- requirements.txt +1 -0
- scripts/retention_metrics.py +1 -8
- scripts/utils.py +0 -1
app.py
CHANGED
|
@@ -65,21 +65,21 @@ def load_all_data():
|
|
| 65 |
|
| 66 |
# all trades profitability
|
| 67 |
# Download the compressed file
|
| 68 |
-
|
| 69 |
repo_id="valory/Olas-predict-dataset",
|
| 70 |
filename="all_trades_profitability.parquet.gz",
|
| 71 |
repo_type="dataset",
|
| 72 |
)
|
| 73 |
|
| 74 |
-
|
|
|
|
| 75 |
|
| 76 |
-
with gzip.open(
|
| 77 |
-
with open(
|
| 78 |
shutil.copyfileobj(f_in, f_out)
|
| 79 |
|
| 80 |
# Now read the decompressed parquet file
|
| 81 |
-
df1 = pd.read_parquet(
|
| 82 |
-
os.remove(parquet_file_path)
|
| 83 |
|
| 84 |
# closed_markets_div
|
| 85 |
closed_markets_df = hf_hub_download(
|
|
@@ -106,18 +106,18 @@ def load_all_data():
|
|
| 106 |
df4 = pd.read_parquet(unknown_df)
|
| 107 |
|
| 108 |
# retention activity
|
| 109 |
-
|
| 110 |
repo_id="valory/Olas-predict-dataset",
|
| 111 |
filename="retention_activity.parquet.gz",
|
| 112 |
repo_type="dataset",
|
| 113 |
)
|
| 114 |
-
|
| 115 |
|
| 116 |
-
with gzip.open(
|
| 117 |
-
with open(
|
| 118 |
shutil.copyfileobj(f_in, f_out)
|
| 119 |
-
df5 = pd.read_parquet(
|
| 120 |
-
os.remove(
|
| 121 |
|
| 122 |
# active_traders.parquet
|
| 123 |
active_traders_df = hf_hub_download(
|
|
@@ -208,6 +208,7 @@ def prepare_data():
|
|
| 208 |
active_traders,
|
| 209 |
all_mech_calls,
|
| 210 |
) = prepare_data()
|
|
|
|
| 211 |
retention_df = prepare_retention_dataset(
|
| 212 |
retention_df=raw_retention_df, unknown_df=unknown_traders
|
| 213 |
)
|
|
|
|
| 65 |
|
| 66 |
# all trades profitability
|
| 67 |
# Download the compressed file
|
| 68 |
+
gz_file_path_trades = hf_hub_download(
|
| 69 |
repo_id="valory/Olas-predict-dataset",
|
| 70 |
filename="all_trades_profitability.parquet.gz",
|
| 71 |
repo_type="dataset",
|
| 72 |
)
|
| 73 |
|
| 74 |
+
parquet_file_path_trades = gz_file_path_trades.replace(".gz", "")
|
| 75 |
+
parquet_file_path_trades = parquet_file_path_trades.replace("all", "")
|
| 76 |
|
| 77 |
+
with gzip.open(gz_file_path_trades, "rb") as f_in:
|
| 78 |
+
with open(parquet_file_path_trades, "wb") as f_out:
|
| 79 |
shutil.copyfileobj(f_in, f_out)
|
| 80 |
|
| 81 |
# Now read the decompressed parquet file
|
| 82 |
+
df1 = pd.read_parquet(parquet_file_path_trades)
|
|
|
|
| 83 |
|
| 84 |
# closed_markets_div
|
| 85 |
closed_markets_df = hf_hub_download(
|
|
|
|
| 106 |
df4 = pd.read_parquet(unknown_df)
|
| 107 |
|
| 108 |
# retention activity
|
| 109 |
+
gz_file_path_ret = hf_hub_download(
|
| 110 |
repo_id="valory/Olas-predict-dataset",
|
| 111 |
filename="retention_activity.parquet.gz",
|
| 112 |
repo_type="dataset",
|
| 113 |
)
|
| 114 |
+
parquet_file_path_ret = gz_file_path_ret.replace(".gz", "")
|
| 115 |
|
| 116 |
+
with gzip.open(gz_file_path_ret, "rb") as f_in:
|
| 117 |
+
with open(parquet_file_path_ret, "wb") as f_out:
|
| 118 |
shutil.copyfileobj(f_in, f_out)
|
| 119 |
+
df5 = pd.read_parquet(parquet_file_path_ret)
|
| 120 |
+
# os.remove(parquet_file_path_ret)
|
| 121 |
|
| 122 |
# active_traders.parquet
|
| 123 |
active_traders_df = hf_hub_download(
|
|
|
|
| 208 |
active_traders,
|
| 209 |
all_mech_calls,
|
| 210 |
) = prepare_data()
|
| 211 |
+
|
| 212 |
retention_df = prepare_retention_dataset(
|
| 213 |
retention_df=raw_retention_df, unknown_df=unknown_traders
|
| 214 |
)
|
requirements.txt
CHANGED
|
@@ -10,4 +10,5 @@ pydantic
|
|
| 10 |
pydantic_core
|
| 11 |
nbformat
|
| 12 |
pytz
|
|
|
|
| 13 |
ipfshttpclient
|
|
|
|
| 10 |
pydantic_core
|
| 11 |
nbformat
|
| 12 |
pytz
|
| 13 |
+
duckdb
|
| 14 |
ipfshttpclient
|
scripts/retention_metrics.py
CHANGED
|
@@ -1,6 +1,5 @@
|
|
| 1 |
import pandas as pd
|
| 2 |
from datetime import datetime, timedelta
|
| 3 |
-
from scripts.utils import DATA_DIR
|
| 4 |
|
| 5 |
|
| 6 |
# Basic Week over Week Retention
|
|
@@ -263,10 +262,4 @@ def prepare_retention_dataset(
|
|
| 263 |
|
| 264 |
|
| 265 |
if __name__ == "__main__":
|
| 266 |
-
|
| 267 |
-
traders_df = pd.read_parquet(DATA_DIR / "all_trades_profitability.parquet")
|
| 268 |
-
unknown_df = pd.read_parquet(DATA_DIR / "unknown_traders.parquet")
|
| 269 |
-
all_traders = prepare_retention_dataset(traders_df, unknown_df)
|
| 270 |
-
# Usage example:
|
| 271 |
-
wow_retention = calculate_wow_retention_by_type(all_traders)
|
| 272 |
-
cohort_retention = calculate_cohort_retention(all_traders)
|
|
|
|
| 1 |
import pandas as pd
|
| 2 |
from datetime import datetime, timedelta
|
|
|
|
| 3 |
|
| 4 |
|
| 5 |
# Basic Week over Week Retention
|
|
|
|
| 262 |
|
| 263 |
|
| 264 |
if __name__ == "__main__":
|
| 265 |
+
print("WIP")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scripts/utils.py
CHANGED
|
@@ -4,7 +4,6 @@ from datetime import datetime, timedelta
|
|
| 4 |
|
| 5 |
SCRIPTS_DIR = Path(__file__).parent
|
| 6 |
ROOT_DIR = SCRIPTS_DIR.parent
|
| 7 |
-
DATA_DIR = ROOT_DIR / "data"
|
| 8 |
TMP_DIR = ROOT_DIR / "tmp"
|
| 9 |
|
| 10 |
|
|
|
|
| 4 |
|
| 5 |
SCRIPTS_DIR = Path(__file__).parent
|
| 6 |
ROOT_DIR = SCRIPTS_DIR.parent
|
|
|
|
| 7 |
TMP_DIR = ROOT_DIR / "tmp"
|
| 8 |
|
| 9 |
|