Spaces:
Sleeping
Sleeping
Fix VES problems
Browse files
app.py
CHANGED
|
@@ -734,6 +734,8 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
| 734 |
metrics_conc = target_df
|
| 735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
|
|
|
|
|
|
| 737 |
eval_text = generate_eval_text("End evaluation")
|
| 738 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
| 739 |
|
|
@@ -872,12 +874,16 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
| 872 |
metrics_df_model['model'] = model
|
| 873 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
| 874 |
|
| 875 |
-
if 'valid_efficency_score' not in metrics_conc.columns
|
| 876 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 877 |
|
| 878 |
-
if 'VES' not in metrics_conc.columns
|
| 879 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 880 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 881 |
eval_text = generate_eval_text("End evaluation")
|
| 882 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model] for model in model_list]
|
| 883 |
|
|
|
|
| 734 |
metrics_conc = target_df
|
| 735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 737 |
+
if 'VES' not in metrics_conc.columns:
|
| 738 |
+
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 739 |
eval_text = generate_eval_text("End evaluation")
|
| 740 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
| 741 |
|
|
|
|
| 874 |
metrics_df_model['model'] = model
|
| 875 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
| 876 |
|
| 877 |
+
if 'valid_efficency_score' not in metrics_conc.columns:
|
| 878 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
| 879 |
|
| 880 |
+
if 'VES' not in metrics_conc.columns:
|
| 881 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
| 882 |
+
|
| 883 |
+
if 'VES' not in metrics_conc.columns and 'valid_efficency_score' not in metrics_conc.columns:
|
| 884 |
+
metrics_conc['VES'] = 0
|
| 885 |
+
metrics_conc['valid_efficency_score'] = 0
|
| 886 |
+
|
| 887 |
eval_text = generate_eval_text("End evaluation")
|
| 888 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model] for model in model_list]
|
| 889 |
|