import gradio as gr
import pandas as pd
# Import our UI factories and the data loader
from ui_components import create_leaderboard_display, create_benchmark_details_display, get_full_leaderboard_data, create_sub_navigation_bar
def build_category_page(CATEGORY_NAME, PAGE_DESCRIPTION):
with gr.Column(elem_id="page-content-wrapper"):
gr.HTML(f'
AstaBench {CATEGORY_NAME} Leaderboard (Aggregate)
', elem_id="main-header")
validation_df, validation_tag_map = get_full_leaderboard_data("validation")
test_df, test_tag_map = get_full_leaderboard_data("test")
with gr.Column(elem_id="validation_nav_container", visible=False) as validation_nav_container:
create_sub_navigation_bar(validation_tag_map, CATEGORY_NAME, validation=True)
with gr.Column(elem_id="test_nav_container", visible=True) as test_nav_container:
create_sub_navigation_bar(test_tag_map, CATEGORY_NAME)
gr.Markdown(PAGE_DESCRIPTION, elem_id="category-intro")
# --- This page now has two main sections: Validation and Test ---
with gr.Tabs():
with gr.Tab("Results: Test Set") as test_tab:
# Repeat the process for the "test" split
if not test_df.empty:
gr.Markdown("**Test Set** results are reserved for final assessment. This helps ensure that the agent generalizes well to unseen problems.")
create_leaderboard_display(
full_df=test_df,
tag_map=test_tag_map,
category_name=CATEGORY_NAME,
split_name="test"
)
create_benchmark_details_display(
full_df=test_df,
tag_map=test_tag_map,
category_name=CATEGORY_NAME,
validation=False,
)
else:
gr.Markdown("No data available for test split.")
with gr.Tab("Results: Validation Set") as validation_tab:
# 1. Load all necessary data for the "validation" split ONCE.
if not validation_df.empty:
gr.Markdown("**Validation Set** results are used during development to tune and compare agents before final testing.")
# 2. Render the main category display using the loaded data.
create_leaderboard_display(
full_df=validation_df,
tag_map=validation_tag_map,
category_name=CATEGORY_NAME,
split_name="validation"
)
# 3. Render the detailed breakdown for each benchmark in the category.
create_benchmark_details_display(
full_df=validation_df,
tag_map=validation_tag_map,
category_name=CATEGORY_NAME,
validation=True,
)
else:
gr.Markdown("No data available for validation split.")
show_validation_js = """
() => {
document.getElementById('validation_nav_container').style.display = 'block';
document.getElementById('test_nav_container').style.display = 'none';
setTimeout(() => { window.dispatchEvent(new Event('resize')) }, 0);
}
"""
# JavaScript to show the TEST nav, hide the VALIDATION nav, AND fix the plots.
show_test_js = """
() => {
document.getElementById('validation_nav_container').style.display = 'none';
document.getElementById('test_nav_container').style.display = 'block';
}
"""
# Assign the pure JS functions to the select events. No Python `fn` is needed.
validation_tab.select(fn=None, inputs=None, outputs=None, js=show_validation_js)
test_tab.select(fn=None, inputs=None, outputs=None, js=show_test_js)
return validation_nav_container, test_nav_container