Skip to content

Commit 8583383

Browse files
fix: Added description and resolved bug in rangeslider (#1990)
* fix: Added description and resolved bug in rangeslider addresses #1987 fixes #1954 * format * remove certificate
1 parent 56f90a7 commit 8583383

File tree

3 files changed

+27
-17
lines changed

3 files changed

+27
-17
lines changed

.gitignore

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -149,3 +149,6 @@ uv.lock
149149
# model loading tests
150150
model_names.txt
151151
mteb/leaderboard/__cached_results.json
152+
153+
# gradio
154+
.gradio/

mteb/leaderboard/app.py

Lines changed: 21 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -156,7 +156,7 @@ def update_task_info(task_names: str) -> gr.DataFrame:
156156

157157

158158
# Model sizes in million parameters
159-
MIN_MODEL_SIZE, MAX_MODEL_SIZE = 0, 10_000
159+
MIN_MODEL_SIZE, MAX_MODEL_SIZE = 0, 100_000
160160

161161

162162
def filter_models(
@@ -399,6 +399,7 @@ def filter_models(
399399
download_summary.click(
400400
download_table, inputs=[summary_table], outputs=[download_summary]
401401
)
402+
402403
with gr.Accordion(
403404
"What do aggregate measures (Rank(Borda), Mean(Task), etc.) mean?",
404405
open=False,
@@ -425,6 +426,25 @@ def filter_models(
425426
Based on community feedback and research findings, This definition could change in the future.
426427
"""
427428
)
429+
with gr.Accordion(
430+
"Why is a model is missing or not showing up?",
431+
open=False,
432+
):
433+
gr.Markdown(
434+
"""
435+
There is multiple reasons why a model might not show up in the leaderboard. Here is a few common reasons:
436+
437+
- **Filter Setting**: It is being filtered out with your current filter. By default we do not show models that are not zero-shot on the benchmark.
438+
You can change this setting in the model selection panel.
439+
- **Removed Derivatives**: While the previous models showed many variant of a models (e.g. quantized, varying embedding sizes) as separate models, we now only show the original model.
440+
- **Missing Results**: The model might not have been run on the tasks in the benchmark. We only display models that have been run on at least one tasks
441+
in the benchmark. For visualizations which require the mean across all tasks, we only display models that have been run on all tasks in the benchmark.
442+
You can see existing results the [results repository](https://github.com/embeddings-benchmark/results). This is also where new results are added using a PR.
443+
- **Missing Metadata**: Currently the we only show models for which we have metadata on in [mteb](https://github.com/embeddings-benchmark/mteb).
444+
You can follow this guide on how to add a [model](https://github.com/embeddings-benchmark/mteb/blob/main/docs/adding_a_model.md) and
445+
see existing implementations [here](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/models).
446+
"""
447+
)
428448
with gr.Tab("Performance per task"):
429449
per_task_table.render()
430450
download_per_task = gr.DownloadButton("Download Table")

mteb/leaderboard/figures.py

Lines changed: 3 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ def parse_float(value) -> float:
6060

6161

6262
def add_size_guide(fig: go.Figure):
63-
xpos = [5 * 1e9] * 4
63+
xpos = [2 * 1e6] * 4
6464
ypos = [7.8, 8.5, 9, 10]
6565
sizes = [256, 1024, 2048, 4096]
6666
fig.add_trace(
@@ -78,25 +78,13 @@ def add_size_guide(fig: go.Figure):
7878
)
7979
)
8080
fig.add_annotation(
81-
text="<b>Embedding Size:</b>",
81+
text="<b>Embedding Size</b>",
8282
font=dict(size=16),
83-
x=np.log10(1.5e9),
83+
x=np.log10(10 * 1e6),
8484
y=10,
8585
showarrow=False,
8686
opacity=0.3,
8787
)
88-
for x, y, size in zip(xpos, np.linspace(7.5, 14, 4), sizes):
89-
fig.add_annotation(
90-
text=f"<b>{size}</b>",
91-
font=dict(size=12),
92-
x=np.log10(x),
93-
y=y,
94-
showarrow=True,
95-
ay=0,
96-
ax=50,
97-
opacity=0.3,
98-
arrowwidth=2,
99-
)
10088
return fig
10189

10290

@@ -127,7 +115,6 @@ def performance_size_plot(df: pd.DataFrame) -> go.Figure:
127115
size="sqrt(dim)",
128116
color="Log(Tokens)",
129117
range_color=[2, 5],
130-
range_x=[8 * 1e6, 11 * 1e9],
131118
range_y=[min(0, min_score * 1.25), max_score * 1.25],
132119
hover_data={
133120
"Max Tokens": True,

0 commit comments

Comments
 (0)