Skip to content

Commit da7d705

Browse files
committed
fix html error
1 parent 6d65246 commit da7d705

File tree

1 file changed

+72
-25
lines changed

1 file changed

+72
-25
lines changed

omnigirl_leaderboard.html

Lines changed: 72 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -28,20 +28,10 @@
2828
table{font-size:small}
2929
}
3030

31-
/* 仅新增的样式 ↓↓↓ */
32-
.icon-cell{
33-
text-align:center;
34-
}
35-
.icon-cell img{
36-
height:1.1em;
37-
}
38-
.icon-cell a{
39-
display:inline-block;
40-
font-size:1.1em;
41-
line-height:1;
42-
text-decoration:none;
43-
}
44-
/* ↑↑↑ */
31+
/* 新增:Org / Site 图标尺寸 */
32+
.icon-cell{text-align:center}
33+
.icon-cell img{height:1.1em}
34+
.icon-cell a{display:inline-block;font-size:1.1em;line-height:1;text-decoration:none}
4535
</style>
4636
</head>
4737

@@ -50,26 +40,24 @@
5040

5141
<h1 class="text-nowrap mt-5">🏆 OmniGIRL Leaderboard 🏆</h1>
5242
<h3 class="fw-light text-nowrap">
53-
<small id="warning">A Multilingual & Multimodal Benchmark for GitHub Issue Resolution<br></small>
43+
<small id="warning">A Multilingual &amp; Multimodal Benchmark for GitHub Issue Resolution<br></small>
5444
</h3>
5545

5646
<!-- 徽章:GitHub / Paper / HF -->
5747
<div class="d-flex flex-row justify-content-center gap-3">
5848
<a href="https://github.com/DeepSoftwareAnalytics/OmniGIRL" target="_blank">
59-
<img src="https://img.shields.io/badge/github-%23121011.svg?style=for-the-badge&logo=github&logoColor=white"
60-
alt="github" class="img-fluid">
49+
<img src="https://img.shields.io/badge/github-%23121011.svg?style=for-the-badge&logo=github&logoColor=white" alt="github" class="img-fluid">
6150
</a>
6251
<a href="https://openreview.net/forum?id=yourISSTA25paperID" target="_blank">
63-
<img src="https://img.shields.io/badge/Paper-ISSTA%2725-a55fed.svg?style=for-the-badge"
64-
alt="paper" class="img-fluid">
52+
<img src="https://img.shields.io/badge/Paper-ISSTA%2725-a55fed.svg?style=for-the-badge" alt="paper" class="img-fluid">
6553
</a>
54+
<!-- 只改了这一行的链接 -->
6655
<a href="https://huggingface.co/datasets/Deep-Software-Analytics/OmniGIRL" target="_blank">
67-
<img src="https://img.shields.io/badge/HuggingFace-%23f9ac00.svg?style=for-the-badge&logo=huggingface"
68-
alt="hf" class="img-fluid">
56+
<img src="https://img.shields.io/badge/HuggingFace-%23f9ac00.svg?style=for-the-badge&logo=huggingface" alt="hf" class="img-fluid">
6957
</a>
7058
</div>
7159

72-
<!-- 五语言按钮 -->
60+
<!-- 语言切换 -->
7361
<div class="btn-group mt-3" role="group" id="Language">
7462
<input type="radio" class="btn-check" name="langradio" id="Full" value="full" checked>
7563
<label class="btn btn-outline-primary" for="Full">Full</label>
@@ -83,7 +71,7 @@ <h3 class="fw-light text-nowrap">
8371
<label class="btn btn-outline-primary" for="TypeScript">TypeScript</label>
8472
</div>
8573

86-
<!-- 排名表 -->
74+
<!-- Leaderboard 表格 -->
8775
<table id="origin" class="table table-striped table-bordered border border-primary border-3 mt-4 w-100">
8876
<thead>
8977
<tr>
@@ -98,8 +86,67 @@ <h3 class="fw-light text-nowrap">
9886
<tbody id="leaderboard-body"></tbody>
9987
</table>
10088

101-
<!-- Notes(略,保持不动) -->
102-
<!-- ... 其余静态内容不变 ... -->
89+
<!-- 📝 Notes -->
90+
<div id="notes" class="w-100">
91+
<h3>📝 Notes</h3>
92+
<div class="inline-block mt-3">
93+
<ol>
94+
<li><strong>OmniGIRL</strong> is a multilingual &amp; multimodal GitHub-issue-resolution benchmark with <strong>959 tasks</strong> spanning four programming languages. Inputs may include text, screenshots, rendered web pages and other modalities.</li>
95+
<li>For realistic evaluation, <em>we recommend</em> that methods automatically examine each task’s raw input to detect available modalities (e.g., embedded webpages, images), retrieve the relevant content by themselves, and invoke the appropriate tools— instead of relying on manual hints. Doing so better assesses a solver’s <strong>general-purpose issue-resolution ability in real-world scenarios</strong>.</li>
96+
<li>Our baseline system is released <em>for research purposes only</em>; please cite OmniGIRL if you use it.</li>
97+
</ol>
98+
</div>
99+
</div>
100+
101+
<!-- 📨 How to Submit -->
102+
<div id="notes" class="w-100">
103+
<h3>📨 How&nbsp;to&nbsp;Submit</h3>
104+
<div class="inline-block mt-3">
105+
<ol>
106+
<li>Prepare a <code>.json</code> or <code>.jsonl</code> file. Each record must contain at least the keys <code>instance_id</code>, <code>model_name_or_path</code>, and <code>model_patch</code>.</li>
107+
<li>Email the file to <a href="mailto:guolh8@mail2.sysu.edu.cn?subject=OmniGIRL%20Submission">guolh8@mail2.sysu.edu.cn</a>.</li>
108+
<li>We will evaluate your submission locally and update the leaderboard once the results are verified.</li>
109+
</ol>
110+
</div>
111+
</div>
112+
113+
<!-- 🤗 More Leaderboards -->
114+
<div id="notes" class="w-100">
115+
<h3>🤗 More Leaderboards</h3>
116+
<div class="inline-block mt-3">
117+
<ol>
118+
<li><a href="https://bigcode-bench.github.io/">BigCodeBench</a></li>
119+
<li><a href="https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard">Big Code Models</a></li>
120+
<li><a href="https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard">Chatbot Arena</a></li>
121+
<li><a href="https://github.com/amazon-science/cceval">CrossCodeEval</a></li>
122+
<li><a href="https://fudanselab-classeval.github.io/">ClassEval</a></li>
123+
<li><a href="https://crux-eval.github.io/leaderboard.html">CRUXEval</a></li>
124+
<li><a href="https://codetlingua.github.io/leaderboard.html">Code Lingua</a></li>
125+
<li><a href="https://evo-eval.github.io/">Evo-Eval</a></li>
126+
<li><a href="https://huggingface.co/spaces/EffiBench/effibench-leaderboard">EffiBench</a></li>
127+
<li><a href="https://github.com/01-ai/HumanEval.jl">HumanEval.jl</a></li>
128+
<li><a href="https://livecodebench.github.io/leaderboard.html">LiveCodeBench</a></li>
129+
<li><a href="https://sparksofagi.github.io/MHPP/">MHPP</a></li>
130+
<li><a href="https://github.com/THUDM/NaturalCodeBench">NaturalCodeBench</a></li>
131+
<li><a href="https://github.com/Leolty/repobench">RepoBench</a></li>
132+
<li><a href="https://www.swebench.com/">SWE-bench</a></li>
133+
<li><a href="https://leaderboard.tabbyml.com/">TabbyML</a></li>
134+
<li><a href="https://llm4softwaretesting.github.io/">TestEval</a></li>
135+
</ol>
136+
</div>
137+
</div>
138+
139+
<!-- 🙏 Acknowledgements -->
140+
<div id="notes" class="w-100 mb-5">
141+
<h3>🙏 Acknowledgements</h3>
142+
<div class="inline-block mt-3">
143+
<ol>
144+
<li>We build on prior work — <strong><a href="https://arxiv.org/abs/2310.06770" target="_blank">SWE-bench</a></strong>, <strong><a href="https://arxiv.org/abs/2407.01489" target="_blank">Agentless</a></strong>, and <strong><a href="https://arxiv.org/abs/2404.05427" target="_blank">AutoCodeRover</a></strong> — which laid the groundwork for this study.</li>
145+
<li>We thank the <strong><a href="https://github.com/evalplus/evalplus" target="_blank">EvalPlus leaderboard</a></strong> team for releasing the elegant page template that inspired this site.</li>
146+
<li>Finally, we are grateful to the <strong>open-source developer community</strong> for their invaluable contributions.</li>
147+
</ol>
148+
</div>
149+
</div>
103150

104151
</div><!-- /#content -->
105152

0 commit comments

Comments
 (0)