@@ -203,8 +203,11 @@ def render_mds(overwrite_previous, subfolder="docs"):
203
203
continue
204
204
for branch_path in glob .glob (os .path .join (org_path , "*.json" )):
205
205
cum_tests_passed = 0
206
+ lite_cum_tests_passed = 0
206
207
repos_resolved = 0
208
+ lite_repos_resolved = 0
207
209
total_duration = 0.0
210
+ lite_total_duration = 0.0
208
211
branch_metrics = json .load (open (branch_path ))
209
212
submission_info = branch_metrics ["submission_info" ]
210
213
split = submission_info ["split" ]
@@ -298,6 +301,10 @@ def render_mds(overwrite_previous, subfolder="docs"):
298
301
cum_tests_passed += pytest_info ["summary" ]["passed" ]
299
302
total_duration += pytest_info ["duration" ]
300
303
repos_resolved += int (resolved )
304
+ if repo_name in SPLIT ["lite" ]:
305
+ lite_cum_tests_passed += pytest_info ["summary" ]["passed" ]
306
+ lite_total_duration += pytest_info ["duration" ]
307
+ lite_repos_resolved += int (resolved )
301
308
if write_submission :
302
309
pytest_details = f"{ pytest_info ['summary' ]['passed' ]} / { pytest_info ['summary' ]['total' ]} "
303
310
duration = f"{ pytest_info ['duration' ]:.2f} "
@@ -331,6 +338,16 @@ def render_mds(overwrite_previous, subfolder="docs"):
331
338
f"{ analysis_link } |"
332
339
f"{ github_link } |"
333
340
)
341
+ if split == "all" :
342
+ leaderboard ["lite" ] += (
343
+ f"\n |{ display_name } |"
344
+ f"{ lite_repos_resolved } |"
345
+ f"{ lite_cum_tests_passed } |"
346
+ f"{ lite_total_duration :.2f} |"
347
+ f"{ submission_date } |"
348
+ f"{ analysis_link } |"
349
+ f"{ github_link } |"
350
+ )
334
351
335
352
leaderboard_filepath = os .path .join (subfolder , "analysis.md" )
336
353
with open (leaderboard_filepath , "w" ) as wf :
0 commit comments