Skip to content

Commit 002f086

Browse files
committed
submitting models
1 parent 874ec4f commit 002f086

File tree

8 files changed

+96
-8
lines changed

8 files changed

+96
-8
lines changed

build/asset-manifest.json

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
{
22
"files": {
33
"main.css": "./static/css/main.be588df3.css",
4-
"main.js": "./static/js/main.cd12942c.js",
4+
"main.js": "./static/js/main.6702bd47.js",
55
"index.html": "./index.html",
66
"main.be588df3.css.map": "./static/css/main.be588df3.css.map",
7-
"main.cd12942c.js.map": "./static/js/main.cd12942c.js.map"
7+
"main.6702bd47.js.map": "./static/js/main.6702bd47.js.map"
88
},
99
"entrypoints": [
1010
"static/css/main.be588df3.css",
11-
"static/js/main.cd12942c.js"
11+
"static/js/main.6702bd47.js"
1212
]
1313
}

build/index.html

Lines changed: 33 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -318,6 +318,27 @@ <h2 class="title is-3">Model Comparisions</h2>
318318
</div>
319319
</section>
320320

321+
<section class="section">
322+
<div class="container is-max-desktop">
323+
<div class="columns is-centered has-text-centered">
324+
<div class="column is-four-fifths">
325+
<h2 class="title is-3">Submitting Custom Models</h2>
326+
<div class="content has-text-justified">
327+
<p>
328+
To submit models to the leaderboard, you can run the evaluation using the evaluation scripts in
329+
<a href="https://github.com/LiveCodeBench/LiveCodeBench">GitHub</a>. Once you have the results,
330+
you can fill out <a href="https://forms.gle/h2abvAHh6UnhWzzd9">this form</a>. You will need to fill out
331+
model details and provide the generated evaluation file with model generations and pass@1 scores. We will
332+
review the submission and add the model to the leaderboard accordingly.
333+
</p>
334+
</div>
335+
</div>
336+
</div>
337+
</div>
338+
</section>
339+
340+
341+
321342

322343
<section class="section" id="BibTeX">
323344
<div class="container is-max-desktop content">
@@ -337,6 +358,18 @@ <h2 class="title">BibTeX</h2>
337358
<div class="columns is-centered">
338359
<div class="column is-8">
339360
<div class="content">
361+
<p>
362+
Please reach out to <a href="mailto:naman_jain@berkeley.edu">naman_jain@berkeley.edu</a> for questions or
363+
feedback on LiveCodeBench. We are also open to collaborations and suggestions for new scenarios to add to
364+
the benchmark. Finally, LiveCodeBench provides one axis of LLM coding evaluations and we recommend the
365+
following leaderboards for measuring code LM ability on various coding tasks, such as
366+
<a href="https://evalplus.github.io/leaderboard.html">EvalPlus Leaderboard</a>,
367+
<a href="https://crux-eval.github.io/leaderboard.html">CruxEval Leaderboard</a>,
368+
<a href="https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard">Chatbot Arena Leaderboard</a>,
369+
<a href="https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard">BigCode Models Leaderboard</a>,
370+
<a href="https://infi-coder.github.io/inficoder-eval/">InfiCoder-Eval</a>, and
371+
<a href="https://leaderboard.tabbyml.com/">TabbyML Leaderboard</a>.
372+
</p>
340373
<p>
341374
The source code from this website is borrowed from <a
342375
href="https://github.com/nerfies/nerfies.github.io">this template</a>!

build/leaderboard.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
<!doctype html><html lang="en"><head><title>LiveCodeBench Leaderboard</title><meta charset="UTF-8"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="description" content="LiveCodeBench Leaderboard"/><link rel="stylesheet" href="bootstrap.min.css"/><link rel="stylesheet" href="./css/bulma.min.css"/><link rel="stylesheet" href="./css/bulma-carousel.min.css"/><link rel="stylesheet" href="./css/bulma-slider.min.css"/><link rel="stylesheet" href="./css/fontawesome.all.min.css"/><link rel="stylesheet" href="https://cdn.jsdelivr.net/gh/jpswalsh/academicons@1/css/academicons.min.css"/><link rel="stylesheet" href="./css/index.css"/><link rel="icon" href="./images/favicon.svg"/><script src="https://ajax.googleapis.com/ajax/libs/jquery/3.5.1/jquery.min.js"></script><script defer="defer" src="./js/fontawesome.all.min.js"></script><script src="./js/bulma-carousel.min.js"></script><script src="./js/bulma-slider.min.js"></script><script src="./js/index.js"></script><script defer="defer" src="./static/js/main.cd12942c.js"></script><link href="./static/css/main.be588df3.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
1+
<!doctype html><html lang="en"><head><title>LiveCodeBench Leaderboard</title><meta charset="UTF-8"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="description" content="LiveCodeBench Leaderboard"/><link rel="stylesheet" href="bootstrap.min.css"/><link rel="stylesheet" href="./css/bulma.min.css"/><link rel="stylesheet" href="./css/bulma-carousel.min.css"/><link rel="stylesheet" href="./css/bulma-slider.min.css"/><link rel="stylesheet" href="./css/fontawesome.all.min.css"/><link rel="stylesheet" href="https://cdn.jsdelivr.net/gh/jpswalsh/academicons@1/css/academicons.min.css"/><link rel="stylesheet" href="./css/index.css"/><link rel="icon" href="./images/favicon.svg"/><script src="https://ajax.googleapis.com/ajax/libs/jquery/3.5.1/jquery.min.js"></script><script defer="defer" src="./js/fontawesome.all.min.js"></script><script src="./js/bulma-carousel.min.js"></script><script src="./js/bulma-slider.min.js"></script><script src="./js/index.js"></script><script defer="defer" src="./static/js/main.6702bd47.js"></script><link href="./static/css/main.be588df3.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
Lines changed: 3 additions & 3 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.
File renamed without changes.
Lines changed: 1 addition & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

public/index_home.html

Lines changed: 33 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -318,6 +318,27 @@ <h2 class="title is-3">Model Comparisions</h2>
318318
</div>
319319
</section>
320320

321+
<section class="section">
322+
<div class="container is-max-desktop">
323+
<div class="columns is-centered has-text-centered">
324+
<div class="column is-four-fifths">
325+
<h2 class="title is-3">Submitting Custom Models</h2>
326+
<div class="content has-text-justified">
327+
<p>
328+
To submit models to the leaderboard, you can run the evaluation using the evaluation scripts in
329+
<a href="https://github.com/LiveCodeBench/LiveCodeBench">GitHub</a>. Once you have the results,
330+
you can fill out <a href="https://forms.gle/h2abvAHh6UnhWzzd9">this form</a>. You will need to fill out
331+
model details and provide the generated evaluation file with model generations and pass@1 scores. We will
332+
review the submission and add the model to the leaderboard accordingly.
333+
</p>
334+
</div>
335+
</div>
336+
</div>
337+
</div>
338+
</section>
339+
340+
341+
321342

322343
<section class="section" id="BibTeX">
323344
<div class="container is-max-desktop content">
@@ -337,6 +358,18 @@ <h2 class="title">BibTeX</h2>
337358
<div class="columns is-centered">
338359
<div class="column is-8">
339360
<div class="content">
361+
<p>
362+
Please reach out to <a href="mailto:naman_jain@berkeley.edu">naman_jain@berkeley.edu</a> for questions or
363+
feedback on LiveCodeBench. We are also open to collaborations and suggestions for new scenarios to add to
364+
the benchmark. Finally, LiveCodeBench provides one axis of LLM coding evaluations and we recommend the
365+
following leaderboards for measuring code LM ability on various coding tasks, such as
366+
<a href="https://evalplus.github.io/leaderboard.html">EvalPlus Leaderboard</a>,
367+
<a href="https://crux-eval.github.io/leaderboard.html">CruxEval Leaderboard</a>,
368+
<a href="https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard">Chatbot Arena Leaderboard</a>,
369+
<a href="https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard">BigCode Models Leaderboard</a>,
370+
<a href="https://infi-coder.github.io/inficoder-eval/">InfiCoder-Eval</a>, and
371+
<a href="https://leaderboard.tabbyml.com/">TabbyML Leaderboard</a>.
372+
</p>
340373
<p>
341374
The source code from this website is borrowed from <a
342375
href="https://github.com/nerfies/nerfies.github.io">this template</a>!

src/index.tsx

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -119,6 +119,28 @@ ReactDOM.render(
119119
<div className="column has-text-centered">
120120
<LeaderboardTabs />
121121
</div>
122+
123+
124+
<section className="section">
125+
<div className="container is-max-desktop">
126+
<div className="columns is-centered has-text-centered">
127+
<div className="column is-four-fifths">
128+
<h2 className="title is-3">Submitting Custom Models</h2>
129+
<div className="content has-text-justified">
130+
<p>
131+
To submit models to the leaderboard, you can run the evaluation using the evaluation scripts in <a href="https://github.com/LiveCodeBench/LiveCodeBench">GitHub</a>. Once you have the results,
132+
you can fill out <a href="https://forms.gle/h2abvAHh6UnhWzzd9">this form</a>. You will need to fill out
133+
model details and provide the generated evaluation file with model generations and pass@1 scores. We will
134+
review the submission and add the model to the leaderboard accordingly.
135+
</p>
136+
</div>
137+
</div>
138+
</div>
139+
</div>
140+
</section>
141+
142+
143+
122144
</div>
123145
</div>
124146
</div>

0 commit comments

Comments
 (0)