Spaces:
Running
Running
Auto-restart every hour
Browse files- app.py +1 -13
- content.py +3 -0
app.py
CHANGED
@@ -306,7 +306,6 @@ def add_new_eval(
|
|
306 |
def refresh():
|
307 |
leaderboard = get_leaderboard()
|
308 |
finished_eval_queue, running_eval_queue, pending_eval_queue = get_eval_table()
|
309 |
-
get_leaderboard(), get_eval_table()
|
310 |
return leaderboard, finished_eval_queue, running_eval_queue, pending_eval_queue
|
311 |
|
312 |
|
@@ -420,18 +419,7 @@ We chose these benchmarks as they test a variety of reasoning and general knowle
|
|
420 |
submission_result,
|
421 |
)
|
422 |
|
423 |
-
# demo.load(
|
424 |
-
# refresh,
|
425 |
-
# inputs=[],
|
426 |
-
# outputs=[
|
427 |
-
# leaderboard_table,
|
428 |
-
# finished_eval_table,
|
429 |
-
# running_eval_table,
|
430 |
-
# pending_eval_table,
|
431 |
-
# ],
|
432 |
-
# )
|
433 |
-
|
434 |
scheduler = BackgroundScheduler()
|
435 |
-
scheduler.add_job(restart_space, 'interval', seconds=
|
436 |
scheduler.start()
|
437 |
demo.launch()
|
|
|
306 |
def refresh():
|
307 |
leaderboard = get_leaderboard()
|
308 |
finished_eval_queue, running_eval_queue, pending_eval_queue = get_eval_table()
|
|
|
309 |
return leaderboard, finished_eval_queue, running_eval_queue, pending_eval_queue
|
310 |
|
311 |
|
|
|
419 |
submission_result,
|
420 |
)
|
421 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
422 |
scheduler = BackgroundScheduler()
|
423 |
+
scheduler.add_job(restart_space, 'interval', seconds=3600)
|
424 |
scheduler.start()
|
425 |
demo.launch()
|
content.py
CHANGED
@@ -1,4 +1,7 @@
|
|
1 |
CHANGELOG_TEXT = f"""
|
|
|
|
|
|
|
2 |
## [2023-05-24]
|
3 |
- Added a baseline that has 25.0 for all values.
|
4 |
- Added CHANGELOG
|
|
|
1 |
CHANGELOG_TEXT = f"""
|
2 |
+
## [2023-05-29]
|
3 |
+
- Auto-restart every hour
|
4 |
+
|
5 |
## [2023-05-24]
|
6 |
- Added a baseline that has 25.0 for all values.
|
7 |
- Added CHANGELOG
|