task_id
stringlengths 8
69
| name
stringlengths 3
64
| suite
stringclasses 6
values | hf_repo
stringclasses 125
values | hf_subset
stringlengths 0
55
| file_path
stringclasses 6
values | line_number
int64 36
22.8k
| variable_name
stringclasses 1
value | is_subtask
bool 2
classes | main_task
stringclasses 122
values | subtask_count
int64 1
1
| suites
listlengths 1
1
| subtasks
listlengths 1
1
| is_standalone
bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
lighteval:blimp:sentential_negation_npi_licensor_present
|
blimp:sentential_negation_npi_licensor_present
|
lighteval
|
blimp
|
sentential_negation_npi_licensor_present
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,819
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:sentential_negation_npi_licensor_present"
] | false
|
helm:blimp:sentential_negation_npi_licensor_present
|
blimp:sentential_negation_npi_licensor_present
|
helm
|
blimp
|
sentential_negation_npi_licensor_present
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,834
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:sentential_negation_npi_licensor_present"
] | false
|
lighteval:blimp:sentential_negation_npi_scope
|
blimp:sentential_negation_npi_scope
|
lighteval
|
blimp
|
sentential_negation_npi_scope
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,852
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:sentential_negation_npi_scope"
] | false
|
helm:blimp:sentential_negation_npi_scope
|
blimp:sentential_negation_npi_scope
|
helm
|
blimp
|
sentential_negation_npi_scope
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,867
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:sentential_negation_npi_scope"
] | false
|
lighteval:blimp:sentential_subject_island
|
blimp:sentential_subject_island
|
lighteval
|
blimp
|
sentential_subject_island
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,885
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:sentential_subject_island"
] | false
|
helm:blimp:sentential_subject_island
|
blimp:sentential_subject_island
|
helm
|
blimp
|
sentential_subject_island
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,900
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:sentential_subject_island"
] | false
|
lighteval:blimp:superlative_quantifiers_1
|
blimp:superlative_quantifiers_1
|
lighteval
|
blimp
|
superlative_quantifiers_1
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,918
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:superlative_quantifiers_1"
] | false
|
helm:blimp:superlative_quantifiers_1
|
blimp:superlative_quantifiers_1
|
helm
|
blimp
|
superlative_quantifiers_1
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,933
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:superlative_quantifiers_1"
] | false
|
lighteval:blimp:superlative_quantifiers_2
|
blimp:superlative_quantifiers_2
|
lighteval
|
blimp
|
superlative_quantifiers_2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,951
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:superlative_quantifiers_2"
] | false
|
helm:blimp:superlative_quantifiers_2
|
blimp:superlative_quantifiers_2
|
helm
|
blimp
|
superlative_quantifiers_2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,966
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:superlative_quantifiers_2"
] | false
|
lighteval:blimp:tough_vs_raising_1
|
blimp:tough_vs_raising_1
|
lighteval
|
blimp
|
tough_vs_raising_1
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,984
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:tough_vs_raising_1"
] | false
|
helm:blimp:tough_vs_raising_1
|
blimp:tough_vs_raising_1
|
helm
|
blimp
|
tough_vs_raising_1
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 5,999
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:tough_vs_raising_1"
] | false
|
lighteval:blimp:tough_vs_raising_2
|
blimp:tough_vs_raising_2
|
lighteval
|
blimp
|
tough_vs_raising_2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,017
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:tough_vs_raising_2"
] | false
|
helm:blimp:tough_vs_raising_2
|
blimp:tough_vs_raising_2
|
helm
|
blimp
|
tough_vs_raising_2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,032
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:tough_vs_raising_2"
] | false
|
lighteval:blimp:transitive
|
blimp:transitive
|
lighteval
|
blimp
|
transitive
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,050
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:transitive"
] | false
|
helm:blimp:transitive
|
blimp:transitive
|
helm
|
blimp
|
transitive
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,065
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:transitive"
] | false
|
lighteval:blimp:wh_island
|
blimp:wh_island
|
lighteval
|
blimp
|
wh_island
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,083
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_island"
] | false
|
helm:blimp:wh_island
|
blimp:wh_island
|
helm
|
blimp
|
wh_island
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,098
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_island"
] | false
|
lighteval:blimp:wh_questions_object_gap
|
blimp:wh_questions_object_gap
|
lighteval
|
blimp
|
wh_questions_object_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,116
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_questions_object_gap"
] | false
|
helm:blimp:wh_questions_object_gap
|
blimp:wh_questions_object_gap
|
helm
|
blimp
|
wh_questions_object_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,131
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_questions_object_gap"
] | false
|
lighteval:blimp:wh_questions_subject_gap
|
blimp:wh_questions_subject_gap
|
lighteval
|
blimp
|
wh_questions_subject_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,149
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_questions_subject_gap"
] | false
|
helm:blimp:wh_questions_subject_gap
|
blimp:wh_questions_subject_gap
|
helm
|
blimp
|
wh_questions_subject_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,164
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_questions_subject_gap"
] | false
|
lighteval:blimp:wh_questions_subject_gap_long_distance
|
blimp:wh_questions_subject_gap_long_distance
|
lighteval
|
blimp
|
wh_questions_subject_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,182
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_questions_subject_gap_long_distance"
] | false
|
helm:blimp:wh_questions_subject_gap_long_distance
|
blimp:wh_questions_subject_gap_long_distance
|
helm
|
blimp
|
wh_questions_subject_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,197
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_questions_subject_gap_long_distance"
] | false
|
lighteval:blimp:wh_vs_that_no_gap
|
blimp:wh_vs_that_no_gap
|
lighteval
|
blimp
|
wh_vs_that_no_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,215
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_vs_that_no_gap"
] | false
|
helm:blimp:wh_vs_that_no_gap
|
blimp:wh_vs_that_no_gap
|
helm
|
blimp
|
wh_vs_that_no_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,230
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_vs_that_no_gap"
] | false
|
lighteval:blimp:wh_vs_that_no_gap_long_distance
|
blimp:wh_vs_that_no_gap_long_distance
|
lighteval
|
blimp
|
wh_vs_that_no_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,248
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_vs_that_no_gap_long_distance"
] | false
|
helm:blimp:wh_vs_that_no_gap_long_distance
|
blimp:wh_vs_that_no_gap_long_distance
|
helm
|
blimp
|
wh_vs_that_no_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,263
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_vs_that_no_gap_long_distance"
] | false
|
lighteval:blimp:wh_vs_that_with_gap
|
blimp:wh_vs_that_with_gap
|
lighteval
|
blimp
|
wh_vs_that_with_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,281
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_vs_that_with_gap"
] | false
|
helm:blimp:wh_vs_that_with_gap
|
blimp:wh_vs_that_with_gap
|
helm
|
blimp
|
wh_vs_that_with_gap
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,296
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_vs_that_with_gap"
] | false
|
lighteval:blimp:wh_vs_that_with_gap_long_distance
|
blimp:wh_vs_that_with_gap_long_distance
|
lighteval
|
blimp
|
wh_vs_that_with_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,314
|
direct_call
| true
|
blimp
| 1
|
[
"lighteval"
] |
[
"lighteval:blimp:wh_vs_that_with_gap_long_distance"
] | false
|
helm:blimp:wh_vs_that_with_gap_long_distance
|
blimp:wh_vs_that_with_gap_long_distance
|
helm
|
blimp
|
wh_vs_that_with_gap_long_distance
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,329
|
direct_call
| true
|
blimp
| 1
|
[
"helm"
] |
[
"helm:blimp:wh_vs_that_with_gap_long_distance"
] | false
|
helm:bold
|
bold
|
helm
|
lighteval/bold_helm
|
all
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,347
|
direct_call
| false
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold"
] | false
|
helm:bold:gender
|
bold:gender
|
helm
|
lighteval/bold_helm
|
gender
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,362
|
direct_call
| true
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold:gender"
] | false
|
helm:bold:political_ideology
|
bold:political_ideology
|
helm
|
lighteval/bold_helm
|
political_ideology
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,377
|
direct_call
| true
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold:political_ideology"
] | false
|
helm:bold:profession
|
bold:profession
|
helm
|
lighteval/bold_helm
|
profession
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,392
|
direct_call
| true
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold:profession"
] | false
|
helm:bold:race
|
bold:race
|
helm
|
lighteval/bold_helm
|
race
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,407
|
direct_call
| true
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold:race"
] | false
|
helm:bold:religious_ideology
|
bold:religious_ideology
|
helm
|
lighteval/bold_helm
|
religious_ideology
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,422
|
direct_call
| true
|
bold
| 1
|
[
"helm"
] |
[
"helm:bold:religious_ideology"
] | false
|
helm:boolq
|
boolq
|
helm
|
lighteval/boolq_helm
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,437
|
direct_call
| false
|
boolq
| 1
|
[
"helm"
] |
[
"helm:boolq"
] | false
|
helm:boolq:contrastset
|
boolq:contrastset
|
helm
|
lighteval/boolq_helm
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,463
|
direct_call
| true
|
boolq
| 1
|
[
"helm"
] |
[
"helm:boolq:contrastset"
] | false
|
lighteval:bigbench:bridging_anaphora_resolution_barqa
|
bigbench:bridging_anaphora_resolution_barqa
|
lighteval
|
tasksource/bigbench
|
bridging_anaphora_resolution_barqa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,489
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:bridging_anaphora_resolution_barqa"
] | false
|
lighteval:bigbench:cause_and_effect
|
bigbench:cause_and_effect
|
lighteval
|
tasksource/bigbench
|
cause_and_effect
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,519
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:cause_and_effect"
] | false
|
lighteval:bigbench:checkmate_in_one
|
bigbench:checkmate_in_one
|
lighteval
|
tasksource/bigbench
|
checkmate_in_one
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,534
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:checkmate_in_one"
] | false
|
lighteval:bigbench:chess_state_tracking
|
bigbench:chess_state_tracking
|
lighteval
|
tasksource/bigbench
|
chess_state_tracking
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,549
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:chess_state_tracking"
] | false
|
lighteval:bigbench:chinese_remainder_theorem
|
bigbench:chinese_remainder_theorem
|
lighteval
|
tasksource/bigbench
|
chinese_remainder_theorem
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,564
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:chinese_remainder_theorem"
] | false
|
lighteval:bigbench:cifar10_classification
|
bigbench:cifar10_classification
|
lighteval
|
tasksource/bigbench
|
cifar10_classification
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,579
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:cifar10_classification"
] | false
|
helm:civil_comments
|
civil_comments
|
helm
|
lighteval/civil_comments_helm
|
all
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,594
|
direct_call
| false
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments"
] | false
|
helm:civil_comments:LGBTQ
|
civil_comments:LGBTQ
|
helm
|
lighteval/civil_comments_helm
|
LGBTQ
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,622
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:LGBTQ"
] | false
|
helm:civil_comments:black
|
civil_comments:black
|
helm
|
lighteval/civil_comments_helm
|
black
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,650
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:black"
] | false
|
helm:civil_comments:christian
|
civil_comments:christian
|
helm
|
lighteval/civil_comments_helm
|
christian
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,678
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:christian"
] | false
|
helm:civil_comments:female
|
civil_comments:female
|
helm
|
lighteval/civil_comments_helm
|
female
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,706
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:female"
] | false
|
helm:civil_comments:male
|
civil_comments:male
|
helm
|
lighteval/civil_comments_helm
|
male
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,734
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:male"
] | false
|
helm:civil_comments:muslim
|
civil_comments:muslim
|
helm
|
lighteval/civil_comments_helm
|
muslim
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,762
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:muslim"
] | false
|
helm:civil_comments:other_religions
|
civil_comments:other_religions
|
helm
|
lighteval/civil_comments_helm
|
other_religions
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,790
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:other_religions"
] | false
|
helm:civil_comments:white
|
civil_comments:white
|
helm
|
lighteval/civil_comments_helm
|
white
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,818
|
direct_call
| true
|
civil_comments
| 1
|
[
"helm"
] |
[
"helm:civil_comments:white"
] | false
|
lighteval:bigbench_lite:code_line_description
|
bigbench_lite:code_line_description
|
lighteval
|
tasksource/bigbench
|
code_line_description
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,846
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:code_line_description"
] | false
|
lighteval:bigbench:codenames
|
bigbench:codenames
|
lighteval
|
tasksource/bigbench
|
codenames
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,861
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:codenames"
] | false
|
lighteval:bigbench:color
|
bigbench:color
|
lighteval
|
tasksource/bigbench
|
color
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,876
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:color"
] | false
|
lighteval:bigbench:common_morpheme
|
bigbench:common_morpheme
|
lighteval
|
tasksource/bigbench
|
common_morpheme
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,896
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:common_morpheme"
] | false
|
helm:commonsenseqa
|
commonsenseqa
|
helm
|
commonsense_qa
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,911
|
direct_call
| false
|
commonsenseqa
| 1
|
[
"helm"
] |
[
"helm:commonsenseqa"
] | true
|
lighteval:bigbench_lite:conceptual_combinations
|
bigbench_lite:conceptual_combinations
|
lighteval
|
tasksource/bigbench
|
conceptual_combinations
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,937
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:conceptual_combinations"
] | false
|
lighteval:bigbench_lite:conlang_translation
|
bigbench_lite:conlang_translation
|
lighteval
|
tasksource/bigbench
|
conlang_translation
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,952
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:conlang_translation"
] | false
|
lighteval:bigbench:contextual_parametric_knowledge_conflicts
|
bigbench:contextual_parametric_knowledge_conflicts
|
lighteval
|
tasksource/bigbench
|
contextual_parametric_knowledge_conflicts
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,967
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:contextual_parametric_knowledge_conflicts"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_125
|
copyright:n_books_1000-extractions_per_book_1-prefix_length_125
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_1-prefix_length_125
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,982
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_125"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_25
|
copyright:n_books_1000-extractions_per_book_1-prefix_length_25
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_1-prefix_length_25
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 6,997
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_25"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_5
|
copyright:n_books_1000-extractions_per_book_1-prefix_length_5
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_1-prefix_length_5
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,012
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_5"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_125
|
copyright:n_books_1000-extractions_per_book_3-prefix_length_125
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_3-prefix_length_125
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,027
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_125"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_25
|
copyright:n_books_1000-extractions_per_book_3-prefix_length_25
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_3-prefix_length_25
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,042
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_25"
] | false
|
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_5
|
copyright:n_books_1000-extractions_per_book_3-prefix_length_5
|
helm
|
lighteval/copyright_helm
|
n_books_1000-extractions_per_book_3-prefix_length_5
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,057
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_5"
] | false
|
helm:copyright:oh_the_places
|
copyright:oh_the_places
|
helm
|
lighteval/copyright_helm
|
oh_the_places
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,072
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:oh_the_places"
] | false
|
helm:copyright:pilot
|
copyright:pilot
|
helm
|
lighteval/copyright_helm
|
pilot
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,087
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:pilot"
] | false
|
helm:copyright:popular_books-prefix_length_10
|
copyright:popular_books-prefix_length_10
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_10
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,102
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_10"
] | false
|
helm:copyright:popular_books-prefix_length_125
|
copyright:popular_books-prefix_length_125
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_125
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,117
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_125"
] | false
|
helm:copyright:popular_books-prefix_length_25
|
copyright:popular_books-prefix_length_25
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_25
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,132
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_25"
] | false
|
helm:copyright:popular_books-prefix_length_250
|
copyright:popular_books-prefix_length_250
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_250
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,147
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_250"
] | false
|
helm:copyright:popular_books-prefix_length_5
|
copyright:popular_books-prefix_length_5
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_5
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,162
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_5"
] | false
|
helm:copyright:popular_books-prefix_length_50
|
copyright:popular_books-prefix_length_50
|
helm
|
lighteval/copyright_helm
|
popular_books-prefix_length_50
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,177
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:popular_books-prefix_length_50"
] | false
|
helm:copyright:prompt_num_line_1-min_lines_20
|
copyright:prompt_num_line_1-min_lines_20
|
helm
|
lighteval/copyright_helm
|
prompt_num_line_1-min_lines_20
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,192
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:prompt_num_line_1-min_lines_20"
] | false
|
helm:copyright:prompt_num_line_10-min_lines_20
|
copyright:prompt_num_line_10-min_lines_20
|
helm
|
lighteval/copyright_helm
|
prompt_num_line_10-min_lines_20
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,207
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:prompt_num_line_10-min_lines_20"
] | false
|
helm:copyright:prompt_num_line_5-min_lines_20
|
copyright:prompt_num_line_5-min_lines_20
|
helm
|
lighteval/copyright_helm
|
prompt_num_line_5-min_lines_20
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,222
|
direct_call
| true
|
copyright
| 1
|
[
"helm"
] |
[
"helm:copyright:prompt_num_line_5-min_lines_20"
] | false
|
lighteval:coqa
|
coqa
|
lighteval
|
stanfordnlp/coqa
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,237
|
direct_call
| false
|
coqa
| 1
|
[
"lighteval"
] |
[
"lighteval:coqa"
] | true
|
lighteval:coqa_bb
|
coqa_bb
|
lighteval
|
coqa
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,266
|
direct_call
| false
|
coqa_bb
| 1
|
[
"lighteval"
] |
[
"lighteval:coqa_bb"
] | true
|
helm:covid_dialogue
|
covid_dialogue
|
helm
|
lighteval/covid_dialogue
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,281
|
direct_call
| false
|
covid_dialogue
| 1
|
[
"helm"
] |
[
"helm:covid_dialogue"
] | true
|
lighteval:bigbench:crash_blossom
|
bigbench:crash_blossom
|
lighteval
|
tasksource/bigbench
|
crash_blossom
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,303
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:crash_blossom"
] | false
|
lighteval:bigbench:crass_ai
|
bigbench:crass_ai
|
lighteval
|
tasksource/bigbench
|
crass_ai
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,318
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:crass_ai"
] | false
|
lighteval:bigbench:cryobiology_spanish
|
bigbench:cryobiology_spanish
|
lighteval
|
tasksource/bigbench
|
cryobiology_spanish
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,333
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:cryobiology_spanish"
] | false
|
lighteval:bigbench:cryptonite
|
bigbench:cryptonite
|
lighteval
|
tasksource/bigbench
|
cryptonite
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,348
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:cryptonite"
] | false
|
lighteval:bigbench:cs_algorithms
|
bigbench:cs_algorithms
|
lighteval
|
tasksource/bigbench
|
cs_algorithms
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,363
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:cs_algorithms"
] | false
|
lighteval:bigbench:dark_humor_detection
|
bigbench:dark_humor_detection
|
lighteval
|
tasksource/bigbench
|
dark_humor_detection
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,378
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:dark_humor_detection"
] | false
|
lighteval:bigbench:discourse_marker_prediction
|
bigbench:discourse_marker_prediction
|
lighteval
|
tasksource/bigbench
|
discourse_marker_prediction
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,423
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:discourse_marker_prediction"
] | false
|
lighteval:bigbench:disfl_qa
|
bigbench:disfl_qa
|
lighteval
|
tasksource/bigbench
|
disfl_qa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,438
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:disfl_qa"
] | false
|
lighteval:drop
|
drop
|
lighteval
|
lighteval/drop_harness
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,453
|
direct_call
| false
|
drop
| 1
|
[
"lighteval"
] |
[
"lighteval:drop"
] | true
|
helm:dyck_language:2
|
dyck_language:2
|
helm
|
lighteval/DyckLanguage
|
2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,497
|
direct_call
| true
|
dyck_language
| 1
|
[
"helm"
] |
[
"helm:dyck_language:2"
] | false
|
helm:dyck_language:3
|
dyck_language:3
|
helm
|
lighteval/DyckLanguage
|
3
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,512
|
direct_call
| true
|
dyck_language
| 1
|
[
"helm"
] |
[
"helm:dyck_language:3"
] | false
|
helm:dyck_language:4
|
dyck_language:4
|
helm
|
lighteval/DyckLanguage
|
4
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,527
|
direct_call
| true
|
dyck_language
| 1
|
[
"helm"
] |
[
"helm:dyck_language:4"
] | false
|
lighteval:bigbench:dyck_languages
|
bigbench:dyck_languages
|
lighteval
|
tasksource/bigbench
|
dyck_languages
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,542
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:dyck_languages"
] | false
|
lighteval:bigbench:elementary_math_qa
|
bigbench:elementary_math_qa
|
lighteval
|
tasksource/bigbench
|
elementary_math_qa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,557
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:elementary_math_qa"
] | false
|
lighteval:bigbench_lite:emoji_movie
|
bigbench_lite:emoji_movie
|
lighteval
|
tasksource/bigbench
|
emoji_movie
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,572
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:emoji_movie"
] | false
|
lighteval:bigbench:emojis_emotion_prediction
|
bigbench:emojis_emotion_prediction
|
lighteval
|
tasksource/bigbench
|
emojis_emotion_prediction
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,592
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:emojis_emotion_prediction"
] | false
|
lighteval:bigbench:empirical_judgments
|
bigbench:empirical_judgments
|
lighteval
|
tasksource/bigbench
|
empirical_judgments
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 7,607
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:empirical_judgments"
] | false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.