mattdl commited on
Commit
a480476
·
verified ·
1 Parent(s): 721a27e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -0
README.md CHANGED
@@ -317,6 +317,24 @@ In this example:
317
 
318
  ### BibTeX
319
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
320
  #### Sentence Transformers
321
  ```bibtex
322
  @inproceedings{reimers-2019-sentence-bert,
 
317
 
318
  ### BibTeX
319
 
320
+ ### JobBERT-v2 paper
321
+ Please cite this paper when using JobBERT-v2:
322
+ ```bibtex
323
+ @article{01K47W55SG7ZRKFG431ESRXC35,
324
+ abstract = {{Labor market analysis relies on extracting insights from job advertisements, which provide valuable yet unstructured information on job titles and corresponding skill requirements. While state-of-the-art methods for skill extraction achieve strong performance, they depend on large language models (LLMs), which are computationally expensive and slow. In this paper, we propose ConTeXT-match, a novel contrastive learning approach with token-level attention that is well-suited for the extreme multi-label classification task of skill classification. ConTeXT-match significantly improves skill extraction efficiency and performance, achieving state-of-the-art results with a lightweight bi-encoder model. To support robust evaluation, we introduce Skill-XL a new benchmark with exhaustive, sentence-level skill annotations that explicitly address the redundancy in the large label space. Finally, we present JobBERT V2, an improved job title normalization model that leverages extracted skills to produce high-quality job title representations. Experiments demonstrate that our models are efficient, accurate, and scalable, making them ideal for large-scale, real-time labor market analysis.}},
325
+ author = {{Decorte, Jens-Joris and Van Hautte, Jeroen and Develder, Chris and Demeester, Thomas}},
326
+ issn = {{2169-3536}},
327
+ journal = {{IEEE ACCESS}},
328
+ keywords = {{Taxonomy,Contrastive learning,Training,Annotations,Benchmark testing,Training data,Large language models,Computational efficiency,Accuracy,Terminology,Labor market analysis,text encoders,skill extraction,job title normalization}},
329
+ language = {{eng}},
330
+ pages = {{133596--133608}},
331
+ title = {{Efficient text encoders for labor market analysis}},
332
+ url = {{http://doi.org/10.1109/ACCESS.2025.3589147}},
333
+ volume = {{13}},
334
+ year = {{2025}},
335
+ }
336
+ ```
337
+
338
  #### Sentence Transformers
339
  ```bibtex
340
  @inproceedings{reimers-2019-sentence-bert,