-
BERT-as-a-Judge: A Robust Alternative to Lexical Methods for Efficient Reference-Based LLM Evaluation
Paper • 2604.09497 • Published • 16 -
artefactory/BERTJudge
0.2B • Updated • 18 • 1 -
artefactory/BERTJudge-Formatted-QCR
0.2B • Updated -
artefactory/BERTJudge-Formatted-CR
0.2B • Updated • 6
AI & ML interests
NLP, Information Retrieval, Computer Vision, Uncertainty Estimation, Trustworthy AI, Bias Estimation, Unbalanced ML, Choice Modeling, Time Series
Recent Activity
View all activity
Papers
BERT-as-a-Judge: A Robust Alternative to Lexical Methods for Efficient Reference-Based LLM Evaluation
Learned Hallucination Detection in Black-Box LLMs using Token-level Entropy Production Rate
Related paper: "Should We Still Pretrain Encoders with Masked Language Modeling?"
-
Should We Still Pretrain Encoders with Masked Language Modeling?
Paper • 2507.00994 • Published • 81 -
MLMvsCLM/610m-mlm30-42k
Feature Extraction • Updated • 12 -
MLMvsCLM/610m-mlm40-42k-2000
Feature Extraction • Updated • 13 -
MLMvsCLM/610m-clm-17k-mlm40-22k
Feature Extraction • Updated • 12
Related paper: "Towards Trustworthy Reranking: A Simple yet Effective Abstention Mechanism" (accepted at TMLR 2024)
Suite of Encoder models EuroBERT
-
EuroBERT: Scaling Multilingual Encoders for European Languages
Paper • 2503.05500 • Published • 81 -
EuroBERT/EuroBERT-210m
Fill-Mask • 0.3B • Updated • 8.65k • 83 -
EuroBERT/EuroBERT-610m
Fill-Mask • 0.8B • Updated • 1.62k • 34 -
EuroBERT/EuroBERT-2.1B
Fill-Mask • 2B • Updated • 1.07k • 67
Related paper: "Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis" (accepted at WMT 2024)
-
Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis
Paper • 2409.20059 • Published • 16 -
artefactory/ALMA-13B-LoRA
Text Generation • 13B • Updated • 5 -
artefactory/ALMA-13B-LoRA-SFT-xCOMET-QE-Multi
Text Generation • 13B • Updated • 3 -
artefactory/ALMA-13B-LoRA-SFT-xCOMET-QE-Multi-No-Base
Text Generation • 13B • Updated • 1
-
BERT-as-a-Judge: A Robust Alternative to Lexical Methods for Efficient Reference-Based LLM Evaluation
Paper • 2604.09497 • Published • 16 -
artefactory/BERTJudge
0.2B • Updated • 18 • 1 -
artefactory/BERTJudge-Formatted-QCR
0.2B • Updated -
artefactory/BERTJudge-Formatted-CR
0.2B • Updated • 6
Related paper: "Should We Still Pretrain Encoders with Masked Language Modeling?"
-
Should We Still Pretrain Encoders with Masked Language Modeling?
Paper • 2507.00994 • Published • 81 -
MLMvsCLM/610m-mlm30-42k
Feature Extraction • Updated • 12 -
MLMvsCLM/610m-mlm40-42k-2000
Feature Extraction • Updated • 13 -
MLMvsCLM/610m-clm-17k-mlm40-22k
Feature Extraction • Updated • 12
Suite of Encoder models EuroBERT
-
EuroBERT: Scaling Multilingual Encoders for European Languages
Paper • 2503.05500 • Published • 81 -
EuroBERT/EuroBERT-210m
Fill-Mask • 0.3B • Updated • 8.65k • 83 -
EuroBERT/EuroBERT-610m
Fill-Mask • 0.8B • Updated • 1.62k • 34 -
EuroBERT/EuroBERT-2.1B
Fill-Mask • 2B • Updated • 1.07k • 67
Related paper: "Towards Trustworthy Reranking: A Simple yet Effective Abstention Mechanism" (accepted at TMLR 2024)
Related paper: "Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis" (accepted at WMT 2024)
-
Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis
Paper • 2409.20059 • Published • 16 -
artefactory/ALMA-13B-LoRA
Text Generation • 13B • Updated • 5 -
artefactory/ALMA-13B-LoRA-SFT-xCOMET-QE-Multi
Text Generation • 13B • Updated • 3 -
artefactory/ALMA-13B-LoRA-SFT-xCOMET-QE-Multi-No-Base
Text Generation • 13B • Updated • 1