Automatic Evaluation Metric described in the paper BERTScore: Evaluating Text Generation with BERT (ICLR 2020). We now support about 130 models (see this spreadsheet for their correlations with human evaluation). Currently, the best model is Microsoft/debate-large-online, please consider using it instead of the default roberta-large in order to have the best correlation with human evaluation.