We support the evaluation of bert model (model 2) and citation-bert model (model 3). Currently we calculate the following metrics:
-
mAP: calculating mAP is relatively slow, if you want to speed up the evaluation, just ignoring mAP by changing
recorder = ResultRecorder(paper_num, include_mAP = True, recall_K = RECALL_K)
to
recorder = ResultRecorder(paper_num, include_mAP = False, recall_K = RECALL_K)
in the evaluation program (
eval_bert.py
andeval_citation_bert.py
). -
mRR: the mean reciprocal rank;
-
Recall@K: the hit rate of the top-K recommendations. Here the values of
K
is specified inrecall_K
item in the configuration file.