Skip to content

sberbank-ai-lab/RuMedBench

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Python 3.7

RuMedBench

A Russian Medical language understanding Benchmark is the set of NLP tasks on medical textual data for the Russian language.

This repository contains code and data to reproduce the results of the paper RuMedBench: A Russian Medical Language Understanding Benchmark.

Tasks Descriptions

  • RuMedTop3* is the task for diagnosis prediction from a raw medical text, including patient symptoms and complaints.

  • RuMedSymptomRec* Given an incomplete medical text, the task is to recommend the best symptom to check or verify.

  • RuMedDaNet is the yes/no question answering task in the range of medical-related domains (pharmacology, anatomy, therapeutic medicine, etc).

  • RuMedNLI is the natural language inference task in the clinical domain. The data is the full translated counterpart of MedNLI data.

  • RuMedNER is the task of named entity recognition in drug-related user reviews. The data is from the RuDReC repo.

*Both tasks are based on the RuMedPrime dataset.

Baselines & Results

We have implemented several baseline models; please see details in the paper.

Accuracy is the base metric for all tasks evaluation. For some tasks, additional metrics are used:

  • RuMedTop3 and RuMedSymptomRec - Hit@3
  • RuMedNER - F1-score

Test results:

Model RuMedTop3 RuMedSymptomRec RuMedDaNet RuMedNLI RuMedNER Overall
Naive 10.58/22.02 1.93/5.30 50.00 33.33 93.66/51.96 35.21
Feature-based 49.76/72.75 32.05/49.40 51.95 59.70 94.40/62.89 58.46
BiLSTM 40.88/63.50 20.24/31.33 52.34 60.06 94.74/63.26 53.87
RuBERT 39.54/62.29 18.55/34.22 67.19 77.64 96.63/73.53 61.44
RuPoolBERT 47.45/70.44 34.94/52.05 71.48 77.29 96.47/73.15 67.20
Human 25.06/48.54 7.23/12.53 93.36 83.26 96.09/76.18 61.89

We define the overall model score as mean over all metric values (with prior averaging in the case of two metrics).

How to Run

Please refer to the code/ directory.

Contact

If you have any questions, please post a Github issue or email the authors.

Citation

@misc{blinov2022rumedbench,
    title={RuMedBench: A Russian Medical Language Understanding Benchmark},
    author={Pavel Blinov and Arina Reshetnikova and Aleksandr Nesterov and Galina Zubkova and Vladimir Kokh},
    year={2022},
    eprint={2201.06499},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}