当前位置: X-MOL 学术Artif. Intell. Med. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
BertMCN: Mapping colloquial phrases to standard medical concepts using BERT and highway network
Artificial Intelligence in Medicine ( IF 7.5 ) Pub Date : 2021-01-07 , DOI: 10.1016/j.artmed.2021.102008
Katikapalli Subramanyam Kalyan 1 , Sivanesan Sangeetha 1
Affiliation  

In the last few years, people started to share lots of information related to health in the form of tweets, reviews and blog posts. All these user generated clinical texts can be mined to generate useful insights. However, automatic analysis of clinical text requires identification of standard medical concepts. Most of the existing deep learning based medical concept normalization systems are based on CNN or RNN. Performance of these models is limited as they have to be trained from scratch (except embeddings). In this work, we propose a medical concept normalization system based on BERT and highway layer. BERT, a pre-trained context sensitive deep language representation model advanced state-of-the-art performance in many NLP tasks and gating mechanism in highway layer helps the model to choose only important information. Experimental results show that our model outperformed all existing methods on two standard datasets. Further, we conduct a series of experiments to study the impact of different learning rates and batch sizes, noise and freezing encoder layers on our model.



中文翻译:

BertMCN:使用 BERT 和高速公路网络将口语短语映射到标准医学概念

在过去的几年里,人们开始以推文、评论和博客文章的形式分享大量与健康相关的信息。所有这些用户生成的临床文本都可以被挖掘以产生有用的见解。然而,临床文本的自动分析需要识别标准医学概念。大多数现有的基于深度学习的医学概念归一化系统都是基于 CNN 或 RNN。这些模型的性能是有限的,因为它们必须从头开始训练(嵌入除外)。在这项工作中,我们提出了一种基于 BERT 和高速公路层的医学概念规范化系统。BERT 是一种预训练的上下文敏感的深度语言表示模型,在许多 NLP 任务中具有先进的性能,并且高速公路层的门控机制有助于模型仅选择重要信息。实验结果表明,我们的模型在两个标准数据集上的表现优于所有现有方法。此外,我们进行了一系列实验来研究不同学习率和批量大小、噪声和冻结编码器层对我们模型的影响。

更新日期:2021-01-16
down
wechat
bug