Selected article for: "machine learning and training set"

Author: Langnickel, Lisa; Fluck, Juliane
Title: We are not ready yet: limitations of transfer learning for Disease Named Entity Recognition
  • Cord-id: kqswzqtp
  • Document date: 2021_7_12
  • ID: kqswzqtp
    Snippet: Intense research has been done in the area of biomedical natural language processing. Since the breakthrough of transfer learning-based methods, BERT models are used in a variety of biomedical and clinical applications. For the available data sets, these models show excellent results – partly exceeding the inter-annotator agreements. However, biomedical named entity recognition applied on COVID-19 preprints shows a performance drop compared to the results on available test data. The question a
    Document: Intense research has been done in the area of biomedical natural language processing. Since the breakthrough of transfer learning-based methods, BERT models are used in a variety of biomedical and clinical applications. For the available data sets, these models show excellent results – partly exceeding the inter-annotator agreements. However, biomedical named entity recognition applied on COVID-19 preprints shows a performance drop compared to the results on available test data. The question arises how well trained models are able to predict on completely new data, i.e. to generalize. Based on the example of disease named entity recognition, we investigate the robustness of different machine learning-based methods – thereof transfer learning – and show that current state-of-the-art methods work well for a given training and the corresponding test set but experience a significant lack of generalization when applying to new data. We therefore argue that there is a need for larger annotated data sets for training and testing.

    Search related documents:
    Co phrase search for related documents
    • Try single phrases listed below for: 1
    Co phrase search for related documents, hyperlinks ordered by date