Microsoft Research develops AI ``BioGPT'' specialized in the biomedical field, capable of answering questions with accuracy comparable to human experts



Modern science is built on the vast amount of scientific knowledge that has been accumulated so far, and it is extremely important to extract important information from the vast amount of previous research. So a team at

Microsoft Research , a computer science lab, has developed BioGPT , an AI trained on a vast amount of medical literature and specialized in the task of answering questions in the biomedical field.

[2210.10341] BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining
https://arxiv.org/abs/2210.10341

GitHub-microsoft/BioGPT
https://github.com/microsoft/BioGPT



BioGPT is a Microsoft language model trained for biomedical tasks
https://the-decoder.com/biogpt-is-a-microsoft-language-model-trained-for-biomedical-tasks/

Microsoft Research Proposes BioGPT: A Domain-Specific Generative Transformer Language Model Pre-Trained on Large-Scale Biomedical Literature - MarkTechPost
https://www.marktechpost.com/2023/01/31/microsoft-research-proposes-biogpt-a-domain-specific-generative-transformer-language-model-pre-trained-on-large-scale-biomedical- literature/

Researchers have long suspected that pre-trained language models could be useful in efficiently extracting relevant knowledge from the vast medical literature. However, the accuracy of applying a normal language model to the biomedical field as it is was not sufficient, so a method of ``training a language model with medical literature'' was devised, and so far, ` `BioBERT '' and ` `PubMedBERT '' have been developed. A pre-trained language model has been developed specifically for the biomedical domain of

BioBERT and PubMedBERT were based on BERT , a natural language processing model developed by Google that excels in contextual understanding, but the performance of OpenAI's GPT model, which excels at generative tasks, in the biomedical field is still under scrutiny. He said he didn't. Therefore, the research team at Microsoft Research decided to develop a language model trained on a wide range of biomedical literature based on `` GPT-2 '' announced in 2019.



In developing a language model specialized in the biomedical field, the research team collected text-based English papers published before 2021 from the

PubMed database, a search engine for academic literature on biomedical science. Using the 15 million pieces of content, including titles and summaries, as a dataset, 200,000 steps of pre-training were performed using eight Nvidia V100 GPUs for AI training, followed by one Nvidia V100 GPU. It seems that 32 steps of fine adjustment were performed. Next, he improved the model using 375 million parameters for tasks such as end-to-end relationship extraction, text generation, question and answer, and document classification.

The AI developed in this way, specialized in the biomedical field, is ``BioGPT,'' and the research team says that it is superior to past models based on BERT in answering questions in the biomedical field and extracting end-to-end relationships. claim. Below is an example of a sentence generated by the research team entering a question sentence into BioGPT.

Input:
The treatment of COVID-19
input:
Treatment for COVID-19

Output BioGPT:
The treatment of COVID-19 with remdesivir (remdesivir for COVID-19) has been approved by the US Food and Drug Administration (FDA) for the treatment of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) infection in patients with confirmed SARS-CoV-2 infection and is currently being evaluated in clinical trials for the treatment of COVID-19 in patients with confirmed SARS-CoV-2 infection, as well as in patients with suspected or confirmed SARS-CoV-2 infection.
BioGPT output:
Treatment of COVID-19 with remdesivir is approved by the U.S. Food and Drug Administration (FDA) for the treatment of severe acute respiratory syndrome coronavirus 2 in patients with confirmed SARS-CoV-2 infection. It is currently being administered to patients with suspected or confirmed SARS-CoV-2 infection as a treatment for COVID-19 and is being evaluated in clinical trials.



In addition, BioGPT scored 78.2% accuracy on the PubMed-provided dataset PubMedQA benchmark, beating the 78% performance of human experts. In addition, 'BioGPTLarge', which scales up BioGPT to the largest available GPT-2 architecture, has an accuracy of 81%, which ranks first in the ranking.



The number of parameters of BioGPTLarge is 1.5 billion, which is significantly lower than ' Flan-PaLM ' with 540 billion parameters and ' Galactica ' with 120 billion parameters. The Decofer, an AI-related web media, said the results show that relatively small, domain-specific language models can compete well with much larger general language models. I said.

in Software,   Science, Posted by log1h_ik