Abstract

Biological knowledgebases are essential resources for biomedical researchers, providing ready access to gene function and genomic data. Professional, manual curation of knowledgebases, however, is labour-intensive and thus high-performing machine learning (ML) methods that improve biocuration efficiency are needed. Here, we report on sentence-level classification to identify biocuration-relevant sentences in the full text of published references for two gene function data types: gene expression and protein kinase activity. We performed a detailed characterization of sentences from references in the WormBase bibliography and used this characterization to define three tasks for classifying sentences as either (i) fully curatable, (ii) fully and partially curatable, or (iii) all language-related. We evaluated various ML models applied to these tasks and found that GPT and BioBERT achieve the highest average performance, resulting in F1 performance scores ranging from 0.89 to 0.99 depending upon the task. Moreover, our inter-annotator agreement analyses and curator timing exercises demonstrated that curators readily converged on classification of high-quality training sentences that take a relatively short period of time to collect, making expansion of this approach to other data types a realistic addition to existing biocuration workflows. Our findings demonstrate the feasibility of extracting biocuration-relevant sentences from full text. Integrating these models into professional biocuration workflows, such as those used by the Alliance of Genome Resources and the ACKnowledge community curation platform, might well facilitate efficient and accurate annotation of the biomedical literature.

Introduction

Biological knowledgebases, hereafter referred to as knowledgebases, are a key resource for biomedical researchers, affording them ready access to gene function and genomic data [1]. Primarily maintained by professional biocurators, knowledgebases adhere to strict data models and standards in accordance with FAIR principles (Findable, Accessible, Interoperable, and Reusable) [2]. An important aspect of FAIR biocuration is preserving the provenance of biological knowledge, which typically involves a manual process of identifying relevant published data, translating that data into models using standardized identifiers and ontologies, and associating the curated data with the original peer-reviewed reference [3]. While manual biocuration provides knowledgebases with detailed, high-quality data, the process is labour-intensive, and the increasing number of references, as well as the complexity of data within those references, has necessitated the development of semi-automated approaches to ensure that experimental findings are incorporated into knowledgebases in a timely manner. Additionally, recent funding cuts to manual biocuration projects emphasize the urgent need to focus on developing biocuration methods that are more efficient, yet still of uncompromising quality [4–6].

Two main approaches to this curation challenge adopted by biological knowledgebases are (i) artificial intelligence (AI) and machine learning (ML) methods and (ii) community curation. AI and ML methods have been successfully employed at several stages of the curation process, including identifying: (i) curation-relevant references, (ii) specific data types within those references, (iii) entities such as genes, alleles, and strains used in experiments, and (iv) sentences for fact extraction (for illustrative examples, see 7–17). In addition, some knowledgebases also rely on authors for curatorial assistance, as these subject matter experts can, with guidance from professional biocurators, provide high-quality curation of their published references [18–23].

The ACKnowledge (Author Curation to Knowledgebase) project combines the strengths of state-of-the-art ML approaches with author expertise in a community curation pipeline [20]. Initially implemented at WormBase [24], the pipeline automatically extracts biological entities such as genes and strains from the full text of recently published scientific articles and predicts the presence of nine curation-relevant data types such as gene expression and catalytic activity. Authors are then presented with a web form containing the extracted information where they can validate and submit data to WormBase.

While the current ACKnowledge system allows authors to contribute valuable information on entities and data types, our ultimate goal is to actively engage them in submitting more comprehensive curation, i.e. fact extraction statements such as ‘gene A is expressed in cell B’ or ‘protein C phosphorylates protein D’. To this end, we are furthering our approach by developing methods that combine ML with community curation to identify relevant text for fact extraction. The identified text will then be presented to authors to assist them in curation. While we are developing these methods to assist community curation, they will also be invaluable in professional biocuration workflows.

Relevant work in the literature on text classification includes the work of Shatkay et al. [25], who introduced a multi-dimensional framework for annotating biomedical sentences, incorporating scientific focus, generic nature, and statement polarity, which paved the way for further research in the field. Other methods have focused on classifying sentences in PubMed abstracts into broad semantic and rhetorical categories, such as Introduction/Background, Methods, Results, and Conclusions, which further aid in text mining tasks. These methods vary widely, including Naive Bayes models, support vector machines (SVMs), Hidden Markov models, Conditional Random Fields, and other methods based on more sophisticated feature engineering (e.g. see 26, 27).

Additional work in sentence identification was performed as part of the BioCreative IV Gene Ontology (GO) task, in which curators collected two types of evidence sentences, summary and experimental, to support GO annotations [28]. Using these evidence sentences as training data, participating teams attempted to locate GO-relevant sentences from full-text articles given relevant gene information in the text retrieval task [29]. The results of this task, however, indicated that much work still needed to be done to identify curation-relevant sentences, as the F1 scores for identifying exact or overlapping evidence sentences were only 0.270 and 0.387, respectively.

The BioCreative V BioC track [30] aimed to develop a collaborative biocurator assistant system for BioGRID using the BioC XML format to manage text, annotations, and relations. The system analysed full-text articles, prioritizing sentences reporting experimental methods and mentions of protein–protein and genetic interactions. BioGRID curators evaluated the system, giving positive feedback on its usability and design, though they suggested improving passage prediction accuracy, with F1 scores ranging from 0.65 to 0.8, depending on the task.

Recent advances in sentence and text classification have further enhanced the capabilities of biocuration. BERT (Bidirectional Encoder Representations from Transformers) is a state-of-the-art language representation model developed by Devlin et al. [31]. This model leverages a transformer architecture to pre-train on a large corpus of text, enabling it to capture deep bidirectional context. BERT proved to be a robust model for text classification, outperforming previous solutions. Models such as BioWordVec [32] and BioSentVec [33] are able to generate word and sentence embeddings specifically for biomedical texts and demonstrated considerable promise. Each of these models was trained on over 30 million documents from PubMed and clinical notes from the MIMIC-III Clinical Database, providing robust embeddings for word and sentence pair similarity tasks. These embeddings can be easily combined with more traditional ML models such as SVMs and logistic regression to perform sentence classification with lightweight models compared to those based on BERT and GPT. BioBERT (Bidirectional Encoder Representations from Transformers for Biomedical Text Mining), a domain-specific language representation model pre-trained on large-scale biomedical corpora, has outperformed BERT and previous state-of-the-art models in a range of biomedical text mining tasks, including named entity recognition (NER), relation extraction, question answering, and text classification [34]. More recently, specialized sentence classification models based on BERT have also been used to classify semantically related sentences describing key aspects of a paper’s content, e.g. Background, Problem, Results, etc. While this work showed that different types of sentences can successfully be classified, it did not address use in subsequent downstream tasks of identification and classification of sentences for the purpose of domain-specific curation such as that employed by professional biocurators who add structured data to knowledgebases [35].

Most recently, Generative Pre-trained Transformer (GPT) models [36] have demonstrated remarkable capabilities primarily in text generation but also in classification tasks. These models leverage extensive pretraining on diverse datasets, making them applicable for various NLP tasks, including sentence classification [37].

In this study, we focus on sentence-level classification for identifying curatable information from full-text articles. This approach simplifies the identification process by treating sentences as discrete units of text that convey a statement and are complete in themselves, rather than attempting to extract relevant information from the entire body of scientific articles. By focusing on individual sentences, we hope to more effectively isolate and analyse specific pieces of information, which may enhance the accuracy and efficiency of our curation efforts. This method also reduces the complexity associated with parsing and interpreting longer, more intricate texts, making it a more manageable and scalable solution for biocuration tasks. In addition, identifying curation-relevant sentences can facilitate manual validation of fact and entity extraction by highlighting the most critical information needed for curation.

To achieve our aims, we first performed a detailed characterization of the types of sentences relevant for curation of gene expression and protein kinase activity, two key data types curated at WormBase. We then grouped these sentences into classes for the purpose of performing biocuration tasks of increasing specificity. Using the identified classes, we trained ML and AI models to automatically classify relevant sentences. GPT and BioBERT models demonstrate superior average performance, with F1 scores ranging from 0.89 to 0.99 depending upon the task.

Our findings demonstrate the feasibility of extracting biocuration-relevant sentences from published scientific references. Integrating these sentence classifiers into the ACKnowledge, as well as professional biocuration platforms, will enhance community and professional biocuration efforts, by focusing work on high-value curation statements, thus maximizing curation efforts.

Methods

Full-text data collection and preprocessing

Source of data

The full text of references was obtained from the Alliance of Genome Resources (The Alliance), which now hosts the Caenorhabditis elegans literature corpus and serves as the primary data source for WormBase curation and ACKnowledge processing. While the Alliance infrastructure supports literature management for multiple model organisms, the present study focuses exclusively on C. elegans. As part of shared infrastructure, the Alliance has developed the Alliance Bibliographic Central (ABC), a platform for managing references, including full-text files, which are primarily stored as PDFs [38]. We accessed these PDFs using the ABC Application Programming Interface (API) and converted them to plain text with GROBID (https://github.com/kermitt2/grobid), a state-of-the-art ML-based library for parsing structured file formats and converting them to standard XML that is also able to identify and extract sentences. We applied XML manipulation to obtain the list of plain text sentences from each article, sourcing them from all sections except the cited references. The high variability in templates across different journals and articles makes it challenging to consistently identify specific sections beyond the cited references, which are always correctly removed by GROBID.

Sentence similarity visualization

We used visual techniques to explore the similarity between sentences and examine whether sentences from the same classes (as defined below) tended to group together. Specifically, we embedded the text sentences using BioSentVec and applied Uniform Manifold Approximation and Projection (UMAP, 39), a dimensionality reduction technique that enabled visualization of sentence representations in two dimensions.

Sentence datasets acquisition

Overview

To create a dataset for sentence classification, we collected both data type-relevant sentences and non-relevant, or ‘negative’, sentences. Negative sentences do not contain language pertinent to the data types of interest. We identified sentences via both random sampling and through manual selection from references in the WormBase bibliography. We included manually selected sentences because random sampling alone did not yield a sufficient number of data type-relevant sentences for our analysis.

Data type-relevant sentences either described experimental results for a given data type, described related information such as experimental details, or contained language related to the data type but not directly relevant to the curation of that reference. A more detailed description of the differences between data type-relevant sentences is provided below.

Random sampling

We collected a set of 1000 randomly selected sentences sourced from 15 manually validated positive and 15 manually validated negative references for the data types. These sentences were then manually reviewed and labelled by a curator according to the criteria outlined below.

Manually extracted data type-relevant sentences

We collected 589 and 494 data type-relevant sentences from 115 and 51 references, respectively, for gene expression and protein kinase activity (Supplementary Table 1). For each of the data type-relevant sentences, curators characterized the type and amount of information it contained (see the Results section for more examples) and collected relevant metadata, such as the section of the reference where the sentence appeared and whether it referred to a figure or a table. While some sentences could fall into more than one category, for the purposes of data analysis, each sentence was consistently classified into the single most representative data class. For example, a sentence that reported both previously published experimental results and novel experimental results was classified solely as ‘Directly reports experimental results’ [e.g.: In addition to the neuronal and intestinal expression previously detected using an acs-1 promoter-driven GFP expression construct (Kniazeva et al. 2004), we observed prominent GFP fluorescence in the somatic gonad of adults but not in that of larvae (Fig. 2A, B; Supplemental Fig. S2A–L, S, T)].All sentences were reviewed collectively by curators to ensure accurate characterization.

Additional randomly selected negative sentences

For each data type, we sourced 500 additional negative sentences by randomly selecting a set of sentences from references that were manually validated negative in our curation database. Curators manually verified that these sentences were indeed negative.

Overall dataset

After merging all the sentences obtained as described above, we obtained an overall dataset from which we removed duplicate sentences resulting from possible overlap between the different sets. The complete dataset is available in Supplementary Table 2 (in Excel format) and also in Comma Separated Value (CSV) format on the Hugging Face Hub at the Alliance Genome Resources account (https://huggingface.co/alliance-genome-account) for ease of reuse and integration with AI tools.

Sentence-level classifiers

We evaluated the performance of traditional supervised learning binary classifiers as well as transformer-based approaches such as BERT and large language models (LLMs) in classifying sentences. For the traditional classifiers, we tested several models, including Logistic Regression, Random Forests, Gradient Boosting, eXtreme Gradient Boosting (XGB), Multi-Layer Perceptron (MLP), SVMs, K-neighbours classifiers, Stochastic Gradient Descent (SGD), and Perceptron classifiers [40–43]. We converted plain text sentences into numeric vectors (sentence embedding, aka sent2vec) to be able to process them with ML methods. We used BioSentVec, a pretrained sent2vec model specifically designed for the biomedical literature [33], which embeds text sentences into 700-dimensional numerical vectors. Then, we fitted each model using stratified K-fold cross-validation with k = 5 and using hyperparameter optimization with randomized search on 100 different random configurations, for a total of 500 fits for each model. Both the hyperparameter optimization and best model selection were based on the F1 score. We calculated the average precision, recall, and F1 score for the best models. We used implementations available from the Python library scikit-learn (scikit-learn.org), one of the most widely used ML libraries in the literature, and from the xgboost library (xgboost.readthedocs.io). For each model type, we provided a range of possible hyperparameters to be selected by randomized search optimization. The code is available at https://github.com/WormBase/curation-sentence-classification.

For the transformer-based approaches, we used BioBERT [33], available from Hugging Face (huggingface.co; dmis-lab/biobert-v1.1), and GPT-4o [44]. For BioBERT, we used the same model for text embedding and for classification. We fine-tuned the classification model on our specific dataset and calculated its performance using the same stratified K-fold cross-validation technique used with the traditional ML models. The number of epochs for each fine-tuning phase during the K-fold cross-validation was set to 5. The code is available at https://github.com/WormBase/huggingface-document-classifier.

For GPT, there are many GPT models available, such as BioGPT, but we focused on the latest models developed by OpenAI to ensure we are arguably leveraging the latest advancements in the field. Two GPT-4o models (gpt-4o-2024-08-06) were fine-tuned separately for gene expression and protein kinase activity classification. Training, validation, and testing datasets were created using a stratified multilabel split, with 70% of the data allocated for training, 15% for validation, and 15% for testing. This stratification ensured balanced representation of all curatable labels across subsets. The datasets were then converted to JSONL format as per OpenAI’s documentation (https://github.com/alliance-genome/agr_sentence_classifier). Both models were trained over three epochs with a batch size of two and a learning rate multiplier of two. Testing of sentence datasets was conducted using custom Python code interfaced with OpenAI’s API, performing five test iterations per dataset. The evaluation workflow incorporated detailed prompts tailored to each data type for accurate classification (Supplementary File S1). The cost for fine-tuning a model was $69.34 for protein kinase activity and $73.17 for gene expression. Running the models against our datasets cost $3.85 for protein kinase activity and $4.12 for gene expression.

Results

Selection of data types

Our analysis focused on two data types curated at WormBase: gene expression and protein kinase activity. We started with these data types for several reasons. First, each of these data types is relatively straightforward for curation because they do not require extensive curatorial interpretation. This contrasts, for example, with genetic interaction data for which curators must sometimes interpret entire paragraphs covering multiple experiments. Secondly, sentences describing these data types can potentially be processed for automatic extraction of entities, e.g. genes, and ontology terms, such as those from the C. elegans Anatomy Ontology, to pre-populate curation forms. Third, we hoped to mitigate curation backlogs as exemplified by the relatively high frequency of flagging for those data types within the ACKnowledge form (e.g. as of November 2024, 336 references were flagged positive for gene expression since 2019, 20.25% of the total number of submissions). Finally, capturing the causal effects between protein kinases and their substrates is a major component of pathway modelling, and we wished to explore how sentence classification could be used to help identify experimental support for protein kinase activity annotations when creating Gene Ontology Causal Activity Models (GO-CAMs, 44).

Gene expression curation

‘Low throughput’ gene expression data curation aims at extracting from the literature evidence of gene expression in a particular tissue or cell component, or during specific stages of development, inferred from a small number of custom-crafted experiments rather than genome-scale analyses (‘High throughput’). Such gene expression experiments include reporter gene analysis, antibody staining, in situ hybridization (ISH), single-molecule fluorescent ISH, Reverse Transcription-Polymerase Chain Reaction (RT-PCR), quantitative Polymerase Chain Reaction (qPCR), northern blot, western blot analysis, and ISH chain reaction.

To make an annotation for gene expression in wild-type conditions, curators need to identify the gene product and its spatial or temporal localization. Ideally, a sentence from which a curator can make an annotation contains the gene name, a keyword indicating expression, the anatomical or cellular location of expression, and/or the relevant life stage. It is worth noting that often this information is scattered across multiple adjoining sentences. Curators also record the absence of expression in specific tissues or life stages when the authors explicitly indicate the absence of the expression signal.

Notably, sentences that contain the aforementioned information may also be negative for curation, such as when authors are reporting results from previously published studies or when they are describing expression in a mutant background.

Protein kinase activity curation

Protein kinases are enzymes that play crucial roles in regulation of biological pathways by phosphorylating specific amino acid residues to activate or inhibit the activities of their protein targets. One of the curatorial goals of WormBase and the GO Consortium is to model such pathways using the GO-CAM framework in which GO molecular functions (MFs) are linked to one another with causal relations, e.g. directly positively regulates [RO:0 002 629], from the Relations Ontology [45].

For protein kinase activity curation, we focused on sentences that report the results of biochemical experiments providing in vitro evidence for protein kinase activity and, whenever possible, the physiologically relevant substrate(s). Fully curatable experimental protein kinase activity sentences thus state the protein kinase and keywords (e.g. phosphorylation, in vitro) that describe the enzymatic activity and type of assay, respectively. Sentences that also state the protein kinase substrate provide additional contextual information that can be captured as part of a GO-CAM or as a standard GO annotation extension [46]. We note, however, that substrate identification is not absolutely required to make a GO MF annotation. As for gene expression curation, full information for a GO MF annotation may be spread across multiple sentences.

Characterization of data type-relevant sentences

To characterize data type-relevant sentences in depth, we analysed the set of manually collected gene expression and protein kinase activity sentences in Supplementary Table 1, which contains additional sentence metadata compared to the randomly selected sentences in our dataset (see the Methods section for more information on the sentence collection process). We performed this analysis because, from our curatorial experience, we have observed that there is a broad range of data type-relevant sentences in the literature, and we wanted to understand the differences in those sentences in the context of curation tasks. Our goal was to define classes that could, in the future, be used for automatic sentence classification and thus aid in fact extraction. Specifically, while some data type-relevant sentences concisely report experimental results in their entirety, it is also the case that experimental results are described across multiple, sometimes non-contiguous, sentences. Also, some sentences may contain data type-relevant language without reporting curatable information.

We therefore characterized sentences across two main axes: (i) the type of information they contained and (ii) the completeness of the information presented for the purpose of creating an annotation. For example, ‘fully curatable’ sentences contain mention of a gene or gene product, language that can be used to select an ontology term for an annotation (C. elegans Anatomy, Life Stage Ontologies and GO Cellular Component for gene expression, and a GO MF term for protein kinase activity) and, where applicable, language that allows a curator to correctly assign an evidence code for the annotation from the Evidence and Conclusion Ontology (e.g. ‘in vitro assay’) [47]. ‘Partially curatable’ sentences are missing one or more critical pieces of information, e.g. the sentence mentions expression in an anatomy term but not the actual gene expressed or mentions the kinase substrate, but not the actual kinase, requiring a curator to seek additional curation details. ‘Related language’ sentences may describe hypotheses or experimental design or contain similar language but not refer to a curatable experiment within that reference.

As shown in Table 1, our analyses led to the identification of three main classes of data type-relevant sentences (Column 1, main sentence class) with several subclasses (Column 2, specific subclass) represented within each main class.

Table 1.

Summary of data type-relevant sentences.

Main sentence classSpecific subclassProtein kinase sentences examplesGene expression sentences examples
Statements that directly report experimental resultsDirectly reports experimental results and is fully curatableWhen this activated PMK-1 was tested in the in vitro kinase assay, it was found to efficiently phosphorylate GST–SKN-1 (Fig. 3A, lane 2).A transcriptional reporter for nlp-18 revealed strong expression in the intestine and three pairs of neurons: a gustatory sensory neuron (ASI); a mechanical sensory neuron (FLP); and an interneuron (RIM; Figs. 2C and S2B).
Directly reports experimental results and is partially curatableWhen the immunoprecipitates were incubated with [γ-32P]ATP, SEK-1(K79R) became phosphorylated (Fig. 3B).We observed expression in five additional neurons close to the nerve ring, including possibly M3, AWB, and ASJ, and three neurons in the tail (Fig. 1E).
Summary statementsSummary statement of experimental results in reference and is fully curatableTaken together, our results demonstrate that UNC-43 is a calcium-dependent kinase capable of phosphorylating CaMKII substrates.In conclusion, gmap1 and gmap-2/3 genes are both expressed in epithelial cells although with non-overlapping expression patterns.
Summary statement of experimental results in reference and is partially curatableThe protein was shown to phosphorylate a model substrate and to undergo autophosphorylation.Expression pattern of the acox-1.1 gene.
Summary statement of previously published experimental resultsPhosphorylation of many proteins by GSK-3 first requires that a serine or threonine four residues C-terminal be phosphorylated by a priming kinase (Biondi and Nebreda, 2003; Fiol et al. 1987).abts-1 is expressed in HSNs from the L4 larval stage onward (Shinkai et al. 2018).
Other language relevant sentencesExperiment was performed; experimental hypothesis or rationaleTo investigate whether MBK-2 has a direct role in OMA-1 degradation, we began by determining whether MBK-2 can phosphorylate OMA-1 in vitro.Due to the structural and functional similarities between GPM6A, M6 and NMGP-1 and because proteolipid proteins are expressed in neurons, we hypothesized that NMGP-1 was expressed in the nervous system of the worm.
Experimental setupCatalytic activity of DKF-1 was quantified by measuring incorporation of 32P radioactivity from [-32P]ATP into Syntide-2 peptide substrate (Calbiochem).Dissected gonads of adult worms were prepared for RNA ISH and processed according to the protocol described in Lee et al. (2006) (Lee and Schedl 2006).
Related experimental resultsCritical residues for DYRK2 phosphorylation are shown in blue and red. *: the residue phosphorylated by DYRK2 in eIF2B and corresponding residues in OMA-1 and OMA-2.As reported above, we observed that the expression and subcellular localization of RAB-11 and SYN-4 in acs-1(RNAi)C17ISO embryos were not different from wild type.
Contains overlapping keywords with the curated data typeAn important question that remains is whether CDK-1 targets WRM-1 in a signal-dependent manner—e.g. by only phosphorylating a subpopulation of WRM-1 on the membrane proximal to the P2/EMS contact site.Expression of CKR-1 in AIB robustly rescued ckr-1 mutant animal's escape defects (Fig. 4C).
Main sentence classSpecific subclassProtein kinase sentences examplesGene expression sentences examples
Statements that directly report experimental resultsDirectly reports experimental results and is fully curatableWhen this activated PMK-1 was tested in the in vitro kinase assay, it was found to efficiently phosphorylate GST–SKN-1 (Fig. 3A, lane 2).A transcriptional reporter for nlp-18 revealed strong expression in the intestine and three pairs of neurons: a gustatory sensory neuron (ASI); a mechanical sensory neuron (FLP); and an interneuron (RIM; Figs. 2C and S2B).
Directly reports experimental results and is partially curatableWhen the immunoprecipitates were incubated with [γ-32P]ATP, SEK-1(K79R) became phosphorylated (Fig. 3B).We observed expression in five additional neurons close to the nerve ring, including possibly M3, AWB, and ASJ, and three neurons in the tail (Fig. 1E).
Summary statementsSummary statement of experimental results in reference and is fully curatableTaken together, our results demonstrate that UNC-43 is a calcium-dependent kinase capable of phosphorylating CaMKII substrates.In conclusion, gmap1 and gmap-2/3 genes are both expressed in epithelial cells although with non-overlapping expression patterns.
Summary statement of experimental results in reference and is partially curatableThe protein was shown to phosphorylate a model substrate and to undergo autophosphorylation.Expression pattern of the acox-1.1 gene.
Summary statement of previously published experimental resultsPhosphorylation of many proteins by GSK-3 first requires that a serine or threonine four residues C-terminal be phosphorylated by a priming kinase (Biondi and Nebreda, 2003; Fiol et al. 1987).abts-1 is expressed in HSNs from the L4 larval stage onward (Shinkai et al. 2018).
Other language relevant sentencesExperiment was performed; experimental hypothesis or rationaleTo investigate whether MBK-2 has a direct role in OMA-1 degradation, we began by determining whether MBK-2 can phosphorylate OMA-1 in vitro.Due to the structural and functional similarities between GPM6A, M6 and NMGP-1 and because proteolipid proteins are expressed in neurons, we hypothesized that NMGP-1 was expressed in the nervous system of the worm.
Experimental setupCatalytic activity of DKF-1 was quantified by measuring incorporation of 32P radioactivity from [-32P]ATP into Syntide-2 peptide substrate (Calbiochem).Dissected gonads of adult worms were prepared for RNA ISH and processed according to the protocol described in Lee et al. (2006) (Lee and Schedl 2006).
Related experimental resultsCritical residues for DYRK2 phosphorylation are shown in blue and red. *: the residue phosphorylated by DYRK2 in eIF2B and corresponding residues in OMA-1 and OMA-2.As reported above, we observed that the expression and subcellular localization of RAB-11 and SYN-4 in acs-1(RNAi)C17ISO embryos were not different from wild type.
Contains overlapping keywords with the curated data typeAn important question that remains is whether CDK-1 targets WRM-1 in a signal-dependent manner—e.g. by only phosphorylating a subpopulation of WRM-1 on the membrane proximal to the P2/EMS contact site.Expression of CKR-1 in AIB robustly rescued ckr-1 mutant animal's escape defects (Fig. 4C).

This table illustrates data type-relevant sentence classification as determined by expert biocurators. The three main relevant sentence classes are defined as (i) sentences that directly report experimental results, with either complete or partially curatable information; (ii) sentences that summarize experimental results, either complete or partial, including statements from reference section headings and/or figure titles; and (iii) sentences that contain language relevant to a data type, but not necessarily used for curation, such as use of reporter fusions to assess phenotypic outcomes for the gene expression data type. Examples of sentences within each category are provided, highlighting the nuances in reporting and summarizing experimental data. A full list of sentence classes with relative examples is presented in Supplementary Table 1.

Table 1.

Summary of data type-relevant sentences.

Main sentence classSpecific subclassProtein kinase sentences examplesGene expression sentences examples
Statements that directly report experimental resultsDirectly reports experimental results and is fully curatableWhen this activated PMK-1 was tested in the in vitro kinase assay, it was found to efficiently phosphorylate GST–SKN-1 (Fig. 3A, lane 2).A transcriptional reporter for nlp-18 revealed strong expression in the intestine and three pairs of neurons: a gustatory sensory neuron (ASI); a mechanical sensory neuron (FLP); and an interneuron (RIM; Figs. 2C and S2B).
Directly reports experimental results and is partially curatableWhen the immunoprecipitates were incubated with [γ-32P]ATP, SEK-1(K79R) became phosphorylated (Fig. 3B).We observed expression in five additional neurons close to the nerve ring, including possibly M3, AWB, and ASJ, and three neurons in the tail (Fig. 1E).
Summary statementsSummary statement of experimental results in reference and is fully curatableTaken together, our results demonstrate that UNC-43 is a calcium-dependent kinase capable of phosphorylating CaMKII substrates.In conclusion, gmap1 and gmap-2/3 genes are both expressed in epithelial cells although with non-overlapping expression patterns.
Summary statement of experimental results in reference and is partially curatableThe protein was shown to phosphorylate a model substrate and to undergo autophosphorylation.Expression pattern of the acox-1.1 gene.
Summary statement of previously published experimental resultsPhosphorylation of many proteins by GSK-3 first requires that a serine or threonine four residues C-terminal be phosphorylated by a priming kinase (Biondi and Nebreda, 2003; Fiol et al. 1987).abts-1 is expressed in HSNs from the L4 larval stage onward (Shinkai et al. 2018).
Other language relevant sentencesExperiment was performed; experimental hypothesis or rationaleTo investigate whether MBK-2 has a direct role in OMA-1 degradation, we began by determining whether MBK-2 can phosphorylate OMA-1 in vitro.Due to the structural and functional similarities between GPM6A, M6 and NMGP-1 and because proteolipid proteins are expressed in neurons, we hypothesized that NMGP-1 was expressed in the nervous system of the worm.
Experimental setupCatalytic activity of DKF-1 was quantified by measuring incorporation of 32P radioactivity from [-32P]ATP into Syntide-2 peptide substrate (Calbiochem).Dissected gonads of adult worms were prepared for RNA ISH and processed according to the protocol described in Lee et al. (2006) (Lee and Schedl 2006).
Related experimental resultsCritical residues for DYRK2 phosphorylation are shown in blue and red. *: the residue phosphorylated by DYRK2 in eIF2B and corresponding residues in OMA-1 and OMA-2.As reported above, we observed that the expression and subcellular localization of RAB-11 and SYN-4 in acs-1(RNAi)C17ISO embryos were not different from wild type.
Contains overlapping keywords with the curated data typeAn important question that remains is whether CDK-1 targets WRM-1 in a signal-dependent manner—e.g. by only phosphorylating a subpopulation of WRM-1 on the membrane proximal to the P2/EMS contact site.Expression of CKR-1 in AIB robustly rescued ckr-1 mutant animal's escape defects (Fig. 4C).
Main sentence classSpecific subclassProtein kinase sentences examplesGene expression sentences examples
Statements that directly report experimental resultsDirectly reports experimental results and is fully curatableWhen this activated PMK-1 was tested in the in vitro kinase assay, it was found to efficiently phosphorylate GST–SKN-1 (Fig. 3A, lane 2).A transcriptional reporter for nlp-18 revealed strong expression in the intestine and three pairs of neurons: a gustatory sensory neuron (ASI); a mechanical sensory neuron (FLP); and an interneuron (RIM; Figs. 2C and S2B).
Directly reports experimental results and is partially curatableWhen the immunoprecipitates were incubated with [γ-32P]ATP, SEK-1(K79R) became phosphorylated (Fig. 3B).We observed expression in five additional neurons close to the nerve ring, including possibly M3, AWB, and ASJ, and three neurons in the tail (Fig. 1E).
Summary statementsSummary statement of experimental results in reference and is fully curatableTaken together, our results demonstrate that UNC-43 is a calcium-dependent kinase capable of phosphorylating CaMKII substrates.In conclusion, gmap1 and gmap-2/3 genes are both expressed in epithelial cells although with non-overlapping expression patterns.
Summary statement of experimental results in reference and is partially curatableThe protein was shown to phosphorylate a model substrate and to undergo autophosphorylation.Expression pattern of the acox-1.1 gene.
Summary statement of previously published experimental resultsPhosphorylation of many proteins by GSK-3 first requires that a serine or threonine four residues C-terminal be phosphorylated by a priming kinase (Biondi and Nebreda, 2003; Fiol et al. 1987).abts-1 is expressed in HSNs from the L4 larval stage onward (Shinkai et al. 2018).
Other language relevant sentencesExperiment was performed; experimental hypothesis or rationaleTo investigate whether MBK-2 has a direct role in OMA-1 degradation, we began by determining whether MBK-2 can phosphorylate OMA-1 in vitro.Due to the structural and functional similarities between GPM6A, M6 and NMGP-1 and because proteolipid proteins are expressed in neurons, we hypothesized that NMGP-1 was expressed in the nervous system of the worm.
Experimental setupCatalytic activity of DKF-1 was quantified by measuring incorporation of 32P radioactivity from [-32P]ATP into Syntide-2 peptide substrate (Calbiochem).Dissected gonads of adult worms were prepared for RNA ISH and processed according to the protocol described in Lee et al. (2006) (Lee and Schedl 2006).
Related experimental resultsCritical residues for DYRK2 phosphorylation are shown in blue and red. *: the residue phosphorylated by DYRK2 in eIF2B and corresponding residues in OMA-1 and OMA-2.As reported above, we observed that the expression and subcellular localization of RAB-11 and SYN-4 in acs-1(RNAi)C17ISO embryos were not different from wild type.
Contains overlapping keywords with the curated data typeAn important question that remains is whether CDK-1 targets WRM-1 in a signal-dependent manner—e.g. by only phosphorylating a subpopulation of WRM-1 on the membrane proximal to the P2/EMS contact site.Expression of CKR-1 in AIB robustly rescued ckr-1 mutant animal's escape defects (Fig. 4C).

This table illustrates data type-relevant sentence classification as determined by expert biocurators. The three main relevant sentence classes are defined as (i) sentences that directly report experimental results, with either complete or partially curatable information; (ii) sentences that summarize experimental results, either complete or partial, including statements from reference section headings and/or figure titles; and (iii) sentences that contain language relevant to a data type, but not necessarily used for curation, such as use of reporter fusions to assess phenotypic outcomes for the gene expression data type. Examples of sentences within each category are provided, highlighting the nuances in reporting and summarizing experimental data. A full list of sentence classes with relative examples is presented in Supplementary Table 1.

Using the data type-relevant sentences in Supplementary Table 1, we performed a UMAP analysis on the sentence embeddings obtained with BioSentVec to see if: (i) the data type-relevant sentences were clearly separable from negative sentences that do not contain relevant language, and (ii) the sentences in the same data type-relevant sentence class are close to each other in the embedding space. For this visual analysis, we included 500 randomly selected negative sentences taken from 15 random references manually validated negative for each data type (see Methods section for more details). As shown in Fig. 1, the negative sentences (i.e. ‘does not contain language’) are separated from the data type-relevant sentences. This suggests that an automated classifier may be used to separate negative sentences from the data type-relevant classes. Resolution of individual data type-relevant sentence classes in the 2D map is not as distinct, but in some cases, e.g. the ‘Summary statements’ and the ‘Experimental setup statements’ for protein kinase activity, suggests that there are defining features in each class that could be exploited for more granular classification, given the much higher dimensionality of the original data compared to the 2D visualization obtained with UMAP. Current results demonstrate potential for further subclass separation with larger training sets.

UMAP visualization of sentence embeddings for gene expression (left panel) and protein kinase activity (right panel). A detailed description of the sentence classes is reported in Table 1.
Figure 1.

UMAP visualization of sentence embeddings for gene expression (left panel) and protein kinase activity (right panel). A detailed description of the sentence classes is reported in Table 1.

Curation tasks for sentence classification

Our sentence analysis illustrated that a range of sentence classes can be relevant for curation. To integrate this information into curatorial workflows, we focused on three classification tasks: (i) identifying fully curatable sentences, (ii) identifying partially curatable sentences, and (iii) identifying sentences with related language. Table 2 presents the relevant sentence classes for each classification task along with potential use cases or applications for sentences within each class. For example, the identification of fully curatable sentences could be used to automatically create annotations or pre-populate fields in curation forms. Partially curatable sentences could be used individually, or collectively, to pre-populate curation forms or serve as curatorial aids for creating a complete annotation. Related language sentences could be used to retrieve experimental or methodological details, or for alerting curators to related experiments that warrant further review of the reference for potential curation.

Table 2.

Task-centric sentence classification.

Datasets compositionSentences classTypes of sentences in this classPossible use cases
T1PT2PT3PClass A: fully curatableSentences containing all the relevant information for an annotation:
–Statements directly reporting experimental results and fully curatable
–Summary statements of experimental results in reference and fully curatable
–identify single sentences from which an annotation might be extracted automatically
–present ‘fully curatable’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
T1N T2P T3PClass B: partially curatableSentences containing partial relevant information for an annotation:
–Statements directly reporting experimental results and partially curatable
–Summary statements of experimental results in reference and partially curatable
–identify a set of sentences that, collectively, could be used to automatically extract an annotation
–present ‘curation relevant’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
 T1NT2N T3PClass C: related languageSentences containing related terms and phrases but not necessarily containing curatable information:
–Summary statement of previously published experimental results
–Experiment was performed; experimental hypothesis or rational
–Experimental setup
–Related experimental result
–Contains overlapping keywords with the curated data type
–identify additional sentences that may provide curation details (e.g. the antibody used for immunohistochemistry, the construct details for reporter gene analysis)
–identify any sentences related to that data type that may help guide curators or readers to potentially relevant work
 T1NT2NT3NNegativeSentences that do not contain data type-related language
Datasets compositionSentences classTypes of sentences in this classPossible use cases
T1PT2PT3PClass A: fully curatableSentences containing all the relevant information for an annotation:
–Statements directly reporting experimental results and fully curatable
–Summary statements of experimental results in reference and fully curatable
–identify single sentences from which an annotation might be extracted automatically
–present ‘fully curatable’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
T1N T2P T3PClass B: partially curatableSentences containing partial relevant information for an annotation:
–Statements directly reporting experimental results and partially curatable
–Summary statements of experimental results in reference and partially curatable
–identify a set of sentences that, collectively, could be used to automatically extract an annotation
–present ‘curation relevant’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
 T1NT2N T3PClass C: related languageSentences containing related terms and phrases but not necessarily containing curatable information:
–Summary statement of previously published experimental results
–Experiment was performed; experimental hypothesis or rational
–Experimental setup
–Related experimental result
–Contains overlapping keywords with the curated data type
–identify additional sentences that may provide curation details (e.g. the antibody used for immunohistochemistry, the construct details for reporter gene analysis)
–identify any sentences related to that data type that may help guide curators or readers to potentially relevant work
 T1NT2NT3NNegativeSentences that do not contain data type-related language

The values in Column 1 (‘Datasets composition’) indicate the task and what classes were used as positive (P) and negative (N) for that task (T). For example, T1P means Task 1 Positive. Fully curatable sentences supply all information and could be used in stand-alone curation. Partially curatable sentences would require additional text to make an annotation. Related language sentences provide hints at relevant experiments but would require additional text to confirm their presence and/or outcome.

Table 2.

Task-centric sentence classification.

Datasets compositionSentences classTypes of sentences in this classPossible use cases
T1PT2PT3PClass A: fully curatableSentences containing all the relevant information for an annotation:
–Statements directly reporting experimental results and fully curatable
–Summary statements of experimental results in reference and fully curatable
–identify single sentences from which an annotation might be extracted automatically
–present ‘fully curatable’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
T1N T2P T3PClass B: partially curatableSentences containing partial relevant information for an annotation:
–Statements directly reporting experimental results and partially curatable
–Summary statements of experimental results in reference and partially curatable
–identify a set of sentences that, collectively, could be used to automatically extract an annotation
–present ‘curation relevant’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
 T1NT2N T3PClass C: related languageSentences containing related terms and phrases but not necessarily containing curatable information:
–Summary statement of previously published experimental results
–Experiment was performed; experimental hypothesis or rational
–Experimental setup
–Related experimental result
–Contains overlapping keywords with the curated data type
–identify additional sentences that may provide curation details (e.g. the antibody used for immunohistochemistry, the construct details for reporter gene analysis)
–identify any sentences related to that data type that may help guide curators or readers to potentially relevant work
 T1NT2NT3NNegativeSentences that do not contain data type-related language
Datasets compositionSentences classTypes of sentences in this classPossible use cases
T1PT2PT3PClass A: fully curatableSentences containing all the relevant information for an annotation:
–Statements directly reporting experimental results and fully curatable
–Summary statements of experimental results in reference and fully curatable
–identify single sentences from which an annotation might be extracted automatically
–present ‘fully curatable’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
T1N T2P T3PClass B: partially curatableSentences containing partial relevant information for an annotation:
–Statements directly reporting experimental results and partially curatable
–Summary statements of experimental results in reference and partially curatable
–identify a set of sentences that, collectively, could be used to automatically extract an annotation
–present ‘curation relevant’ sentences to the authors in the ACKnowledge form and to biocurators in curation tools to aid them in creating an annotation
 T1NT2N T3PClass C: related languageSentences containing related terms and phrases but not necessarily containing curatable information:
–Summary statement of previously published experimental results
–Experiment was performed; experimental hypothesis or rational
–Experimental setup
–Related experimental result
–Contains overlapping keywords with the curated data type
–identify additional sentences that may provide curation details (e.g. the antibody used for immunohistochemistry, the construct details for reporter gene analysis)
–identify any sentences related to that data type that may help guide curators or readers to potentially relevant work
 T1NT2NT3NNegativeSentences that do not contain data type-related language

The values in Column 1 (‘Datasets composition’) indicate the task and what classes were used as positive (P) and negative (N) for that task (T). For example, T1P means Task 1 Positive. Fully curatable sentences supply all information and could be used in stand-alone curation. Partially curatable sentences would require additional text to make an annotation. Related language sentences provide hints at relevant experiments but would require additional text to confirm their presence and/or outcome.

Sentence classifiers—traditional ML classifiers and transformer-based approaches

To train and fine-tune automated sentence classifiers for the three tasks defined above, we used the complete dataset (Supplementary Table 2—‘Complete Sentence Datasets’), which includes data type-relevant sentences manually collected and randomly selected, as well as randomly selected negative sentences. The sentence count for each task in the dataset is summarized in Table 3. The number of sentences is similar for both data types. As we further discuss in the Discussion and Conclusion section, the order of magnitude of positive sentences is sufficient to obtain significant results for all tasks for both data types, but it may be on the low side for some of the evaluated models, especially for Tasks 1 and 2.

Table 3.

Number of positive sentences in the dataset for the three curation tasks for sentence classification.

Data typeTotal number of sentencesNumber of positive sentences
Task 1Task 2Task 3
Gene expression1986310442786
Protein kinase activity1932289379664
Data typeTotal number of sentencesNumber of positive sentences
Task 1Task 2Task 3
Gene expression1986310442786
Protein kinase activity1932289379664
Table 3.

Number of positive sentences in the dataset for the three curation tasks for sentence classification.

Data typeTotal number of sentencesNumber of positive sentences
Task 1Task 2Task 3
Gene expression1986310442786
Protein kinase activity1932289379664
Data typeTotal number of sentencesNumber of positive sentences
Task 1Task 2Task 3
Gene expression1986310442786
Protein kinase activity1932289379664

We evaluated both traditional ML models and transformer-based approaches such as GPT and BERT. For the traditional models, we determined the best-performing one based on the F1 score, as described in the Methods section. The average precision, recall, and F1 score, calculated on the five-folds from cross-validation (as described in the Methods section) for each evaluated model, data type, and task, are reported in Table 4, with the values of the best performing methods highlighted in bold. A bar chart representation of the average F1 score is depicted in Fig. 2. For traditional ML classifiers, only the best model, selected through F1 score-based model selection, is reported in Table 4.

Barcharts representing the average F1 score for the different models tested for the three curation tasks.
Figure 2.

Barcharts representing the average F1 score for the different models tested for the three curation tasks.

Table 4.

Average and standard deviation for precision, recall, and F1 score of the classifiers for gene expression and protein kinase activity for the three tasks.

Data typeModel typeAvg. precisionAvg. recallAvg. F1 score
Task 1
Gene expressionTraditional (SVM)0.718 ± 0.2580.9 ± 0.0310.762 ± 0.196
GPT-4o0.867 ± 0.0120.991 ± 0.0120.925 ± 0.009
BioBERT0.845 ± 0.0160.916 ± 0.0550.879 ± 0.028
Protein kinase activityTraditional (SVM)0.688 ± 0.2540.878 ± 0.0900.726 ± 0.178
GPT-4o0.837 ± 0.0080.959 ± 0.010.894 ± 0.007
BioBERT0.865 ± 0.0780.903 ± 0.0440.880 ± 0.033
Task 2
Gene expressionTraditional (MLP)0.844 ± 0.1950.843 ± 0.0870.822 ± 0.107
GPT-4o0.951 ± 0.0070.933 ± 0.0080.942 ± 0.007
BioBERT0.888 ± 0.0360.939 ± 0.0340.912 ± 0.025
Protein kinase activityTraditional (Logistic)0.703 ± 0.2210.915 ± 0.0400.770 ± 0.152
GPT-4o0.886 ± 0.0060.966 ± 0.000.924 ± 0.003
BioBERT0.889 ± 0.0370.908 ± 0.0270.897 ± 0.016
Task 3
Gene expressionTraditional (SVM)0.943 ± 0.0310.855 ± 0.1470.888 ± 0.083
GPT-4o0.966 ± 0.000.949 ± 0.000.957 ± 0.00
BioBERT0.948 ± 0.0160.940 ± 0.0250.944 ± 0.011
Protein kinase activityTraditional (SVM)0.966 ± 0.0220.941 ± 0.0550.952 ± 0.024
GPT-4o0.96 ± 0.0081 ± 0.000.98 ± 0.004
BioBERT0.982 ± 0.0100.991 ± 0.0090.987 ± 0.008
Data typeModel typeAvg. precisionAvg. recallAvg. F1 score
Task 1
Gene expressionTraditional (SVM)0.718 ± 0.2580.9 ± 0.0310.762 ± 0.196
GPT-4o0.867 ± 0.0120.991 ± 0.0120.925 ± 0.009
BioBERT0.845 ± 0.0160.916 ± 0.0550.879 ± 0.028
Protein kinase activityTraditional (SVM)0.688 ± 0.2540.878 ± 0.0900.726 ± 0.178
GPT-4o0.837 ± 0.0080.959 ± 0.010.894 ± 0.007
BioBERT0.865 ± 0.0780.903 ± 0.0440.880 ± 0.033
Task 2
Gene expressionTraditional (MLP)0.844 ± 0.1950.843 ± 0.0870.822 ± 0.107
GPT-4o0.951 ± 0.0070.933 ± 0.0080.942 ± 0.007
BioBERT0.888 ± 0.0360.939 ± 0.0340.912 ± 0.025
Protein kinase activityTraditional (Logistic)0.703 ± 0.2210.915 ± 0.0400.770 ± 0.152
GPT-4o0.886 ± 0.0060.966 ± 0.000.924 ± 0.003
BioBERT0.889 ± 0.0370.908 ± 0.0270.897 ± 0.016
Task 3
Gene expressionTraditional (SVM)0.943 ± 0.0310.855 ± 0.1470.888 ± 0.083
GPT-4o0.966 ± 0.000.949 ± 0.000.957 ± 0.00
BioBERT0.948 ± 0.0160.940 ± 0.0250.944 ± 0.011
Protein kinase activityTraditional (SVM)0.966 ± 0.0220.941 ± 0.0550.952 ± 0.024
GPT-4o0.96 ± 0.0081 ± 0.000.98 ± 0.004
BioBERT0.982 ± 0.0100.991 ± 0.0090.987 ± 0.008
Table 4.

Average and standard deviation for precision, recall, and F1 score of the classifiers for gene expression and protein kinase activity for the three tasks.

Data typeModel typeAvg. precisionAvg. recallAvg. F1 score
Task 1
Gene expressionTraditional (SVM)0.718 ± 0.2580.9 ± 0.0310.762 ± 0.196
GPT-4o0.867 ± 0.0120.991 ± 0.0120.925 ± 0.009
BioBERT0.845 ± 0.0160.916 ± 0.0550.879 ± 0.028
Protein kinase activityTraditional (SVM)0.688 ± 0.2540.878 ± 0.0900.726 ± 0.178
GPT-4o0.837 ± 0.0080.959 ± 0.010.894 ± 0.007
BioBERT0.865 ± 0.0780.903 ± 0.0440.880 ± 0.033
Task 2
Gene expressionTraditional (MLP)0.844 ± 0.1950.843 ± 0.0870.822 ± 0.107
GPT-4o0.951 ± 0.0070.933 ± 0.0080.942 ± 0.007
BioBERT0.888 ± 0.0360.939 ± 0.0340.912 ± 0.025
Protein kinase activityTraditional (Logistic)0.703 ± 0.2210.915 ± 0.0400.770 ± 0.152
GPT-4o0.886 ± 0.0060.966 ± 0.000.924 ± 0.003
BioBERT0.889 ± 0.0370.908 ± 0.0270.897 ± 0.016
Task 3
Gene expressionTraditional (SVM)0.943 ± 0.0310.855 ± 0.1470.888 ± 0.083
GPT-4o0.966 ± 0.000.949 ± 0.000.957 ± 0.00
BioBERT0.948 ± 0.0160.940 ± 0.0250.944 ± 0.011
Protein kinase activityTraditional (SVM)0.966 ± 0.0220.941 ± 0.0550.952 ± 0.024
GPT-4o0.96 ± 0.0081 ± 0.000.98 ± 0.004
BioBERT0.982 ± 0.0100.991 ± 0.0090.987 ± 0.008
Data typeModel typeAvg. precisionAvg. recallAvg. F1 score
Task 1
Gene expressionTraditional (SVM)0.718 ± 0.2580.9 ± 0.0310.762 ± 0.196
GPT-4o0.867 ± 0.0120.991 ± 0.0120.925 ± 0.009
BioBERT0.845 ± 0.0160.916 ± 0.0550.879 ± 0.028
Protein kinase activityTraditional (SVM)0.688 ± 0.2540.878 ± 0.0900.726 ± 0.178
GPT-4o0.837 ± 0.0080.959 ± 0.010.894 ± 0.007
BioBERT0.865 ± 0.0780.903 ± 0.0440.880 ± 0.033
Task 2
Gene expressionTraditional (MLP)0.844 ± 0.1950.843 ± 0.0870.822 ± 0.107
GPT-4o0.951 ± 0.0070.933 ± 0.0080.942 ± 0.007
BioBERT0.888 ± 0.0360.939 ± 0.0340.912 ± 0.025
Protein kinase activityTraditional (Logistic)0.703 ± 0.2210.915 ± 0.0400.770 ± 0.152
GPT-4o0.886 ± 0.0060.966 ± 0.000.924 ± 0.003
BioBERT0.889 ± 0.0370.908 ± 0.0270.897 ± 0.016
Task 3
Gene expressionTraditional (SVM)0.943 ± 0.0310.855 ± 0.1470.888 ± 0.083
GPT-4o0.966 ± 0.000.949 ± 0.000.957 ± 0.00
BioBERT0.948 ± 0.0160.940 ± 0.0250.944 ± 0.011
Protein kinase activityTraditional (SVM)0.966 ± 0.0220.941 ± 0.0550.952 ± 0.024
GPT-4o0.96 ± 0.0081 ± 0.000.98 ± 0.004
BioBERT0.982 ± 0.0100.991 ± 0.0090.987 ± 0.008

Across all tasks, GPT-4o and BioBERT showed strong performances, while traditional ML methods generally underperformed relative to these more advanced architectures. It is worth noting that the statistical variation of traditional ML models is rather high, at least for the first two tasks. This observation indicates that the datasets may not be large enough for these tasks. Nonetheless, the current results already give an indication of their performances. The low standard deviation of GPT and BioBERT shows that these models have stable performances on different subsets of the dataset, and this is an indication that the amount of sentences used for fine-tuning is sufficient.

Finding the right size for fine-tuning datasets: degradation analysis

We performed a detailed analysis to assess the minimum amount of fine-tuning data required to achieve reliable performance while minimizing the time investment of curators. This analysis is relevant, e.g. to add another data type to the classification analysis, or expand the classification for that data type to other organisms. Given that collecting and curating these data demands significant curator time, our goal was to evaluate the performance drop-off at each reduction level. By doing so, we aimed to identify the minimum amount of fine-tuning data necessary for reliable performance without significantly sacrificing accuracy, thereby optimizing the trade-off between curator effort and classifier effectiveness. Given the better performances provided by BioBERT compared to traditional ML models, and the fact that it is currently less expensive to fine-tune than OpenAI GPT models, we decided to take BioBERT as the reference model for this analysis.

To perform the analysis, we started with the full dataset (100%) and progressively reduced the size in 10% increments, down to 10% (Fig. 3). This approach allowed us to monitor how precision, recall, and F1 score were impacted as the available data decreased.

Comparison of average F1 scores across different dataset sizes for the three curation tasks: Task 1 (fully curatable), Task 2 (partially curatable), and Task 3 (language related) for the gene expression and protein kinase activity datasets.
Figure 3.

Comparison of average F1 scores across different dataset sizes for the three curation tasks: Task 1 (fully curatable), Task 2 (partially curatable), and Task 3 (language related) for the gene expression and protein kinase activity datasets.

To identify fully curatable sentences, using around 40% of our original dataset for protein kinase activity is sufficient for optimal performance. For gene expression, even though the graph does not clearly show a plateau, the gain in F1 score is marginal after 70% of the dataset. Partially curatable sentences require even less data, as low as 10% for protein kinase activity, and around 50% of the dataset for gene expression. The classification models seem to be able to correctly identify language-related sentences with a small set of fine-tuning sentences for both data types, but provide the best performances after about 30%–40% of our dataset. We note that even though the standard deviation is large with a small sample size, it reduces as the number of sentences increases.

These results suggest that while a significant portion of our dataset is needed to achieve high performance, there is a point of diminishing returns where additional data does not significantly improve the model. These thresholds can help guide curators in deciding the minimum amount of fine-tuning data necessary for effective model training while balancing the time and resources required for data curation.

Finally, to better understand the effort involved in generating fine-tuning data, we evaluated three new papers each for gene expression and protein kinase activity. On average, a professional biocurator was able to extract ~120 high-quality training sentences in 1 h (Table 5). This indicates that assembling sufficient fine-tuning data is a feasible and practical task, further supporting the utility of the degradation analysis in guiding efficient curation efforts.

Table 5.

Shown are the results of curator timing exercises for collecting curation-relevant sentences from three different publications for the associated data types.

Data typePMIDNumber of sentence collectedTotal time to collect sentences (min)Rate of sentence collection (sentences/min)
Gene expression38986620102551.9
Gene expression3955047189521.7
Gene expression3967143670561.3
Protein kinase activity39962268128582.2
Protein kinase activity4030539076332.3
Protein kinase activity4049755762252.5
Data typePMIDNumber of sentence collectedTotal time to collect sentences (min)Rate of sentence collection (sentences/min)
Gene expression38986620102551.9
Gene expression3955047189521.7
Gene expression3967143670561.3
Protein kinase activity39962268128582.2
Protein kinase activity4030539076332.3
Protein kinase activity4049755762252.5
Table 5.

Shown are the results of curator timing exercises for collecting curation-relevant sentences from three different publications for the associated data types.

Data typePMIDNumber of sentence collectedTotal time to collect sentences (min)Rate of sentence collection (sentences/min)
Gene expression38986620102551.9
Gene expression3955047189521.7
Gene expression3967143670561.3
Protein kinase activity39962268128582.2
Protein kinase activity4030539076332.3
Protein kinase activity4049755762252.5
Data typePMIDNumber of sentence collectedTotal time to collect sentences (min)Rate of sentence collection (sentences/min)
Gene expression38986620102551.9
Gene expression3955047189521.7
Gene expression3967143670561.3
Protein kinase activity39962268128582.2
Protein kinase activity4030539076332.3
Protein kinase activity4049755762252.5

Inter-annotator agreement analysis

To assess the reliability and consistency of our sentence classification approach, we conducted an inter-annotator agreement analysis. Two experienced curators independently classified sentences from four papers representing the two data types under study: protein kinase activity (PMIDs 40272473 and 40410380) and gene expression (PMIDs 39671436 and 39550471). We calculated the % agreement between annotators for the individual curation tasks (T1P, T2P, and T3P, see Table 2) using the Jaccard coefficient (Table 6). The % agreement averaged on all papers for both data types is 90% for T1P, 82% for T2P, and 88% for T3P. The highest percentage agreement for T1P is consistent with the stricter criteria for classifying a sentence as fully curatable, as compared to partially curatable and related language sentences, where broader criteria can be subject to curator’s interpretation. The lower percentage of agreement for T2Ps was somewhat unexpected, but we note that variability within individual papers with regard to the totality of experiments performed can make it more or less difficult to distinguish subclasses, particularly class B (partially curatable) from class C (related language) (see Table 2), which accounted for the majority of interannotator differences in T2P classification.

Table 6.

Inter-annotator agreement for sentence classification across curation tasks and papers.

Data typePMIDT1P agreementT2P agreementT3P agreement
Protein kinase activity4027247380%68%88%
Protein kinase activity4041038094%86%94%
Gene expression3967143695%88%80%
Gene expression3955047190%87%88%
Average percentages90%82%88%
Data typePMIDT1P agreementT2P agreementT3P agreement
Protein kinase activity4027247380%68%88%
Protein kinase activity4041038094%86%94%
Gene expression3967143695%88%80%
Gene expression3955047190%87%88%
Average percentages90%82%88%

Agreement percentages are shown for each paper (identified by PMID) within the two curation tasks: protein kinase activity and gene expression.

Table 6.

Inter-annotator agreement for sentence classification across curation tasks and papers.

Data typePMIDT1P agreementT2P agreementT3P agreement
Protein kinase activity4027247380%68%88%
Protein kinase activity4041038094%86%94%
Gene expression3967143695%88%80%
Gene expression3955047190%87%88%
Average percentages90%82%88%
Data typePMIDT1P agreementT2P agreementT3P agreement
Protein kinase activity4027247380%68%88%
Protein kinase activity4041038094%86%94%
Gene expression3967143695%88%80%
Gene expression3955047190%87%88%
Average percentages90%82%88%

Agreement percentages are shown for each paper (identified by PMID) within the two curation tasks: protein kinase activity and gene expression.

Overall, these results demonstrate excellent inter-annotator reliability across our three-tier classification system. The strong agreement levels validate that our classification criteria are sufficiently well defined to enable consistent annotation between curators, supporting the robustness of our approach for training automated sentence classification models and its potential applicability to other curation contexts.

Integrating sentence classification into professional and community curation workflows

To maximize the impact of our sentence classification work, we are actively integrating the output into both professional and community curation workflows. At present, extracted sentences identified as fully and partially curatable are presented to professional curators on the ACKnowledge dashboard, allowing them to rapidly assess and annotate relevant data (Fig. 4, panel A). As a next step, we plan to extend this functionality to community curation by embedding the extracted sentences directly into the ACKnowledge submission form. This will enable authors to view and confirm sentence-level evidence supporting the presence of specific data types in their manuscripts at the time of submission (Fig. 4, panel B). In addition, we plan on presenting the sentences on a dedicated PDF viewer that highlights curatable content directly within the context of the full-text article (Fig. 4, panel C). This will allow both authors and curators to easily locate and review relevant statements in situ, streamlining the validation process and reducing the need to cross-reference between external tools and source documents. We anticipate that this workflow will empower authors to contribute more effectively by verifying key statements, thereby enhancing data quality and reducing manual effort across the curation lifecycle. Further, this functionality could be added to professional biocurator tools to support more efficient and accurate literature triage and annotation.

Current and planned integration of sentence classification output into curation workflows. (A) Extracted sentences identified as fully or partially curatable are presented to professional curators via a dedicated interface, allowing for rapid evaluation and annotation of key data types. (B) Future integration into the ACKnowledge submission form will present curatable sentences to authors, enabling direct validation of sentence-level evidence during manuscript submission, and supporting more comprehensive community curation. (C) A dedicated PDF viewer will highlight relevant sentences within the full text of the article, enabling authors and curators to efficiently locate, review, and validate statements without cross-referencing external tools. The paper displayed is Meng et al., PMID: 32668210.
Figure 4.

Current and planned integration of sentence classification output into curation workflows. (A) Extracted sentences identified as fully or partially curatable are presented to professional curators via a dedicated interface, allowing for rapid evaluation and annotation of key data types. (B) Future integration into the ACKnowledge submission form will present curatable sentences to authors, enabling direct validation of sentence-level evidence during manuscript submission, and supporting more comprehensive community curation. (C) A dedicated PDF viewer will highlight relevant sentences within the full text of the article, enabling authors and curators to efficiently locate, review, and validate statements without cross-referencing external tools. The paper displayed is Meng et al., PMID: 32668210.

Discussion and conclusions

Manual curation of the biomedical literature provides detailed, high-quality annotations for biological knowledgebases and the scientific communities that rely on them daily for research support. However, this process is labour-intensive, and the current trend of decreased funding for knowledgebases makes it imperative that manual biocuration is augmented with high-performing ML methods.

In this work, we explored, for the purpose of developing automated or semi-automated curation pipelines, the manual biocuration process for two key data types, gene expression and protein kinase activity, curated by WormBase and other knowledgebases. To this end, we collected a variety of sentences used by biocurators to support detailed annotation and characterized the amount and type of information they contained. This characterization allowed us to define three sentence classification tasks for the identification of (i) fully curatable sentences; (ii) partially curatable sentences; and (iii) language-related sentences.

We evaluated multiple ML models to automate the classification of sentences according to these defined tasks. Among them, GPT and BioBERT models show the best average performances with very low standard deviation, even with a limited number of sentences used for fine-tuning.

For implementation into biocuration workflows, we will consider the relative merits of each approach. Traditional ML classifiers, while showing lower performances, still achieved medium to high scores, indicating their potential utility in various curatorial tasks, especially when optimizing resource use and supporting real-time applications is critical.

BERT models tend to be slower in both training and prediction phases compared to traditional ML models, often requiring Graphics Processing Units for efficient computation. This can become a limiting factor, especially for real-time applications or when working on very large datasets.

Large GPT models can be expensive to train and deploy, often requiring significant computational resources or high cost through proprietary APIs. On the other hand, they are very flexible models that can be used for a wide range of tasks in various applications.

Our analysis highlights the value of having well-defined tasks and high-quality datasets as a baseline for training domain-specific models, such as BioBERT, or fine-tuning LLMs. All in all, our results indicate that it is possible to accurately classify sentences according to curatorial need, and suggest that implementing these classifiers into existing biocuration workflows has the potential to be of significant benefit to biocuration.

Limitations

Our current datasets, consisting of ~2000 sentences each, might be on the lower end of the recommended dataset size for traditional ML models. A common rule of thumb suggests having at least ten times as many data points as there are features in the dataset [48]. Given that our sentence embeddings generated with BioWordVec result in around 700 dimensions per embedding vector, we would ideally need at least 7000 sentences to achieve significant results. The current dataset size might therefore limit the model’s ability to generalize well and could lead to overfitting. This is reflected in the higher standard deviation in the precision, recall, and F-score for traditional ML models in Table 3, especially for Tasks 1 and 2, which have less positive sentences. On the other hand, GPT and BERT models do not necessarily need large datasets for fine-tuning, due to the pre-trained nature of the models.

In the current study, we focused on analysing two specific data types: gene expression and protein kinase activity, given their importance in WormBase curation. However, we recognize the importance of expanding our analysis to encompass a broader range of data types, some of which may not be as straightforward for curation as gene expression and protein kinase activity. Therefore, we are committed to extending our research to include additional data types in future studies, and our degradation analysis can serve as a guide for the minimum training sample size.

Future work to enhance the extraction of curatable information from scientific articles will focus on developing hybrid models that combine sentence-level classification with contextual information from full-text analysis, e.g. the location of the sentence within the article, or the content of surrounding sentences. Such models could leverage the simplicity and manageability of sentence classifiers while incorporating the contextual richness of full-text analysis. We will also leverage NLP techniques such as NER to improve classification accuracy. The rapid improvements of LLMs also hold promise for high-quality fact extraction, and we will explore using them for suggesting annotations.

By expanding the scope of our analysis and incorporating these advanced techniques, we aim to achieve a more comprehensive and accurate extraction of curatable information from scientific literature.

As mentioned before, we will also explore ways to more rapidly create sentence datasets. One possible approach is to facilitate sentence identification and classification as part of the curatorial process. To this end, the new curation tools being developed by the Alliance of Genome Resources provide an opportunity to integrate this functionality from the outset into a curatorial framework that will be used by multiple model organism curators. Given the high recall shown by GPT models, another approach is to employ these AI models, combined with prompt engineering, to more rapidly identify and characterize curation-relevant sentences for new data types. Lastly, we will keep abreast of developments in other ML-enhanced curation resources such as PubTator 3 [49] and iTextMine [50] that also identify supporting text for annotations. At present, our curation pipelines require faster processing of full text and more specific relation extraction than are currently available through these resources, but we will monitor them for future enhancements that could help to augment and expand on the work described here. For all approaches, we hope to systematically test the lower limits of training data to obtain the highest performance with minimum manual effort.

Lastly, to maximize the utility of the classifiers, we will integrate their output into existing curation tools such as the ACKnowledge community curation and professional biocuration platforms to aid in fact extraction.

In conclusion, our study demonstrates that sentence classification using advanced ML models, such as BioBERT and GPT, has the potential to enhance the efficiency and accuracy of biocuration. By leveraging these insights, we aim to streamline the curation process and improve the quality of annotations in biomedical knowledgebases, ultimately supporting the scientific community more effectively.

Acknowledgements

We thank Chris Grove, Ranjana Kishore, and Pengyuan Li for critical feedback on this work. Their insightful comments and constructive suggestions greatly improved the quality and clarity of this publication.

Author contributions

Daniela Raciti—Conceptualization, Data curation, Funding acquisition, Investigation, Validation, Writing—original draft, Writing—review & editing.

Kimberly Van Auken—Conceptualization, Data curation, Funding acquisition, Investigation, Validation, Writing—original draft, Writing—review & editing.

Valerio Arnaboldi—Conceptualization, Formal Analysis, Funding acquisition, Investigation, Methodology, Software, Writing—original draft, Writing—review & editing.

Christopher J. Tabone—Formal Analysis, Methodology, Software, Writing—original draft, Writing—review & editing.

Hans-Michael Muller—Formal Analysis, Investigation, Methodology, Software, Writing—review & editing.

Paul W. Sternberg—Funding acquisition, Supervision, Writing—review & editing.

Conflict of interest

The authors declare there are no competing interests.

Funding

The ACKnowledge project is funded by RO1 OD023041 from the National Library of Medicine of the National Institutes of Health. The Alliance of Genome Resources is funded by U24HG010859 from the National Human Genome Research Institute of the National Institutes of Health and the National Heart, Lung and Blood Institute of the National Institutes of Health. WormBase and FlyBase are funded by the National Human Genome Research Institute of the National Institutes of Health by U24HG002223 and U41HG000739, respectively. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Data availability

The data underlying this article are available in Hugging Face at https://huggingface.co/alliance-genome-account. The code is available in GitHub at https://github.com/WormBase/curation-sentence-classification, https://github.com/WormBase/huggingface-document-classifier, and https://github.com/alliance-genome/agr_sentence_classifier.

References

1.

Wood
 
V
,
Sternberg
 
PW
,
Lipshitz
 
HD
.
Making biological knowledge useful for humans and machines
.
Genetics
.
2022
;
220
:
iyac001
.

2.

Wilkinson
 
MD
,
Dumontier
 
M
,
Aalbersberg
 
IJ
 et al.  
The FAIR Guiding Principles for scientific data management and stewardship
.
Sci Data
.
2016
;
3
:
160018
.

3.

International Society for Biocuration
.
Biocuration: distilling data into knowledge
.
PLoS Biol
.
2018
;
16
:
e2002846
.

4.

Hayden
 
EC
.
Concern over funding cuts for model organism databases
.
Nature
.
2016
.

5.

Kaiser
 
J
.
BIOMEDICAL RESOURCES. Funding for key data resources in jeopardy
.
Science
.
2016
;
351
:
14
.

6.

Alliance of Genome Resources Consortium
.
The Alliance of Genome Resources: building a modern data ecosystem for model organism databases
.
Genetics
.
2019
;
213
:
1189
96
.

7.

Van Auken
 
K
,
Jaffery
 
J
,
Chan
 
J
 et al.  
Semi-automated curation of protein subcellular localization: a text mining-based approach to Gene Ontology (GO) Cellular Component curation
.
BMC Bioinf
.
2009
;
10
:
228
.

8.

Fang
 
R
,
Schindelman
 
G
,
Van Auken
 
K
 et al.  
Automatic categorization of diverse experimental information in the bioscience literature
.
BMC Bioinf
.
2012
;
13
:
16
.

9.

Liu
 
W
,
Laulederkind
 
SJ
,
Hayman
 
GT
 et al.  
OntoMate: a text-mining tool aiding curation at the Rat Genome Database
.
Database
.
2015
;
2015
:
bau129
.

10.

Torii
 
M
,
Arighi
 
CN
,
Li
 
G
 et al.  
RLIMS-P 2.0: a generalizable rule-based information extraction system for literature mining of protein phosphorylation information
.
IEEE/ACM Trans Comput Biol Bioinf
.
2015
;
12
:
17
29
.

11.

Tudor
 
CO
,
Ross
 
KE
,
Li
 
G
 et al.  
Construction of phosphorylation interaction networks by text mining of full-length articles using the eFIP system
.
Database
.
2015
;
2015
:
bav020
.

12.

König
 
C
,
Shaim
 
I
,
Vellido
 
A
 et al.  
Using machine learning tools for protein database biocuration assistance
.
Sci Rep
.
2018
;
8
:
10148
.

13.

Nenni
 
MJ
,
Fisher
 
ME
,
James-Zorn
 
C
 et al.  
Xenbase: facilitating the use of Xenopus to model human disease
.
Front Physiol
.
2019
;
10
:
154
.

14.

Teodoro
 
D
,
Knafou
 
J
,
Naderi
 
N
 et al.  
UPCLASS: a deep learning-based classifier for UniProtKB entry publications
.
Database
.
2020
;
2020
:
baaa026
.

15.

Karimi
 
K
,
Agalakov
 
S
,
Telmer
 
CA
 et al.  
Classifying domain-specific text documents containing ambiguous keywords
.
Database
.
2021
;
2021
:
baab062
.

16.

Li
 
P
,
Jiang
 
X
,
Zhang
 
G
 et al.  
Utilizing image and caption information for biomedical document classification
.
Bioinformatics
.
2021
;
37
:
i468
76
.

17.

Ringwald
 
M
,
Richardson
 
JE
,
Baldarelli
 
RM
 et al.  
Mouse Genome Informatics (MGI): latest news from MGD and GXD
.
Mamm Genome
.
2022
;
33
:
4
18
.

18.

Berardini
 
TZ
,
Li
 
D
,
Muller
 
R
 et al.  
Assessment of community-submitted ontology annotations from a novel database-journal partnership
.
Database
.
2012
;
2012
:
bas030
.

19.

Bunt
 
SM
,
Grumbling
 
GB
,
Field
 
HI
 et al.  
Directly e-mailing authors of newly published papers encourages community curation
.
Database
.
2012
;
2012
:
bas024
.

20.

Arnaboldi
 
V
,
Raciti
 
D
,
Van Auken
 
K
 et al.  
Text mining meets community curation: a newly designed curation platform to improve author experience and participation at WormBase
.
Database
.
2020
;
2020
:
baaa006
.

21.

Lock
 
A
,
Harris
 
MA
,
Rutherford
 
K
 et al.  
Community curation in PomBase: enabling fission yeast experts to provide detailed, standardized, sharable annotation from research publications
.
Database
.
2020
;
2020
:
baaa028
.

22.

Larkin
 
A
,
Marygold
 
SJ
,
Antonazzo
 
G
 et al.  
FlyBase: updates to the Drosophila melanogaster knowledge base
.
Nucleic Acids Res
.
2021
;
49
:
D899
907
.

23.

Wang
 
Y
,
Wang
 
Q
,
Huang
 
H
 et al.  
A crowdsourcing open platform for literature curation in UniProt
.
PLoS Biol
.
2021
;
19
:
e3001464
.

24.

Sternberg
 
PW
,
Van Auken
 
K
,
Wang
 
Q
 et al.  
WormBase 2024: status and transitioning to Alliance infrastructure
.
Genetics
.
2024
;
227
:
iyae050
.

25.

Shatkay
 
H
,
Pan
 
F
,
Rzhetsky
 
A
 et al.  
Multi-dimensional classification of biomedical text: toward automated, practical provision of high-utility text to diverse users
.
Bioinformatics
.
2008
;
24
:
2086
93
.

26.

Agarwal
 
S
,
Yu
 
H
.
Automatically classifying sentences in full-text biomedical articles into Introduction, Methods, Results and Discussion
.
Bioinformatics
.
2009
;
25
:
3174
80
.

27.

Agibetov
 
A
,
Blagec
 
K
,
Xu
 
H
 et al.  
Fast and scalable neural embedding models for biomedical sentence classification
.
BMC Bioinf
.
2018
;
19
:
541
.

28.

Van Auken
 
K
,
Schaeffer
 
ML
,
McQuilton
 
P
 et al.  
BC4GO: a full-text corpus for the BioCreative IV GO task
.
Database
.
2014
;
2014
:
bau074
.

29.

Mao
 
Y
,
Van Auken
 
K
,
Li
 
D
 et al.  
Overview of the gene ontology task at BioCreative IV
.
Database
.
2014
;
2014
:
bau086
.

30.

Kim
 
S
,
Islamaj Doğan
 
R
,
Chatr-Aryamontri
 
A
 et al.  
BioCreative V BioC track overview: collaborative biocurator assistant task for BioGRID
.
Database
.
2016
;
2016
:
baw121
.

31.

Devlin
 
J
,
Chang
 
MW
,
Lee
 
K
 et al.  
BERT: pre-training of deep bidirectional transformers for language understanding
.
Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
 
Minneapolis, Minnesota.
 
Association for Computational Linguistics
 
4171
4186
.
1
:
2019
.

32.

Zhang
 
Y
,
Chen
 
Q
,
Yang
 
Z
 et al.  
BioWordVec, improving biomedical word embeddings with subword information and MeSH
.
Sci Data
.
2019
;
6
:
52
.

33.

Chen
 
Q
,
Peng
 
Y
,
Lu
 
Z
.
BioSentVec: creating sentence embeddings for biomedical texts
. In:
The 7th IEEE International Conference on Healthcare Informatics
.
Xi'an
:
IEEE
,
2019
.

34.

Lee
 
J
,
Yoon
 
W
,
Kim
 
S
 et al.  
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
.
Bioinformatics
.
2020
;
36
:
1234
40
.

35.

Brack
 
A
,
Entrup
 
E
,
Stamatakis
 
M
 et al.  
Sequential sentence classification in research papers using cross-domain multi-task learning
.
Int J Digit Libr
.
2024
;
25
:
377
400
.

36.

Radford
 
A
,
Narasimhan
 
K
,
Salimans
 
T
 et al.  
Improving language understanding by generative pre-training
.
OpenAI
.
2018
. https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf  
(17 December 2024, date last accessed).

37.

Brown
 
TB
,
Mann
 
B
,
Ryder
 
N
 et al.  
Language models are few-shot learners
. In:
Proceedings of the 34th International Conference on Neural Information Processing Systems
.
1877
901
.
Red Hook, NY
:
Curran Associates Inc
,
2020
. https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf

38.

Alliance of Genome Resources Consortium
.
Updates to the Alliance of Genome Resources central infrastructure
.
Genetics
.
2024
;
227
:
iyae049
.

39.

McInnes
 
L
,
Healy
 
J
,
Melville
 
J
.
UMAP: Uniform Manifold Approximation and Projection
.
J Open Source Software
.
2018
;
3
:
861
. .

40.

Mohri
 
M
,
Rostamizadeh
 
A
,
Talwalkar
 
A
.
Foundations of Machine Learning
.
Cambridge, MA
:
MIT Press
,
2018
.

41.

Kowsari
 
K
,
Meimandi
 
KJ
,
Heidarysafa
 
M
 et al.  
Text classification algorithms: a survey
.
Information
.
2019
;
10
:
150
.

42.

Chen
 
T
,
Guestrin
 
C
.
XGBoost: a scalable tree boosting system
. In:
Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
.
785
94
.
New York, NY
:
ACM
,
2016
.

43.

Bottou
 
L
.
Large-scale machine learning with stochastic gradient descent
. In
Proceedings of COMPSTAT'2010
.
177
86
.
Heidelberg
:
Physica-Verlag HD
,
2010
.

44.

Open
 
AI
,
Achiam
 
J
,
Adler
 
S
 et al.  
GPT-4 technical report
.
arXiv
.
2024
.

45.

Thomas
 
PD
,
Hill
 
DP
,
Mi
 
H
 et al.  
Gene ontology causal activity modeling (GO-CAM) moves beyond GO annotations to structured descriptions of biological functions and systems
.
Nat Genet
.
2019
;
51
:
1429
33
.

46.

Huntley
 
RP
,
Harris
 
MA
,
Alam-Faruque
 
Y
 et al.  
A method for increasing expressivity of Gene Ontology annotations using a compositional approach
.
BMC Bioinf
.
2014
;
15
:
155
.

47.

Nadendla
 
S
,
Jackson
 
R
,
Munro
 
J
 et al.  
ECO: the evidence and conclusion ontology, an update for 2022
.
Nucleic Acids Res
.
2022
;
50
:
D1515
21
.

48.

Alwosheel
 
A
,
Van Cranenburgh
 
S
,
Chorus Caspar
 
G
.
Is your dataset big enough? Sample size requirements when using artificial neural networks for discrete choice analysis
.
J Choice Model
.
2018
;
28
:
28167
182
.

49.

Wei
 
CH
,
Allot
 
A
,
Lai
 
PT
 et al.  
PubTator 3.0: an AI-powered literature resource for unlocking biomedical knowledge
.
Nucleic Acids Res
.
2024
;
52
:
W540
46
.

50.

Ren
 
J
,
Li
 
G
,
Ross
 
K
 et al.  
iTextMine: integrated text-mining system for large-scale knowledge extraction from the literature
.
Database
.
2018
;
2018
:
bay128
.

Author notes

Daniela Raciti, Kimberly M. Van Auken and Valerio Arnaboldi Contributed equally.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.

Supplementary data