Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 318
Filtrar
1.
Heliyon ; 10(16): e35865, 2024 Aug 30.
Artículo en Inglés | MEDLINE | ID: mdl-39220956

RESUMEN

The digital era has expanded social exposure with easy internet access for mobile users, allowing for global communication. Now, people can get to know what is going on around the globe with just a click; however, this has also resulted in the issue of fake news. Fake news is content that pretends to be true but is actually false and is disseminated to defraud. Fake news poses a threat to harmony, politics, the economy, and public opinion. As a result, bogus news detection has become an emerging research domain to identify a given piece of text as genuine or fraudulent. In this paper, a new framework called Generative Bidirectional Encoder Representations from Transformers (GBERT) is proposed that leverages a combination of Generative pre-trained transformer (GPT) and Bidirectional Encoder Representations from Transformers (BERT) and addresses the fake news classification problem. This framework combines the best features of both cutting-edge techniques-BERT's deep contextual understanding and the generative capabilities of GPT-to create a comprehensive representation of a given text. Both GPT and BERT are fine-tuned on two real-world benchmark corpora and have attained 95.30 % accuracy, 95.13 % precision, 97.35 % sensitivity, and a 96.23 % F1 score. The statistical test results indicate the effectiveness of the fine-tuned framework for fake news detection and suggest that it can be a promising approach for eradicating this global issue of fake news in the digital landscape.

2.
Data Brief ; 56: 110855, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39286413

RESUMEN

With the soaring demand for healthcare systems, chatbots are gaining tremendous popularity and research attention. Numerous language-centric research on healthcare is conducted day by day. Despite significant advances in Arabic Natural Language Processing (NLP), challenges remain in natural language classification and generation due to the lack of suitable datasets. The primary shortcoming of these models is the lack of suitable Arabic datasets for training. To address this, authors introduce a large Arabic Healthcare Dataset (AHD) of textual data. The dataset consists of over 808k questions and answers across 90 categories, offered to the research community for Arabic computational linguistics. Authors anticipate that this rich dataset would make a great aid for a variety of NLP tasks on Arabic textual data, especially for text classification and generation purposes. Authors present the data in raw form. AHD is composed of main dataset scraped from medical website, which is Altibbi website. AHD is made public and freely available at http://data.mendeley.com/datasets/mgj29ndgrk/5.

3.
Data Brief ; 56: 110849, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39286414

RESUMEN

Our study aims to collect data to understand ideological and extreme bias in text articles shared across various online communities, particularly focusing on the language used in subreddits associated with extremism and targeted violence. Initially, we gathered data from related online communities, specifically the r/Liberal and r/Conservative communities on Reddit, utilizing the Reddit Pushshift API to collect URLs shared within these subreddits. Our aim was to gather news, opinion, and feature articles, resulting in a corpus of 226,010 articles. We also curated a balanced subset of 45,108 articles and annotated 4000 articles to validate their relevance, facilitating understanding of language usage within ideological Reddit communities and insights into ideological bias in media content. Expanding beyond binary ideologies, we introduced a new category termed "Restricted" to encompass articles shared in private or banned subreddits. This third category encompasses articles shared in restricted, privatized, quarantined, or banned subreddits characterized by radicalized and extremist ideologies. This expansion yielded a large dataset of 377,144 articles. Additionally, we included articles from subreddits with unspecified ideologies, creating a holdout set of 922,522 articles. In total, our combined dataset of 1.3 million articles collected from 55 different subreddits will assist in examining radicalized communities and providing discourse analysis in associated subreddits, enhancing understanding of the language used in articles shared within radicalized Reddit communities and offering insights into extreme bias in media content. In summary, we collected 1.52 million articles to understand ideological and extreme bias, providing a comprehensive dataset that aids in understanding language usage within text articles posted in ideological and extreme Reddit communities.

4.
Sci Rep ; 14(1): 21805, 2024 Sep 18.
Artículo en Inglés | MEDLINE | ID: mdl-39294183

RESUMEN

Accurate identification and classification of equipment defects are essential for assessing the health of power equipment and making informed maintenance decisions. Traditional defect classification methods, which rely on subjective manual records and intricate defect descriptions, have proven to be inefficient. Existing approaches often evaluate equipment status solely based on defect grade. To enhance the precision of power equipment defect identification, we developed a multi-label classification dataset by compiling historical defect records. Furthermore, we assessed the performance of 11 established multi-label classification methods on this dataset, encompassing both traditional machine learning and deep learning methods. Experimental results reveal that methods considering label correlations exhibit significant performance advantages. By employing balanced loss functions, we effectively address the challenge of sample imbalance across various categories, thereby enhancing classification accuracy. Additionally, segmenting the power equipment defect classification task, which involves numerous labels, into label recall and ranking stages can substantially improve classification performance. The dataset we have created is available for further research by other scholars in the field.

5.
Heliyon ; 10(17): e36861, 2024 Sep 15.
Artículo en Inglés | MEDLINE | ID: mdl-39296200

RESUMEN

Text classification involves annotating text data with specific labels and is a crucial research task in the field of natural language processing. Chinese text classification presents significant challenges due to the complex semantics of the language, difficulties in semantic feature extraction, and the interleaving and irregularity of lexical features. Traditional methods often struggle to manage the relationships between words and sentences in Chinese, hindering the model's ability to capture deep semantic information and resulting in poor classification performance. To address these issues, a Chinese text classification method based on utterance information enhancement and feature fusion is proposed. This method first embeds the text into a unified space and obtains feature representations of word vectors and sentence vectors using the BERT (Bidirectional Encoder Representations from Transformers) pre-trained language model. Subsequently, an utterance information enhancement module is constructed to perform syntactic enhancement and feature extraction on the sentence information within the text. Additionally, a feature fusion strategy is introduced to combine the enhanced sentence-level information features with the word-level features extracted by the Bi-GRU (Bidirectional Gated Recurrent Unit network), culminating in the classification output. This approach effectively enhances the feature representation of Chinese text and significantly filters out irrelevant and noisy information. Evaluations on several Chinese datasets demonstrate that the proposed method surpasses existing mainstream classification models in terms of classification accuracy and F1 value, validating its effectiveness and feasibility.

6.
J Appl Stat ; 51(13): 2592-2626, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39290353

RESUMEN

Optimizing text preprocessing and text classification algorithms is an important, everyday task in large organizations and companies and it usually involves a labor-intensive and time-consuming effort. For example, the filtering and sorting of a large number of electronic mails (emails) are crucial to keeping track of the received information and converting it automatically into useful and profitable knowledge. Business emails are often unstructured, noisy, and with many abbreviations and acronyms, which makes their handling a challenging procedure. To overcome those challenges, a two-step classification approach is proposed, along with a two-cycle labeling procedure in order to speed up the labeling process. Every step incorporates a heuristic classification approach to assign emails to predefined classes by comparing several classification and text vectorization algorithms. These algorithms are compared and evaluated using the F1 score and balanced accuracy. The implementation of the proposed algorithm is demonstrated in a shipbroker agent operating in Greece with excellent performance, improving organization and administration while reducing expenses.

7.
Stud Health Technol Inform ; 317: 210-217, 2024 Aug 30.
Artículo en Inglés | MEDLINE | ID: mdl-39234724

RESUMEN

INTRODUCTION: Human and veterinary medicine are practiced separately, but literature databases such as Pubmed include articles from both fields. This impedes supporting clinical decisions with automated information retrieval, because treatment considerations would not ignore the discipline of mixed sources. Here we investigate data-driven methods from computational linguistics for automatically distinguishing between human and veterinary medical texts. METHODS: For our experiments, we selected language models after a literature review of benchmark datasets and reported performances. We generated a dataset of around 48,000 samples for binary text classification, specifically designed to differentiate between human medical and veterinary subjects. Using this dataset, we trained and fine-tuned classifiers based on selected transformer-based models as well as support vector machines (SVM). RESULTS: All trained classifiers achieved more than 99% accuracy, even though the transformer-based classifiers moderately outperformed the SVM-based one. DISCUSSION: Such classifiers could be applicable in clinical decision support functions that build on automated information retrieval.


Asunto(s)
Procesamiento de Lenguaje Natural , Máquina de Vectores de Soporte , Humanos , Medicina Veterinaria , Almacenamiento y Recuperación de la Información/métodos , Animales
8.
Front Artif Intell ; 7: 1363531, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39109323

RESUMEN

Deep learning models have achieved state-of-the-art performance for text classification in the last two decades. However, this has come at the expense of models becoming less understandable, limiting their application scope in high-stakes domains. The increased interest in explainability has resulted in many proposed forms of explanation. Nevertheless, recent studies have shown that rationales, or language explanations, are more intuitive and human-understandable, especially for non-technical stakeholders. This survey provides an overview of the progress the community has achieved thus far in rationalization approaches for text classification. We first describe and compare techniques for producing extractive and abstractive rationales. Next, we present various rationale-annotated data sets that facilitate the training and evaluation of rationalization models. Then, we detail proxy-based and human-grounded metrics to evaluate machine-generated rationales. Finally, we outline current challenges and encourage directions for future work.

9.
Stud Health Technol Inform ; 316: 1780-1784, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176562

RESUMEN

Radiology reports contain crucial patient information, in addition to images, that can be automatically extracted for secondary uses such as clinical support and research for diagnosis. We tested several classifiers to classify 1,218 breast MRI reports in French from two Swiss clinical centers. Logistic regression performed better for both internal (accuracy > 0.95 and macro-F1 > 0.86) and external data (accuracy > 0.81 and macro-F1 > 0.41). Automating this task will facilitate efficient extraction of targeted clinical parameters and provide a good basis for future annotation processes through automatic pre-annotation.


Asunto(s)
Neoplasias de la Mama , Imagen por Resonancia Magnética , Humanos , Femenino , Neoplasias de la Mama/diagnóstico por imagen , Francia , Sistemas de Información Radiológica , Registros Electrónicos de Salud , Procesamiento de Lenguaje Natural , Suiza , Minería de Datos
10.
Stud Health Technol Inform ; 316: 374-375, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176755

RESUMEN

There is a rapid growth in the volume of data in the cancer field and fine-grained classification is in high demand especially for interdisciplinary and collaborative research. There is thus a need to establish a multi-label classifier with higher resolution to reduce the burden of screening articles for clinical relevance. This research trains a multi-label classifier with scalability for classifying literature on cancer research directly at the publication level. Firstly, a corpus was divided into a training set and a testing set at a ratio of 7:3. Secondly, we compared the performance of classifiers developed by "PubMedBERT + TextRNN" and "BioBERT + TextRNN" with ICRP CT. Finally, the classifier was obtained based on the optimal combination "PubMedBERT + TextRNN", with P= 0.952014, R=0.936696, F1=0.931664. The quantitative comparisons demonstrate that the resulting classifier is fit for high-resolution classification of cancer literature at the publication level to support accurate retrieving and academic statistics.


Asunto(s)
Neoplasias , Neoplasias/clasificación , Humanos , PubMed , Minería de Datos/métodos
11.
Stud Health Technol Inform ; 316: 560-564, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176804

RESUMEN

The goal of this paper is to build an automatic way to interpret conclusions from brain molecular imaging reports performed for investigation of cognitive disturbances (FDG, Amyloid and Tau PET) by comparing several traditional machine learning (ML) techniques-based text classification methods. Two purposes are defined: to identify positive or negative results in all three modalities, and to extract diagnostic impressions for Alzheimer's Disease (AD), Fronto-Temporal Dementia (FTD), Lewy Bodies Dementia (LBD) based on metabolism of perfusion patterns. A dataset was created by manual parallel annotation of 1668 conclusions of reports from the Nuclear Medicine and Molecular Imaging Division of Geneva University Hospitals. The 6 Machine Learning (ML) algorithms (Support Vector Machine (Linear and Radial Basis function), Naive Bayes, Logistic Regression, Random Forrest, and K-Nearest Neighbors) were trained and evaluated with a 5-fold cross-validation scheme to assess their performance and generalizability. The best classifier was SVM showing the following accuracies: FDG (0.97), Tau (0.94), Amyloid (0.98), Oriented Diagnostic (0.87 for a diagnosis among AD, FTD, LBD, undetermined, other), paving the way for a paradigm shift in the field of data handling in nuclear medicine research.


Asunto(s)
Disfunción Cognitiva , Tomografía de Emisión de Positrones , Humanos , Disfunción Cognitiva/diagnóstico por imagen , Disfunción Cognitiva/clasificación , Encéfalo/diagnóstico por imagen , Aprendizaje Automático , Enfermedad de Alzheimer/diagnóstico por imagen , Enfermedad de Alzheimer/clasificación , Procesamiento de Lenguaje Natural , Máquina de Vectores de Soporte , Sensibilidad y Especificidad , Suiza , Reproducibilidad de los Resultados
12.
Stud Health Technol Inform ; 316: 1979-1983, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176881

RESUMEN

Electronic health data concerning implantable medical devices (IMD) opens opportunities for dynamic real-world monitoring to assess associated risks related to implanted materials. Due to population ageing and expanding demands, total hip, knee, and shoulder arthroplasties are increasing. Automating the collection and analysis of orthopedic device features could benefit physicians and public health policies enabling early issue detection, IMD monitoring and patient safety assessment. A machine learning tool using natural language processing (NLP) was developed for the automated extraction of operation information from medical reports in orthopedics. A corpus of 959 orthopaedic operative reports from 5 centres was manually annotated using the Prodigy software® with a strong inter-annotator agreement of 0.80. Data to extract concerned key clinical and procedure information (n= 9) selected by a multidisciplinary group based on the French health authority checklist. Performances parameters of the NLP model estimated an overall strong precision and recall of respectively 97.0 and 96.0 with a F1-score 96.3. Systematic monitoring of orthopedic devices could be ensured by an automated tool, leveraging clinical data warehouses. Traceability of medical devices with implantation modalities will allow detection of implant factors leading to complications. The evidence from real-world data could provide concrete and dynamic insights to surgeons and infectious disease specialists concerning implant follow-up, guiding therapeutic decision-making, and informing public health policymakers. The tool will be applied on clinical data warehouses to automate information extraction and presentation, providing feedback on mandatory information completion and contents of operative reports to support improvements, and thereafter implant research projects.


Asunto(s)
Registros Electrónicos de Salud , Aprendizaje Automático , Procesamiento de Lenguaje Natural , Francia , Humanos , Procedimientos Ortopédicos
13.
Stud Health Technol Inform ; 316: 846-850, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176925

RESUMEN

Text classification plays an essential role in the medical domain by organizing and categorizing vast amounts of textual data through machine learning (ML) and deep learning (DL). The adoption of Artificial Intelligence (AI) technologies in healthcare has raised concerns about the interpretability of AI models, often perceived as "black boxes." Explainable AI (XAI) techniques aim to mitigate this issue by elucidating AI model decision-making process. In this paper, we present a scoping review exploring the application of different XAI techniques in medical text classification, identifying two main types: model-specific and model-agnostic methods. Despite some positive feedback from developers, formal evaluations with medical end users of these techniques remain limited. The review highlights the necessity for further research in XAI to enhance trust and transparency in AI-driven decision-making processes in healthcare.


Asunto(s)
Inteligencia Artificial , Procesamiento de Lenguaje Natural , Humanos , Aprendizaje Automático , Registros Electrónicos de Salud/clasificación , Aprendizaje Profundo
14.
Artículo en Inglés | MEDLINE | ID: mdl-39121174

RESUMEN

OBJECTIVES: Large language models (LLMs) have demonstrated remarkable success in natural language processing (NLP) tasks. This study aimed to evaluate their performances on social media-based health-related text classification tasks. MATERIALS AND METHODS: We benchmarked 1 Support Vector Machine (SVM), 3 supervised pretrained language models (PLMs), and 2 LLMs-based classifiers across 6 text classification tasks. We developed 3 approaches for leveraging LLMs: employing LLMs as zero-shot classifiers, using LLMs as data annotators, and utilizing LLMs with few-shot examples for data augmentation. RESULTS: Across all tasks, the mean (SD) F1 score differences for RoBERTa, BERTweet, and SocBERT trained on human-annotated data were 0.24 (±0.10), 0.25 (±0.11), and 0.23 (±0.11), respectively, compared to those trained on the data annotated using GPT3.5, and were 0.16 (±0.07), 0.16 (±0.08), and 0.14 (±0.08) using GPT4, respectively. The GPT3.5 and GPT4 zero-shot classifiers outperformed SVMs in a single task and in 5 out of 6 tasks, respectively. When leveraging LLMs for data augmentation, the RoBERTa models trained on GPT4-augmented data demonstrated superior or comparable performance compared to those trained on human-annotated data alone. DISCUSSION: The results revealed that using LLM-annotated data only for training supervised classification models was ineffective. However, employing the LLM as a zero-shot classifier exhibited the potential to outperform traditional SVM models and achieved a higher recall than the advanced transformer-based model RoBERTa. Additionally, our results indicated that utilizing GPT3.5 for data augmentation could potentially harm model performance. In contrast, data augmentation with GPT4 demonstrated improved model performances, showcasing the potential of LLMs in reducing the need for extensive training data. CONCLUSIONS: By leveraging the data augmentation strategy, we can harness the power of LLMs to develop smaller, more effective domain-specific NLP models. Using LLM-annotated data without human guidance for training lightweight supervised classification models is an ineffective strategy. However, LLM, as a zero-shot classifier, shows promise in excluding false negatives and potentially reducing the human effort required for data annotation.

15.
J Med Internet Res ; 26: e50236, 2024 Aug 01.
Artículo en Inglés | MEDLINE | ID: mdl-39088259

RESUMEN

BACKGROUND: Patients increasingly rely on web-based physician reviews to choose a physician and share their experiences. However, the unstructured text of these written reviews presents a challenge for researchers seeking to make inferences about patients' judgments. Methods previously used to identify patient judgments within reviews, such as hand-coding and dictionary-based approaches, have posed limitations to sample size and classification accuracy. Advanced natural language processing methods can help overcome these limitations and promote further analysis of physician reviews on these popular platforms. OBJECTIVE: This study aims to train, test, and validate an advanced natural language processing algorithm for classifying the presence and valence of 2 dimensions of patient judgments in web-based physician reviews: interpersonal manner and technical competence. METHODS: We sampled 345,053 reviews for 167,150 physicians across the United States from Healthgrades.com, a commercial web-based physician rating and review website. We hand-coded 2000 written reviews and used those reviews to train and test a transformer classification algorithm called the Robustly Optimized BERT (Bidirectional Encoder Representations from Transformers) Pretraining Approach (RoBERTa). The 2 fine-tuned models coded the reviews for the presence and positive or negative valence of patients' interpersonal manner or technical competence judgments of their physicians. We evaluated the performance of the 2 models against 200 hand-coded reviews and validated the models using the full sample of 345,053 RoBERTa-coded reviews. RESULTS: The interpersonal manner model was 90% accurate with precision of 0.89, recall of 0.90, and weighted F1-score of 0.89. The technical competence model was 90% accurate with precision of 0.91, recall of 0.90, and weighted F1-score of 0.90. Positive-valence judgments were associated with higher review star ratings whereas negative-valence judgments were associated with lower star ratings. Analysis of the data by review rating and physician gender corresponded with findings in prior literature. CONCLUSIONS: Our 2 classification models coded interpersonal manner and technical competence judgments with high precision, recall, and accuracy. These models were validated using review star ratings and results from previous research. RoBERTa can accurately classify unstructured, web-based review text at scale. Future work could explore the use of this algorithm with other textual data, such as social media posts and electronic health records.


Asunto(s)
Algoritmos , Internet , Procesamiento de Lenguaje Natural , Humanos , Femenino , Masculino , Médicos , Relaciones Médico-Paciente , Juicio , Adulto , Persona de Mediana Edad
16.
J Affect Disord ; 366: 445-458, 2024 Dec 01.
Artículo en Inglés | MEDLINE | ID: mdl-39214375

RESUMEN

Emotions are integral to human social interactions, with diverse responses elicited by various situational contexts. Particularly, the prevalence of negative emotional states has been correlated with negative outcomes for mental health, necessitating a comprehensive analysis of their occurrence and impact on individuals. In this paper, we introduce a novel dataset named DepressionEmo designed to detect 8 emotions associated with depression by 6037 examples of long Reddit user posts. This dataset was created through a majority vote over inputs by zero-shot classifications from pre-trained models and validating the quality by annotators and ChatGPT, exhibiting an acceptable level of inter-rater reliability between annotators. The correlation between emotions, and linguistic analysis are conducted on DepressionEmo. Besides, we provide several text classification methods classified into two groups: machine learning methods such as SVM, XGBoost, and LightGBM; and deep learning methods such as BERT, BART, GAN-BERT, and T5. Despite achieving the same F1 Macro score of 0.76 as BART, the pretrained BERT model, bert-base-uncased, stands out as the most efficient model in our experiments due to its lower number of parameters. Across all emotions, the highest F1 Macro value is achieved by suicide intent, indicating a certain value of our dataset in identifying emotions in individuals with depression symptoms through text analysis. The curated dataset is publicly available at: https://github.com/abuBakarSiddiqurRahman/DepressionEmo.


Asunto(s)
Depresión , Emociones , Aprendizaje Automático , Humanos , Depresión/clasificación , Depresión/psicología , Reproducibilidad de los Resultados , Aprendizaje Profundo , Conjuntos de Datos como Asunto
17.
Sci Rep ; 14(1): 19614, 2024 Aug 23.
Artículo en Inglés | MEDLINE | ID: mdl-39179733

RESUMEN

Text classification plays a major role in research such as sentiment analysis, opinion mining, and customer feedback analysis. Text classification using hypergraph algorithms is effective in capturing the intricate relationships between words and phrases in documents. The method entails text preprocessing, keyword extraction, feature selection, text classification, and performance metric evaluation. Here, we proposed a Hypergraph Attention Layer with Logistic Regression (HGATT_LR) for text classification in the Amazon review data set. The essential keywords are extracted by utilizing the Latent Dirichlet Allocation (LDA) technique. To build a hypergraph attention layer, feature selection based on node-level and edge-level attention is assessed. The resultant features are passed as an input of Logistic regression for text classification. Through a comparison analysis of different text classifiers on the Amazon data set, the performance metrics are assessed. Text classification using hypergraph Attention Network has been shown to achieve 88% accuracy which is better compared to other state-of-the-art algorithms. The proposed model is scalable and may be easily enhanced with more training data. The solution highlights the utility of hypergraph approaches for text classification as well as their applicability to real-world datasets with complicated interactions between text parts. This type of analysis will empower the business people will improve the quality of the product.

18.
JMIR AI ; 3: e56932, 2024 Aug 06.
Artículo en Inglés | MEDLINE | ID: mdl-39106099

RESUMEN

BACKGROUND: Despite their growing use in health care, pretrained language models (PLMs) often lack clinical relevance due to insufficient domain expertise and poor interpretability. A key strategy to overcome these challenges is integrating external knowledge into PLMs, enhancing their adaptability and clinical usefulness. Current biomedical knowledge graphs like UMLS (Unified Medical Language System), SNOMED CT (Systematized Medical Nomenclature for Medicine-Clinical Terminology), and HPO (Human Phenotype Ontology), while comprehensive, fail to effectively connect general biomedical knowledge with physician insights. There is an equally important need for a model that integrates diverse knowledge in a way that is both unified and compartmentalized. This approach not only addresses the heterogeneous nature of domain knowledge but also recognizes the unique data and knowledge repositories of individual health care institutions, necessitating careful and respectful management of proprietary information. OBJECTIVE: This study aimed to enhance the clinical relevance and interpretability of PLMs by integrating external knowledge in a manner that respects the diversity and proprietary nature of health care data. We hypothesize that domain knowledge, when captured and distributed as stand-alone modules, can be effectively reintegrated into PLMs to significantly improve their adaptability and utility in clinical settings. METHODS: We demonstrate that through adapters, small and lightweight neural networks that enable the integration of extra information without full model fine-tuning, we can inject diverse sources of external domain knowledge into language models and improve the overall performance with an increased level of interpretability. As a practical application of this methodology, we introduce a novel task, structured as a case study, that endeavors to capture physician knowledge in assigning cardiovascular diagnoses from clinical narratives, where we extract diagnosis-comment pairs from electronic health records (EHRs) and cast the problem as text classification. RESULTS: The study demonstrates that integrating domain knowledge into PLMs significantly improves their performance. While improvements with ClinicalBERT are more modest, likely due to its pretraining on clinical texts, BERT (bidirectional encoder representations from transformer) equipped with knowledge adapters surprisingly matches or exceeds ClinicalBERT in several metrics. This underscores the effectiveness of knowledge adapters and highlights their potential in settings with strict data privacy constraints. This approach also increases the level of interpretability of these models in a clinical context, which enhances our ability to precisely identify and apply the most relevant domain knowledge for specific tasks, thereby optimizing the model's performance and tailoring it to meet specific clinical needs. CONCLUSIONS: This research provides a basis for creating health knowledge graphs infused with physician knowledge, marking a significant step forward for PLMs in health care. Notably, the model balances integrating knowledge both comprehensively and selectively, addressing the heterogeneous nature of medical knowledge and the privacy needs of health care institutions.

19.
PeerJ Comput Sci ; 10: e2206, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39145211

RESUMEN

With the advent and improvement of ontological dictionaries (WordNet, Babelnet), the use of synsets-based text representations is gaining popularity in classification tasks. More recently, ontological dictionaries were used for reducing dimensionality in this kind of representation (e.g., Semantic Dimensionality Reduction System (SDRS) (Vélez de Mendizabal et al., 2020)). These approaches are based on the combination of semantically related columns by taking advantage of semantic information extracted from ontological dictionaries. Their main advantage is that they not only eliminate features but can also combine them, minimizing (low-loss) or avoiding (lossless) the loss of information. The most recent (and accurate) techniques included in this group are based on using evolutionary algorithms to find how many features can be grouped to reduce false positive (FP) and false negative (FN) errors obtained. The main limitation of these evolutionary-based schemes is the computational requirements derived from the use of optimization algorithms. The contribution of this study is a new lossless feature reduction scheme exploiting information from ontological dictionaries, which achieves slightly better accuracy (specially in FP errors) than optimization-based approaches but using far fewer computational resources. Instead of using computationally expensive evolutionary algorithms, our proposal determines whether two columns (synsets) can be combined by observing whether the instances included in a dataset (e.g., training dataset) containing these synsets are mostly of the same class. The study includes experiments using three datasets and a detailed comparison with two previous optimization-based approaches.

20.
Med Phys ; 2024 Jul 09.
Artículo en Inglés | MEDLINE | ID: mdl-38981056

RESUMEN

BACKGROUND: A comprehensive collection of data on doses in adult computed tomography procedures in Australia has not been undertaken for some time. This is largely due to the effort involved in collecting the data required for calculating the population dose. This data collection effort can be greatly reduced, and the coverage increased, if the process can be automated without major changes to the workflow of the imaging facilities providing the data. Success would provide a tool to determine a truly national assessment of the dose incurred through diagnostic imaging in Australia. PURPOSE: The aims of this study were to develop an automated tool to categorize electronic records of imaging procedures into a standardized set of broad procedure types, to validate the tool by applying it to data collected from nine facilities, and to assess the feasibility of applying the automated tool to compute population dose and determine the data manipulations required. METHODS: A rule-based classifier was implemented capitalizing on semantic and clinical rules. The keyword list was initially built from 609 unique study descriptions. It was then refined using an additional 414 unique study descriptions. The classifier was then tested on an additional 1198 unique study descriptions. Input from a radiologist provided the ground truth for the refinement of the classifier. RESULTS: From a sample of 238 139 studies containing 2794 unique study descriptions, the classifier correctly classified 2789 study types with only five misclassifications, demonstrating the feasibility of automating the process and the need for data pre-processing. Dose statistics for 21 categories were compiled using the 238 139 studies. CONCLUSION: The classifier achieved excellent classification results using the testing data supplied by the facilities. However, since all data supplied were from public facilities, the performance of the classifier may be biased. The performance of the classifier is yet to be tested on a more representative mix of private and public facilities.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA