Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 150
Filtrar
1.
Heliyon ; 10(17): e36892, 2024 Sep 15.
Artículo en Inglés | MEDLINE | ID: mdl-39281495

RESUMEN

Sarcasm in Sentiment Analysis (SA) is important due to the sense of sarcasm in sentences that differs from their literal meaning. Analysis of Arabic sarcasm still has many challenges like implicit indirect idioms to express the opinion, and lack of Arabic sarcasm corpus. In this paper, we proposed a new detecting model for sarcasm in Arabic tweets called the ArSa-Tweet model. It is based on implementing and developing Deep Learning (DL) models to classify tweets as sarcastic or not. The development of our proposed model consists of adding main improvements by applying robust preprocessing steps before feeding the data to the adapted DL models. The adapted DL models are LSTM, Multi-headed CNN-LSTM-GRU, BERT, AraBert-V01, and AraBert-V02. In addition, we proposed ArSa-data as a golden corpus that consists of Arabic tweets. A comparative process shows that our proposed ArSa-Tweet method has the most impact accuracy rate based on deploying the AraBert-V02 model, which obtains the best performance results in all accuracy metrics when compared with other methods.

2.
Proc Natl Acad Sci U S A ; 121(38): e2322764121, 2024 Sep 17.
Artículo en Inglés | MEDLINE | ID: mdl-39250662

RESUMEN

Are members of marginalized communities silenced on social media when they share personal experiences of racism? Here, we investigate the role of algorithms, humans, and platform guidelines in suppressing disclosures of racial discrimination. In a field study of actual posts from a neighborhood-based social media platform, we find that when users talk about their experiences as targets of racism, their posts are disproportionately flagged for removal as toxic by five widely used moderation algorithms from major online platforms, including the most recent large language models. We show that human users disproportionately flag these disclosures for removal as well. Next, in a follow-up experiment, we demonstrate that merely witnessing such suppression negatively influences how Black Americans view the community and their place in it. Finally, to address these challenges to equity and inclusion in online spaces, we introduce a mitigation strategy: a guideline-reframing intervention that is effective at reducing silencing behavior across the political spectrum.


Asunto(s)
Racismo , Medios de Comunicación Sociales , Humanos , Negro o Afroamericano , Algoritmos
3.
Acad Radiol ; 2024 Sep 07.
Artículo en Inglés | MEDLINE | ID: mdl-39245597

RESUMEN

RATIONALE AND OBJECTIVE: To compare the performance of large language model (LLM) based Gemini and Generative Pre-trained Transformers (GPTs) in data mining and generating structured reports based on free-text PET/CT reports for breast cancer after user-defined tasks. MATERIALS AND METHODS: Breast cancer patients (mean age, 50 years ± 11 [SD]; all female) who underwent consecutive 18F-FDG PET/CT for follow-up between July 2005 and October 2023 were retrospectively included in the study. A total of twenty reports from 10 patients were used to train user-defined text prompts for Gemini and GPTs, by which structured PET/CT reports were generated. The natural language processing (NLP) generated structured reports and the structured reports annotated by nuclear medicine physicians were compared in terms of data extraction accuracy and capacity of progress decision-making. Statistical methods, including chi-square test, McNemar test and paired samples t-test, were employed in the study. RESULTS: The structured PET/CT reports for 131 patients were generated by using the two NLP techniques, including Gemini and GPTs. In general, GPTs exhibited superiority over Gemini in data mining in terms of primary lesion size (89.6% vs. 53.8%, p < 0.001) and metastatic lesions (96.3% vs 89.6%, p < 0.001). Moreover, GPTs outperformed Gemini in making decision for progress (p < 0.001) and semantic similarity (F1 score 0.930 vs 0.907, p < 0.001) for reports. CONCLUSION: GPTs outperformed Gemini in generating structured reports based on free-text PET/CT reports, which is potentially applied in clinical practice. DATA AVAILABILITY: The data used and/or analyzed during the current study are available from the corresponding author on reasonable request.

4.
Cureus ; 16(8): e66209, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-39233986

RESUMEN

Extended reality (XR) simulations are becoming increasingly common in educational settings, particularly in medical education. Advancing XR devices to enhance these simulations is a booming field of research. This study seeks to understand the value of a novel, non-wearable mixed reality (MR) display during interactions with a simulated holographic patient, specifically in taking a medical history. Twenty-one first-year medical students at the University of North Carolina at Chapel Hill participated in the virtual patient (VP) simulations. On a five-point Likert scale, students overwhelmingly agreed with the statement that the simulations helped ensure they were progressing along learning objectives related to taking a patient history. However, they found that, at present, the simulations can only partially correct mistakes or provide clear feedback. This finding demonstrates that the novel hardware solution can help students engage in the activity, but the underlying software may need adjustment to attain sufficient pedagogical validity.

5.
Asian J Psychiatr ; 100: 104168, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39111087

RESUMEN

INTRODUCTION: Medical decision-making is crucial for effective treatment, especially in psychiatry where diagnosis often relies on subjective patient reports and a lack of high-specificity symptoms. Artificial intelligence (AI), particularly Large Language Models (LLMs) like GPT, has emerged as a promising tool to enhance diagnostic accuracy in psychiatry. This comparative study explores the diagnostic capabilities of several AI models, including Aya, GPT-3.5, GPT-4, GPT-3.5 clinical assistant (CA), Nemotron, and Nemotron CA, using clinical cases from the DSM-5. METHODS: We curated 20 clinical cases from the DSM-5 Clinical Cases book, covering a wide range of psychiatric diagnoses. Four advanced AI models (GPT-3.5 Turbo, GPT-4, Aya, Nemotron) were tested using prompts to elicit detailed diagnoses and reasoning. The models' performances were evaluated based on accuracy and quality of reasoning, with additional analysis using the Retrieval Augmented Generation (RAG) methodology for models accessing the DSM-5 text. RESULTS: The AI models showed varied diagnostic accuracy, with GPT-3.5 and GPT-4 performing notably better than Aya and Nemotron in terms of both accuracy and reasoning quality. While models struggled with specific disorders such as cyclothymic and disruptive mood dysregulation disorders, others excelled, particularly in diagnosing psychotic and bipolar disorders. Statistical analysis highlighted significant differences in accuracy and reasoning, emphasizing the superiority of the GPT models. DISCUSSION: The application of AI in psychiatry offers potential improvements in diagnostic accuracy. The superior performance of the GPT models can be attributed to their advanced natural language processing capabilities and extensive training on diverse text data, enabling more effective interpretation of psychiatric language. However, models like Aya and Nemotron showed limitations in reasoning, indicating a need for further refinement in their training and application. CONCLUSION: AI holds significant promise for enhancing psychiatric diagnostics, with certain models demonstrating high potential in interpreting complex clinical descriptions accurately. Future research should focus on expanding the dataset and integrating multimodal data to further enhance the diagnostic capabilities of AI in psychiatry.


Asunto(s)
Inteligencia Artificial , Trastornos Mentales , Psiquiatría , Humanos , Trastornos Mentales/diagnóstico , Psiquiatría/métodos , Manual Diagnóstico y Estadístico de los Trastornos Mentales , Procesamiento de Lenguaje Natural , Toma de Decisiones Clínicas/métodos , Adulto
6.
Front Hum Neurosci ; 18: 1421435, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39165904

RESUMEN

Neurolinguistic assessments play a vital role in neurological examinations, revealing a wide range of language and communication impairments associated with developmental disorders and acquired neurological conditions. Yet, a thorough neurolinguistic assessment is time-consuming and laborious and takes valuable resources from other tasks. To empower clinicians, healthcare providers, and researchers, we have developed Open Brain AI (OBAI). The aim of this computational platform is twofold. First, it aims to provide advanced AI tools to facilitate spoken and written language analysis, automate the analysis process, and reduce the workload associated with time-consuming tasks. The platform currently incorporates multilingual tools for English, Danish, Dutch, Finnish, French, German, Greek, Italian, Norwegian, Polish, Portuguese, Romanian, Russian, Spanish, and Swedish. The tools involve models for (i) audio transcription, (ii) automatic translation, (iii) grammar error correction, (iv) transcription to the International Phonetic Alphabet, (v) readability scoring, (vi) phonology, morphology, syntax, semantic measures (e.g., counts and proportions), and lexical measures. Second, it aims to support clinicians in conducting their research and automating everyday tasks with "OBAI Companion," an AI language assistant that facilitates language processing, such as structuring, summarizing, and editing texts. OBAI also provides tools for automating spelling and phonology scoring. This paper reviews OBAI's underlying architectures and applications and shows how OBAI can help professionals focus on higher-value activities, such as therapeutic interventions.

7.
Data Brief ; 55: 110690, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-39109169

RESUMEN

The Languages of the Indian subcontinent are less represented in current NLP literature. To mitigate this gap, we present the IndicDialogue dataset, which contains subtitles and dialogues in 10 major Indic languages: Hindi, Bengali, Marathi, Telugu, Tamil, Urdu, Odia, Sindhi, Nepali, and Assamese. This dataset is sourced from OpenSubtitles.org, with subtitles pre-processed to remove irrelevant tags, timestamps, square brackets, and links, ensuring the retention of relevant dialogues in JSONL files. The IndicDialogue dataset comprises 7750 raw subtitle files (SRT), 11 JSONL files, 6,853,518 dialogues, and 42,188,569 words. It is designed to serve as a foundation for language model pre-training for low-resource languages, enabling a wide range of downstream tasks including word embeddings, topic modeling, conversation synthesis, neural machine translation, and text summarization.

8.
Stud Health Technol Inform ; 316: 909-913, 2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39176940

RESUMEN

Electronic Health Records (EHRs) contain a wealth of unstructured patient data, making it challenging for physicians to do informed decisions. In this paper, we introduce a Natural Language Processing (NLP) approach for the extraction of therapies, diagnosis, and symptoms from ambulatory EHRs of patients with chronic Lupus disease. We aim to demonstrate the effort of a comprehensive pipeline where a rule-based system is combined with text segmentation, transformer-based topic analysis and clinical ontology, in order to enhance text preprocessing and automate rules' identification. Our approach is applied on a sub-cohort of 56 patients, with a total of 750 EHRs written in Italian language, achieving an Accuracy and an F-score over 97% and 90% respectively, in the three extracted domains. This work has the potential to be integrated with EHR systems to automate information extraction, minimizing the human intervention, and providing personalized digital solutions in the chronic Lupus disease domain.


Asunto(s)
Registros Electrónicos de Salud , Lupus Eritematoso Sistémico , Procesamiento de Lenguaje Natural , Humanos , Enfermedad Crónica , Minería de Datos/métodos
9.
BMC Med Inform Decis Mak ; 24(1): 195, 2024 Jul 16.
Artículo en Inglés | MEDLINE | ID: mdl-39014417

RESUMEN

BACKGROUND: Despite the significance and prevalence of acute respiratory distress syndrome (ARDS), its detection remains highly variable and inconsistent. In this work, we aim to develop an algorithm (ARDSFlag) to automate the diagnosis of ARDS based on the Berlin definition. We also aim to develop a visualization tool that helps clinicians efficiently assess ARDS criteria. METHODS: ARDSFlag applies machine learning (ML) and natural language processing (NLP) techniques to evaluate Berlin criteria by incorporating structured and unstructured data in an electronic health record (EHR) system. The study cohort includes 19,534 ICU admissions in the Medical Information Mart for Intensive Care III (MIMIC-III) database. The output is the ARDS diagnosis, onset time, and severity. RESULTS: ARDSFlag includes separate text classifiers trained using large training sets to find evidence of bilateral infiltrates in radiology reports (accuracy of 91.9%±0.5%) and heart failure/fluid overload in radiology reports (accuracy 86.1%±0.5%) and echocardiogram notes (accuracy 98.4%±0.3%). A test set of 300 cases, which was blindly and independently labeled for ARDS by two groups of clinicians, shows that ARDSFlag generates an overall accuracy of 89.0% (specificity = 91.7%, recall = 80.3%, and precision = 75.0%) in detecting ARDS cases. CONCLUSION: To our best knowledge, this is the first study to focus on developing a method to automate the detection of ARDS. Some studies have developed and used other methods to answer other research questions. Expectedly, ARDSFlag generates a significantly higher performance in all accuracy measures compared to those methods.


Asunto(s)
Algoritmos , Registros Electrónicos de Salud , Aprendizaje Automático , Procesamiento de Lenguaje Natural , Síndrome de Dificultad Respiratoria , Humanos , Síndrome de Dificultad Respiratoria/diagnóstico , Unidades de Cuidados Intensivos , Persona de Mediana Edad , Masculino , Femenino
10.
Cureus ; 16(5): e60044, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38854210

RESUMEN

Background Clinical trial matching, essential for advancing medical research, involves detailed screening of potential participants to ensure alignment with specific trial requirements. Research staff face challenges due to the high volume of eligible patients and the complexity of varying eligibility criteria. The traditional manual process, both time-consuming and error-prone, often leads to missed opportunities. Recently, large language models (LLMs), specifically generative pre-trained transformers (GPTs), have become impressive and impactful tools. Utilizing such tools from artificial intelligence (AI) and natural language processing (NLP) may enhance the accuracy and efficiency of this process through automated patient screening against established criteria. Methods Utilizing data from the National NLP Clinical Challenges (n2c2) 2018 Challenge, we utilized 202 longitudinal patient records. These records were annotated by medical professionals and evaluated against 13 selection criteria encompassing various health assessments. Our approach involved embedding medical documents into a vector database to determine relevant document sections and then using an LLM (OpenAI's GPT-3.5 Turbo and GPT-4) in tandem with structured and chain-of-thought prompting techniques for systematic document assessment against the criteria. Misclassified criteria were also examined to identify classification challenges. Results This study achieved an accuracy of 0.81, sensitivity of 0.80, specificity of 0.82, and a micro F1 score of 0.79 using GPT-3.5 Turbo, and an accuracy of 0.87, sensitivity of 0.85, specificity of 0.89, and micro F1 score of 0.86 using GPT-4. Notably, some criteria in the ground truth appeared mislabeled, an issue we couldn't explore further due to insufficient label generation guidelines on the website. Conclusion Our findings underscore the potential of AI and NLP technologies, including LLMs, in the clinical trial matching process. The study demonstrated strong capabilities in identifying eligible patients and minimizing false inclusions. Such automated systems promise to alleviate the workload of research staff and improve clinical trial enrollment, thus accelerating the process and enhancing the overall feasibility of clinical research. Further work is needed to determine the potential of this approach when implemented on real clinical data.

11.
Cureus ; 16(4): e57728, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38711724

RESUMEN

Clinical Decision Support Systems (CDSS) are essential tools in contemporary healthcare, enhancing clinicians' decisions and patient outcomes. The integration of artificial intelligence (AI) is now revolutionizing CDSS even further. This review delves into AI technologies transforming CDSS, their applications in healthcare decision-making, associated challenges, and the potential trajectory toward fully realizing AI-CDSS's potential. The review begins by laying the groundwork with a definition of CDSS and its function within the healthcare field. It then highlights the increasingly significant role that AI is playing in enhancing CDSS effectiveness and efficiency, underlining its evolving prominence in shaping healthcare practices. It examines the integration of AI technologies into CDSS, including machine learning algorithms like neural networks and decision trees, natural language processing, and deep learning. It also addresses the challenges associated with AI integration, such as interpretability and bias. We then shift to AI applications within CDSS, with real-life examples of AI-driven diagnostics, personalized treatment recommendations, risk prediction, early intervention, and AI-assisted clinical documentation. The review emphasizes user-centered design in AI-CDSS integration, addressing usability, trust, workflow, and ethical and legal considerations. It acknowledges prevailing obstacles and suggests strategies for successful AI-CDSS adoption, highlighting the need for workflow alignment and interdisciplinary collaboration. The review concludes by summarizing key findings, underscoring AI's transformative potential in CDSS, and advocating for continued research and innovation. It emphasizes the need for collaborative efforts to realize a future where AI-powered CDSS optimizes healthcare delivery and improves patient outcomes.

12.
ESC Heart Fail ; 2024 May 13.
Artículo en Inglés | MEDLINE | ID: mdl-38741373

RESUMEN

AIMS: Worsening heart failure (WHF) events occurring in non-inpatient settings are becoming increasingly recognized, with implications for prognostication. We evaluate the performance of a natural language processing (NLP)-based approach compared with traditional diagnostic coding for non-inpatient clinical encounters and left ventricular ejection fraction (LVEF). METHODS AND RESULTS: We compared characteristics for encounters that did vs. did not meet WHF criteria, stratified by care setting [i.e. emergency department (ED) and observation stay]. Overall, 8407 (22%) encounters met NLP-based criteria for WHF (3909 ED visits and 4498 observation stays). The use of an NLP-derived definition adjudicated 3983 (12%) of non-primary HF diagnoses as meeting consensus definitions for WHF. The most common diagnosis indicated in these encounters was dyspnoea. Results were primarily driven by observation stays, in which 2205 (23%) encounters with a secondary HF diagnosis met the WHF definition by NLP. CONCLUSIONS: The use of standard claims-based adjudication for primary diagnosis in the non-inpatient setting may lead to misclassification of WHF events in the ED and overestimate observation stays. Primary diagnoses alone may underestimate the burden of WHF in non-hospitalized settings.

13.
Front Psychiatry ; 15: 1389597, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38803678

RESUMEN

Background: Individuals at clinical high risk (CHR) for psychosis experience subtle emotional disturbances that are traditionally difficult to assess, but natural language processing (NLP) methods may provide novel insight into these symptoms. We predicted that CHR individuals would express more negative emotionality and less emotional language when compared to controls. We also examined associations with symptomatology. Methods: Participants included 49 CHR individuals and 42 healthy controls who completed a semi-structured narrative interview. Interview transcripts were analyzed using Linguistic Inquiry and Word Count (LIWC) to assess the emotional tone of the language (tone -the ratio of negative to positive language) and count positive/negative words used. Participants also completed clinical symptom assessments to determine CHR status and characterize symptoms (i.e., positive and negative symptom domains). Results: The CHR group had more negative emotional tone compared to healthy controls (t=2.676, p=.009), which related to more severe positive symptoms (r2=.323, p=.013). The percentages of positive and negative words did not differ between groups (p's>.05). Conclusions: Language analyses provided accessible, ecologically valid insight into affective dysfunction and psychosis risk symptoms. Natural language processing analyses unmasked differences in language for CHR that captured language tendencies that were more nuanced than the words that are chosen.

14.
Med Ref Serv Q ; 43(2): 196-202, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38722609

RESUMEN

Named entity recognition (NER) is a powerful computer system that utilizes various computing strategies to extract information from raw text input, since the early 1990s. With rapid advancement in AI and computing, NER models have gained significant attention and been serving as foundational tools across numerus professional domains to organize unstructured data for research and practical applications. This is particularly evident in the medical and healthcare fields, where NER models are essential in efficiently extract critical information from complex documents that are challenging for manual review. Despite its successes, NER present limitations in fully comprehending natural language nuances. However, the development of more advanced and user-friendly models promises to improve work experiences of professional users significantly.


Asunto(s)
Almacenamiento y Recuperación de la Información , Procesamiento de Lenguaje Natural , Almacenamiento y Recuperación de la Información/métodos , Humanos , Inteligencia Artificial
15.
Drug Discov Today ; 29(6): 104018, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38723763

RESUMEN

Text summarization is crucial in scientific research, drug discovery and development, regulatory review, and more. This task demands domain expertise, language proficiency, semantic prowess, and conceptual skill. The recent advent of large language models (LLMs), such as ChatGPT, offers unprecedented opportunities to automate this process. We compared ChatGPT-generated summaries with those produced by human experts using FDA drug labeling documents. The labeling contains summaries of key labeling sections, making them an ideal human benchmark to evaluate ChatGPT's summarization capabilities. Analyzing >14000 summaries, we observed that ChatGPT-generated summaries closely resembled those generated by human experts. Importantly, ChatGPT exhibited even greater similarity when summarizing drug safety information. These findings highlight ChatGPT's potential to accelerate work in critical areas, including drug safety.


Asunto(s)
Etiquetado de Medicamentos , United States Food and Drug Administration , Humanos , Estados Unidos , Procesamiento de Lenguaje Natural , Efectos Colaterales y Reacciones Adversas Relacionados con Medicamentos
16.
Sensors (Basel) ; 24(10)2024 May 15.
Artículo en Inglés | MEDLINE | ID: mdl-38793999

RESUMEN

The complexity and the criticality of automotive electronic implanted systems are steadily advancing and that is especially the case for automotive software development. ISO 26262 describes requirements for the development process to confirm the safety of such complex systems. Among these requirements, fault injection is a reliable technique to assess the effectiveness of safety mechanisms and verify the correct implementation of the safety requirements. However, the method of injecting the fault in the system under test in many cases is still manual and depends on an expert, requiring a high level of knowledge of the system. In complex systems, it consumes time, is difficult to execute, and takes effort, because the testers limit the fault injection experiments and inject the minimum number of possible test cases. Fault injection enables testers to identify and address potential issues with a system under test before they become actual problems. In the automotive industry, failures can have serious hazards. In these systems, it is essential to ensure that the system can operate safely even in the presence of faults. We propose an approach using natural language processing (NLP) technologies to automatically derive the fault test cases from the functional safety requirements (FSRs) and execute them automatically by hardware-in-the-loop (HIL) in real time according to the black-box concept and the ISO 26262 standard. The approach demonstrates effectiveness in automatically identifying fault injection locations and conditions, simplifying the testing process, and providing a scalable solution for various safety-critical systems.

17.
Front Robot AI ; 11: 1362463, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38726067

RESUMEN

The condition for artificial agents to possess perceivable intentions can be considered that they have resolved a form of the symbol grounding problem. Here, the symbol grounding is considered an achievement of the state where the language used by the agent is endowed with some quantitative meaning extracted from the physical world. To achieve this type of symbol grounding, we adopt a method for characterizing robot gestures with quantitative meaning calculated from word-distributed representations constructed from a large corpus of text. In this method, a "size image" of a word is generated by defining an axis (index) that discriminates the "size" of the word in the word-distributed vector space. The generated size images are converted into gestures generated by a physical artificial agent (robot). The robot's gesture can be set to reflect either the size of the word in terms of the amount of movement or in terms of its posture. To examine the perception of communicative intention in the robot that performs the gestures generated as described above, the authors examine human ratings on "the naturalness" obtained through an online survey, yielding results that partially validate our proposed method. Based on the results, the authors argue for the possibility of developing advanced artifacts that achieve human-like symbolic grounding.

18.
Sci Rep ; 14(1): 10262, 2024 May 04.
Artículo en Inglés | MEDLINE | ID: mdl-38704434

RESUMEN

As sustainability emerges as a crucial factor in the development of modern enterprises, integrating environmental, social, and governance (ESG) information into financial assessments has become essential. ESG indicators serve as important metrics in evaluating a company's sustainable practices and governance effectiveness, influencing investor trust and future growth potential, ultimately affecting stock prices. This study proposes an innovative approach that combines ESG sentiment index extracted from news with technical indicators to predict the S&P 500 index. By utilizing a deep learning model and exploring optimal window sizes, the study explores the best model through mean absolute percentage error (MAPE) as an evaluation metric. Additionally, an ablation test clarifies the influence of ESG and its causality with the S&P 500 index. The experimental results demonstrate improved predictive accuracy when considering ESG sentiment compared to relying solely on technical indicators or historical data. This comprehensive methodology enhances the advantage of stock price prediction by integrating technical indicators, which consider short-term fluctuations, with ESG information, providing long-term effects. Furthermore, it offers valuable insights for investors and financial market experts, validating the necessity to consider ESG for financial assets and introducing a new perspective to develop investment strategies and decision-making processes.

19.
MethodsX ; 12: 102737, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38774687

RESUMEN

In the digital age, the proliferation of health-related information online has heightened the risk of misinformation, posing substantial threats to public well-being. This research conducts a meticulous comparative analysis of classification models, focusing on detecting health misinformation. The study evaluates the performance of traditional machine learning models and advanced graph convolutional networks (GCN) across critical algorithmic metrics. The results comprehensively understand each algorithm's effectiveness in identifying health misinformation and provide valuable insights for combating the pervasive spread of false health information in the digital landscape. GCN with TF-IDF gives the best result, as shown in the result section. •The research method involves a comparative analysis of classification algorithms to detect health misinformation, exploring traditional machine learning models and graph convolutional networks.•This research used algorithms such as Passive Aggressive Classifier, Random Forest, Decision Tree, Logistic Regression, Light GBM, GCN, GCN with BERT, GCN with TF-IDF, and GCN with Word2Vec were employed. Performance Metrics: Accuracy: for Passive Aggressive Classifier: 85.75 %, Random Forest: 86 %, Decision Tree: 81.30 %, Light BGM: 83.29 %, normal GCN: 84.53 %, GCN with BERT: 85.00 %, GCN with TR-IDF: 93.86 % and GCN with word2Vec: 81.00 %•Algorithmic performance metrics, including accuracy, precision, recall, and F1-score, were systematically evaluated to assess the efficacy of each model in detecting health misinformation, focusing on understanding the strengths and limitations of different approaches. The superior performance of Graph Convolutional Networks (GCNs) with TF-IDF embedding, achieving an accuracy of 93.86.

20.
Syst Rev ; 13(1): 107, 2024 Apr 15.
Artículo en Inglés | MEDLINE | ID: mdl-38622611

RESUMEN

BACKGROUND: Abstract review is a time and labor-consuming step in the systematic and scoping literature review in medicine. Text mining methods, typically natural language processing (NLP), may efficiently replace manual abstract screening. This study applies NLP to a deliberately selected literature review problem, the trend of using NLP in medical research, to demonstrate the performance of this automated abstract review model. METHODS: Scanning PubMed, Embase, PsycINFO, and CINAHL databases, we identified 22,294 with a final selection of 12,817 English abstracts published between 2000 and 2021. We invented a manual classification of medical fields, three variables, i.e., the context of use (COU), text source (TS), and primary research field (PRF). A training dataset was developed after reviewing 485 abstracts. We used a language model called Bidirectional Encoder Representations from Transformers to classify the abstracts. To evaluate the performance of the trained models, we report a micro f1-score and accuracy. RESULTS: The trained models' micro f1-score for classifying abstracts, into three variables were 77.35% for COU, 76.24% for TS, and 85.64% for PRF. The average annual growth rate (AAGR) of the publications was 20.99% between 2000 and 2020 (72.01 articles (95% CI: 56.80-78.30) yearly increase), with 81.76% of the abstracts published between 2010 and 2020. Studies on neoplasms constituted 27.66% of the entire corpus with an AAGR of 42.41%, followed by studies on mental conditions (AAGR = 39.28%). While electronic health or medical records comprised the highest proportion of text sources (57.12%), omics databases had the highest growth among all text sources with an AAGR of 65.08%. The most common NLP application was clinical decision support (25.45%). CONCLUSIONS: BioBERT showed an acceptable performance in the abstract review. If future research shows the high performance of this language model, it can reliably replace manual abstract reviews.


Asunto(s)
Investigación Biomédica , Procesamiento de Lenguaje Natural , Humanos , Lenguaje , Minería de Datos , Registros Electrónicos de Salud
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA