Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 12 de 12
Filtrar
Más filtros











Base de datos
Intervalo de año de publicación
1.
Biomed Phys Eng Express ; 10(3)2024 Mar 18.
Artículo en Inglés | MEDLINE | ID: mdl-38350128

RESUMEN

The paper aims to explore the current state of understanding surrounding in silico oral modelling. This involves exploring methodologies, technologies and approaches pertaining to the modelling of the whole oral cavity; both internally and externally visible structures that may be relevant or appropriate to oral actions. Such a model could be referred to as a 'complete model' which includes consideration of a full set of facial features (i.e. not only mouth) as well as synergistic stimuli such as audio and facial thermal data. 3D modelling technologies capable of accurately and efficiently capturing a complete representation of the mouth for an individual have broad applications in the study of oral actions, due to their cost-effectiveness and time efficiency. This review delves into the field of clinical phonetics to classify oral actions pertaining to both speech and non-speech movements, identifying how the various vocal organs play a role in the articulatory and masticatory process. Vitaly, it provides a summation of 12 articulatory recording methods, forming a tool to be used by researchers in identifying which method of recording is appropriate for their work. After addressing the cost and resource-intensive limitations of existing methods, a new system of modelling is proposed that leverages external to internal correlation modelling techniques to create a more efficient models of the oral cavity. The vision is that the outcomes will be applicable to a broad spectrum of oral functions related to physiology, health and wellbeing, including speech, oral processing of foods as well as dental health. The applications may span from speech correction, designing foods for the aging population, whilst in the dental field we would be able to gain information about patient's oral actions that would become part of creating a personalised dental treatment plan.


Asunto(s)
Boca , Habla , Humanos , Anciano , Boca/fisiología , Habla/fisiología , Fonética
2.
IEEE Trans Pattern Anal Mach Intell ; 46(2): 805-822, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-37851557

RESUMEN

Automatically recognising apparent emotions from face and voice is hard, in part because of various sources of uncertainty, including in the input data and the labels used in a machine learning framework. This paper introduces an uncertainty-aware multimodal fusion approach that quantifies modality-wise aleatoric or data uncertainty towards emotion prediction. We propose a novel fusion framework, in which latent distributions over unimodal temporal context are learned by constraining their variance. These variance constraints, Calibration and Ordinal Ranking, are designed such that the variance estimated for a modality can represent how informative the temporal context of that modality is w.r.t. emotion recognition. When well-calibrated, modality-wise uncertainty scores indicate how much their corresponding predictions are likely to differ from the ground truth labels. Well-ranked uncertainty scores allow the ordinal ranking of different frames across different modalities. To jointly impose both these constraints, we propose a softmax distributional matching loss. Our evaluation on AVEC 2019 CES, CMU-MOSEI, and IEMOCAP datasets shows that the proposed multimodal fusion method not only improves the generalisation performance of emotion recognition models and their predictive uncertainty estimates, but also makes the models robust to novel noise patterns encountered at test time.

3.
Front Neurogenom ; 4: 994969, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-38234474

RESUMEN

Background: While efforts to establish best practices with functional near infrared spectroscopy (fNIRS) signal processing have been published, there are still no community standards for applying machine learning to fNIRS data. Moreover, the lack of open source benchmarks and standard expectations for reporting means that published works often claim high generalisation capabilities, but with poor practices or missing details in the paper. These issues make it hard to evaluate the performance of models when it comes to choosing them for brain-computer interfaces. Methods: We present an open-source benchmarking framework, BenchNIRS, to establish a best practice machine learning methodology to evaluate models applied to fNIRS data, using five open access datasets for brain-computer interface (BCI) applications. The BenchNIRS framework, using a robust methodology with nested cross-validation, enables researchers to optimise models and evaluate them without bias. The framework also enables us to produce useful metrics and figures to detail the performance of new models for comparison. To demonstrate the utility of the framework, we present a benchmarking of six baseline models [linear discriminant analysis (LDA), support-vector machine (SVM), k-nearest neighbours (kNN), artificial neural network (ANN), convolutional neural network (CNN), and long short-term memory (LSTM)] on the five datasets and investigate the influence of different factors on the classification performance, including: number of training examples and size of the time window of each fNIRS sample used for classification. We also present results with a sliding window as opposed to simple classification of epochs, and with a personalised approach (within subject data classification) as opposed to a generalised approach (unseen subject data classification). Results and discussion: Results show that the performance is typically lower than the scores often reported in literature, and without great differences between models, highlighting that predicting unseen data remains a difficult task. Our benchmarking framework provides future authors, who are achieving significant high classification scores, with a tool to demonstrate the advances in a comparable way. To complement our framework, we contribute a set of recommendations for methodology decisions and writing papers, when applying machine learning to fNIRS data.

4.
Pediatr Res ; 2022 Mar 03.
Artículo en Inglés | MEDLINE | ID: mdl-35241791

RESUMEN

Advances in neonatal care have resulted in improved outcomes for high-risk newborns with technologies playing a significant part although many were developed for the neonatal intensive care unit. The care provided in the delivery room (DR) during the first few minutes of life can impact short- and long-term neonatal outcomes. Increasingly, technologies have a critical role to play in the DR particularly with monitoring and information provision. However, the DR is a unique environment and has major challenges around the period of foetal to neonatal transition that need to be overcome when developing new technologies. This review focuses on current DR technologies as well as those just emerging and further over the horizon. We identify what key opinion leaders in DR care think of current technologies, what the important DR measures are to them, and which technologies might be useful in the future. We link these with key technologies including respiratory function monitors, electoral impedance tomography, videolaryngoscopy, augmented reality, video recording, eye tracking, artificial intelligence, and contactless monitoring. Encouraging funders and industry to address the unique technological challenges of newborn care in the DR will allow the continued improvement of outcomes of high-risk infants from the moment of birth. IMPACT: Technological advances for newborn delivery room care require consideration of the unique environment, the variable patient characteristics, and disease states, as well as human factor challenges. Neonatology as a speciality has embraced technology, allowing its rapid progression and improved outcomes for infants, although innovation in the delivery room often lags behind that in the intensive care unit. Investing in new and emerging technologies can support healthcare providers when optimising care and could improve training, safety, and neonatal outcomes.

5.
Image Vis Comput ; 83-84: 87-99, 2019.
Artículo en Inglés | MEDLINE | ID: mdl-31762527

RESUMEN

A baby's gestational age determines whether or not they are premature, which helps clinicians decide on suitable post-natal treatment. The most accurate dating methods use Ultrasound Scan (USS) machines, but these are expensive, require trained personnel and cannot always be deployed to remote areas. In the absence of USS, the Ballard Score, a postnatal clinical examination, can be used. However, this method is highly subjective and results vary widely depending on the experience of the examiner. Our main contribution is a novel system for automatic postnatal gestational age estimation using small sets of images of a newborn's face, foot and ear. Our two-stage architecture makes the most out of Convolutional Neural Networks trained on small sets of images to predict broad classes of gestational age, and then fuses the outputs of these discrete classes with a baby's weight to make fine-grained predictions of gestational age using Support Vector Regression. On a purpose-collected dataset of 130 babies, experiments show that our approach surpasses current automatic state-of-the-art postnatal methods and attains an expected error of 6 days. It is three times more accurate than the Ballard method. Making use of images improves predictions by 33% compared to using weight only. This indicates that even with a very small set of data, our method is a viable candidate for postnatal gestational age estimation in areas were USS is not available.

6.
IEEE Trans Pattern Anal Mach Intell ; 40(9): 2037-2050, 2018 09.
Artículo en Inglés | MEDLINE | ID: mdl-28858786

RESUMEN

Linear regression is a fundamental building block in many face detection and tracking algorithms, typically used to predict shape displacements from image features through a linear mapping. This paper presents a Functional Regression solution to the least squares problem, which we coin Continuous Regression, resulting in the first real-time incremental face tracker. Contrary to prior work in Functional Regression, in which B-splines or Fourier series were used, we propose to approximate the input space by its first-order Taylor expansion, yielding a closed-form solution for the continuous domain of displacements. We then extend the continuous least squares problem to correlated variables, and demonstrate the generalisation of our approach. We incorporate Continuous Regression into the cascaded regression framework, and show its computational benefits for both training and testing. We then present a fast approach for incremental learning within Cascaded Continuous Regression, coined iCCR, and show that its complexity allows real-time face tracking, being 20 times faster than the state of the art. To the best of our knowledge, this is the first incremental face tracker that is shown to operate in real-time. We show that iCCR achieves state-of-the-art performance on the 300-VW dataset, the most recent, large-scale benchmark for face tracking.


Asunto(s)
Cara/anatomía & histología , Procesamiento de Imagen Asistido por Computador/métodos , Modelos Lineales , Reconocimiento de Normas Patrones Automatizadas/métodos , Algoritmos , Bases de Datos Factuales , Humanos , Aprendizaje Automático
7.
Artículo en Inglés | MEDLINE | ID: mdl-29606917

RESUMEN

The field of Automatic Facial Expression Analysis has grown rapidly in recent years. However, despite progress in new approaches as well as benchmarking efforts, most evaluations still focus on either posed expressions, near-frontal recordings, or both. This makes it hard to tell how existing expression recognition approaches perform under conditions where faces appear in a wide range of poses (or camera views), displaying ecologically valid expressions. The main obstacle for assessing this is the availability of suitable data, and the challenge proposed here addresses this limitation. The FG 2017 Facial Expression Recognition and Analysis challenge (FERA 2017) extends FERA 2015 to the estimation of Action Units occurrence and intensity under different camera views. In this paper we present the third challenge in automatic recognition of facial expressions, to be held in conjunction with the 12th IEEE conference on Face and Gesture Recognition, May 2017, in Washington, United States. Two sub-challenges are defined: the detection of AU occurrence, and the estimation of AU intensity. In this work we outline the evaluation protocol, the data used, and the results of a baseline method for both sub-challenges.

8.
IEEE Trans Affect Comput ; 7(4): 435-451, 2016.
Artículo en Inglés | MEDLINE | ID: mdl-30906508

RESUMEN

Pain-related emotions are a major barrier to effective self rehabilitation in chronic pain. Automated coaching systems capable of detecting these emotions are a potential solution. This paper lays the foundation for the development of such systems by making three contributions. First, through literature reviews, an overview of how pain is expressed in chronic pain and the motivation for detecting it in physical rehabilitation is provided. Second, a fully labelled multimodal dataset (named 'EmoPain') containing high resolution multiple-view face videos, head mounted and room audio signals, full body 3D motion capture and electromyographic signals from back muscles is supplied. Natural unconstrained pain related facial expressions and body movement behaviours were elicited from people with chronic pain carrying out physical exercises. Both instructed and non-instructed exercises were considered to reflect traditional scenarios of physiotherapist directed therapy and home-based self-directed therapy. Two sets of labels were assigned: level of pain from facial expressions annotated by eight raters and the occurrence of six pain-related body behaviours segmented by four experts. Third, through exploratory experiments grounded in the data, the factors and challenges in the automated recognition of such expressions and behaviour are described, the paper concludes by discussing potential avenues in the context of these findings also highlighting differences for the two exercise scenarios addressed.

9.
IEEE Trans Cybern ; 44(2): 161-74, 2014 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-23757539

RESUMEN

Both the configuration and the dynamics of facial expressions are crucial for the interpretation of human facial behavior. Yet to date, the vast majority of reported efforts in the field either do not take the dynamics of facial expressions into account, or focus only on prototypic facial expressions of six basic emotions. Facial dynamics can be explicitly analyzed by detecting the constituent temporal segments in Facial Action Coding System (FACS) Action Units (AUs)-onset, apex, and offset. In this paper, we present a novel approach to explicit analysis of temporal dynamics of facial actions using the dynamic appearance descriptor Local Phase Quantization from Three Orthogonal Planes (LPQ-TOP). Temporal segments are detected by combining a discriminative classifier for detecting the temporal segments on a frame-by-frame basis with Markov Models that enforce temporal consistency over the whole episode. The system is evaluated in detail over the MMI facial expression database, the UNBC-McMaster pain database, the SAL database, the GEMEP-FERA dataset in database-dependent experiments, in cross-database experiments using the Cohn-Kanade, and the SEMAINE databases. The comparison with other state-of-the-art methods shows that the proposed LPQ-TOP method outperforms the other approaches for the problem of AU temporal segment detection, and that overall AU activation detection benefits from dynamic appearance information.


Asunto(s)
Biometría/métodos , Cara/anatomía & histología , Cara/fisiología , Expresión Facial , Interpretación de Imagen Asistida por Computador/métodos , Imagen Multimodal/métodos , Reconocimiento de Normas Patrones Automatizadas/métodos , Algoritmos , Inteligencia Artificial , Simulación por Computador , Humanos , Modelos Biológicos , Modelos Estadísticos , Fotograbar/métodos , Reproducibilidad de los Resultados , Sensibilidad y Especificidad
10.
IEEE Trans Pattern Anal Mach Intell ; 35(5): 1149-63, 2013 May.
Artículo en Inglés | MEDLINE | ID: mdl-23520256

RESUMEN

We propose a new algorithm to detect facial points in frontal and near-frontal face images. It combines a regression-based approach with a probabilistic graphical model-based face shape model that restricts the search to anthropomorphically consistent regions. While most regression-based approaches perform a sequential approximation of the target location, our algorithm detects the target location by aggregating the estimates obtained from stochastically selected local appearance information into a single robust prediction. The underlying assumption is that by aggregating the different estimates, their errors will cancel out as long as the regressor inputs are uncorrelated. Once this new perspective is adopted, the problem is reformulated as how to optimally select the test locations over which the regressors are evaluated. We propose to extend the regression-based model to provide a quality measure of each prediction, and use the shape model to restrict and correct the sampling region. Our approach combines the low computational cost typical of regression-based approaches with the robustness of exhaustive-search approaches. The proposed algorithm was tested on over 7,500 images from five databases. Results showed significant improvement over the current state of the art.


Asunto(s)
Identificación Biométrica/métodos , Cara/anatomía & histología , Procesamiento de Imagen Asistido por Computador/métodos , Análisis de Regresión , Bases de Datos Factuales , Humanos , Máquina de Vectores de Soporte
11.
IEEE Trans Syst Man Cybern B Cybern ; 42(1): 28-43, 2012 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-21926026

RESUMEN

Past work on automatic analysis of facial expressions has focused mostly on detecting prototypic expressions of basic emotions like happiness and anger. The method proposed here enables the detection of a much larger range of facial behavior by recognizing facial muscle actions [action units (AUs)] that compound expressions. AUs are agnostic, leaving the inference about conveyed intent to higher order decision making (e.g., emotion recognition). The proposed fully automatic method not only allows the recognition of 22 AUs but also explicitly models their temporal characteristics (i.e., sequences of temporal segments: neutral, onset, apex, and offset). To do so, it uses a facial point detector based on Gabor-feature-based boosted classifiers to automatically localize 20 facial fiducial points. These points are tracked through a sequence of images using a method called particle filtering with factorized likelihoods. To encode AUs and their temporal activation models based on the tracking data, it applies a combination of GentleBoost, support vector machines, and hidden Markov models. We attain an average AU recognition rate of 95.3% when tested on a benchmark set of deliberately displayed facial expressions and 72% when tested on spontaneous expressions.


Asunto(s)
Inteligencia Artificial , Expresión Facial , Interpretación de Imagen Asistida por Computador/métodos , Reconocimiento de Normas Patrones Automatizadas/métodos , Fotograbar/métodos , Técnica de Sustracción , Grabación en Video/métodos , Humanos
12.
Med Image Comput Comput Assist Interv ; 11(Pt 1): 486-93, 2008.
Artículo en Inglés | MEDLINE | ID: mdl-18979782

RESUMEN

This paper presents a new framework for the analysis of anatomical connectivity derived from diffusion tensor MRI. The framework has been applied to estimate whole brain structural networks using diffusion data from 174 adult subjects. In the proposed approach, each brain is first segmented into 83 anatomical regions via label propagation of multiple atlases and subsequent decision fusion. For each pair of anatomical regions the probability of connection and its strength is then estimated using a modified version of probabilistic tractography. The resulting brain networks have been classified according to age and gender using non-linear support vector machines with GentleBoost feature extraction. Classification performance was tested using a leave-one-out approach and the mean accuracy obtained was 85.4%.


Asunto(s)
Algoritmos , Inteligencia Artificial , Encéfalo/anatomía & histología , Imagen de Difusión por Resonancia Magnética/métodos , Interpretación de Imagen Asistida por Computador/métodos , Vías Nerviosas/anatomía & histología , Reconocimiento de Normas Patrones Automatizadas/métodos , Adulto , Anciano , Anciano de 80 o más Años , Simulación por Computador , Femenino , Humanos , Aumento de la Imagen/métodos , Masculino , Persona de Mediana Edad , Modelos Biológicos , Modelos Estadísticos , Análisis Multivariante , Reproducibilidad de los Resultados , Sensibilidad y Especificidad
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA