You are viewing the site in preview mode

Skip to main content

Predicting neurodevelopmental disorders using machine learning models and electronic health records – status of the field

Abstract

Machine learning (ML) is increasingly used to identify patterns that could predict neurodevelopmental disorders (NDDs), such as autism spectrum disorder (ASD) and attention-deficit hyperactivity disorder (ADHD). One key source of multilevel data for ML prediction models includes population-based registers and electronic health records. These can contain rich information on individual and familial medical histories and socio-demographics. This review summarizes studies published between 2010–2022 that used ML algorithms to develop predictive models for NDDs using population-based registers and electronic health records. A literature search identified 1191 articles, of which 32 were retained. Of these, 47% developed ASD prediction models and 25% ADHD models. Classical ML methods were used in 82% of studies and in particular tree-based prediction models performed well. The sensitivity of the models was lower than 75% for most studies, while the area under the curve (AUC) was greater than 75%. The most important predictors were patient and familial medical history and sociodemographic factors. Using private in-house datasets makes comparing and validating model generalizability across studies difficult. The ML model development and reporting guidelines were adopted only in a few recently reported studies. More work is needed to harness the power of data for detecting NDDs early.

Introduction

Neurodevelopmental disorders (NDDs) are childhood conditions impacting cognitive development, motor functioning, and higher-order executive functions, such as emotions, language, and memory. The Diagnostic and Statistical Manual of Mental Disorders (DSM-5) specifies that the symptoms of NDDs may be present in early development but not manifest fully into core diagnostic features till later in life [1]. NDDs are lifelong and encompass multiple conditions, including autism spectrum disorder (ASD), attention deficit hyperactivity disorder (ADHD), intellectual disability (ID), communication disorders, and motor disorders. NDDs frequently co-occur in individuals, and their deficit or excessive symptoms may overlap amongst conditions. Currently, there are no known biomarkers for any NDD [2, 3], and the diagnosis is given based on their phenotypic manifestations by trained clinicians observing an individual over time. The subjectivity in the diagnostic process and the lack of trained clinicians, especially in remote regions, leads to delays in diagnosis, depriving early intervention opportunities. Early diagnosis of NDDs can lead to early interventions, improved prognosis, and treatment outcomes [4].

Machine learning (ML) methods have the potential to manage the myriad of these complexities by developing robust prediction models for NDDs using various data [3]. If designed right, the power of prediction models comes from their generalization ability to predict unseen individual outcomes reliably across populations. The splurge in a massive amount of multimodal, high-dimensional healthcare data warrants advanced analytical approaches for mining them for subtle patterns, designing prediction models, and characterizing medical conditions. The learning process from the training data can be supervised or unsupervised, and the resulting learned information is captured in a model. The model thus developed is further applied for deriving insights into new unseen data and should generalize well across populations [5]. In the supervised learning process, the goal is to learn a mapping between the provided input–output label pairs, while in the unsupervised learning, in the absence of target labels, the aim is to identify hidden patterns in the data. The field of ML has made tremendous progress in the last two decades in natural language processing (NLP), computer vision (CV), and speech processing domains. In addition, ML models have shown promising results of predictive diagnostic abilities in cancer [6, 7], and cardiovascular diseases [8].

In the past few decades, an increasing number of nationwide population-based registers and digitized electronic health records (EHRs) have been administered to develop national-wide healthcare systems, social services, and research [9]. Such an ensemble of registers contains a large volume of rich information that can be exploited for developing robust and intelligent ML approaches to develop prediction, prognosis, and treatment response models for multiple health conditions [10,11,12,13,14,15,16,17] The EHRs contain information such as patient’s medical history, clinical visits, prescribed medicines, and sociodemographic details that may provide insights into the development of conditions.

ML models have shown great promise in screening and diagnostic prediction of NDDs using different modalities, such as clinical and behavioural assessments, genetics, and brain imaging techniques such as Electroencephalography (EEG) and Magnetic Resource Imaging (MRI) [3]. A common challenge with these modalities is the unavailability of labeled large sample sizes for building effective ML models. EHRs and population-based registers do not generally have this limitation and, therefore, could potentially have data to help construct NDD prediction models. However, the lack of easy access to EHR data stored in unstructured formats hinders the ability to perform analyses more efficiently. Despite this, studies utilizing ML methods for processing population-based registers and EHRs in the context of NDDs are emerging.

In this review, we sought to identify studies employing ML models with EHRs and population-based registers for predicting NDD. Our objectives included reporting the variables used for prediction and assessing model performance across studies. Additionally, we synthesized recommendations for future research based on our findings. We hypothesized that existing studies would exhibit heterogeneity in design, sample sizes, predictor variables, and outcome measurements. Such variability could challenge these models' comprehensive conclusions, generalizability, and clinical implementation. We provide a summary and evaluation of the current state of ML-based prediction models for NDD diagnosis, considering their quality and performance. We also address the challenges and potential avenues for the successful integration of ML in enhancing neurodevelopmental disorders diagnosis. To our knowledge, this is the first review focused on ML models leveraging EHRs and population-based registers for NDD prediction.

Methods

Protocol and information sources

We performed this scoping review according to the five-stage framework of Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA) Checklist and Explanation [18,19,20]. We restricted the search to the articles to include peer-reviewed and published work between 2010 – 2022 followed with a narrow search for the first months of 2023. A literature search was performed in Medline, Embase, Cochrane Library, Web of Science and PsycInfo databases. A simplified search was also done in Google Scholar. An information specialist from Karolinska Institutet University library assisted in our initial literature search to screen articles for the final review.

Search strategy and inclusion/exclusion criteria

The search strategy was developed in Medline (Ovid). For each search concept, the Medical Subject Headings (MeSH-terms) and free text terms were identified. The search was then translated into the other databases. No language restriction was applied. De-duplication was done using the method described by Bramer et al. [21].

Additionally, we compared DOIs of the identified articles as final step. After the initial searches, we performed a snowball search to check references and citations of eligible studies from the database searches using the tool—connectedpapers.com—and manually verifying the references section from selected articles in this study. A simplified search was also done in Google Scholar where the first 320 records were screened. The last search of articles was conducted 2022–09-27. To account for very recent published works, we conducted a simplified search for new articles published between Oct 2022—Mar 2023. The full search strategies for all databases are available in the Supplementary Material (Table S1—Table S5).

Our objective was to review articles that had developed prediction models for NDDs using ML methods and population-based registers, and EHRs. We included articles using classical ML and deep learning (DL) algorithms for model development. We excluded studies establishing associations between multiple factors and disease conditions. Also, we excluded grouping studies involving the identification of subcategories or comorbidities of NDDs.

Selection of sources of evidence

The articles matching the inclusion criteria from the literature searches were exported into the Zotero citation tool. Thereafter, both authors sorted articles based on title, abstract, and inclusion criteria. The selected articles are further assessed by reviewing the full text of the article to match the inclusion criteria.

Data extraction

To process the selected articles for information needed in the review, we first extracted a table of items from each publication and used them to summarize the study design and findings. These included data items, such as aim, sample size, predictive variables used in the study, method of handling unbalanced data, sparsity, heterogeneity, employed ML methods, their findings, limitations, performance variables and the used datasets.

Results

Our search strategy resulted in 1191 records, from which 302 were removed in the first step as these were meeting abstracts or proceedings and 358 as duplicate records (Fig. 1). After removing 683 from the initial list of 1191, we were left with 508. From this, we excluded records that did not use medical records (n = 93), did not use machine learning (n = 142), review articles (n = 29), no NDD (n = 215), records not found (n = 4), and system development/correspondence (n = 2). We further included 9 articles from other sources, 32 of which met the inclusion criteria and were included in this review, as summarized in the PRISMA flow diagram (Fig. 1). The included studies were conducted in 13 different countries. Most studies came from the USA (59%), followed by Denmark (6%). There is a single study each from the UK, Sweden, Germany, Finland, Switzerland, Netherlands, Egypt, Brazil, Israel, Thailand, and Australia (Fig. 2a, Table S6).

Fig. 1
figure 1

PRISMA flow diagram. After screening 1191 articles, 32 were retained for the review

Fig. 2
figure 2

Distribution of reviewed articles across multiple factors. (a) number of articles from different countries, (b) usage of categories of predictor variables in studies, (c) ML algorithms used in studies, and (d) multiple co-occurring NDD conditions in studies

The sample sizes in the studies varied significantly between 50 participants and a dataset of 4.5 million subjects (Table 1). Prediction models were developed for a single diagnosis and a combination of NDDs. Approximately 47% of included studies developed ASD prediction models, 28% focused on ADHD models, and the remaining were individual studies for other NDDs (Fig. 2b, Table 1).

Table 1 Main characteristics of the reviewed articles

There were only three studies [13, 17, 27] that used Swedish and Danish nationwide population-based registers for the model development, and the rest of the studies used EHRs. Of the studies, 82% employed classical ML methods, while the remaining 18% applied DL methods (Fig. 2c, Table S7). The majority of the studies (60%) were published in the last three years.

The comorbid medical conditions, sociodemographics, and parent medical history were the most commonly used predictor variables (Fig. 2d, Table S8). In terms of performance results, the AUC metric was greater than 75% in most studies, while the sensitivity metric was less than 75% (Table 1, Table S9), indicating a need for further advancements. In the studies reviewed here, we did not find clear association between larger sample sizes and higher performance.

Input preprocessing for addressing data quality issues

The data used in studies were from population-based registers, EHRs or medical records (diseases, medications, lab tests, procedures, clinical notes) of patients and family members, and insurance claim forms which all are characterized by missing values, the high dimensionality of records, heterogeneity, imbalance case–control categories, errors, and systematic biases. Data imputation is one of the commonly employed methods for handling missing data. In the reviewed studies, imputation was done in different ways, including using the random forest-based methods to impute the values [24], populating the missing data with the average value for continuous variables and mode value for discrete variables [36], replacing missing values either with zeros or unknown status values [42] and use chain equations from remaining predictors to fill missing values. Mikolas and colleagues filtered the features and participants with more than 20% of missing values [35], and Garcia-Argibay, et al. [17] included features with less than 10% missing values.

The case–control class/category imbalance problem was addressed using different tools, including downsampling the number of controls [27] or upsampling the number of cases [32] by randomly generating new samples between each positive sample and its nearest neighbors, employing the Synthetic Minority Over-sampling Technique (SMOTE) method to increase the cases [17, 24] or to assign more weights to cases while training a model [34].

There are no standard ways to process multimodal data from population-based registers and EHRs for generating an effective representation. Schuler et al. [50] proposed a generalized low-rank modeling framework to form efficient representations. This low-rank modeling framework was further used for downstream clustering applications. The advent and success of deep neural networks for efficient representation learning have led to the emergence of new studies to form efficient representations for EHRs data. Landi et al. [51] have proposed a representation learning model using word embeddings, convolutional neural networks, and autoencoders to transform patient history in EHRs into a set of low-dimensional vectors. They evaluated the generated representation for patient stratification across various conditions using clustering methods and demonstrated the effectiveness of the representation. Miotto et al. [52] have applied an unsupervised DL method using denoising autoencoders to generate a representation—Deep Patient—for each patient record in EHRs. They demonstrated the effectiveness of this representation by developing risk prediction models for various diseases for patients. Most studies included in this review have used custom methods to convert the input records into a multidimensional numerical vector, while some have grouped certain factors into features and used them as input to the model.

Prediction features and ML methods

The predictor variables used in included studies were comorbid medical conditions from ICD-9 and ICD-10 codes, health problems, medical screening data, prescribed medications of a child, parental medical histories, medications, extended family history of mental and non-mental health conditions, socio-demographics, hospital admission/discharge, outpatient visit events, clinical notes and medical claims (Fig 2d, Table S8).

The predictor variables or features were processed multiple ways to generate a unique numerical representation for each subject before training an ML model. For example, Onishchenko et al. [22] developed digital biomarkers for ASD from past medical conditions. Autism comorbid risk score (ACoR) was estimated in the early years of a child with ASD comorbidities. The score was further conditioned on current screening scores to reduce the false positive rate. A diagnostic history model using time-series patient data across 17 disease categories was developed for each patient. Chen et al. [31] developed an ASD prediction model for young children at 18 mo, 24 mo, and 30 mo using medical claims data. They have examined all diagnosis and procedure codes of a child's medical encounters and used Clinical Classifications Software (CCS) software to form different disease categories. The total number of encounters for each CCS category, sex, and encounters of emergency department visits were used as predictor variables for the model. Ejlskov et al. [27] examined the feasibility of using extended family history of mental and non-mental conditions to predict the ASD risk. A large national Denmark cohort of medical history data of three generations of family members is used for developing ML models. Morbidity indicators across 73 disorders, including mental, cardiometabolic, neurologic, congenital defects, autoimmune, and asthma of family members, were used as predictor variables. Allesøe et al. [13] have employed a DL model for mental disorder prediction, including NDDs using nationwide register data, family and patient medical histories, birth-related measurements, and genetics.

Most studies (82%) trained one or more ASD prediction models using classical ML methods. The most commonly used methods were logistic regression and random forests. Onishchenko et al. [22] used Sequence Likelihood Defect (SLD) to measure the deviations in the observed time-series diagnostic events across positive and control cohorts. It was shown that this approach resulted in better performance than state-of-the-art ML algorithms. The proposed approach has few model parameters to learn, unlike conventional deep neural networks with a large set of parameters. Yuan et al. [32] developed an ASD prediction model from medical claim forms. The medical claim forms are preprocessed to extract textual content, and natural processing techniques were applied to derive text features. These features were used to build a Support Vector Machine (SVM) classifier for ASD prediction. There were nine studies utilizing DL methods. Tran et al. [40] investigated the feasibility of using a short textual description of clinical notes to predict the risk of future multiple mental conditions, including ADHD. The baseline model was developed using the SVM classifier and compared with deep network models, such as Convolutional Neural Networks (CNN) and Recurrent Neural Networks with Hierarchical Attention (ReHAN). The detailed list of ML algorithms used in included studies is presented in Table S7. Many studies have developed an ensemble of models using multiple ML algorithms.

The model evaluation methods used in studies involved k-fold cross-validation techniques and/or train-valid-test split methods. The performance of the algorithms was reported using the commonly used metrics, such as sensitivity, specificity, the area under the curve (AUC), positive predictive value (PPV), and accuracy. Half of the studies have reported AUC performance greater than 75%. While most studies have reported superior performance in some metrics, the sensitivity results were relatively low, as the majority showed a sensitivity of less than 75%. Not all studies have reported performance results across all evaluation metrics in a consistent way. The summary of performance evaluation metrics reported in studies is shown in Table 1, Table S9.

ML model interpretability for feature importances and generalization

Model interpretability is key to healthcare problems as it helps identify influencing variables for decision-making needed by clinicians. It is worth mentioning that most of the studies in this review have addressed the interpretability aspects in sufficient detail. The findings from the included studies and the dominating predictor variables influencing model performance varied across studies. For example, for ASD models, Betts et al. [16] identified gender, maternal age at birth, delivery analgesia, maternal prenatal tobacco disorders, and low 5-min APGAR score as dominant risk factors for ASD. Rahman et al. [24] found that the features derived from predictor variables, such as parental age and parental medications, contributed to a better ML model performance. These predictors agree with prior studies. However, they noted that the performance metrics varied across different ML models, with no one clear model outperforming all metrics. Ejlskov et al. [27] found that the best-performing ML model—extreme gradient boosting (XGB)—identified indicators across mental conditions (ASD, ADHD, neurotic/stress disorders) and non-mental conditions (obesity, hypertension, and asthma) of family members. The study concluded that a comprehensive family history of mental and non-mental conditions could better predict ASD than considering only the immediate family history of ASD. Hassan et al. [25] aim was to identify etiological factors of ASD using subject and family medical histories. Among the 81 family history attributes, six of them—father anxiety, sibling PDD-NOS, father autism disorder, sibling learning disability, father development delay, and mother autism disorder—were highly predictive of ASD. The attributes from the subject medical history that were highly predictive of ASD were atypical language development, age at 3-word sentences, age of first words, disrupted sleep patterns, dietary, gastrointestinal problems, allergies, low birth weight, and ADHD. Gender comparisons highlighted unique and overlapping conditions. One of the significant findings from this study was that parental and sibling developmental delays were strongly associated with ASD. Chen et al. [31] found that for prediction at ages 24 months and 30 months, 30–40 predictor variables were sufficient to achieve stable prediction performance, whereas, for early prediction at 18 months, the model needed 50 predictor variables. For prediction at age 24 months, the identified important variables included sex, developmental and nervous system disorders, psychological and psychiatric services, respiratory system infections and symptoms, gastrointestinal-related diagnosis, ear and eye infections, perinatal conditions, and emergency department visits. Lerthattasilp et al. [33] developed a logistic regression-based ASD prediction model. The experiments identified delayed speech, a history of avoiding eye contact, a history of not showing objects to others, poor response when the clinician draws attention, and low frequency of social interaction as the influencing predictor variables.

Garcia-Argibay et al. [17] developed an ADHD prediction model from population-based Swedish national registers. They found that parents' criminal history, male sex, relative with ADHD, number of academic subjects failed, and speech/learning disabilities were the top features contributing to the model performance. Shi et al. [34] developed ADHD and LD prediction models. The main findings from the study were that complex ML models using ICD-9 codes perform well in ADHD identification. However, they did not offer significant differences compared to using a simple model with a single family of ICD9 codes for ADHD. For LD identification, the utility of clinical diagnostic codes was limited. Mikolas et al. [35] developed a predictive model to detect individuals with ADHD comorbid with psychiatric conditions in a population. The findings from the study were: (a) age, gender, and accuracy/reaction time were more critical than other features, and (b) The ADHD core symptoms reported by parents/teachers did not carry the degree of importance as commonly assumed. Instead, combining symptoms across different domains had strong predictive power for ADHD diagnosis. Elujide et al. [38] developed an ADHD prediction model and found that the factors influencing the model were sex, age, occupation, and marital status. van Dokkum et al. [49] developed a prediction model of development delay at age 4. The perinatal, parental, and child growth milestones of 1st two years were used as predictor variables. They found that sex, maternal educational level, pre-existing maternal obesity, smiling, speaking 2 to 3-word sentences, standing, and BMI z score at one year were features of high importance for the prediction model. Allesøe et al. [13] developed a cross-diagnostic mental disorder diagnosis prediction model and found that previous mental disorders and age were the most important predictors for multi-diagnostic prediction. In summary, the most common predictor categories of NDDs across studies were patient and familial medical history and sociodemographic factors. The specific predictor variables in these categories vary across studies, making it harder to draw more detailed conclusions.

Most studies have used cross-validation and train-validation-test techniques to report model performance. While these methods provide sufficient information about the model performance in a single dataset, model generalizability can be validated using multiple independent datasets across sites and populations. For example, Onishchenko et al. [22] have used data from the Truven dataset for training models and an independent UCM database for validation. Lingren et al. [29] have used cohorts from Boston Children's Hospital, Cincinnati Children's Hospital and Medical Center, The Children's Hospital of Philadelphia, and the Vanderbilt University Medical Center in their model validation. The ADHD prediction model developed by Caye et al. [37] was validated using three external birth cohorts. Koivu et al. [42] developed a Down Syndrome prediction model using datasets from Canada and the UK for training a model and validating the model using an independent dataset from Canada. In summary, studies utilizing cohorts from distinct populations and sites for model validation for generalizability are emerging. The results of individual studies are summarized in Table 1.

Limitations in the included studies

While all studies reported superior performance of proposed individual models, there were not many performance comparisons across studies. There is some overlap of influencing predictive variables across studies; however, the variations in experimental conditions and target sample populations make it harder to form conclusive evidence. For instance, the sample set size varies from fifty participants to millions across studies. The risk of bias, either due to sex, gender, the proportion of cases vs. controls, target site, and populations, was not sufficiently discussed in most studies. More studies following standardized protocols and using common data are critical for reproducible research and moving toward the clinical utility of such models [53].

Discussion

We reviewed the current status of the application of ML to develop prediction models for NDDs using population-based registers and EHRs. More studies started emerging in the last few years. More recently, Engelhard et al. [15] proposed an early autism prediction model from the EHRs data collected before age one year and showed its promise for integration with other screening tools. The predictive models of NDDs show promising results and open up possibilities for further detailed studies for model development and clinical adaptation. Open research questions and challenges need to be addressed before we see clinical translational value from such ML models.

Our analysis of the findings from the studies presented in this work reveals several key insights. The studies that achieved better performance metrics (AUC > 0.8) employed deep neural networks, classical tree based ML models and used subject-wise representation from past medical history for model training. This observation holds across prediction models for ASD, ADHD, and other neurodevelopmental disorders. For studies with limited dataset sizes and noticeable class imbalances, the strategy to balance dataset sizes between cases and controls using methods, such as SMOTE, resulted in improved performance [17, 24]. While only a few studies leveraged population-based registers, they consistently showcased commendable ML predictive performance, underscoring the value of these registers in predicting NDDs.

Many studies use simple regression models to derive associations between predictor variables and outcomes. Such classical statistical-based association studies often were limited by their ability to operate on a handful of predictor variables and challenges with dealing with non-linear relationships. The outcomes from association studies have less clinical translational value than efficient predictive models on large datasets with tens and thousands of predictor variables [54]. Prediction models using DL methods show great promise toward this goal [55]. One of the key initial steps in using deep neural networks working with high-dimensional, longitudinal population-based registers and EHRs is to efficiently learn a representation capturing the complex non-linear inter-relationships present in the data. There is good progress in successfully designing restricted Boltzmann machines, deep convolutional networks, recurrent neural networks, attention-based transformer networks, variational autoencoders, and deep feed-forward networks for effective representation learning [56,57,58]. These advancements can be effectively utilized for building better prediction models. We noticed from our review here, that the tree based classical machine learning models can also be a good substitute when large sample sizes are available.

While most studies reviewed in this work used curated private EHRs datasets, only a few studies leverage population-based registers; the private datasets prevent open and reproducible research, and the results reported are difficult to replicate due to a lack of dataset availability, limiting scientific progress. There are efforts to collect, curate, anonymize, and publicly share population-based registers and EHRs datasets to advance the research area [59,60,61]. In addition, standard protocols must be employed across cohorts, and a standard set of measures must be collected.

The methods, tools, and protocols used to collect population-based registers and EHRs vary significantly across sites and populations. Also, clinicians' subjectivity in assigning disease codes to individuals leads to more inconsistency of records across sites and populations [3, 13]. These problems result in large datasets having missing and inconsistent data for a specific set of individuals. To address these challenges, imputation techniques and representation learning methods are used to preprocess data before training ML models. It is necessary to have well-balanced datasets across cases and controls to train such models. This will not be the case with medical records, as the proportion of cases will be significantly less than controls. However, a significant limitation in studies utilizing EHR is the missingness of data and how it is imputed before developing the models. Missingness in EHR data can arise from various sources, including inconsistent data entry, differences in healthcare practices, and patient non-compliance. Addressing the reasons for missing data is crucial and it was not done in standard manner in the reviewed articles. If the reasons are not understood, an improper imputation can introduce biases and affect the validity of the ML models' predictions. Further work is needed in this space to bring consistency in overcoming these challenges. ML model development and reporting results using standard guidelines, such as TRIPOD [62], GREMLIN [63] and STROBE [64] are essential.

Specific to ML model development for healthcare, sample sizes, hyperparameter tuning, overfitting, model complexity, model interpretability, and generalization are key challenges to be addressed carefully [55]. Recently, de Hond et al. [53] reviewed the current guidelines and quality criteria used in artificial intelligence-based prediction models in healthcare and reported a lack of stricter guidelines for three phases of the predictive model pipeline. Based on our review, we note that not all studies have sufficiently discussed these training considerations. Also, the reporting of performance results varies across studies. Recently, Baker, et al. [65] observed similar varying results across studies in their systematic review on the applicability of machine learning in NDD prediction and understanding. One of the commonly used metrics to report the model performance is the AUC.

Translating these ML models into real-world implementations involves completing three steps [66]. Once the model has been successfully validated with in-house datasets, it needs to undergo external validations, i.e., testing the model performance across different sites, datasets, and populations for its generalizability, followed by implementation trials. Model interpretability is key to explaining the rationale for the model's decision to assist clinicians in making informed decisions. Currently, not many DL studies have focused on addressing the interpretability aspects of the model in the context of NDDs. In summary, more work is needed to implement ML models utilizing population-based registers and EHRs for NDDs. Future studies should focus on enabling public data availability for reproducible research, more standardization in data representation, experiment conditions, and performance reporting. The advancements in DL approaches and explainable AI should be explored for better performance and model interpretability.

Conclusion

In conclusion, this review thoroughly analyzes prior studies leveraging population-based registers and EHRs for ML-driven prediction of NDDs. Our systematic exploration spanned data preprocessing, model selection, and evaluation stages. While most studies employed classical ML techniques, only a handful adopted deep learning. Nearly all delved into model interpretability, spotlighting key influencing factors. Notable predictive variables from registers and EHRs include patient and family medical history and sociodemographic details. However, model sensitivity generally lagged behind other metrics. Challenges like data sparsity, inconsistent disease coding, and heterogeneity often went unaddressed before model creation. Importantly, the majority of the data utilized was private.

A limited number of studies use these registers and EHRs for ML in NDDs, but these are emerging. Differing datasets and experimental setups hinder direct performance comparisons between them. Future work should prioritize data availability, standardize data representation and performance metrics, and explore advancements in deep learning and explainable AI for enhanced outcomes and interpretability.

Data availability

All collected data from article is presented in the main and supplemental material.

References

  1. Association AP. Diagnostic and statistical manual of mental disorders (5th ed., text rev.) (2022). https://doiorg.publicaciones.saludcastillayleon.es/10.1176/appi.books.9780890425787.

  2. Cortese S, et al. Candidate diagnostic biomarkers for neurodevelopmental disorders in children and adolescents: a systematic review. World Psychiatry Off J World Psychiatric Assoc (WPA). 2023;22(1):129–49. https://doiorg.publicaciones.saludcastillayleon.es/10.1002/wps.21037.

    Article  Google Scholar 

  3. Uddin M, Wang Y, WoodburySmith, M. Artificial intelligence for precision medicine in neurodevelopmental disorders. npj Digit Med 2019 2(1);1–10. https://www.nature.com/articles/s41746-19-0191-0. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41746-019-0191-0. Number: 1 Publisher: Nature Publishing Group.

  4. Ismail FY, Fatemi A, Johnston MV. Cerebral plasticity: Windows of opportunity in the developing brain. Eur J Paediatr Neurol EJPN Off J Eur Paediatr Neurol Soc. 2017;21(1):23–48. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ejpn.2016.07.007.

    Article  Google Scholar 

  5. Hastie T, Tibshirani R, Friedman J. The elements of statistical learning springer series in statistics (Springer, New York, NY, 2009). http://link.springer.com/10.1007/978-0-387-84858-7.

  6. Esteva A, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017 542(7639);115-118. https://www.nature.com/articles/nature21056.10.1038/nature21056. Number: 7639 Publisher: Nature Publishing Group.

  7. Ehteshami Bejnordi B, et al. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. JAMA. 2017;318(22):2199–210. https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jama.2017.14585.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Poplin R, et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nature Biomedical Engineering 2018 2(3);158-164. https://www.nature.com/articles/s41551-018-0195-0. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41551-018-0195-0. Number: 3 Publisher: Nature Publishing Group.

  9. Smoller JW. The Use of Electronic Health Records for Psychiatric Phenotyping and Genomics. American journal of medical genetics. Part B, Neuropsychiatric genetics: the official publication of the International Society of Psychiatric Genetics 2018 177(7);601-612. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6440216/.https://doiorg.publicaciones.saludcastillayleon.es/10.1002/ajmg.b.32548 .

  10. Szklo M. Population-based cohort studies. Epidemiol Rev. 1998;20(1):81–90. https://doiorg.publicaciones.saludcastillayleon.es/10.1093/oxfordjournals.epirev.a017974.

    Article  CAS  PubMed  Google Scholar 

  11. Canova C, Cantarutti A. PopulationBased Birth Cohort Studies in Epidemiology. International Journal of Environmental Research and Public Health 2020 17(15);5276. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7432312/. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/ijerph17155276.

  12. Allesøe RL, et al. Deep learning–based integration of genetics with registry data for stratification of schizophrenia and depression. Science Advances 2022 8(26);eabi7293. https://www.science.org/doi/full/10.1126/sciadv.abi7293. https://doiorg.publicaciones.saludcastillayleon.es/10.1126/sciadv.abi7293. Publisher: American Association for the Advancement of Science.

  13. Allesøe RL, et al. Deep learning for cross-diagnostic prediction of mental disorder diagnosis and prognosis using Danish Nationwide Register and Genetic Data. JAMA Psychiatry (2022). https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jamapsychiatry.2022.4076

  14. Song J, et al. The impact of educational attainment, intelligence and intellectual disability on schizophrenia: a Swedish population-based register and genetic study. Mol Psychiatry 2022 27(5);2439-2447. https://www.nature.com/articles/s41380-022-01500-2. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41380-022-01500-2. Number: 5 Publisher: Nature Publishing Group.

  15. Engelhard MM, et al. Predictive value of early autism detection models based on electronic health record data collected before age 1 year. JAMA Network Open. 2023;6(2):e2254303. https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jamanetworkopen.2022.54303.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Betts KS, Chai K, Kisely S, Alati R. Development and validation of a machine learning-based tool to predict autism among children. Autism Research n/a (n/a) (2023). https://onlinelibrary.wiley.com/doi/abs/10.1002/aur.2912.10.1002/aur.2912. Eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/aur.291

  17. Garcia-Argibay M, et al. Predicting childhood and adolescent attentiondeficit/hyperactivity disorder onset: a nationwide deep learning approach. Mol Psychiatry 2023 28(3);1232-1239. https://www.nature.com/articles/s41380-022-01918-8. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41380-022-01918-8. Number: 3 Publisher: Nature Publishing Group.

  18. Tricco AC, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Internal Med 2018 169(7);467-473. https://www.acpjournals.org/doi/10.7326/M18-0850. https://doiorg.publicaciones.saludcastillayleon.es/10.7326/M18-0850. Publisher: American College of Physicians

  19. Page MJ, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ 2021 372;n71. https://www.bmj.com/content/372/bmj.n71. https://doiorg.publicaciones.saludcastillayleon.es/10.1136/bmj.n71. Publisher: British Medical Journal Publishing Group Section: Research Methods and; Reporting.

  20. Haddaway NR, Page MJ, Pritchard CC, McGuinness LA. PRISMA2020: An R package and Shiny app for producing PRISMA 2020-compliant flow diagrams, with interactivity for optimised digital transparency and Open Synthesis. Campbell Syst Rev. 2022;18(2):e1230. https://doiorg.publicaciones.saludcastillayleon.es/10.1002/cl2.1230.10.1002/cl2.1230. (Publisher: John Wiley & Sons, Ltd).

    Article  PubMed  PubMed Central  Google Scholar 

  21. Bramer WM, Giustini D, de Jonge GB, Holland L, Bekhuis T. De-duplication of database search results for systematic reviews in EndNote. J Med Library Association: JMLA 2016 104(3);240–243. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4915647/. https://doiorg.publicaciones.saludcastillayleon.es/10.3163/1536-5050.104.3.014.

  22. Onishchenko D, et al. Reduced false positives in autism screening via digital biomarkers inferred from deep comorbidity patterns. Sci Adv 2021 7(41);eabf0354. https://www.science.org/doi/10.1126/sciadv.abf0354. https://doiorg.publicaciones.saludcastillayleon.es/10.1126/sciadv.abf0354. Publisher: American Association for the Advancement of Science.

  23. Bishop-Fitzpatrick L, et al. Using machine learning to identify patterns of lifetime health problems in decedents with autism spectrum disorder. Autism Res. 2018;11(8):1120–8. https://doiorg.publicaciones.saludcastillayleon.es/10.1002/aur.1960.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Rahman R, et al. Identification of newborns at risk for autism using electronic medical records and machine learning. Eur Psychiatry. 2020;63(1):e22. https://doiorg.publicaciones.saludcastillayleon.es/10.1192/j.eurpsy.2020.17.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Hassan MM, Mokhtar HMO. Investigating autism etiology and heterogeneity by decision tree algorithm. Inform Med Unlocked 2019 16;100215. https://www.sciencedirect.com/science/article/pii/S2352914819300541. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.imu.2019.100215.

  26. Maenner MJ, Yeargin-Allsopp M, Braun KVN, Christensen DL, Schieve LA. Development of a machine learning algorithm for the surveillance of autism spectrum disorder. Plos One 2016 11(12);e0168224. https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0168224. https://doiorg.publicaciones.saludcastillayleon.es/10.1371/journal.pone.0168224. Publisher: Public Library of Science.

  27. Ejlskov L, et al. Prediction of autism risk from family medical history data using machine learning: a national cohort study from Denmark. Biological Psychiatry Global Open Science 2021 1(2);156-164. https://www.sciencedirect.com/science/article/pii/S2667174321000136. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bpsgos.2021.04.007.

  28. Alexeeff SE, et al. Medical conditions in the first years of life associated with future diagnosis of ASD in children. J Autism Dev Disord. 2017;47(7):2067–79. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s10803-017-3130-4.

    Article  PubMed  PubMed Central  Google Scholar 

  29. Lingren T, et al. Electronic health record based algorithm to identify patients with autism spectrum disorder. PLoS One. 2016;11(7):e0159621. https://doiorg.publicaciones.saludcastillayleon.es/10.1371/journal.pone.0159621.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  30. Leroy G, et al. Automated extraction of diagnostic criteria from electronic health records for autism spectrum disorders: development, evaluation, and application. J Med Internet Res. 2018;20(11):e10497. https://doiorg.publicaciones.saludcastillayleon.es/10.2196/10497.

    Article  PubMed  PubMed Central  Google Scholar 

  31. Chen YH, Chen QS, Kong L, Liu GD. Early detection of autism spectrum disorder in young children with machine learning using medical claims data. Bmj Health Care Inform 2022 29(1). https://doiorg.publicaciones.saludcastillayleon.es/10.1136/bmjhci-2022-100544.

  32. Yuan J, Holtz C, Smith T, Luo J. Autism spectrum disorder detection from semistructured and unstructured medical data. EURASIP J Bioinform Syst Biol 2017 2017;3. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5288414/. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13637-017-0057-1.

  33. Lerthattasilp T, Tanprasertkul C, Chunsuwan I. Development of clinical prediction rule for diagnosis of autistic spectrum disorder in children. Mental Illness 2020 12(1);7-16. https://ovidsp.ovid.com/ovidweb.cgi?T=JS&CSC=Y&NEWS=N&PAGE=fulltext&D=pmnm&AN=32742626. https://doiorg.publicaciones.saludcastillayleon.es/10.1108/MIJ-01-2020-0001.

  34. Shi Y, et al. Utility of medical record diagnostic codes to ascertain attentiondeficit/hyperactivity disorder and learning disabilities in populations of children. BMC Pediatr. 2020;20(1):510. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12887-020-02411-3.

    Article  PubMed  PubMed Central  Google Scholar 

  35. Mikolas P, et al. Training a machine learning classifier to identify ADHD based on real-world clinical data from medical records. Scient Rep. 2022;12(1):12934. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-022-17126-x.

    Article  CAS  Google Scholar 

  36. Chen T, Antoniou G, Adamou M, Tachmazidis I, Su P. Automatic diagnosis of attention deficit hyperactivity disorder using machine learning. Appl Artif Intell. 2021;35(9):657–69. https://doiorg.publicaciones.saludcastillayleon.es/10.1080/08839514.2021.1933761. (Taylor & Francis eprint:).

    Article  Google Scholar 

  37. Caye A, et al. A risk calculator to predict adult attention-deficit/hyperactivity disorder: generation and external validation in three birth cohorts and one clinical sample. Epidemiol Psychiatr Sci. 2019;29:e37. https://doiorg.publicaciones.saludcastillayleon.es/10.1017/S2045796019000283.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  38. Elujide I, et al. Application of deep and machine learning techniques for multi-label classification performance on psychotic disorder diseases. Informatics in Medicine Unlocked 2021 23. https://www.embase.com/search/results?subaction=viewrecord&id=L2011532322&from=export. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.imu.2021.100545.

  39. Morris SM, et al. Predictive modeling for clinical features associated with neurofibromatosis type 1. Neurol Clin Pract 2021 11(6);497-505. https://ovidsp.ovid.com/ovidweb.cgi?T=JS&CSC=Y&NEWS=N&PAGE=fulltext&D=pmnm&AN=34987881. https://doiorg.publicaciones.saludcastillayleon.es/10.1212/CPJ.0000000000001089 .

  40. Tran T, Kavuluru R. Predicting mental conditions based on "history of present illness" in psychiatric notes with deep neural networks. J Biomed Inform 2017 75;S138-S148. https://www.embase.com/search/results?subaction=viewrecord&id=L617182454&from=export. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jbi.2017.06.010.

  41. Movaghar A, et al. Artificial intelligence–assisted phenotype discovery of fragile X syndrome in a population-based sample. Genetics in Medicine 2021 23(7);1273-1280. https://linkinghub.elsevier.com/retrieve/pii/S1098360021050450. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41436-021-01144-7.

  42. Koivu A, Korpim¨aki T, Kivel¨aP, Pahikkala T, Sairanen M. Evaluation of machine learning algorithms for improved risk assessment for Down's syndrome. Comp Biol Med 2018 98;1-7. https://linkinghub.elsevier.com/retrieve/pii/S0010482518301112. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2018.05.004.

  43. Gui L, et al. Longitudinal study of neonatal brain tissue volumes in preterm infants and their ability to predict neurodevelopmental outcome. NeuroImage 2019 185;728-741. http://proxy.kib.ki.se/login?url=https://ebscohostsearch.publicaciones.saludcastillayleon.es/login.aspx?direct=true&db=psyh&AN=2018-29772-001&site=ehost-livecristina.borradoritolsa@hcuge.ch. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.neuroimage.2018.06.034.

  44. Randolph DA. et al. Outcomes of extremely low birth weight infants with acidosis at birth. Archives of disease in childhood. Fetal Neonatal edition 2014 99(4);F263-F268. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4274605/. https://doiorg.publicaciones.saludcastillayleon.es/10.1136/archdischild-2013-304179

  45. Hill DL, et al. Identifying communication impaired pediatric patients using detailed hospital administrative data. Hosp Pediatr 2016 6(8);456-467. https://ovidsp.ovid.com/ovidweb.cgi?T=JS&CSC=Y&NEWS=N&PAGE=fulltext&D=med13&AN=27381628. https://doiorg.publicaciones.saludcastillayleon.es/10.1542/hpeds.2015-0154.

  46. Pruett DG, et al. Identifying developmental stuttering and associated comorbidities in electronic health records and creating a phenome risk classifier. J Fluency Disord 2021 68(6);105847. https://ovidsp.ovid.com/ovidweb.cgi?T=JS&CSC=Y&NEWS=N&PAGE=fulltext&D=med19&AN=33894541. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jfludis.2021.105847.

  47. Shaw DM, et al. Phenome risk classification enables phenotypic imputation and gene discovery in developmental stuttering. Am J Hum Genet 2021 108(12);2271-2283. https://www.embase.com/search/results?subaction=viewrecord&id=L2015735761&from=export.https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ajhg.2021.11.004.

  48. Shrot S, et al. Prediction of tuberous sclerosis-associated neurocognitive disorders and seizures via machine learning of structural magnetic resonance imaging. Neuroradiology. 2022;64(3):611–20. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s00234-021-02789-6.

    Article  PubMed  Google Scholar 

  49. van Dokkum NH, Reijneveld SA, Heymans MW, Bos AF, de Kroon MLA. Development of a prediction model to identify children at risk of future developmental delay at age 4 in a population-based setting. Int J Environ Res Public Health 2020 17(22);8341. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7698029/. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/ijerph17228341.

  50. Schuler A, et al. Discovering patient phenotypes using gener-alized low rank models. Pacific Symposium Biocomputing 2016 21;144. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4836913/.

  51. Landi I, et al. Deep representation learning of electronic health records to unlock patient stratification at scale. npj Digit Med 2020 3(1). https://www.embase.com/search/results?subaction=viewrecord&id=L2005603948&from=export. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41746-020-0301-z.

  52. Miotto R, Li L, Kidd BA, Dudley JT. Deep patient: an unsupervised representation to predict the future of patients from the electronic health records. Scient Rep 2016 6(1);26094. https://www.nature.com/articles/srep26094. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/srep26094. Number: 1 Publisher: Nature Publishing Group.

  53. de Hond AAH, et al. Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. npj Digital Medicine 2022 5(1);1-13. https://www.nature.com/articles/s41746-021-00549-7. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41746-021-00549-7. Number: 1 Publisher: Nature Publishing Group.

  54. Bzdok D, Varoquaux G, Steyerberg EW. Prediction, not association, paves the road to precision medicine. JAMA Psychiatry. 2021;78(2):127–8. https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jamapsychiatry.2020.2549.

    Article  PubMed  Google Scholar 

  55. Koppe G, Meyer-Lindenberg A, Durstewitz D. Deep learning for small and big data in psychiatry. Neuropsychopharmacology 2021 46(1);176-190. https://www.nature.com/articles/s41386-020-0767-z. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41386-020-0767-z. Number: 1 Publisher: Nature Publishing Group.

  56. Si Y, et al. Deep representation learning of patient data from Electronic Health Records (EHR): A systematic review. J Biomed Inform. 2021;115:103671. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jbi.2020.103671.

    Article  PubMed  Google Scholar 

  57. Chen P. Representation Learning for Electronic Health Records: A Survey. J Phys Confer Ser 2020 1487(1);012015. https://iopscience.iop.org/article/10.1088/1742-6596/1487/1/012015. https://doiorg.publicaciones.saludcastillayleon.es/10.1088/1742-6596/1487/1/012015.

  58. Ayala Solares JR, et al. Deep learning for electronic health records: A comparative review of multiple deep neural architectures. J Biomed Inform 2020 101;103337. https://www.sciencedirect.com/science/article/pii/S1532046419302564. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jbi.2019.103337.

  59. Clinical Practice Research Datalink | CPRD. URL https://www.cprd.com/node/120.

  60. Feliciano P, et al. SPARK: A US Cohort of 50,000 Families to accelerate autism research. Neuron 2018 97(3);488-493. https://www.cell.com/neuron/abstract/S0896-6273(18)30018-7. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.neuron.2018.01.015. Publisher: Elsevier.

  61. Sudlow C, et al. UK biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS Med. 2015;12(3):e1001779. https://doiorg.publicaciones.saludcastillayleon.es/10.1371/journal.pmed.1001779.

    Article  PubMed  PubMed Central  Google Scholar 

  62. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement. Ann Intern Med. 2015;162(1):55–63. https://doiorg.publicaciones.saludcastillayleon.es/10.7326/M14-0697.

    Article  PubMed  Google Scholar 

  63. Quinn TP, et al. Signal from noise: using machine learning to distil knowledge from data in biological psychiatry (2022). https://psyarxiv.com/dz7gt/.

  64. von Elm E, et al. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement: guidelines for reporting observational studies. Int J Surg (London, England). 2014;12(12):1495–9. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ijsu.2014.07.013.

    Article  Google Scholar 

  65. Baker S, Kandasamy Y. Machine learning for understanding and predicting neurodevelopmental outcomes in premature infants: a systematic review. Pediatr Res. 2023;93:293–9. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41390-022-02120-w.

    Article  PubMed  Google Scholar 

  66. Chekroud AM, et al. The promise of machine learning in predicting treatment outcomes in psychiatry. World Psychiatry 2021 20(2);154-170. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8129866/. https://doiorg.publicaciones.saludcastillayleon.es/10.1002/wps.20882 .

Download references

Acknowledgements

We want to thank the University Library of Karolinska Institutet and their information services representative for their support with the search strategy and for identifying the initial set of articles from multiple search databases. We want to thank all Tammimies lab members for their help.

Funding

Open access funding provided by Karolinska Institute. This study was supported by grants from the Swedish Foundation for Strategic Research (FFL18-0104), The Swedish Brain Foundation (FO2021-0073), Strategic Research Area Neuroscience (StratNeuro), and The Swedish Foundation for International Cooperation in Research and Higher Education (STINT).

Author information

Authors and Affiliations

Authors

Contributions

Shyam Sundar Rajagopalan (SR) led the review process, worked with the university library information specialist for the data extraction, screened the articles, analyzed the studies, and wrote the first draft. Kristiina Tammimies (KT) conceived the review, defined the scope, screened the articles, provided feedback on the quality of the review, assisted with writing and finally curated the content.

Corresponding authors

Correspondence to Shyam Sundar Rajagopalan or Kristiina Tammimies.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that there are no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Rajagopalan, S.S., Tammimies, K. Predicting neurodevelopmental disorders using machine learning models and electronic health records – status of the field. J Neurodevelop Disord 16, 63 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s11689-024-09579-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s11689-024-09579-0

Keywords