banner
News center
We promise to provide top-notch goods, timely shipping, and conscientious assistance.

A synthetic data generation system for myalgic encephalomyelitis/chronic fatigue syndrome questionnaires

Aug 05, 2023

Scientific Reports volume 13, Article number: 14256 (2023) Cite this article

2 Altmetric

Metrics details

Artificial intelligence or machine-learning-based models have proven useful for better understanding various diseases in all areas of health science. Myalgic Encephalomyelitis or chronic fatigue syndrome (ME/CFS) lacks objective diagnostic tests. Some validated questionnaires are used for diagnosis and assessment of disease progression. The availability of a sufficiently large database of these questionnaires facilitates research into new models that can predict profiles that help to understand the etiology of the disease. A synthetic data generator provides the scientific community with databases that preserve the statistical properties of the original, free of legal restrictions, for use in research and education. The initial databases came from the Vall Hebron Hospital Specialized Unit in Barcelona, Spain. 2522 patients diagnosed with ME/CFS were analyzed. Their answers to questionnaires related to the symptoms of this complex disease were used as training datasets. They have been fed for deep learning algorithms that provide models with high accuracy [0.69–0.81]. The final model requires SF-36 responses and returns responses from HAD, SCL-90R, FIS8, FIS40, and PSQI questionnaires. A highly reliable and easy-to-use synthetic data generator is offered for research and educational use in this disease, for which there is currently no approved treatment.

Myalgic encephalomyelitis, commonly called chronic fatigue syndrome (ME/CFS), is a serious, complex, and chronic multisystem illness of unknown etiology, often triggered by a persistent viral infection (for this reason, it is also known as post-viral fatigue syndrome). ME/CFS affects as many as 17 to 24 million people worldwide, and its prevalence is expected to double by 20301. It is characterized by unexplained and persistent post-exertional fatigue that is not relieved by rest. It is exacerbated by physical and mental exertion and other core symptoms such as cognitive, immunometabolic, autonomic, and neuroendocrine dysfunction2. It produces severe disability in patients, significantly interfering with their work activity and their daily life tasks3. In addition to fatigue, these patients have characteristic inflammatory and muscular symptoms, sleep dysfunction, and altered cognitive functions4. The symptomatic muscle blocks symptoms such as pain, generalized muscle weakness, fatigue after physical exertion, neurological symptoms (sensory hypersensitivity, ataxia, dysmetria, visual disturbances, and motor incoordination), neurocognitive symptoms (alterations in memory, concentration, calculation, task planning). The autonomic block (cephalic instability, dizziness, fainting spells, excessive sweating, orthostatic hypotension, tremor or alterations in intestinal rhythm), immunoinflammatory symptoms (low-grade fever, sore throat, recurrent canker sores, polyarthralgia, morning numbness, infections such as herpes or candida) and deficiency symptoms in the production of cellular metabolic energy. Sleep disturbances have been relevant since their description as their clinical entity. In all versions of the different ME/CFS diagnostic criteria, sleep disorders have played a key role, especially the presence of unrefreshing sleep and the importance of the Pittsburgh Sleep Quality Index (PSQI) questionnaire in the assessment of the severity of alterations in sleep quality and its association with fatigue, pain, psychopathology, and neurovegetative dysfunction5. ME/CFS, together with the symptomatic complexity that it presents, as a consequence of its multisystemic nature, is associated with different comorbid phenomena such as fibromyalgia, sicca syndrome, myofascial syndrome, psychopathology, ligament hyperlaxity, fasciitis plantar, degenerative vertebral disease or mechanical, shoulder tendinopathy, multiple chemical sensitivity, epicondylitis, carpal tunnel syndrome, osteoporosis, hypercholesterolemia, hypertriglyceridemia, vascular risk, endometriosis, thyroiditis, with a higher prevalence than that observed in patients not affected by ME/CFS6.

In the study of ME/CFS, after the diagnosis and assessment of comorbid phenomena, it is essential to quantify and assess fatigue, quality of life, or anxiety/depression psychopathology using a battery of clinically self-administered questionnaires. Today there are few units specialized in ME/CFS in the world, with a relatively low number of duly documented cases and a lack of publicly available data compared with other disorders. Moreover, unfortunately, there are no commercially available diagnostic tests, no specific lab biomarkers, and no targeted FDA-approved drugs for ME/CFS7. Therefore, each subject to be diagnosed with ME/CFS must undergo a Fukuda criteria evaluation and procedure that each unit has established using batteries of validated self-administered questionnaires. As stated before, it is important to evaluate the disabling fatigue perception, sleep problems, and health-related quality of life using self-administered questionnaires such as the fatigue impact scale FIS408 and FIS89, PSQI10, and Short Form Health Survey (SF-36)11, Symptom Checklist-90-revised (SCL 90 R) psychological inventory12, hospital anxiety and depression scale (HAD)13. Ongoing placebo-controlled clinical trials to evaluate the clinical benefits of drugs on ME/CFS symptoms14 have changed some questionnaire scores from baseline to final study as a primary endpoint.

There is no consensus on the number and type of questionnaires that should be carried out, so not all units record the same number per subject. Consequently, it is complex for ME/CFS units to have many records of the questionnaires necessary to efficiently approach large longitudinal and multicenter studies of patients with this pathology using the latest advances in data analysis, such as Machine Learning techniques.

Machine Learning is a particular method of data analytics that automates model building as it relates to the development of models. Over the last years, it has been proven great performance of machine learning supervised algorithms in several clinical applications15 to diagnose and treat diseases. Supervised learning involves training machine learning-based algorithms using labeled input datasets requiring however, to be efficient and get optimal results, a large number of records are needed. The learning occurs by comparing results with the expected outputs to identify errors and change the model's weights to infer knowledge. There are few publications, and all of them very recent, that refers to the application of different machine learning techniques in ME/CFS: seeking a new biomarker16, clustering17, or discovering the relationship between depression and ME/CFS18, using neural networks seeking omic biomarkers16 or neural networks classifiers19. While they all make important steps forward in understanding ME/CSF, the limited sample size makes generalization and translation of their findings to clinical practice or other datasets difficult. Also, as stated before, when there are no clear biomarkers to follow the evolution of the illness, like in ME/CSF, quality-of-life questionnaires are used to measure it14. There are several lines of investigation, such as clustering20 or finding relations between blood measurements with questionnaires data21.

Therefore, there is an increasing demand to access large repositories of high-quality health datasets for better and more reliable predictions from supervised machine learning algorithms. Anonymized electronic health records are bought and sold by insurance22 and clinical groups23. However, they are limited in size or content, might be incomplete, and their applications might be restricted. This problem can be overcome using synthetic datasets coming from simulations24,25. Synthetic datasets are generated to create data for improving the sample size of existing cohorts or filling in the missing values, preserving privacy while keeping the real data characteristics. Synthetic data generators preserve the statistical properties of the original. However, they do not reveal any information regarding real people and offer several benefits, such as overcoming real data usage restrictions of data sharing and patient consent. There is a need for developing synthetic datasets that would complement real-world data for various reasons26: ease of access, cost-efficiency, test-efficiency, patient privacy protection, completeness, and validation capabilities, handling missingness, complex interactions between variables, resulting sensitivity analysis statistics from latest classifiers and graphical modeling and resampling27. A common application of synthetic data generation in medicine is image generation simulating diseases. It helps to test and benchmark the performance and accuracy of different algorithms. Some recent applications are in the simulation of skin lesions28, brain atrophy in aging or Dementia29, generation of PET MRI scans for Alzheimer's disease30, tumor generation in the brain31, or breast cancer32.

This work aims to generate a robust and reliable synthetic data generator for ME/CFS questionnaires to produce high-fidelity and risk-free health care records, enhance existing public and private ME/CFS datasets for investigation and educational use, and are free of legal, privacy, security, and intellectual property restrictions.

This prospective cross-sectional study includes 2,522 subjects diagnosed with ME/CFS from the Vall d’Hebron University Hospital, Barcelona, Spain, 90.5% females (mean age 48.11 ± 10.31 years) and 9.5% males (mean age 44.41 ± 11.35 years). Data for SF-36, HAD, FIS8, FIS40, SCL 90 R, and PSQI questionnaires has been obtained and recorded from 2008 to 2021. See Table 1 for final records. Patients were eligible to participate if they were 18 years, had a confirmed diagnosis of ME/CFS, met the Fukuda33 and Carruthers criteria34, and provided signed written informed consent and ethics committee approval. The data collected were anonymized in a database to which only those designated for the study had access, and in no case was any information known that could reveal or infer the participant’s identity.

Graph theory was used to analyze the relationships between the subscales of each questionnaire. A graph is a collection of nodes (also called vertices) joined together in pairs by edges (undirected) or arcs (directed)35. The graph structure allows us to capture the pattern of interactions between the nodes (individuals or entities). Graph (or network) analysis is used to study relationships between individuals to discover knowledge about global and local structures. The study of structure networks helps to decide the optimal order36.

In this work, the graph nodes are defined as all subscales, and the edges are defined as moderate or strong correlations between nodes (subscales). The linear correlation between two subscales is represented by \(corr\left(i,j\right)\), and Pearson correlation is defined as moderate or strong if \(corr\left(i,j\right)\ge 0.5\) 37 in case of direct correlation. An \(edge\left(i,j\right)\) is defined if \(abs\left(corr\left(i,j\right)\right)\ge 0.5\).

The relation of subscales between each test is related in Table 2. Each subscale has been classified according to the area to which it has been defined and named as the subject. Thirty-eight subscales, six tests, and twelve subjects form the dataset to create the relationship between them to the graph.

The study of the relationships mentioned above should indicate the order to generate our machine learning models. The SF-36 is prevalent and will be used in our model as initial data. The rest of the order will be given by the relationships between the different tests so that those with a stronger relationship are consecutive in the model. The strength of the relationship is measured in terms of the percentage of connections between the test nodes.

Real data of all of six questionnaires are required to train and build the models. First, an input matrix represents the validated answers of a number of patients, where n is the number of validated responses and f the number of questions. That is the first training data. As predicted, it has to be a second questionnaire which the same n and f1 questions. The model must generate a predicted matrix with the same dimension. The next step has as input matrix the initial matrix concatenated with the last predicted matrix and the second questionnaire response matrix for prediction, as shown in Fig. 1.

Modeling schema. The model requires a questionnaire answers matrix as an input value. The output is the other five questionnaires. Each questionnaire has a different number of questions and subscales. A simple sum of the number of questions calculates most subscales. For example, the SF-36 input dimension is n × 36, where n is the number of patients who answered the SF-36 questionnaire. The output is n × 186, where 186 is all five questionnaire answers.

Classification and regression models can be used. The goal is to provide 186 output dimensions that must be calculated step by step. The output is compared with the real data set to validate the model. The results are validated using the t-student test. The strategy is to validate one questionnaire. The next step is concatenating the questionnaire answers matrix as input with the final output with different models. It has tested machine learning and deep learning algorithms step by step. The validation system has measured whether real and synthetic data come from the same populations within t-student statistics. The models tested have been regressors and classifiers. The comparison between XGBoost and Deep Neural Networks (DNN)38,39 shows that both models offer similar performance in structured data.

The F1-score can be interpreted as a harmonic mean of precision and recall, where an F1-score reaches its best value at one and worst score at zero. The relative contribution of precision and recall to the F1-score are equal. The formula for the F1 score is:

and operating,

where TP is the number of true positives, FN is the number of false negatives, and FP is the number of false positives. Better performance means lower FN and FP values, and better precision and recall mean better F1 performance. In imbalanced data, greater accuracy than the F1 score indicates that some labels perform poorly. Recall is defined by the ratio \(recall=\frac{tp}{tp+fn}\)40. Accuracy is defined if \(\left(y,\widehat{y}\right)\) as a (sample, predicted), then the fraction of correct predictions over samples is defined as

Mean error is defined as the ratio of overall value questionnaire predicted versus comprehensive value sample questionnaire. The series for t-student value is defined as the sum of all answers of each questionnaire variable, predicted, and sample data.

The authors declare that the procedures followed were by the regulations of the responsible Clinical Research Ethics Committee and by those of the World Medical Association and the Helsinki Declaration. The research protocols were approved by the Ethics Committee of the Vall d'Hebron University Hospital, the first “Population-based Registry of Patients with Chronic Fatigue Syndrome” approved on 18/10/2006.

In our proposed model, an \(edge\left(i,j\right)\) is defined if \(abs\left(corr\left(i,j\right)\right)\ge 0.5\) which indicates moderate or strong direct and indirect correlation. The 2370 registers were validated, and the Pearson correlation analyzed 38 questionnaire subscales. The subject of each subscale represents networks with each node (9) shown in Fig. 2. Mental, depression, and anxiety are strongly correlated with physical subjects. SF-36 emotional subscales are relational with anxiety, depression, and mental subscales (SCL 90 R and HAD questionnaires). As can be seen, HAD and SCL 90 R are strongly correlated. The node's size is related to the degree of the node, i.e. the number of incident edges.

Subscales relationship graph. Color nodes represent the test to which nodes belong. The percentage in the legend represents the number of nodes versus the total.

In supplementary material, the second network analyzes the subscales as a node and the same relationship as an edge. The SF-36 subscales (green) have strong relationships with HAD (magenta) and FIS8, and FIS40 (strong-green and red, respectively). SCL 90 R (brown) has a strong relationship with HAD. Furthermore, PSQI (blue) has no relationship except the total psqi value. The strength of the relationship is measured in terms of the percentage of relationships between the test nodes. The initial test is SF-36, and its nodes have relationships with 100% of HAD’s nodes (3 of 3) and only 25% of SCL 90 R (4 of 12). HAD’s nodes have a 100% relationship with SCL 90 R’s nodes. SCL 90 R has a relationship with the unique FIS8 node, which has relationships with all four FIS40 nodes. The last test with few relations is PSQI. Consequently, the order decided according to aforementioned relationships is: HAD, SCL 90 R, FIS8, FIS40, and PSQI.

A test comparison between XGBoost, Classifier and XGBoost Regressor using SF-36 as training data and HAD as a target with 2321 validated registers, is provided in supplementary material Fig. 3. The hyperparameter defines how our model works41. The parameters tuned were max_depth, gamma, reg_alpha, reg_lambda, colsample_bytree, min_child_weight, subsample, n_estimators and eta. Hyperopt has been used for hyperparameter tuning41. Both must be trained for each question; therefore 14 models have to be trained. The order on a set predicted value is {0, 1, 2, 3}, and the trained value is {1, 2, 3}, where in both cases, greater values show worse health status. Regressor predicted rounded to compare between real data. The results of the model are analyzed with XGBoost and the regression and classification are compared. The mean regression error is much higher than the classification error (32.50% vs. 3.16%). Therefore, the regression model is discarded in the following analyses (the results are available in the supplementary material, Table S1). Total connections have been 32,494 (2321 registers × 14 questions HAD questionnaire) and “1” and “2” answers are 67.25% of the total. The model tends to reduce the mean error, so the model predicted 70% more “1” than real and rare predicted, “3” (For more information, see Table S3 in the Supplementary Material).

Imbalanced data occur where one or more class labels have a very high number of observations, and the other has a lower one. The main problem is to increase accurate predictions of the minority class. To consider the skewed distribution of classes of different weights, classes with weights result in a penalty and a minor update of the model coefficients. The model based on the Keras library is more flexible, and for each question, it can be considered as the difference of the unbalanced data. The main difference between the Keras classifier model is the usage of the recall value, which helps to reduce the aforementioned problem with imbalanced data (for more information, see Table S3 in the Supplementary Material). For each class, do

where n is the number of valid registers, classes is the number of classes, and \(coun{t}_{i}\) is the support of ith class. Results comparison 1st questions of HAD (for more information, see Table S1–S3 in the Supplementary Material). The answer “0” has 66 (2.8%) support, and the answer “3” has 562 (24.21%) support. Minority-weighted label classes tend to be underrepresented with a low recall rate, 0.00 in the first case. These biases produce worse synthetic quality data for posterior analysis. Table 3 shows the results once corrected by the configuration in our model, improving the results significantly in those responses with low representation.

Building the model needs five steps, as depicted in Fig. 3. The first step requires an SF-36 questionnaire input matrix with 3019 registers which HAD questionnaire had the same. The output is a HAD synthetic matrix. The second step requires an input matrix of SF-36 + HAD (synthetic data) and produces synthetic SCL 90 R responses and so on. The results are detailed in Tables 4, 5.

Keras Classifier algorithm schema.

Given the SF-36 questionnaire data can create using a new model, synthetic responses from other questionnaires inform the impact of fatigue, psychological phenomena, and sleep dysfunction. The lack of risk-free health data is an issue in ME/SFC hospital units and investigators. This open-source project offers a tool to generate risk-free synthetic data for the health IT and clinical community to use, experiment, and create more synthetic data. The quality based on validation tests did not cover projects or research focused on clinical discovery. Synthetic data can be an alternative to ground truth when data access is restricted and an excellent alternative to machine learning training/testing datasets26.

The SF-36 includes one multi-item scale that assesses eight health concepts: (1) limitations in physical activities because of health problems; (2) limitations in social activities because of physical or emotional problems; (3) limitations in usual role activities because of physical health problems; (4) bodily pain; (5) general mental health (psychological distress and well-being); (6) limitations in usual role activities because of emotional problems; (7) vitality (energy and fatigue), and (8) general health perceptions and is one of the most used quality life questionnaires used and evaluated41. The other five questionnaires used in this work complement most information about the quality of life of ME/SFC patients.

The questionnaires can be answered quickly and are regularly available in primary care and specialized medical consultations. Some applications offer automated analyzed results that inform essential information about patient health conditions.

The graph theory has been used to decide the order of the modeling cascade. Although a deeper analysis of these relationships should be the subject of another, more specific work, in this case, it informs us of the order used in our model. These relationships will characterize our model, which will be more robust with more records analyzed. Our dataset is unusually great in SFC, which becomes robust to our models.

Our synthetic dataset generator applications fill in missing data of real datasets from any other five questionnaires. For those, ME/SFC dataset clinical units with SF-36 questionnaire answers but missing others could build a complete dataset.

(1) Single-center trial. (2) Unit of reference in diagnosing and treating CFS/ME, which may be biased towards more severe cases and a longer evolution time than studies in primary care. (3) No information is available on parameters such as the results of the two-day ergometric test for assessing exercise intolerance, a neuropsychological battery for assessing cognitive impairment, and neurovegetative dysfunction, e.g., heart rate variability. (4) That this is a prospective study with cross-sectional data collection. It is not a longitudinal study.

Synthetic patients can be simulated with models of ME/CFS questionnaires data and corresponding standards of care to produce risk-free realistic synthetic healthcare records at scale. An open-source generator offers high-fidelity synthetic data for investigation and educational use, free of legal, privacy, security, and intellectual property restrictions.

GitHub is an online platform where researchers and software developers share their work with the scientific community. The following link shares the work described here. The datasets generated and/or analyzed during the current study are available in the SFCSyntheticDataGenerator repository, https://github.com/mlacasa/SFCSyntheticDataGenerator

Lim, E.-J. et al. Systematic review and meta-analysis of the prevalence of chronic fatigue syndrome/myalgic encephalomyelitis (CFS/ME). J. Transl. Med. 18, 100. https://doi.org/10.1186/s12967-020-02269-0 (2020).

Article PubMed PubMed Central Google Scholar

Morris, G. et al. Myalgic encephalomyelitis/chronic fatigue syndrome: From pathophysiological insights to novel therapeutic opportunities. Pharmacol. Res. 148, 104450. https://doi.org/10.1016/j.phrs.2019.104450 (2019).

Article CAS PubMed Google Scholar

Castro-Marrero, J. et al. Unemployment and work disability in individuals with chronic fatigue syndrome/myalgic encephalomyelitis: A community-based cross-sectional study from Spain. BMC Public Health 19, 840. https://doi.org/10.1186/s12889-019-7225-z (2019).

Article PubMed PubMed Central Google Scholar

Maes, M. & Twisk, F. N. Why myalgic encephalomyelitis/chronic fatigue syndrome (ME/CFS) may kill you: Disorders in the inflammatory and oxidative and nitrosative stress (IO&NS) pathways may explain cardiovascular disorders in ME/CFS. Neuroendocrinol Lett. 30, 677–693 (2009).

PubMed Google Scholar

Castro-Marrero, J. et al. Poor self-reported sleep quality and health-related quality of life in patients with chronic fatigue syndrome/myalgic encephalomyelitis. J. Sleep Res. 27, e12703. https://doi.org/10.1111/jsr.12703 (2018).

Article PubMed Google Scholar

Castro-Marrero, J. et al. Comorbidity in chronic fatigue syndrome/myalgic encephalomyelitis: A nationwide population-based cohort study. Psychosomatics 58, 533–543. https://doi.org/10.1016/j.psym.2017.04.010 (2017).

Article PubMed Google Scholar

Castro-Marrero, J., Sáez-Francàs, N., Santillo, D. & Alegre, J. Treatment and management of chronic fatigue syndrome/myalgic encephalomyelitis: All roads lead to Rome. Br. J. Pharmacol. 174, 345–369. https://doi.org/10.1111/bph.13702 (2017).

Article CAS PubMed PubMed Central Google Scholar

Fisk, J. D. et al. Measuring the functional impact of fatigue: initial validation of the fatigue impact scale. Clin. Infect. Dis. 18(Suppl 1), S79-83. https://doi.org/10.1093/clinids/18.supplement_1.s79 (1994).

Article PubMed Google Scholar

Fisk, J. D. & Doble, S. E. Construction and validation of a fatigue impact scale for daily administration (D-FIS). Qual. Life Res. 11, 263–272. https://doi.org/10.1023/a:1015295106602 (2002).

Article PubMed Google Scholar

Buysse, D. J., Reynolds, C. F. 3rd., Monk, T. H., Berman, S. R. & Kupfer, D. J. The pittsburgh sleep quality index: A new instrument for psychiatric practice and research. Psychiatry Res. 28, 193–213. https://doi.org/10.1016/0165-1781(89)90047-4 (1989).

Article CAS PubMed Google Scholar

Alonso, J., Prieto, L. & Antó, J. M. The Spanish version of the SF-36 health survey (the SF-36 health questionnaire): An instrument for measuring clinical results. Med. Clin. 104, 771–776 (1995).

CAS Google Scholar

McGregor, N. R. et al. A preliminary assessment of the association of SCL-90-R psychological inventory responses with changes in urinary metabolites in patients with chronic fatigue syndrome. J. Chronic Fatigue Syndr. 3, 17–37. https://doi.org/10.1300/J092v03n01_03 (1997).

Article Google Scholar

Castresana, C., Perez, A. G. -E., de Rivera, J. L. G. Hospital anxiety and depression scale y psicopatología afectiva. Anales de psiquiatría. pp. 126–130. (1995) Available: https://www.academia.edu/download/51823551/95_A138_03.pdf

Castro-Marrero, J. et al. Effect of dietary coenzyme Q10 plus NADH supplementation on fatigue perception and health-related quality of life in individuals with myalgic encephalomyelitis/chronic fatigue syndrome: A prospective, randomized, double-blind placebo-controlled trial. Nutrients https://doi.org/10.3390/nu13082658 (2021).

Article PubMed PubMed Central Google Scholar

Watson, D. S. et al. Clinical applications of machine learning algorithms: Beyond the black box. BMJ 364, l886. https://doi.org/10.1136/bmj.l886 (2019).

Article PubMed Google Scholar

Kitami, T. et al. Deep phenotyping of myalgic encephalomyelitis/chronic fatigue syndrome in Japanese population. Sci. Rep. 10, 19933. https://doi.org/10.1038/s41598-020-77105-y (2020).

Article CAS PubMed PubMed Central Google Scholar

Słomko, J. et al. Autonomic phenotypes in chronic fatigue syndrome (CFS) are associated with illness severity: A cluster analysis. J. Clin. Med. Res. 9, 254. https://doi.org/10.3390/jcm9082531 (2020).

Article Google Scholar

Zhang, F. et al. Artificial intelligence based discovery of the association between depression and chronic fatigue syndrome. J. Affect. Disord. 250, 380–390. https://doi.org/10.1016/j.jad.2019.03.011 (2019).

Article PubMed Google Scholar

Hanson, S. J., Gause, W. & Natelson, B. Detection of immunologically significant factors for chronic fatigue syndrome using neural-network classifiers. Clin. Diagn. Lab. Immunol. 8, 658–662. https://doi.org/10.1128/CDLI.8.3.658-662.2001 (2001).

Article CAS PubMed PubMed Central Google Scholar

Levine, P. H. et al. Clinical, epidemiologic, and virologic studies in four clusters of the chronic fatigue syndrome. Arch. Intern. Med. 152, 1611–1616 (1992).

Article CAS PubMed Google Scholar

Asprusten, T. T., Sletner, L. & Wyller, V. B. B. Are there subgroups of chronic fatigue syndrome? An exploratory cluster analysis of biological markers. J. Transl. Med. 19, 48. https://doi.org/10.1186/s12967-021-02713-9 (2021).

Article PubMed PubMed Central Google Scholar

Hunter, P. The big health data sale: as the trade of personal health and medical data expands, it becomes necessary to improve legal frameworks for protecting patient anonymity, handling consent and ensuring the quality of data. EMBO Rep. 17, 1103–1105. https://doi.org/10.1525/embr.201642917 (2016).

Article CAS PubMed PubMed Central Google Scholar

Tate, A. R. et al. Exploiting the potential of large databases of electronic health records for research using rapid search algorithms and an intuitive query interface. J. Am. Med. Inform. Assoc. 21, 292–298. https://doi.org/10.1136/amiajnl-2013-001847 (2014).

Article PubMed Google Scholar

Moniz, L. et al. Construction and validation of synthetic electronic medical records. Online J. Public Health Inform. https://doi.org/10.5210/ojphi.v1i1.2720 (2009).

Article PubMed PubMed Central Google Scholar

Weiss, J. C., Page, D. Forest-based point process for event prediction from electronic health records. Machine learning and knowledge discovery in databases. Springer Berlin Heidelberg. 547–562 (2013). https://doi.org/10.1007/978-3-642-40994-3_35

Wang, Z., Myles, P. & Tucker, A. Generating and evaluating cross-sectional synthetic electronic healthcare data: Preserving data utility and patient privacy. Comput. Intell. 37, 819–851. https://doi.org/10.1111/coin.12427 (2021).

Article MathSciNet Google Scholar

Tucker, A., Wang, Z., Rotalinti, Y. & Myles, P. Generating high-fidelity synthetic patient data for assessing machine learning healthcare software. NPJ. Digit. Med. 3, 147. https://doi.org/10.1038/s41746-020-00353-9 (2020).

Article PubMed PubMed Central Google Scholar

Qin, Z., Liu, Z., Zhu, P. & Xue, Y. A GAN-based image synthesis method for skin lesion classification. Comput. Methods Programs Biomed. 195, 105568–105616. https://doi.org/10.1016/j.cmpb.2020.105568 (2020).

Article PubMed Google Scholar

Ravi, D. et al. Degenerative adversarial neuroimage nets for brain scan simulations: Application in ageing and dementia. Med. Image Anal. 75, 102257. https://doi.org/10.1016/j.media.2021.102257 (2022).

Article PubMed PubMed Central Google Scholar

Islam, J. & Zhang, Y. GAN-based synthetic brain PET image generation. Brain Inform. 7, 3. https://doi.org/10.1186/s40708-020-00104-2 (2020).

Article PubMed PubMed Central Google Scholar

Li, Q., Yu, Z., Wang, Y. & Zheng, H. TumorGAN: A multi-modal data augmentation framework for brain tumor segmentation. Sensors https://doi.org/10.3390/s20154203 (2020).

Article PubMed PubMed Central Google Scholar

Tien, H.-J., Yang, H.-C., Shueng, P.-W. & Chen, J.-C. Cone-beam CT image quality improvement using Cycle-Deblur consistent adversarial networks (Cycle-Deblur GAN) for chest CT imaging in breast cancer patients. Sci. Rep. 11, 1133. https://doi.org/10.1038/s41598-020-80803-2 (2021).

Article CAS PubMed PubMed Central Google Scholar

Fukuda, K. et al. The chronic fatigue syndrome: A comprehensive approach to its definition and study. Ann. Intern. Med. 121, 953–959. https://doi.org/10.7326/0003-4819-121-12-199412150-00009 (1994).

Article CAS PubMed Google Scholar

Carruthers, B. M. et al. Myalgic encephalomyelitis: International consensus criteria. J. Int. Med. 2415, 327–338. https://doi.org/10.1111/j.1365-2796.2011.02428.x (2011).

Article Google Scholar

Newman, M. Networks. Oxford University Press; 2018. Available: https://play.google.com/store/books/details?id=YdZjDwAAQBAJ

Hagberg, A., Swart, P. S., Chult, D. Exploring network structure, dynamics, and function using NetworkX. Los Alamos National Lab.(LANL), Los Alamos, NM (United States); (2008). Available: https://www.osti.gov/biblio/960616

Suchowski, M. A. An analysis of the impact of an outlier on correlation coefficients across small sample data where rho is non -zero. Western Michigan UniversityProQuest Dissertations Publishing, Degree Year. p. 3007026. Available: https://search.proquest.com/openview/5d1cbf13c930b7358050381ebab41a85/1?pq-origsite=gscholar&cbl=18750&diss=y

Ferreira, L., Pilastri, A., Martins, C. M., Pires, P. M., Cortez, P. A comparison of AutoML tools for machine learning, deep learning and XGBoost. 2021 International Joint Conference on Neural Networks (IJCNN). pp. 1–8. (2021). https://doi.org/10.1109/IJCNN52387.2021.9534091

Park, D. J. et al. Development of machine learning model for diagnostic disease prediction based on laboratory tests. Sci. Rep. 11, 7567. https://doi.org/10.1038/s41598-021-87171-5 (2021).

Article ADS CAS PubMed PubMed Central Google Scholar

Pedregosa, F. et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).

MathSciNet MATH Google Scholar

Chen, T., Guestrin, C. XGBoost: A scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York, NY, USA: Association for Computing Machinery; pp. 785–794 (2016). https://doi.org/10.1145/2939672.2939785

Download references

ADaS Lab - E-Health Center, Universitat Oberta de Catalunya, Rambla del Poblenou, 156, 08018, Barcelona, Spain

Marcos Lacasa, Ferran Prados & Jordi Casas-Roma

Center for Medical Image Computing, University College London, London, UK

Ferran Prados

National Institute for Health Research Biomedical Research Centre at UCL and UCLH, London, UK

Ferran Prados

Department of Neuroinflammation, Queen Square MS Center, UCL Institute of Neurology, Faculty of Brain Sciences, University College London, London, UK

Ferran Prados

ME/CFS Unit, Division of Rheumatology, Vall d’Hebron Hospital Research Institute Universitat Autònoma de Barcelona, Barcelona, Spain

José Alegre

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

You can also search for this author in PubMed Google Scholar

All authors contributed to the study's conception and design. M.L. and J.A performed material preparation, data collection, and analysis. M.L. wrote the first draft of the manuscript, and all authors commented on previous versions. All authors read and approved the final manuscript.

Correspondence to Marcos Lacasa.

The authors declare no competing interests.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

Lacasa, M., Prados, F., Alegre, J. et al. A synthetic data generation system for myalgic encephalomyelitis/chronic fatigue syndrome questionnaires. Sci Rep 13, 14256 (2023). https://doi.org/10.1038/s41598-023-40364-6

Download citation

Received: 22 March 2023

Accepted: 09 August 2023

Published: 31 August 2023

DOI: https://doi.org/10.1038/s41598-023-40364-6

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.