v
Search
Advanced

Publications > Journals > Exploratory Research and Hypothesis in Medicine> Article Full Text

  • OPEN ACCESS

Multimodal Machine Learning Framework for Cardiovascular Risk Stratification in Adult Obesity: A Cross-sectional Study

  • Pedro Ribeiro1,
  • João Alexandre Lobo Marques2,
  • Marconi Pereira Brandão3,
  • Octávio Barbosa Neto3,
  • Camila Ferreira Leite3 and
  • Pedro Miguel Rodrigues1,* 
 Author information 

Abstract

Background and objectives

Cardiovascular diseases account for approximately 80% of all deaths caused by known medical conditions, making them the leading cause of mortality worldwide. The present study investigates the use of electrocardiogram (ECG) non-linear features and different topological medical features (heart rate, anthropometry, blood, glucose, and lipid profile, and heart rate variability) to discriminate between different Framingham Cardiovascular Risk Scale status groups in adult obesity using machine learning.

Methods

We conducted a cross-sectional study between November 2023 and May 2024 in Fortaleza, Ceará, Brazil. Based on the Framingham Cardiovascular Risk Scale, patients were categorized into three cardiovascular risk groups: Low (22 participants), Moderate (14 participants), and High (17 participants). From ECG signals at two different positions (ECG_Down and ECG_UP), 27 non-linear features were extracted using multi-band analysis. Additionally, 42 medical features provided by physicians were included. From a pool of 19 machine learning classifiers, models were trained and tested within a nested leave-one-out cross-validation procedure using information solely from ECG, solely from medical features, and combining both (multimodal), respectively, to distinguish between Low vs. Moderate, Low vs. High, Moderate vs. High, and All vs. All.

Results

The multimodal model presented the best results for every comparison group, reaching (1) 88.89% Accuracy and 0.8831 area under the curve (AUC) for Low vs. Moderate; (2) 97.44% Accuracy and 0.9706 AUC for Low vs. High; (3) 93.55% Accuracy and an AUC of 0.9412 for Moderate vs. High; (4) 86.79% Accuracy and 0.9346 AUC for All vs. All.

Conclusions

The multimodal model outperformed single-source models in cardiovascular risk classification. ECG-derived non-linear features, especially from ECG_Down, were key drivers, with medical features adding complementary value. The results support its potential use in clinical triage and diagnosis.

Keywords

Adult obesity, Framingham risk score, Cardiovascular risk, Non-linear features, Multimodal, Machine-learning

Introduction

Cardiovascular diseases (CVDs) are the primary cause of death worldwide.1,2 Until the 1940s, the risk factors associated with CVD were not known in a context in which these diseases had already caused many casualties among Americans.3 Thus, during times of epidemiological transition, the treatment and prevention of CVD did not have precise directions.4 The need for a study to investigate the causes of the increasing burden of CVD was recognized.5

The Framingham study was initiated following 5,209 adults free from overt CVD, representing about 19% of Framingham’s population.6 The main objective was to follow the included participants for the development of cardiac heart failure for 20 years.3 The first results of this cohort emerged four years after the beginning of the study and revealed that high blood pressure, high cholesterol levels, and overweight were associated with the development of new-onset coronary heart disease.6 Although being overweight had stood out as a factor related to the development of CVD, obesity was not very prevalent at that time. It was only at the end of the 1970s that its prevalence reached epidemic levels worldwide,7 and today it represents one of the most serious public health challenges.8

Obesity is a complex disease of multiple aetiologies, with its own pathophysiologies, comorbidities, and disabilities, with direct or indirect influence on CVD, affecting endothelial and myocyte function, as well as enhancing major cardiovascular risk factors like diabetes, hypertension, and hyperlipidaemia.9 Thus, the problem lies in the fact that obesity evokes or exacerbates CVD. In addition, parameters indicative of autonomic nervous system (ANS) imbalance are commonly observed in these individuals, with increased sympathetic and reduced parasympathetic nervous activity.10

The ANS imbalance, or dysautonomia, often progresses to ANS dysfunction and is one of the most overlooked and misdiagnosed conditions.11 The ANS plays a major role in the integrated regulation of food intake, involving satiety signals and energy expenditure; thus, ANS dysregulation might favor body weight gain. Conversely, obesity might trigger alterations in the sympathetic regulation of cardiovascular function, thus favoring the development of cardiovascular complications and events.12 Neural mechanisms have been involved in the pathogenesis of obesity, particularly sympathovagal imbalance, and the relative prevalence of sympathetic activity has been suggested to play a pivotal role in this complex bi-directional relationship.13

Heart rate variability (HRV) is a non-invasive method to evaluate the modulation of the ANS on the electrophysiological sinoatrial node. By describing the oscillations between consecutive electrocardiogram (ECG) R-R intervals, HRV can measure the physiological link between the ANS and the heart.14 Thus, HRV is a relevant indicator of cardiovascular autonomic dysregulation in individuals with obesity.10,15

From information from the Framingham cohort, cardiovascular risk prediction models for different CVDs have emerged.16 Variables such as age, serum cholesterol, systolic blood pressure, cigarette smoking, left ventricular hypertrophy on ECG, and glucose intolerance were included in the score.16 The 10-year risk estimates used in the 1998 score provided a convenient way to classify individuals as having low, intermediate, or high risk for future coronary heart disease.3,17

Measuring the risk of a future adverse event in low-income countries from a simple ECG may be a good ally for therapeutic adjustment and directing behaviors for individuals at risk of developing CVD, especially when considering long-term clinical follow-up. Simplifying the process that defines cardiovascular risk based on isolated information (i.e., ECG data) can facilitate the identification of a larger number of individuals at risk. Recent advances in artificial intelligence strategies incorporating machine learning (ML) are gaining new applications in the clinical context, including disease prognosis,18–20 and may be tools to be incorporated into the definition of cardiovascular risk.

We found in state-of-the-art 11 ML studies using Framingham-style clinical variables for cardiovascular risk detection that most ML models rely on classic risk factors (age, sex, systolic blood pressure/diastolic blood pressure,21–31 total and high-density lipoprotein cholesterol, smoking, diabetes), sometimes extending to body mass index (BMI), medications, HbA1c, family history, estimated glomerular filtration rate, and heart rate (HR). Only one study (Yang et al.29) used a broader feature set (49 electronic medical record variables). Classifiers are varied but traditional: Support Vector Machine appears in 4/11 studies, Random Forest in 3/11, and single instances of Logistic Regression, XGBoost, Neural Networks, and gradient boosted trees/proportional hazards regression. Validation is predominantly simple: 7/11 use hold-out splits, and 5/11 use cross-validation; no study in the table reports external validation. Cohort sizes and balance vary widely from very small and imbalanced (e.g., Dogan et al.22: 504 vs 20; Navarini et al.26: 18 vs 115) to large and balanced (e.g., Sajeev et al.28: 23,152 vs 23,152; Quesada et al.23: 5,837 vs 5,837). Performance is heterogeneous. Reported accuracies (n = 9) range from 65.41% to 93.01%, with a mean of 80.21% and a median of 82.35%. Reported area under the curves (AUCs) (n = 9) span 0.6333 to 0.9220, with a mean of 0.764 and a median of 0.751. The highest AUC is from Yang et al.29 (stroke, XGBoost, 0.9220) on a moderately sized, hold-out cohort; the highest accuracy (93.01%) is from Dogan et al.22 but on a tiny, highly imbalanced dataset (504 low-risk vs 20 high-risk), which likely inflates accuracy. Among larger or balanced cohorts, AUCs cluster in the mid-0.7s to mid-0.8s (Alaa et al.24: 0.774; Sajeev et al.28: 0.852; Cho et al.30: 0.751; Chun et al.31: 0.836), while Quesada et al.23 is a lower outlier at 0.6333 despite balance. Methodologically, 10/11 studies reduce risk to binary endpoints (Low vs. High or event vs. no event), limiting calibration and clinical interpretability relative to full, continuous risk scoring; several explicitly note unbalanced data (at least 5/11) and small sample sizes (3/11).

This study aimed to develop and validate a multimodal ML model that integrates ECG non-linear parameters with medical features (HR, anthropometry, blood glucose lipid profile, and HRV) to classify cardiovascular risk severity in alignment with the Framingham risk score. Specifically, we sought to quantify the incremental value of multimodal integration by comparing ECG-only and medical features-only models with the combined model, to assess the relative contribution of each feature source (including ECG lead positions) to discrimination, and to make our curated dataset publicly available to support transparency, reproducibility, and further research.

Materials and methods

In this section, we describe each step of the experimental study and the introduction of the database. Figure 1 illustrates the proposed methodology, structured into five main phases: (1) data collection and curation of a database; (2) signal normalization and filtering: to ensure quality and consistency; (3) ECG multi-band decomposition via Discrete Wavelet Transform (DWT) and feature extraction: extraction of 27 non-linear features per 1-second segment across five decomposition levels, followed by dimensionality reduction using six statistical functions; (4) statistical analysis: refinement of the ECG-based feature set; and (5) classification evaluation metrics: integration of the optimized ECG features with 42 additional medical parameters, including HR, anthropometric measures, blood glucose lipid profile (BGLP), and HRV, followed by normalization and processing through a ML pipeline comprising 19 classifiers, resulting in a comprehensive classification report. More details and specific explanations for each phase are provided in the following subsections.

Methodology workflow diagram.
Fig. 1  Methodology workflow diagram.

Experimental setup

This work was conducted using a MacBook Pro 14 with an M1 Pro chip (8-core CPU, 14-core GPU) and 16 GB of RAM, utilizing MATLAB® and Python coding languages. MATLAB®, version R2023b, was used to extract the non-linear characteristics of the ECG signals, organize them with different topological medical features, and compress and organize the data for ML tasks. Python (version 3.9.12) was used to design, train/test, and obtain discrimination reports from ML models.

Database collection and curation

The database is a cross-sectional study with a quantitative approach and was conducted at the Hospital Universitário Walter Cantídio in Fortaleza, the capital of the State of Ceará, Brazil. The research sample was selected by convenience, and the research was conducted from November 2023 to May 2024 after approval by the institutional Research Ethics Committees (CAAE: 74256823.4.0000.5054 and 74256823.4.3001.5045). The ethical principles recommended by the Declaration of Helsinki and Resolution 466/12 of the Brazilian National Health Council were followed.

Individuals of both sexes aged 30 years or older with a previous nosological diagnosis of obesity, asymptomatic for heart disease, and who underwent laboratory tests (lipid profile and fasting blood glucose) within a maximum period of six months from the interview and data collection were included. Participants in the acute phase of any disease and those with an inability to communicate verbally or cognitive deficits were not included.

A total of 60 patients were initially screened based on the inclusion criteria mentioned above. Of these, five were excluded due to the presence of cardiac disease, discrepancies in diagnostic or comorbidity information, or refusal to provide complete information during the interview, leaving 55 patients eligible for further evaluation. Following anthropometric measurements, vital signs assessment, cardiovascular risk stratification using the Framingham score, and ECG acquisition, two additional patients were excluded due to high interference in HRV measurements. Consequently, the final structured collected database includes data from a total of 53 participants with no missing data (Fig. 2).

Flowchart of the participant recruitment and selection process.
Fig. 2  Flowchart of the participant recruitment and selection process.

Participant flow diagram showing the recruitment, inclusion criteria, and exclusions of the cross-sectional study.

ECG recording was performed noninvasively using the PowerLab data capture hardware system, and HRV parameters were consolidated using lead II of the ECG (Labchart Pro version 7.3.4, Brazil) beat by beat in the CM5 position and analyzed by software (MATLAB® 6.1.1.450 Release 12.1.2001). The collection was performed with a sampling frequency of 1,000 Hz.

ECG acquisition was performed at rest in a climate-controlled room in the morning to minimize circadian HR variations.32 For this purpose, all volunteers were previously instructed to abstain from stimulant drugs, caffeine, tobacco, alcohol, ingestion of high-fat foods, and physical activity for at least 24 h beforehand. Recordings were taken from 07:00 to 11:00 a.m. to avoid any hemodynamic effect on HRV. Participants were instructed not to talk during the assessment, thus avoiding interference that could affect the capture of the HR signal.33,34 Initially, the participant remained at rest on a stretcher in a supine position (ECG_D) for 5 minutes. They were then asked to perform the active postural maneuver (APM), during which they were instructed to stand up abruptly and remain in an orthostatic position (ECG_UP) without movement for 5 minutes until the end of the measurement.35,36

The use of APM was considered because it is a technique with potential sensitivity for assessing vagal and cardiac sympathetic responses.35 The use of APM causes reflex stimulation of the baroreceptors and contraction of the muscles of the lower limbs, thus changing the individual’s position from supine to bipedal, favoring the acquisition of higher delta HRV values.36

In the HRV analyses performed at rest and standing, the stability of the tracings was highlighted, excluding the phase related to the movement during posture change. This is because such movement generates intense instability in the signal due to high cardiocirculatory stress in progress, and therefore, it was not considered for validating the interpretation of the variable information.34,37 Thus, for interpretation purposes, the indexes used in the analysis were: standard deviation of RR intervals (hereinafter referred to as SDRR), square root of the mean of the squares of successive differences (hereinafter referred to as RMSSD), percentage of successive RR intervals with differences greater than 50 ms (hereinafter referred to as pRR50), low-frequency power (LF), high-frequency power (HF), low/high-frequency ratio (LF/HF), short-term variability (SD1), long-term variability (SD2), ratio between standard deviations 1 and 2 (SD1/SD2), and ratio between standard deviations 2 and 1 (SD2/SD1).

The volunteers were then assessed and stratified for cardiovascular risk using the Framingham score with information on sex, age, total and fractional cholesterol, blood pressure, smoking, and diabetes.38 Based on the percentage estimate, cardiovascular risk was classified into three categories: Low (<5%), Moderate (5% to 20% for men and 5% to 10% for women), and High (>20% for men and >10% for women).38–40

The Low cardiovascular risk class included 22 participants, of whom 95.45% (21 out of 22) were females, with an average age of 35.77 years and an average BMI of 42.96 Kg/m2. The Moderate cardiovascular risk class included 14 participants, of whom 71.43% (10 out of 14) were women, with an average age of 48.43 years and a mean BMI of 39.26 Kg/m2.

The High cardiovascular risk class comprised 17 participants, of which 88.24% (15 out of 17) were women, with an average age of 54.24 years and an average BMI of 45.03 Kg/m2. The number of participants results in a minor class imbalance, with a maximum ratio of 60:40, which is generally not considered critical for ML tasks’ performance, as suggested by Thabtah et al.41Table S1 in the Supplement provides additional details regarding the participants’ characteristics.

The database is registered at Mendeley Data - DOI: 10.17632/z8mrvy259n.1, and it includes:

  • An Excel file (Framingham Patients Information.xlsx), which contains each patient’s gender, age, Framingham risk score, and a set of medical features grouped into four categories:

    • HR: resting HR, HR during APM, average HR, systolic and diastolic blood pressure, mean arterial pressure, resting double product, estimated double product during APM, and average double product.

    • Anthropometry: BMI, abdominal circumference, waist circumference, and neck circumference.

    • BGLP: LDL cholesterol, HDL cholesterol, and glycemia.

    • HRV: average RR interval, SDRR, RMSSD, pRR50, LF Power, HF Power, LF/HF Power ratio, SD1, SD2, SD1/SD2, and SD2/SD1 ratios.

  • A folder named “ECG Signals”, containing raw ECG recordings for each patient in.mat format.

Note that in this study, we additionally combined non-linear features extracted from the ECG signals with the clinical data provided in the Excel file to enhance the analysis of cardiovascular risk.

Signal normalization

The ECG signals, x(n), were loaded into MATLAB® and normalized according to the Root Mean Square normalization formula.42

x(n)=x(n)n=0N1x2(n)N,
where N represents the signal’s length. Then, the average value was removed from the entire signal.

Signal filtering

The signals were sampled at a frequency rate of 1,000 Hz, and an elliptic band-pass filter of order 16 with frequency cut-offs at 1 Hz and 40 Hz,43 a steepness of 0.85, and a stop band attenuation of 60 dB was applied to the signal. Figure 3 presents the 2 seconds per class after signal filtering.

2-second electrocardiogram signal example of each cardiovascular risk class after normalization and filtering.
Fig. 3  2-second electrocardiogram signal example of each cardiovascular risk class after normalization and filtering.

ECG Multi-band decomposition via wavelet transform and feature extraction

The DWT is a robust tool for analyzing finite-energy discrete-time signals. It decomposes a signal into a family of base functions constructed from a small set of prototype sequences and their time shifts, enabling compact, time-frequency-localized representations that are well-suited to nonstationary signals. Decomposition and perfect reconstruction are performed with an octave-band, critically decimated filter bank, following the framework introduced by Malvar et al.44 and extended by Vetterli et al.45 When focusing on positive frequencies, each sub-band in the transform is confined to a specific range:

Wk={[0,π2S],m=0,[π2Sm+1,π2Sm],m=1,2,...,S,
where S is the number of levels, S+1 is the number of sub-bands, and π is the normalized angular frequency equivalent to half the sampling rate. The DWT utilizes an analysis scale function,  ϕ˜1(n), and an analysis wavelet function,  ψ˜1(n), defined as:
 ϕ˜1(n)=hLP(n)
and
 ψ˜1(n)=hHP(n),
where hLP(n) and hHP(n) are the impulse responses of the low-pass and high-pass analysis filters, respectively.

The following recursion formulas are defined:

ϕ˜i+1(n)=ϕ˜i(n/2)*ϕ˜1(n),ψ˜i+1(n)=ϕ˜i(n)*ψ˜1(n/2i),
where the symbol “*” denotes the convolution operation. The analysis filter for the mth sub-band is expressed as:
 hm(n)={ϕ˜S(n),m=0ψ˜S+1m(n)m=1,2,...,S.

The mth sub-band signal is computed as:

xm(n){k=x(k)hm(2Snk),m=0k=x(k)hm(2Sm+1nk)m=1,2,...,S.

In this study, the DWT was used to decompose each ECG signal into sub-bands (xm(n)) up to level five (S = 5). The Symlet7 wavelet was selected for its strong performance in ECG analysis up to five decomposition levels.19,25 To maintain consistency with the original sampling rate, the sub-band signals xm(n) were resampled to the native rate using a wavelet interpolation method. For example, Figure 4 provides a visual representation of the multiscale analysis of a 1-second ECG segment performed using the DWT.

Multiscale analysis of a 1s electrocardiogram segment using the discrete wavelet transform.
Fig. 4  Multiscale analysis of a 1s electrocardiogram segment using the discrete wavelet transform.

Feature extraction

After that, 27 non-linear features (see Table 1 for more information46–68) were collected from the signal’s full-band and sub-bands every 1 s within a sliding rectangular non-overlapping window analysis. Then, the resulting time series of features per feature and band were compressed over time, respectively, by six distinct statistical functions: average (Avg), standard deviation (Std), 95th percentile (P95), variance (Var), median (Med), and kurtosis (Kur).19

Table 1

Extracted features with corresponding equations and definitions

FeatureDefinitionEquation
Approximate Entropy (AE)AE assesses the likelihood that similar patterns within the data remain consistent when additional data points are included. A lower AE suggests more regular or predictable data, while a higher AE indicates greater complexity or irregularityAE(m,r)=limNΘm(r)Θm+1(r), θ is the Heaviside step function and can be defined as Θm(r)=iln[Cim(r)]Nm+1, where r denotes a predetermined tolerance, N is the number of data points in the time series, and m is the dimension46
Correlation Dimension (CD)CD measures self-similarity, with higher values indicating a high degree of complexity and less similarityCD=limM2i=1Mkj=i+kMΘ(l|XiXj|)M2, where θ(x) is the Heaviside step function, Xi and Xj are position vectors on the attractor, l is the distance under consideration, k is the summation offset, and M is the number of reconstructed vectors from x(n)46
Detrended Fluctuation Analysis (DFA)DFA measures power scaling observed through R/S analysisDFA(n)=k=1N[y(k)yn(k)]2N, where N is the length, yn(k) is the local trend, and y(k) is defined as y(k)=i=1k[x(i)x¯], with x(i) as the inter-beat interval and x as its average47
Energy (En)En represents the system capacity for performing work48En=n=0N1|x(n)|2
Higuchi Fractal Dimension (H)H estimates the fractal dimension of a time series49H=ln(L(k))ln(1k),
where k is the number of composed sub-series, and L(k) is the average curve size
Hurst Exponent (EH)EH quantifies how chaotic or unpredictable a time series isKq(τ)~(τν)qEH(q),
with Kq(τ)=|X(t+τ)X(t)|q|X(t)|q,
where q is the order moments of the distribution increments, ν is the time resolution, τ is the incorporation time delay of the attractor, and t is the period of a given time series signal X(t)50
Katz Fractal Dimension (K)K estimates the fractal dimensions through a waveform analysis of a time series50K=log(n)log(n)+log(maxn((n1)2+(x(n)x(1))2)n=2N1+(x(n1)x(n))2),
Lyapunov Exponent (ELy)ELy evaluates the system’s predictability and sensitivity to changeELy(x0)=limnk=1nln|f(xk1)|n, where f′ is the derivative of f51
Logarithmic Entropy (LE)LE quantifies the average amount of information (in bits) needed to represent each event in the probability distribution. Higher logarithmic entropy values indicate greater unpredictability or randomness, while lower values suggest more certainty or order48LE=n=1Nlog2|x(n)|2
Shannon Entropy (SE)SE are measured in bits when the base-2 logarithm (log2) is used, quantifying the average bits required to represent each outcome in a probability distribution. Higher entropy values indicate greater uncertainty, unpredictability, or randomness, while lower values suggest more order or certainty48SE=n=1N|x(n)|2log2|x(n)|2
Sample Entropy (SampEn)where m is the length of the vector, r is the tolerance, N is the total size of the vector, and n is a vector portion that is being analysed52SampEn(m,r,N)=log(nm+1nm)
Fuzzy Entropy (FuzzEn)where r is the resolution and m is the dimension53FuzzEn(x,m,r,ne,d)=ln(ψm+1,d(ne,r)ψm,d(ne,r)) where ψm,d (ne,r) is defined as ψm,d(ne,r)=Λ=1Nmdλ=1,λΛNmdexp(ΔΛ,λne)(Nmd)(Nmd1)
Kolmogorov Entropy (K2En)where Cd(ε) is the correlation integral54K2En=limε0limdlnCd(ε)Cd+1(ε)τ
Permutation Entropy (PermEn)where P is the probability distribution for each symbol55PermEn(m)=v=1kPvlnPv
Conditional Entropy (CondEn)where E is the Shannon Entropy, and L is the dimension56CondEn(LL1)=E(L)E(L1)
Distribution Entropy (DistEn)where p is the probability of each bin57DistEn=t=1Mptlog2(pt)log2(M)
Dispersion Entropy (DispEn)where πv0v1vm1 is the dispersion pattern and cm is the potential dispersion patterns58DispEn=n=1cmp(πv0v1vm1)ln(p(πv0v1vm1)) where p(πv0v1vm1) can be defined as p(πv0v1vm1)=#{i|iN(m1)d,Zim,c}N(m1)
Spectral Entropy (SpecEn)where S(f) is the power spectrum and fn is the upper limit Frequency59SpecEn=f=0fnS(f)logS(f)
Symbolic Dynamic Entropy (SyDyEn)where p(l) is the probability of occurrence and c(l) is the frequency of occurrence60SyDyEn=l=14mp(l)logp(l) and p(l) is defined as p(l)=c(l)Nm+1
Increment Entropy (IncrEn)where p(wn) is the frequency of each unique word (wn), q is the quantifying precision, and m is the dimension61IncrEn=n=12q+1)mp(wn)logp(wn)m1 and p(wn) is defined as p(wn)=q(wn)Nm
Cosine Similarity Entropy (CoSiEn)where B(ε)(m)(rCSE) is the global probability of occurrences of similar patterns and rCSE is the tolerance62CoSiEn=[B(ε)(m)(rCSE)log2B(ε)(m)(rCSE)      +(1B(ε)(m)(rCSE))      log2(1B(ε)(m)(rCSE))]
Phase Entropy (PhasEn)where the p(i) is the probability distribution and k is the maximum sector of i63PhasEn=i=1kp(i)logp(i)logk
Slope Entropy (SlopEn)SlopEn is defined as the computation of a sub-sequence of symbols where d is the difference between two consecutive samples, and γ and δ are thresholds64{d>γ2dγ&d>δ1|d|δ0d<δ&d>γ1d<γ2
Bubble Entropy (BubbEn)where Hswapsm is the conditional Rényi entropy65BubbEn=Hswapsm+1Hswapsmlogm+1m1
Gridded Distribution Entropy (GridEn)where j is the jth grid, the pj is the probability of each grid, b is the number of points in the jth grid, and N is the length of the RR intervals66GridEn=j=1n2pjlogpj where pj is defined as pj=bNm
Entropy of Entropy (EnofEn)where l is the level index, yj(τ) is the sequence, and N/τ is the number of representative states for each original time series67EnofEn=l=1s2pllnpl and pl is defined as pl=Numberofyj(τ)lnlN/τ
Attention Entropy (AttnEn)where ai,j are the attention weight probabilities between the positions i and j and ai,j is the averaged weights68AttnEn=j=0d2ai,jlogai,j and ai,j is defined as ai,j=eai,jjeai,j

At the end of the feature extraction from ECG analysis, we obtained a total of 1,944 non-linear features (27 non-linear features extracted from [×] 6 bands (ECG full-band + sub-bands), compacted by [×] 6 data compressors in [×] 2 different positions).

Statistical analysis

To optimize the feature selection process, we performed a total of 1,944 statistical analyses with the Kruskal-Wallis statistical test using MATLAB® Statistics and Machine Learning Toolbox, which evolved all the group’s parameter distributions of features. Only those analyses showing significant differences (p < 0.05) between class distributions were selected, and the others were discarded. For multiclass distribution, we applied the Bonferroni correction. The percentage of statistical analyses with significant differences was approximately 3.1% (60 out of 1,944). Additionally, we combined the selected ECG features with the medical features (defined in Section 2.2), initially provided by health professionals, for use in the ML discrimination stage. Table 2 shows the number of features per origin after filtering.

Table 2

Number of features per origin

# of features
Non-linear features
  ECG_D39
  ECG_UP21
Medical features
  Heart rate9
  Anthropometry4
  Blood glucose lipid Profile3
  Heart rate Variability26

Classification evaluation metrics

To train and test the ML models using a nested leave-out cross-validation procedure, we utilized data derived from three sources: ECG-only, medical features-only, and a combination of both as a multimodal model. Cross-validation techniques allow the entire dataset to be used in classification while preventing data leakage between training and testing sets. This method is beneficial for deriving classification conclusions from small datasets.69

For all models, data were pre-filtered using the analysis of variance F-value feature selector, which selects feature hierarchy based on their F-statistic power. This approach maximizes model performance by iteratively feeding the model with different feature combinations to identify the optimal set for each discriminative task (Low vs. Moderate, Low vs. High, Moderate vs. High, and All vs. All). Specifically:

  • For ECG-only models, the best combination of features ranged from 1 to 60.

  • For medical features-only models, the best combination of features ranged from 1 to 42.

  • For multimodal models, the best combination of features ranged from 1 to 102.

This approach resulted in three optimized models, each responsible for one of the four discrimination tasks previously identified. These models were selected from a pool of 19 pre-designed Scikit-learn ML classifiers,70 as shown in Table 3, with the hyperparameter tuning step selected within Nested leave-one-out cross-validation (LOOCV) due to the constraints of our dataset size. Given the sample size, we therefore relied on rigorous internal validation, using Nested LOOCV (a strategy widely used in the literature for small-sample studies) with all model selection and tuning confined to the training folds,71–75 and we report performance aggregated across folds along with calibration and sensitivity analyses. While this does not replace external validation, it provides an unbiased estimate of out-of-sample performance.76 The selection of the best results was based on their AUC performance for distinguishing between comparison groups.

Table 3

19 Scikit-learn machine learning classifiers configuration

Classifier – Scikit-learn class (Abbreviation)Hyperparameters
AdaBoostClassifier (AdaBoost)n_estimators = 50, learning_rate = 1.0, algorithm = ‘SAMME.R’ + Default
BaggingClassifier (BaggC)n_estimators = 10, max_samples = 1.0, bootstrap = true + Default
DecisionTreeClassifier (DeTreeC)max_depth = 5, criterion = ‘gini’, splitter = ‘best’, min_samples_split = 2 + Default
ExtraTreesClassifier (ExTreeC)n_estimators = 300, criterion = ‘gini’, max_features = ‘auto’, bootstrap=false + Default
GaussianNB (GauNB)priors = none, var_smoothing = 1e-9, store_covariance = false + Default
GaussianProcess Classifier (GauPro)1.0 * RBF (1.0), optimizer = ‘fmin_l_bfgs_b’, max_iter_predict = 100, copy_X_train = true + Default
GradientBoosting Classifier (GradBoost)loss = ‘log loss’, learning_rate = 0.1, n_estimators = 100 + Default
KNearestNeighbors Classifier (KNN)n_neighbors = 5, weights = ‘uniform’, algorithm = ‘auto’ + Default
LinearDiscriminant Analysis (LinDis)solver = ‘svd’, shrinkage = none, priors = none + Default
LinearSVC (LinSVC)penalty = ‘l2’, loss = ‘squared_hinge’, dual = True + Default
LogisticRegression (LogReg)solver = “lbfgs”, penalty = ‘l2’, C = 1.0, max_iter = 100 + Default
LogisticRegressionCV (LogRegCV)cv = 3, penalty = ‘l2’, solver = ‘lbfgs’, max_iter = 100 + Default
MLPClassifier (MLP)α = 1, max iter = 1,000, hidden_layer_sizes = 100, activation = ‘relu’, solver = ‘adam’ + Default
OneVsRestClassifier (OvsR)estimator = LinearSVC(random_state = 0), n_jobs = none, verbose = false + Default
QuadraticDiscriminantAnalysis (QuadDis)reg_param = 0.0, priors = none, store_covariance = false + Default
RandomForest Classifier (RF)max_depth = 5, n_estimators = 300, criterion = ‘gini’, bootstrap = true, min_samples_split = 2 + Default
SGDClassifier (SGD)max iter = 100, tol = 0.001, penalty = ‘l2’, loss = ‘hinge’, alpha = 0.0001 + Default
SGDClassifierMod (SGDCMod)penalty = ‘l2’, loss = ‘hinge’, alpha = 0.0001 + Default
Support-vector Machines (SVC)γ = “auto”, kernel = ‘Radial Basis Function’, C = 1.0, probability = false + Default

AUC quantifies a binary classifier’s ability to distinguish between positive and negative instances across all decision thresholds by comparing the true positive rate (TPR) to the false positive rate (FPR). AUC ranges from 0 to 1, where 1 indicates a perfect classifier and 0.5 indicates performance equivalent to random guessing. Because it summarizes performance across thresholds, AUC provides a single, threshold-independent measure that is especially useful for model comparison under class imbalance.77 In addition to AUC, we report the following metrics: Accuracy (Acc), Precision (Prec; positive predictive value), Recall (Rec; sensitivity), Specificity (Spec), F1 − Score, and negative predictive value (NPV). Let TP, TN, FP, and FN denote true positives, true negatives, false positives, and false negatives, respectively.

Acc is the proportion of correctly classified instances among all instances.78

Acc=TP+TNTP+TN+FP+FN×100%,

Prec, the proportion of predicted positives that are truly positive,79 can be defined as

Prec=TPTP+FP×100%,

NPV shows the proportion of predicted negatives that are truly negative and can be defined as80

NPV=TNTN+FN×100%,

Rec represents the proportion of actual positives that are correctly identified and is defined as79

Rec=TPTP+FN/×100%,

Spec is the proportion of actual negatives that are correctly identified.81

Spec=TNTN+FP×100%

F1 − Score is the harmonic mean of Prec and Rec, balancing both types of error,82 and it is defined as

F1Score=2×Prec×RecPrec+Rec×100%.

Results

We divided these results into four subsections: (1) the performance of the ECG-only model, trained and tested exclusively with ECG features; (2) the performance of the medical features-only model, trained and tested on medical features; (3) the performance of the multimodal models, trained with both ECG and medical features; and (4) a comparison between the performance of the three classes.

ECG-only models’ performance

Table 4 shows the classification report for the ECG-only models. Models trained exclusively on ECG-derived features achieved strong discrimination across all comparison groups. For Low vs. Moderate, the best model using 100% of features from the ECG_D position reached an Acc of 80.36%, Rec of 86.36%, Prec of 82.61%, Spec of 71.43%, F1 − Score of 84.44%, NPV of 76.92%, and an AUC of 0.7890. For Low vs. High, the best model, trained with 83.3% of features from ECG_D and the remainder from ECG_UP, attained an Acc of 97.44%, Rec of 100%, Prec of 95.65%, Spec of 94.12%, F1 − Score of 97.78%, NPV of 100%, and an AUC of 0.9706. For Moderate vs. High, performance included an Acc of 93.55%, Rec of 85.71%, Prec of 100%, Spec of 100%, F1 − Score of 92.31%, NPV of 89.47%, and an AUC of 0.9286, with 65.2% of discriminative information coming from ECG_D. In the All vs. All comparison, the model achieved an Acc of 86.79%, Rec of 86.79%, Prec of 86.92%, Spec of 93.06%, F1 − Score of 86.86%, NPV of 93.47%, and an AUC of 0.9346. Collectively, these results indicate that the ECG_D position contributed the most informative features for maximizing discrimination, with ECG_UP playing a secondary role, particularly in the Moderate vs. High comparison.

Table 4

Best classification results for the electrocardiogram (ECG)-only models per comparison group

Comparison group# of featuresFeature combination (category percentage)ClassifierAcc (%)Rec (%)Prec (%)Spec (%)F1 − Score (%)Negative predictive value (%)Area under the curve
Low vs. Moderate28Electrocardiogram_D (100%) & Electrocardiogram _UP (0%)DeTreeC80.5686.3682.6171.4384.4476.920.7890
Low vs. High18Electrocardiogram _D (≈ 83.3%) & Electrocardiogram _UP (≈ 16.7%)SGD97.4410095.6594.1297.781000.9706
Moderate vs. High23Electrocardiogram _D (≈ 65.2%) & Electrocardiogram _UP (≈ 34.8%)MLP93.5585.7110010092.3189.470.9286
All vs. All47Electrocardiogram _D (≈ 68.1%) & Electrocardiogram _UP (≈ 31.9%)LinSVC86.7986.7986.9293.0686.8693.470.9346

Medical features models’ performance

Table 5 illustrates the classification results for the medical features-based models. When trained exclusively on medical parameters (HR, anthropometry, BGLP, and HRV), models showed competitive but generally lower performance than ECG-based models, except in one case. For Low vs. Moderate, Acc was 83.33%, Rec 86.36%, Prec 86.36%, Spec 78.57%, F1 − Score 86.36%, NPV 78.57%, and AUC 0.8247. For Low vs. High, Acc reached 87.18%, with Rec 86.36%, Prec 90.48%, Spec 88.24%, F1 − Score 88.24%, NPV 83.33%, and AUC 0.8730. For Moderate vs. High, Acc was 80.65%, Rec 92.86%, Prec 72.22%, Spec 70.59%, F1 − Score 81.25%, NPV 92.31%, and AUC 0.8172. In the All vs. All comparison, Acc was 60.38%, Rec 60.38%, Prec 60.03%, Spec 77.66%, F1 − Score 60.20%, NPV 80.17%, and AUC 0.6626. HRV features were the most influential among the medical parameters, contributing 57.1%, 61.5%, 79.3%, and 80% of the discriminative information for Low vs. Moderate, Low vs. High, Moderate vs. High, and All vs. All, respectively. Comparing ECG-based and medical features-only models, the medical features-only model was 3.45% more accurate for Low vs. Moderate but was less accurate by 10.78% to 26.42% for Low vs. High, Moderate vs. High, and All vs. All, indicating stronger ECG-driven discrimination, particularly in binary tasks involving the High class and in multiclass settings.

Table 5

Best classification results for the medical features-only models per comparison group

Comparison group# of featuresFeature combination (category percentage)ClassifierAcc (%)Rec (%)Prec (%)Spec (%)F1 − Score (%)Negative predictive value (%)Area under the curve
Low vs. Moderate7Heart rate (≈28.6%) & Anthropometry (0%) & Blood glucose lipid profile (≈14.3%) & Heart rate variability (≈57.1%)QuadDis83.3386.3686.3678.5786.3678.570.8247
Low vs. High26Heart rate (≈27%) & Anthropometry (≈11.5%) & Blood glucose lipid profile (0%) & Heart rate variability (≈61.5%)QuadDis87.1886.3690.4888.2488.3783.330.8730
Moderate vs. High29Heart rate (0%) & Anthropometry (≈13.8%) & Blood glucose lipid profile(≈6.9%) & Heart rate variability (≈79.3%)LinDis80.6592.8672.2270.5981.2592.310.8172
All vs. All5Heart rate (0%) & Anthropometry (0%) & Blood glucose lipid profile (20%) & Heart rate variability (80%)SGDMod60.3860.3860.0377.6660.2080.170.6626

Multimodal models’ performance

Table 6 corresponds to the best classification results for the multimodal models (ECG features + medical features). Combining ECG and medical features yielded the best overall results in selected comparisons. For Low vs. Moderate, the multimodal model achieved an Acc of 88.89%, Rec 90.91%, Prec 90.91%, Spec 85.71%, F1 − Score 90.91%, NPV 85.71%, and AUC 0.8831, improving accuracy by 10.34% over the ECG-only model. For Low vs. High, the combined model matched the ECG-only performance with an Acc of 97.44%, Rec 100%, Prec 95.65%, Spec 94.12%, F1 − Score 97.78%, NPV 100%, and AUC 0.9706. For Moderate vs. High, Acc was 93.55%, Rec 100%, Prec 87.50%, Spec 88.24%, F1 − Score 93.33%, NPV 100%, and AUC 0.9412. For All vs. All, performance was comparable to ECG-only: Acc 86.79%, Rec 86.79%, Prec 86.92%, Spec 93.06%, F1 − Score 86.86%, NPV 93.47%, and AUC 0.9346. These patterns reflect that feature selection often discarded medical variables in the Low vs. High and All vs. All settings, effectively aligning the combined model’s behavior with ECG-only models in those comparisons.

Table 6

Best classification results for the multimodal models per comparison group

Comparison group# of featuresFeature combination (category percentaege)ClassifierAcc (%)Rec (%)Prec (%)Spec (%)F1 − Score (%)Negative predictive value (%)Area under the curve
Low vs. Moderate33Electrocardiogram_D (0%) & Electrocardiogram _UP (≈51.5%) & Heart Rate (0%) & Anthropometry (≈3%) & Blood Glucose lipid profile (≈3%) & Heart rate variability (≈42.5%)LinDis88.8990.9190.9185.7190.9185.710.8831
Low vs. High18Electrocardiogram_D (≈83.3%) & Electrocardiogram_UP (≈16.7%) & Heart rate (0%) & Anthropometry (0%) & Blood glucose lipid profile (0%) & Heart rate variability (0%)SGD97.4410095.6594.1297.781000.9706
Moderate vs. High31Electrocardiogram_D (≈83.3%) & Electrocardiogram_UP (≈16.7%) & Heart Rate (0%) & Anthropometry (0%) & Blood glucose lipid profile (0%) & Heart rate variability (0%)LinDis93.5510087.5088.2493.331000.9412
All vs. All47Electrocardiogram_D (≈68.1%) & Electrocardiogram_UP (≈31.9%) & Heart rate (0%) & Anthropometry (0%) & Blood glucose lipid profile (0%) & Heart rate variability (0%)LinSVC86.7986.7986.9293.0686.8693.470.9346

Comparison between different types of models’ performance

Across models, Figure 5 summarizes AUC behavior between the approaches presented in Tables 46, for each comparison group: adding ECG to medical features improved AUC in every group (by 7.1% to 41.1%), while adding medical features to ECG improved AUC for Low vs. Moderate and Moderate vs. High but not for Low vs. High or All vs. All due to the feature selection process. Confusion matrices in Figure 6 show a small number of misclassifications: in Low vs. Moderate, two participants per class were misclassified; in Low vs. High, one High participant was predicted as Low; in Moderate vs. High, two High participants were misclassified; and in All vs. All, most errors occurred between Low and Moderate. Receiver-operating characteristic curves (Fig. 7) approached near-perfect classification with AUCs of 0.8831, 0.9706, 0.9412, and 0.9346 for Low vs. Moderate, Low vs. High, Moderate vs. High, and All vs. All, respectively.

Area under the curve (AUC) comparison results, between the electrocardiogram-only models, the medical features-only models, and the multimodal models, per comparison groups.
Fig. 5  Area under the curve (AUC) comparison results, between the electrocardiogram-only models, the medical features-only models, and the multimodal models, per comparison groups.
Confusion matrices’ best prediction results per study group comparison.
Fig. 6  Confusion matrices’ best prediction results per study group comparison.

(a) Low vs. Moderate; (b) Low vs. High; (c) Moderate vs. High; (d) All vs. All.

Best prediction results for the receiver-operating characteristic curves per study group comparison.
Fig. 7  Best prediction results for the receiver-operating characteristic curves per study group comparison.

(a) Low vs. Moderate; (b) Low vs. High; (c) Moderate vs. High; (d) All vs. All.

Discussion

The results demonstrate that ECG-derived information, particularly from the ECG_D position, is the dominant discriminative signal. Compared with medical features-only models, the ECG-only model achieved higher Acc in Low vs. High (97.44% vs. 87.18%; AUC +11.2% relative), Moderate vs. High (93.55% vs. 80.65%; AUC +13.6%), and All vs. All (86.79% vs. 60.38%; AUC +41.1%), with the sole exception of Low vs. Moderate, where the medical features-only model slightly surpassed the ECG-only model in Acc (83.33% vs. 80.36%; AUC +4.5% relative). Multimodal integration delivered its largest gains for adjacent strata, improving Low vs. Moderate Acc by 10.6% relative over ECG-only (88.89% vs. 80.36%) and raising AUC by 11.9% vs. ECG-only and 7.1% vs. medical features-only. For Moderate vs. High, AUC increased by 1.4% vs. ECG-only and 15.2% vs. medical features-only, with accuracy unchanged. In Low vs. High and All vs. All, multimodal performance matched ECG-only (0% change in Acc and AUC) because feature selection discarded medical variables, suggesting that ECG alone captured the dominant signal or that selection favored parsimony over complementarity. The multimodal findings underscore both the potential and the limitations of feature selection in combined pipelines. In Low vs. Moderate and Moderate vs. High, combining modalities improved discrimination (AUC +11.9% and +1.4% versus ECG-only; +7.1% and +15.2% versus medical features-only), with a notable Acc gain only for Low vs. Moderate (+10.3%), consistent with the idea that adjacent categories benefit from complementary features. In contrast, for Low vs. High and All vs. All, the feature selection procedure eliminated medical variables, yielding no change versus ECG-only models (0% for both AUC and Acc), although AUC still improved substantially over medical features-only baselines (+11.2% and +41.1%, respectively), indicating either that ECG features alone sufficiently captured strong cardiovascular risk evidence or that the selection algorithm favored parsimony over potential complementarity. This behavior highlights an important methodological consideration: aggressive feature selection can optimize performance while inadvertently suppressing clinically relevant complementarity. Alternative selection strategies or model families designed to leverage multimodal complementarity may further improve performance in settings where ECG already dominates. Error patterns align with clinical plausibility. Misclassifications most frequently occurred between Low and Moderate, reflecting the expected continuum in risk phenotypes. High Spec and Prec in the High class suggest a favorable profile for ruling in higher risk, while strong Rec and NPV in several settings support safe rule-out. Nonetheless, even small numbers of misclassifications in the High class warrant attention; model calibration and threshold selection should consider clinical consequences and decision costs.

Table 6 indicates that LinDis, LinSVC, and SGD classifiers achieved the best performance. All three are linear classifiers: LinDis (linear discriminant, LDA-style) fits class-conditional densities and applies Bayes’ rule to produce a linear decision boundary; LinSVC is a support vector machine with a linear kernel; and SGDClassifier with loss = ‘hinge’ optimizes the linear SVM (hinge) objective using stochastic gradient descent, yielding a linear large-margin classifier trained via SGD. The superiority of these linear models likely reflects a favorable bias-variance trade-off. On small or noisy datasets, non-linear kernels (e.g., polynomial or radial basis function) can capture spurious patterns unless heavily regularized and carefully tuned, which degrades generalization. Linear models impose a simpler hypothesis class that emphasizes dominant trends over noise and are straightforward to regularize (e.g., via C/alpha or early stopping), resulting in more stable out-of-sample performance.83 The similar results from LinSVC and SGD further suggest that the margin-based linear decision boundary, rather than solver-specific nuances, drives the gains.83

It should be noted that, although we report in Tables 46 the best-performing model for each classification method and comparison based on AUC, we have included Figure S1 to provide a comprehensive overview of the AUC distributions across all combinations of feature sets and comparisons per classifier. These boxplots clearly illustrate that (1) performance differences between classifiers are relatively minor, (2) the results are consistent across models, indicating no evidence of cherry-picking, and (3) the choice of classifier does not significantly affect overall performance, thereby reinforcing the robustness and reliability of our findings.

In comparison with the state-of-the-art (Table 7),21–31 prior works predominantly reported binary tasks, with 66.67% using unbalanced datasets and 41.67% using cross-validation. Among 11 studies, five used an equivalent Low vs. High risk formulation.22,23,25,27,28 Relative to these, our best overall results (Table 6) showed Acc gains of 4.43% to 32.03%, though direct comparison must be interpreted cautiously due to differences in databases and validation protocols. Nevertheless, external comparison with published studies is encouraging but must be interpreted with caution. Reported Acc improvements over prior work in Low vs. High tasks suggest that our approach, particularly with ECG_D features and, where beneficial, multimodal integration, is competitive with the state-of-the-art. However, differences in datasets, class balance, and validation protocols can substantially affect headline metrics. A robust external validation, harmonized evaluation protocols, and transparent reporting will be essential to confirm generalizability.

Table 7

State-of-the-art literature report on cardiovascular risk detection systems based on the Framingham risk scale, with information about the database, the comparison groups, the features extracted, used classifiers, limitations, and Accuracy.

RefYearSourceFeatures extractedComparison group (number of participants)ClassifierLimitationsValidationAccuracyArea under the curve
Unnikrishnan et al.212016Medical FeaturesAge, body mass index, current smoker, gender, total cholesterol, systolic blood pressure, high density lipoprotein cholesterol, diabetes, medication for hypertension, retinopathy, diastolic blood pressureNo-Cardiovascular Diseases (382) vs. Cardiovascular Diseases (128)Support Vector MachineLimited assessment of risk score, just Cardiovascular Diseases vs. No-Cardiovascular Diseases. Unbalanced dataset. Small datasetHold-on82.35%0.71
Dogan et al.222018Medical FeaturesAge, gender, total cholesterol, High density lipoprotein cholesterol, systolic blood pressure, diastolic blood pressure, hemoglobin A1c, and smoking statusLow-Risk (504) vs. High-Risk (20)Ensemble of Random ForestLimited assessment of risk score, just Low vs. High. Unbalanced datasetHold-on93.01%
Quesada et al.232019Medical FeaturesAge, sex, total cholesterol, systolic blood pressure, tobacco use, diastolic blood pressure, high density lipoprotein cholesterol, and the presence of diabetesLow-Risk (5,837) vs. High-Risk (5,837)Random ForestLimited assessment of risk score, just Low vs. HighHold-on80.9%0.6333
Alaa et al.242019Medical FeaturesGender, age, systolic blood pressure, treatment for hypertension, smoking status, history of diabetes, and body mass index5-year risk of Cardiovascular Diseases (4,801) vs. No-5-year risk of Cardiovascular Diseases (4,801)AutoPrognosisLimited assessment of risk score, just Cardiovascular Diseases vs. No Cardiovascular DiseasesCross-validation0.774
Chen et al.252020Medical FeaturesAge, sex, systolic blood pressure, diastolic blood pressure, total cholesterol, high density lipoprotein cholesterol, diabetes status, smoking statusLow-Risk (1,036) vs. High-Risk (983)Support Vector MachineLimited assessment of risk score, just Low vs. HighCross-validation85.11%
Navarini et al.262020Medical FeaturesAge, sex, systolic blood pressure, total cholesterol, smoking status, and hypertension treatmentCardiovascular Diseases (18) vs. No- Cardiovascular Diseases (115)Random ForestLimited assessment of risk score, just Cardiovascular Diseases vs. No Cardiovascular Diseases. Unbalanced dataset. Small datasetCross-validation65.41%0.7297
Jamthikar et al.272020Medical FeaturesAge, Sex, glycated hemoglobin, low density lipoprotein cholesterol, high density lipoprotein cholesterol, total cholesterol, triglyceride, systolic blood pressure, diastolic blood, pressure, hypertension, estimated glomerular filtration rate, and family historyLow-Risk (22) vs. High-Risk (180)Support Vector MachineLimited assessment of risk score, just Low vs. High. Unbalanced dataset. Small datasetCross-validation65.41%0.67
Sajeev et al.282021Medical FeaturesAge, sex, total cholesterol, High density lipoprotein cholesterol, systolic blood pressure, hypertension medication, diabetes, and smoking statusLow-Risk (23,152) vs. High-Risk (23,152)Logistic RegressionLimited assessment of risk score, just Low vs. HighCross-validation0.852
Yang et al.292021Medical Features49 Medical records’ featuresStroke (2,648) vs. No-Stroke (3,337)XGBoostLimited assessment of risk score, just Stroke vs. No-StrokeHold-on84.78%0.9220
Cho et al.302021Medical FeaturesAge, sex, systolic blood pressure, total cholesterol, high density lipoprotein cholesterol, smoking status, history of diabetes, and antihypertensive medication use5-year risk of Cardiovascular Diseases (1,862) vs. No-5-year risk of Cardiovascular Diseases (50,327)Neural NetworkLimited assessment of risk score, just Cardio Vascular Diseases vs. No Cardio Vascular Diseases. Unbalanced datasetHold-on0.751
Chun et al.312021Medical FeaturesAge, smoking status, coronary heart disease, diabetes, blood pressure-lowering treatment, systolic blood pressure-untreated, and systolic blood pressure-treatedStroke (532) vs. No-Stroke (6,185)Gradient Boosted Trees and proportional hazards regressionLimited assessment of risk score, just Stroke vs. No Stroke. Unbalanced datasetHold-on80%0.836
Present study2025Electrocardiogram + Medical features102 featuresLow (22) vs. Moderate (14)LinDisSmall dataset.Cross-validation88.89%0.8831
Low (22) vs. High (17)SGD97.44%0.9706
Moderate (14) vs. High (17)LinDis93.55%0.9412
Low (22) vs. Moderate (14) vs. High (17)LinSVC86.79%0.9346

Overall, the evidence supports ECG as the primary source of discriminative information for cardiovascular risk detection, with ECG_D emerging as the most informative position and ECG_UP contributing to specific pairwise comparisons. Medical parameters (especially HRV) add value when discriminating adjacent risk levels, and their utility can be amplified by careful feature selection and model design choices that preserve multimodal complementarity.

Study limitations

While the findings are promising, several limitations warrant caution. First, the dataset is relatively small, meaning that even a few misclassifications can significantly affect the reported metrics. Without confidence intervals, the true uncertainty may be greater than the point estimates suggest. Second, generalizability may be constrained by the specific acquisition protocol, electrode positions (ECG_D and ECG_UP), device characteristics, and preprocessing choices. These factors, along with the operational definitions of risk groups, may not transfer to other populations, settings, or hardware. Moreover, medical parameters (particularly HRV) are sensitive to transient influences such as autonomic state, medications, and recording conditions, introducing potential confounding if not fully controlled.

Methodological choices may also introduce optimism. The observation that the feature selection step frequently discarded medical variables for feeding the multimodal model suggests that complementary multimodal information may have been underutilized; alternative integration strategies might yield different results. Finally, we did not evaluate model calibration or clinical utility (e.g., decision-curve analysis), and no external or prospective validation was conducted.

Future directions

Future efforts should prioritize updating the database to ensure the generalization of the results. This will also enable the use of a hold-out validation process for classification, which can provide a more straightforward evaluation of model performance on unseen data compared to cross-validation.

Additionally, it should prioritize the use of large and more balanced multi-center cohorts, harmonized acquisition protocols, rigorously nested model selection with calibration assessment, alternative feature selection and integration strategies, and external validation to substantiate robustness and clinical applicability.

Conclusions

In a cohort of 53 patients, we extracted 27 non-linear ECG features from two positions and 42 physician-curated clinical features and, using nested LOOCV with analysis of variance F-value selection, trained models to discriminate Framingham risk strata. Across Low vs. Moderate, Low vs. High, Moderate vs. High, and All vs. All tasks, the multimodal model consistently outperformed ECG-only and medical features-only models, achieving 86–97% Acc with AUCs up to 0.97. ECG-derived non-linear features, especially from the ECG_D position, were the principal drivers of discrimination, while medical features provided complementary gains, indicating the proposed multimodal approach is a promising tool to support clinical triage.

Supporting information

Supplementary material for this article is available at https://doi.org/10.14218/ERHM.2025.00037 .

Table S1

Summary table with the subjects’ characteristics.

(DOCX)

Fig. S1

Area under the curve (AUC) Boxplots of every classifier performance per category per comparison group. (a) AUC Boxplots for the comparison Low vs. Moderate; (b) AUC Boxplots for the comparison Low vs. High; (c) AUC Boxplots for the comparison Moderate vs. High; (d) AUC Boxplots for the comparison All vs. All.

(TIF)

Declarations

Acknowledgement

This work was supported by National Funds from FCT - Fundação para a Ciência e a Tecnologia through project UIDB/50016/2020.

Ethical statement

The research was approved by the Institutional Research Ethics Committees (CAAE: 74256823.4.0000.5054 and 74256823.4.3001.5045). The ethical principles recommended by the Declaration of Helsinki (as revised in 2024) and Resolution 466/12 of the Brazilian National Health Council were followed. All patients provided consent beforehand.

Data sharing statement

The dataset used to support the findings of this study has been deposited in Mendeley Data (doi:10.17632/z8mrvy259n.1). The dataset is currently under embargo and will be publicly available on February 11, 2026.

Funding

No funding was received.

Conflict of interest

The authors have no conflicts of interest related to this publication.

Authors’ contributions

Conceptualization, drafting of manuscript, investigation, manuscript editing (PR, JALM, PMR), validation (CL, MB, ON, JALM, PMR), critical revision of the manuscript for important intellectual content (PR, CL, MB, ON, JALM, PMR), study supervision, and funding acquisition (JALM, PMR). All authors have approved the final version and publication of the manuscript.

References

  1. Roth GA, Mensah GA, Fuster V. The Global Burden of Cardiovascular Diseases and Risks: A Compass for Global Action. J Am Coll Cardiol 2020;76(25):2980-2981 View Article PubMed/NCBI
  2. Tsao CW, Aday AW, Almarzooq ZI, Anderson CAM, Arora P, Avery CL, et al. Heart Disease and Stroke Statistics-2023 Update: A Report From the American Heart Association. Circulation 2023;147(8):e93-e621 View Article PubMed/NCBI
  3. Mahmood SS, Levy D, Vasan RS, Wang TJ. The Framingham Heart Study and the epidemiology of cardiovascular disease: a historical perspective. Lancet 2014;383(9921):999-1008 View Article PubMed/NCBI
  4. Teo KK, Rafiq T. Cardiovascular Risk Factors and Prevention: A Perspective From Developing Countries. Can J Cardiol 2021;37(5):733-743 View Article PubMed/NCBI
  5. Mendis S. The contribution of the Framingham Heart Study to the prevention of cardiovascular disease: a global perspective. Prog Cardiovasc Dis 2010;53(1):10-14 View Article PubMed/NCBI
  6. Andersson C, Nayor M, Tsao CW, Levy D, Vasan RS. Framingham Heart Study: JACC Focus Seminar, 1/8. J Am Coll Cardiol 2021;77(21):2680-2692 View Article PubMed/NCBI
  7. Conway B, Rene A. Obesity as a disease: no lightweight matter. Obes Rev 2004;5(3):145-151 View Article PubMed/NCBI
  8. Koliaki C, Dalamaga M, Liatis S. Update on the Obesity Epidemic: After the Sudden Rise, Is the Upward Trajectory Beginning to Flatten?. Curr Obes Rep 2023;12(4):514-527 View Article PubMed/NCBI
  9. Welsh A, Hammad M, Piña IL, Kulinski J. Obesity and cardiovascular health. Eur J Prev Cardiol 2024;31(8):1026-1035 View Article PubMed/NCBI
  10. Yadav RL, Yadav PK, Yadav LK, Agrawal K, Sah SK, Islam MN. Association between obesity and heart rate variability indices: an intuition toward cardiac autonomic alteration - a risk of CVD. Diabetes Metab Syndr Obes 2017;10:57-64 View Article PubMed/NCBI
  11. Novak P. Autonomic Disorders. Am J Med 2019;132(4):420-436 View Article PubMed/NCBI
  12. Guarino D, Nannipieri M, Iervasi G, Taddei S, Bruno RM. The Role of the Autonomic Nervous System in the Pathophysiology of Obesity. Front Physiol 2017;8:665 View Article PubMed/NCBI
  13. Zapparoli L, Devoto F, Giannini G, Zonca S, Gallo F, Paulesu E. Neural structural abnormalities behind altered brain activation in obesity: Evidence from meta-analyses of brain activation and morphometric data. Neuroimage Clin 2022;36:103179 View Article PubMed/NCBI
  14. Forte G, Favieri F, Casagrande M. Heart Rate Variability and Cognitive Function: A Systematic Review. Front Neurosci 2019;13:710 View Article PubMed/NCBI
  15. Williams SM, Eleftheriadou A, Alam U, Cuthbertson DJ, Wilding JPH. Cardiac Autonomic Neuropathy in Obesity, the Metabolic Syndrome and Prediabetes: A Narrative Review. Diabetes Ther 2019;10(6):1995-2021 View Article PubMed/NCBI
  16. Kannel WB, McGee D, Gordon T. A general cardiovascular risk profile: the Framingham Study. Am J Cardiol 1976;38(1):46-51 View Article PubMed/NCBI
  17. Wilson PW, D’Agostino RB, Levy D, Belanger AM, Silbershatz H, Kannel WB. Prediction of coronary heart disease using risk factor categories. Circulation 1998;97(18):1837-1847 View Article PubMed/NCBI
  18. Fan X, Li Y, He Q, Wang M, Lan X, Zhang K, et al. Predictive Value of Machine Learning for Recurrence of Atrial Fibrillation after Catheter Ablation: A Systematic Review and Meta-Analysis. Rev Cardiovasc Med 2023;24(11):315 View Article PubMed/NCBI
  19. Ribeiro P, Marques JAL, Pordeus D, Zacarias L, Leite CF, Sobreira-Neto MA, et al. Machine learning-based cardiac activity non-linear analysis for discriminating COVID-19 patients with different degrees of severity. Biomed Signal Process Control 2024;87(Pt A):105558 View Article
  20. Fleuren LM, Klausch TLT, Zwager CL, Schoonmade LJ, Guo T, Roggeveen LF, et al. Machine learning for the prediction of sepsis: a systematic review and meta-analysis of diagnostic test accuracy. Intensive Care Med 2020;46(3):383-400 View Article PubMed/NCBI
  21. Unnikrishnan P, Kumar DK, Poosapadi Arjunan S, Kumar H, Mitchell P, Kawasaki R. Development of Health Parameter Model for Risk Prediction of CVD Using SVM. Comput Math Methods Med 2016;2016:3016245 View Article PubMed/NCBI
  22. Dogan MV, Beach SRH, Simons RL, Lendasse A, Penaluna B, Philibert RA. Blood-Based Biomarkers for Predicting the Risk for Five-Year Incident Coronary Heart Disease in the Framingham Heart Study via Machine Learning. Genes (Basel) 2018;9(12):641 View Article PubMed/NCBI
  23. Quesada JA, Lopez-Pineda A, Gil-Guillén VF, Durazo-Arvizu R, Orozco-Beltrán D, López-Domenech A, et al. Machine learning to predict cardiovascular risk. Int J Clin Pract 2019;73(10):e13389 View Article PubMed/NCBI
  24. Alaa AM, Bolton T, Di Angelantonio E, Rudd JHF, van der Schaar M. Cardiovascular disease risk prediction using automated machine learning: A prospective study of 423,604 UK Biobank participants. PLoS One 2019;14(5):e0213653 View Article PubMed/NCBI
  25. Chen YS, Cheng CH, Chen SF, Jhuang JY. Identification of the Framingham Risk Score by an Entropy-Based Rule Model for Cardiovascular Disease. Entropy (Basel) 2020;22(12):1406 View Article PubMed/NCBI
  26. Navarini L, Caso F, Costa L, Currado D, Stola L, Perrotta F, et al. Cardiovascular Risk Prediction in Ankylosing Spondylitis: From Traditional Scores to Machine Learning Assessment. Rheumatol Ther 2020;7(4):867-882 View Article PubMed/NCBI
  27. Jamthikar A, Gupta D, Saba L, Khanna NN, Araki T, Viskovic K, et al. Cardiovascular/stroke risk predictive calculators: a comparison between statistical and machine learning models. Cardiovasc Diagn Ther 2020;10(4):919-938 View Article PubMed/NCBI
  28. Sajeev S, Champion S, Beleigoli A, Chew D, Reed RL, Magliano DJ, et al. Predicting Australian Adults at High Risk of Cardiovascular Disease Mortality Using Standard Risk Factors and Machine Learning. Int J Environ Res Public Health 2021;18(6):3187 View Article PubMed/NCBI
  29. Yang Y, Zheng J, Du Z, Li Y, Cai Y. Accurate Prediction of Stroke for Hypertensive Patients Based on Medical Big Data and Machine Learning Algorithms: Retrospective Study. JMIR Med Inform 2021;9(11):e30277 View Article PubMed/NCBI
  30. Cho SY, Kim SH, Kang SH, Lee KJ, Choi D, Kang S, et al. Pre-existing and machine learning-based models for cardiovascular risk prediction. Sci Rep 2021;11(1):8886 View Article PubMed/NCBI
  31. Chun M, Clarke R, Cairns BJ, Clifton D, Bennett D, Chen Y, et al. Stroke risk prediction using machine learning: a prospective cohort study of 0.5 million Chinese adults. J Am Med Inform Assoc 2021;28(8):1719-1727 View Article PubMed/NCBI
  32. Souza Neto EG, Peixoto JVC, Rank Filho C, Petterle RR, Fogaça RTH, Wolska BM, et al. Effects of High-Intensity Interval Training and Continuous Training on Exercise Capacity, Heart Rate Variability and Isolated Hearts in Diabetic Rats. Arq Bras Cardiol 2023;120(1):e20220396 View Article PubMed/NCBI
  33. Ernst G. Hidden Signals-The History and Methods of Heart Rate Variability. Front Public Health 2017;5:265 View Article PubMed/NCBI
  34. Vanderlei LC, Pastre CM, Hoshi RA, Carvalho TD, Godoy MF. Basic notions of heart rate variability and its clinical applicability. Rev Bras Cir Cardiovasc 2009;24(2):205-217 View Article PubMed/NCBI
  35. Dídimo N dos S, Cruz LAS, Do Carmo GS, Da Rocha RB, Cardoso VS. Parasympathetic autonomic modulation of heart rate and glycemic profile in type 2 Diabeteses patients (in Portuguese). Braz J Dev 2022;8(10):67735-67749 View Article
  36. Paschoal MA, Volanti VM, Pires CS, Fernandes FC. Heart rate variability in different age groups (in Portuguese). Braz J Phys Ther 2006;10(4):413-419 View Article
  37. Nunan D, Donovan G, Jakovljevic DG, Hodges LD, Sandercock GR, Brodie DA. Validity and reliability of short-term heart-rate variability from the Polar S810. Med Sci Sports Exerc 2009;41(1):243-250 View Article PubMed/NCBI
  38. Simão AF, Precoma DB, Andrade JP, Correa Filho H, Saraiva JFK, Oliveira GMM, Murro ALB, et al. First Brazilian Guideline on Cardiovascular Prevention (in Portuguese). Arq Bras Cardiol 2013;101(6 Suppl 2):1-63
  39. Mensegere AL, Sundarakumar JS, Diwakar L, Issac TG, SANSCOG Study Team. Relationship between Framingham Cardiovascular Risk Score and cognitive performance among ageing rural Indian participants: a cross-sectional analysis. BMJ Open 2023;13(11):e074977 View Article PubMed/NCBI
  40. Chen X, Tu Q, Wang D, Liu J, Qin Y, Zhang Y, et al. Effectiveness of China-PAR and Framingham risk score in assessment of 10-year cardiovascular disease risk in Chinese hypertensive patients. Public Health 2023;220:127-134 View Article PubMed/NCBI
  41. Thabtah F, Hammoud S, Kamalov F, Gonsalves A. Data imbalance in classification: Experimental evaluation. Inf Sci 2020;513:429-441 View Article
  42. Lux RL, Sower CT, Allen N, Etheridge SP, Tristani-Firouzi M, Saarel EV. The application of root mean square electrocardiography (RMS ECG) for the detection of acquired and congenital long QT syndrome. PLoS One 2014;9(1):e85689 View Article PubMed/NCBI
  43. Kalidas V, Tamil L. Real-time QRS detector using Stationary Wavelet Transform for Automated ECG Analysis. 2017 IEEE 17th International Conference on Bioinformatics and Bioengineering (BIBE); 2017 Oct 23-25. Washington, DC, USA, Piscataway, NJ, USA: IEEE; 2017:457-461 View Article
  44. Malvar H. Signal Processing with Lapped Transforms. Norwood, Massachusetts: Artech House; 1992
  45. Vetterli M, Kovaèeviæ J. Wavelets and Subband Coding. Englewood Cliffs, New Jersey: Prentice Hall; 1995
  46. Caesarendra W, Kosasih B, Tieu K, Moodie CAS. An application of nonlinear feature extraction - A case study for low speed slewing bearing condition monitoring and prognosis. 2013 IEEE/ASME International Conference on Advanced Intelligent Mechatronics; 2013 Jul 09-12. Wollongong, NSW, Australia, Piscataway, NJ, USA: IEEE; 2013:1713-1718 View Article
  47. Hardstone R, Poil SS, Schiavone G, Jansen R, Nikulin VV, Mansvelder HD, et al. Detrended fluctuation analysis: a scale-free view on neuronal oscillations. Front Physiol 2012;3:450 View Article PubMed/NCBI
  48. Sundararajan D. Discrete wavelet transform a signal processing approach. 1st ed. New Jersey: John Wiley & Sons; 2015 View Article
  49. Silva M, Ribeiro P, Bispo BC, Rodrigues PM. Detection of Alzheimer’s Disease through Nonlinear Parameters of Speech Signals (in Portuguese). Proceedings of the 41st Brazilian Symposium on Telecommunications and Signal Processing; 2023 Oct 8-11. São José dos Campos, SP, Brazil, São Paulo: Brazilian Society of Telecommunications; 2023:1-5 View Article
  50. Garcia A, Garcia C, Villasenor-Pineda L, Montoya O. Biosignal Processing and Classification Using Computational Learning and Intelligence Principles, Algorithms, and Applications. London: Academic Press; 2022
  51. Silva G, Batista P, Rodrigues PM. COVID-19 activity screening by a smart-data-driven multi-band voice analysis. J Voice 2025;39(3):602-611 View Article PubMed/NCBI
  52. Richman JS, Lake DE, Moorman JR. Sample entropy. Methods Enzymol 2004;384:172-184 View Article PubMed/NCBI
  53. Azami H, Li P, Arnold SE, Escudero J, Humeau-Heurtier A. Fuzzy entropy metrics for the analysis of biomedical signals: Assessment and comparison. IEEE Access 2019;7:104833-104847 View Article
  54. Gao L, Wang J, Chen L. Event-related desynchronization and synchronization quantification in motor-related EEG by Kolmogorov entropy. J Neural Eng 2013;10(3):036023 View Article PubMed/NCBI
  55. Bian C, Qin C, Ma QD, Shen Q. Modified permutation-entropy analysis of heartbeat dynamics. Phys Rev E Stat Nonlin Soft Matter Phys 2012;85(2 Pt 1):021906 View Article PubMed/NCBI
  56. Porta A, Baselli G, Liberati D, Montano N, Cogliati C, Gnecchi-Ruscone T, et al. Measuring regularity by means of a corrected conditional entropy in sympathetic outflow. Biol Cybern 1998;78(1):71-78 View Article PubMed/NCBI
  57. Li P, Liu C, Li K, Zheng D, Liu C, Hou Y. Assessing the complexity of short-term heartbeat interval series by distribution entropy. Med Biol Eng Comput 2015;53(1):77-87 View Article PubMed/NCBI
  58. Rostaghi M, Azami H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process Lett 2016;23(5):610-614 View Article
  59. Inouye T, Shinosaki K, Sakamoto H, Toi S, Ukai S, Iyama A, et al. Quantification of EEG irregularity by use of the entropy of the power spectrum. Electroencephalogr Clin Neurophysiol 1991;79(3):204-210 View Article PubMed/NCBI
  60. Wang J, Li T, Xie R, Wang XM, Cao YY. Fault feature extraction for multiple electrical faults of aviation electro-mechanical actuator based on symbolic dynamics entropy. 2015 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC); 2015 Sep 19-22. Ningbo, China, Piscataway, NJ, USA: IEEE; 2015:1-6 View Article
  61. Liu X, Wang X, Zhou X, Jiang A. Appropriate use of the increment entropy for electrophysiological time series. Comput Biol Med 2018;95:13-23 View Article PubMed/NCBI
  62. Chanwimalueang T, Mandic DP. Cosine similarity entropy: Self-correlation-based complexity analysis of dynamical systems. Entropy 2017;19(12):652 View Article
  63. Rohila A, Sharma A. Phase entropy: a new complexity measure for heart rate variability. Physiol Meas 2019;40(10):105006 View Article PubMed/NCBI
  64. Cuesta-Frau D, Schneider J, Bakštein E, Vostatek P, Spaniel F, Novák D. Classification of Actigraphy Records from Bipolar Disorder Patients Using Slope Entropy: A Feasibility Study. Entropy (Basel) 2020;22(11):1243 View Article PubMed/NCBI
  65. Manis G, Aktaruzzaman M, Sassi R. Bubble Entropy: An Entropy Almost Free of Parameters. IEEE Trans Biomed Eng 2017;64(11):2711-2718 View Article PubMed/NCBI
  66. Yan C, Li P, Liu C, Wang X, Yin C, Yao L. Novel gridded descriptors of poincaré plot for analyzing heartbeat interval time-series. Comput Biol Med 2019;109:280-289 View Article PubMed/NCBI
  67. Hsu CF, Wei SY, Huang HP, Hsu L, Chi S, Peng CK. Entropy of entropy: Measurement of dynamical complexity for biological systems. Entropy 2017;19(10):550 View Article
  68. Zhang J, Zhao Y, Li H, Zong C. Attention With Sparsity Regularization for Neural Machine Translation and Summarization. IEEE/ACM Trans. Audio Speech Lang Process 2019;27(3):507-518 View Article
  69. Morin K, Davis JL. Cross-validation: What is it and how is it used in regression?. Commun Stat Theory Methods 2016;46(11):5238-5251 View Article
  70. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in Python. J Mach Learn Res 2011;12:2825-2830
  71. Allgaier J, Pryss R. Cross-Validation Visualized: A Narrative Guide to Advanced Methods. Mach Learn Knowl Extr 2024;6(2):1378-1388 View Article
  72. Sushentsev N, Rundo L, Abrego L, Li Z, Nazarenko T, Warren AY, et al. Time series radiomics for the prediction of prostate cancer progression in patients on active surveillance. Eur Radiol 2023;33(6):3792-3800 View Article PubMed/NCBI
  73. Geroldinger A, Lusa L, Nold M, Heinze G. Leave-one-out cross-validation, penalization, and differential bias of some prediction model performance measures-a simulation study. Diagn Progn Res 2023;7(1):9 View Article PubMed/NCBI
  74. Bradshaw TJ, Huemann Z, Hu J, Rahmim A. A Guide to Cross-Validation for Artificial Intelligence in Medical Imaging. Radiol Artif Intell 2023;5(4):e220232 View Article PubMed/NCBI
  75. Yates LA, Aandahl Z, Richards SA, Brook BW. Cross validation for model selection: A review with examples from ecology. Ecol Monogr 2023;93(1):e1557 View Article
  76. Varma S, Simon R. Bias in error estimation when using cross-validation for model selection. BMC Bioinformatics 2006;7:91 View Article PubMed/NCBI
  77. Nahm FS. Receiver operating characteristic curve: overview and practical use for clinicians. Korean J Anesthesiol 2022;75(1):25-36 View Article PubMed/NCBI
  78. Doðan O. Data Linkage Methods for Big Data Management in Industry 4.0. In: Öner, Yüregir O (eds). Optimizing Big Data Management and Industrial Systems With Intelligent Techniques. Optimizing Big Data Management and Industrial Systems With Intelligent Techniques. Hershey, PA: IGI Global; 2019:108-127 View Article
  79. Ting KM. Precision and Recall. Encyclopedia of Machine Learning and Data Mining. Boston, MA: Springer; 2017:990-991 View Article
  80. Safari S, Baratloo A, Elfil M, Negida A. Evidence Based Emergency Medicine Part 2: Positive and negative predictive values of diagnostic tests. Emerg (Tehran) 2015;3(3):87-88 PubMed/NCBI
  81. Sidey-Gibbons JAM, Sidey-Gibbons CJ. Machine learning in medicine: a practical introduction. BMC Med Res Methodol 2019;19(1):64 View Article PubMed/NCBI
  82. Goutte C, Gaussier E. A Probabilistic Interpretation of Precision, Recall and F-Score, with Implication for Evaluation. In: Losada DE, Fernández-Luna JM (eds). Advances in Information Retrieval. 27th European Conference on IR Research, ECIR 2005; 2025 March 21-23. Santiago de Compostela, Spain, Berlin, Heidelberg: Springer; 2005:345-359 View Article
  83. Mehta P, Wang CH, Day AGR, Richardson C, Bukov M, Fisher CK, et al. A high-bias, low-variance introduction to Machine Learning for physicists. Phys Rep 2019;810:1-124 View Article PubMed/NCBI

About this Article

Cite this article
Ribeiro P, Marques JAL, Brandão MP, Neto OB, Leite CF, Rodrigues PM. Multimodal Machine Learning Framework for Cardiovascular Risk Stratification in Adult Obesity: A Cross-sectional Study. Explor Res Hypothesis Med. 2026;11(1):e00037. doi: 10.14218/ERHM.2025.00037.
Copy        Export to RIS        Export to EndNote
Article History
Received Revised Accepted Published
July 16, 2025 September 4, 2025 September 28, 2025 November 6, 2025
DOI http://dx.doi.org/10.14218/ERHM.2025.00037
  • Exploratory Research and Hypothesis in Medicine
  • pISSN 2993-5113
  • eISSN 2472-0712
Back to Top

Multimodal Machine Learning Framework for Cardiovascular Risk Stratification in Adult Obesity: A Cross-sectional Study

Pedro Ribeiro, João Alexandre Lobo Marques, Marconi Pereira Brandão, Octávio Barbosa Neto, Camila Ferreira Leite, Pedro Miguel Rodrigues
  • Reset Zoom
  • Download TIFF