Play all audios:
ABSTRACT The COVID-19 outbreak put a significant pressure on limited healthcare resources. The specific number of people that may be affected in the near future is difficult to determine. We
can therefore deduce that the corona virus pandemic’s healthcare requirements surpassed available capacity. The Internet of Things (IoT) has emerged an crucial concept for the advancement
of information and communication technology. Since IoT devices are used in various medical fields like real-time tracking, patient data management, and healthcare management. Patients can be
tracked using a variety of tiny-powered and lightweight wireless sensor nodes which use the body sensor network (BSN) technology, one of the key technologies of IoT advances in healthcare.
This gives clinicians and patients more options in contemporary healthcare management. This study report focuses on the conditions for vacating beds available for COVID-19 patients. The
patient’s health condition is recognized and categorised as positive or negative in terms of the Coronavirus disease (COVID-19) using IoT sensors. The proposed model presented in this paper
uses the ARIMA model and Transformer model to train a dataset with the aim of providing enhanced prediction. The physical implementation of these models is expected to accelerate the process
of patient admission and the provision of emergency services, as the predicted patient influx data will be made available to the healthcare system in advance. This predictive capability of
the proposed model contributes to the efficient management of healthcare resources. The research findings indicate that the proposed models demonstrate high accuracy, as evident by its low
mean squared error (MSE), root mean squared error (RMSE), and mean absolute error (MAE). SIMILAR CONTENT BEING VIEWED BY OTHERS HEART FAILURE PATIENTS MONITORING USING IOT-BASED REMOTE
MONITORING SYSTEM Article Open access 06 November 2023 SMART OXYGEN MONITORING IN HOSPITALS: A PILOT STUDY DURING COVID-19 Article Open access 25 January 2025 A NOVEL HEALTH MONITORING
SYSTEM FOR VITAL SIGNS USING IOT Article Open access 19 August 2024 INTRODUCTION Because of the potential for SARS coronavirus to spread, it is critical to separate healthcare services for
COVID and non-covid patients. Treating both groups in the same facility could result in the infection spreading rapidly. However, this approach requires extensive testing with many
individuals and the availability of hospitals with critical care infrastructure that can be turned into COVID facilities. The high dependency unit (HDU) and intensive care unit (ICU) should
have more beds in the COVID facilities. As a result, ventilators, vital monitors, pulse oximeters, infusion pumps, non-invasive ventilators, and a continual supply of oxygen are in high
demand. Aside from treating moderate to severe symptomatic patients, another important consideration is the infrastructure to reduce the burden on already overburdened healthcare facilities
caused by asymptomatic or mildly infected cases, which account for nearly 80% of infected patients. All over the world, life expectancy has grown steadily and has been affected by this
disease outbreak. In the last few decades, the COVID-19 pandemic has resulted in a significant increase in critically ill older people and other patients1. Ensuring a good standard of life
for older people has become a major social problem. Innovative healthcare innovations and resources are being enabled by the rapid proliferation of information and communication
technologies, which promise to address the ex-listing issues. IoT has evolved as one of the most influential connectivity concepts of the twenty-first century and industry 4.02. Because of
its interaction and computational capacities, all artifacts in our everyday lives are part of the web in the IoT world. IoT broadens and expands the definition of the Internet3. IoT enables
seamless interactions between various types of devices, such as medical sensors, video surveillance, and household appliances4. As a result, IoT has become more productive in various
sectors, including health services or healthcare systems. IoT in healthcare refers to a variety of low-cost sensors (wearable, implanted, and environmental) that enable older people to
access current medical healthcare services from anywhere, at any time5. Furthermore, it significantly increases the quality of life of elderly patients and other patients who need urgent
treatment. One of the most imperative technologies is the IoT-oriented, modern health system in the body sensor network (BSN)6. It is essentially a series of low- and lightweight wireless
sensor networks that control the functions of the human body and the environment. BSN nodes are used to gather confidential information (critical to life) and can function in extreme
conditions in the healthcare system. BSN has made patient tracking possible by improving its use in healthcare applications. Several wireless health studies and initiatives have recently
been suggested ensuring a continuous patient tracking process for the health monitoring system, in-ambulatory, in-clinical and open-air environment7. The body sensor network allows contact
with the base station between invasive procedures. An on-body network of sensor systems, on the other hand, enables the communication between the wearable and the non-invasive sensors8. When
a BSN Care server collects data from the local processing unit from an individual (which uses several bio-sensors), the BSN server then feeds and analyses the BSN data into the database9.
Depending on the level of anomalies, it could then communicate with the family members, the local doctor, or even an emergency department of a regional health facility10. When a person (not
necessarily a patient) wears multiple biosensors on their body, the BSN server receives regular updates via local processing units from these sensors11. These data are used to find the
severity in the patients and reflect the requirements of early-stage treatment of the patients; this leads to the need for the availability of beds and ventilators in that particular health
care location. LITERATURE REVIEW This section discusses the various literatures reviewed in the same field. The three parameters has been considered for the literature review analysis, and
the same has been discussed in the Table 1, and these are High dimentional data, demographic information, and hospital admission management.As, in the context of COVID-19 healthcare
management, high-dimensional data, demographic information, and hospital admission management play crucial roles in optimizing patient care and resource allocation. High-dimensional data
refers to complex datasets with numerous features, such as medical history, symptoms, lab results, imaging scans, and real-time monitoring from IoT healthcare devices. This data helps in
predicting disease severity, identifying high-risk patients, and improving treatment strategies using AI and machine learning. Demographic information, including age, gender, ethnicity,
socioeconomic status, and pre-existing conditions, is essential for understanding vulnerable populations, tracking infection trends, and prioritizing vaccination efforts, which also adds the
weitage into the dataset for getting better results. It enables healthcare providers and policymakers to design targeted interventions for different demographic groups. Hospital admission
management, on the other hand, involves tracking bed and ventilator availability, ICU occupancy, and patient discharge planning. Efficient management of hospital resources during COVID-19
ensures that healthcare systems are not overwhelmed, improves patient flow, and enhances overall care delivery. AI-driven forecasting models can leverage these factors to predict patient
surges, optimize resource allocation, and support data-driven decision-making in pandemic response efforts and also for deciding on many more healthcare-related queries. Table 1 presents the
parameters selected for performing the literature review. Various machine learning and artificial intelligence techniques have been extensively utilized to analyze vast amounts of
COVID-19-related data, enabling accurate predictions of disease evolution, patient risk assessment, and healthcare resource optimization. These advanced computational methods focuses on
high-dimensional datasets, incorporating medical history, laboratory-tests findings, imaging results, and real-time patient monitoring to enhance diagnostic accuracy and treatment
strategies. Additionally, AI-driven models assist in forecasting infection trends, identifying potential outbreaks, and optimizing hospital admission management to prevent healthcare system
overload. A comprehensive discussion and detailed analysis of these methodologies, along with their impact on COVID-19 healthcare management, have been presented and systematically
highlighted in Table 2. Liew et al.12 utilized Random Forest (RF), XGBoost, and AdaBoost on a selection of 14 parameters to enhance predictive modeling. Similary, in another research article
by Xiang et al.13 applied SHAP (Shapley Additive Explanations) values to quantify the impact of each variable on predictions, with non-normal data reported using the median and
interquartile range (IQR), while normal data were expressed as mean ± standard deviation (SD). Likewise, Park et al.14 developed and validated a Random Forest-based machine learning model to
predict delirium in hospitalized COVID-19 patients. A neural network-based fusion approach was introduced by Farahat et al.15, where Gibbs energy estimates from three Markov–Gibbs models
were combined to optimize patient decision-making. Elaraby et al.16 designed a cloud-based platform leveraging modern AWS services to ensure high availability, elasticity, coverage,
configurability, and reliability. Meanwhile, Cai et al.17 employed Deep Long Short-Term Memory (LSTM) models to capture dynamic changes in disease severity, clinical outcomes, and Spike
antibody levels. In an effort to enhance interpretability, Solayman et al.18 incorporated Explainable Artificial Intelligence (AI) using the LIME (Local Interpretable Model-agnostic
Explanations) framework to provide transparency in prediction results. For classification tasks, Li et al.19 developed a Gradient Boosting Decision Tree (GBDT) and Logistic Regression (LR)
model for binary classification. Hasan et al.20 focused on forecasting COVID-19 trends using data acquisition techniques and the ARIMA model to predict active cases, positive case rates, and
death rates. Additionally, Famiglini et al.21 demonstrated the effectiveness of the Support Vector Machine (SVM) model, which exhibited strong performance in discrimination, calibration,
and utility. To address segmentation, diagnosis, and prediction challenges, Liu et al.22 employed widely used Convolutional Neural Network (CNN) models, including ResNet, DenseNet, VGG,
Inception, EfficientNet, U-Net, V-Net, and U-Net++. Lastly, Xu et al.23 assessed model performance using tenfold cross-validation and external tests, where datasets were divided into 9 parts
for training and 1 part for validation, ensuring robust evaluation. These studies collectively contribute to the growing body of research aimed at utilizing AI-driven methodologies for
improving COVID-19 diagnosis, prediction, and patient management. MOTIVATION AND NOVELTY The coronavirus disease (COVID-19) outbreak has spread globally, posing a significant threat to
global healthcare organizations. Due to this, the healthcare systems in India are facing lots of problems. As every work is done using a manual process. The Invention of IoT can prove to be
a complete factor for the Indian healthcare organization. This motivated us to come up with a proposed scheme. The scheme monitors the patient’s current health recovery process without the
physical presence of the human being. The scheme manages the total count of available beds and ventilators in the healthcare system. Also, it calculates how many patients can be accommodated
in the healthcare organizations for better services, helping the elderly patients get managed treatment in the hospital. RESEARCH CONTRIBUTION The role of disease prevention and control is
more difficult as the COVID-19 epidemic has spread across the world. In this paper, we present a technological framework for reacting to the COVID-19 epidemic from a health informative
perspective to include theoretical and practical references to other countries on how health information technology can respond to the COVID-19 epidemic for various public health emergencies
and disasters. IoT is used to identify the availability of beds and ventilators in healthcare organizations. Hence, the following are the contributions of this research paper based on the
proposed framework IOT-BVM. The proposed approach smoothly solved the allotment of beds in the healthcare organization. The availability of the beds supports the emergency services if the
patient’s information is present in advance, informs the healthcare systems regarding their capacity, and makes the information available on demand.The IoT-based framework and its analysis
system will quickly emanate messages to the required person to decide on a time. The performance of the proposed scheme is validated against the mentioned methodologies. ORGANIZATION In this
section, the introduction of the research article has been presented, encompassing an in-depth discussion of the literature review, motivation, and contributions. “Methodology” section
focuses on the methodology, providing a comprehensive overview. Subsequently, “Proposed framework and its analysis” section elaborates on the proposed architecture, including its flowchart
and algorithms, followed by a discussion of the experimental results. In “Conclusion” and “Future work” sections, the conclusion and future directions have been addressed, concluding with
the references. METHODOLOGY IoT-based bright pulse oximeter: A physician can track a patient remotely using an oximeter connected to an Internet of Things (IoT) device. This acts as a
real-time warning system that can warn patients if they’re in danger. IoT-based body temperature measurement device (temperature sensor): The unit is small and convenient, and it runs on a
3.3 V battery cell. We can forecast a time series using the ARIMA model by looking at the past values of the series. A time series is a set of measurements taken at regular intervals.
Forecasting is the next step in the process, and it involves predicting the series’ future values. The fundamental business planning, sourcing, and development activities are all guided by
forecasting24. Any forecasting errors can affect the entire supply chain or any business context. As a result, it’s essential to get the predictions right to save money and achieve success.
A time series can now be divided into two forms of forecasting25. Univariate Time Series Forecasting is when you use only the past values of a time series to estimate future matters.
Multi-Variate Time Series Forecasting is when you use predictors other than the series (also known as exogenous variables) to forecast. ARIMA, which stands for “Auto-Regressive Integrated
Moving Average,” is a classification method that “explains” a particular time series based on its previous values, i.e., its hick-ups and stagnated. Prediction error, results in an equation
that can be used to forecast future information. An ARIMA model is represented by 3 terms: a, b, and c. Where a= the AR term’s sequence, b= the MA term’s sequence and c= the number of
differences that must be present in order for the time series to become stationary. A pure Auto Regressive (AR only) model is one in which _X__t_ is solely determined by its own lags.
_X__t_, in other words, is a feature (or function) of the lags of _X_′_t_. $$X_{t} = \alpha + \beta_{1} X_{t - 1} + \beta_{2} X_{t - 2} + \cdots + \beta_{p} X_{t - p} + \varepsilon_{1}$$ (1)
where _X__t_1 is the series’ lag1, _β_ is the lag1 coefficient calculated by the model, and _α_ is the intercept term estimated by the model. A pure Moving Average (MA only) model, on the
other hand, is one in which _X__t_ is solely determined by the lagged forecast errors. $$X_{t} = \alpha + \varepsilon_{t} + \phi_{{1}} \varepsilon_{t - 1} + \phi_{{2}} \varepsilon_{t - 2} +
\cdots + \phi_{q} \varepsilon_{t - q}$$ (2) The errors of the autoregressive models of the corresponding lags are represented by the error terms. The errors _ε__t_ and _ε__t−_1 are the
results of the Equations (3) and (4). $$X_{t} = \beta_{{1}} X_{t - 1} + \beta_{{2}} X_{t - 2} + \cdots + \beta_{0} X_{0} + \varepsilon_{t}$$ (3) $$X_{t - 1} \beta_{{1}} X_{t - 2} +
\beta_{{2}} X_{t - 3} + \cdots + \beta_{0} X_{0} + \varepsilon_{t - 1}$$ (4) AR and MA were the models of the ARIMA. So here’s what an ARIMA model’s equation looks like. An ARIMA model is
one in which the time series has been differenced at least once to make it stationary, and the AR and MA terms have been combined. The equation is presented in Equation (5). $$X_{t} = \alpha
+ \beta_{{1}} X_{t - 1} + \beta_{{2}} X_{t - 2} + \cdots + \beta_{p} Xt - p\varepsilon_{t} + \phi_{{1}} \varepsilon_{t - 1} + \phi_{{2}} \varepsilon_{t - 2} + \phi_{{3}} \varepsilon_{t - 3}
+ \cdots + \phi_{q} \varepsilon_{t - q}$$ (5) The second model, which has been utilized for predicting future patient admissions, has been computed and analyzed using the Transformer model.
The implementation process has been systematically carried out, ensuring that the model’s parameters and configurations are optimized for accurate predictions. Furthermore, the
effectiveness of the Transformer model has been evaluated based on its ability to process sequential data and capture complex patterns within the dataset. It has the following layers: *
_Data preparation and embeddings_ The original Transformer design relies on embeddings to represent discrete tokens. For a time-series application, numerical values are normalized and then
projected into a continuous embedding space. A linear transformation is frequently used to map each time-step value into an embedding vector. Positional information, which is critical for
sequence modeling, is preserved by adding positional encodings. These encodings impose an ordering on the input data, ensuring that temporal relationships are recognized by the model. *
_Encoder module_ The encoder is composed of multiple layers, each containing a multi-head self-attention mechanism and a position-wise feed-forward network. Within each encoder layer,
self-attention computes the relevance of each time step to every other time step in the input sequence. This operation is facilitated by query, key, and value projections, which are combined
to form attention scores. Multi-head attention extends this process by allowing multiple sets of projections to capture different types of temporal dependencies simultaneously. After the
attention sublayer, a feed-forward network is applied to each embedding vector independently. Residual connections and normalization are included to stabilize training and retain information
from earlier layers. * _Decoder module (Adapted for Regression)_ In a standard Transformer, the decoder generates outputs step by step for language modeling tasks. For time-series
regression, the decoder can be configured to receive a shifted version of the target sequence (e.g., previous time steps or placeholder values) to predict subsequent values. The decoder
follows a structure similar to the encoder but incorporates an additional cross-attention mechanism. This mechanism uses the encoder outputs as keys and values while the decoder states serve
as queries, allowing the model to focus on relevant information from the encoded sequence. The final output is often produced by a linear layer that maps the decoder’s hidden states to
numeric predictions rather than discrete tokens. * _Training objective_ Since the goal involves predicting continuous patient counts, a regression loss function such as mean squared error or
mean absolute error is employed. Backpropagation is used to adjust the parameters within the encoder-decoder blocks, including attention weights, feed-forward layers, and embedding
transformations. Gradient-based optimization algorithms, such as Adam, are commonly utilized to achieve stable convergence. * _Temporal attention and long-range dependencies_ A primary
advantage of this architecture lies in its ability to capture long-range dependencies without relying solely on recurrent connections. By computing attention scores across all time steps
simultaneously, the model can learn complex temporal relationships in patient count data, which may include seasonal trends or abrupt changes. Multi-head attention further refines these
relationships by enabling the model to attend to different aspects of the sequence in parallel. * _Inference and forecast generation_ Once the model has been trained, forecast values are
produced by feeding the most recent observed data (or previously predicted outputs) into the encoder-decoder structure. The decoder iteratively predicts subsequent time steps until the
desired forecast horizon is reached. In practice, teacher forcing or scheduled sampling can be applied during training to balance stability and predictive accuracy. The transformer model
architecture has been discussed in the result section. PROPOSED MODEL AND ALGORITHMS The world’s current COVID-19 pandemic requires a patient safety policy that places a greater focus on
epidemiology, particularly identifying effective population-based developmental and learning programs and addressing the consequences26. The healthcare facilities and programs are still in
the early stages of growth, with skill shortages, sickness absence, insufficient facilities, and low-quality care among the challenges27,28. Even though the government’s determination and
the National Health Mission, sufficient and sustainable healthcare remains a mirage29. The rural healthcare system is beset by a persistent shortage of medical practitioners, which
negatively affects the overall accessibility of care for rural residents30. In certain places, the healthcare infrastructure is insufficient or unprepared to deal with the pandemic like
COVID-19 transmission31,32. To deal with such issues33,34 and to solve the problem, the proposed scheme has been implemented in this research paper. The utilization of IoT devices
significantly contributed to patient monitoring during the COVID-19 pandemic. Smart sensors were employed to continuously measure vital health parameters such as body temperature, oxygen
levels, heart rate, and respiratory function. Data collected through wearable devices and remote monitoring systems were transmitted to healthcare professionals, enabling timely medical
responses. By facilitating remote assessments, these technologies helped reduce direct patient contact, thereby lowering the risk of virus transmission and their will not be any direct
contacts with the patients. Cloud-based platforms were used for data storage and analysis, supporting predictive insights and early detection of health deterioration. AI-driven algorithms
were integrated into IoT systems to identify irregularities and notify medical personnel, ensuring swift intervention. Additionally, automated medication dispensers and intelligent inhalers
were implemented to enhance treatment adherence. IoT-enabled tracking mechanisms further assisted in contact tracing and monitoring quarantine compliance. As a result, hospital resources
were effectively managed, and patient care was significantly enhanced throughout the pandemic35. Figure 1 depicts the architecture of the IoT-based healthcare management system for the
benefit of the patients and as well as for the hospital management system. Here the patient’s health is being monitored using the body area sensor network. For the sake of simplicity, only
parameters are being considered here, i.e., the patients’ body temperature and oxygen level through the IoT devices. This information is then passed to the Local Process Unit (LPU). After
that, the data is passed via the Internet to the various stakeholders of the healthcare management systems, such as such emergency ward, physician, and care center of the BSN, where the
parameters related to the patient’s details are maintained and observed for any credulity of the patients. The illustrated architecture is composed of several layers, each intended to manage
patients who use IoT devices during the COVID-19 period. In the first layer, body sensors are placed on individuals to collect vital signs. The readings are then received and filtered by a
local processing unit. Next, the data are transferred via the internet to a central repository, where they are stored in a database. Critical cases are flagged, and positive results are
tracked automatically. This information is then shared with healthcare services, such as emergency departments, physicians, and care centers, to enable ongoing monitoring. One notable
advantage of this layered design is its ability to provide real-time updates on patient health, which supports faster intervention and more informed clinical decisions. In addition, the need
for in-person visits can be reduced, lowering the potential for virus transmission. The architecture can also be scaled up if the number of patients increases, offering flexibility in times
of high demand. However, certain drawbacks must be acknowledged. A reliable internet connection is required for seamless data flow, and any disruption can affect timely updates. Moreover,
patient privacy must be protected through robust data security measures36. Costs related to installing and maintaining sensors, databases, and other infrastructure can also be considerable.
Despite these challenges, the layered approach continues to offer valuable support for remote patient care and broader public health initiatives. Fig. 1 shows the architecture of the
proposed IOT-BVM scheme. The working of the proposed scheme, IOT-BVM, is expressed in flowcharts 1 and 2. Figure 2 depicts the working mechanism of flowchart 1 and the operations of the
process that is being undergone inside the hospital management. Initially, the details of the patients are identified who are presently admitted to the hospital. The patients are being
monitored using the IoT-based environment, and it is being checked whether or not the temperature and oxygen level are in the correct range. If it’s not proper, continuous monitoring will be
done. If the monitored information results in the details that indicate that the patients are being recovered and if they have completed the 14 days of the isolation period, then they are
to be in the process of getting discharged from the hospital, and its information must be passed to the stakeholders of the hospital system. The outcome of this algorithm is 1 the total
number of available beds and is denoted as “V”, The algorithm 1 represents herein. The time complexity of the algorithm 1 can be analyzed by looking at the steps involved in the procedure.
Let n represent the number of hospitalized patients. In Step 3, retrieving the details of the patients is a constant-time operation, O(1), since it’s just an input data retrieval. In Step 4,
fetching the IoT-based body temperature (T) and oxygen level readings (O) for each patient requires iterating through each patient’s data. This operation takes O(n) time, where nn is the
number of patients, because it involves fetching and storing data for all patients. Step 5 consists of checking each patient’s body temperature and oxygen level. The comparison of these
values is a constant-time operation, O(1), for each patient. If a patient meets the specified conditions (temperature < 99°F and oxygen level > 96%), the algorithm either discharges
the patient or continue’s monitoring based on whether the isolation period has passed. This decision-making process also takes constant time, O(1), per patient. If the conditions are not
met, an alarm is generated and the hospital management is notified, which is again an O(1) operation. Since these steps are executed for each patient, Step 5 has a time complexity of O(n).
Finally, Step 6 involves repeating the evaluation process until all patients are discharged. Since each patient is processed only once, the overall time complexity of this loop remains O(n).
Thus, considering all steps together, the total time complexity of the algorithm is O(n), where nn is the number of hospitalized patients. This indicates that the algorithm’s execution time
increases linearly with the number of patients, making it a highly efficient solution for managing patient data in real time. The system performance of this algorithm can be evaluated based
on its efficiency, responsiveness, and scalability. In terms of efficiency, the algorithm performs well because it operates in linear time, O(n). Each step of the procedure involves
essential operations, such as retrieving data, comparing patient conditions, and triggering the necessary actions based on those comparisons. As the number of patients increases, the
system’s performance remains proportional to the data size, making it well-suited for managing larger datasets without excessive delays. Regarding responsiveness, the algorithm is designed
to make decisions promptly. It evaluates each patient’s condition and, based on the results, either discharges the patient, continues monitoring, or raises an alarm. This ensures that the
hospital management system can respond quickly to any changes in a patient’s condition, improving patient care and hospital workflow. The ability to react immediately is crucial for
real-time healthcare applications, where fast decision-making can directly impact patient outcomes. In terms of scalability, the algorithm is capable of handling an increasing number of
patients without significant performance degradation. Since the time complexity is linear, the system scales efficiently with the number of patients. For very large datasets, the processing
time will increase proportionally, but the algorithm will still function adequately as long as the hospital’s infrastructure is equipped to manage the additional data. Finally, in terms of
resource utilization, the algorithm relies mainly on data retrieval and condition checks, both of which are low-complexity operations. This results in minimal resource consumption, making
the system efficient in terms of CPU and memory usage. However, as the dataset grows, additional infrastructure might be needed to handle the increased data processing load. Hence, the
algorithm offers efficient performance with linear time complexity, ensuring that it can handle the increasing scale of patient data without significant delays. Its quick decision-making
process enhances responsiveness, and its scalability makes it suitable for use in larger hospitals or healthcare facilities. Overall, the system is designed to perform effectively in
real-time environments, balancing efficiency, scalability, and resource usage. Flowchart 2 is the subpart of the proposed scheme and is represented in Fig. 3. Figure 3 depicts the scenario
where the comparison of the available beds (V) and the positive cases and its prediction of the Confirmed cases (A) are taken into consideration in that particular region. Then this is
compared to get the idea that the vacant beds are available in enough away so that the patients can be admitted and will be given good services by the healthcare systems. In case the
available beds are not present, then the nearby hospital vacancies beds data is to be identified so that the patients can be accommodated in the hospitals; the same is depicted in Algorithm
2. The outcome of the mentioned approach solves the problem of managing the patients in this pandemic situation. The services can be improved as the data will be available in advance, and
the hospital management will have enough time to arrange the physicians, medical devices, staff, and vacant beds. This will also aid the healthcare systems to be prepared in advance with the
availability of medical equipment. The following section discusses the performance evaluations of the proposed scheme. The time complexity of the algorithm 2 can be analyzed by evaluating
the operations involved in each step. Let n represent the number of critical COVID-19 patients and mm represent the number of available vacant beds in the region. In Step 3, retrieving the
details of the COVID-19 patients for a specified region is an O(1) operation, assuming the data is readily available. Step 4 involves assigning the number of critical patients (denoted as A)
from the retrieved data, which is a simple assignment and thus takes constant time, O(1). In Step 5, the algorithm compares the number of critical patients A with the available vacant beds
VV. This comparison is a constant-time operation, O(1). If the number of critical patients is less than the available beds, the algorithm allocates beds by admitting the patients to
hospitals in the region, which again takes constant time, O(1). However, if the number of critical patients exceeds the available beds, the algorithm needs to identify hospitals with vacant
beds, which requires scanning the list of hospitals. This step takes O(m) time, where mm is the number of hospitals. Once the hospitals with available beds are identified, assigning patients
to these hospitals takes O(1) per patient. Finally, Step 6 involves repeating the process until all patients are discharged. Since each patient is evaluated once, this loop runs O(n) times.
Given that the algorithm may require scanning all hospitals in the worst case, the overall time complexity of the algorithm is O(n⋅m), where n is the number of critical patients and mm is
the number of hospitals. This time complexity suggests that as the number of critical patients and hospitals increases, the algorithm’s execution time will scale accordingly. The system
performance of this algorithm can be assessed in terms of efficiency, responsiveness, and scalability. In terms of efficiency, the algorithm performs well for smaller datasets, as most
operations are constant-time, O(1), including retrieving data and assigning patients to hospitals. The primary area where performance may degrade is when identifying hospitals with available
beds, which requires scanning the list of hospitals. As the number of hospitals grows, this step can become a bottleneck, increasing the overall time complexity to O(n⋅m). Therefore, while
the algorithm is efficient for smaller datasets, its performance may decrease with a larger number of hospitals or patients. In terms of responsiveness, the algorithm ensures quick
decision-making when allocating beds to critical patients. If the number of critical patients is less than the number of vacant beds, they can be admitted to the available hospitals in the
region, which ensures that patients are admitted promptly. If there are more critical patients than vacant beds, the algorithm identifies nearby hospitals with available beds and directs
patients accordingly. This ensures that the system is responsive and that no critical patient is left without attention, making it highly suitable for real-time applications in healthcare
settings. Regarding scalability, the algorithm can handle a growing number of patients, but its scalability is affected by the number of hospitals in the region. As the number of hospitals
increases, the algorithm’s performance may suffer due to the need to scan a larger list of hospitals for available beds. This means that while the algorithm is scalable to a degree, it may
require optimizations, such as indexing hospitals or clustering them based on location or availability, to improve efficiency when the number of hospitals or patients grows significantly.
Lastly, in terms of resource utilization, the algorithm is generally efficient, as it primarily performs basic comparisons and assignments. However, the process of identifying vacant beds in
hospitals can be resource-intensive, especially if the list of hospitals is large. As the dataset grows, the system may need more resources to manage the increased computational load.
Therefore, the algorithm is suitable for smaller-scale applications, but for larger datasets, it might require infrastructure improvements or optimization techniques to ensure optimal
performance. Hence, while the algorithm offers an efficient and responsive solution for allocating beds to critical patients, its time complexity of O(n⋅m) makes it less efficient as both
the number of patients and hospitals increases. The system performs well for smaller datasets but may face challenges with scalability and resource usage when handling larger volumes of
data. To ensure better performance at scale, optimizations such as data indexing or improved searching algorithms could be considered. PROPOSED FRAMEWORK AND ITS ANALYSIS The world’s current
COVID-19 Pandemic requires a patient safety policy that places a greater focus on epidemiology, particularly identifying effective population-based developmental and learning programs and
addressing the consequences36. The healthcare facilities and programs are still in the early stages of growth, with skill shortages, sickness absence, insufficient facilities, and
low-quality care among the challenges. However, the government’s determination and the National Health Mission that adequate and sustainable healthcare remain a mirage. Keeping these things
in the mind, the said approach is proposed. In this section, the performance evaluation has been discussed and is as follows. Figures 4 and 5 represents the conventional (X-ray image to see
the impact of the infection) methods to identify the criticalness of the patient’s health. But every time doing the X-ray of the target patients is not good for their health. Hence,
monitoring the parameters such as oxygen level and temperature is another solution implemented using IoT-based devices and theories. IoT (Internet of Things) can also be used to further
automate the process of hospital management and patient monitoring37. Using wearable body sensors, e.g., pulse oximeters and temperature sensors, patients’ vital signs can be monitored
continuously in real time38. The sensors send data wirelessly to a central gateway, which processes and aggregates the data before sending it to a decision-support system. The cloud system
then runs the decision logic described, e.g., checking if the patient’s temperature is below 99°F and oxygen levels above 96%. If these are the criteria, the system automatically checks if
the 14-day isolation period has elapsed. After confirmation, it sends a discharge notification, updates the hospital bed management system to indicate the newly released bed, and informs the
concerned staff. If the vitals of the patient exceed the safe threshold, the IoT system automatically raises an alarm, alerting hospital management through SMS, email, or dashboard alerts.
This is an automated process that saves time wasted in manual checks, reduces human error, and allows timely intervention. Further, the merging of IoT data with electronic health records
(EHR) and hospital management systems allows enhanced resource planning and trend analysis. Overall, IoT deployment automates the entire process, allowing better patient care, operational
efficiency, and bed turn-over rates without compromising on high accuracy for monitoring critical health parameters. Table 3 below indicates the various aspects of the IoT based patients
monitoring system. The results are based on the historical data taken through the KAGGLE. Figure 6 shows the plot of patients who were admitted to the hospitals. The dataset utilized in this
research was sourced from Kaggle and provides a comprehensive view of key COVID-19 metrics, offering crucial insights for understanding the progression and impact of the pandemic. The
dataset contains several time-series plots that capture important trends over time, each serving a specific purpose in analyzing the dynamics of the pandemic. One of the key visualizations
is the plot of patients admitted to hospitals, which illustrates the fluctuations in hospital admissions throughout the observed period. This plot is essential in understanding how
healthcare facilities were impacted as the number of cases surged. Additionally, the dataset includes a plot of recovered cases, which provides a clear picture of the recovery trends and
helps to evaluate the effectiveness of treatment and healthcare interventions during different phases of the pandemic. A plot of the death rate within the hospital is also included, offering
critical insights into the severity of the disease and the challenges faced by healthcare systems. Moreover, the dataset tracks the total test results and the increasing number of cases,
showing the relationship between testing efforts and the rise in reported cases. This is particularly important for understanding how testing infrastructure scaled and its role in detecting
the spread of the virus. Finally, the dataset presents a plot of COVID-19 positive cases, reflecting the overall trajectory of the virus in the population and providing an essential context
for understanding how the pandemic evolved over time. Together, these time-series plots offer a rich dataset for analyzing the various facets of the COVID-19 pandemic, and they form the
foundation for our research, enabling a detailed exploration of trends, correlations, and patterns within the data. The preprocessing of the dataset was carried out using several essential
steps to ensure its suitability for machine learning (ML) and deep learning (DL) models. First, we began by handling missing data. For any rows with missing values, we applied imputation
techniques, such as mean or median imputation, to replace missing numerical data. Rows with excessive missing information were removed to maintain the dataset’s integrity. Categorical
features were transformed using one-hot encoding to convert non-numeric values into a format compatible with machine learning algorithms. Next, the focus has been given on normalizing
numerical features. We used Min–Max scaling to ensure that all numerical values fell within a uniform range (0 to 1), preventing any one feature from disproportionately influencing model
training due to differences in scale. Given that the dataset includes time-series data, it was critical to preserve the chronological order of the observations. We formatted the data to
ensure it accurately reflected the temporal dependencies in the dataset. For time-series forecasting, we used a sliding window approach to create sequences of past data points to predict
future outcomes. This allowed the models to capture the dynamic trends over time, such as hospital admissions, recoveries, and death rates. Feature engineering played a key role in enriching
the dataset. We created lag features, capturing past values of vital variables like patient admissions and recoveries, which help the models account for historical trends and dependencies.
Additional temporal features, such as day-of-week or month, were also generated to capture any seasonal patterns. Finally, the dataset was split into training and testing subsets, typically
using an 80/20 split. K-fold cross-validation was implemented to evaluate the models robustly, ensuring they were not overfitting to any particular subset of the data. By preprocessing the
dataset in these ways, we ensured that it was ready for modelling. RESULTS AND DISCUSSIONS _Test bed setup for experimental evaluation_ The experimental evaluations were conducted on the
environment equipped with an Intel(R) Xeon(R) CPU E5-2620 v2 operating at 2.10 GHz, featuring 6 physical cores and 12 logical processors. The system was supported by 128 GB of DDR4 memory.
Storage was configured with five 300 GB drives. For GPU-based computations, an NVIDIA Tesla K80 graphics card with 128 GB memory capacity was utilized. The dataset used here was downloaded
from the publicly hosted COVID-19 time series data by the Center for Systems Science and Engineering (CSSE) at Johns Hopkins University, URL:
https://github.com/CSSEGISandData/COVID19/tree/master/csse_covid_19_data/csse_covid_19_time_series. Cumulative confirmed United States cases from the dataset were extracted. After
preprocessing, there was data aggregation by date, and the cumulative case counts were converted to daily new cases by differencing two consecutive days’ values. Through this, analysis of
the spread of COVID-19 over time was made in a more defined manner. Preprocessing was then followed by a split of the dataset into the training and the testing set by an 80/20 ratio. A
sliding window approach was used for the generation of sequence, where for each input, there were 10 consecutive days of daily new case counts in an attempt to predict the next day’s value.
These were reshaped accordingly to fit into the input parameters of deep learning models. The model development was centered on three deep learning architectures: LSTM, GRU, and Transformer.
Hyperparameter tuning was used for each of the models in order to provide optimal performance, and evaluation included metrics such as Mean Absolute Error (MAE), Root Mean Squared Error
(RMSE), and Loss. The LSTM model was constructed with 64 units of memory, dropout of 0.2 to prevent overfitting, and trained for over 50 epochs with the Adam optimizer at learning rate
0.0001. The Mean Squared Error (MSE) as the loss function was employed. The GRU model also employed a similar structure, swapping the LSTM layer for a GRU layer but utilizing the same
hyperparameters such as the number of units, dropout rate, optimizer, and loss function. Mechanisms with 4 attention heads that permit the simultaneous attention of different components of
the input sequence. A light Transformer was created with the hyperparameters tuned including the number of layers, embedding size, and dropout rate. Training was conducted for 50 epochs
using the Adam optimizer and MSE loss function. Throughout the training phase, there was a consistent drop in loss over epochs, and the models were exhibiting increasingly better
performance. After training, the models were tested on the test set specified. The Transformer model demonstrated better results on all metrics of evaluation, closely followed by the LSTM
and GRU models. Visual plots were created to represent the intercorrelations of the evaluation metrics. Graphical plots of MAE vs. Loss, RMSE vs. Loss, and RMS vs. Loss revealed similar
trends, justifying the seen hierarchy of performance between the models. Moreover, values for the x-axis and y-axis were easily identified and shown for each graph to enhance
interpretability. The thorough evaluation pointed out that deep learning models, particularly the Transformer model, hold enormous potential in forecasting trends in COVID-19 cases,
demonstrating considerable enhancements over conventional sequential models. _Discussion on Hyperparameter Selection and Tuning_ In the LSTM model, a number of hyperparameters were selected
with care to maximize performance and avoid overfitting. In particular, the usage included the utilization of a total of 64 LSTM units, which allowed for a balance between learning capacity
and computational cost. This number of units was empirically selected to provide sufficient memory capacity without making the model excessively complicated. In another effort to enhance
generalization, a dropout rate of 0.2 was utilized, where 20% of the neurons were randomly dropped out during training, thereby minimizing the chances of overfitting. The model was trained
for 50 epochs, a time period which was selected based on observation of the trend in loss, such that this number of iterations allowed for sufficient accommodation of temporal patterns
without inducing overfitting. A batch size of 32 was utilized, as it was discovered to offer the best balance between computational cost and training stability. The Adam optimizer, with a
learning rate of 0.0001, was utilized due to its adaptive learning characteristics and ability to deal with sparse gradients, features highly sought after in time series forecasting
activities. Model optimization was conducted using the Mean Squared Error (MSE) loss function, which is generally known for suitability in regression tasks, where large prediction errors
were given a proportionately larger penalty. The same hyperparameters were used in the GRU model for consistency and to allow fair comparisons. The setup of the GRU network was 64 units, a
reflection of the structure of the LSTM network because the same number has been found to offer an optimal balance between training time and model performance. Another 0.2 dropout rate was
applied to prevent overfitting. Training was carried out for 50 epochs, with a batch size of 32, to allow for uniform updating of the gradients and efficient memory management. The Adam
optimizer, using a learning rate of 0.0001, offered fast convergence and responsiveness to the changing learning needs of the GRU model. Like the LSTM network, the MSE loss function was used
to reduce the gap between the real and predicted values, with accuracy in regression tasks in mind. For the Transformer model, some modifications were made to accommodate its architecture,
the same has been depicted in Fig. 4. The model used 4 attention heads, which allowed it to attend to various sections of the sequence data simultaneously, thereby capturing complex
interdependencies well. Two multi-head attention layers were utilized, balancing between depth and computational tractability. A hidden size of 128 was chosen to offer sufficient
representational capacity, enabling the model to learn complex patterns without overfitting. A dropout rate of 0.2 was applied uniformly to further generalize the model. Training was carried
out for 50 epochs, and a batch size of 32 was utilized, just like in the recurrent models. The Adam optimizer with a learning rate of 0.0001 facilitated efficient and adaptive optimization
during training. MSE was once again used as the loss function to maintain uniformity across models as well as to appropriately penalize large prediction errors using PyTorch or TensorFlow
can be provided for time-series forecasting with transformers. DISCUSSIONS OF THE RESULT ACHIEVED: The performance of the deep learning models—LSTM, GRU, and Transformer—was evaluated based
on their Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and Loss values after training and testing over the COVID-19 daily new cases dataset and the same has been depicted in the
Fig. 5, 6, 7, 8, 9, and 10. The figigure indicates the values as, the LSTM model achieved a MAE of 4213.76, a RMSE of 6050.09, and a corresponding Loss value of 78,426,017.38. These values
suggest that while the LSTM was reasonably effective in capturing the temporal dependencies of the COVID-19 daily new cases, there were still notable prediction errors. Mathematically, the
RMSE being higher than the MAE indicates that larger errors occurred occasionally because RMSE penalizes larger errors more than MAE. Similarly, the GRU model achieved a MAE of 4580.59, a
RMSE of 7061.71, with a Loss value of 91,656,112.95. The GRU’s MAE and RMSE were slightly higher than the LSTM’s, which suggests that while the GRU was able to model the temporal structure
to some extent, it underperformed relative to the LSTM model. The difference between MAE and RMSE for GRU also points toward the occurrence of significant errors on some days, negatively
impacting the model’s reliability. In contrast, the Transformer model achieved a MAE of 4018.71, a RMSE of 5838.42, with a Loss value of 67,516,675.40. These numerical results demonstrate a
superior performance in comparison to both the LSTM and GRU models. The MAE was the lowest among the three models, indicating that the Transformer made the smallest errors on average.
Furthermore, the RMSE for the Transformer was also the lowest, highlighting that the Transformer was not only minimizing average errors but also ensuring that extreme prediction deviations
were fewer compared to LSTM and GRU. Mathematically, the following observations has been observed: MAE comparison:\({\text{MAE}}_{{{\text{Transformer}}}} < {\text{MAE}}_{{{\text{LSTM}}}}
< {\text{ MAE}}_{{{\text{GRU}}}}\). Substituting the actual values: 4018.71 < 4213.76 < 4580.59. RMSE comparison:\({\text{RMSE}}_{{{\text{Transformer}}}} < {\text{
RMSE}}_{{{\text{LSTM}}}} < {\text{ RMSE}}_{{{\text{GRU}}}}\). Substituting the actual values: 5838.42 < 6050.09 < 7061.71. This mathematically validates that the Transformer model
yielded consistently lower errors both in terms of MAE and RMSE. In addition, considering the Loss values, which indicates the Mean Squared Error (MSE), the Transformer again achieved a
minimum value (67,516,675.40), followed by LSTM (78,426,017.38) and GRU (91,656,112.95). A lower loss confirms that the model’s predictions were closer to the actual values on average.
_Visual Interpretation of Results_ The plots of MAE vs. Loss and RMSE vs. Loss further corroborated the numerical observations. In these graphs, the Transformer’s data points were
consistently located lower than those of LSTM and GRU models, visually indicating better performance. The MAE vs. Loss graph showed a near-linear trend for the Transformer, implying that its
error reduction was consistently aligned with minimizing loss. Similarly, the RMSE vs. Loss plot depicted that the Transformer maintained lower root mean squared deviations compared to
other models. The fine mini-grid style used in the graphs enabled a clearer visualization of minor variations in performance, and the x- and y-axis values extracted demonstrated a visible
gap where the Transformer’s metrics outperformed others.Thus, based on the mathematical analysis of MAE, RMSE, and Loss, and the visual evidence from the plotted graphs, it can be
conclusively proven that the Transformer model provided the most accurate and reliable forecasts for daily COVID-19 new cases. This superiority can be attributed to the Transformer’s ability
to capture complex long-range dependencies in the data without the sequential bottlenecks that hinder recurrent architectures like LSTM and GRU. The same has been depicted in the Table 4 To
statistically compare the performance of the deep learning models (LSTM, GRU, and Transformer) with regard to the Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and Loss, a
pair-wise comparison through a statistical test such as the paired t-test or Wilcoxon signed-rank test can be done. The test would identify if the differences in the measures of performance
between the models that are being observed are statistically significant or by chance. Because these models are tested on the same dataset, a paired t-test would be suitable to compare their
performance on each of the metrics.First, the null hypothesis (H₀) can be stated as “there is no significant difference between the MAE, RMSE, and Loss values of the models,” and the
alternative hypothesis (H₁) would be that “there is a significant difference between the performance of at least one model.” The paired t-test would then compare the MAE, RMSE, and Loss
between the models pairwise (LSTM vs. GRU, LSTM vs. Transformer, and GRU vs. Transformer) for each measure. Since the Transformer model performed better than both LSTM and GRU in all three
measures (lower MAE, RMSE, and Loss), the outcome of these tests will most likely have large differences, particularly between the Transformer and the other two models. The Statistical
significance can be tested using a _p_-value cutoff (usually _p_ < 0.05). If the _p_-value is less than this value for a particular metric comparison, the null hypothesis can be rejected,
meaning that the performance difference is statistically significant. For instance, if the Transformer-LSTM comparison has a _p_-value less than 0.05 for the MAE or RMSE, we can safely
conclude that the performance of the Transformer model is considerably superior to LSTM in terms of prediction accuracy and error reduction. Additionally, effect size measures like Cohen’s d
may also be computed to quantify the size of the difference in performance. The larger the value of Cohen’s d, the larger the difference in model performance would be, thereby
substantiating the argument that the Transformer model performs better than LSTM and GRU. According to the numeral discrepancies shown in the MAE, RMSE, and Loss values, the statistical
testing and effect size calculation provided strong evidence to affirm the performance superiority of the Transformer model in estimating COVID-19 daily new cases. Visual evidence from the
graphs shown, based on which the Transformer consistently outperforms, is that which confirms these statistical findings and supports the ability of the model to tackle intricate dependency
in the data. _Challenges_ The findings obtained through this study bring forth a number of challenges from the perspective of model performance and interpretation. As much as the Transformer
model outperformed both the LSTM model and the GRU model by way of MAE, RMSE, and Loss, an observation was made that there were huge errors committed in the prediction, especially where the
models were dealing with complex patterns in the COVID-19 daily new cases dataset. The LSTM model, though efficient in handling temporal dependencies, had greater MAE and RMSE values than
the Transformer, which implies that greater prediction errors were experienced during certain periods. This indicates that the ability of the model to generalize over various trends was
compromised, resulting in deviations from true values. The GRU model also suffered from the same problems, with greater MAE and RMSE than the LSTM, indicating that GRU was unable to capture
the temporal dynamics with the same precision. The vast disparity between the MAE and RMSE for LSTM and GRU also indicated that sporadic sheer errors were severely punished by the RMSE,
which affected their overall usability. Such models, limited by their sequential natures, could have struggled with learning long-distance dependencies without experiencing large errors when
sudden changes in the data, like spikes or sudden drops, were present. The Transformer model, however, although uniformly demonstrating better performance compared to all other models, also
struggled to reduce prediction errors to a degree that would be deemed negligible in real-world use. Regardless of its capacity for capturing long-distance dependencies and the lower Loss
figures, even predictions by the Transformer showed fluctuations which could potentially undermine real-world forecast accuracy. Such challenges are multiplied by the erratic and complicated
nature of the data on the COVID-19 pandemic, making precise prediction especially troublesome. In addition, while the graphical results do indeed show convincing outperformance of the
Transformer model, it is a challenge to interpret and graph results across various measures (MAE, RMSE, and Loss) and thus ensure the most important performance measures are emphasized.
These challenges necessitate further tuning of model designs and further development of different approaches, like ensemble learning or composite models, for even better, more precise
prediction of dynamic datasets like COVID-19 cases. CONCLUSION In this research paper, the prediction of new COVID-19 cases on a daily basis was tackled with the implementation of deep
learning models, i.e., LSTM, GRU, and Transformer models. Following appropriate preprocessing of the dataset and input sequence preparation, the models were created, trained, and tested
using standard performance measures such as MAE, RMSE, and Loss. The findings show that the Transformer model outperformed the LSTM and GRU models in all evaluation metrics. The capacity to
learn complex temporal relationships without the constraint of processing sequential data was a significant strength. Visual examination of the models’ performance also corroborated the
findings, with the Transformer recording lower error values and showing improved generalization on new data. It can be concluded from the experimental findings that Transformer-based models
offer a very efficient solution to time series forecasting problems such as COVID-19 case prediction. The findings show that future research in epidemic forecasting can benefit a great deal
from the use and extension of Transformer models. FUTURE WORK This section highlights the future work on COVID-19 datasets and also highlights the privacy and data security challenges in
healthcare IoT frameworks. * _Integration of Heterogeneous Data Sources_44 Future efforts must focus on merging data from diverse origins—such as hospital electronic health records, public
health databases, social determinants, and real-time IoT sensor feeds. This integration enables comprehensive analyses, facilitates cross-validation of insights, and supports more accurate
predictive models. By combining structured and unstructured data, researchers can capture a fuller picture of COVID-19 dynamics, ultimately leading to improved public health responses. *
_Standardization of Data Collection Protocols_45 Uniform data collection methods are critical to ensure consistency and interoperability. Establishing standard operating procedures across
healthcare institutions minimizes variability in data quality. Consistent protocols allow datasets from various regions and sources to be compared effectively, reducing errors during data
aggregation and enabling robust multi-center research collaborations. * _Advanced Anonymization Techniques_46 Given the sensitive nature of health data, it is imperative to develop
sophisticated anonymization methods that safeguard individual identities while preserving data utility. Techniques such as differential privacy, k-anonymity, and synthetic data generation
can help maintain confidentiality. Future work should focus on balancing data utility with privacy, ensuring that analytical insights remain valid without exposing personal information. *
_Real-Time Analytics Frameworks_28 Implementing frameworks capable of processing and analyzing data in real time is crucial for timely decision-making during health crises. These systems
would monitor outbreak trends as they develop, allowing public health officials to deploy resources more effectively. Future frameworks should support streaming data inputs from various
sources, thus enabling rapid response strategies during emerging outbreaks. * _Interoperability of IoT Devices_47 Healthcare IoT ecosystems consist of numerous devices—from wearable monitors
to smart sensors—each potentially using different communication protocols. Ensuring that these devices can seamlessly share data is essential. Future research should promote the development
of universal standards and middleware solutions that bridge protocol differences, thus fostering an integrated and responsive healthcare network. * _Scalable Data Infrastructure_48 As
COVID-19 datasets grow in volume and complexity, the underlying infrastructure must scale accordingly. Cloud computing and distributed systems can support high-volume data processing and
storage. Future infrastructure development should focus on flexibility and scalability, ensuring that both current and future data needs are met without compromising performance or
reliability. * _Enhanced Predictive Modeling_49 Machine learning and statistical models are vital for forecasting outbreak trends and resource requirements. Future work should refine these
models to incorporate evolving data streams and emerging viral variants. Continuous model validation and updates are necessary to ensure that predictions remain accurate over time, thus
enhancing preparedness for future waves or pandemics. * _Secure Data Sharing Protocols_50 Facilitating the secure exchange of data between institutions is paramount for collaborative
research. Developing robust protocols that encrypt data during transmission and authenticate users can help prevent unauthorized access. Future initiatives should work towards establishing
standardized, secure data sharing agreements that protect sensitive information while fostering innovation and collaboration. * _Privacy-Preserving Data Mining_51 Privacy-preserving
techniques—such as federated learning and secure multiparty computation—allow multiple entities to analyze data collectively without the need for centralization. These methods minimize the
risk of large-scale breaches by keeping raw data at its source. Future work should focus on refining these approaches to ensure they are both secure and effective in yielding actionable
insights. * _Robust Encryption for Data Protection_52 Protecting healthcare data at rest and in transit requires state-of-the-art encryption standards. Employing strong encryption algorithms
and regularly updating them in response to emerging vulnerabilities is essential. Future initiatives should also explore end-to-end encryption protocols specifically tailored for IoT
networks to ensure that sensitive information remains confidential. * _Dynamic Consent Management Systems_53 Patient consent for data usage should be adaptable to changes over time. Dynamic
consent models allow individuals to modify their data sharing preferences as circumstances or regulatory frameworks evolve. Future systems must integrate user-friendly interfaces and secure
mechanisms for consent management, ensuring that patients retain control over their personal health information throughout its lifecycle. * _Real-Time Intrusion Detection_37,54 With the
growing number of connected devices, healthcare IoT networks are increasingly vulnerable to cyberattacks. Real-time intrusion detection systems (IDS) can monitor network traffic, identify
suspicious behavior, and alert administrators instantly. Future efforts should prioritize the development and deployment of IDS that are optimized for the unique characteristics of
healthcare environments. * _Blockchain-Based Data Integrity_41,55 Blockchain technology offers a decentralized approach to recording data transactions in an immutable ledger. By
incorporating blockchain, healthcare systems can verify the authenticity and integrity of data, reducing the risk of tampering. Future applications may include tracking data provenance and
ensuring that each transaction is securely recorded, thereby enhancing overall trust in the system. * _Compliance with Global Privacy Regulations_56 Data handling practices must adhere to
international privacy regulations, such as GDPR, HIPAA, and other regional standards. Achieving and maintaining compliance can be challenging due to differing requirements across
jurisdictions. Future work should focus on developing adaptive compliance frameworks that automatically adjust to regional regulatory changes, ensuring that privacy standards are
consistently met. * _Continuous Risk and Vulnerability Assessments_57 Cyber threats evolve constantly, necessitating regular vulnerability assessments and risk evaluations. Ongoing security
audits can identify weak points within the system and prompt timely countermeasures. Future initiatives should implement automated risk assessment tools that continuously monitor system
integrity and flag potential security breaches before they occur. * _Ethical Integration of Artificial Intelligence_58 The application of AI in healthcare requires a strong ethical
foundation to prevent biases and ensure equitable treatment of all patient groups. Integrating ethical guidelines into AI development can help mitigate risks associated with discriminatory
practices. Future work should emphasize transparency, fairness, and accountability in AI systems, ensuring that these technologies serve all communities without prejudice. * _Adaptive
Security Measures Against Evolving Threats_59 The cybersecurity landscape is dynamic, with new threats emerging regularly. Security systems must be designed to adapt rapidly to these
changes. Future security strategies should include machine-learning-based threat detection and automated patch management systems, ensuring that defenses are continuously updated to
counteract novel attack vectors. * _Ensuring Data Accuracy and Validation_60 Reliable analysis hinges on the accuracy of the underlying data. Ensuring data quality involves rigorous
validation processes, regular audits, and the use of error-detection algorithms. Future work should focus on developing standardized validation protocols and employing cross-verification
techniques to maintain the integrity of COVID-19 datasets. * _Cost-Effective Security Implementations_37,61 While robust security measures are essential, they must also be economically
viable—especially for smaller healthcare providers with limited budgets. Future research should explore cost-effective solutions that do not compromise on security. This may include
open-source security tools, cloud-based security services, and scalable security architectures that can be tailored to the financial constraints of various institutions. * _User Education
and Training_62 The human factor remains one of the most significant vulnerabilities in data security. Regular training programs can equip healthcare personnel with the necessary skills to
recognize and mitigate security threats. Future initiatives should develop comprehensive education campaigns that emphasize best practices in data handling, cyber hygiene, and incident
response, thereby reducing the risk of human-induced breaches. * _Edge Computing for Reduced Latency_63 Processing data at the edge of the network—closer to where it is generated—can
significantly reduce latency and enhance the speed of critical decision-making. Edge computing enables real-time data analysis, which is particularly valuable in time-sensitive healthcare
scenarios. Future work should investigate integrating edge devices with centralized systems to create a hybrid architecture that balances speed and data centralization. * _Resilient Backup
and Recovery Systems_64 Ensuring data continuity in the event of system failures or cyberattacks requires robust backup and recovery protocols. These systems must be designed to restore data
quickly and accurately, minimizing downtime and preventing data loss. Future efforts should focus on creating resilient backup architectures that incorporate redundancy, regular testing,
and automated recovery processes. * _Adoption of Standardized Security Frameworks_40,65 The implementation of industry-recognized security frameworks—such as ISO/IEC 27001 or NIST—can
provide a systematic approach to managing data security risks. Standardized frameworks ensure that best practices are followed consistently across all levels of an organization. Future work
should encourage the adoption of these frameworks to streamline security processes and foster a culture of continuous improvement in risk management. * _Cross-Sector Collaboration on Data
Security_66 The complexity of healthcare IoT systems often necessitates collaboration between public institutions, private enterprises, and technology vendors. Such cross-sector partnerships
can drive the development of innovative security solutions and promote shared learning. Future initiatives should focus on creating collaborative networks that facilitate knowledge
exchange, joint research projects, and the pooling of resources to tackle common security challenges. * _Continuous Monitoring and Auditing Mechanisms_42,67 Ongoing surveillance of data
access and usage is essential for early detection of anomalies and ensuring regulatory compliance. Automated monitoring systems, coupled with periodic audits, can help identify suspicious
activity and trigger prompt remedial actions. Future work should invest in advanced auditing tools that leverage artificial intelligence to continuously scan for deviations from established
security norms, thereby maintaining the integrity of healthcare IoT ecosystems. DATA AVAILABILITY The datasets generated during and/or analysed during the current study are not publicly
available but are available from the corresponding author on reasonable request. REFERENCES * Konwar, A. N. & Borse, V. Current status of point-of-care diagnostic devices in the Indian
healthcare system with an update on COVID-19 pandemic. _Sens. Int._ 1, 100015 (2020). Article PubMed PubMed Central Google Scholar * Prasad, V. K. & Bhavsar, M. D. Monitoring and
prediction of SLA for IoT based cloud. _Scalable Comput. Pract. Exp._ 21(3), 349–358 (2020). Article Google Scholar * Javaid, M. & Khan, I. H. Internet of Things (IoT) enabled
healthcare helps to take the challenges of COVID-19 Pandemic. _J. Oral Biol. Craniofacial Res._ 11(2), 209–214 (2021). Article Google Scholar * Prasad, V. K., Tanwar, S. & Bhavsar, M.
D. Advance cloud data ana- lytics for 5G enabled IoT. In _Blockchain for 5G-Enabled IoT_ 159 (Springer, 2021). Chapter Google Scholar * Rana, A. K., Salau, A. O., Sharma, S., Tayal, S.
& Gupta, S. _Internet of Things: Energy, Industry, and Healthcare_ 1–330 (CRC Press, 2021). https://doi.org/10.1201/9781003140443. Book Google Scholar * Rout, N. K. et al. Estimation
of COVID-19 secondary effects on cardiac anomalies using deep transfer learning approach. _Sādhanā_ 50, 91. https://doi.org/10.1007/s12046-025-02739-5 (2025). Article CAS Google Scholar *
Schwickert, L. et al. Fall detection with body-worn sensors. _Z. Für Gerontol. Und Geriatr._ 46(8), 706–719 (2013). Article CAS Google Scholar * He, J., Geng, Y. & Pahlavan, K.
Modeling indoor TOA ranging error for body mounted sensors. In _2012 IEEE 23rd International Symposium on Personal, Indoor and Mobile Radio Communications-(PIMRC)_. (IEEE, 2012). * Mancini,
M. et al. Mobility lab to assess balance and gait with synchronized body-worn sensors. _J. Bioeng. Biomed. Sci._ https://doi.org/10.4172/2155-9538.S1-007 (2011). Article PubMed PubMed
Central Google Scholar * Horak, F. B. & Mancini, M. Objective biomarkers of balance and gait for Parkinson’s disease using body-worn sensors. _Mov. Disord._ 28(11), 1544–1551 (2013).
Article PubMed PubMed Central Google Scholar * Otto, C., Milenkovic, A., Sanders, C. & Jovanov, E. System architecture of a wireless body area sensor network for ubiquitous health
monitoring. _J. Mob. Multimed._ 1(4), 307–326 (2006). Google Scholar * Liew, C.-H., Ong, S.-Q. & Ng, D.-E. Utilizing machine learning to predict hospital admissions for pediatric
COVID-19 patients (PrepCOVID-Machine). _Sci. Rep._ 15(1), 3131 (2025). Article CAS PubMed PubMed Central Google Scholar * Xiang, Z. et al. Machine learning based prediction models for
the prognosis of COVID-19 patients with DKA. _Sci. Rep._ 15(1), 2633 (2025). Article CAS PubMed PubMed Central Google Scholar * Park, H. Y. et al. Application of machine learning for
delirium prediction and analysis of associated factors in hospitalized COVID-19 patients: A comparative study using the Korean Multidisciplinary cohort for delirium prevention (KoMCoDe).
_Int. J. Med. Inform._ 195, 105747 (2025). Article PubMed Google Scholar * Farahat, I. S. et al. An AI-based novel system for predicting respiratory support in COVID-19 patients through
CT imaging analysis. _Sci. Rep._ 14(1), 851 (2024). Article ADS CAS PubMed PubMed Central Google Scholar * Elaraby, M. E., Ewees, A. A. & Anter, A. M. A robust IoT-based cloud
model for COVID-19 prediction using advanced machine learning technique. _Biomed. Signal Process. Control_ 87, 105542 (2024). Article Google Scholar * Cai, J. et al. Developing deep LSTMs
with later temporal attention for predicting COVID-19 severity, clinical outcome, and antibody level by screening serological indicators over time. _IEEE J. Biomed. Health Inform._ 28, 4204
(2024). Article PubMed Google Scholar * Solayman, S., Aumi, S. A., Mery, C. S., Mubassir, M. & Khan, R. Automatic COVID-19 prediction using explainable machine learning techniques.
_Int. J. Cogn. Comput. Eng._ 4, 36–46 (2023). Google Scholar * Li, S. et al. Development and external evaluation of predictions models for mortality of COVID-19 patients using machine
learning method. _Neural Comput. Appl._ 35, 1–10 (2023). CAS Google Scholar * Hasan, I., Dhawan, P., Rizvi, S. A. & Dhir, S. Data analytics and knowledge management approach for
COVID-19 prediction and control. _Int. J. Inf. Technol._ 15(2), 937–954 (2023). PubMed Google Scholar * Famiglini, L., Campagner, A., Carobene, A. & Cabitza, F. A robust and
parsimonious machine learning method to predict ICU admission of COVID-19 patients. _Med. Biol. Eng. Comput._ https://doi.org/10.1007/s11517-022-02543-x (2022). Article PubMed PubMed
Central Google Scholar * Liu, T., Siegel, E. & Shen, D. Deep learning and medical image analysis for COVID-19 diagnosis and prediction. _Annu. Rev. Biomed. Eng._ 24(1), 179–201 (2022).
Article CAS PubMed Google Scholar * Xu, W. et al. Risk factors analysis of COVID-19 patients with ARDS and prediction based on machine learning. _Sci. Rep._ 11(1), 2933 (2021). Article
ADS CAS PubMed PubMed Central Google Scholar * Singh, S., Parmar, K. S. & Kumar, J. Development of multi-forecasting model using Monte Carlo simulation coupled with wavelet
denoising-ARIMA model. _Math. Comput. Simul._ 230, 517–540 (2025). Article MathSciNet MATH Google Scholar * Wang, X., Feng, M., Qiu, J., Jinjin, Gu. & Zhao, J. From news to forecast:
Integrating event analysis in llm-based time series forecasting with reflection. _Adv. Neural. Inf. Process. Syst._ 37, 58118–58153 (2025). Google Scholar * Prasad, V. K. et al. ABV-COVID:
An ensemble forecasting model to predict availability of beds and ventilators for COVID-19 like pandemics. _Ieee Access_ 10, 74131–74151 (2022). Article PubMed Google Scholar * He, Y. et
al. Foundation model for advancing healthcare: Challenges, opportunities and future directions. _IEEE Rev. Biomed. Eng._ 18, 172 (2024). Article Google Scholar * Sharma, N. & Kaushik,
P. Integration of AI in healthcare systems—A discussion of the challenges and opportunities of integrating AI in healthcare systems for disease detection and diagnosis. In _AI in Disease
Detection: Advancements and Applications_ 239–263 (Wiley, 2025). Chapter Google Scholar * Fernandez-Gonzalo, R., Schneider, S., Heer, M. & Liphardt, A. M. The search for the ultimate
exercise countermeasure to preserve crew health and ensure mission success in long-duration spaceflight. _Exp. Physiol._ https://doi.org/10.1113/EP091737 (2025). Article PubMed Google
Scholar * Golestani, R., Farahani, F. K. & Peters, P. Exploring barriers to accessing health care services by young women in rural settings: A qualitative study in Australia, Canada,
and Sweden. _BMC Public Health_ 25(1), 213 (2025). Article PubMed PubMed Central Google Scholar * Tan, C. Y., Jang, S. T., Lam, S. M., An, A. Q. & Lo, U. K. Teaching and learning
challenges due to the COVID-19 pandemic: A systematic review. _Educ. Res. Rev._ 47, 100667 (2025). Article Google Scholar * Stoltenberg, A. S., Ranieri, V., Dahlen, H. K. N., Nastouli, E.,
SAFER Investigators Byott Matt 9 Edwards Sarah JL 8 Frampton Daniel 10 Gilson Richard 11 Hayward Andrew 12 Heaney Jude 9 Houlihan Catherine 9 Lorencatto Fabiana 4 Manley Ed 13 Michie Susan
4 Spyer Moira 9 Vora Nina 11 Walker Naomi 13 Nastouli Eleni 4 5, Roger Almvik, Sarah JL Edwards, and Sunjeev K. Kamboj. Perceived coercion amongst healthcare workers during the COVID-19
pandemic. _Sci. Rep_. 15 (1), 4701 (2025). * Roos, P. et al. Protecting oneself while supporting the organisation: A longitudinal exploratory study of healthcare workers’ coping strategies
and organisational resilience processes in the first year of the COVID-19 pandemic. _Saf. Sci._ 181, 106702 (2025). Article Google Scholar * Papa, A. Gaps in mental health care–Seeking
among health care providers during the COVID-19 pandemic—United States, September 2022–May 2023. In _MMWR. Morbidity and Mortality Weekly Report_ 74 (2025). * Komasawa, M. et al.
Determinants maintaining healthcare personnel’s motivation during COVID-19 pandemic in Uganda. _Sci. Rep._ 15(1), 3373 (2025). Article CAS PubMed PubMed Central Google Scholar *
Khadidos, A. O., Manoharan, H., Khadidos, A. O., Selvarajan, S. & Singh, S. Synthetic healthcare data utility with biometric pattern recognition using adversarial networks. _Sci. Rep._
15(1), 9753 (2025). Article CAS PubMed PubMed Central Google Scholar * Fuladi, S. et al. A reliable and privacy-preserved federated learning framework for real-time smoking prediction
in healthcare. _Front. Comput. Sci._ 6, 1494174 (2025). Article Google Scholar * Selvarajan, S. et al. Diagnostic behavior analysis of profuse data intrusions in cyber physical systems
using adversarial learning techniques. _Sci. Rep._ 15(1), 7287 (2025). Article CAS PubMed PubMed Central Google Scholar * Fuladi, S. et al. Linear regressive weighted Gaussian kernel
liquid neural network for brain tumor disease prediction using time series data. _Sci. Rep._ 15(1), 5912 (2025). Article Google Scholar * Natarajan, M., Bharathi, A., Varun, C. S. &
Selvarajan, S. Quantum secure patient login credential system using blockchain for electronic health record sharing framework. _Sci. Rep._ 15(1), 4023 (2025). Article CAS PubMed PubMed
Central Google Scholar * Ghadi, Y. Y. et al. The role of blockchain to secure internet of medical things. _Sci. Rep._ 14(1), 18422 (2024). Article CAS PubMed PubMed Central Google
Scholar * Ghadi, Y. Y. et al. Enhancing patient healthcare with mobile edge computing and 5G: Challenges and solutions for secure online health tools. _J. Cloud Comput._ 13(1), 93 (2024).
Article Google Scholar * Mazhar, T. et al. Analysis of integration of IoMT with blockchain: Issues, challenges and solutions. _Discover Internet Things_ 4(1), 1–36 (2024). Article Google
Scholar * Batta, P. Advancements and challenges in multimodal data fusion for bioinformatics AI. In _Multimodal Data Fusion for Bioinformatics Artificial Intelligence_ 1–12 (Wiley, 2025).
Google Scholar * Ganatra, H. A. Machine learning in pediatric healthcare: Current trends, challenges, and future directions. _J. Clin. Med._ 14(3), 807 (2025). Article CAS PubMed PubMed
Central Google Scholar * Natarajan, A. & Shanthi, N. Optimizing healthcare big data privacy with scalable subtree-based L-Anonymization in cloud environments. _Wirel. Netw._ 31, 1–16
(2025). Article Google Scholar * Rahman, Md Habibur, Tanjila Islam, Mohammad Hamid Hasan Amjad, Md Shihab Sadik Shovon, Md Estehad, Md Rahatul Ashakin Chowdhury, Bayazid Hossain, Proshanta
Kumar Bhowmik, Md Nurullah, & Rahman, A. Impact of Internet of Things (IoT) on Healthcare in transforming patient care and overcoming operational challenges. * Tlemçani, K. et al.
Empowering Diabetes management through blockchain and edge computing: A systematic review of healthcare innovations and challenges. _IEEE Access_ 13, 14426 (2025). Article Google Scholar *
Olalekan Kehinde, A. Leveraging machine learning for predictive models in healthcare to enhance patient outcome management. _Int. Res. J. Mod. Eng. Technol. Sci._ 7(1), 1465 (2025). Google
Scholar * Ogundokun, R.O., Lukman, A.F., Kibria, G.B.M., Awotunde, J.B. & Aladeitan, B.B Predictive modelling of COVID-19 confirmed cases in Nigeria. _Infect. Disease Model_. 5, 543–548
(2020). * Mehla, R., Garg, R. & Khan, M. A. Privacy-preserving solution for data sharing in IoT based smart consumer electronic devices for healthcare. _IEEE Trans. Consum. Electron._
https://doi.org/10.1109/TCE.2025.3540642 (2025). Article Google Scholar * Kothai, G., Harish, P., Ajinesh, D. & Sathyanarayanan, V. Optimizing data privacy and scalability in health
monitoring: Leveraging, encryption, blockchain, and cloud technologies. In _Critical Phishing Defense Strategies and Digital Asset Protection_ 245–270 (IGI Global Scientific Publishing,
2025). Chapter Google Scholar * Lay, W., Gasparini, L., Siero, W. & Hughes, E. K. A rapid review of the benefits and challenges of dynamic consent. _Res. Ethics_ 21(1), 180–202 (2025).
Article Google Scholar * Naghib, A., Gharehchopogh, F. S. & Zamanifar, A. A comprehensive and systematic literature review on intrusion detection systems in the internet of medical
things: Current status, challenges, and opportunities. _Artif. Intell. Rev._ 58(4), 1–88 (2025). Article Google Scholar * Arya, S., Aggarwal, S. & Ansar, S. A. Transforming healthcare
with blockchain and digital twins: innovations in IoMT for smart hospitals. In _Blockchain and Digital Twin for Smart Hospitals_ 497–510 (Elsevier, 2025). Chapter Google Scholar * Prasad,
V.K. et al. CIA-CVD: cloud based image analysis for COVID-19 vaccination distribution. _J. Cloud Comput._ 12, 163. https://doi.org/10.1186/s13677-023-00539-y (2023). * Czekster, R. M.,
Webber, T., Furstenau, L. B. & Marcon, C. Dynamic risk assessment approach for analysing cyber security events in medical IoT networks. _Internet Things_ 29, 101437 (2025). Article
Google Scholar * ElHassan, B. T. & Arabi, A. A. Ethical forethoughts on the use of artificial intelligence in medicine. _Int. J. Ethics Syst._ 41(1), 35–44 (2025). Article Google
Scholar * Deon, L. & Best, T. Zero trust security and cloud security: A modern approach to cyberattack prevention. (2025). * Rahim, M. J., Afroz, A. & Akinola, O. Predictive
analytics in healthcare: Big data, better decisions. _Int. J. Sci. Res. Mod. Technol._ 4(1), 1–21 (2025). Google Scholar * Surya, N. & Someshwar, H. P. Low-Cost telerehabilitation in
low-and middle-income countries (LMICs): Overcoming barriers to access and improving healthcare delivery. _NeuroRehabilitation_ https://doi.org/10.1177/10538135241303349 (2025). Article
PubMed Google Scholar * Davids, J. A. & Hughes, M. A usability study of a set of multidisciplinary professional standards for health professions educators within a healthcare setting.
_Discov. Med._ 2(1), 1–17 (2025). Article Google Scholar * Anand, J. & Karthikeyan, B. Dynamic priority-based task scheduling and adaptive resource allocation algorithms for efficient
edge computing in healthcare systems. _Results Eng._ 25, 104342 (2025). Article Google Scholar * Oluwagbade, E. & Odumbo, O. R. Building resilient healthcare distribution networks:
Adapting to crises, securing supplies and improving scalability. (2025). * Misra, P., Panigrahi, N., Gopal Krishna Patro, S., Salau, A. O. & Aravinth, S. S. PETLFC: Parallel ensemble
transfer learning based framework for COVID-19 differentiation and prediction using deep convolutional neural network models. _Multimed. Tools Appl._ 83(5), 14211–14233.
https://doi.org/10.1007/s11042-023-16084-4 (2024). Article Google Scholar * Penttilä, V., den Hond, F., Fougère, M. & Solitander, N. Decontesting corporate responsibility for
collaboration: A case study of a cross-sector coalition lobbying for human rights regulation. _Bus. Soc._ https://doi.org/10.1177/00076503251317253 (2025). Article Google Scholar *
Kathole, A. B., Mirge, A. S. & Jadhao, A. P. Advancing electronic medical data protection through blockchain innovation. In _Blockchain and Digital Twin for Smart Hospitals_ 193–205
(Elsevier, 2025). Chapter Google Scholar Download references FUNDING Authors declare no funding for this research. AUTHOR INFORMATION AUTHORS AND AFFILIATIONS * Computer Science
Department, Institute of Technology, Nirma University, Ahmedabad, India Vivek Kumar Prasad & Divyang Yadav * Department of Computer Science, Rajendra University, Balangir, Odisha, India
Debabrata Dansana * School of Engineering, Sreenidhi University, Hyderabad, Telangana, 501301, India S. Gopal Krishna Patro * Department of Electrical/Electronics and Computer Engineering,
Afe Babalola University, Ado-Ekiti, Nigeria Ayodeji Olalekan Salau * NIST Institute of Science and Technology (Autonomous), Berhampur, India Brojo Kishore Mishra * Saveetha School of
Engineering, Saveetha Institute of Medical and Technical Sciences, Chennai, Tamil Nadu, India Ayodeji Olalekan Salau * Chitkara University Institute of Engineering and Technology, Chitkara
University, Rajpura, Punjab, India Ayodeji Olalekan Salau Authors * Vivek Kumar Prasad View author publications You can also search for this author inPubMed Google Scholar * Debabrata
Dansana View author publications You can also search for this author inPubMed Google Scholar * S. Gopal Krishna Patro View author publications You can also search for this author inPubMed
Google Scholar * Ayodeji Olalekan Salau View author publications You can also search for this author inPubMed Google Scholar * Divyang Yadav View author publications You can also search for
this author inPubMed Google Scholar * Brojo Kishore Mishra View author publications You can also search for this author inPubMed Google Scholar CONTRIBUTIONS Vivek Kumar Prasad:
Conceptualization, Methodology, Software, Writing- Original draft preparation Debabrata Dansana: Data curation, Visualization, Investigation. S Gopal Krishna Patro: Investigation,
Visualization, Investigation. Ayodeji Olalekan Salau: Data curation, Methodology, Writing- Reviewing, Editing and Validation. Divyang Yadav: Data curation, Visualization, Investigation.
Brojo Kishore Mishra: Validation. CORRESPONDING AUTHOR Correspondence to Ayodeji Olalekan Salau. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests.
ADDITIONAL INFORMATION PUBLISHER’S NOTE Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. RIGHTS AND PERMISSIONS OPEN
ACCESS This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and
reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you
modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in
this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative
Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a
copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Prasad, V.K., Dansana, D., Patro, S.G.K. _et al._
IoT-based bed and ventilator management system during the COVID-19 pandemic. _Sci Rep_ 15, 19163 (2025). https://doi.org/10.1038/s41598-025-03144-y Download citation * Received: 11 January
2025 * Accepted: 19 May 2025 * Published: 31 May 2025 * DOI: https://doi.org/10.1038/s41598-025-03144-y SHARE THIS ARTICLE Anyone you share the following link with will be able to read this
content: Get shareable link Sorry, a shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature SharedIt content-sharing initiative
KEYWORDS * COVID-19 * Health care * IoT * Body sensor network * Cloud computing * Prediction * ARIMA model * Trasnformer