Optimizing IoT Fog Computing Preprocessing
Optimizing IoT Fog Computing Preprocessing
Corresponding Author:
Nandini Gowda Puttaswamy
Department of Computer Science and Engineering, Sapthagiri College of Engineering
Bengaluru, India
Email: nandini.educator1@[Link]
1. INTRODUCTION
The internet of things (IoT) has dramatically transformed how we interact with the physical world,
integrating intelligence into everyday objects and enabling them to communicate and make decisions.
This widespread adoption of IoT has led to the generation of massive amounts of data at the edge of the
network, necessitating innovative approaches to data processing and management. Fog computing, which
extends cloud computing to the edge of the network, has emerged as a pivotal technology in this context.
It aims to reduce latency, improve bandwidth utilization, and enhance the overall efficiency of IoT systems
by processing data closer to its source [1], [2].
Figure 1 shows a security architecture involving three entities such as the user, a cloud server, and a
trusted third party. The working principal centers around mutual authentication, a security mechanism
ensuring that both the user and the cloud server verify each other's identities before initiating any
communication. Here, the trusted third party plays a crucial role, possibly as a certificate authority or
authentication server, that both the user and the cloud server trust. This entity could facilitate the exchange of
credentials or cryptographic keys that enable mutual authentication [3]. Upon successful authentication,
a secure channel is established between the user and the cloud server, allowing for safe data exchange,
service requests, and transactions, all under the supervision of the trusted third party to prevent unauthorized
access and ensure data integrity and confidentiality. This framework is fundamental to preserving security in
cloud computing, where data and resources are accessed over potentially insecure networks [4], [5].
Recent trends in IoT and fog computing highlight a shift towards more autonomous, intelligent
systems capable of real-time decision-making. However, the sheer volume and velocity of data generated by
IoT devices present significant challenges in real-time data preprocessing. Traditional cloud-centric models
often fail to meet the requirements of latency-sensitive applications, leading to a research gap in developing
more efficient, adaptive, and scalable real-time data preprocessing methods within the fog computing
paradigm [6].
The application of machine learning algorithms in optimizing these preprocessing tasks holds
promise in bridging this gap. By leveraging machine learning, systems can dynamically adapt to changing
data patterns and network conditions, ensuring efficient data processing and resource utilization. However,
despite its potential, the integration of machine learning into fog computing for IoT systems is still in its
nascent stages, with several challenges to overcome. These include ensuring data privacy, managing resource
constraints, and maintaining system adaptability in highly dynamic environments [7]–[10].
The convergence of IoT, fog computing, and machine learning opens up new avenues for research
and development. By addressing the current limitations and harnessing the strengths of these technologies,
we can pave the way for more responsive, efficient, and intelligent IoT systems. Such advancements have
profound implications across various sectors, including healthcare, smart cities, and industrial automation,
where real-time data processing and decision-making are crucial.
In exploring the landscape of real-time data preprocessing in IoT-based fog computing
environments, several noteworthy contributions have been made in recent years. The integration of machine
learning algorithms for enhancing efficiency and adaptability has been a focal point of research. However,
while these studies have laid a solid foundation, they also highlight various challenges and limitations that
warrant further investigation. Varun et al. [11] presented a framework leveraging convolutional neural
networks (CNNs) for data preprocessing in fog computing nodes. Their method significantly improved data
processing speeds by automatically filtering irrelevant data before it reached the cloud. However, a notable
drawback is the study acknowledged the high computational overhead of CNNs, making it less viable for
devices with limited processing capabilities. Gowrishankar et al. [12] introduced an adaptive algorithm based
on reinforcement learning that dynamically allocates resources in fog computing environments to optimize
data preprocessing tasks. Their approach demonstrated improved system adaptability and resource efficiency.
However, a drawback of the study is that the complexity of the algorithm led to difficulties in real-time
implementation, especially in highly volatile IoT environments. Marković et al. [13] proposed a novel data
anonymization technique within the fog layer to address privacy concerns during the preprocessing of
sensitive information. While their method effectively enhanced data privacy, a drawback was that found to
introduce latency, particularly with large datasets, which could compromise the real-time processing
requirements of IoT applications.
Khan et al. [14] explored the use of edge-based machine learning models to preprocess data locally,
reducing the need for data transmission to the cloud. Their work showed promising results in decreasing
latency and bandwidth usage. However, a drawback highlighted in the study was the challenge of
Optimizing real-time data preprocessing in IoT-based fog computing using … (Nandini Gowda Puttaswamy)
1902 ISSN: 2252-8938
maintaining model accuracy over time without regular updates, which could require significant data transfers,
thus negating some of the benefits. Saravanan et al. [15] developed a distributed ledger technology (DLT)-
based approach for secure data preprocessing in fog computing, aiming to improve both transparency and
security. While their solution effectively addressed trust issues, a drawback was that it introduced substantial
computational and storage overhead, questioning its scalability in larger IoT deployments. These studies
illustrate the dynamic and evolving nature of research in real-time data preprocessing within IoT-based fog
computing environments. They underscore the critical balance between enhancing processing efficiency,
ensuring privacy and security, and maintaining system adaptability and scalability. As such, they highlight
the need for innovative solutions that can address these multifaceted challenges in a holistic manner.
2. PROPOSED METHOD
Figure 2 shows the proposed methodology, to establish a multi-tiered IoT-based fog computing
model. Data collection commences with harvesting raw inputs from IoT devices, simulating a high-velocity
data stream. The preprocessing phase involves algorithmic noise filtering, feature extraction, and
normalization to prepare datasets for machine learning application [16], [17]. We select machine learning
algorithms suited to real-time analytics, emphasizing decision efficiency and computational lightness.
Supervised learning models are trained on a partitioned dataset, employing cross-validation to mitigate
overfitting while optimizing performance parameters [18]–[20].
Figure 2. The proposed methodology of real-time data preprocessing in IoT-based fog computing using
machine learning algorithms (IoT-FCML)
Post-training, machine learning models are embedded within fog nodes. Their performance is
assessed through key metrics: latency, throughput, and resource allocation. These are benchmarked against
conventional preprocessing paradigms to evaluate the efficacy and improvements our machine learning-
driven method offers. Security protocols are integral, ensuring data integrity and confidentiality. The system
undergoes iterative optimization, responsive to empirical data and user-centric feedback, striving for
enhanced operational excellence within the fog computing sphere [21]–[23].
optimization, the processed data can be sent to the cloud data center for further analysis or long-term storage.
The cloud layer offers more extensive computational resources and storage capacity, suitable for complex
analytics and historical data analysis that the fog layer cannot perform due to resource constraints.
Finally, the performance analysis phase evaluates the efficiency and effectiveness of the
preprocessing and optimization steps. This analysis considers factors like latency, throughput, and resource
utilization, ensuring that the system meets the real-time processing requirements of IoT applications. The
proposed method leverages the strengths of fog computing-proximity to data sources and reduced latency,
with the extensive processing power of cloud computing, providing a balanced and optimized approach to
data management in IoT networks.
Optimizing real-time data preprocessing in IoT-based fog computing using … (Nandini Gowda Puttaswamy)
1904 ISSN: 2252-8938
Where 𝐷(𝑡) is the total system demand at time 𝑡, and 𝑑𝑖 (𝑡) is the demand of the 𝑖 𝑡ℎ IoT device at time 𝑡.
Where 𝐿 is the total latency, 𝐿𝑓𝑜𝑔 is the processing latency in the fog layer, 𝐿𝑛𝑒𝑡𝑤𝑜𝑟𝑘 is the network latency,
and 𝐿𝑐𝑙𝑜𝑢𝑑 is the processing latency in the cloud layer. The goal is to minimize L, especially 𝐿𝑓𝑜𝑔 as it's the
first processing layer for real-time data.
Where 𝑇 is the throughput, 𝑉 is the volume of processed data, and 𝑅 is the available resources. Maximizing 𝑇
indicates improved system performance.
Where 𝑃(𝑑𝑖 ) is the privacy-preserving function for data 𝑑𝑖 from the 𝑖 𝑡ℎ IoT device, and 𝑘 is the encryption
key. This equation doesn't directly reduce latency or improve throughput but is essential for ensuring data
confidentiality.
Where 𝐸 is the efficiency, 𝑈 is the utilization of resources, and 𝑅 is the total available resources. 𝐸 should be
maximized under the constraint that 𝑈 ≤ 𝑅, ensuring no resource is over-utilized.
Where 𝛼, 𝛽, 𝛾 are weighting factors indicating the importance of each objective (latency, throughput, and
efficiency).
2.3. Proposed optimizing real-time data preprocessing in IoT-based fog computing using machine
learning algorithms
Creating an overarching mathematical equation that encapsulates the optimization of IoT real-time
data preprocessing using machine learning, while taking into account factors such as system demand, latency,
throughput, data privacy, and resource efficiency, involves synthesizing the individual objectives into a
singular objective function. This unified equation aims to balance these multiple aspects through weighted
parameters, reflecting their relative importance to the system's overall performance and objectives. Optimize
proposed algorithm as calculated using (7).
1
Optimize(𝑂) = 𝑤1 × (∑𝑛 ) + 𝑤2 ∑𝑛𝑖=0 𝑇𝑖 + 𝑤3 ∑𝑛𝑖=0 𝐸𝑖 − 𝑤4 ∑𝑛𝑖=0 𝐷𝑖 (𝑡) − 𝑤5 ∑𝑛𝑖=0 𝐶(𝑃𝑖 ) (7)
𝑖=0 𝐿𝑖
𝐿𝑖 , 𝑇𝑖 , 𝐸𝑖 , 𝐷𝑖 and 𝐶(𝑃𝑖 ) now represent the latency, throughput, resource efficiency, system demand, and cost
of privacy for the 𝑖 𝑡ℎ IoT device, respectively. The sums ∑𝑛𝑖=0. aggregate the contributions of each device
from the 0th to the nth, offering a comprehensive view of the entire IoT ecosystem. The optimization
objective (𝑂) now directly accounts for the performance and demands of each individual device, ensuring
that the optimization strategy is effective across the entire network of IoT devices.
Table 2 demonstrates that the proposed optimization method surpasses the conventional methods
across all evaluated performance metrics. It emphasizes the effectiveness of the proposed method in lowering
latency, boosting throughput, improving resource efficiency, and reducing the overhead involved in securing
data privacy. Figure 4 presents a performance comparison of the proposed method with conventional
methods in relation to system demand.
Table 3 encompasses a broader set of performance metrics beyond efficiency, including latency,
throughput, data privacy overhead, resource utilization, scalability, and reliability. It provides a clear
comparison between the proposed optimization method and the other conventional methods. Figure 5
presents a performance comparison of the proposed method with conventional methods in relation to
efficiency.
Optimizing real-time data preprocessing in IoT-based fog computing using … (Nandini Gowda Puttaswamy)
1906 ISSN: 2252-8938
Figure 4. The performance analysis of the proposed method compared to conventional methods in relation to
system demand
Figure 5. The performance analysis of the proposed method compared to conventional methods in relation to
efficiency
Table 4 provides a comparison of various data privacy-related performance metrics across the
proposed optimization method and the three conventional methods. The metrics include the overheads for
data encryption and anonymization, compliance with privacy policies, overhead for secure data transmission,
and latency due to data access controls. Figure 6 presents a performance comparison of the proposed method
with conventional methods in relation to data privacy.
Figure 6. The performance analysis of the proposed method compared to conventional methods in relation to
data privacy
4. CONCLUSION
The paper presented a IoT-FCML for real-time data preprocessing in IoT-based fog computing,
showing marked improvements over traditional approaches. Specifically, the proposed method enhanced
latency by approximately 0.26%, increased throughput by up to 0.32%, improved resource efficiency by
0.20%, and reduced data privacy overhead by 0.64%, reflecting significant advancements in both
performance and security. These enhancements signify a substantial step forward in developing adaptive,
efficient IoT systems, particularly in dynamic and resource-constrained fog computing environments.
The integration of machine learning algorithms has proven to be a pivotal factor in the system's ability to
dynamically adjust to varying data streams and operational demands, ultimately leading to smarter, more
responsive IoT infrastructures. With these results, the paper sets a precedent for future research to expand
upon, indicating a bright horizon for the intersection of IoT, fog computing, and intelligent data processing
techniques. This research promises advancements in machine learning algorithms tailored for IoT scalability,
sophisticated privacy preservation techniques, enhanced resource allocation strategies, and the exploration of
edge computing integration. These developments aim to bolster the IoT ecosystem, enabling it to handle
growing data volumes and complexity with greater efficiency and security.
ACKNOWLEDGMENTS
The author would like to thank East Point College of Engineering and Technology, Sapthagiri
College of Engineering, BNM Institute of Technology, Visvesvaraya Technological University (VTU),
Belagavi, for all the support and encouragement provided by them to take up this research work and publish
this paper.
FUNDING INFORMATION
Authors state no funding involved.
Name of Author C M So Va Fo I R D O E Vi Su P Fu
Nandini Gowda Puttaswamy ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓
Anitha Narasimha Murthy ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓ ✓
INFORMED CONSENT
We have obtained informed consent from all individuals included in this study.
ETHICAL APPROVAL
The research related to human use has been conducted in compliance with all relevant national
regulations and institutional policies, in accordance with the tenets of the Helsinki Declaration, and has been
approved by the authors' institutional review board.
DATA AVAILABILITY
The authors confirm that the data supporting the findings of this study are available within the article
and its supplementary materials.
REFERENCES
[1] S. Jha and D. Tripathy, “Low latency consistency based protocol for fog computing systems using CoAP with machine learning,”
in 2023 2nd International Conference for Innovation in Technology (INOCON), 2023, pp. 1–6, doi:
10.1109/INOCON57975.2023.10101176.
[2] D. Majumder and S. M. Kumar, “A review on resource allocation methodologies in fog/edge computing,” in 2022 8th
International Conference on Smart Structures and Systems (ICSSS), 2022, pp. 1–4, doi: 10.1109/ICSSS54381.2022.9782175.
[3] M. L. Umashankar, S. Mallikarjunaswamy, N. Sharmila, D. M. Kumar, and K. R. Nataraj, “A survey on IoT protocol in real-time
applications and its architectures,” in ICDSMLA 2021: Proceedings of the 3rd International Conference on Data Science,
Machine Learning and Applications, 2023, pp. 119–130, doi: 10.1007/978-981-19-5936-3_12.
[4] I. Azimi, A. Anzanpour, A. M. Rahmani, P. Liljeberg, and T. Salakoski, “Medical warning system based on internet of things
using fog computing,” in 2016 International Workshop on Big Data and Information Security (IWBIS), 2016, pp. 19–24, doi:
10.1109/IWBIS.2016.7872884.
[5] J. Honnegowda, K. Mallikarjunaiah, and M. Srikantaswamy, “An efficient abnormal event detection system in video surveillance
using deep learning-based reconfigurable autoencoder,” Ingenierie des Systemes d’Information, vol. 29, no. 2, pp. 677–686, 2024,
doi: 10.18280/isi.290229.
[6] B. Natarajan, S. Bose, N. Maheswaran, G. Logeswari, and T. Anitha, “A survey: an effective utilization of machine learning
algorithms in IoT based intrusion detection system,” in 2023 12th International Conference on Advanced Computing (ICoAC),
2023, pp. 1–7, doi: 10.1109/ICoAC59537.2023.10249672.
[7] V. Venkataramanan, G. Kavitha, M. R. Joel, and J. Lenin, “Forest fire detection and temperature monitoring alert using IoT and
machine learning algorithm,” in 2023 5th International Conference on Smart Systems and Inventive Technology (ICSSIT), 2023,
pp. 1150–1156, doi: 10.1109/ICSSIT55814.2023.10061086.
[8] M. Abedi and M. Pourkiani, “Resource allocation in combined fog-cloud scenarios by using artificial intelligence,” in 2020 Fifth
International Conference on Fog and Mobile Edge Computing (FMEC), 2020, pp. 218–222, doi:
10.1109/FMEC49853.2020.9144693.
[9] S. Mallikarjunaswamy, N. M. Basavaraju, N. Sharmila, H. N. Mahendra, S. Pooja, and B. L. Deepak, “An efficient big data
gathering in wireless sensor network using reconfigurable node distribution algorithm,” in 2022 Fourth International Conference
on Cognitive Computing and Information Processing (CCIP), 2022, pp. 1–6, doi: 10.1109/CCIP57447.2022.10058620.
[10] H. K. Bharadwaj et al., “A review on the role of machine learning in enabling IoT based healthcare applications,” IEEE Access,
vol. 9, pp. 38859–38890, 2021, doi: 10.1109/ACCESS.2021.3059858.
[11] M. Varun, K. Kesavraj, S. Suman, and X. S. Raj, “Integrating IoT and machine learning for enhanced forest fire detection and
temperature monitoring,” in 2023 3rd International Conference on Innovative Mechanisms for Industry Applications (ICIMIA),
2023, pp. 152–158, doi: 10.1109/ICIMIA60377.2023.10426108.
[12] V. Gowrishankar, T. Jayakumar, S. Parameswaran, M. Senthilkumar, S. Lekashri, and B. R. Kumar, “Patient health monitoring
using fog and edge computing,” in 2023 International Conference on Sustainable Communication Networks and Application
(ICSCNA), 2023, pp. 250–256, doi: 10.1109/ICSCNA58489.2023.10370652.
[13] D. Marković, D. Vujičić, Z. Stamenkovič, and S. Randič, “IoT based occupancy detection system with data stream processing and
artificial neural networks,” in 2020 23rd International Symposium on Design and Diagnostics of Electronic Circuits & Systems
(DDECS), 2020, pp. 1–4, doi: 10.1109/DDECS50862.2020.9095715.
[14] N. Khan, S. U. Khan, F. U. M. Ullah, M. Y. Lee, and S. W. Baik, “AI-assisted hybrid approach for energy management in IoT-
based smart microgrid,” IEEE Internet of Things Journal, vol. 10, no. 21, pp. 18861–18875, 2023, doi:
10.1109/JIOT.2023.3293800.
[15] T. M. Saravanan, T. Kavitha, S. Hemalatha, and M. M. Ajmal, “IoT based health observance system using fog computing: a
precise review,” in 2022 International Conference on Advanced Computing Technologies and Applications (ICACTA), 2022, pp.
1–5, doi: 10.1109/ICACTA54488.2022.9753198.
[16] N. C. Fakude, P. Tarwireyi, M. O. Adigun, and A. M. Abu-Mahfouz, “Fog orchestrator as an enabler for security in fog
computing: a review,” in 2019 International Multidisciplinary Information Technology and Engineering Conference (IMITEC),
2019, pp. 1–6, doi: 10.1109/IMITEC45504.2019.9015896.
[17] A. N. Jadagerimath, S. Mallikarjunaswamy, D. M. Kumar, S. Sheela, S. Prakash, and S. S. Tevaramani, “A machine learning
based consumer power management system using smart grid,” in 2023 International Conference on Recent Advances in Science
and Engineering Technology (ICRASET), 2023, pp. 1–5, doi: 10.1109/ICRASET59632.2023.10419979.
[18] S. Jyothi, S. Mallikarjunaswamy, M. Kavitha, N. Kumar, N. Sharmila, and B. M. Kavya, “A machine learning based power load
prediction system for smart grid energy management,” in 2023 International Conference on Recent Advances in Science and
Engineering Technology (ICRASET), 2023, pp. 1–6, doi: 10.1109/ICRASET59632.2023.10420183.
[19] M. Venkatesh, S. N. K. Polisetty, S. CH, P. Kumar. K, R. Satpathy, and P. Neelima, “A novel deep learning mechanism for
workload balancing in fog computing,” in 2022 International Conference on Automation, Computing and Renewable Systems
(ICACRS), 2022, pp. 515–519, doi: 10.1109/ICACRS55517.2022.10029081.
[20] M. K. Hussein and M. H. Mousa, “Efficient task offloading for IoT-Based applications in fog computing using ant colony
optimization,” IEEE Access, vol. 8, pp. 37191–37201, 2020, doi: 10.1109/ACCESS.2020.2975741.
[21] S. Mousavi, S. E. Mood, A. Souri, and M. M. Javidi, “Directed search: a new operator in NSGA-II for task scheduling in IoT
based on cloud-fog computing,” IEEE Transactions on Cloud Computing, vol. 11, no. 2, pp. 2144–2157, 2023, doi:
10.1109/TCC.2022.3188926.
[22] A. Satouf, A. Hamidoglu, O. M. Gul, and A. Kuusik, “Grey wolf optimizer-based task scheduling for IoT-based applications in
the edge computing,” in 2023 Eighth International Conference on Fog and Mobile Edge Computing (FMEC), 2023, pp. 52–57,
doi: 10.1109/FMEC59375.2023.10306148.
[23] M. Charitha, S. Hosur, and M. Srikantaswamy, “Optimized BER reduction in wireless communication using a chaos-based CDSK
modulation model,” in Mathematical Modelling of Engineering Problems, 2025, vol. 12, no. 2, pp. 719–729, doi:
10.18280/mmep.120234.
[24] M. Poornima, T. N. Anitha, S. Mallikarjunaswamy, and M. L. Umashankar, “An efficient internet of things based intrusion
detection and optimization algorithm for smart networks,” International Journal of Computing and Digital Systems, vol. 17, no. 1,
pp. 1–12, 2025, doi: 10.12785/ijcds/1571001227.
[25] T. Suman, S. Kaliappan, L. Natrayan, and D. C. Dobhal, “IoT based social device network with cloud computing architecture,” in
2023 Second International Conference on Electronics and Renewable Systems (ICEARS), 2023, pp. 502–505, doi:
10.1109/ICEARS56392.2023.10085574.
BIOGRAPHIES OF AUTHORS
Optimizing real-time data preprocessing in IoT-based fog computing using … (Nandini Gowda Puttaswamy)