Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Development and Field Test of Integrated Electronics Piezoelectric Accelerometer Based on Lead-Free Piezoelectric Ceramic for Centrifugal Pump Monitoring
Previous Article in Journal
Deep Recyclable Trash Sorting Using Integrated Parallel Attention
Previous Article in Special Issue
Design and Validity of a Smart Healthcare and Control System for Electric Bikes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Prediction of Resident ADL Using Edge-Based Time-Series Ambient Sound Recognition

by
Cheolhwan Lee
,
Ah Hyun Yuh
and
Soon Ju Kang
*
School of Electronic and Electrical Engineering, College of IT Engineering, Kyungpook National University, 80 Daehakro, Bukgu, Daegu 41566, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6435; https://doi.org/10.3390/s24196435
Submission received: 26 August 2024 / Revised: 25 September 2024 / Accepted: 2 October 2024 / Published: 4 October 2024
(This article belongs to the Special Issue Internet of Medical Things and Smart Healthcare)

Abstract

:
To create an effective Ambient Assisted Living (AAL) system that supports the daily activities of patients or the elderly, it is crucial to accurately detect and differentiate user actions to determine the necessary assistance. Traditional intrusive methods, such as wearable or object-attached devices, can interfere with the natural behavior of patients and may lead to resistance. Furthermore, non-intrusive systems that rely on video or sound data processed by servers or the cloud can generate excessive data traffic and raise concerns about the security of personal information. In this study, we developed an edge-based real-time system for detecting Activities of Daily Living (ADL) using ambient noise. Additionally, we introduced an online post-processing method to enhance classification performance and extract activity events from noisy sound in resource-constrained environments. The system, tested with data collected in a living space, achieved high accuracy in classifying ADL-related behaviors in continuous events and successfully generated user activity logs from time-series sound data, enabling further analyses such as ADL assessments. Future work will focus on enhancing detection accuracy and expanding the range of detectable behaviors by integrating the activity logs generated in this study with additional data sources beyond sound.

1. Introduction

The increasing average life expectancy and the diversification of family structures have led to a significant rise in the number of elderly individuals living alone. Consequently, elderly care, particularly for individuals with conditions like dementia who face challenges in daily living, has become a critical concern. Traditionally, these challenges have been managed through facilities and human caregivers. However, as the population in need of assistance rapidly grows, this approach may not be sustainable in the long term.
To address these issues, research into Ambient Assisted Living (AAL) systems is advancing. AAL systems use intelligent objects to assist individuals with daily activities [1,2,3,4]. The success of AAL relies on technology capable of detecting and classifying user actions by observing behaviors through smart objects integrated into the living environment.
Activities of Daily Living (ADL) are a standard metric used to evaluate an individual’s ability to perform routine tasks, such as eating, maintaining hygiene, and moving around, and to identify areas where assistance may be needed. ADL assessments are also used as diagnostic tools for tracking cognitive disorders like dementia [5,6]. Traditionally, ADL measurements have been gathered through surveys completed by patients or caregivers. However, this method is often subjective and prone to inaccuracies due to memory reliance. Therefore, efforts are underway to develop technological solutions for more objective ADL detection [7,8,9].
IoT systems designed to recognize ADL frequently use wearable devices or sensors attached to objects to gather data through direct interaction with users [4,10,11]. While this intrusive approach can provide accurate behavioral data, ensuring consistent device usage among elderly individuals is challenging. Moreover, these devices may interfere with natural behavior, leading to discomfort and resistance. The specialized nature of such systems also results in configuration challenges and scalability limitations.
In contrast, non-intrusive systems estimate behaviors indirectly from user actions, keeping devices hidden and allowing users to maintain their routines with minimal disruption [7,8,12]. These systems are more attractive for long-term monitoring as they avoid the discomfort and inconsistency associated with wearable devices. However, analyzing time-series data such as sound with deep learning models often results in fragmented and error-prone inferences, requiring post-processing techniques to correct these issues and produce reliable activity logs.
Typically, deep learning models send data to servers or the cloud for processing, which raises concerns about privacy and generates significant network traffic, particularly when sensitive information like health data or conversations is transmitted [13]. To mitigate these concerns, there is a growing trend towards edge-based systems, where data are processed locally, reducing privacy risks and minimizing network traffic while maintaining real-time performance.
Handling time-series data in edge computing environments presents unique challenges. Edge devices are typically resource-constrained, with limited processing power and memory compared to centralized servers. Time-series data, such as ambient noise, require continuous real-time processing, which can overwhelm edge devices if not efficiently optimized. Furthermore, the nature of time-series data makes them prone to fragmentation, especially in environments with overlapping sounds and background noise. This can result in inaccurate or incomplete inferences, complicating the classification of ADL-related events. Therefore, robust techniques are needed to manage this fragmentation and ensure efficient data processing within the limited computational resources of edge devices.
Additionally, edge-based processing requires real-time inference, meaning that latency must be minimized without sacrificing accuracy. The need for continuous data processing in noisy environments creates a significant trade-off between low latency and reliable classification. Achieving this balance involves designing models and algorithms that operate efficiently within the constraints of edge hardware while maintaining high accuracy.
In this study, we propose a real-time ADL analysis system that utilizes ambient noise in an edge computing environment. We also introduce an online post-processing method to organize fragmented inference results and convert them into ADL events. The system comprises edge nodes installed in individual spaces, along with a hub device that manages the nodes and collects data. Each edge node performs data collection, inference, and post-processing locally, transmitting only event information to the hub device, enabling high-level analysis of behavioral data from the entire house.
The main contributions of this study are as follows:
  • Development of an easy-to-install, edge-based, non-intrusive real-time behavior detection system for seamless integration into existing living spaces.
  • Proposal of post-processing techniques to improve inference accuracy and abstract results for high-level analysis in resource-constrained environments.

2. Related Works

2.1. Human Activity Recognition

Research aimed at measuring human behavior through technology is being conducted for various purposes as in Table 1, such as healthcare and elderly care [2,14]. These approaches can generally be divided into two categories: methods that use sensor data from devices carried by individuals, such as wearables or smartphones, and methods that install multiple sensors in an environment to analyze the data collected from that space. More complex methods may involve analyzing video or audio data.
Wearable devices or smartphones typically use accelerometer data to directly track human movement [15,16,28] or physiological data such as PPG (photoplethysmography) and ECG (electrocardiogram) [18,19,29]. Smartphones can also leverage built-in sensors such as WiFi signal strength, temperature, and magnetometers. In the realm of smart home systems, research often combines motion sensors with environmental data or directly interacts with objects equipped with measurement devices [10,21]. Some studies also use environmental data alongside sensor data to observe behavior through changes in factors like air quality resulting from user actions [7,11,30].
Video-based methods have been widely researched for behavior analysis using RGB video data [22,31]. However, privacy concerns limit their suitability for home environments. As a result, alternative methods using ambiguous video formats such as LiDAR or thermal imaging are being explored [12,23]. Studies that utilize sound, particularly in environmental sound classification, are advantageous for non-intrusive systems. Sound can capture information from an entire space regardless of sensor orientation, making it ideal for constructing non-intrusive systems [8,24,32].

2.2. Time-Series Ambient Sound Classification with Edge AI Technologies

Unlike voice data, environmental sound data lack structured features like phonemes or distinct frequency bands [33]. This makes it difficult to apply many of the signal processing techniques typically used for speech. Environmental sound data have numerous applications, from detecting outdoor events to identifying indoor behaviors or anomalies. Datasets such as UrbanSound8k [34], AudioSet [35], and DESED (Domestic Environment Sound Event Detection Dataset) [36] encompass a wide variety of environmental sounds. Research using deep learning models on these datasets is ongoing [25,37,38]. While these datasets are useful for developing models, real-world environments often introduce additional noises that are not accounted for in the datasets. Thus, it is essential to develop methods to handle these extraneous noises effectively.
As AI technology progresses, efforts are being made to implement it on devices like smartphones, Raspberry Pi, and Microcontroller Units (MCUs). New MCUs with higher clock speeds and greater memory are being developed to support deep learning models, which typically require substantial memory and matrix operations [39]. Unlike server-based systems, edge devices have limited resources and are closely connected to their physical environment. Research is focusing on optimizing models to operate efficiently in edge environments with minimal resource consumption [40]. There is also ongoing research into extracting features efficiently from sensor data [25,26].
Additionally, these advancements enable the development of new services. For example, previously complex data processing models are now feasible on edge devices [27,41]. Furthermore, time-series sensor data, often handled by embedded systems, are also being explored for use in edge computing applications [17,20].

3. Methodology

In this study, we aimed to develop an edge-based system capable of real-time detection of human actions by utilizing noise data generated by people’s activities in daily living environments. The issues that needed to be addressed in implementing the proposed system are as follows:
  • Real-Time Processing of Time-Series Sound Signals
    It should be capable of processing and inferring time-series sound signals in real time. To achieve this, we designed an edge device with a hybrid architecture consisting of low-power cores and high-performance cores, and we developed and implemented software architecture to effectively utilize it. Additionally, we implemented an AI model architecture suitable for use on edge devices and applied optimization methods such as quantization to enable real-time operation with low resource consumption.
  • Ability to Identify Relevant Sounds in Overlapping Noise
    The system must be able to detect sounds related to the actions of interest in environments where various noises and meaningful data are overlapping. In this study, we built a dataset for model training using sound data collected directly from the actual environment where the system is to be installed. Furthermore, by applying a post-processing technique using voting, we improved the accuracy for continuous sound data and rectified inference results where complex signals were mixed, enabling us to identify meaningful signals.
  • Generation of Outputs Integrable with Other Data
    The detected action events must be produced in an output format that allows integration with other data. We interpreted the action data identified through post-processing according to the installation environment and generated action event logs.

4. System Overview

We aimed to establish a system that should meet the following criteria:
  • The installation and management should be simple, with minimal device configuration.
  • All data processing should occur within the edge device, without the need for external data transmission.
  • The system should be capable of operating in real time, even under exceptional circumstances.
The proposed system consists of several edge node devices and a hub device, as shown in Figure 1. Each edge node is installed in a designated space and collects data within its coverage area, indicating the location of event occurrences. The hub device connects to all edge nodes, enabling it to monitor time synchronization, battery levels, and the connection status for each node. When a sound exceeding a certain threshold is detected, the edge node samples the sound, extracts features, performs inference and post-processing to generate an event log, and transmits it to the hub. The hub stores these event logs with timestamps and locations, allowing for high-level analysis, such as examining interactions between behaviors across multiple spaces. If needed, the hub can also transmit data externally.
In addition, the system functions as a platform for collecting sound data related to activities, an essential step in building the foundational model. It gathers sound data tied to behavior, along with corresponding location and timestamp information. These data can later be compared with participant records for labeling purposes.

5. Edge-Based System Architecture

The edge nodes are designed to operate wirelessly, eliminating the need for power or network cables. They function on an event-driven basis, activating and operating only when there is a substantial noise level, and remaining in standby mode during quiet periods to conserve power. This design enables effortless installation in optimal locations for behavior detection, irrespective of the physical layout of the existing living space. The system operation process is shown in Figure 2.
The edge nodes are designed to utilize two types of cores: a low-power MCU and a high-performance MCU. While a high-performance MCU is necessary to generate real-time inference results, continuously operating it would consume significant power, making it unsuitable for prolonged use. To address this, the edge node is designed with a parallel structure that leverages a low-power core for data collection, pre-processing, and communication, and a high-performance core dedicated solely to model inference. The low-power core collects data, performs pre-processing, and conducts a preliminary analysis to determine if the data require inference. If inference is needed, the pre-processed data are sent to the high-performance core via an I2C line. The high-performance core awakens upon an I2C interrupt, uses the model to perform inference on the received data, and carries out online post-processing. If the result needs to be transmitted, it signals the low-power core via an interrupt to indicate that the data are ready. The low-power core then reads these data and transmits them to the hub device via BLE. The specifications of each core are shown in Table 2.
The edge nodes remain in a sleep state until an input exceeds the specified threshold. When sound is detected, the microphone module triggers an interrupt, waking the nodes for data collection. During data acquisition, the edge nodes continuously monitor the energy value of the incoming signal. If the average energy of the frame consistently remains below the threshold, they cease reception and return to sleep mode. The energy calculation formula is illustrated in Equation (1).
E f r a m e = n = 0 N 1 | x | N , N = f r a m e _ l e n g t h
The hub device maintains information about each connected device in a class-based structure. Each device class includes details such as the device type, name, and location. As an inference process is generated for each connected device, the system can efficiently handle data from multiple sources simultaneously. The processed results are subsequently forwarded to a logging process for recording, ensuring the chronological order of events at various locations.

6. Sound Sampling and Pre-Processing

A commonly used pre-processing technique for audio data involves generating spectrograms to extract both temporal and spectral characteristics from periodic signals. Spectrograms are typically created by applying a windowing function to waveform data at regular intervals and then performing Fast Fourier Transform (FFT) operations, as described in Equation (2). However, this windowing operation can be computationally intensive, and edge nodes equipped with MCUs may struggle to achieve real-time processing. Therefore, it is essential to implement a structure that performs calculations frame by frame during data reception, with the results being stored in a buffer for subsequent processing.
X k = n = 0 N 1 x n e 2 π i n k n , k = 0 , 1 , , N 1
In our previous research, we verified that the time–frequency resolution of environmental sound data has a relatively small impact on accuracy [42]. Building on this finding, we configured the audio sampling buffer size to match the hop distance between FFT operations, incorporating a 50% window overlap. As shown in Figure 3, we have ensured real-time processing capabilities at the edge node by performing the pre-processing without windowing.
The resulting spectrogram has dimensions of 512 × 32 , which is too large to be directly used as input for deep learning models. Consequently, Mel-spectrograms and MFCCs (Mel-Frequency Cepstral Coefficients) are commonly employed as alternatives to traditional spectrograms. The Mel-spectrogram simplifies the frequency range by applying a Mel-scale filter bank (as shown in Equation (3)) to the spectrogram.
m = 2595 log 10 1 + f 700
In our research, we applied 48 Mel-filters and used a log-transformed 48 × 32 Mel-spectrogram as the input. In general, 40 Mel-filters are commonly used in speech recognition. This is because 40 filters provide sufficient resolution to represent the frequency range of 80 Hz to 8000 Hz, which contains the key features of speech signals. However, unlike speech signals, noise signals can have characteristics across all frequency bands and may carry significant information even in regions that humans cannot hear. To account for this, a greater number of filters than typically used was employed, and based on the reduction factor through pooling in the model, we utilized n × 2 4 filters.
This allowed us to compress the feature data to 6 KB per second. To maintain consistency, we implemented pre-processing operations on the edge node to match the results of the Mel-spectrogram feature extraction function found in the widely used librosa library [43], which is commonly used in sound analysis. While environments like servers typically utilize float-type features directly as input, in edge environments, it is common to employ quantized data to reduce the size and computational load of both the input and model parameters. By quantizing float32 data to int8, the input size is significantly reduced to 1.5 KB per second, with some sacrifice in accuracy. The low-power core then prepares for inference by transmitting these data to the performance core via I2C.

7. Training and Deploying Model

7.1. Characteristics of Behavioral Sound Data

In this study, we utilize a model that infers and identifies sounds based on one-second audio clips. However, sounds occurring in real environments have diverse characteristics, making it challenging to classify all sounds uniformly within a one-second timeframe. For instance, as shown in Figure 4, the sound of flowing water exhibits a consistent and repetitive pattern, whereas sounds like flushing appear briefly over just a few frames. To accurately infer a variety of sound patterns using the same model, it is essential to consider the characteristics of each sound throughout various stages, from preparing the training data to post-processing.
We broadly categorized sounds into two types: continuous sounds and short-duration sounds, applying different methods for data collection and processing for each. Continuous sounds can be easily collected by dividing the continuous data into one-second segments without any special processing, allowing us to gather a large amount of data. These sounds also have the characteristic of consistently producing the same inference results during the continuous inference process. On the other hand, short-duration sounds, where the entire event occurs within one second, require padding and only allow for the collection of a single piece of training data per event. Therefore, strategies to overcome the limited amount of data are needed. These sounds also typically appear briefly, often only once or twice during the inference process.
Table 3 shows the sounds selected for collection, considering the characteristics of various sounds, the corresponding ADL (Activities of Daily Living) categories, and how they are classified by the model. The data collected based on behavioral events and the classification labels used by the model do not always align. This discrepancy reflects challenges in distinguishing various events using the model in real-world environments or cases where multiple types of sounds occur within a single event. For example, the “water flowing” event is categorized into “watering1” and “watering2” to distinguish between the fast, strong sound of water during a shower and the gentle sound of water during handwashing. Additionally, events like chopping, dropping, and dish clanging have very similar sound characteristics, so they were grouped and trained as a single “hitting” event.

7.2. Building Dataset with Real-World Sound Data

Various datasets, such as AudioSet [35], have extensively covered environmental sound data and have been actively utilized in research. However, when a model trained on these datasets is deployed in real-world environments, its performance often decreases due to the effects of domain shift [44]. To mitigate the impact of domain shift, models are typically designed to be more generalizable by using more features as input and increasing model depth. However, models operating in edge environments must produce inference results within a limited time using constrained resources, and they can only utilize quantized data that may have lost fine details. Consequently, the effects of domain shift are particularly pronounced in edge environments. To overcome this, it is necessary to train the base model with data collected using microphones with the same specifications as those installed on the edge nodes, and to adapt the model by collecting data from the actual deployment site before distribution.
In this study, we independently carried out the process from data collection and labeling to model training and deployment to implement a system that operates in real-world environments. We obtained consent from participants to install our system in their actual homes, allowing us to collect sound data and label them for use as a training dataset. The dataset is broadly divided into two categories: data for training the base model and data for specializing the model for the installation site. The data for base model training include some portions of publicly available datasets and were collected from various locations such as laboratories and homes using different types of recording equipment. The dataset for model specialization consists of 1–2 days of data collected using edge nodes in the same environment where the system is to be installed, and it is divided into training and testing data. The model training uses a combined training dataset of base and domain-specific data depending on the location where the edge nodes will be installed.
Domain-specific data collection on edge nodes is conducted in the following manner. When the edge node initially establishes a BLE connection with the hub device, it synchronizes with the hub’s time information and then counts time on its own. The edge node is equipped with an SD card for collecting data for analysis purposes. During data collection mode, the pre-processing and inference processes are not active. Collected data are stored on the SD card in WAV format, with the file name reflecting the time the file was created.
Data recording is triggered similarly to inference mode: recording starts when a sound exceeding a certain threshold is detected and stops when low energy levels are maintained for more than 5 s, using an event-driven approach. Participants are asked to record their actions and corresponding times separately to aid in the labeling process afterward.
Table 4 shows the number of data points in the base dataset and the domain dataset collected from the test environment. The main difference between the base dataset and the domain dataset lies in their diversity. The base dataset includes publicly available datasets, data collected from various types of recording equipment, and data gathered from multiple locations. On the other hand, the domain dataset consists only of sounds recorded by the installed edge devices at a single location and does not include all labeled sounds. For example, it would be hard to collect Microwave sounds in a bathroom environment.
The role of the base dataset is to train the overall feature extractor of the model, while the domain dataset’s role is to teach the model the characteristics of sounds that occur in a specific environment. The actual training is conducted by combining the base dataset and domain dataset in a 1500:250 ratio. If this process is not followed, it is not possible to properly classify real data, as in Figure 5. The edge device we propose is designed to be permanently installed in a specific location to provide services, where it is more important to accurately classify specific data in a particular domain than to achieve generic performance across all environments. For further work, we plan to conduct research on systematizing the process of tuning the base model into a domain-specific model.

7.3. Handling Imbalanced Datasets

As illustrated in Table 4, the original dataset exhibits significant variations in the number of data samples per label. Training models on such imbalanced data can lead to overfitting on specific labels [45]. To mitigate this issue, we resampled the training dataset to contain an equal number of 1500 samples for each label. For labels with fewer than 1500 clips, we applied oversampling techniques, repetitively extracting samples to reach the desired quantity [46]. Similarly, for the domain dataset, resampling was carried based on 250 samples and added to the base dataset. Subsequently, we augmented the resampled dataset by introducing data with some time and frequency information randomly removed [47]. This augmentation process generated two augmented data samples for each clip. As a result, we obtained a total of over 4500 training data samples for each label, ensuring robust model training.

7.4. Training and Deploying Deep Learning Model for Edge Nodes

To determine the foundational model for the system, we tested several model architectures to find the optimal balance between accuracy and inference performance. The models we considered for testing included a 1D CNN, 2D CNN, and TCN (Temporal Convolutional Network), LSTM (Long Short-Term Memory), and ResNet base model for handling time-series data. Based on the test results, we selected a ResNet-based model as the foundational model and structured it as shown in Figure 6.
Edge environments often lack the resources necessary for model training. Therefore, model creation is typically carried out in high-performance environments like servers, after which the trained model is deployed at the edge. To optimize the model for the edge environment, optimization processes are integrated into the training phase.
One commonly used optimization technique is quantization, which involves quantizing model parameters to integers. Quantization can reduce the model’s accuracy to some extent, but it significantly improves runtime performance by reducing the amount of floating-point operations. Quantization of model parameters is an essential process for running the model on a microcontroller using TensorFlow Lite for Microcontrollers (TFLM). Additionally, the edge node architecture designed in this study includes a process of exchanging features and results between the low-power core and the performance core. Through quantization, communication data can be reduced by up to one-fourth, enabling smoother communication and real-time operation.
Quantized models are stored in the flatbuffer format and can be loaded and used by the TFLite interpreter running on the performance core. These files contain all the necessary information for inference, including the model’s input/output formats, model weights, and quantization parameters.

8. Post-Processing

In this study, the model we employed infers 1-second audio clips and provides results. When using such a model for continuous analysis of ongoing sounds, it generates a significant amount of event information in a short timeframe. In scenarios where multiple events occur simultaneously or when there is ambient noise, it may exhibit a pattern where various events change rapidly. To measure a user’s ADL, obtaining reliable information about the action, including the type of action, time of occurrence, and duration, is crucial. However, when multiple labels change randomly, it becomes challenging to obtain accurate action and timing information. To address this issue, we applied a voting-based post-processing method, as illustrated in Figure 7, to refine the action data. Additionally, based on this, we transformed the occurring actions into event logs to facilitate higher-level analysis.
Voting, a method commonly used in ensemble models, combines results from multiple models to produce a single outcome when making inferences on the same data. We use it to ensure consistent results from a single model by assuming that consecutive input data frames share the same label. This assumption is grounded in our analysis of recorded data, which revealed that instances of actions changing multiple times within a 1-second window were extremely rare, except for brief events like drops or door slams. Furthermore, a short event’s occurrence did not necessarily imply the interruption of a continuous event. For example, during dishwashing, both the sound of water flowing and the sound of dishes clanging can occur simultaneously without one interrupting the other.
Traditional voting methods can be categorized as hard voting and soft voting. In hard voting, the final decision is based on the most frequent result from multiple inferences. However, this approach can potentially lose information from events with slightly lower confidence, especially when multiple events occur simultaneously. On the other hand, soft voting computes the average of all inference results, which ensures that events with lower confidence still contribute to the final decision. However, soft voting may not effectively handle short-duration events due to its reliance on the overall average.
To overcome the drawbacks of both methods, we independently assess the occurrence of each event without linking it to other labels. This approach allows us to provide confidence values for additional analysis. Moreover, we classify events into continuous and momentary categories and process them differently to prevent the loss of information from short-duration events.
In this section, we refer to the results inferred from a single post-processing buffer as an “event”, and we represent the abstracted behavioral data that we aim to obtain through post-processing as an “activity”.
We segmented the continuous audio signal into 1-second frames with a 50% overlap, performing inference every 0.5 s. These frames were then grouped into sets of seven frames (equivalent to 3.5 s) and subjected to windowing. Within each window, we counted the number of inferences with probabilities exceeding a specific threshold for each label. For continuous events, if the count exceeded half of the window size, we deemed the event to have occurred. The confidence score was computed by averaging the probabilities within the subsequent processing window. In the case of momentary events, we assumed they had occurred if the inference result exceeded the threshold value. The confidence score was determined based on the number of detections within the frame, resulting in an average probability.
This process aims to improve the reliability of continuous events and preserve the occurrence of momentary events. Based on this, we transform the inference results into an activity log. Activity information includes start and end times, duration, occurrence location, and confidence. The occurrence location refers to the spatial unit where the data are transmitted from the installed edge node, while confidence represents the average confidence of the continuous events that compose the activity. Each event has a t o l e r a n c e value for each label, which determines the time interval allowed to consider consecutive identical events as a single activity. For continuous activity, we set t o l e r a n c e to 5 s or more to prevent sustained activity from being interrupted by short activities. For momentary activity, we set it to be shorter, typically 3 to 5 counts, to report it as they occur.
When a sequence of events exceeds the t o l e r a n c e value, indicating the end of an activity, we record the end time and calculate the duration and confidence of the activity. Subsequently, if necessary, events with excessively short duration or low confidence can be further filtered. The detailed post-processing procedure is outlined in Algorithm 1.
Algorithm 1 Post-processing
 1:
repeat
 2:
    store prediction results in post-processing buffer
 3:
until buffer is full
 4:
▹ Process for smoothing events and enhancing accuracy for continuous activity
 5:
c o u n t count the prediction results for each label > t h r e s h o l d  ▹ filter out the events with low probability
 6:
if label is continuous event then
 7:
    if  c o u n t > b u f f e r _ s i z e / 2  then
 8:
        set e v e n t occurred
 9:
         c o n f i d e n c e mean of the predictions for the entire buffer
10:
    end if
11:
else if label is instant event then
12:
    set e v e n t occurred
13:
     c o n f i d e n c e  mean of the predictions > t h r e s h o l d
14:
end if
15:
▹ Process for generating abstracted activity logs
16:
if new  e v e n t  is occurred then
17:
    store  s t a r t _ t i m e
18:
    reset  t o l e r e n c e
19:
end if
20:
e v e n t _ c o u n t + +
21:
s u m _ c o n f i d e n c e s u m _ c o n f i d e n c e + c o n f i d e n c e
22:
if the continuous event that had occurred previously did not take place this time then
23:
     t o l e r a n c e
24:
    if  t o l e r a n c e < 0  then
25:
        store  e n d _ t i m e
26:
         d u r a t i o n e n d _ t i m e s t a r t _ t i m e
27:
         a c t i v i t y _ c o n f i d e n c e s u m _ c o n f i d e n c e / e v e n t _ c o u n t
28:
        generate the activity log
29:
    end if
30:
end if
31:
dequeue buffer

9. Evaluations

9.1. Effects of Applying Resampling, Augmentation, and Quantization

In this study, resampling and augmentation techniques were applied to enhance the model’s classification performance and prevent overfitting. Quantization was also employed as an optimization method to ensure efficient operation in an edge environment. These techniques have a direct impact on the model’s classification performance, and their effects were analyzed prior to model development. For evaluation, we used the ResNet18 model, which serves as the foundation for our approach, as the baseline. This section presents the classification results based on the base dataset, excluding the domain-specific dataset.
Table 5 shows the impact of each process on classification performance. When no techniques are applied, the model shows relatively poor learning outcomes. Applying resampling and augmentation individually results in significant accuracy improvements, with the highest accuracy achieved when all techniques are applied together. Quantization, on the other hand, leads to a slight decrease in classification performance.

9.2. Model Classification and Runtime Performance

To address training data variations due to varying action event durations, we used both sampling and augmentation techniques. Table 6 displays results, comparing scenarios with only sampling, only augmentation, and both techniques. We evaluated various models and performed performance assessment with int8 quantization using TensorFlow Lite’s tool. LSTM performance was assessed before quantization, since it lacks quantization support.
The experimental results showed that the LeNet5 model, with its relatively high proportion of Dense layers, achieved the highest accuracy. The ResNet series models also demonstrated strong performance, comparable to other models, while utilizing significantly fewer parameters and consuming less memory. However, despite the lower number of parameters, the actual execution speed did not improve as much as anticipated. This may be attributed to the higher proportion of CNN operations and the deeper layers compared to other models.
The significant difference in memory usage is due to the Dense layers in the model, which have a large number of parameters relative to the execution time. We adjusted the model, while maintaining its structure for comparative analysis, to achieve affordable accuracy while reducing the execution speed. Memory usage can be seen as a characteristic of the model’s structure and does not always correlate with execution time.
The TCN model exhibited some weaknesses in processing Mel-spectrograms. On the other hand, the LSTM model achieved the highest classification performance but could not be deployed on edge nodes, as TensorFlow Lite (TfLite) does not support quantization for LSTM models. Although our model did not achieve the best classification performance, it was confirmed to be suitable for deployment on edge nodes due to its minimal memory usage and its ability to meet the 500 ms time limit required for real-time inference, even with a 50% overlap. The model training graph and classification results are shown in Figure 8. The phenomenon of the validation accuracy initially surpassing the training accuracy is likely due to the application of augmentation only to the training dataset.

9.3. Post-Processing

9.3.1. Improving Continuous Event Classification via Voting

The post-processing carried out in this study serves the purpose of enhancing the accuracy and duration estimation of events, even in cases where brief misclassifications occur during long events or when new events emerge. To validate this, we sequentially inferred files longer than 10 s from the test dataset to assess their accuracy. Given that the test dataset comprises data consistently labeled under the same category, effective post-processing should rectify errors in the inference results, resulting in improved accuracy.
We conducted experiments on 5364 clips obtained through pre-processing with a 75% overlap, resulting in a total of 137 files longer than 10 s. As the post-processing group processed seven clips, the number of results after post-processing is 4470.
Table 7 shows the performance improvement for classifying continuous events after voting. Due to imbalances in the test data caused by variations in file length for each event, accuracy measurements utilize the overall average of label-specific classification accuracy. The overall low F1 score is attributed to the fact that only data longer than a certain length were selected, resulting in a lack of data for short events.
The experimental results showcased enhancements in both accuracy and F1 score, all while maintaining the same loss value. These findings suggest that misclassified labels in the inference results of continuous data were effectively rectified. Consequently, it has been verified that post-processing can improve the reliability of continuous events.

9.3.2. Extracting Activity Logs from Noisy Data

Figure 9 and Table 8 depict the initial inference results alongside the results after implementing post-processing on noisy data obtained from a real-world environment. In the pre-processing phase, we presented outcomes for most instances where the count surpassed the threshold within the window, omitting event types and confidence evaluations.
As depicted in Figure 9, the initial inference results show a multitude of unrelated scenarios. However, after applying the post-processing step, the results are predominantly organized based on labels associated with specific actions. Although some misclassified labels may still emerge in noisy environments, they can be further filtered during the log generation process.
Activity log generation involves considering various factors, including event type, confidence level, installation location, and duration, based on the post-processing results. Initially, activities with low confidence can be excluded using the final confidence value. For continuous activities, a tolerance period is set to 5 s for situations where the sound briefly stops or short events occur without interrupting the ongoing event. Additionally, in cases where the installation location is identified as a bathroom, labels such as “eating” or “TV” cannot occur and will be automatically excluded from consideration.
The resulting log data are presented in Table 8. The “Airutils” label represents the sound of a ventilation fan operating in the bathroom, which is detected when no other sounds are present. The “Hitting” label is classified as a short event and is processed separately; thus, it can be observed that it appears briefly between other continuous events. This indicates that even during ongoing continuous events, short sounds, such as the clinking of a cup while brushing teeth or the closing of a door, can be effectively detected. It is evident that the generated logs closely resemble real events. Sounds such as “gargling”, which are not included in existing classification categories, were classified as “speech” due to their similarity to human voices. This highlights the importance of incorporating an outlier detection mechanism to eliminate labels that do not pertain to the intended classification targets in real-world settings.

9.4. Predictability and Stability of the System

Predictability of the Pre-Processing Process in the Edge Node

The edge node returns a PDM buffer containing 512 samples. Each time this buffer is received, it is combined with the previous buffer for pre-processing before being transmitted to the hub device. Therefore, to ensure real-time performance, pre-processing and transmission must be completed before the next PDM buffer is returned. Since the data collected by the edge node consist of 16,000 samples per second, the processing must be completed within 32 milliseconds to ensure reliable real-time execution. To verify this, we measured the time it takes from when the input PDM buffer is received, which indicates the collection of one frame of data, to the completion of pre-processing and data transmission. Table 9 presents the average processing speed and standard deviation from 100 executions.
Figure 10 shows the actual execution time for 100 iterations of pre-processing and transmission on the edge nodes. The experimental results in Table 9 show that the pre-processing and transmission process takes an average of 14.21 ms with a deviation of 13 ns, confirming stable real-time operation. Here, the average time spent on data transmission was 0.4 ms, which had minimal impact on the overall execution time. This is attributed to the architecture of the nRF5340 MCU, where the Application core only stores the generated packets from pre-processing results in shared memory, while the actual transmission is performed in parallel by the Network core.

10. Conclusions

In this study, we have developed a real-time system for analyzing residents’ activities based on noise sound information generated in residential areas. To achieve this, we implemented the proposed system in real-life living environments, collected a dataset for training purposes, and built a baseline model that achieved an accuracy over 90% for inference. Additionally, we introduced a post-processing method to enhance event classification, resulting in a classification accuracy of up to 96% for sustained events. Based on this, we have proposed a method for generating log data for high-level analysis.
In our research, we tailored the system to adapt to the installation environment based on field-collected data to enhance model performance. In future research, to ensure system versatility, we aim to implement a system that can adapt to varying environments using techniques like transfer learning. Moreover, since it is impractical to classify all sounds in a given space, we plan to incorporate outlier detection technology to filter out sounds beyond the scope of classification.
For high-level analysis, we successfully abstracted inference results to generate log data. In future research, we intend to integrate these data with activity information obtained from sources other than sound to accurately detect activities and identify abnormal behaviors. Ultimately, our goal is to implement an AAL system that leverages detected behaviors and information on anomalous behavior to provide tailored recommendations for improving the ability to perform ADL.

Author Contributions

Conceptualization, C.L. and S.J.K.; Methodology, C.L. and A.H.Y.; Software, C.L.; Validation, C.L.; Formal analysis, C.L.; Investigation, C.L.; Resources, C.L.; Data curation, C.L. and A.H.Y.; Writing—original draft, C.L.; Writing—review & editing, C.L.; Visualization, C.L.; Supervision, S.J.K.; Project administration, S.J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1A6A1A03025109).

Institutional Review Board Statement

This study was approved by the Institutional Review Board of Kyungpook National University (approval No. 2018-0149). All procedures in this study were explained to the participants and written informed consent was obtained prior to participation.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this article are not readily available because the data provider did not grant permission for sharing. Requests to access the datasets should be directed to authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Costa, R.; Carneiro, D.; Novais, P.; Lima, L.; Machado, J.; Marques, A.; Neves, J. Ambient Assisted Living. In Proceedings of the 3rd Symposium of Ubiquitous Computing and Ambient Intelligence 2008, Salamanca, Spain, 22–24 October 2008; Corchado, J.M., Tapia, D.I., Bravo, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 86–94. [Google Scholar]
  2. Maskeliūnas, R.; Damaševičius, R.; Segal, S. A Review of Internet of Things Technologies for Ambient Assisted Living Environments. Future Internet 2019, 11, 259. [Google Scholar] [CrossRef]
  3. Marques, G.; Bhoi, A.K.; Hareesha, K. IoT in Healthcare and Ambient Assisted Living; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  4. Stavropoulos, T.G.; Papastergiou, A.; Mpaltadoros, L.; Nikolopoulos, S.; Kompatsiaris, I. IoT Wearable Sensors and Devices in Elderly Care: A Literature Review. Sensors 2020, 20, 2826. [Google Scholar] [CrossRef] [PubMed]
  5. Katz, S. Assessing self-maintenance: Activities of daily living, mobility, and instrumental activities of daily living. J. Am. Geriatr. Soc. 1983, 31, 721–727. [Google Scholar] [CrossRef] [PubMed]
  6. Bucks, R.S.; Ashworth, D.L.; Wilcock, G.K.; Siegfried, K. Assessment of Activities of Daily Living in Dementia: Development of the Bristol Activities of Daily Living Scale. Age Ageing 1996, 25, 113–120. [Google Scholar] [CrossRef]
  7. Gambi, E.; Temperini, G.; Galassi, R.; Senigagliesi, L.; De Santis, A. ADL Recognition Through Machine Learning Algorithms on IoT Air Quality Sensor Dataset. IEEE Sens. J. 2020, 20, 13562–13570. [Google Scholar] [CrossRef]
  8. Liang, D.; Thomaz, E. Audio-Based Activities of Daily Living (ADL) Recognition with Large-Scale Acoustic Embeddings from Online Videos. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–18. [Google Scholar] [CrossRef]
  9. Pires, I.M.; Marques, G.; Garcia, N.M.; Pombo, N.; Flórez-Revuelta, F.; Spinsante, S.; Teixeira, M.C.; Zdravevski, E. Recognition of Activities of Daily Living and Environments Using Acoustic Sensors Embedded on Mobile Devices. Electronics 2019, 8, 1499. [Google Scholar] [CrossRef]
  10. Park, Y.J.; Jung, S.Y.; Son, T.Y.; Kang, S.J. Self-Organizing IoT Device-Based Smart Diagnosing Assistance System for Activities of Daily Living. Sensors 2021, 21, 785. [Google Scholar] [CrossRef]
  11. Hoang, M.L.; Carratù, M.; Paciello, V.; Pietrosanto, A. Body Temperature—Indoor Condition Monitor and Activity Recognition by MEMS Accelerometer Based on IoT-Alert System for People in Quarantine Due to COVID-19. Sensors 2021, 21, 2313. [Google Scholar] [CrossRef]
  12. Naik, K.; Pandit, T.; Naik, N.; Shah, P. Activity Recognition in Residential Spaces with Internet of Things Devices and Thermal Imaging. Sensors 2021, 21, 988. [Google Scholar] [CrossRef]
  13. Khan, M.A.; Din, I.U.; Majali, T.; Kim, B.S. A Survey of Authentication in Internet of Things-Enabled Healthcare Systems. Sensors 2022, 22, 9089. [Google Scholar] [CrossRef] [PubMed]
  14. Onthoni, D.D.; Sahoo, P.K. Artificial-Intelligence-Assisted Activities of Daily Living Recognition for Elderly in Smart Home. Electronics 2022, 11, 4129. [Google Scholar] [CrossRef]
  15. Khatun, M.A.; Yousuf, M.A.; Ahmed, S.; Uddin, M.Z.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Khan, A.; Azad, A.; Moni, M.A. Deep CNN-LSTM with Self-Attention Model for Human Activity Recognition Using Wearable Sensor. IEEE J. Transl. Eng. Health Med. 2022, 10, 2700316. [Google Scholar] [CrossRef]
  16. Mekruksavanich, S.; Jitpattanakul, A. LSTM Networks Using Smartphone Data for Sensor-Based Human Activity Recognition in Smart Homes. Sensors 2021, 21, 1636. [Google Scholar] [CrossRef]
  17. Zebin, T.; Scully, P.J.; Peek, N.; Casson, A.J.; Ozanyan, K.B. Design and Implementation of a Convolutional Neural Network on an Edge Computing Smartphone for Human Activity Recognition. IEEE Access 2019, 7, 133509–133520. [Google Scholar] [CrossRef]
  18. Afzali Arani, M.S.; Costa, D.E.; Shihab, E. Human Activity Recognition: A Comparative Study to Assess the Contribution Level of Accelerometer, ECG, and PPG Signals. Sensors 2021, 21, 6997. [Google Scholar] [CrossRef]
  19. Bondugula, R.K.; Udgata, S.K.; Sivangi, K.B. A novel deep learning architecture and MINIROCKET feature extraction method for human activity recognition using ECG, PPG and inertial sensor dataset. Appl. Intell. 2023, 53, 14400–14425. [Google Scholar] [CrossRef]
  20. Agarwal, P.; Alam, M. A Lightweight Deep Learning Model for Human Activity Recognition on Edge Devices. Procedia Comput. Sci. 2020, 167, 2364–2373. [Google Scholar] [CrossRef]
  21. Zhang, H.; Xiao, Z.; Wang, J.; Li, F.; Szczerbicki, E. A Novel IoT-Perceptive Human Activity Recognition (HAR) Approach Using Multihead Convolutional Attention. IEEE Internet Things J. 2020, 7, 1072–1080. [Google Scholar] [CrossRef]
  22. Franco, A.; Magnani, A.; Maio, D. A multimodal approach for human activity recognition based on skeleton and RGB data. Pattern Recognit. Lett. 2020, 131, 293–299. [Google Scholar] [CrossRef]
  23. Bouchard, K.; Maitre, J.; Bertuglia, C.; Gaboury, S. Activity Recognition in Smart Homes using UWB Radars. Procedia Comput. Sci. 2020, 170, 10–17. [Google Scholar] [CrossRef]
  24. Jobanputra, C.; Bavishi, J.; Doshi, N. Human Activity Recognition: A Survey. Procedia Comput. Sci. 2019, 155, 698–703. [Google Scholar] [CrossRef]
  25. Elliott, D.; Otero, C.E.; Wyatt, S.; Martino, E. Tiny Transformers for Environmental Sound Classification at the Edge. arXiv 2021, arXiv:2103.12157. [Google Scholar] [CrossRef]
  26. Mohaimenuzzaman, M.; Bergmeir, C.; West, I.; Meyer, B. Environmental Sound Classification on the Edge: A Pipeline for Deep Acoustic Networks on Extremely Resource-Constrained Devices. Pattern Recognit. 2023, 133, 109025. [Google Scholar] [CrossRef]
  27. Elliott, D.; Martino, E.; Otero, C.E.; Smith, A.; Peter, A.M.; Luchterhand, B.; Lam, E.; Leung, S. Cyber-Physical Analytics: Environmental Sound Classification at the Edge. In Proceedings of the 2020 IEEE 6th World Forum on Internet of Things (WF-IoT), New Orleans, LA, USA, 2–16 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  28. Micucci, D.; Mobilio, M.; Napoletano, P. UniMiB SHAR: A Dataset for Human Activity Recognition Using Acceleration Data from Smartphones. Appl. Sci. 2017, 7, 1101. [Google Scholar] [CrossRef]
  29. Qiu, S.; Zhao, H.; Jiang, N.; Wang, Z.; Liu, L.; An, Y.; Zhao, H.; Miao, X.; Liu, R.; Fortino, G. Multi-sensor information fusion based on machine learning for real applications in human activity recognition: State-of-the-art and research challenges. Inf. Fusion 2022, 80, 241–265. [Google Scholar] [CrossRef]
  30. Fonollosa, J.; Rodriguez-Lujan, I.; Shevade, A.V.; Homer, M.L.; Ryan, M.A.; Huerta, R. Human activity monitoring using gas sensor arrays. Sens. Actuators B Chem. 2014, 199, 398–402. [Google Scholar] [CrossRef]
  31. Beddiar, D.R.; Nini, B.; Sabokrou, M.; Hadid, A. Vision-based human activity recognition: A survey. Multimed. Tools Appl. 2020, 79, 30509–30555. [Google Scholar] [CrossRef]
  32. Jung, M.; Chi, S. Human activity classification based on sound recognition and residual convolutional neural network. Autom. Constr. 2020, 114, 103177. [Google Scholar] [CrossRef]
  33. Alías, F.; Socoró, J.C.; Sevillano, X. A Review of Physical and Perceptual Feature Extraction Techniques for Speech, Music and Environmental Sounds. Appl. Sci. 2016, 6, 143. [Google Scholar] [CrossRef]
  34. Salamon, J.; Jacoby, C.; Bello, J.P. A Dataset and Taxonomy for Urban Sound Research. In Proceedings of the 22nd ACM International Conference on Multimedia, MM ’14, New York, NY, USA, 3–7 November 2014; pp. 1041–1044. [Google Scholar] [CrossRef]
  35. Gemmeke, J.F.; Ellis, D.P.W.; Freedman, D.; Jansen, A.; Lawrence, W.; Moore, R.C.; Plakal, M.; Ritter, M. Audio Set: An ontology and human-labeled dataset for audio events. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 776–780. [Google Scholar] [CrossRef]
  36. Serizel, R.; Turpault, N.; Shah, A.; Salamon, J. Sound Event Detection in Synthetic Domestic Environments. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 86–90. [Google Scholar] [CrossRef]
  37. Demir, F.; Turkoglu, M.; Aslan, M.; Sengur, A. A new pyramidal concatenated CNN approach for environmental sound classification. Appl. Acoust. 2020, 170, 107520. [Google Scholar] [CrossRef]
  38. Turpault, N.; Serizel, R.; Wisdom, S.; Erdogan, H.; Hershey, J.R.; Fonseca, E.; Seetharaman, P.; Salamon, J. Sound Event Detection and Separation: A Benchmark on Desed Synthetic Soundscapes. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 840–844. [Google Scholar] [CrossRef]
  39. Ray, P.P. A review on TinyML: State-of-the-art and prospects. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 1595–1623. [Google Scholar] [CrossRef]
  40. Surianarayanan, C.; Lawrence, J.J.; Chelliah, P.R.; Prakash, E.; Hewage, C. A Survey on Optimization Techniques for Edge Artificial Intelligence (AI). Sensors 2023, 23, 1279. [Google Scholar] [CrossRef] [PubMed]
  41. Zhu, J.; Lou, X.; Ye, W. Lightweight Deep Learning Model in Mobile-Edge Computing for Radar-Based Human Activity Recognition. IEEE Internet Things J. 2021, 8, 12350–12359. [Google Scholar] [CrossRef]
  42. Lee, C.; Kang, H.M.; Jeon, Y.; Kang, S.J. Ambient Sound Analysis for Non-Invasive Indoor Activity Detection in Edge Computing Environments. In Proceedings of the 2023 IEEE Symposium on Computers and Communications (ISCC), Gammarth, Tunisia, 9–12 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  43. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 6–12 July 2015; Volume 8, pp. 18–25. [Google Scholar]
  44. Farahani, A.; Voghoei, S.; Rasheed, K.; Arabnia, H.R. A Brief Review of Domain Adaptation. In Advances in Data Science and Information Engineering; Stahlbock, R., Weiss, G.M., Abou-Nasr, M., Yang, C.Y., Arabnia, H.R., Deligiannidis, L., Eds.; Springer: Cham, Switzerland, 2021; pp. 877–894. [Google Scholar]
  45. Liu, A.; Ghosh, J.; Martin, C. Generative Oversampling for Mining Imbalanced Datasets. DMIN 2007, 7, 66–72. [Google Scholar]
  46. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  47. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. Specaugment: A simple data augmentation method for automatic speech recognition. arXiv 2019, arXiv:1904.08779. [Google Scholar]
Figure 1. System overview.
Figure 1. System overview.
Sensors 24 06435 g001
Figure 2. Edge-based system architecture overview.
Figure 2. Edge-based system architecture overview.
Sensors 24 06435 g002
Figure 3. PDM sampling and buffer management for pre-processing.
Figure 3. PDM sampling and buffer management for pre-processing.
Sensors 24 06435 g003
Figure 4. Features of different sound types.
Figure 4. Features of different sound types.
Sensors 24 06435 g004
Figure 5. Classification result of Domain1 test dataset.
Figure 5. Classification result of Domain1 test dataset.
Sensors 24 06435 g005
Figure 6. ResNet-based model structure.
Figure 6. ResNet-based model structure.
Sensors 24 06435 g006
Figure 7. Post-processing overview.
Figure 7. Post-processing overview.
Sensors 24 06435 g007
Figure 8. Learning graph and classification results.
Figure 8. Learning graph and classification results.
Sensors 24 06435 g008
Figure 9. Post-processing noisy data.
Figure 9. Post-processing noisy data.
Sensors 24 06435 g009
Figure 10. Elapsed time of pre-processing in the edge node.
Figure 10. Elapsed time of pre-processing in the edge node.
Sensors 24 06435 g010
Table 1. Summary of related works.
Table 1. Summary of related works.
MethodRef.DataContribution
Smartphone-based HAR [15]Accel, Gyro, Lin Acc (H-Activity, MHEALTH, UCI-HAR)CNN-LSTM with self-attention for 99.93% accuracy on H-Activity.
 [16]Accel (UCI-HAR)4-layer CNN-LSTM, +2.24% accuracy.
 [17]Accel (5-class data)CNN, 96.4% accuracy, 8-bit quantization for speed.
Wearable HAR [18]3D-ACC, ECG, PPG (PPG-DaLiA)Feature fusion, +3.00% F1 score.
 [19]HR, Motion dataMINIROCKET for feature extraction, superior results at lower cost.
 [20]Wearable sensor (6 activities)Lightweight DL model, optimized for edge devices.
IoT-Based Monitoring [21]Env Sensors (Temp, Humidity, Flame)IoT platform for real-time ADL monitoring.
 [11]MEMS Accel, Temp/Hum sensorsReal-time quarantine monitoring with alerts.
Environmental Sensor HAR [7]Air quality sensorsADL detection using air composition.
Video/Multimodal HAR [22]Skeleton, RGB (Kinect)RGB and skeleton fusion for activity recognition.
Thermal and IoT Activity Recog. [12]RGB, Thermal (IoT)3D thermal model for appliance activity recognition.
UWB Radar-based HAR [23]UWB Radar (15 ADLs)80% accuracy with UWB radar for ADL recognition.
Audio-based HAR [24]Audio embeddings (Online videos)Audio-based ADL recognition, 64.2% (Top-1) accuracy.
Edge-based Sound Class. [25]Mel spectrograms (ESC-50, Office)Tiny Transformer, 99.85% fewer params than CNN.
 [26]Raw audio (ESC-10, ESC-50, UrbanSound8K)ACDNet, 97.22% size reduction with strong accuracy.
Cyber–Physical Edge Analytics [27]Env Sound (6 office sounds)Ensemble DNN for event detection with low latency.
Edge-based HAR [20]Wearable sensor (6 activities)Lightweight model, outperforming existing DL techniques.
Table 2. Hardware specification.
Table 2. Hardware specification.
Low-Power MCUPerformance MCU
ProcessorArm Cortex-M33Arm Cortex-M7
Clock64 MHz550 MHz
Memory512 KB1 MB
Power source3.7 V Li-ion Battery
Table 3. List of sound events recorded for ADL analysis.
Table 3. List of sound events recorded for ADL analysis.
LocationADLEventsLabels
KitchenFeeding/CookingDish ClangingHitting
Boiling/FryingCooking
MicrowaveMicrowave
ChoppingHitting
Smoke ExtractorAirutils
BathroomBathingWater FlowingWatering1/Watering2
ToothbrushingBrushing
ToiletingFlushingFlushing
PeeingPeeing
EverywhereCleaningVacuumingAirutils
CommunicationSpeechSpeech
TVTV
TransferringDoorHitting
Anomalous EventsDropHitting
Table 4. Data counts for base and example domain datasets.
Table 4. Data counts for base and example domain datasets.
LabelBaseDomain1Domain2Domain3
airutils3970347637423
brushing1560 39453
cooking3070 1032
flushing66021280
hitting1290 324
microwave1930 832
peeing920230128
speech4220 596
tv5920 1253
watering130601265
watering218401951978938
Table 5. Performance depending on dataset processing.
Table 5. Performance depending on dataset processing.
ProcessAccuracyLossF1
None0.8430.4750.829
Resampling0.9370.1830.938
Augmentation0.940.1720.939
Res+Aug0.9500.1630.951
Quantization0.9420.1900.941
Table 6. Model classification and runtime performance on performance core.
Table 6. Model classification and runtime performance on performance core.
ModelAccuracyLossF1 ScoreParametersMemory (Bytes)Runtime (ms)
Our Model0.9320.2160.93431,38647,936484
LeNet50.9430.1760.949303,762311,272740
VGG130.9150.2570.911246,980269,336731
ResNet180.9420.1900.94146,85469,872841
TCN0.8670.3340.885245,740396,7361451
LSTM (unquantized)0.9500.1580.945141,580Quantization Unavailable
Table 7. Performance with post-processing.
Table 7. Performance with post-processing.
AccLossF1
Original0.9300.2310.744
Processed0.9630.2350.776
Table 8. Generated logs.
Table 8. Generated logs.
A. Handwashing and Brushing Teeth
Start (s)End (s)LabelDuration (s)Reliability
0.008.00hitting8.000.80
13.5015.25airutils1.750.64
7.2526.25watering219.000.97
28.2530.75airutils2.500.73
30.2535.75hitting5.500.83
34.5059.00brushing24.500.79
68.2569.25brushing1.000.68
72.7574.00airutils1.250.65
73.2575.25hitting2.000.38
89.5090.50speech1.000.70
96.5098.00airutils1.500.68
97.75102.25hitting4.500.73
62.0095.75watering233.750.87
B. Taking Shower and Drying
Start (s)End (s)LabelDuration (s)Reliability
8.2513.25airutils5.000.67
0.0014.50hitting14.500.70
13.7514.75speech1.000.70
15.00127.00watering2112.000.91
127.25132.25speech5.000.80
133.50141.25background7.750.74
161.50162.25peeing0.750.73
141.50164.50airutils23.000.73
148.25161.00hitting12.750.70
Table 9. Average time and standard deviation for pre-processing in the edge node.
Table 9. Average time and standard deviation for pre-processing in the edge node.
AverageDeviation
14.209 ms13.129 ns
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, C.; Yuh, A.H.; Kang, S.J. Real-Time Prediction of Resident ADL Using Edge-Based Time-Series Ambient Sound Recognition. Sensors 2024, 24, 6435. https://doi.org/10.3390/s24196435

AMA Style

Lee C, Yuh AH, Kang SJ. Real-Time Prediction of Resident ADL Using Edge-Based Time-Series Ambient Sound Recognition. Sensors. 2024; 24(19):6435. https://doi.org/10.3390/s24196435

Chicago/Turabian Style

Lee, Cheolhwan, Ah Hyun Yuh, and Soon Ju Kang. 2024. "Real-Time Prediction of Resident ADL Using Edge-Based Time-Series Ambient Sound Recognition" Sensors 24, no. 19: 6435. https://doi.org/10.3390/s24196435

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop