GeriatricCare 4.0: A Novel 3D Context-Based CareVision Framework for Fall Detection, Fall Classification and Fall Alerts for Elderlies

GeriatricCare 4.0: A Novel 3D Context-Based CareVision Framework for Fall Detection, Fall Classification and Fall Alerts for Elderlies

Shital Patel* Amit Lathigara

Department of Computer and Engineering, R K University, Kasturbadham, Rajkot 360020, Gujarat, India

Corresponding Author Email: 
spatel232@gmail.com
Page: 
661-669
|
DOI: 
https://doi.org/10.18280/ria.380230
Received: 
10 October 2023
|
Revised: 
18 November 2023
|
Accepted: 
29 December 2023
|
Available online: 
24 April 2024
| Citation

© 2024 The authors. This article is published by IIETA and is licensed under the CC BY 4.0 license (http://creativecommons.org/licenses/by/4.0/).

(This article is part of the Special Issue the Impact of AI on Decision-Making)

OPEN ACCESS

Special Issue
Abstract: 

Falls can cause severe injuries, if an elderly person has a "long-life." In the presented research work, we have applied the widely known benchmark dataset "L2ei" and the customized "Geriatric-2000" dataset, which contains more than 6,000 elderly fall video sequences. The 25 human skeleton features are extracted using the customized 3D Open Pose methodology. The proposed research work presents a customized 3D Context-based LSTM-CNN enabled CareVision framework for fall position classification for elderlies. Furthermore, the proposed research work is compared with other customized AI-enabled computer vision approaches such as Fine KNN, Medium KNN, Decision Tree, long short-term memory network (LSTM), Bi-LSTM and Recurrent Neural Network (RNN). The proposed 3D CareVision Framework has achieved an accuracy of 98.23 percent and a ROC value of 0.96. The indicated results demonstrate the efficiency and reliability of the proposed 3D CareVision Framework for elderly fall position classifications. The proposed 3D CareVision Framework will assist elderly personnels in case of emergencies and notify house members by sending emergency fall alerts.

Keywords: 

fall position classification, fall alert, geriatric care, fall forecasting, elderly health, preventive care, ageing

1. Introduction and Related Work

A recent report on aging and health predicts that by 2050, there will be twice as many senior citizens over 60 outnumbering other age groups [1, 2]. As a result, concerns about the health and safety of the elderly are mounting. Falling is one of the factors that contributes to fatalities and serious injuries in the elderly. An immediate approach to elderly patients may help keep them out of danger or perhaps save their lives. Therefore, it’s essential to offer senior citizens medical aid following a fall in order to protect their health. Whether it comes to geriatric care or people with medical disorders that make them prone to falling, fall detection technology is used to automatically detect whether a person has fallen [3]. The primary objective of fall detection systems is to offer early help to the individual, such as alerting family members, caretakers, or emergency services, in order to ensure prompt medical attention and avoid any potential harmful fall-related effects [4, 5].

Figure 1 represents classification of fall detection approaches. As shown in Figure 1, data from human actions is used in approaches to sensor-based statistical falls, either in the form of inertial sensors or other sorts of sensors. The term "vision-based fall detection" refers to a technique for detecting and classifying fall positions that watches people’s movements with cameras or other visual sensors. In the presented research work, we have aimed at elderly fall detection and forecasting using a customized 3D CareVision Framework. Falls can have serious effects and necessitate rapid medical attention. Vision-based fall detection systems may automatically recognize and respond to probable falls by utilizing computer vision algorithms, machine learning, and image processing techniques [6]. Anitha and fellow researchers have proposed a visual fall recognition system using an LSTM based fall detection approach using human skeleton key points. However, they did not propose a complete 3D elderly care vision framework for fall detection and forecasting [7]. Amsaprabhaa and fellow researchers have implemented a fused vision based fall detection approach using spatiotemporal data and skeletal gait features. However, they did not propose 3D vision-based fall detection, fall forecasting, and fall alert approaches. Furthermore, they did not discuss any ideas related to aging care, fall alert systems [8]. Wang and his team have discussed ideas related to human motion recognition using a triboelectric nanogenerator. However, they did not discuss any ideas related to vision-based elderly fall detection [9]. Quan and fellow researchers have proposed vision based tactile sensing for large data sets. However, they did not discuss or propose any frameworks for fall detection and forecasting [10]. Lian and his team have proposed a video surveillance based approach for fall detection. However, they did not discuss any ideas related to 3D vision-based fall detection methodologies [11]. Zhang and fellow researchers [12] have discussed image-processing based fall detection in bus compartments. However, they did not discuss anything related to elderly care, elderly fall detection, and fall alerts. Pian and his research team have proposed a multimodal approach for fall detection. The approach was focused on achieving data privacy and security for the fall detection data. However, they did not discuss any ideas related to elderly falls, fall alerts, fall detection, or forecasting [13]. Wang and fellow researchers have implemented CNN-based fall detection using radar signals. However, they did not discuss anything related to elderly fall detection, or aging care [14]. Li and his team have discussed a frame-prediction based fall detection approach using pre-captured surveillance video samples. However, they did not discuss any ideas related to 3D-vision based fall detection for elderlies [15]. De and fellow researchers have proposed an indoor body activity based classification approach for fall detection [16]. However, they did not discuss any ideas related to elderly care, and elderly fall alerts. Mobsite and her team [17] have proposed posture based fall detection system for monitoring human activities. However, they did not discuss any ideas related to vision based fall detection and forecasting, and elderly care. Wang and fellow researchers have performed a detailed comparative analysis of various AI methodologies for fall detection of human movements. However, the researchers [18] did not discuss any ideas related to vision based fall detection, IoT enabled fall detection and forecasting, or monitoring elderly people for emergency situations. Sofia and her fellow researchers have proposed an SVM-enabled fall detection methodology for wheelchair based fall detection. However, they did not discuss any ideas related to elderly care, vision-based fall detection and forecasting [19]. Wu and fellow researchers [20] have proposed a floor vibration based wireless fall detection methodology. However, they did not discuss any ideas related to vision based fall detection and forecasting, and elderly care. Li and his research team have discussed a time-series data based fall detection approach using key points. The approach was focused on achieving data privacy and security of the fall detection data. However, they did not discuss any ideas related to elderly falls, fall alerts, fall detection [21]. Yousuf and his fellow researchers [22] have proposed a ubiquitous sensor-based architecture for fall detection. However, they did not discuss any ideas related to elderly care, vision based fall detection. Rezaei and fellow researchers [23] have proposed a human fall detection approach using radar sensors. However, the researchers did not discuss any ideas related to vision based fall detection and forecasting. Mojidra and fellow team members [24] have proposed an AI-based vision approach for crack detection and forecasting. However, they did not discuss any ideas related to vision-based fall detection and forecasting, and elderly care.

Figure 1. A representation of various ambient-based, wearable-based and vision-based fall detection methodologies

In this study, we have proposed a novel 3D Carevision Framework for analyzing and evaluating elderly fall detection movements in various directions and notifying family members in emergency situations. In conducted experiments, we have discussed a customized iterative conformal geometric algebra methodology, a customized 3D Open Pose-based human skeleton detection, and an AI-enabled CareVision Posture Classification Framework for fall detection and forecasting. In recent times, fellow researchers have carried out scatter attempts and proposed vision-based fall detection methodologies [25-34]. However, a complete 3D CareVision framework for fall detection, fall alerts, and forecasting for elderlies has remained an open research problem for fellow researchers.

1.1 Objectives and contributions

The primary objectives and major contributions of the proposed 3D CareVision Framework are the following:

  • The elderly video dataset creation termed "Geriatric-2000", along with the benchmark Le2i fall video sequences dataset. In the presented research work, we have applied the widely known benchmark dataset "L2ei" and the customized "Geriatric-2000" dataset, which contains more than 6,000 elderly fall video sequences. The 25 human skeleton features are extracted using the customized 3D Open Pose methodology.
  • Analysis and Detection of twenty-five human skeleton points using the customized 3D Open Pose Methodology have been discussed as described in section 2.
  • Design and Development of a Customized 3D Open Pose CareVision Fall Estimation and Detection Framework for Fall Detection of Elderlies.
  • Design and Development of a Customized LSTM-CNN Enabled Context-based 3D CareVision Framework for Fall position classifications.
  • The proposed 3D CareVision Framework will assist elderly personnels in case of emergencies and notify house members by sending emergency fall alerts.
2. 3D Human Skeleton Points Model and Description of Limb Parts

A human body skeleton model is a representation of the skeletal system, which serves as the body’s framework and allows for movement, support, and safety. Bones, cartilage, ligaments, and tendons make up the skeleton [7]. The axial skeleton and the appendicular skeleton are two of its main components. A basic description of these pieces and their components is provided below:

2.1 Upper limbs

The shoulder joins the upper arm (humerus) to the body via a ball-and-socket joint. It permits a large range of motion.

2.1.1 Arm

The arm is located between the forearm, wrist, shoulder, and elbow.

2.1.2 Forearm

The radius and ulna are the two bones that make up the forearm.

2.1.3 Wrist

Carpals, a group of several tiny bones, make form the wrist. It gives the hand flexibility.

2.1.4 Hand

The palm, fingers, and thumb make up the hand. Except for the thumb, which has two phalanges, each finger has three phalanges. The hand is necessary for manipulating and gripping objects.

2.2 Lower limbs

2.2.1 Hip

The hip joint joins the femur, or thigh bone, to the pelvic. A ball-and-socket joint, it bears the weight of the body and permits mobility.

2.2.2 Thigh

Femur bone makes up the thigh, which is the upper leg. Strong movements like walking and running are caused by it.

2.2.3 Knee

The complicated hinge joint in the knee allows the leg to be bent and straightened. Ligaments and tendons support it.

2.2.4 Leg

The leg is made up of the lower leg’s tibia and fibula bones. These bones aid in movement and offer structural support.

2.2.5 Ankle

The joint at the ankle joins the leg to the foot. The foot can move up and down thanks to it.

2.2.6 Foot

The toes, arch, and heel all make up the foot. With the exception of the big toe, which has just two, each toe contains three phalanges. During motions like walking and sprinting, the foot offers stability, balance, and propulsion. In terms of mobility, support, manipulation, and interaction with the environment, each component of a limb performs a specific role. Figure 2 depicts 2D skeleton model, 3D human skeleton model, and the corresponding 3D human skeleton model.

Figure 2. Representation of 2D and 3D human skeleton points model

3. Elderly CareVision 3D OpenPose Fall Estimation and Detection Methodology

Multi-person skeleton information can be detected in real-time by the 3D OpenPose human key node recognition system. It employs the feature vector affinity parameter to determine the hot spot map of human key nodes after using the top-down human body attitude estimate technique to identify the locations of key points on the human body. OpenPose is capable of realizing posture estimate as well as human movement, facial expression, and finger movement. It has outstanding robustness and is appropriate for both one person and several people. The crucial point coordinates are extracted and normalized after the posture estimation method tracks and outputs the key point coordinates of the human 3D skeleton [25, 26]. H={h0, h1, · · · , h13} denotes the joint position set for ease of representation. The Joint Coordinates ( hc ) are defined as: Establish the node j’s position at time t as hc(t)=(xtn, ytn), n∈{0, 1, · · · , 13}. In conducted experiments, the surveillance camera enables 3D OpenPose CareVision Estimation methodology to gather data on twenty-five human key nodes. Figure 3 represents a processing of an elderly frame using the 3D human skeleton model.

Figure 3. Elderly video frame processing using 3D Human skeleton model

Figure 4. Video sequences of elderly falling process using 3D CareVision estimation and fall detection methodology

The center of gravity of the human body will shift vertically during a quick fall, as depicted in Figure 4. The center of the hip joint can mirror this feature and serve as a representation of the body’s center of gravity. The longitudinal coordinates of the hip joint center point of each frame of the image are determined by processing the joint point data from the Open Pose. It is identified once every five adjacent frames with a time interval of 0.25 seconds because the transition from standing posture to falling posture takes place in a very brief amount of time.

Hips coordinates can be defined as: $\mathrm{h} c_1 \mathrm{o}(\mathrm{n})=\left(\mathrm{p}_{\mathrm{n} 10}, \mathrm{q}_{\mathrm{n} 10}\right)$ and $\mathrm{x}_{10}(\mathrm{t})=\left(\mathrm{x}_{\mathrm{n} 10}, \mathrm{y}_{\mathrm{n} 10}\right)$. The center of the hip coordinates at the time $\mathrm{n} 1$ is $q n_1=\frac{q_{n 1} 10+q_{n 1} 10}{2}$ and the $\mathrm{y}$-coordinate at time $\mathrm{n} 2$ is $q n_2=\frac{q_{n 2} 10+q_{n 2} 10}{2}$. The hip center descent velocity can be defined as:

$\Delta_n=n_2-n_1$          (1)

$\mathrm{dv}=\frac{\left|q_{n 2}-q_{n 1}\right|}{\Delta n}$       (2)

where, the value of z is greater than or equal to the descent velocity $\overline{d v}$, the elderly fall is detected. According to the conducted experiments, we have set 0.0085 p/q as the fall threshold of the falling velocity of the detected center of the hip joint of an individual, where $\mathrm{z} \geq \underline{\mathrm{z}} X_1=1$.

$X_1=\left\{\begin{array}{l}0 ; z<\overline{d v} \\ 1 ; z \geq \overline{d v}\end{array}\right.$      (3)

The human body’s tendency to tilt is the most noticeable aspect of falling, and this tendency will continue to increase. In this study, a human centerline HC is created in order to express the characteristics of the body’s constant tilt during the process of human fall (Let l be the link between midpoint mp and joint mp0, and let mp12 be the midway of joint mp12 and joint point mp25). Where angle between the ground and the human centerline is defined as θ.

The proposed customized 3D Open pose methodology detects the joint points $0,1,2, \ldots 25$ are $m p_0(t)=\left(p_n, q_n\right)$, $m p_{10}(\mathrm{t})=\left(\mathrm{p}_{\mathrm{n} 10}, \mathrm{q}_{\mathrm{n} 10}\right)$ and $\mathrm{mp}_{13}(\mathrm{n})=\left(\mathrm{p}_{\mathrm{q} 13}, \mathrm{q}_{\mathrm{n} 13}\right)$ respectively. Therefore, $\overline{m p}=\frac{m p_{10}+m p_{13}}{2}, \overline{m p}(\mathrm{n})=\left(p_n, q_n\right)$. The angle between the ground and the human centerline at time $\mathrm{n}$ can be defined as,

$\theta_n=\arctan \left|\frac{q_{n 0}-\overline{q_n}}{p_{n 0}-\overline{p_n}}\right|$       (4)

$X_2=\left\{\begin{array}{l}0 ; \theta \geq \theta_0 \\ 1 ; \theta<\theta_0\end{array}\right.$       (5)

when θ<θ00=45◦) M2=1, can be considered as the occurrence of an elderly fall event. The body tilt, which becomes more pronounced when one is falling, is the most noticeable aspect of a human body. The center line HC of the human body is established in this study to reflect the characteristics of the body’s continuous tilt during the process of human fall (let the link between the midpoint of joint mp and joint mp0 be g, and let the midpoint of joint mp12 and joint point mp13 be h). The outer rectangle of the human body has a width to height ratio of HP=Width/Height. The outer rectangle of the target will change as the human body hits it; the change in the length-to-height ratio is an important parameter in this situation.

$X_3=\left\{\begin{array}{l}1 ; H P \geq N \\ 0, H P<N\end{array}\right.$       (6)

where, the threshold value is N. The width-to-height ratio HP of an elderly body when walking properly is less than 1, whereas it is more than 1 when falling occurs, according to the elderly fall occurrence event. The occurrence of the elderly fall event is satisfied when,

$\mathrm{HP} \geq \mathrm{N}, X_3=1$       (7)

Figure 5. Elderly standing process after fall using 3D

CareVision Estimation and Fall Detection methodology

No fall alert will be generated if an elderly person can stand up on their own after falling. The majority of fall detection relies on the examination of the fall process, without considering the fact that most people rise up on their own shortly after the occurrence of a fall event. As depicted in Figure 5, getting up after falling might be seen as the opposite of falling. The slowness of the entire procedure compared to a fall is the only distinction. Algo. 1 describes the step-wise 3D Open Pose CareVision Fall Estimation and Detection Methodology.

Algorithm 1 The proposed 3D Open Pose CareVision Fall Estimation and Detection Methodology

Input: The continuous elderly video frames EI in the video. In total, 6000 elderly fall video sequences and 25 human skeleton points were considered.

Output: The result of elderly fall detection EIAF, if a fall is abnormal, or EINF, if a fall is normal.

Function ElderlyFallDetection (VideoSequences [EIRGB, EIDepth])

Step 1. Capture the current elderly video sequence EIn

Step 2. Extract the human body features, RGB and depth images EI Depth∈Hc×t×3, EI Depth∈Hc×h, Extract feature: Hc:=VideoSequences (EIRGB, EIDepth). The feature extraction was carried out using the customized 3D Openpose Carevision Framework.

Step 3. Detect twenty-five human joints using 3D Open-pose Estimation, detect joint human body coordinates: Hc:=ρt(Hc).

Step 4. Construct a 3D Human Body Set (Hc) from the processed elderly fall video sequences.

Step 5. The elderly fall detection EIAF, if a fall is abnormal, else EINF, if a fall is normal.

4. Context Based CareVision Fall Detection and Classification For Elderlies

By analyzing RGB images or depth images that are obtained using a single camera system, multi-camera system, or depth camera systems, vision-based fall detection systems are made possible. Despite extensive research utilizing both RGB and depth systems [25-34], the majority of these systems are unable to distinguish between falls and fall-like actions while standing, sitting, bending, lying down, etc. Systems for detecting falls are anticipated to be quick. A slight delay or error in judgment could have serious repercussions.

4.1 Data collection

In conducted experiments, we have used widely known L2ei dataset which contains more than 6,000 fall video sequences, and the customized elderly dataset termed as “Getriatric-2000”, which contains more than 2,000 elderly fall video sequences. https://github.com/YifeiYang210/Fall_Detection_dataset.

Systems for detecting falls are anticipated to be quick. A slight delay or error in judgment could have serious repercussions. By utilizing a single camera-based system, our suggested architecture intends to reduce the computational complexity of fall detection and interior monitoring systems (we have solely used the dataset’s lateral camera input for creation of a customized “Geriatric-2000” dataset). Furthermore, we have divided the number of elderly video sequences in training and testing set with the ratio of 80 and 20 percent, respectively. Additionally, we replace the ambient sensors with digital fall information gathered via cameras. This makes it more appropriate for real-time solutions because wearable gadget sensors are quite inconvenient for the user. Figure 3 depicts the architecture of the suggested approach, and Table 1 provides a summary of the major points taken from the human skeleton.

Table 1. Accuracy Comparison Metrics

List of Methodologies

Accuracy (in Percentage)

The 3D CareVision LSTM-CNN methodology

98.23

FineKNN

84.21

Medium KNN

85.09

Decision Tree

87

ANN

89.91

LSTM

90.86

Bi-LSTM

91.5

RNN

92.4

4.2 The proposed customized LSTM-CNN enabled context-based 3D CareVision framework for fall detection and classification for elderlies

Figure 6. The proposed context-based CareVision fall detection and fall classification framework

A Time Distributed convolution neural network is fed to the feature vector. A set of features are arranged sequentially as a consequence of the sequence learning issue being solved with the help of the Time Distributed wrapper. We employed a window with a size of 64 and an overlap with a size of 48 to provide input to the CNN input layer. We could give the CNN input layer an equalized feature vector in this method. Recurrent neural networks (RNN), a class of neural networks that are frequently employed for models whose issues are characterized by data sequences, including long-term and short-term memory networks (LSTM). The common network parameters over multiple points of the sequence under analysis set these two types of networks apart from other neural networks. Each of the output items is influenced by the outputs of the preceding elements. The model can consider a sequence as a whole rather than just an individual element thanks to the special persistence of the characteristics found in the data. In conducted experiments, we have added 3 layers, which contain a convolution layer and 2 hidden layers, with the kernel size of 3x3. RNNs, unlike other networks, make decisions for future time slots based on feedback from previous time slot activation. Figure 6 depicts the architecture of the proposed 3D Context-based LSTM-CNN Enabled Fall Detection and Fall Classification approach. As shown in Figure 6, The elderly fall video sequences collected from the benchmark L2ei dataset and the customized Geriatric-2000 dataset will be fed to the 3D Open Pose estimation and fall detection methodology, as described in Algo. 1. The input elderly fall video sequences will be converted into the corresponding 3D human skeleton as shown in Figure 3. The 3D Open Pose methodology will extract twenty-five human skeleton features such as head, neck, elbow, hand, foot, etc. The extracted elderly features will be sent to the customized 3D Context-based LSTMCNN-enabled CareVision Framework. The extracted elderly fall featured will be fed to CNN (Convolution Neural Network) to identify whether a fall event has occurred or not. The elderly fall event has been termed ADLs in Figure 6. If an elderly fall is detected, it will be sent to the corresponding LSTM layers to classify elderly falls in various positions: (i) elderly fall in a standing position, (ii) Elderly fall in a sitting position, (iii) Elderly fall in a bending position, (iv)Elderly fall in a lying position. In the proposed study, the velocity threshold and tilt angle threshold were set based on the conducted elderly fall experimentations. The step-wise procedure of elderly fall classification according to various positions is described in Algo. 2.

Algorithm 2 The proposed Context-based 3D CareVision Fall Detection and Classification Methodology

input: Elderly Fall Video Sequences of L2ei and Getriatric-2000 dataset

output: Elderly fall in a standing position (ElderlyFallstanding)

or

Elderly fall in a sitting position (ElderlyFallsitting)

or

Elderly fall in a bending position (ElderlyFallbending)

or

Elderly fall in a lyingposition (ElderlyFalllying)

Function: Elderly Fall Position Classfication (VideoSequences [EIFallVideoSequnces, ElderlyFallposition])

1. Capture the current elderly fall video sequences ElderlyFallp

2. Build LSTM-CNN Enabled Deep Learning Model

3. Initiate LSTM-CNN Enabled Deep Learning Model Training using Elderly Fall Sequences for m=LSTMepoch1, LSTMepoch2, ..., LSTMepochn do

4. Model (n)=Input Pre-Processed Elderly Fall Video Sequences

Compile LSTM-CNN Enabled Deep Learning Model model←loss=Categorical Cross Entropy

model←Metric=Accuracy

model←Optimizer=Adam

TrainLSTM−CNN Enabled Deep Learning Model

model←Dataset

Store LSTM−CNN Enabled

Fall Classification Results =Elderly fall in a standing position (ElderlyFallstanding)

or

Elderly fall in a sitting position (ElderlyFallsitting)

or

Elderly fall in a bending position (ElderlyFallbending)

or

Elderly fall in a lying position (ElderlyFalllying)

5. Result and Discussions

In conducted experiments, we have applied elderly video sequences, generated the corresponding 25 3D human skeleton joint points, and trained the proposed 3D CareVision LSTM-CNN Enabled Framework for fall detection and fall position classification of elderly as described in Section 4.2. The customized LSTM-CNN enabled deep learning model is compared with the customized machine learning methodologies such as Fine KNN (K-Nearest Neighbors Algorithm), Medium KNN, Decision Tree, and customized deep learning approaches such as artificial neural network (ANN), long short-term memory network (LSTM), and Recurrent Neural Network (RNN), which have received corresponding accuracy of 84.21, 85.09, 87, 89.91, 90.86, 91.5, and 92.4. The customized 3D CareVision LSTM-CNN-enabled Framework for fall detection and fall position classification Framework has achieved an accuracy of 98.23 percent and a ROC value of 0.96. Figure 7 and Figure 8 depict the training and testing accuracy and loss representation of the proposed 3D Context-based LSTM-CNN-enabled Fall Detection and Fall Classification approach. Figure 9 represents the ROC curve representation of the proposed CareVision, and the customized machine learning methodologies such as Fine KNN (K-Nearest Neighbors Algorithm), Medium KNN, Decision Tree, and customized deep learning approaches such as artificial neural network (ANN), long short-term memory network (LSTM), and Recurrent Neural Network (RNN). Figure 10, Figure 11, Figure 12, Figure 13 represents the fall position classification results of the proposed 3D Context-based LSTM-CNN-enabled Fall Detection and Fall classification approach in standing, sitting, bending, and lying positions. Table 1 describes the metrics comparison of the customized CareVision LSTM-CNN framework, with the customized machine learning methodologies such as Fine KNN (K-Nearest Neighbors Algorithm), Medium KNN, Decision Tree, and customized deep learning approaches such as artificial neural network (ANN), long short-term memory network (LSTM), and Recurrent Neural Network (RNN).

Figure 7. Training and validation accuracy representation of the proposed 3D CareVision LSTM-CNN enabled framework

Figure 8. Training and validation loss representation of the proposed 3D CareVision LSTM-CNN enabled framework

Figure 9. ROC curve representation of proposed 3D CareVision LSTM-CNN enabled framework

Figure 10. Fall classification results in a standing position

Figure 11. Fall classification results in a sitting position

Figure 12. Fall classification results in a bending position

Figure 13. Fall classification results in a lying position

6. Conclusions

Elderly Falls can result in severe injuries, including death if an elderly person has a “long-life.” As a result, a trustworthy fall detection system is needed to send out an emergency fall alerts for first aid. Fellow researchers have done scatter attempts in designing a complete AI-enabled fall detection systems to analyze geriatric fall detection events. However, a complete AI-enabled fall detection and fall position classification has remained an open research problem. In the proposed research work, we have designed a customized 3D Context-based LSTM-CNN enabled CareVision framework for fall detection and fall position classification for elderlies. In conducted experiments, we have applied the widely known benchmark dataset “L2ei” and the customized “Geriatric-2000” dataset, which contains more than 6000 elderly fall video sequences. In the initial stage, the 25 human skeleton features are extracted using the customized 3D Open Pose methodology. The extracted elderly fall featured will be fed to CNN (Convolution Neural Network) to identify whether a fall event has occurred or not. If an elderly fall is detected, it will be sent to the corresponding LSTM layers to classify elderly falls in various positions: (i) elderly fall in a standing position, (ii) Elderly fall in a sitting position, (iii) Elderly fall in a bending position, (iv)Elderly fall in a lying position. The proposed customized 3D Open Pose methodology can detect 25 human skeleton points. However, the existing human pose estimations methods can only detect 12 to 17 human skeleton points. For this reason, in the conducted experiments, we have designed a customized 3D Openpose methodology.

Furthermore, the proposed customized LSTM-CNN enabled deep learning model is compared with the customized machine learning methodologies such as Fine KNN (K-Nearest Neighbors Algorithm), Medium KNN, Decision Tree, and customized deep learning approaches such as artificial neural network (ANN), long short-term memory network (LSTM), and Recurrent Neural Network (RNN), which have received corresponding accuracy of 84.21, 85.09, 87, 89.91, 90.86, 91.5, and 92.4. The customized 3D CareVision LSTM-CNN-enabled Framework for fall detection and fall position classification Framework has achieved an accuracy of 98.23 percent and a ROC value of 0.96 and outclassed other machine learning and deep learning methodologies. The indicated results demonstrate the efficiency and reliability of the proposed CareVision Framework for elderly fall classification in positions such as standing, sitting, bending, and lying. The customized 3D CareVision LSTM-CNN-enabled Framework for fall detection and fall position classification Framework has outperformed existing methodologies such as, FineKNN, Medium KNN, Decision Tree, ANN, LSTM, RNN as described in Table 1.

The proposed 3D CareVision LSTM-CNN-enabled Framework can provide approximate fall event alerts in case of emergencies. In the future, the proposed fall detection and fall position classification framework can be enhanced to achieve more accurate fall event description and responsive fall alerts.

  References

[1] Kulurkar, P., Kumar Dixit, C., Bharathi, V.C., Monikavishnuvarthini, A., Dhakne, A., Preethi, P. (2023). AI based elderly fall prediction system using wearable sensors: A smart home-care technology with IOT. Measurement: Sensors, 25: 100614. https://doi.org/10.1016/j.measen.2022.100614

[2] Shen, M., Tsui, K.L., Nussbaum, M.A., Kim, S., Lure, F. (2023). An indoor fall monitoring system: Robust, multistatic radar sensing and explainable, feature-resonated deep neural network. IEEE Journal of Biomedical and Health Informatics, 27(4): 1891-1902. https://doi.org/10.1109/JBHI.2023.3237077

[3] Patel, W.D., Pandya, S., Koyuncu, B., Ramani, B., Bhaskar, S., Ghayvat, H. (2018). NXTGeUH: LoRaWAN based NEXT generation ubiquitous healthcare system for vital signs monitoring & falls detection. In 2018 IEEE Punecon, IEEE, pp. 1-8. https://doi.org/10.1109/PUNECON.2018.8745431

[4] Elagovan, R., Perumal, T., Krishnan, S. (2023). Fall detection systems at night. Computer, 56(6): 44-51. https://doi.org/10.1109/MC.2022.3200404

[5] Wang, B., Zhang, H., Guo, Y.X. (2022). Radar-Based soft fall detection using pattern contour vector. IEEE Internet of Things Journal, 10(3): 2519-2527. https://doi.org/10.1109/JIOT.2022.3213693

[6] Saadeh, W., Butt, S.A., Altaf, M.A.B. (2019). A patient-specific single sensor IoT-based wearable fall prediction and detection system. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 27(5): 995-1003. https://doi.org/10.1109/TNSRE.2019.2911602

[7] Inturi, A.R., Manikandan, V.M., Garrapally, V. (2023). A novel vision-based fall detection scheme using keypoints of human skeleton with long short-term memory network. Arabian Journal for Science and Engineering, 48(2): 1143-1155. https://doi.org/10.1007/s13369-022-06684-x

[8] Amsaprabhaa, M. (2023). Multimodal spatiotemporal skeletal kinematic gait feature fusion for vision-based fall detection. Expert Systems with Applications, 212: 118681. https://doi.org/10.1016/j.eswa.2022.118681

[9] Wang, S., Gao, J., Lu, F., Wang, F., You, Z., Huang, M., Fang, W., Liu, X., Li, Y., Liu, Y. (2023). Human motion recognition by a shoes-floor triboelectric nanogenerator and its application in fall detection. Nano Energy, 108: 108230. https://doi.org/10.1016/j.nanoen.2023.108230

[10] Luu, Q.K., Nguyen, N.H. (2023). Simulation, learning, and application of vision-based tactile sensing at large scale. IEEE Transactions on Robotics. https://doi.org/10.1109/TRO.2023.3245983

[11] Wu, L., Huang, C., Zhao, S., Li, J., Zhao, J., Cui, Z., Yu, Z., Xu Y., Zhang, M. (2023). Robust fall detection in video surveillance based on weakly supervised learning. Neural Networks, 163: 286-297. https://doi.org/10.1016/j.neunet.2023.03.042

[12] Zhang, X., Ji, J., Wang, L., He, Z., Liu, S. (2023). Image‐based fall detection in bus compartment scene. IET Image Processing, 17(4): 1181-1194. https://doi.org/10.1049/ipr2.12705

[13] Qi, P., Chiaro, D., Piccialli, F. (2023). FL-FD: Federated learning-based fall detection with multimodal data fusion. Information Fusion, 101890. https://doi.org/10.1016/j.inffus.2023.101890

[14] Wang, P., Li, Q., Yin, P., Wang, Z., Ling, Y., Gravina, R., Li, Y. (2023). A convolution neural network approach for fall detection based on adaptive channel selection of UWB radar signals. Neural Computing and Applications, 35(22): 15967-15980. https://doi.org/10.1007/s00521-021-06795-w

[15] Li, S., Song, X. (2023). Future frame prediction network for human fall detection in surveillance videos. IEEE Sensors Journal. https://doi.org/10.1109/JSEN.2023.3276891

[16] De, A., Saha, A., Kumar, P., Pal, G. (2023). Fall detection approach based on combined two-channel body activity classification for innovative indoor environment. Journal of Ambient Intelligence and Humanized Computing, 14(9): 11407-11418. https://doi.org/10.1007/s12652-022-03714-2

[17] Mobsite, S., Alaoui, N., Boulmalf, M., Ghogho, M. (2023). Semantic segmentation-based system for fall detection and post-fall posture classification. Engineering Applications of Artificial Intelligence, 117: 105616. https://doi.org/10.1016/j.engappai.2022.105616

[18] Wang, D., Li, Z. (2023). Comparison of four machine learning algorithms for a pre-impact fall detection system. Medical & Biological Engineering & Computing, 1-14. https://doi.org/10.1007/s11517-023-02853-8

[19] Sheikh, S.Y., Jilani, M.T. (2023). A ubiquitous wheelchair fall detection system using low-cost embedded inertial sensors and unsupervised one-class SVM. Journal of Ambient Intelligence and Humanized Computing, 14(1): 147-162. https://doi.org/10.1007/s12652-021-03279-6

[20] Wu, K., Huang, Y., Qiu, M., Peng, Z., Wang, L. (2023). Toward device-free and user-independent fall detection using floor vibration. ACM Transactions on Sensor Networks, 19(1): 1-20. https://doi.org/10.1145/3519302

[21] Li, J., Gao, M., Li, B., Zhou, D., Zhi, Y., Zhang, Y. (2023). KAMTFENet: A fall detection algorithm based on keypoint attention module and temporal feature extraction. International Journal of Machine Learning and Cybernetics, 14(5): 1831-1844. https://doi.org/10.1007/s13042-022-01730-4

[22] Yousuf, S., Kadri, M.B. (2023). A ubiquitous architecture for wheelchair fall anomaly detection using low-cost embedded sensors and isolation forest algorithm. Computers and Electrical Engineering, 105: 108518. https://doi.org/10.1016/j.compeleceng.2022.108518

[23] Rezaei, A., Mascheroni, A., Stevens, M.C., Argha, R., Papandrea, M., Puiatti, A., Lovell, N.H. (2023). Unobtrusive human fall detection system using mmWave radar and data driven methods. IEEE Sensors Journal, 23(7): 7968-7976. https://doi.org/10.1109/JSEN.2023.3245063

[24] Mojidra, R., Li, J., Mohammadkhorasani, A., Moreu, F., Bennett, C., Collins, W. (2023). Vision-based fatigue crack detection using global motion compensation and video feature tracking. Earthquake Engineering and Engineering Vibration, 1-21. https://doi.org/10.1007/s11803-023-2156-1

[25] Liu, J., Mu, X., Liu, Z., Li, H. (2023). Human skeleton behavior recognition model based on multi-object pose estimation with spatiotemporal semantics. Machine Vision and Applications, 34(3): 44. https://doi.org/10.1007/s00138-023-01396-0

[26] Zhang, J., Wu, C., Wang, Y. (2020). Human fall detection based on body posture spatio-temporal evolution. Sensors, 20(3): 946. https://doi.org/10.3390/s20030946

[27] Liu, C.P., Li, J.H., Chu, E.P., Hsieh, C.Y., Liu, K.C., Chan, C.T., Tsao, Y. (2023). Deep learning-based fall detection algorithm using ensemble model of coarse-fine CNN and GRU networks. arXiv Preprint arXiv: 2304.06335. https://doi.org/10.48550/arXiv.2304.06335

[28] Matos-Carvalho, J.P., Correia, S.D., Tomic, S. (2023). Sensitivity analysis of LSTM networks for fall detection wearable sensors. In 2023 6th Conference on Cloud and Internet of Things (CIoT), Lisbon, Portugal, pp. 112-118. https://doi.org/10.1109/CIoT57267.2023.10084906

[29] Jain, R., Semwal, V.B. (2022). A novel feature extraction method for preimpact fall detection system using deep learning and wearable sensors. IEEE Sensors Journal, 22(23): 22943-22951. https://doi.org/10.1109/JSEN.2022.3213814

[30] Ribeiro, N.F., Santos, C.P. (2021). Two fall-related and kinematic data-based approaches for an instrumented conventional cane. IEEE Transactions on Human-Machine Systems, 51(5): 554-563. https://doi.org/10.1109/THMS.2021.3097984

[31] Lu, N., Wu, Y., Feng, L., Song, J. (2018). Deep learning for fall detection: Three-dimensional CNN combined with LSTM on video kinematic data. IEEE Journal of Biomedical and Health Informatics, 23(1): 314-323. https://doi.org/10.1109/JBHI.2018.2808281

[32] Bouazizi, M., Ye, C., Ohtsuki, T. (2021). 2-D LIDAR-based approach for activity identification and fall detection. IEEE Internet of Things Journal, 9(13): 10872-10890. https://doi.org/10.1109/JIOT.2021.3127186

[33] Li, H., Shrestha, A., Heidari, H., Le Kernec, J., Fioranelli, F. (2019). Bi-LSTM network for multimodal continuous human activity recognition and fall detection. IEEE Sensors Journal, 20(3): 1191-1201. https://doi.org/10.1109/JSEN.2019.2946095

[34] Li, X., Chen, P., Jing, L., He, Z., Yu, G. (2022). SwissLog: Robust anomaly detection and localization for interleaved unstructured logs. IEEE Transactions on Dependable and Secure Computing, 20(4): 2762-2780. https://doi.org/10.1109/TDSC.2022.3162857