Detecting Abnormal Health Conditions in Smart Home Using a Drone
Abstract
Nowadays, detecting aberrant health issues is a difficult process. Falling, especially among the elderly, is a severe concern worldwide. Falls can result in deadly consequences, including unconsciousness, internal bleeding, and often times, death. A practical and optimal, smart approach of detecting falling is currently a concern. The use of vision-based fall monitoring is becoming more common among scientists as it enables senior citizens and those with other health conditions to live independently. For tracking, surveillance, and rescue, unmanned aerial vehicles use video or image segmentation and object detection methods. The Tello drone is equipped with a camera and with this device we determined normal and abnormal behaviors among our participants. The autonomous falling objects are classified using a convolutional neural network (CNN) classifier. The results demonstrate that the systems can identify falling objects with a precision of 0.9948.
Index Terms:
Ryze Tello, CNN, Fall detectionI Introduction
According to the World Health Organization (WHO), over 680,000 fatal falls occur worldwide each year, with the majority of victims being individuals over the age of 65[1]. Falls are the leading cause of fatal injuries in older adults and are especially dangerous for those living independently. In order to address this issue, the purpose of our study is to detect falls within the home. With this, we hope to reduce the number of deaths associated with falls and allowing the elderly and others that are prone to falls to still live independently.
An unmanned aerial vehice (UVA), often referred to as a drone is an aircraft without any humans on board. Drones are often used in a number of situations such as emergency rescue, asset protection, and wildlife monitoring, but they can do more. Introducing a drone into the home increases vigilance that allows victims to receive the help that they need in an appropriate amount of time. More specifically, the Tello drone consists of a built-in camera that will be used for tracking, surveillance, and rescue. Through our study, we believe that a drone can be an effective smart home device to help reduce mortality rates and the need of in-home aids for the elderly and others prone to falls.
I-A Proposed Approach
Although our current solution only solves the fall detection aspect, we envisioned our project to have an end-to-end solution based on person detection and fall classification.
As seen in Figure 1 below, our solution accounts for all possible outcomes of person detection and fall classification. The drone will constantly be scanning the area and taking images and videos and if a person is detected we move on to feature extraction to gather all relevant information on the human’s movement. If a fall is detected the user will be asked whether or not they need help. If the human answers yes or does not give an answer at all, an alarm will be triggered, contacting emergency personnel along with family members programmed into the system. However, if a person or a fall is not detected, the drone will go back to the beginning phase to continue its analysis. Lastly, if there is any uncertainty, the drone will re-evaluate by moving to a new location to reassess the situation.
As previously mentioned, our current solution does not offer an end-to-end solution, but this serves as the framework we envisioned our project to follow.
I-B Related Work
Previous studies have presented fall detection systems used to detect patient falls. While other studies have designed a UAV to monitor patients’ behavior in order to alert emergency personnel for victims of falls. This section covers previous work related to UAVs and other fall detection systems such as robots.
I-B1 Work Related to UAV Systems
Previous studies have designed UAVs to transport medical supplies and medication to patients[2]. Other studies have adopted UAVs to have an advanced first aid system to monitor elderly patients with heart conditions putting them at risk of falls. These authors proposed a wearable sensor-based fall detection device (FDD) designed for monitoring HR and detecting falls, and a call emergency center (CEC) to receive messages from the FDD and plan the UAV’s path[3]. The FDD is attached to the patient’s upper arm and once it detects a fall and the HR is measured as abnormal, the FDD sends a message to the CEC which includes the patient’s personal information. The advanced first aid system provides first aid to the patients that are prone to falling[3]. The first aid package is then delivered to the patient via the UAV based on the GPS coordinates of the FDD.
The results illustrated excellent performance where the proposed FDD had an accuracy of 99.2% when distinguishing between falls and normal activities. Furthermore, in terms of HR monitoring, results show that the proposed algorithm for the FDD was successful and can be used for monitoring the elderly who have atrial fibrillation (irregular heartbeat), which causes falls[3]. Moreover, the GPS was accurate at determining the patient’s location at the time of the fall.
In the future, researchers are interested in the development of the entire system, including the design of a system that works autonomously to receive patient information and send a UAV without human intervention[3].
I-B2 Work Related to Robots
Previous studies have developed mobile robots that monitors one person in a single view with constant supervision, while other studies proposed a more robust system. Saturnino Maldonado-Bascon et al., proposed autonomous mobile-patrol robots that integrate privacy protection and only necessary supervision. The robot was designed to autonomously patrol an indoor environment and when a fall is detected, an alarm is activated[4]. This study also believes that vision-based systems offer more advantages than wearable sensor-based systems. However, in vision-based systems, cameras play a critical role. In order to solve the problem of occlusion, the system should be equipped with multiple cameras in order to monitor and assess a number of angles[4]. Recent advancements in Computer Vision support this idea more boldly[5, 6, 7]. This study combines the YOLOv3 algorithm based on a convolutional neural network (CNN) for person detection and a support vector machine (SVM) for fall classification[4]. The basis for this study’s fall detection approach was an assistance robot named LOLA, that was designed entirely by the research team. This device was developed to monitor and help the elderly and others with functional disabilities living alone. Not only is LOLA an assistive robot, but it also serves as a rollator for assistance with walking or as a table[4].
When evaluating the performance of the end-to-end solution, results show that the fall classifier detected 390 true positives, 1 false negative, and 0 false positives, indicating values of 100% and 99.74% for precision and recall, respectively[4]. Overall, results indicate that the system had a high success rate in fall detection.
In the future, researchers hope to investigate and improve upon occlusion detection and possibly merging person detection and fall classification into one CNN[4].
II Data Collection
In our study we utilized two datasets. The first dataset is available to the public and was used to train our model. The second dataset is our primary dataset that was collected from two male participants aged 12 and 16 years old. We collected the data in two locations:
-
•
living room one
-
•
living room two
Furthermore, we followed IRB standards since we were collecting data from human participants.


II-A Design of Hardware
We used a Ryze Tello drone as a low-cost hardware platform to perform our study. The Ryze Tello drone costs approximately 100 USD and is a small and light weight drone that weighs approximately 80 g (with propellers and battery). The full specifications are given below:
-
•
Dimensions: 98×92.5×41 mm
-
•
Propeller: 3 inches
-
•
Built-in Functions: Range Finder, Barometer, LED, Vision System, 2.4 GHz 802.11n Wi-Fi, 720p Live View
-
•
Port: Micro USB Charging Port
-
•
Detachable Battery: 1.1Ah/3.8V
The drone can be operated both indoors and outdoors. A single camera, an IMU with a 3-axis gyroscope, 3-axis accelerometer, and 3-axis magnetometer, and a pressure and IR-based altitude sensors are all incorporated into the drone. The front-facing camera features an 8-megapixel sensor and can record video at a resolution of at 30 frames per second.
II-B Device Setup and Requirements
The Tello drone is a ready-to-use device that requires no additional setup. In order to control the drone, a compatible mobile phone is required along with the ’Tello’ app and wifi. The drone is a battery charged device, lasting approximately 12-13 minutes.
II-C Study Design
As previously mentioned, our study consisted of two male participants, aged 12 and 16 years old. This study took place in a single-family home with approximately 10 inches ceilings. The images and videos were captured in two different rooms within the home, one with more open space and one with less space. The study was carried out over a span of two weeks, with sessions ranging from 15 minutes to an hour depending on drone complications, participant poses and assessments, and retakes.
II-D Participant Instructions
The instructions that our participants followed were straightforward. Using our own intuition, we first determined what positions people usually are in after a fall and took note of it. We then showed our participants an image of the position and instructed them to reenact it. Once our participants were in position, we powered up the drone, captured the videos and images, assessed them and then asked our participants to move on to the next position.
III Proposed Methodology
In this section we briefly discuss fall detection systems and methods.
III-A Convolutional Neural Network (CNN)
We propose a deep learning-based approach for detecting whether or not a person has fallen using RGB photographs. For feature extraction, we employ a CNN algorithm111The proposed idea is taken from [11] with 6 blocks and a header. Conv2D Layer with ReLU activation and MaxPooling2D are included in each block. We begin with 16 troops and increase by 2 units every 2 blocks:
We flatten the result after feature extraction, add a Dense layer of 32 units, and utilize sigmoid activation for classification.
The CNN model uses visual representations based on
the RGB image and segmentation and learns high-level embedding features for fall recognition. We explain the individual components of the conceptual structure in detail below.
III-B Training and Implementation
Initialize the weights of the nodes in the FallNet network model at the start. The weights for the convolutional layer are started in this model using the weights from the following embedding layer, which have a zero-mean gaussian distribution (SD= 0.01 and bias=0) [8]. We utilize the Adam optimizer with a learning rate of 0.0001 and a loss function of Binary Crossentropy to train our model. The proposed system is built on the basis of Torch, Keras libraries[9, 10]. Figure 6 depicts the convolution layers, which range from 1-6 convolutions.
IV Experimental Results
IV-A Datasets
At first, we utilized this publicly available dataset data to train our model for fall detection. The acquired data was then used to validate the pre-trained model. The accuracy and loss plot was obtained after 5 epochs.
IV-B Evaluation Metrics
The loss is the value that a neural network tries to minimize in deep learning: it’s the difference between the ground truth and the forecasts. The neural network learns to decrease this distance by altering weights and biases in a way that minimizes the loss.
In regression problems, for example, you have a continuous target, such as height. The discrepancy between your predictions and the actual height is what you wish to minimize. You can use mean absolute error as a loss to tell the neural network what it should be minimizing.
It’s a little more complicated in terms of classification, but it’s still comparable. Probability is used to predict classes and the neural network minimizes the possibility of assigning a low probability to the real class in classification. Typically, the loss is categorical crossentropy.
The difference between loss and validation loss is that the former is applied to the train set, while the latter is applied to the test set. As a result, the latter is a useful indicator of how the model performs on data that hasn’t been seen before. Use validation data=[] or validation split=0.2 to get a validation set.
To avoid overfitting, it’s advisable to rely on the validation loss. When the model fits the training data too closely, overfitting occurs, and the loss continues to decrease while the val loss remains stale or grows. When the validation loss stops lowering, you can use EarlyStopping in Keras to halt training.
After performing 5 epochs we get the following accuracy and loss plot from training and testing data.
Epoch | Loss | Accuracy | Validation Loss | Validation Accuracy |
---|---|---|---|---|
Table I shows that the model achieves complete accuracy after 5 epochs. Our model performs better on the unseen data, as seen in figures 7 and 8. On testing data, our model is more accurate. Furthermore, the loss of model is minimal. Figure 9 also displays the prediction rate of . That means our model accurately detects the fall.
V Conclusion
V-A Challenges and Limitations
During the duration of our study, we faced some challenges and had some limitations along the way. This may be due to the inexpensive nature of the Tello drone, however during the initial stages of understanding how to fly the drone, the propellers would often fall off on it’s own. After flying the drone several times, we encountered takeoff complications, where the drone could not takeoff from the ground. Even after re-calibrating the device multiple times, we could not successfully address the issue, causing us to utilize a secondary Tello drone.
Secondly, many of the training datasets available publicly required permission to use and given the time constraint of the project, we could not rely on waiting to be approved for access.
Also, for a number of poses we wanted our participants to re-enact, we could not achieve certain angles based on the height limitations of the drone in order to ensure safe flying.
Lastly, the drone’s camera quality often gave us issues when testing the images with the CNN model, sometimes being incapable of determining what the image consisted of.
V-B Takeaway
We used a small drone to detect falls in this experiment. We wanted to test if the Tello drone could be effective in monitoring smart health on a broader scale. Based on our findings, we can safely conclude that we can deploy a Tello drone to monitor the residence. There are some drawbacks as well, because the CNN model requires a large dataset with uniformly distributed classes. Also, the drone’s range is limited, and the video resolution is only 720p with an 8-megapixel sensor. The weather circumstances contribute to hazy photos, and we are unable to fly it in strong winds. It is, nevertheless, quite inexpensive and convenient. We feel that this drone can still be used to monitor the residence in an interior scenario.
V-C Learning Experience
As machine learning and smart home devices, such as a drone are new topics that we had not previously explored, we learned a great deal throughout the duration of the project. First and foremost, it was critical to work together as a team from start to finish. We played on our strengths and assisted the other when in need of help. Furthermore, we learned the importance of trouble shooting because although technology is revolutionary, there will still be some challenges along the way. Being able to turn to one another and other resources when needed allowed us to achieve the best results that we could.
In terms of the drone, we determined that a drone can be an effective device to detect falls in the home if created in a robust fashion. Although drones are initially thought of to be used in the entertainment industry to capture footage in areas where humans cannot reach, it can be applied to many other disciplines such as healthcare. A drone can be used in the home to detect falls, minimize the need for an in-home aid, and reduce unwanted stress.
This project serves as an eye opener that smart home technology can be adapted to make the world a more safer place.
V-D Future Work
In the future, we hope to explore whether the drone is capable alerting appropriate officials in the event of an incident. Furthermore, we would be interested in exploring the possibilities of having an autonomous drone or autonomous vehicle, taking away the burden of having to fly the drone manually. We aim to monitor the system for potential attacks and devise strategies to mitigate them, recognizing that autonomous drones or vehicles are susceptible to security threats[12, 13]. Furthermore, we would like to introduce the idea of having the drone dock itself when it needs to be charged. Lastly, since we did not include mask R-CNN in our analysis, comparing mask R-CNN and CNN algorithms in the future can prove beneficial to the effectiveness of our system.
Acknowledgment
I would like to express my deepest gratitude and appreciation to Professor Dr. Nirmalya Roy for his invaluable guidance, mentorship, and expertise throughout the course of this research endeavor. Furthermore, I extend my heartfelt thanks to Mariam Yekini for her diligent and meticulous data collection efforts. Mariam’s dedication, attention to detail, and competence played a crucial role in ensuring the quality and accuracy of the data used in this study. Her tireless work was instrumental in the successful execution of this research project.
References
- [1] World Health Organization. (2021, April 26). Falls. WHO. https://www.who.int/news-room/fact-sheets/detail/falls
- [2] Eichleay, M., Evens, E., Stankevitz, K., & Parker, C. (2019). Using the unmanned aerial vehicle delivery decision tool to consider transporting medical supplies via drone. Global Health: Science and Practice, 7(4), 500-506. https://doi.org/10.9745/ghsp-d-19-00119
- [3] Fakhrulddin, S. S., Gharghan, S. K., Al-Naji, A., & Chahl, J. (2019). An advanced first aid system based on unmanned aerial vehicles and a wireless body area sensor network for elderly persons in outdoor environments. Sensors, 19(13), 28. https://doi.org/10.3390/s19132955
- [4] Maldonado-Bascón, S., Iglesias-Iglesias, C., Martín-Martín, P., & Lafuente-Arroyo, S. (2019). Fallen people detection capabilities using assistive robot. Electronics, 8(9), 20. https://doi.org/10.3390/electronics8090915
- [5] Maloy Kumar Devnath, Avijoy Chakma, Mohammad Saeid Anwar, Emon Dey, Zahid Hasan, Marc Conn, Biplab Pal, Nirmalya Roy, A Systematic Study on Object Recognition Using Millimeter-wave Radar, in 2023 IEEE International Conference on Smart Computing (SMARTCOMP), 2023, pp. 57-64, https://doi.org/10.1109/SMARTCOMP58114.2023.00025
- [6] Mohammad Saeid Anwar, Emon Dey, Maloy Kumar Devnath, Indrajeet Ghosh, Naima Khan, Jade Freeman, Timothy Gregory, Niranjan Suri, Kasthuri Jayarajah, Sreenivasan Ramasamy Ramamurthy, Nirmalya Roy, HeteroEdge: Addressing Asymmetry in Heterogeneous Collaborative Autonomous Systems, in 2023 IEEE 20th International Conference on Mobile Ad Hoc and Smart Systems (MASS), 2023, pp. 575-583, https://doi.org/10.1109/MASS58611.2023.00077
- [7] Subrata Kumer Paul, Md Abul Ala Walid, Rakhi Rani Paul, Md Jamal Uddin, Md Sohel Rana, Maloy Kumar Devnath, Ishaat Rahman Dipu, Md Momenul Haque, An Adam based CNN and LSTM approach for sign language recognition in real time for deaf people, Bulletin of Electrical Engineering and Informatics, vol. 13, no. 1, pp. 499–509, 2024,https://doi.org/10.11591/eei.v13i1.6059
- [8] Sagar Chhetri et al. “Deep learning for a vision-based fall detection system: Enhanced optical dynamic flow”. In: Computational Intelligence 37.1 (2021), pp. 578–595.
- [9] Campbell G. Zweig MH. “Receiver-operating characteristic (ROC) plots: a fundamental evaluation tool in clinical medicine”. In: Erratum in: Clin Chem 39.4 (1993), pp. 561–77. doi: https://pubmed.ncbi.nlm.nih.gov/8472349/.
- [10] Francois Chollet et al. Keras. 2015. https://github.com/fchollet/keras.
- [11] KananMahammadli. (n.d.). Kananmahammadli/fall-detection-using-CNN-architecture: Detecting whether a person has fallen or not based on images. GitHub. https://github.com/KananMahammadli/Fall-Detection-using-CNN-Architecture
- [12] Riadul Islam, Maloy K Devnath, Manar D Samad, Syed Md Jaffrey Al Kadry, ”GGNB: Graph-based Gaussian naive Bayes intrusion detection system for CAN bus”, Vehicular Communications, vol. 33, p. 100442, 2022, https://doi.org/10.1016/j.vehcom.2022.100442
- [13] Maloy Kumar Devnath, GCNIDS: Graph Convolutional Network-Based Intrusion Detection System for CAN Bus, UMBC Student Collection, 2023, https://doi.org/10.48550/arXiv.2309.10173