Next Article in Journal
A Neural Circuit Model of the Striatum Resolves the Conflict between Context and Dominance Apparent in the Prefrontal Cortex
Previous Article in Journal
An Assessment of Physicochemical Properties and Microbial Count in Dairy Wastewater in Savar Area, Bangladesh
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Autonomous Movement of Wheelchair by Cameras and YOLOv7 †

by
Md Abdul Baset Sarker
,
Ernesto Sola-Thomas
,
Collin Jamieson
and
Masudul H. Imtiaz
*
Department of Electrical and Computer Engineering, Clarkson University, Potsdam, NY 13699, USA
*
Author to whom correspondence should be addressed.
Presented at the 3rd International Electronic Conference on Applied Sciences, 1–15 December 2022; Available online: https://asec2022.sciforum.net/.
Eng. Proc. 2023, 31(1), 60; https://doi.org/10.3390/ASEC2022-13834
Published: 9 February 2022
(This article belongs to the Proceedings of The 3rd International Electronic Conference on Applied Sciences)

Abstract

:
A wheelchair can provide limited but crucial mobility to an injured or disabled individual. This paper presents the first stage of the development of a smart wheelchair which is the customization of a manually controlled wheelchair with a novel implementation of octascopic vision. This relatively inexpensive design of an autonomous wheelchair consists of two monochromic camera arrays (each having four cameras) placed around the frame of the wheelchair to achieve a view of 360 degrees. The initial research goal was to design a wheelchair controlled by the embedded processor, allowing the wheelchair to navigate autonomously around an indoor facility with and without human intervention. Additionally, it was intended to allow those previously denied access to the world of automatic wheelchairs because of a low personal income. Through the testing of wheelchair functionality, (a) a large dataset of octascopic images was captured from this wheelchair, and (b) a YOLOv7-based object detection model was developed to avoid obstacles and autonomously control the movement. This paper presents the camera placement and the obstacle detection model using octascopic images. All the project design files have been granted an open-source license and can be reproduced publicly.

1. Introduction

According to the World Health Organization, 15% of the world’s population lives with some form of physical disability [1]. Currently, 61 million adults in the US live with a disability [2]. It means one out of four adults in the US has disability issues. This muscular illness, injury, or disability may cause depression, significant decreases in motivation, and loss of independence for many sufferers. Responding to this, a few mobility-assisting devices have become available in the market, such as wheelchairs. Hence, our primary research motivation was to give those people their autonomy back, allowing them to move safely when and where they want without requiring assistance from caregivers or others.
Studies have shown that cameras are widely used in cars for parking, lane detection, autonomous movement, vision-enabled prosthetic hand, etc. [3,4,5,6,7,8,9,10,11,12]. A combination of LiDAR and cameras also used in navigation of wheelchair [13,14,15,16,17,18,19]. Additionally, it is also noticeable that the Convolution neural network (CNN) is also popular among researchers [20,21,22,23]. YOLOv7 gained popularity in object detection among researchers. The benefits of YOLOv7 include quick convergence, excellent precision, and robust customization. It may also be easily transferred to embedded devices due to its powerful real-time processing capabilities, and low hardware computing needs [24].
Although many projects have been found for autonomous navigation for cars in the outdoor environment, there is less study on eight camera-based autonomous wheelchairs for indoor environments. Because of the structure of the wheelchair, it is harder to place a LiDAR-based system to cover ~360 degrees. Our target was to work in this area to build a robust solution for the autonomous wheelchair using a vision-based solution with multiple cameras and a deep neural network.
The contribution of this study is the presentation of a vision sensor (eight OmniVision camera combination) integrated smart wheelchair which is capable of detecting obstacles from the captured octascopic images. This is completed by implementing the YOLOv7, a deep learning-based object detection model. The project’s design files have been made available in [25]. The design concept is that the eight cameras cover ~360 degrees view and while moving the object detection model will detect any object and send the appropriate commands to the motor controller to move the wheels accordingly.
This paper is organized as follows: Section 2 describes camera placement detail. Description of YOLOv7 deep learning model training and accuracy in Section 3. The discussion is in Section 4, and the conclusion is in Section 5.

2. Camera Installation

The next step was to install eight 1-megapixel monochromic cameras, each having a view angle of 75 degrees. Six cameras were placed in the front, with three on each side, and two cameras were placed in the back for rear view. The cameras were placed in such a way that they were close to achieving ~360 degrees of view. In each array of three cameras placed in the front, one camera faces back at a 15-degree offset from the wheelchair. At the rear of the wheelchair, two cameras were placed 180 degrees from the cameras viewing the front of the wheelchair. Figure 1(left) shows the top view and coverage of the cameras. There were some overlaps, but we tried to reduce blind spots.
There was no camera mounting point in the original wheelchair. Custom parts were made to set the cameras into the wheelchair. Because the Jetson Nano only has two digital camera interface MIPI ports, the eight cameras were connected through two Arducam quadrascopic monochrome camera arrays. Figure 1(right) shows a set of Arducam quadrascopic camera arrays. The right three cameras and right-back camera are connected to one array, and the left three and left-back cameras are connected to another Arducam camera array. Then two Arducams quadrascopic arrays were connected to the Jetson Nano MIPI camera ports. Each camera array needs a 5V DC supply connected to the buck converter output.

3. Obstacle Detection Model

Since the study goal was real-time object detection, we employed YOLOv7 [26], the most recent release of the single-stage object detector YOLO (You Only Look Once [27]) models. As they are small and trainable on a single GPU, YOLO models are better for modeling object detection. Comparing its cohorts, YOLOv7 provides faster (>5 FPS on a V100 GPU) and greater accuracy in predicting the bounding box of the objects (obstacles in our study), thus taking state-of-the-art to new heights. The detailed architecture is provided in [26]. The YOLOv7 network is defined in PyTorch, and training scripts, data loaders, and utility scripts are written in Python. The original model was trained to detect the generic 80 classes in the COCO dataset. For training, we have used custom classes (human, chair, table, pole stairs, trash can, door, notice board with stand, cart, dead end, and wall) and annotated using open-source library Labelimg [28]. Figure 2 shows an example of the annotated image that we collected using our wheelchair.
In our study, we re-trained the YOLOv7 model for 10 object classes. The images were resized to 1280 × 400 before training. While training, a batch size of 16, 100 epochs, and adam optimizer were chosen with multi-scale and hyperparameter evaluation enabled. Figure 3 shows the trained model’s output that it detected surrounding objects.
We have trained the model with our custom annotated dataset for 200 epochs and obtained 92 mAP. Figure 4 shows the confusion matrix of the model. where we can see the model successfully detected objects in most cases. There is some false detection that can be seen from the matrix those are due to less annotated data. We believe adding more annotated data can increase the model’s accuracy.

4. Discussion

This study presents a novel application of vision sensor arrays to obtain an octascopic view of the environment and detect obstacles for autonomous movement; this could be a promising future for wheelchair technology. The design features simplicity and flexibility, and it depends on the commercially available off-the-shelf components. Open-source components and software may enable the community to push this design further as new techniques and approaches become available. Additionally, the end users can install the components to their wheelchairs and convert them to autonomous ones. In our case, we bought the commercially available wheelchair for USD 150, and the total cost required to convert the commercially available wheelchair is USD 1300. In our previous study, we implemented electroencephalogram (EEG) [29] on a wheelchair, and to provide autonomous movement we have implemented the camera-based solution.
Here, we have used eight cameras to obtain ~360 degrees views that will help to detect if any object comes from any direction. If there is no way to go to the front side, then the system can determine the path by analyzing another camera. Another benefit of having multiple cameras is that if one camera is damaged or blocked, the other cameras that cover that area can be used as a substitution as we plan to run all the processing into embedded computers. We have used 1MP cameras.
More work on efficient model development for object detection needs to be completed to ensure low power and robust operation. The object detection model that will run the wheelchair needs to be perfected. This study was limited to obstacle detection, an ROS-based wheelchair operating system is required to introduce faster movement and automatic course correction. As a Jetson-based system is implemented and the battery may run out soon, a wireless charging system is required to be installed.
A large human study is required to test the wheelchair with disabled individuals. In addition to object avoidance, some other implementations can be performed, such as voice commands, brain-computer control, etc., to support the user with low or no mobility of arms to reach the desired destination. Innovations like this mitigate the gap between technology and humanity, arguably the purpose of technology.

5. Conclusions

The design of an autonomous wheelchair implementing octoscopic vision has great potential as it is capable of achieving ~360 degrees of view from its surroundings, other technologies, such as LiDAR, have flaws in that placement of LiDAR in the wheelchair is harder, and they can easily be blocked by the parts of wheelchair. By implementing an octoscopic array of cameras covering some partial view of the wheelchair should not considerably affect its performance as it still has the feed from the rest of the cameras. To increase the accuracy, we need to add more images to train the model. A ROS-based wheelchair operating system is needed to enable faster movement and automatic course correction because this study was restricted to obstacle detection. We will study Wi-Fi-RTT on the wheelchair for localization. Our modified wheelchair is cheaper than the cost of a standard electronic wheelchair. As the project design files are open source, anyone can download and implement them.

Author Contributions

Conceptualization, M.A.B.S.; methodology, M.A.B.S.; software, M.A.B.S. and E.S.-T.; data curation, C.J.; writing—original draft preparation, M.A.B.S.; supervision, M.H.I.; project administration, M.H.I.; funding acquisition, M.H.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Clarkson University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Not applicable.

Conflicts of Interest

The authors declare that no conflict of interest.

References

  1. WHO. World Report on Disability 2011. 2011. Available online: https://www.who.int/teams/noncommunicable-diseases/sensory-functions-disability-and-rehabilitation/world-report-on-disability (accessed on 9 November 2022).
  2. CDC. Disability Impacts All of Us. 2018. Available online: https://www.cdc.gov/ncbddd/disabilityandhealth/infographic-disability-impacts-all.html (accessed on 10 November 2022).
  3. Cindy, X.; Collange, F.; Jurie, F.; Martinet, P. Object tracking with a pan-tilt-zoom camera: Application to car driving assistance. In Proceedings of the 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164), Seoul, Repubic of Korea, 21–26 May 2001; Volume 2, pp. 1653–1658. [Google Scholar] [CrossRef]
  4. Balali, V.; Golparvar-Fard, M. Segmentation and recognition of roadway assets from car-mounted camera video streams using a scalable non-parametric image parsing method. Autom. Constr. 2015, 49, 27–39. [Google Scholar] [CrossRef]
  5. Acharya, S.; Tracey, C.; Rafii, A. System design of time-of-flight range camera for car park assist and backup application. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 24–26 June 2008; pp. 1–6. [Google Scholar] [CrossRef]
  6. Chern, M.Y.; Hou, P.C. The lane recognition and vehicle detection at night for a camera-assisted car on highway. In Proceedings of the 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422), Taipei, Taiwan, 10 November 2003; Volume 2, pp. 2110–2115. [Google Scholar] [CrossRef]
  7. Khan, B.; Hanafi, M.; Mashohor, S. Automated road marking detection system for autonomous car. In Proceedings of the 2015 IEEE Student Conference on Research and Development (SCOReD), Piscataway, NJ, USA, 13–14 December 2015; pp. 398–401. [Google Scholar] [CrossRef]
  8. Sun, T.; Tang, S.; Wang, J.; Zhang, W. A robust lane detection method for autonomous car-like robot. In Proceedings of the 2013 Fourth International Conference on Intelligent Control and Information Processing (ICICIP), Beijing, China, 9–11 June 2013; pp. 373–378. [Google Scholar] [CrossRef]
  9. Okuyama, T.; Gonsalves, T.; Upadhay, J. Autonomous Driving System based on Deep Q Learnig. In Proceedings of the 2018 International Conference on Intelligent Autonomous Systems (ICoIAS), Singapore, 1–3 March 2018; pp. 201–205. [Google Scholar] [CrossRef]
  10. Cho, M.G. A Study on the Obstacle Recognition for Autonomous Driving RC Car Using LiDAR and Thermal Infrared Camera. In Proceedings of the 2019 Eleventh International Conference on Ubiquitous and Future Networks (ICUFN), Zagreb, Croatia, 2–5 July 2019; pp. 544–546. [Google Scholar] [CrossRef]
  11. Sarker, M.A.B.; Sola, P.S.-T.; Jones, A.; Laing, E.; Thomas, E.S.; Imtaz, M.H. Vision Controlled Sensorized Prosthetic Hand. In Proceedings of the Interdisciplinary Conference on Mechanics, Computers and Electrics (ICMECE 2022), Barcelona, Spain, 6–7 October 2022. [Google Scholar]
  12. Caracciolo, M.; Casciotti, O.; Lloyd, C.D.; Sola-Thomas, E.; Weaver, M.; Bielby, K.; Sarker, M.A.B.; Imtiaz, M.H. Autonomous Navigation System from Simultaneous Localization and Mapping. In Proceedings of the 2022 IEEE Microelectronics Design Test Symposium (MDTS), Albany, NY, USA, 23–26 May 2022. [Google Scholar] [CrossRef]
  13. Sola-Thomas, E.; Sarker, M.A.B.; Caracciolo, M.V.; Owen, L.; Christopher, D.; Imtiaz, M.H. Design of an Initial Prototype of the AI Wheelchair. In Proceedings of the 2021 IEEE Microelectronics Design & Test Symposium (MDTS), Albany, NY, USA, 18–21 May 2021. [Google Scholar] [CrossRef]
  14. Horn, O.; Kreutner, M. Smart wheelchair perception using odometry, ultrasound sensors, and camera. Robotica 2009, 27, 303–310. [Google Scholar] [CrossRef]
  15. Nguyen, J.S.; Su, S.W.; Nguyen, H.T. Spherical vision cameras in a semi-autonomous wheelchair system. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, Honolulu, HI, USA, 15–18 September 2010; pp. 4064–4067. [Google Scholar] [CrossRef]
  16. Nguyen, T.H.; Nguyen, J.S.; Pham, D.M.; Nguyen, H.T. Real-Time Obstacle Detection for an Autonomous Wheelchair Using Stereoscopic Cameras. In Proceedings of the 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007; pp. 4775–4778. [Google Scholar] [CrossRef]
  17. Nguyen, J.S.; Su, S.W.; Nguyen, H.T. Experimental study on a smart wheelchair system using a combination of stereoscopic and spherical vision. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 4597–4600. [Google Scholar] [CrossRef]
  18. Nguyen, J.S.; Nguyen, T.N.; Tran, Y.; Su, S.W.; Craig, A.; Nguyen, H.T. Real-time performance of a hands-free semi-autonomous wheelchair system using a combination of stereoscopic and spherical vision. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 3069–3072. [Google Scholar] [CrossRef]
  19. Şahin, H.; Kavsaoğlu, A.R. Autonomously Controlled Intelligent Wheelchair System for Indoor Areas. In Proceedings of the 2021 3rd International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), Ankara, Turkey, 11–13 June 2021; pp. 1–6. [Google Scholar] [CrossRef]
  20. Sonata, I.; Heryadi, Y.; Lukas, L.; Wibowo, A. Autonomous car using CNN deep learning algorithm. J. Phys. Conf. Ser. 2021, 1869, 012071. [Google Scholar] [CrossRef]
  21. Thammachantuek, I.; Kosolsomnbat, S.; Ketcham, M. Comparison of Machine Learning Algorithm’s Performance Based on Decision making in Autonomous Car. In Proceedings of the 2018 International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP), Pattaya, Thailand, 15–17 November 2018; pp. 1–6. [Google Scholar] [CrossRef]
  22. Ahmad, I.; Pothuganti, K. Design & implementation of real time autonomous car by using image processing & IoT. In Proceedings of the 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT), Tirunelveli, India, 20–22 August 2020; pp. 107–113. [Google Scholar] [CrossRef]
  23. Nurhandayani, K.; Purwanto, D.; Mardiyanto, R. Development of Obstacle Detection Based on Region Convolutional Neural Network for Autonomous Car. In Proceedings of the 2021 International Seminar on Intelligent Technology and Its Applications (ISITIA), Online, 21–22 July 2021; pp. 354–358. [Google Scholar] [CrossRef]
  24. Li, S.; Li, Y.; Li, Y.; Li, M.; Xu, X. YOLO-FIRI: Improved YOLOv5 for Infrared Image Object Detection. IEEE Access 2021, 9, 141861–141875. [Google Scholar] [CrossRef]
  25. Sola-Thomas, E. Smart Throne. 2021. Available online: https://osf.io/n9af7/ (accessed on 9 November 2022).
  26. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  27. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  28. heartexlabs. labelImg. 2022. Available online: https://github.com/heartexlabs/labelImg (accessed on 10 November 2022).
  29. Stoyell, G.; Seybolt, A.; Griebel, T.; Sood, S.; Sarker, M.A.B.; Khondker, A.; Imtiaz, M.H. Implementation of a Mind-Controlled Wheelchair. In Proceedings of the St. Lawrence Section Annual Conference, Syracuse, NY, USA, 25–26 March 2022. [Google Scholar]
Figure 1. (left) The illustration of the coverage of the wheelchair camera. (right) Arducam camera array.
Figure 1. (left) The illustration of the coverage of the wheelchair camera. (right) Arducam camera array.
Engproc 31 00060 g001
Figure 2. An example of an annotated image.
Figure 2. An example of an annotated image.
Engproc 31 00060 g002
Figure 3. Object detection model output.
Figure 3. Object detection model output.
Engproc 31 00060 g003
Figure 4. Confusion matrix.
Figure 4. Confusion matrix.
Engproc 31 00060 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sarker, M.A.B.; Sola-Thomas, E.; Jamieson, C.; Imtiaz, M.H. Autonomous Movement of Wheelchair by Cameras and YOLOv7. Eng. Proc. 2023, 31, 60. https://doi.org/10.3390/ASEC2022-13834

AMA Style

Sarker MAB, Sola-Thomas E, Jamieson C, Imtiaz MH. Autonomous Movement of Wheelchair by Cameras and YOLOv7. Engineering Proceedings. 2023; 31(1):60. https://doi.org/10.3390/ASEC2022-13834

Chicago/Turabian Style

Sarker, Md Abdul Baset, Ernesto Sola-Thomas, Collin Jamieson, and Masudul H. Imtiaz. 2023. "Autonomous Movement of Wheelchair by Cameras and YOLOv7" Engineering Proceedings 31, no. 1: 60. https://doi.org/10.3390/ASEC2022-13834

Article Metrics

Back to TopTop