Next Article in Journal
Impact of Rice Expansion on Traditional Wetland Management in the Tropical Highlands of Ethiopia
Previous Article in Journal
Studying the Effect of Straw Returning on the Interspecific Symbiosis of Soil Microbes Based on Carbon Source Utilization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of a Personal Mobility System with Autonomous Driving for Agricultural Work by the Physically Challenged and the Vulnerable

1
Department of Mechanical Engineering, Inha University, 100 Inharo, Michuholgu, Incheon 22212, Korea
2
Department of Construction Machinery Engineering, Inha University, 100 Inharo, Michuholgu, Incheon 22212, Korea
*
Author to whom correspondence should be addressed.
Agriculture 2022, 12(7), 1054; https://doi.org/10.3390/agriculture12071054
Submission received: 13 June 2022 / Revised: 5 July 2022 / Accepted: 15 July 2022 / Published: 19 July 2022
(This article belongs to the Section Agricultural Technology)

Abstract

:
Many issues have recently arisen as a result of the aging population and dwindling labor force. To solve these problems, including reduced convenience and productivity, research on the mobility of agricultural machinery and logistics involving autonomous driving is being conducted. However, research on improving individual mobility is scarce. Therefore, an autonomous driving personal mobility system is developed in this study for the physically challenged and other vulnerable groups with mobility difficulties. The personal mobility system can help the physically challenged and the vulnerable to overcome their limited mobility and perform agricultural work. An additional feature is a standing function that allows the performance of tasks too difficult to do while seated. In addition, the autonomous driving function guides the personal mobility system along the work route and stops near agricultural crops. In this study, the design and structural stability of the personal mobility system were analyzed through CAD modeling and structural analysis, with LiDAR applied for autonomous driving and machine learning for agricultural work. This personal mobility system is expected to improve mobility for the physically challenged and the vulnerable, allowing them to play a role and participate more fully in society.

1. Introduction

As the world’s population ages, the number of senior citizens grows, and the super-aged society, including the vulnerable and the physically challenged, deepens. Around 901 million people aged 60 and up will live on the planet by 2022, accounting for 12.3% of the global population. By 2030, this will increase to 1.4 billion or 16.5%, and by 2050, it will increase to 2.1 billion, or 21.5% of the global population [1]. Vulnerable groups such as the elderly and the physically challenged have difficulty with indoor and outdoor activities because moving and working independently is inconvenient compared to the general population. They are also vulnerable to both severe and minor mishaps. As a result, advanced countries have been working on initiatives that use assistive technology in various sectors, focusing on personal mobility.
In the agriculture sector, considerable research is being conducted on the work autonomy and autonomous driving of agricultural robots. These robots aim to solve social problems such as the decline in agricultural productivity owing to population aging and the corresponding decrease in human power. In addition, the development of smart mobility systems is crucial to expand the convenience and range of activities of vulnerable groups. Further, the development and demonstration of autonomous wheelchairs are being promoted to improve the convenience of movement for the vulnerable population. In highly functional electric wheelchairs, a controller that implements intelligent algorithms for learning and judgment is applied instead of a simple feedback controller that uses an existing joystick. Personal mobility is taken for granted by the public, but a personal mobility system is a necessary mode of transportation for the elderly and physically challenged, who are particularly vulnerable. Increased mobility is closely related to improvement in the overall quality of life and affects simple transportation and work and health [2].
Among the various personal mobility systems, the standing wheelchair has medical advantages, such as helping blood circulation and activating the digestive function by lifting the patient who is always sitting into a standing position. A standing position also allows communicating with others at eye level [3]. In addition, the standing wheelchair allows someone to perform tasks that are difficult to do while sitting [4].
Research on autonomous vehicles is constantly underway. Especially since the Fourth Industrial Revolution, the importance of autonomous driving systems has become even greater. Autonomous driving in automobiles is mainly enjoyed by people who are not physically challenged, but the need for technological developments for those who are physically challenged is increasing [5]. Many studies are underway that are focused on autonomous driving for agricultural robots to solve problems in agriculture. In addition, there is development and demonstration of autonomous wheelchairs to improve movement for the vulnerable. For highly functional electric wheelchairs, a controller that implements intelligent algorithms for learning and judgment is applied instead of a simple feedback controller that uses a joystick. Processor performance has greatly improved, with video-processing-based algorithms and the like easily realized [6].
Prassler et al. introduced a robotic wheelchair that provides increased mobility, autonomy, and independence for people with physical restrictions. It has a wide-area navigation control structure to support autonomous driving in a wide and crowded area. Its hierarchical driving structure consists of a basic control level, a tactical level, and a strategic level [7].
In order to drive autonomously to a given destination while avoiding obstacles, an efficient route must be generated. Therefore, it is necessary to generate the shortest route and consider the number of calculations required for route generation, passenger safety, and ride quality. Existing global path generation algorithms, such as Dijkstra’s and A *, primarily focus on finding the shortest path. However, Dijkstra’s algorithm and A * are slower to calculate, occupy more memory, and are inefficient, compared to other algorithms.
Furthermore, with the A * algorithm, a route cannot be generated to deal with sudden changes, so it is not suitable for high-speed driving. The dynamic window approach is a regional route generation and obstacle avoidance algorithm that can select the most appropriate speed. It produces a smoother path than other methods but has a disadvantage in that it cannot easily reach a locally optimal solution. To make up for these shortcomings, and a mix of algorithms is often used, rather than writing just one [8].
In this study on the personal mobility system, a standing function was added to an electric wheelchair to develop a personal mobility system for the physically challenged and the vulnerable. Its structural stability was also analyzed through structural analysis. Convenience functions such as standing and rotating were applied. In addition, autonomous driving was applied to smart mobility systems such as wheelchairs for the convenience of the agricultural population, and deep learning technology using riders and vision cameras was utilized for route tracking and image processing. This study followed a campus map for the experiment’s progress and used tomatoes as the fruit tracked to recreate a rural environment.
This research makes it possible to improve the mobility of the physically challenged and vulnerable, who find it inconvenient to carry out agricultural work. It is also expected that participating in agricultural work more easily can help them play a meaningful role as a member of society.

2. Design and Analysis

2.1. Design

The personal mobility system developed in this paper was designed with reference to the existing electric wheelchair [4]. A front-wheel-drive system was used. The rear wheels are a caster system, left and right front wheels are firewood, and space for the motor is secured through a bevel gear. Steering is possible by using a different RPM for each motor, and the standing function is implemented with an electric actuator. The kinematic structure for the standing function is composed of a four-bar link so the angle between the backrest and the seat can be increased. CAD modeling of the personal mobility system is shown in Figure 1. Acceleration and constant velocity torque were calculated, and an appropriate motor was selected. To calculate the torque at a maximum speed of 5 km / h , the weight ( m ), diameter of the drive wheel ( D ), RPM , and acceleration time ( t ) are required (summarized in Table 1). Weight is the sum of the weight of the personal mobility system (approximately 50 kg ) and the maximum user weight (100 kg ). For the load moment of inertia, Equation (1) is used, and the acceleration and constant velocity torques can be obtained using Equations (2) and (3), respectively. Therefore, the required motor torque is the sum of the acceleration and constant velocity torques and can be obtained at 308.454 kg · cm by applying a safety factor of 1.5. As a result, a motor with 3000   RPM , 11.2 kg · cm , and 1/30 reducer was selected.
J = m · D 2 / 8
T a = J · 2 π · R P M 60 · g · t
T m = μ · m · D / 4

2.2. Analysis

After designing the personal mobility system, structural analysis was performed to examine the structural stability with someone aboard. In order to shorten the analysis time, parts not required for structural analysis were excluded but are still shown in Figure 2.
The materials for the personal mobility system developed in this study, detailed in Table 2, were aluminum 5052 and MC nylon. MC nylon was used for the seat, and aluminum 5052 was used for the rest of the parts. According to the analysis results, aluminum 5052 was more suitable than MC nylon and was used. For boundary conditions, a fixed support was applied to the front wheel shaft and the upper surface of the rear wheel casters, with 1000N applied to the seat.
Figure 3 shows the boundary condition, Figure 3a presents the condition for force, and Figure 3b shows the fixed condition. The analysis results for seating and standing modes are shown in Figure 4a,b, respectively. From checking the maximum equivalent stress, the seating mode is 29.584 MPa in the four-bar link, and standing mode is 75.875 MPa for the footrest. This is lower than the 193 MPa yield strength of aluminum 5052 and satisfies the safety factors of 6.52 and 2.54. The deformation and stress values are summarized in Table 3.

3. System Overview

3.1. Hardware

The motor and battery were attached to the lower end of the prototype so it would not be difficult to move. The configuration and wiring attached to the lower end of the prototype is shown in Figure 5. The motor, motor driver, and battery are mounted at the bottom to drive the prototype. The attachment of the lower end is shown in Figure 6. The two motors are attached to the battery and the embedded board via the motor driver. A driver for control was installed according to specifications, and a lithium-ion battery supplies power to the two motors. A cushion-shaped chair was attached to give a sense of stability to the prototype, and a device for autonomous driving was added, as shown in Figure 7 and described in Section 3.2. A prototype based on the design was prepared for experimental verification and is shown in Figure 7. This study calculated the acceleration and constant velocity torque for the drive motor and used a BLDC motor equipped with a reducer, as described in Section 2.1.

3.2. Software

The software used for driving was the open-source ROS (Robot Operating System) melodic version of the Ubuntu 18.04 operating system. The MD motor drivers and packages were used to drive the motor with ROS. First, the ROSCORE instruction was used to execute the master such that connections and message communication between nodes could be performed. The Velodyne LiDAR package was used to receive LiDAR point cloud values. The utilized CPU was an Intel Core i7-7700HQ, 2.8 GHz, with the NVIDIA GTX1050Ti GPU. The dataset used and learned in YOLO v3 included approximately 550 screened tomato images. A modeling-based prototype, shown in Figure 6, was designed for the experiment. The sensors for autonomous driving were attached to the prototype, and their heights and positions are depicted in Figure 7. As shown in Figure 7, the LiDAR was mounted at a height of 1700 mm, and the vision camera was mounted on the armrest for storage. An Odyssey board was attached to the seat. Simultaneous localization and mapping (SLAM) is a method that generally uses various sensors to create a map of the surrounding environment to estimate one’s position [9]. The master is run using roscore for node-to-node connections and communications. To receive LiDAR point cloud values, this study used the Velodyne LiDAR package to receive topics and used the advanced implementation of Lidar Odometry and Mapping (A-LOAM) to recognize the surrounding environment. LOAM is used to analyze the surroundings simultaneously with position estimation [10].

3.3. Deep-Learning Algorithm for the Autonomous System

The RRT and A* algorithms for autonomous driving are described in this section. This study determines the shortest path between the beginning and goal places without colliding with any obstacles in order to discover and design the best route. In this investigation, an algorithm was chosen and used. The following is a quick summary of the two algorithms. The A* algorithm first calculates the shortest path from the initial node (beginning point) to the destination node (target point). The operation method of the A* algorithm used in the experiment is shown in Figure 8. The RRT method is a sampling-based path-planning technique that creates points at random locations throughout the state space and utilizes this as a propensity to expand the tree quickly from the beginning point to generate a path that leads to the destination.
The A-LOAM algorithm using a LiDAR was created. At this time, if there is an obstacle on the route over which the personal mobility system travels, a route must be generated around the obstacle. The RRT algorithm and the A* algorithm was compared, and the final algorithm to be used was selected. The A* algorithm is the least costly and looks at the nodes adjacent to a particular node in the state space and goes from the starting point to the target point. At each iteration of its main loop, A* needs to determine which of its paths to extend. It does so based on the cost of the path and an estimate of the cost required to extend the path all the way to the goal. Specifically, A* selects the path that minimizes.
f ( n ) = g ( n ) + h ( n )
h ( n ) is a heuristic function that estimates the cost of the cheapest path from n to the objective, where n is the next node on the path, g ( n ) is the cost of the path from the start node to n , and g ( n ) is the cost of the path from the start node to n . The RRT method is far faster than the A* algorithm; nevertheless, the RRT algorithm’s computed path is significantly longer than the A* algorithm, which is a significant drawback. The effects of such distance measuring technologies are the subject of ongoing research. Velodyne LiDAR data was translated into 2D laser data using an algorithm to create a map, and trials were undertaken on straight roadways to test the reaction speed. The RRT [11] method first locates a configuration space, then enters an initial configuration state ( q i n i t ) and a goal configuration state ( q g o a l ), selecting any feasible q from q init and connecting them with edges. The technique is then continued until the q target is attained, at which point the best path is created. The operation method of the RRT algorithm used in the experiment is shown in Figure 9 [12]. An experiment was conducted to evaluate the performance of the prototype in an environment with trees. Ripe tomatoes and unprocessed tomatoes were attached to trees to construct an experimental environment. To distinguish tomatoes, the vision camera uses Intel’s RealSense D435 (a 3D camera). After driving the prototype, YOLO v3-Tiny used the learned weight file to detect harvestable tomatoes in an image, distinguishing between harvestable (red and ripe) tomatoes and unripen (green) tomatoes. This system extracted distance information from the center point of the tomato’s boundary box, as detected by the 3D camera. Object recognition and detection was performed using deep learning. The deep learning algorithm used is YOLO v3, plus a deep neural network learning layer and multilayer perceptron (MLP) consisting of an output layer structure and multiple hidden layers [13,14,15]. YOLO v3 has been applied to various types of algorithms such as the recurrent neural network, although YOLO v4 and YOLO v5 were recently released [16,17]. The structure of YOLO is like a region-based convolutional neural network (R-CNN) in that it consists of bounding box classification, positioning, deduplication, and post-processing, but it defines a series of object detection procedures as one regression problem [18,19]. Although these one-stage-detection systems reduce the accuracy, the detection speed (in frames per second) is greatly increased so that the object can be recognized in real time at a rapid detection speed. This has evoked a large wave of cognitive deep learning. According to the trends in artificial-intelligence-related research for the visually impaired, YOLO showed the highest number of uses at 2.7% as an application technology for domestic research design and development implementation, and YOLO v2 showed the accuracy and speed of an R-CNN [20]. The improved Fast R-CNN and the Faster R-CNN showed higher accuracy and speed [21,22]. From the third version of YOLO used in various projects, this study selected and used the Tiny version, which has a slightly lower accuracy but a relatively small memory capacity and a high detection speed.
The convolutional network process in Figure 10 populates weight data with the image. Therefore, a pre-trained weight file is needed for the target object. In YOLO training, the larger the number of images in the dataset for each class (the target objects to be detected), and the more diverse the angle and lighting environment, the more accurate the detection will be in various actual environments. In this study, 550 images out of about 900 images from Kaggle’s free tomato dataset, which is a machine learning platform, were selected. To use the selected dataset for training, a bounding box is drawn around the target object to be detected from each image.

3.4. Experiment

In this study, an algorithm that constructed a map using LiDAR and reached the target point was used for experimentation. First, a map of the location where the occupant wants to drive was created through mapping. The Velodyne data were changed to a two-dimensional (2D) laser for generating the map. To apply each algorithm and select the optimum algorithm, in the experiment, obstacles were set using safety cones on the campus and the mean of the values for traveling 10 times and reaction speed at each position were obtained. First, the state of point cloud recognition before and after filtering by the rider appears, as shown in Figure 9. Point clouds and maps can be viewed through RVIZ, a visualization tool for ROS, and obstacles, trees, and human shapes can be observed through the point cloud. In addition, to derive these data values, and the values of the results of the experiments with the algorithm applied are listed in Table 4.
Table 4. Algorithm reaction speeds during driving (①, ②, ③). The route environment of each symbol is shown in Figure 11.
Table 4. Algorithm reaction speeds during driving (①, ②, ③). The route environment of each symbol is shown in Figure 11.
Position ①Position ②Position ③
A*1.368 (s)0.478 (s)1.962 (s)
RRT1.498 (s)0.871 (s)2.689 (s)
Figure 12 shows the tomato recognition in the experimental environment. Table 5 lists the recognition results of the final experiment for the tomato positions. The system did not recognize unripe tomatoes, and the recognition rate of the last tomatoes was approximately 40%. It was confirmed that other crops can be easily processed for recognition. For this, the system was configured to detect harvestable fruit. The results of running in an ROS environment using the tomato weight file learned for the experiment indicated that when the harvestable tomato distance is 600 mm (depth 800 mm), topic ‘T’ is published to the subscriber node, and when 1 m (depth 800 mm), the message is not published.

4. Conclusions

In this study, a personal mobility system was developed for the vulnerable and physically challenged, and the stability of the system was examined through structural analysis. The results revealed that the safety factors of 6.52 and 2.54 were met, and that stability was not a concern even when a user weighing up to 100 kg was onboard. The RRT algorithm is much quicker than the A* method, although the RRT algorithm’s computed path is longer than the A* algorithm. Furthermore, because the distance measuring method utilized affects these algorithms, they have been regularly explored (e.g., Manhattan and Euclidean distances). In YOLO, 550 datasets are usually adequate for object recognition, rather than hundreds of photos for a single class. The system did not recognize unripe tomatoes, and the recognition rate of the last tomatoes was approximately 40%. It was confirmed that other crops can be easily processed for recognition. For this, the system was configured to detect harvestable fruit. As a result, depending on the angle, illumination, and type of item, a poor identification rate was seen in repeated testing. Supplementary training utilizing high-quality data with excellent picture quality in a real agricultural context can enhance the identification rate. The use of machine learning techniques by increasing the dataset should be included in this thesis, and more experiments with route-running experiment data are necessary. In this study, Velodyne LiDAR is primarily utilized for route driving and obstacle avoidance. Sensor fusion is largely employed in self-driving and self-working vehicles. However, in the case of a vision camera, an inaccuracy in the identification rate arises owing to the light and angle during sensor fusion. Future research must use Velodyne LiDAR and make use of the variation in reflectance between the surfaces of fruits and leaves to compensate for the identification rate issue. Using Velodyne LiDAR in Sweden, an algorithm based on apple reflectivity was created. It is conceivable to operate a personal mobility system in an agricultural area with one sensor and no sensor fusion if a fruit recognition algorithm employing Velodyne LiDAR is applied. Therefore, in future research the aim is to develop an algorithm that can apply autonomous driving and autonomous work using Velodyne LiDAR. Algorithm development using one sensor is expected to contribute to the development of more convenient and efficient smart mobility systems.

Author Contributions

Conceptualization, E.K. and K.H.; methodology, E.K.; software, E.K.; validation, E.K. and K.H.; formal analysis, E.K.; investigation, E.K.; resources, K.H.; data curation, E.K.; writing—original draft preparation, E.K.; writing—review and editing, C.-H.L.; visualization, E.K.; supervision, C.-H.L.; project administration, C.-H.L.; funding acquisition, C.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was carried out with the support of “Cooperative Research Program for Agriculture Science and Technology Development (Project No. PJ015312022022)” Rural Development Administration, Republic of Korea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Scobie, J. Global AgeWatch Index 2015 Insight Report; HelpAge International: London, UK, 2015; pp. 5–6. [Google Scholar]
  2. Kim, A.J.; Kweon, O.S. The Effect of User Interface of Power Wheelchair for People with Spinal Cord Injuries on Usability. Des. Work. 2018, 1, 12–22. [Google Scholar]
  3. Song, C.Y.; Yoon, H.J.; Lee, C. Development of Standing and Gait Assistive Wheelchair. J. Manuf. Eng. Technol. 2013, 22, 587–592. [Google Scholar] [CrossRef]
  4. Han, K.H.; Lee, C.-H. A Study on the Development of Personal Mobility for the Vulnerable Group. J. Drive Control. 2021, 18, 35–42. [Google Scholar]
  5. Jung, M.W.; Lee, S.H. Object Perception Algorithm based on LiDAR for Autonomous Vehicle. Korean Soc. Electron. Eng. Conf. 2017, 37, 639–642. [Google Scholar]
  6. Choi, J.H.; Kim, M.H.; Choi, B.J. Development of a platform based on image signal processing for autonomous driving of electric wheelchairs. J. Korean Inst. Intell. Syst. 2018, 28, 231–236. [Google Scholar]
  7. Prassler, E.; Scholz, J.; Fiorini, P. A robotics wheelchair for crowded public environments. IEEE Robot. Autom. Mag. 2001, 8, 38–45. [Google Scholar] [CrossRef]
  8. Lu, K.; Zhang, Y.; Li, H. Research Status and Development Trend of Path Planning Algorithm for Unmanned Vehicles. IOP Conf. Ser. J. Phys. Conf. Ser. 2019, 1213, 032006. [Google Scholar] [CrossRef]
  9. Lee, Y.M.; Seo, Y.D. Vision-Based SLAM in Augmented/Mixed Reality. J. Korea Multimed. Soc. 2009, 13, 12–20. [Google Scholar]
  10. Zhang, J.; Singh, S. LOAM: Lidar Odometry and Mapping in Real-time (or Low-drift and Real-time Lidar Odometry and Mapping). Auton. Robot. 2017, 41, 401–416. [Google Scholar] [CrossRef]
  11. Lavalle, S.M. Rapidly-exploring random trees: A new tool for path planning. Comput. Sci. Dept. Oct 1998, 98-11. [Google Scholar]
  12. Tak, H.T.; Park, C.G.; Lee, S.C. Improvement of RRT*-Smart Algorithm for Optimal Path Planning and Application of the Algorithm in 2 & 3-Dimension Environment. J. Korean Soc. Aviat. Aeronaut. 2019, 27, 1–8. [Google Scholar]
  13. Mikolov, T.; Kombrink, S.; Burget, L.; Cernocky, J.; Khudanpur, S. Extensions of Recurrent Neural Network Language Model. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5528–5531. [Google Scholar]
  14. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Bochkovskiy, A.; Wang, C.-Y.; Mark Liao, H.-Y. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision, and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  17. Bea, S.Y. Research Trends on Related to Artificial Intelligence for the Visually Impaired: Focused on Domestic and Foreign Research in 1993-2020. J. Korea Contents Assoc. 2020, 20, 688–701. [Google Scholar]
  18. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  19. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  20. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. arXiv 2015, arXiv:1506.01497. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Lee, Y.H.; Kim, Y. Comparison of CNN and YOLO for Object Detection. J. Semicond. Disp. Technol. 2020, 19, 85–92. [Google Scholar]
  22. Quigley, M.; Conley, K.; Faust, J.; Foote, T.; Leibs, J.; Bergr, E.; Wheeler, R.; Ng, A. ROS: An open-source Robot Operating System. ICRA Workshop Open-Source Softw. 2009, 3, 5. [Google Scholar]
Figure 1. CAD modeling of the personal mobility system (a) driving and (b) standing modes.
Figure 1. CAD modeling of the personal mobility system (a) driving and (b) standing modes.
Agriculture 12 01054 g001
Figure 2. Modeling of the personal mobility system for structural analysis: (a) seating mode and (b) standing mode.
Figure 2. Modeling of the personal mobility system for structural analysis: (a) seating mode and (b) standing mode.
Agriculture 12 01054 g002
Figure 3. Boundary conditions: (a) force (b) fixed.
Figure 3. Boundary conditions: (a) force (b) fixed.
Agriculture 12 01054 g003
Figure 4. Structural analysis results: (a) seating and (b) standing modes.
Figure 4. Structural analysis results: (a) seating and (b) standing modes.
Agriculture 12 01054 g004
Figure 5. Configuration and wiring attached to the lower end of the prototype.
Figure 5. Configuration and wiring attached to the lower end of the prototype.
Agriculture 12 01054 g005
Figure 6. Prototype of the personal mobility system: (a) seating and (b) standing modes.
Figure 6. Prototype of the personal mobility system: (a) seating and (b) standing modes.
Agriculture 12 01054 g006
Figure 7. Prototype showing the mounting positions of the camera and LiDAR sensor. Distance data in mm .
Figure 7. Prototype showing the mounting positions of the camera and LiDAR sensor. Distance data in mm .
Agriculture 12 01054 g007
Figure 8. Operation method of the A* algorithm.
Figure 8. Operation method of the A* algorithm.
Agriculture 12 01054 g008
Figure 9. Operation method of the RRT algorithm.
Figure 9. Operation method of the RRT algorithm.
Agriculture 12 01054 g009
Figure 10. Flowchart of the crop recognition system.
Figure 10. Flowchart of the crop recognition system.
Agriculture 12 01054 g010
Figure 11. Establishment of the autonomous driving test environment: (a) experimental environment; (b) 3D map collected point cloud; (c) experimental environment for tomato recognition.
Figure 11. Establishment of the autonomous driving test environment: (a) experimental environment; (b) 3D map collected point cloud; (c) experimental environment for tomato recognition.
Agriculture 12 01054 g011
Figure 12. Tomato recognition.
Figure 12. Tomato recognition.
Agriculture 12 01054 g012
Table 1. Parameters required to calculate torque at a maximum speed of 5 km / h .
Table 1. Parameters required to calculate torque at a maximum speed of 5 km / h .
ParameterValue
Weight
(Sum of personal mobility system and maximum user weight)
150 kg
Diameter of the driving wheel28.21 cm
M a x i m u m   s p e e d 94.03 rpm
Acceleration time1.5 s
Table 2. Properties of the materials.
Table 2. Properties of the materials.
Aluminum 5052MC Nylon
Density 2680   kg / m 3 1160   kg / m 3
Young’s modulus 70.3   GPa 3.432   GPa
Poisson’s ratio0.330.4
Yield strength 193   MPa 96   MPa
Table 3. Deformation and stress values from structural analysis.
Table 3. Deformation and stress values from structural analysis.
Seating ModeStanding Mode
Deformation 14.923   mm 3.413   mm
Stress 29.584   MPa 75.875   MPa
Safety factor6.522.54
Table 5. The structural analysis results; T: tomatoes located in Figure 8c.
Table 5. The structural analysis results; T: tomatoes located in Figure 8c.
Recognition
(O)
T1T2T3T4T5T6T7T8
TEST 1O--O--OO
TEST 2O--O---O
TEST 3O--O---O
TEST 4O--O---O
TEST 5O--O--OO
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ko, E.; Han, K.; Lee, C.-H. Development of a Personal Mobility System with Autonomous Driving for Agricultural Work by the Physically Challenged and the Vulnerable. Agriculture 2022, 12, 1054. https://doi.org/10.3390/agriculture12071054

AMA Style

Ko E, Han K, Lee C-H. Development of a Personal Mobility System with Autonomous Driving for Agricultural Work by the Physically Challenged and the Vulnerable. Agriculture. 2022; 12(7):1054. https://doi.org/10.3390/agriculture12071054

Chicago/Turabian Style

Ko, EunByul, KwangHo Han, and Chul-Hee Lee. 2022. "Development of a Personal Mobility System with Autonomous Driving for Agricultural Work by the Physically Challenged and the Vulnerable" Agriculture 12, no. 7: 1054. https://doi.org/10.3390/agriculture12071054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop