Next Article in Journal
Taxation of Fiat Money Using Dynamic Control
Previous Article in Journal
Knowledge Management as a Domain, System Dynamics as a Methodology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Floor Path Model for Route Tracking of Autonomous Vehicles

Department of Industrial Engineering, Sakarya University, Sakarya 54050, Turkey
*
Author to whom correspondence should be addressed.
Systems 2022, 10(3), 83; https://doi.org/10.3390/systems10030083
Submission received: 13 February 2022 / Revised: 5 May 2022 / Accepted: 8 June 2022 / Published: 15 June 2022
(This article belongs to the Topic Methods for Data Labelling for Intelligent Systems)

Abstract

:
Real-time route tracking is an important research topic for autonomous vehicles used in industrial facilities. Traditional methods such as copper line tracking on the ground, wireless guidance systems, and laser systems are still used in route tracking. In this study, a deep-learning-based floor path model for route tracking of autonomous vehicles is proposed. A deep-learning floor path model and algorithm have been developed for highly accurate route tracking, which avoids collisions of vehicles and follows the shortest route to reach the destination. The floor path model consists of markers. Routes in the floor path model are created by using these markers. The floor path model is transmitted to autonomous vehicles as a vector by a central server. The server dispatches the target marker address to the vehicle to move. The vehicle calculates all possible routes to this address and chooses the shortest one. Marker images on the selected route are processed using image processing and classified with a pre-trained deep-CNN model. If the classified image and the image on the selected route are the same, the vehicle proceeds toward its destination. While the vehicle moves on the route, it sends the last classified marker to the server. Other autonomous vehicles use this marker to determine the location of this vehicle. Other vehicles on the route wait to avoid a collision. As a result of the experimental studies we have carried out, the route tracking of the vehicles has been successfully achieved.

1. Introduction

Autonomous vehicles, which are programmed to move materials from one department to another without human intervention, are seen as part of smart manufacturing and reduce the workload. Determining route tracking strategies and real-time route tracking is an important research topic for autonomous vehicles used in industrial production areas [1,2]. It is important to follow a flexible route suitable for the variable obstacle structure depending on the working environment. The static and prior knowledge of the working environment or the dynamic and unknown state are the deciding parameters in the trajectory planning characteristic of the autonomous vehicle. Static trajectory planners are recommended when there is a fixed route allocated to the vehicle and this route is not interrupted by any obstacles. Dynamic route planners are recommended for environments with no map information or constantly changing obstacle structures. Traditional methods such as copper line tracking on the ground, wireless guidance systems, and laser systems are used in route tracking.
The route design is the first and most important step in developing an autonomous vehicle system. This process is vital in balancing workload in the production areas and reducing traffic density. The route consists of guideways used to move loads of vehicles from one area to another [3], either unidirectionally or bidirectionally [4].
Deep learning is a method that can make decisions using data characteristics and reduce dependence on experts [5]. CNN is a widely used type of deep-learning method. It is a multi-layered neural network model consisting of convolutional, rectified linear units (ReLUs), pooling, and fully connected layers [6]. Camera images can be analyzed quickly by using deep-learning methods. The routes of vehicles used in production environments can be designed flexibly. The images of these designed routes can be easily obtained with vehicle cameras and determined using deep-learning models. Thus, vehicles can analyze routes and go to their destination without human intervention. Smolyansky et al. present a micro aerial vehicle system for outdoor environments such as forests based on a deep neural network called TrailNet [7]. Wang et al. obtained road signs using a camera mounted onto an AGV. They applied these signals as input to a neural network they developed and performed the visual-based path tracking of the AGV [8].
In the present study, an image processing and a deep-learning-based indoor positioning system are proposed to solve the real-time route tracking and collision problems of autonomous vehicles used in indoor production environments [9,10]. In the proposed system, a floor path model consisting of markers was developed for the ground. The markers used in the floor path model are formed from a combination of letters and numbers. Routes are created using these markers. First, the floor path model is sent as a vector to autonomous vehicles by a server. The target marker address is required for the vehicle to move. The routes to the destination address and the distances of these routes are calculated by the vehicle. The closest route to the destination is selected from among the calculated routes. If there is no other vehicle on the selected route within four marker route distances from the vehicle, the vehicle model is activated. If there is another vehicle on the selected route, the same checks are made for the other route closer to the target. If no empty route is found, the vehicle model is suspended. While the autonomous vehicle moves on the route, the images obtained from the camera mounted on the autonomous vehicle model (AVM) are processed with pre-processing image methods. These images are given as input to the CNN model trained using the transfer learning method. The network output is compared with the corresponding element of the target route vector to ensure that the autonomous vehicle model follows the route. If the route vector information of the network output is equal, the address information from other vehicles is checked. If the vehicle whose address information is received is four markers away from the relevant vehicle, the relevant vehicle waits; otherwise, it will continue. These processes continue until the vehicle reaches its destination. The floor path model and algorithm used the created routes, allowing the autonomous vehicle to reach its destination by following the most appropriate route without colliding with another vehicle.
The main contributions of our work can be summarized as follows:
  • We design a deep-learning-based floor path model for route tracking of autonomous vehicles.
  • It is a collision-free indoor positioning system for autonomous route tracking vehicles, based on the CNN deep-learning model.
  • A new floor road model and algorithm based on indoor positioning has been developed for autonomous vehicles.
  • The vehicles calculate all the routes to the target location in the floor path model and choose the shortest path and follow the route with high accuracy.
  • The experiments we have carried out confirmed the validity of our model.
The rest of this paper is structured as follows. Section 2 summarizes the related works. In Section 3, the subtitles of convolutional neural network, preparing the data sets, and obtaining the region of interest in the frame were presented under the materials and methods. Section 4 describes the proposed the deep-learning-based floor path model for route tracking. Section 5 presents experimental results. Section 6 summarizes results and discuss the results of the study. Finally, Section 7 provides a conclusion and proposes potential future works.

2. Related Work

In recent years, studies on the development of autonomous vehicles have been performed. Different problems are encountered in autonomous vehicles used in production environments. One of these problems is real-time route tracking. Another problem is that the vehicle moving on the route can collide with another vehicle. Some of the studies on autonomous vehicles have focused on route-tracking strategies. These strategies are important for automated guided vehicles (AGVs) used in production areas. Collision avoidance is a critical problem in the motion planning and control of AGVs [11,12]. A collision can occur when AGVs move in the same direction on the same road and in the opposite direction on the same road [13]. Guney et al. designed a controller aimed at resolving motion conflicts between AGVs. They used the traditional right-handed two-way traffic system based on the motion characteristics of AGVs in the controller design. They prevented collisions by determining the right of way for AGVs that would encounter each other at an intersection [14]. Zhao et al. proposed a dynamic resource reservation (DRR) based method that supports collision avoidance for AGVs. In this method, the layout is divided into square blocks of the same size. AGVs moving on each square shared their points. These source points were extracted from the guide paths in real time in the algorithm. The authors then took advantage of collision and crash prevention to make decisions for dynamic resource reservations. These two parameters allowed AGVs to travel without crashes while improving time efficiency. The simulation results in the study showed that the travel time was shortened between 11% and 28% [15]. Guo et al. developed a path-planning model to avoid collisions for AGVs. The authors calculated the travel speed, run time, and collision distance parameters of the AGV relative to its minimum inhibition rate. They performed several experiments to compare the average blocking rate, wait time, and completion time of different AGVs; run time under different interactive protocols; and average blocking rate of different scale maps. As a result of the experiments, they showed that the acceleration control method could resolve the collision faster and reduce the secondary collision probability [16].
AGVs follow a guidance system that contains the routing information in the environment. The guidance technologies used by AGVs are wire/inductive guidance, optical line guidance, magnetic tape guidance, laser navigation guidance, wireless sensor network guidance, vision guidance, barcode guidance, ultrasonic guidance, and GPS guidance [17,18,19].
Murakami, K. used a time–space network to formulate AGV routing as a mixed-integer linear programming problem. In this network, he handled the routing of AGVs and materials separately. He experimentally tested his system [20]. Lu et al. addressed the solution to the routing problem of AGVs using a deep-reinforcement-learning algorithm. They modeled the routing problem of AGVs as a Markov decision process and enabled real-time routing [21]. Holovatenko et al. proposed a mathematical algorithm and tracking system, including energy efficiency, that creates and follows the desired route between two points on a specific map [22]. Different studies based on autonomous driving methods of AGVs have been reported in the literature. The essential function of the autonomous driving method is machine vision. Some of the problems based on machine vision include floor path detection and strip marking detection [23,24,25]. Bayona et al. used computer vision for line tracking with a camera mounted on a model AGV. The vehicle was mainly successful in following a straight line on the floor in their tests [26]. Zhu et al. proposed a vehicle path tracking method that tracks the route by processing the image taken from the camera. They first applied a median filter to the image to clarify the road condition by reducing the picture’s noise and binarizing the image to reduce the computational complexity. They reduced image distortion and low path tracking accuracy issues [27]. Lee et al. proposed a vision-based AGV navigation system that uses a camera to detect triangular markers placed at different locations on the floor. Using the image’s hue and saturation values to identify markers in an image and choosing markers of a specific size, they achieved a 98.8% recognition rate [28]. Revilloud et al. developed a new strip detection and prediction algorithm. They used noise filters to improve the prediction of strip markers on the road. They detected the road in the image with 99.5% success using image processing and lane marking algorithms [29]. Liu, Y. et al. proposed a robot motion trajectory control using the ant colony algorithm and machine vision. They have shown that the system developed with experimental research can make reasonable trajectory planning in complex environments [30]. Lina et al. proposed a Kalman-based Single Shot Multi-Box Detector (K-SSD) target location strategy using a convolutional neural network (CNN) to improve position accuracy and mobile robot speed during automatic navigation. The experimental results showed that their proposed strategy increased the vehicle’s route detection and location accuracy in situations such as light differences and scene changes [31]. Sun et al. used images from the camera in the vehicle to find the path of an autonomous vehicle in a warehouse. They trained the Faster Region-based CNN model with these images. The label and shelf feet in the image were identified using the trained CNN, inverse perspective mapping (IPM) algorithm, and the support vector machine (SVM) algorithm. They obtained a virtual guideline from the camera images using shelf legs and labels in specific areas. They did not use real-time image processing or navigation in their work [32]. Ryota et al. developed an AGV transport system using deep-reinforced learning. They used map information for route planning in their transport system. They tested the validity of the method with the help of a simulator [33].

3. Materials and Methods

3.1. Convolutional Neural Network

A CNN is a deep-learning algorithm that enables high-accuracy image classification. The traditional CNN is usually composed of four layers, as shown in Figure 1. These are the convolution, ReLU, pooling, and fully connected layers. The convolution layer extracts a property map by applying filters to data from the input layer. The ReLU layer is used to eliminate the negative values of data from the convolution layer. The pooling layer reduces the input size while retaining the key features of the data. The fully connected layer is a neural network layer composed entirely of interconnected neurons and is used to classify the image [34,35].
Deep-learning systems require large quantities of data to train a model. Obtaining a trained model is a big problem when there are very small amounts of data for training. Moreover, considerable time is needed to train the model. The weights of a model trained in a similar field are transferred using the transfer learning method to solve problems. Smaller amounts of data can be used, and model training takes less time using this method [36,37].
This study designed the CNN model using the transfer learning method according to the single graphics processing unit (GPU) using the Keras library in Python. In the training of the CNN model, 160 × 160 × 3 size marker images obtained by data methods were used. The original images were reproduced by using data augmentation methods. The transfer learning method was used due to the small amount of our data set. The studies were carried out using the Mobile Net [38] deep-learning CNN model trained with Imagenet. MobileNet is a small, low latency model and can be run efficiently on mobile devices. Mobilenet models can achieve the success of popular models in terms of latency and accuracy.

3.2. Preparing the Dataset

Eight images were taken for each marker from the vehicle model camera to prepare the training dataset. The markers were placed at different distances from the vehicle camera, and four different sizes of images were taken. A total of 544 images were obtained from 17 different marker images. Many sample images are needed to train a CNN model [39]. For this reason, new images were obtained by using data augmentation from 544 original images. For data augmentation, rotating, shifting, shearing, zooming, and brightness methods were used. These data augmentation methods were applied to an image sequentially. Images identical to each other were deleted from the generated dataset. As a result of these operations, 18,085 images were obtained for the dataset. The result of the data augmentation applied to a sample image is shown in Figure 2.

3.3. Obtaining the Region of Interest in the Frame

Images taken from the camera are sent to the computer. Color images are converted to gray-level images to find the region of interest (ROI). They were processed using Gaussian blur, Otsu, and Canny edge detection algorithms. Otsu’s method is one of the most widely used thresholding algorithms in the literature. The values of all pixels in the image are taken, and their variance is calculated according to the distribution criteria with this method. Pixels are clustered according to these variance values [40,41]. A smoothing filter is applied to the image, and the filtered image is derived in the first step of the Canny algorithm. As a result of this process, single-pixel edge lines are obtained depending on directional gradient vectors [42]. Gray image transformation, Gaussian filter blurring, Otsu segmentation, and Canny edge detection are applied to images acquired from the camera sequentially. The ROI fields in the processed image are calculated, and the marker image having the largest area is cropped. The steps for obtaining the ROI are shown in Figure 3. The cropped marker image is applied to the CNN model to classify the marker information.

4. Deep-Learning-Based Floor Path Model for Route Tracking

The floor path model designed for the proposed system in the present study consists of discrete cellular markers. The other model component is the vehicle with a camera following the routes in the proposed floor path model. Markers are a basic structure consisting of a letter and a number. The numbers to be added to the model are obtained from increasing and decreasing letter–number pairs. A letter is added first, and numbers up to nine are added after the letter. After A and 1 are added side by side, the value A1 to be written on the marker is obtained. This value also indicates the location address of the marker. The location of the vehicle can be determined using this address. For the subsequent marker address, the number 2 is added to the letter A. This process continues until the A9 marker address is obtained. Therefore, nine markers are obtained from one letter. There is no letter change as long as it continues in the same direction on the floor. A sequence of two letters and one number is used when no more single-character letters are left to create the markers. The floor path model is obtained using these markers.
While creating routes on the floor path model, increasing markers from the starting location to the target location are selected. Routes created from markers, including increasing and decreasing letter–number pairs, also constitute ascending or descending routes. One advantage of using ascending or descending markers is that when the autonomous vehicle is disconnected from the system, it can reach the starting point by following the markers in the descending direction. The first route created by this method is a vector containing markers A1, A2, A3, A4, A5, D1, and D2. The D1 marker address is used because the vehicle’s direction will change after the A5 address in this route vector. The routes are composed of increasing or decreasing marker vectors, enabling the vehicle model to follow these routes. Bold lines are drawn around the markers to facilitate the detection of the marker. Intersections are formed at a 90-degree angle. There are twenty-five centimeters between the markers.
Special autonomous vehicles suitable for the proposed model were developed. A Raspberry Pi 3 board, Raspberry Pi motor driver board, 640 × 480 pixel color Raspberry Pi camera, and 7.4 V 1750 mAh Li-po battery were installed on the developed AVM. Furthermore, the AVM movement and the change in direction were controlled using several four direct current (DC) motors connected to the AVM wheels. The vehicle can drive at two different speeds. The camera on the vehicle has a single viewing angle and does not move left or right. The AVM used in the present study is shown in Figure 4.
First, in the proposed system, the floor path model is sent as a vector to autonomous vehicles by a server. The target marker address is required for the vehicle to move, as shown in Figure 5. The routes and approximate distances to the destination address are calculated by the vehicle. The closest route to the destination address is selected among the calculated routes. The destination vector of the selected route, which contains the markers to the destination, is created. The vehicle model is activated if there are no other vehicles on this route at least four marker distances from the selected vehicle. The four marker distances used were determined as the safe distance that prevents vehicles from colliding. Each vehicle broadcasts the last address information correctly classified by the deep-learning model on its route. If there is another vehicle on the selected route, the same checks are performed for the other route closer to the target. If no empty route is found, the vehicle model is suspended. One frame is taken per second from the video image obtained by the AVM camera in the image acquisition step. This image may contain more than one marker. The color image taken from the AVM camera is converted to a gray level image in the image processing step. By applying Gaussian blur, Canny edge detection, and Otsu algorithms to the gray level image, the frames of the distance markers in the image are blurred and the near ones are sharpened. Among these markers, the marker having the largest area is determined by comparing the area information of the markers in the image. This marker is cropped from the color image.
The cropped image is given as input to the pre-trained CNN model in the next algorithm step. The marker image classified by the CNN model is compared with the corresponding marker information in the target vector. Suppose the marker information is taken from the output of the network and this information is equal to the target vector marker information. In this case, the information as to whether there is another vehicle on the route is determined by using the address information sent by the vehicle moving on the route. If another vehicle is on the route, the new vehicle to follow the route waits to avoid a collision. If there is no other vehicle on the route, the vehicle moves along the route. If two vehicles have the same marker on their route, the movements of the vehicles are as follows: whichever vehicle is closer to the common marker, that vehicle continues to move on the route; the other vehicle continues to move when the address information of the common marker is not received.
In this way, collisions of vehicles are prevented. The markers in the target vector must be in ascending order for the model vehicle to move. If these two markers are different, the vehicle model turns to the right to see the images of the marker on the floor path and the operation of the system heads into the image acquisition step. The right turn of the vehicle model is important for the detection of other markers. These operations continue until the last marker in the target vector is reached. In the case of an error, the vehicle returns to its starting address using the markers in the vector.
The ascending order addressing method is used to calculate routes to the destination marker. The vehicles calculate routes and approximate distance to the target marker using incremental addressing from the vectors in the floor path model and keep these routes as vectors. The pseudo-code that calculates routes to the destination marker and their lengths is shown in Algorithm 1:
Algorithm 1. The pseudo-code that calculates routes to the destination marker and their lengths.
Let i = 0, k = 0, z = 0; RouteAll: array, RouteLocal:array
While I <= Vektor.Count
  Let j = 0
  Let V = Vector[i, j++]
  RouteLocal[k++] = V
  If V has sub vector
   Add all SubVectors to RouteLocal
  RouteAll[z++] = RouteLocal
  i++
End While
Let Route: array, LengthRoute: array
i = 0, k = 0
While I <= RouteAll.Count
  If RouteAll[i] Contains TagetMarker
   Route[k] = RouteT[i]
   LengthRoute[k] = Route[k].Count * Distance
  i++
End While
Here Vector contains all the markers in the floor path model received from the server. SubVector denotes subroutes going in different directions from a marker on the route. RouteAll shows all routes that will be created, while RouteLocal contains vectors and subroute vectors that will branch from a marker. Route addresses all routes to the target marker. LengthRoute is the length of each route to the destination marker and Distance indicates the distance between two markers.
Let the vehicle be given the F3 marker address as the target and the vehicle will go to the A3 marker position. The vehicle will calculate the route vectors to marker F3 as follows: the first route is A4, A5, A6, F1, F2, and F3. The second route is B1, B2, E1, E2, F2, and F3. The approximate length of each route is calculated by multiplying the number of markers by the distance between two markers. The route with the shorter distance is selected as the route to the destination. For a new scenario, the floor model has markers B1, B2, B3, C1, C2, C3, C4, D1, D2, D3, F1, F2, F3, F4, G1, G2, G3, H1. Let the vehicle be given the B1 marker address as the target and the vehicle will go to the G2 marker position. The vehicle will calculate the route vectors to marker G3 as follows: the first route is B2, B3, B4, B5, and G2. The second route is B2, B3, D1, D2, F1, F2, F3, F4, G1, and G2. The shortest route with markers B2, B3, B4, B5, and G2 will be selected.
The working principle of the designed system can be briefly summarized as follows, shown in Figure 6. All autonomous vehicles in an area act in communication with a central server. The server first sends the Floor Path Model to the vehicles as a vector. Then, the user or server gives the destination marker address to the vehicle that will move. The tool calculates all possible routes to this address and chooses the shortest one. On this selected route, the vehicle proceeds to reach the destination. While the vehicle moves on the route, it sends the last marker it reads to the server. Other autonomous vehicles know that there is a vehicle on that route by communication provided in this way.

5. Experimental Results

The performance of the deep-learning-based positioning system was tested by experiments. Different routes were created and used in the present study. For example, the first route consists of markers A1, A2, A3, A4, A5, D1, D2; the second route consists of markers A1, A2, A3, A4, A5, A6, F1, F2, F3, G1, G2; and the third route consists of markers A1 A2, A3, B1, B2, E1, E2, F2, F3, G1, G2. The studies were carried out for two different speed levels of the vehicle model for each route. The completion times of the routes and speed levels of the vehicle model are given in Table 1.
The completion time of the route followed by the vehicle model varies according to the distance between the markers. The linear distance between the markers was taken as 25 cm. Images from the vehicle camera can be seen clearly by using the speed values given as 0.22 m/s and 0.27 m/s in Table 1. However, when higher speeds are used, images from the vehicle camera become distorted.
Images of markers A1, A2, A3, A4, A5, A6, B1, B2, D1, D2, E1, E2, F1, F2, F3, G1, G2 were augmented using the data augmentation methods. Of these data, 71% were reserved for training, 25% for validation, and 4% for testing. The CNN model was obtained using transfer learning and was trained with original and augmented images together. The images combined with the fully connected layer were classified into 17 different classes by the soft-max function. The steps to obtain the image of interest on the route by processing the images from the vehicle camera are shown in Figure 7.
The application of the Otsu and Canny algorithms together with the image gave more successful results than the application of the Canny algorithm alone to obtain the markers in an image taken from the vehicle model camera.
In the image where both algorithms are applied, there is more edge distortion at markers and objects far from the camera, and the success of detecting markers closer to the camera increases. Firstly, Gaussian blur was applied to the image in order to increase the distortion. As shown in Figure 8, when images Figure 8a,b are compared, it is clearly seen that edge distortions occur in Figure 8a.

6. Results and Discussion

MobileNet is a model trained with an ImageNet dataset. The weights of this model were obtained by using transfer learning. The training set used in the study was different from the ImageNet dataset. For this reason, the weights of the model layers were updated with the new images used during the training. All layers were used in the MobileNet model except the last fully connected layers. The batch normalization layer was added to the classification layer in the model. Batch normalization acts as a regularizer, in some cases eliminating the need for Dropout [43]. Batch normalization was used to reduce overfitting.
The low processor power in model vehicles accelerates the process of image processing. Therefore, the image size for model training was set to 160 × 160 × 3 pixels. The initial learning rate was 0.0001, and the loss function was selected as categorical cross-entropy. Because the number of classes was more than two, the categorical cross-entropy function was used. If the loss value cannot be reduced three times in a row with the determined learning rate, the learning rate is reduced by the specified delta value. The minimum delta value determined was 0.0001. There are classes that are similar to each other in the dataset. This causes problems in the performance graph while training the model. When the batch-size value is large, it takes more time for the model tool to process the data. When this value is small, the model learns the noises and moves away from generalization. Therefore, a mini-batch size of 32 was selected for each epoch at the end of the tests. Adam was used as the optimizer. Softmax was chosen because the activation function of the last layer consists of 17 classes. Model training was ended when the validation loss value was less than 0.009. Model training was completed in 16 epochs.
Training and loss graphics of the model are given in Figure 9. It can be seen from the figure that the training loss and validation loss curves follow each other and progress at very close levels. To evaluate the performance of our model, we used the precision rate, recall rate, and F1-score metrics.
These performance metrics are given below:
P r e c i s i o n = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   P o s i t i v e        
R e c a l l = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   N e g a t i v e                  
F 1 s c o r e = 2   ×   P r e c i s i o n   ×   R e c a l l P r e c i s i o n + R e c a l l          
The precision metric shows how many of the values positively predicted by the model are positive. The recall metric shows how many of the values that the model should positively predict are positively estimated. Because there was more than one class in the dataset in the study, precision, recall, and F1-score metrics were also used besides the classification accuracy metric for model performance. It is seen that all other markers except the E1 marker were classified correctly in the confusion matrix shown in Figure 10a. When the matrix was examined, it was seen that the marker that should be classified as E1 was incorrectly classified as F1. Thirty-four of the 35 E1 markers were correctly classified. This is because the E1 and F1 markers are similar. To test the trained deep-learning model, a total of 796 marker images, which were never given before, were entered as input to the model. As seen in Figure 10b, the model classified the F1 marker image 49 times as F1 and once as E1.
The results obtained when the test set images were applied to the trained model, which are shown in Figure 11. It is seen here that the A4 image on the left has the lowest ratio of 88.81%, and the B1 image on the right has 76.35%. Both markers are correctly classified by the trained model. It is also seen that the model classifies all markers correctly. The performance metrics of the trained model are given in Table 2. It can be seen in the table that marker E1 has the lowest value with a 0.97 recall metric value. It is seen that the F1 score value of the E1 marker is 0.99. The F1 marker has a 0.97 precision ratio and a 0.98 F1 score rate.
It can be seen from Table 2 that the worst performance of the model is 97%. The average performance of each metric is 99.82%.
The ROI image obtained by pre-processing from the view from the AVM camera was given as input to the trained model. The latency times of the model to classify this image are given in Table 3. It can be seen from the table that the average latency time is 0.01449 s.

7. Conclusions

Traditional floor path methods used in production areas have a limited number of road routes, which can cause traffic congestion and workload imbalance. In the present study, an indoor positioning system based on a floor path model and algorithm have been developed for autonomous vehicles. A floor path model consisting of markers was developed. All routes to a given destination on the floor path model are created using the dynamic route planner. The vehicle chooses the shortest alternative route with the least number of vehicles among these routes and goes to its destination without colliding with another vehicle. Markers on the floor path model can be detected with a low-resolution camera. Because the floor path model is created with increasing and decreasing markers, the vehicle model can easily follow a consecutive increasing or decreasing route. Traffic density can be minimized and workload can be balanced by using these alternative routes. Vehicles on the floor path model broadcast their marker address for all vehicles. If there is another vehicle on the vehicle’s route, it waits until there are at least four markers between the vehicles in the proposed algorithm. When the floor path model and algorithm are used, it is not allowed to have more than one vehicle at a marker address and collisions of vehicles are prevented. It was shown that it is possible to create multiple routes to the destination using the proposed floor path model.
When the results of the CNN model trained using transfer learning were examined, it was seen that 34 of the 35 E1 markers were correctly classified. The lowest recall metric value of 0.97 and an F1-score metric of 0.99 were obtained for the E1 marker containing 35 E1 marker images in the test set. The lowest precision metric of 0.97 and an F1-score metric of 0.98 were obtained for the F1 marker containing 29 F1 marker test images. When the test set images were applied to the trained CNN model, the lowest rate was classified as the A4 marker image with 88.81% and the highest rate with the G2 marker image with 99.91%. In addition, it was observed that the model correctly classified all the markers.
In the present study, by using the indoor positioning system, the route determined among many alternative routes was selected and followed. The results showed that the developed floor path model, algorithm, and deep-learning model can be used successfully both in a real production environment and in the service sector areas such as storage and healthcare.
One of the potential future works is that the experimentally designed and modeled pavement path model can be applied in real industrial environments. The camera was fixed in the autonomous vehicle used in the present study. Action cameras could be used in future studies.

Author Contributions

Conceptualization, I.C.; Data curation, M.E.; Formal analysis, M.E.; Investigation, M.E. and I.C.; Methodology, M.E. and I.C.; Project administration, I.C.; Software, M.E.; Supervision, I.C.; Validation, M.E.; Visualization, I.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The study did not report any data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jung, S.H.; Woo, D.G.; Han, J.D.; Shin, J.K.; Kim, K.R. CNN Based Factory Lane Marker Recognition for Indoor Path Tracking of Automated Guided Vehicle. In Symposium of the Korean Institute of Communications and Information Sciences; Korea Institute of Communication Sciences: Daejeon, Korea, 2018; pp. 296–297. [Google Scholar]
  2. Ngo, H.Q.T.; Nguyen, T.P.; Nguyen, H. Research and Develop of AGV Platform for the Logistics Warehouse Environment. In Proceedings of the Future Technologies Conference, Vancouver, BC, Canada, 15–16 November 2018; Springer: Cham, Switzerland, 2018; pp. 455–465. [Google Scholar] [CrossRef]
  3. Xing, W.; Peihuang, L.; Jun, Y.; Xiaoming, Q.; Dunbing, T. Intersection Recognition and Guide-Path Selection for a Vision-Based AGV in a Bidirectional Flow Network. Int. J. Adv. Robot. Syst. 2014, 11, 39. [Google Scholar] [CrossRef] [Green Version]
  4. Gutta, P.R.; Chinthala, V.S.; Manchoju, R.V.; Mvn, V.C.; Purohit, R. A Review on Facility Layout Design of an Automated Guided Vehicle in Flexible Manufacturing System. Mater. Today Proc. 2018, 5, 3981–3986. [Google Scholar] [CrossRef]
  5. Yao, L.; Miller, J. Tiny ImageNet Classification with Convolutional Neural Networks. 2015. Available online: http://cs231n.stanford.edu/reports/2015/pdfs/leonyao_final.pdf (accessed on 27 September 2021).
  6. Yu, X.; Wu, X.; Luo, C.; Ren, P. Deep learning in remote sensing scene classification: A data augmentation enhanced convolutional neural network framework. GISci. Remote Sens. 2017, 54, 741–758. [Google Scholar] [CrossRef] [Green Version]
  7. Smolyanskiy, N.; Kamenev, A.; Smith, J.; Birchfield, S. Toward low-flying autonomous MAV trail navigation using deep neural networks for environmental awareness. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4241–4247. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, Q.; Wang, C. Exploration of port intelligent AGV path tracking based on vision. J. Intell. Fuzzy Syst. 2020, 38, 1281–1285. [Google Scholar] [CrossRef]
  9. Wang, Y.; Ding, H.; Yuan, J.; Chen, H. Output-feedback triple-step coordinated control for path following of autonomous ground vehicles. Mech. Syst. Signal Process. 2019, 116, 146–159. [Google Scholar] [CrossRef]
  10. Cil, I.; Arisoy, F.; Özgürbüz, E.; Cil, A.Y.; Kılınç, H. Indoor positioning technology selection using a combined AHP and PROMETHEE method at SEDEF shipyard. J. ETA Marit. Sci. 2022, 10, 2–17. [Google Scholar] [CrossRef]
  11. Cheong, H.-W.; Lee, H. Concept Design of AGV (Automated Guided Vehicle) Based on Image Detection and Positioning. Procedia Comput. Sci. 2018, 139, 104–107. [Google Scholar] [CrossRef]
  12. Zhou, Y.; Hu, H.; Liu, Y.; Ding, Z. Collision and Deadlock Avoidance in Multirobot Systems: A Distributed Approach. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 1712–1726. [Google Scholar] [CrossRef]
  13. Małopolski, W. A sustainable and conflict-free operation of AGVs in a square topology. Comput. Ind. Eng. 2018, 126, 472–481. [Google Scholar] [CrossRef]
  14. Guney, M.A.; Raptis, I.A. Dynamic prioritized motion coordination of multi-AGV systems. Robot. Auton. Syst. 2020, 139, 103534. [Google Scholar] [CrossRef]
  15. Zhao, Y.; Liu, X.; Wang, G.; Wu, S.; Han, S. Dynamic Resource Reservation Based Collision and Deadlock Prevention for Multi-AGVs. IEEE Access 2020, 8, 82120–82130. [Google Scholar] [CrossRef]
  16. Guo, K.; Zhu, J.; Shen, L. An Improved Acceleration Method Based on Multi-Agent System for AGVs Conflict-Free Path Planning in Automated Terminals. IEEE Access 2020, 9, 3326–3338. [Google Scholar] [CrossRef]
  17. Vale, A.; Ventura, R.; Lopes, P.; Ribeiro, I. Assessment of navigation technologies for automated guided vehicle in nuclear fusion facilities. Robot. Auton. Syst. 2017, 97, 153–170. [Google Scholar] [CrossRef]
  18. Lynch, L.; Newe, T.; Clifford, J.; Coleman, J.; Walsh, J.; Toal, D. Automated Ground Vehicle (AGV) and Sensor Technologies—A Review. In Proceedings of the 2018 12th International Conference on Sensing Technology (ICST), Limerick, Ireland, 4–6 December 2018; pp. 347–352. [Google Scholar] [CrossRef]
  19. Oyekanlu, E.A.; Smith, A.C.; Thomas, W.P.; Mulroy, G.; Hitesh, D.; Ramsey, M.; Kuhn, D.J.; Mcghinnis, J.D.; Buonavita, S.C.; Looper, N.A.; et al. A Review of Recent Advances in Automated Guided Vehicle Technologies: Integration Challenges and Research Areas for 5G-Based Smart Manufacturing Applications. IEEE Access 2020, 8, 202312–202353. [Google Scholar] [CrossRef]
  20. Murakami, K. Time-space network model and MILP formulation of the conflict-free routing problem of a capacitated AGV system. Comput. Ind. Eng. 2020, 141, 106270. [Google Scholar] [CrossRef]
  21. Lu, C.; Long, J.; Xing, Z.; Wu, W.; Gu, Y.; Luo, J.; Huang, Y. Deep Reinforcement Learning for Solving AGVs Routing Problem. In Verification and Evaluation of Computer and Communication Systems; Springer: Cham, Switzerland, 2020; pp. 222–236. [Google Scholar]
  22. Holovatenko, I.; Pysarenko, A. Energy-Efficient Path-Following Control System of Automated Guided Vehicles. J. Control. Autom. Electr. Syst. 2021, 32, 390–403. [Google Scholar] [CrossRef]
  23. Hu, C.; Wang, R.; Yan, F.; Chen, N. Output Constraint Control on Path Following of Four-Wheel Independently Actuated Autonomous Ground Vehicles. IEEE Trans. Veh. Technol. 2015, 65, 4033–4043. [Google Scholar] [CrossRef]
  24. John, V.; Liu, Z.; Mita, S.; Guo, C.; Kidono, K. Real-time road surface and semantic lane estimation using deep features. Signal Image Video Process. 2018, 12, 1133–1140. [Google Scholar] [CrossRef]
  25. Mahaleh, M.B.B.; Mirroshandel, S.A. Harmony search path detection for vision based automated guided vehicle. Robot. Auton. Syst. 2018, 107, 156–166. [Google Scholar] [CrossRef]
  26. Bayona, J.L.D.; Pilonieta, J.D.G.; Lache, R.M.M.; Prada, S.R. Development of a control system for path following applications in an AGV using computer vision. In Proceedings of the 2020 IX International Congress of Mechatronics Engineering and Automation (CIIMA), Cartagena, Colombia, 4–6 November 2020; pp. 1–6. [Google Scholar] [CrossRef]
  27. Zhu, Y.; Guo, N. Unmanned vehicle route tracking method based on video image processing. Jordan J. Mech. Ind. Eng. 2020, 14, 139–147. [Google Scholar]
  28. Lee, J.; Hyun, C.-H.; Park, M. A Vision-Based Automated Guided Vehicle System with Marker Recognition for Indoor Use. Sensors 2013, 13, 10052–10073. [Google Scholar] [CrossRef] [PubMed]
  29. Revilloud, M.; Gruyer, D.; Rahal, M.-C. A lane marker estimation method for improving lane detection. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 289–295. [Google Scholar] [CrossRef]
  30. Liu, Y.; Zhang, X.; Qu, T.; Yin, D.; Deng, S. Intelligent robot motion trajectory planning based on machine vision. Int. J. Syst. Assur. Eng. Manag. 2022. [Google Scholar] [CrossRef]
  31. Lin, W.; Ren, X.; Hu, J.; He, Y.; Li, Z.; Tong, M. Fast, robust and accurate posture detection algorithm based on Kalman filter and SSD for AGV. Neurocomputing 2018, 316, 306–312. [Google Scholar] [CrossRef]
  32. Sun, Y.; Su, T.; Tu, Z. Faster R-CNN based autonomous navigation for vehicles in warehouse. In Proceedings of the 2017 IEEE International Conference on Advanced Intelligent Mechatronics (AIM), Munich, Germany, 3–7 July 2017; pp. 1639–1644. [Google Scholar] [CrossRef]
  33. Kamoshida, R.; Kazama, Y. Acquisition of Automated Guided Vehicle Route Planning Policy Using Deep Reinforcement Learning. In Proceedings of the 2017 6th IEEE International Conference on Advanced Logistics and Transport (ICALT), Bali, Indones, 24–27 July 2017; pp. 1–6. [Google Scholar]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  35. Cao, J.; Pang, Y.; Li, X.; Liang, J. Randomly translational activation inspired by the input distributions of ReLU. Neurocomputing 2018, 275, 859–868. [Google Scholar] [CrossRef]
  36. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A survey of transfer learning. J. Big Data 2016, 3, 1345–1459. [Google Scholar] [CrossRef] [Green Version]
  37. Day, O.; Khoshgoftaar, T.M. A survey on heterogeneous transfer learning. J. Big Data 2017, 4, 29. [Google Scholar] [CrossRef]
  38. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  39. Han, D.; Liu, Q.; Fan, W. A new image classification method using CNN transfer learning and web data augmentation. Expert Syst. Appl. 2018, 95, 43–56. [Google Scholar] [CrossRef]
  40. Goh, T.Y.; Basah, S.N.; Yazid, H.; Safar, M.J.A.; Saad, F.S.A. Performance analysis of image thresholding: Otsu technique. Measurement 2018, 114, 298–307. [Google Scholar] [CrossRef]
  41. Canny, J.F. A Computational Approach To Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–714. [Google Scholar] [CrossRef] [PubMed]
  42. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, ICML, Lille, France, 6–11 July 2015; Volume 1, pp. 448–456. [Google Scholar]
  43. Ben Khalifa, A.; Alouani, I.; Mahjoub, M.A.; Ben Amara, N.E. Pedestrian detection using a moving camera: A novel framework for foreground detection. Cogn. Syst. Res. 2020, 60, 77–96. [Google Scholar] [CrossRef]
Figure 1. Traditional CNN topology.
Figure 1. Traditional CNN topology.
Systems 10 00083 g001
Figure 2. Data augmentation methods in producing markers to be used in training the CNN model.
Figure 2. Data augmentation methods in producing markers to be used in training the CNN model.
Systems 10 00083 g002
Figure 3. Image processing steps. (a) Acquired image, (b) Graying, (c) Blurring, (d) Otsu and Canny, (e) Marker detection, (f) Cropped marker image.
Figure 3. Image processing steps. (a) Acquired image, (b) Graying, (c) Blurring, (d) Otsu and Canny, (e) Marker detection, (f) Cropped marker image.
Systems 10 00083 g003
Figure 4. Autonomous vehicle model.
Figure 4. Autonomous vehicle model.
Systems 10 00083 g004
Figure 5. The developed route tracking flow chart.
Figure 5. The developed route tracking flow chart.
Systems 10 00083 g005
Figure 6. The block diagram of the system.
Figure 6. The block diagram of the system.
Systems 10 00083 g006
Figure 7. Image processing steps. (a) Acquired image, (b) Graying, (c) Blurring, (d) Otsu and Canny, (e) Marker detection, (f) Cropped marker image.
Figure 7. Image processing steps. (a) Acquired image, (b) Graying, (c) Blurring, (d) Otsu and Canny, (e) Marker detection, (f) Cropped marker image.
Systems 10 00083 g007
Figure 8. Application of Otsu and Canny algorithms. (a) Image with Otsu and Canny algorithms applied, (b) Image with Canny algorithm applied.
Figure 8. Application of Otsu and Canny algorithms. (a) Image with Otsu and Canny algorithms applied, (b) Image with Canny algorithm applied.
Systems 10 00083 g008
Figure 9. Train and loss graphics for the model.
Figure 9. Train and loss graphics for the model.
Systems 10 00083 g009
Figure 10. Confusion matrix for the CNN model (a) Model test result 1; (b) Model test result 2.
Figure 10. Confusion matrix for the CNN model (a) Model test result 1; (b) Model test result 2.
Systems 10 00083 g010aSystems 10 00083 g010b
Figure 11. The test set results for the trained model (a) Model classification result 1; (b) Model classification result 2.
Figure 11. The test set results for the trained model (a) Model classification result 1; (b) Model classification result 2.
Systems 10 00083 g011
Table 1. Route tracking times at different speeds of the vehicle model.
Table 1. Route tracking times at different speeds of the vehicle model.
Route NumberSpeed (m/s)Route Completion Time (s)Average Completion Time (s)
10.2215.4115.77
16.35
15.56
0.2715.0914.72
14.15
14.93
20.2222.8922.58
21.87
22.97
0.2720.0519.08
18.31
18.87
30.2230.4031.15
31.45
31.60
0.2728.4828.44
27.48
29.35
Table 2. Performance metrics of the model.
Table 2. Performance metrics of the model.
LabelPrecisionRecallF1-ScoreNumber of Test Images
A11.001.001.0045
A21.001.001.0044
A31.001.001.0037
A41.001.001.0047
A51.001.001.0047
A61.001.001.0046
B11.001.001.0048
B21.001.001.0042
D11.001.001.0048
D21.001.001.0056
E11.000.970.9935
E21.001.001.0026
F10.971.000.9829
F21.001.001.0021
F31.001.001.0048
G11.001.001.0039
G21.001.001.0047
Average99.8299.899.8
Table 3. Test latency times for the model.
Table 3. Test latency times for the model.
Measurement NumberModel Latency Times (Second)
10.01447
20.01401
30.01511
40.01440
50.01447
Average0.01449
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Erginli, M.; Cil, I. Deep-Learning-Based Floor Path Model for Route Tracking of Autonomous Vehicles. Systems 2022, 10, 83. https://doi.org/10.3390/systems10030083

AMA Style

Erginli M, Cil I. Deep-Learning-Based Floor Path Model for Route Tracking of Autonomous Vehicles. Systems. 2022; 10(3):83. https://doi.org/10.3390/systems10030083

Chicago/Turabian Style

Erginli, Mustafa, and Ibrahim Cil. 2022. "Deep-Learning-Based Floor Path Model for Route Tracking of Autonomous Vehicles" Systems 10, no. 3: 83. https://doi.org/10.3390/systems10030083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop