Next Article in Journal
Low-Cost Portable System for Measurement and Representation of 3D Kinematic Parameters in Sport Monitoring: Discus Throwing as a Case Study
Next Article in Special Issue
Mutual-Aided INS/Vision Navigation System Analysis and Optimization Using Sequential Filtering with State Recalculation
Previous Article in Journal
GPS Spoofing Detection Method for Small UAVs Using 1D Convolution Neural Network
Previous Article in Special Issue
A Review on Machine Learning Applications for Solar Plants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Path Generator with Unpaired Samples Employing Generative Adversarial Networks

by
Javier Maldonado-Romo
1,2,*,
Alberto Maldonado-Romo
3 and
Mario Aldape-Pérez
2,*
1
Institute of Advanced Materials and Sustainable Manufacturing, Tecnologico de Monterrey, Mexico City 14380, Mexico
2
Centro de Innovación y Desarrollo Tecnológico en Cómputo, Instituto Politécnico Nacional, Unidad Profesional Adolfo López Mateos, Juan de Dios Bátiz s/n esq. Miguel Othón de Mendizábal, Mexico City 07700, Mexico
3
Centro de Investigación en Computación, Instituto Politécnico Nacional, Unidad Profesional Adolfo López Mateos, Juan de Dios Bátiz s/n esq. Miguel Othón de Mendizábal, Mexico City 07700, Mexico
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(23), 9411; https://doi.org/10.3390/s22239411
Submission received: 19 September 2022 / Revised: 12 November 2022 / Accepted: 25 November 2022 / Published: 2 December 2022

Abstract

:
Interactive technologies such as augmented reality have grown in popularity, but specialized sensors and high computer power must be used to perceive and analyze the environment in order to obtain an immersive experience in real time. However, these kinds of implementations have high costs. On the other hand, machine learning has helped create alternative solutions for reducing costs, but it is limited to particular solutions because the creation of datasets is complicated. Due to this problem, this work suggests an alternate strategy for dealing with limited information: unpaired samples from known and unknown surroundings are used to generate a path on embedded devices, such as smartphones, in real time. This strategy creates a path that avoids virtual elements through physical objects. The authors suggest an architecture for creating a path using imperfect knowledge. Additionally, an augmented reality experience is used to describe the generated path, and some users tested the proposal to evaluate the performance. Finally, the primary contribution is the approximation of a path produced from a known environment by using an unpaired dataset.

1. Introduction

Different academic domains, as well as daily life, have been incredibly touched by interactive multimedia systems [1]. For instance, they provide capabilities for robotic exploration in order to locate obstacles in a physical area [2]. In the realm of e-learning, materials are displayed to help students understand abstract concepts [3]. The use of virtual elements in systems for training in the healthcare industry was described in [4,5]. Therefore, the primary contribution of an interactive multimedia system is the enhancement of real-world settings with virtual components. Such technologies are described by the concepts of augmented reality (AR) and mixed reality (MR). AR technology allows computer-generated virtual visuals to perfectly overlay real-world objects in real time [6]. MR, on the other hand, enables concurrent work in both the physical and virtual domains, hence minimizing domain transition costs [7]. The primary distinction between MR and AR is that MR provides a more immersive experience.
Since MR requires a comprehension of the surroundings, specific equipment that contains sensors to sense the environment is required [8]. In order to provide a real-time user-interaction experience, the system must handle a great volume of data, which generates numerous important opportunities and difficulties [9]. Consequently, the current equipment is intended for a limited audience due primarily to its technological qualities, which limit the gradual extension of this experience on a large scale [10]. Nevertheless, unlike in MR, an immersive experience in AR is unnecessary, which reduces the prices and requirements for equipment features. For instance, mobile devices, such as smartphones, are an integral part of people’s everyday lives, generating a vast presence in conducting a variety of tasks. In 2020, there were at least 50 billion gadgets [11]. These gadgets have provided diverse interactive experiences, including AR experiences, mostly through markers and telemetry [12].
Existing technologies that enable interactive AR experiences, such as the ARCore and ARKit frameworks, make it possible to comprehend interactions with the environment in real time [13]. These frameworks concentrate on gathering information from flat surfaces to display virtual elements, but these frameworks disregard the fact that items may produce an accident between the user and the surrounding environment [14,15]. Therefore, current solutions lack an approach to this fundamental obstacle when providing this type of application: an environment’s security. The authors of [16] defined a collision as the avoidance of a path or trajectory between two configurations embedded in the cost field while motion restrictions are considered. Similarly, free space allows agents to interact freely with the environment to avoid accidents.
For this reason, environmental comprehension is essential to the real-time processing of the environment. In order to apply solutions that enhance the features of constrained equipment, data processing that employs machine learning (ML) techniques is required [17]. For instance, the authors of [18] emulated the behavior of a sophisticated sensor by using machine learning. Deep convolutional neural networks (DCNNs) reduce the number of operations and can operate on devices with limited computational resources [19]. In addition, this technique was supplemented by generative adversarial networks (GANs) to create depth image samples [20]. These enabled the production of a depth image via competition between two networks.
The creation of a specific dataset is a significant constraint of existing methods. Such datasets comprise paired samples, since the predicted sample and source image have linked attributes. This constraint creates a disadvantage for this type of dataset in terms of data collection. For example, RGB-D data are required in order to perceive the scene, but this strategy could be more effective in undetermined scenes, since depth information is necessary to estimate samples. Since the predicted images are unconnected to the source images, this problem involves unpaired samples [21]. However, the concept of the neural transfer style, which involves learning the features from one sample to another, is an alternative method for merging sample features [22]. Cycle-GAN and pix2pix are two concepts for combining the properties of one domain with those of another [23]. In order to interact with a known environment without prior information, an unknown domain modifies its transfer style.
As a result of the above, this proposal offers a path planner generator for avoiding collisions by employing GANs. In addition, the architecture purports to be a substitute for interacting with numerous settings in order to minimize design time and enable the use of smartphones and other devices with few resources. Consequently, the following hypothesis is investigated in this study. Since the GAN permits the transfer of samples from one domain to another, it could infer the properties between domains to connect samples from a known situation with paths through an unfamiliar environment.
This work investigates and explains the influence of inferring features from a known scenario to an unknown scenario in order to approximate a generic solution for creating pathways in real time on low-cost devices.
The remaining sections of this manuscript are structured as follows. Section 2 describes the context and research gaps. The suggested work is introduced in Section 3. Section 4 illustrates the outcomes and analyses of the experiments. The conclusion is then stated in Section 5.

2. Background and Research Gaps

In the research field of autonomous intelligent systems, path planning is a problem where the system determines the optimum collision-free path between two points. Usually, the implemented architecture incorporates the perception and the planning modules separately, forming a complex architecture to offer a real-time performance [24,25]. Nevertheless, the performance in the real world is inefficient in terms of time and resources [26]. As a result, virtual simulators expedite the design process. Simulators such as Airsim help evaluate agents’ dual systems [27]. Therefore, simulators generate samples of possible scenarios for training algorithms that require real-time perception based on ML techniques [28].
The perception of the environment requires specialized sensors, but the current tools require previous interactions to be effective. For instance, ARcore needs to move the device to detect key features to build a 3D environment described by the ORB-SLAM approach [29]. Once the system scans the scene, flat surfaces are detected, offering real-time interaction with the scenery, but the system positions virtual elements in places that are difficult to access or that may cause accidents. For example, Figure 1 shows a virtual element on a flat surface, offering a limited experience in terms of interaction with the environment because the selection of the surface is random, generating some inconvenience in the experience. However, mobile devices require high computer performance but are limited in computer power and battery. Therefore, AR and MR experiences require understanding the environment to determine the best place to display the virtual elements.
Since specific sensors are required and avoid previous interaction with the environment, different proposals have offered alternative solutions, such as replacing the sensor with ML algorithms [30]. Existing evidence where ML replaces specialized smartphone sensors, as described in [31], is used to accurately replace a depth sensor from an image taken from the device. Likewise, another alternative solution is to employ simulators and ML approaches to connect a sample-limited environment with a known environment [32]. Thus, ML techniques provide novel features to the current systems.
On the other hand, the high computing power on an embedded device is high in time and resource needs, limiting the roll-out on generic devices. Therefore, the end-to-end approach reduces external elements such as sensors since the perception of the scenario is considered within a simulation [33]. This approach reduces the conventional architecture and additional sensors to generate the necessary information to perceive the environment.
Although considerable efforts have been made to create alternatives through ML techniques, there is a severe issue where the current solutions focus on particular problems. In other words, a dataset is required for each environment. Since the current solutions implement datasets in the training environment based on specific environments, we have detected that GANs have been used to perform feature blending between unknown domains through images [34,35]. Therefore, GANs can offer novel solutions by exploring domain changing through style transfer between two domains.
The current work aims to contribute to the analysis of the behavior in feature inference between an image acquired from an unknown domain and translate it to a controlled scenario to generate a safe path and evaluate its behavior and determine the reliability of deploying virtual elements in augmented reality applications in real time. Furthermore, the conventional architecture described by Figure 2a is composed of two main modules, the perception and the planner. Likewise, the conventional architecture is reduced for being employed on embedded devices only utilizing a single module, as Figure 2b illustrates.

3. Proposed Work

The present proposal takes the principal features of cycle-GAN and pix2pix models since these approaches generate a new sample according to features of an unknown domain. Based on this approach, the first step is to define the available domain to generate the dataset with known paths connected to the unknown environment. The referent environment is known because it includes the depth data for each sample. In contrast, the depth value in the second scenario is unknown because the depth data is unavailable. As depicted in Figure 3, the virtual simulator generates paired samples, and the samples of the unknown scenario are unpaired because the relation between data is unfamiliar.
In order to describe the performance of the experiment, four environments were defined. An environment represents a domain since a domain contains elements with similar features. Therefore, each environment has different features, such as illumination, color distribution, and textures, as shown in Figure 4. The following analysis describes the performance for each domain according to features such as the average color and dominant color employing the k-means approach [36]. These outcomes are described in Table 1. It is observed that it is challenging to determine whether transfer and regular features exists in relation with the known environment because the unknown environment A has a bad behavior since the dominant color tends to be black. Consequently, the interpretation of data is limited to determining the relation of approximation to the expected environment.
In order to describe the data, each image is plotted in a three-dimensional space. Therefore, the sample is converted into a 3D point. This analysis employs a deep convolutional neuronal network (DCNN) model to plot a sample in 3D space; Figure 5 describes the characteristics. For this case, the model initializes the variables arbitrarily. Consequently, the distribution of the elements in space varies based on the random values, and Tanh has been implemented as a transfer function because the data are distributed throughout several spatial regions (Figure 6a). According to Figure 6b, the transfer style approximates the unknown sample to a known sample. This behavior shows that the GAN trough transfer style positively affects the data between two domains.
Due to the features of each domain depicted in Figure 6c, the unknown domains are distributed throughout the 3D space. Red, for instance, indicates a recognized environment, while green, blue, and cyan represent unknown environments. As is shown in Figure 6d, the approximation from unknown domains to known domains is near since the data concentration covers a similar space region after applying the GAN architecture. Similarly, the centroid of each domain and its distance from the known environment are determined. According to Table 2, the centroid for each domain is close to the known environment centroid. Therefore, the GAN approximates the characteristics between unidentified space domains.
This proposal adds a step to the GAN architecture once the features have been approximated: a hierarchical clustering (HC) [37] where the closest sample is obtained based on the distance between the generated sample and the known samples. Since the accessible domain contains depth information for determining a path, the generated sample must approximate a known sample. Furthermore, the rapidly exploring random tree (RRT) algorithm establishes the collision-free path [38], as is shown in Figure 7.
To generate a path, the calculus of variations describes the composition of a path [39] as the sum of the distances between two consecutive places defined by Equations (1) and (2). Consequently, a path consists of a collection of lines in space. Likewise, one characteristic of a path is its capacity to avoid obstacles. Path planning is the shortest distance between m obstacles O and the best value with the highest level of free collisions in a series of points p of length n. In this sense, the maximum optimization problem is changed into the minimum optimization problem by adding a negative sign to the value, as stated by Equation (3) [40].
d i s t a n c e ( p i , p i + 1 ) = ( x i x i + 1 ) 2 + ( y i y i + 1 ) 2
l e n g t h ( p ) = i = 0 n d i s t a n c e ( p i , p i + 1 )
p a t h ( p ) = min m i n D i s t a n c e ( p i p i + 1 , O j )
The following step is to develop path generators based on ML using an end-to-end approach. Regarding this study, [41], we have presented two methods for developing a path-planning generator. The first method analyzes the input sample based on the close distance between the centroid and each cluster level until a sample with the minimum distance is obtained, as shown in Figure 8a). Figure 8b) describes the second approach, which is an autoencoder composed of two types of networks: a DCNN and a recurrent neural network (RNN).
Since the second method is an autoencoder, the encoder made of DCNN has the features shown in Figure 9 to extract a characteristic vector, and the RNN is responsible for producing a series of points in 3D space. Moreover, a vocabulary is necessary because it was inspired by the picture caption algorithm [42]. The vocabulary consists of discrete samples of the three-dimensional space, and the 2-meter space is split into 20 cm for each step. As a result, the vocabulary decreases the number of discrete-sized samples on each side. Once discrete samples have been gathered, each trajectory calculates the frequency of each node. In this instance, at least 1000 potential values were reduced to 183. Consider that the total number of samples depends on the size of the discrete sample and that the samples generated vary based on the RRT algorithm’s generated pathways.
According to [43], a real-time system generates at least ten frames per second. In order to offer a real-time experience, the architecture must be simplified to be deployed on a mobile device. The transfer learning approach replaces a complex design with fewer features to solve the same problem [44]. As demonstrated in Figure 10, the GAN architecture with HC is simplified to a DCNN since a smartphone can execute this type of model without any difficulty.
The methodology is summarized in Figure 11, which depicts the implementation of a lengthy process in an embedded device using unpaired samples from two environments. The paths provided by the RRT algorithm in a known environment are described in Figure 11a. Figure 11b depicts how the following step constructs the path using HC and autoencoder. Likewise, Figure 11c uses the GAN style generator to convert the characteristics across two domains employing the HC approach to approximate a known path. In order to use the design on embedded devices, Figure 11d implements the transfer learning approach, which describes the 200 samples of the optimized architecture saved in an unknown environment. Figure 11e illustrates the relationship between an unfamiliar and known path for unpaired samples. In addition, an augmented reality system for smartphones is proposed to display the behavior of the development of paths, which generates a path in real time.
The performance of this proposal is evaluated based on the behavior in creating an expected vector by a machine learning model and the collision-free coefficient that characterizes the viability of the generated path.

4. Experimental Phase and Analysis

The proposed architecture was implemented in a g4dn.xlarge instance in Amazon Web Services (AWS) with the following specifications: 4 VCPU XEON 8259CL 2.5 GHz, 16 GB RAM, 125 GB SDD storage, with NVIDIA Tesla T4 GPU with 320 Tensor Core with 16 GB RAM. The introduced architecture was implemented in Tensorflow 2.4. Training time was 1 h and 10 min. The implementation on the smartphone employs Tensorflow-lite. Along with the experimental phase, 300 samples were used for the GAN architecture; finally, 50 unknown samples were used for the experiment. There are three distinct categories of physical items in the surrounding area. Two of these things are chairs of various colors and sizes, while the last object is a table. Microsoft Kinect V1 has a minimum perception range of 40 cm when configured for a 4 m range [45]. Therefore, things must be larger than 40 cm to be perceived at a range of up to 4 m. Consequently, we have both flat and curved surfaces to measure data consistently.
The RRT algorithm returns a vector representing each sample’s path. The following experiment compares the autoencoder’s and HC’s performance between known and created vectors. The behavior is based on the Euclidean distance (Equation (4)), the Manhattan distance (Equation (5)), and the cosine similarity (Equation (6)) used to examine the difference between the predicted vector x and the generated vector y. The free collision coefficient of Equation (7) quantifies whether or not at least one node generates an inadequate path.
e u c l i d e a n = i = 0 k ( x i y i ) 2
m a n h a t t a n = i = 0 k ( x i y i )
c o s i n e s i m i l a r i t y = x · y x y
C f r e e c o l l i s i o n = 1 i = 1 N s a m p l e s i f e x i s t s c o l l i s i o n c = 1 e l s e c = 0 N s a m p l e s
The experiment assesses the behavior in four contexts to develop a path. Furthermore, the suggested method consists of the original model and the transfer learning method, with 50 samples for each model. Comparing the created and predicted vectors, Table 3 illustrates the behavior of path generation. A proposed collision-free coefficient describes the behavior because the coefficient decreases when a possible collision is likely in the path. In addition, distance-measuring devices such as ARcore and the Kinect sensor have been used as references. It is crucial to mention that ARcore and Kinect sensor measurements were only conducted once.
According to the statistics, the model with the best performance is the HC with transfer learning. Given the HC approach’s characteristics, it approximates a near-optimal solution because the training samples are sufficient for the scenario features. However, this algorithm lacks efficiency because more samples are required when the scenario expands. On the other hand, it has been observed that the use of TL reduces characteristic vector resolution for each sample because the vector is normalized, and the word size change directly influences path generation. The autoencoder with TL exhibits this behavior because a fluctuation in the model’s word size is another value that causes the error to grow; an alternative option is to increase the number of samples when performing inference in the TL model. ARcore and Kinect sensor offers the best performance, but integrating these technologies is challenging because a path must be established online and requires additional considerations, for example, previous environment exploration for ARcore and located objects in at least 40cm for the Kinect sensor.
Since HC has the best performance in this experiment, it has been implemented on a Moto X4 smartphone with a Qualcomm 630 processor. In addition, the proposal employs the PointCloud function of ARCore to determine when the input data are updated. Thus, the path is formed when the features of the current image are relevant to the new image, obtaining a maximum sampling rate of 30 frames on the test device. Figure 12 illustrates a path that depicts a 2-meter-long path devoid of collisions. Observe that a route is provided to avoid obstacles in the environment. The test evaluation was conducted in two unknown environments using an AR tool. The path avoids physical elements, implementing a domain change with a transfer style.
Once the systems display virtual elements, a user experiment describes the performance employing a virtual coin route to guide the users to evaluate the interaction with the physical world. Figure 13a describes an unsafe distance between barriers that could result in an accident. Figure 13b depicts the safe location for virtual objects to avoid traversing the obstruction. Real-time deployment of virtual elements in a physical environment is depicted in Figure 14. The position for each coin that composes a virtual path prevents potential collisions.
According to [46], there are three important factors to consider when evaluating the performance of a 3D application: the participation of representative users, the environment of the evaluation, and the sorts of findings generated. The evaluation is based on the user’s interaction with the proposed system. Eight users conducted four system evaluations in this investigation and monitored user behavior. Table 4 details the number of potential collisions between each user and the obstruction.
Over time, each user adjusts to the system based on the outcomes of their user experience. Once the user has used the system for the first time, the user typically mistrusts the location of the virtual elements and maintains a keen awareness of the real world. In other words, the user looks over the mobile device out of concern for a potential collision. The experiment demonstrates how individuals adapted to a system with a high number of potential collisions; each user had at least three collisions, although in some cases, their fear of colliding through the system decreased to zero or one. Consequently, the user focuses on the device’s screen, reducing the fear of colliding with an obstruction, and the system enhances the safety of traveling in a controlled environment.
Therefore, this study presents an alternate technique for providing a path generator for displaying virtual elements with minimal information, which can be implemented in uncertain environments. Likewise, a novel alternative has been established for deriving a general solution from restricted information to avoid specific data information in the environment.

5. Conclusions and Future Work

This research presents a way to resolve the problem of building unpaired datasets for indoor exploration using augmented reality applications on limited devices using the transfer style. This proposal provides an alternative to expanding the limits of specific datasets to specific solutions. Therefore, the transfer style enables connection to unknown domains based on a known environment, which contains the generated paths, and the GAN approximates a potential solution according to sharing features. Furthermore, it is crucial to note that this method delivers a real-time understanding of the physical world because the virtual elements are displayed in a safe location. Due to this behavior, this approach expands the potential implementations of GANs.
Although technologies such as Arcore and Kinect sensors have been utilized to evaluate the possibility of collisions, these technologies require the online execution of a path planner generator. Consequently, the development and implementation time is costly, but the precision is superior. However, this proposal helps reduce the number of external sensors and avoids knowing the environment in advance, two of the most significant limitations of the technologies, as mentioned earlier. In addition, the experience is real-time, and the most innovative aspect is the execution on constrained devices such as smartphones.
According to user experience, the user improves with time while learning to concentrate on the immersive experience without relying on prior knowledge of the environment. However, additional concerns have emerged, such as the minimum number of available samples required to ensure optimal behavior and how this type of solution compares to physical sensors. On the other hand, the concept of meta-learning could supplement this work by enabling the creation of a generic solution for scenarios that share similar features without the need for several samples of the same objects.
The objective of future work will be to address the challenges caused by the introduced work and to supplement this proposal with other methods, such as meta-learning. Meta-learning might be applied to fine-tune potential global solutions whose characteristics are adequate to facilitate the current disadvantages of the present effort.

Author Contributions

Conceptualization, J.M.-R., A.M.-R. and M.A.-P.; software, J.M.-R. and A.M.-R.; validation, M.A.-P.; formal analysis, J.M.-R., A.M.-R. and M.A.-P.; investigation, J.M.-R. and A.M.-R.; resources, M.A.-P.; writing–original draft preparation, J.M.-R., A.M.-R. and M.A.-P.; writing–review and editing, J.M.-R. and M.A.-P.; visualization, J.M.-R., A.M.-R. and M.A.-P.; supervision, J.M.-R. and M.A.-P. and project administration, J.M.-R. and M.A.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported in part by the Secretaría de Investigación y Posgrado (SIP) and in part by the Comisión de Operación y Fomento de Actividades Académicas (COFAA) of the Instituto Politécnico Nacional (IPN).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, K.; Billinghurst, M.; Bruder, G.; Duh, H.B.L.; Welch, G.F. Revisiting Trends in Augmented Reality Research: A Review of the 2nd Decade of ISMAR (2008–2017). IEEE Trans. Vis. Comput. Graph. 2018, 24, 2947–2962. [Google Scholar] [CrossRef]
  2. Hernández, J.D.; Sobti, S.; Sciola, A.; Moll, M.; Kavraki, L.E. Increasing Robot Autonomy via Motion Planning and an Augmented Reality Interface. IEEE Robot. Autom. Lett. 2020, 5, 1017–1023. [Google Scholar] [CrossRef]
  3. Kularbphettong, K.; Vichivanives, R.; Roonrakwit, P. Student Learning Achievement through Augmented Reality in Science Subjects. In Proceedings of the 2019 11th International Conference on Education Technology and Computers, Amsterdam, Netherlands, 28–31 October 2019; Association for Computing Machinery: New York, NY, USA, 2019. ICETC 2019. pp. 228–232. [Google Scholar] [CrossRef]
  4. Serrano Vergel, R.; Morillo Tena, P.; Casas Yrurzum, S.; Cruz-Neira, C. A Comparative Evaluation of a Virtual Reality Table and a HoloLens-Based Augmented Reality System for Anatomy Training. IEEE Trans. Hum.-Mach. Syst. 2020, 50, 337–348. [Google Scholar] [CrossRef]
  5. Qian, L.; Wu, J.Y.; DiMaio, S.P.; Navab, N.; Kazanzides, P. A Review of Augmented Reality in Robotic-Assisted Surgery. IEEE Trans. Med. Robot. Bionics 2020, 2, 1–16. [Google Scholar] [CrossRef]
  6. Zhou, F.; Duh, H.B.L.; Billinghurst, M. Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR. In Proceedings of the 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, 15–18 September 2008; pp. 193–202. [Google Scholar] [CrossRef] [Green Version]
  7. Kent, L.; Snider, C.; Gopsill, J.; Hicks, B. Mixed reality in design prototyping: A systematic review. Des. Stud. 2021, 77, 101046. [Google Scholar] [CrossRef]
  8. Speicher, M.; Hall, B.D.; Nebeling, M. What is Mixed Reality? In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI ’19), Glasgow, UK, 4–9 May 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1–15. [Google Scholar] [CrossRef]
  9. Lee, L.H.; Braud, T.; Hosio, S.; Hui, P. Towards Augmented Reality Driven Human-City Interaction: Current Research on Mobile Headsets and Future Challenges. ACM Comput. Surv. 2021, 54, 1–38. [Google Scholar] [CrossRef]
  10. Costanza, E.; Kunz, A.; Fjeld, M. Mixed Reality: A Survey. In Human Machine Interaction: Research Results of the MMI Program; Lalanne, D., Kohlas, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 47–68. [Google Scholar] [CrossRef]
  11. Evans, D. How the Next Evolution of the Internet Is Changing Everything—Networking, Cloud, and Cybersecurity Solutions, CISCO. 2011. Available online: https://www.cisco.com/c/dam/en_us/about/ac79/docs/innov/IoT_IBSG_0411FINAL.pdf (accessed on 2 June 2022).
  12. Blanco-Novoa, O.; Fernandez-Carames, T.M.; Fraga-Lamas, P.; Vilar-Montesinos, M.A. A Practical Evaluation of Commercial Industrial Augmented Reality Systems in an Industry 4.0 Shipyard. IEEE Access 2018, 6, 8201–8218. [Google Scholar] [CrossRef]
  13. Nowacki, P.; Woda, M. Capabilities of ARCore and ARKit Platforms for AR/VR Applications. In Engineering in Dependability of Computer Systems and Networks. DepCoS-RELCOMEX 2019. Advances in Intelligent Systems and Computing; Zamojski, W., Mazurkiewicz, J., Sugier, J., Walkowiak, T., Kacprzyk, J., Eds.; Springer: Cham, Germany, 2019; pp. 358–370. [Google Scholar] [CrossRef]
  14. Wang, A.I. Systematic literature review on health effects of playing Pokémon Go. Entertain. Comput. 2021, 38, 100411. [Google Scholar] [CrossRef]
  15. Richards, K.; Wong, K.; Khan, M. Augmented reality game related injuries. New Horizons Clin. Case Rep. 2017, 2, 27. [Google Scholar] [CrossRef] [Green Version]
  16. Mao, R.; Gao, H.; Guo, L. A novel collision-free navigation approach for multiple nonholonomic robots based on Orca and linear MPC. Math. Probl. Eng. 2020, 2020, 1–16. [Google Scholar] [CrossRef]
  17. Ajani, T.S.; Imoize, A.L.; Atayero, A.A. An Overview of Machine Learning within Embedded and Mobile Devices–Optimizations and Applications. Sensors 2021, 21, 4412. [Google Scholar] [CrossRef] [PubMed]
  18. Hasinoff, S.W.; Sharlet, D.; Geiss, R.; Adams, A.; Barron, J.T.; Kainz, F.; Chen, J.; Levoy, M. Burst Photography for High Dynamic Range and Low-Light Imaging on Mobile Cameras. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
  19. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  20. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Advances in Neural Information Processing Systems; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Montreal, Quebec, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  21. Tripathy, S.; Kannala, J.; Rahtu, E. Learning Image-to-Image Translation Using Paired and Unpaired Training Samples; Springer: Berlin/Heidelberg, Germany, 2019; pp. 51–66. [Google Scholar] [CrossRef] [Green Version]
  22. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar] [CrossRef]
  23. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks; IEEE International Conference on Computer Vision (ICCV); Venice, Italy, 22–29 October 2017; IEEE: New York, NY, USA, 2017; pp. 2242–2251. [Google Scholar] [CrossRef] [Green Version]
  24. Leven, D.; Sharir, M. An Efficient and Simple Motion Planning Algorithm for a Ladder Moving in Two-Dimensional Space amidst Polygonal Barriers (Extended Abstract). In Proceedings of the First Annual Symposium on Computational Geometry (SCG ’85), Baltimore, Maryland, USA, 5–7 June 1985; Association for Computing Machinery: New York, NY, USA, 1985; pp. 221–227. [Google Scholar] [CrossRef]
  25. Schwartz, J.T.; Sharir, M. On the “piano movers” problem. II. General techniques for computing topological properties of real algebraic manifolds. Adv. Appl. Math. 1983, 4, 298–351. [Google Scholar] [CrossRef] [Green Version]
  26. Hu, X.; Tang, B.; Chen, L.; Song, S.; Tong, X. Learning a Deep Cascaded Neural Network for Multiple Motion Commands Prediction in Autonomous Driving. IEEE Trans. Intell. Transp. Syst. 2021, 22, 7585–7596. [Google Scholar] [CrossRef]
  27. Shah, S.; Dey, D.; Lovett, C.; Kapoor, A. AirSim: High-Fidelity Visual and Physical Simulation for Autonomous Vehicles. In Proceedings of the FSR, First Online, 3 November 2017. [Google Scholar]
  28. Shamwell, E.J.; Lindgren, K.; Leung, S.; Nothwang, W.D. Unsupervised Deep Visual-Inertial Odometry with Online Error Correction for RGB-D Imagery. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2478–2493. [Google Scholar] [CrossRef]
  29. Mur-Artal, R.; Montiel, J.M.M.; Tardós, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  30. Luo, S.; Guo, T. Image-to-Image Transfer Makes Chaos to Order. In Proceedings of the 2021 International Conference on Multimedia Retrieval (ICMR ’21), Taipei, Taiwan, 21–24 August 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 236–243. [Google Scholar] [CrossRef]
  31. Wadhwa, N.; Garg, R.; Jacobs, D.E.; Feldman, B.E.; Kanazawa, N.; Carroll, R.; Movshovitz-Attias, Y.; Barron, J.T.; Pritch, Y.; Levoy, M. Synthetic depth-of-field with a single-camera mobile phone. ACM Trans. Graph. 2018, 37, 1–13. [Google Scholar] [CrossRef] [Green Version]
  32. Maldonado-Romo, J.; Aldape-Pérez, M.; Rodríguez-Molina, A. Analysis of Depth and Semantic 373 Mask for Perceiving a Physical Environment Using Virtual Samples Generated by a GAN. IEEE Access 2022, 10, 5595–5607. [Google Scholar] [CrossRef]
  33. Wu, T.; Luo, A.; Huang, R.; Cheng, H.; Zhao, Y. End-to-End Driving Model for Steering Control of Autonomous Vehicles with Future Spatiotemporal Features. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 950–955. [Google Scholar] [CrossRef]
  34. Chen, X.; Xu, C.; Yang, X.; Song, L.; Tao, D. Gated-GAN: Adversarial Gated Networks for Multi-Collection Style Transfer. IEEE Trans. Image Process. 2019, 28, 546–560. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Gao, X.; Tian, Y.; Qi, Z. RPD-GAN: Learning to Draw Realistic Paintings With Generative Adversarial Network. IEEE Trans. Image Process. 2020, 29, 8706–8720. [Google Scholar] [CrossRef]
  36. Mignotte, M. Segmentation by Fusion of Histogram-Based K-Means Clusters in Different Color Spaces. IEEE Trans. Image Process. 2008, 17, 780–787. [Google Scholar] [CrossRef] [Green Version]
  37. Huang, W.; Ribeiro, A. Hierarchical Clustering Given Confidence Intervals of Metric Distances. IEEE Trans. Signal Process. 2018, 66, 2600–2615. [Google Scholar] [CrossRef]
  38. Wang, W.; Gao, H.; Yi, Q.; Zheng, K.; Gu, T. An Improved RRT* Path Planning Algorithm for Service Robot. In Proceedings of the 2020 IEEE 4th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chongqing, China, 12–14 June 2020; Volume 1, pp. 1824–1828. [Google Scholar] [CrossRef]
  39. Zabarankin, M.; Uryasev, S.; Murphey, R. Aircraft routing under the risk of detection. Nav. Res. Logist. (NRL) 2006, 53, 728–747. [Google Scholar] [CrossRef]
  40. Xue, Y.; Sun, J.Q. Solving the Path Planning Problem in Mobile Robotics with the Multi-Objective Evolutionary Algorithm. Appl. Sci. 2018, 8, 1425. [Google Scholar] [CrossRef] [Green Version]
  41. Maldonado-Romo, J.; Aldape-Pérez, M.; Rodríguez-Molina, A. Path planning generator with metadata through a domain change by Gan between physical and Virtual Environments. Sensors 2021, 21, 7667. [Google Scholar] [CrossRef] [PubMed]
  42. Poghosyan, A.; Sarukhanyan, H. Short-term memory with read-only unit in neural image caption generator. In Proceedings of the 2017 Computer Science and Information Technologies (CSIT), Lviv, Ukraine, 5–8 September 2017; pp. 162–167. [Google Scholar] [CrossRef]
  43. Handa, A.; Newcombe, R.A.; Angeli, A.; Davison, A.J. Real-Time Camera Tracking: When is High Frame-Rate Best? In Proceedings of the Computer Vision—ECCV 2012, Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 222–235. [Google Scholar]
  44. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
  45. Zainuddin, N.; Mustafah, Y.; Shawgi, Y.; Rashid, N. Autonomous Navigation of Mobile Robot Using Kinect Sensor. In Proceedings of the 2014 International Conference on Computer and Communication Engineering, Kuala Lumpur, Malaysia, 23–25 September 2014; pp. 28–31. [Google Scholar] [CrossRef]
  46. Bowman, D.A.; Kruijff, E.; LaViola, J.J.; Poupyrev, I. 3D User Interfaces: Theory and Practice; Addison Wesley Longman Publishing Co., Inc.: Redwood City, CA, USA, 2004. [Google Scholar]
Figure 1. Displays of virtual elements on flat surfaces using the Pokemon Go game. The virtual elements are displayed through obstacles within the scenario that could cause accidents because the user requires interaction with the virtual objects.
Figure 1. Displays of virtual elements on flat surfaces using the Pokemon Go game. The virtual elements are displayed through obstacles within the scenario that could cause accidents because the user requires interaction with the virtual objects.
Sensors 22 09411 g001
Figure 2. Path generator architectures. (a) Conventional architecture. (b) Reduced architecture.
Figure 2. Path generator architectures. (a) Conventional architecture. (b) Reduced architecture.
Sensors 22 09411 g002
Figure 3. Type of samples for the known paired domain and the unknown domain unpaired. The paired samples have a 2D image and depth data to calculate the occupied obstacles. On the other hand, unpaired samples have two principal samples for indistinct environments.
Figure 3. Type of samples for the known paired domain and the unknown domain unpaired. The paired samples have a 2D image and depth data to calculate the occupied obstacles. On the other hand, unpaired samples have two principal samples for indistinct environments.
Sensors 22 09411 g003
Figure 4. Unknown environments with different features, such as illumination, texture, and color. However, the environments involve furniture as a principal element. (a) It is an Android environment emulator, and the unreal store downloaded (b,c).
Figure 4. Unknown environments with different features, such as illumination, texture, and color. However, the environments involve furniture as a principal element. (a) It is an Android environment emulator, and the unreal store downloaded (b,c).
Sensors 22 09411 g004
Figure 5. Features of the DCNN model to describe samples in a 3D space.
Figure 5. Features of the DCNN model to describe samples in a 3D space.
Sensors 22 09411 g005
Figure 6. Samples plotted in a 3D space. (a) A DCNN plots a sample in 3D space. (b) Implementation of the transfer style approach. (c) Three-dimensional description of the distribution of samples between different domains. (d) Graphical description of the samples with transfer style.
Figure 6. Samples plotted in a 3D space. (a) A DCNN plots a sample in 3D space. (b) Implementation of the transfer style approach. (c) Three-dimensional description of the distribution of samples between different domains. (d) Graphical description of the samples with transfer style.
Sensors 22 09411 g006
Figure 7. Change in domains to achieve a path in a known scenario. In the first stage (X), the sample from the unknown domain is translated to the known domain (Y). Subsequently, the samples from the known domain are approximated to generate the path (Y’).
Figure 7. Change in domains to achieve a path in a known scenario. In the first stage (X), the sample from the unknown domain is translated to the known domain (Y). Subsequently, the samples from the known domain are approximated to generate the path (Y’).
Sensors 22 09411 g007
Figure 8. Path generator approaches. (a) A hierarchical cluster is composed of the centroid of each set to avoid analyzing whole samples. (b) Autoencoder comprises a deep convolutional neuronal network as the encoder and a recurrent neuronal network as the decoder.
Figure 8. Path generator approaches. (a) A hierarchical cluster is composed of the centroid of each set to avoid analyzing whole samples. (b) Autoencoder comprises a deep convolutional neuronal network as the encoder and a recurrent neuronal network as the decoder.
Sensors 22 09411 g008
Figure 9. Features of the DCNN model to extract characteristic vector.
Figure 9. Features of the DCNN model to extract characteristic vector.
Sensors 22 09411 g009
Figure 10. Transfer learning reduces the number of operations since a complex problem with many features is replaced by a model with a smaller number of features.
Figure 10. Transfer learning reduces the number of operations since a complex problem with many features is replaced by a model with a smaller number of features.
Sensors 22 09411 g010
Figure 11. Proposed architecture to generate a path in augmented reality applications on embedded devices employing unpaired samples. (a) Virtual path associated with each virtual sample. (b) Three machine learning approaches to generate a virtual path. (c) Generative adversarial network to connect unknown samples with the known environment. (d) Transfer learning to reduce the architecture to employ embedded devices. (e) The connection between unknown samples with a virtual environment generates a path with free collisions.
Figure 11. Proposed architecture to generate a path in augmented reality applications on embedded devices employing unpaired samples. (a) Virtual path associated with each virtual sample. (b) Three machine learning approaches to generate a virtual path. (c) Generative adversarial network to connect unknown samples with the known environment. (d) Transfer learning to reduce the architecture to employ embedded devices. (e) The connection between unknown samples with a virtual environment generates a path with free collisions.
Sensors 22 09411 g011
Figure 12. The path generated for augmented reality experiences in real-time. A line indicates the path to display virtual elements to avoid physical objects.
Figure 12. The path generated for augmented reality experiences in real-time. A line indicates the path to display virtual elements to avoid physical objects.
Sensors 22 09411 g012
Figure 13. Location of virtual elements. (a) Unsafe location. (b) Safe location.
Figure 13. Location of virtual elements. (a) Unsafe location. (b) Safe location.
Sensors 22 09411 g013
Figure 14. Virtual elements on the path to avoid obstacles in an unknown environment.
Figure 14. Virtual elements on the path to avoid obstacles in an unknown environment.
Sensors 22 09411 g014
Table 1. Description of each domain employing average and dominate color.
Table 1. Description of each domain employing average and dominate color.
Average ColorDominate Color
MeanSTDMeanSTD
Known environment110.9818, 110.6834,125.324411.1791,11.2014, 6.3755114.2768, 171.9477, 170.842518.5680, 12.0359, 8.7037
Transfer features84.0189, 84.2476, 83.50525.0146, 5.4588, 5.511711.9112, 12.0532, 11.97453.3174, 3.4040, 3.5008
119.9242, 125.1150, 123.813412.2781, 6.5528, 7.4241179.4007, 214.4179, 193.915248.8878, 7.6791, 13.3808
128.9802, 143.6930, 140.151113.6080, 10.2819, 9.3439174.8663, 208.5393, 175.389814.4137, 9.6241, 11.1316
Normal features120.8573, 120.9270, 120.845312.9963, 12.9853, 12.986883.6730, 83.6655, 83.619016.7216, 16.6732, 16.5330
110.5359, 97.0221, 112.803824.3627, 28.4992, 6.5405109.7473, 97.8742, 196.922127.4893, 28.5040, 14.3613
163.2307, 149.6583, 184.57554.9271, 6.2114, 2.9584231.9854, 207.9542, 217.79504.0251, 4.4566, 6.5562
Table 2. Features description for each unknown domain.
Table 2. Features description for each unknown domain.
DomainTransfer FeaturesNormal Features
CentroidDistanceCentroidDistance
Known0.1565, 0.3084, −0.6504-0.1565, 0.3084, −0.6504-
Unknown A0.1741, 0.3108, −0.59180.06120.1934, 0.3004, −0.36970.2832
Unknown B0.1549, 0.3215, −0.65290.01340.1553, 0.3260, −0.52490.1267
Unknown C0.1365, 0.3254, −0.68110.04030.1410, 0.3047, −0.57730.0748
Table 3. Performance of each model to generate a path with 50 evaluations, where (↓) Low is better and ↑ Up is better.
Table 3. Performance of each model to generate a path with 50 evaluations, where (↓) Low is better and ↑ Up is better.
EnvironmentModelAccuracy
Euclidean
Distance
(Mean-Sdt)
(↓)
Accuracy
Manhattan
Distance
(Mean-Sdt)
(↓)
Accuracy
Cosine
Similarity
(Mean-Sdt)
(↑)
Free
Collision
Coefficient
(↑)
AAutoencoder2.4481 ± 0.74584.9868 ± 1.63540.7896 ± 0.08960.78
HC1.3789 ± 0.34893.4579 ± 0.89780.8928 ± 0.04800.88
TL-Autoencoder7.7589 ± 1.789511.4587 ± 2.47560.5478 ± 0.07890.52
TL-HC2.1458 ± 0.74583.2458 ± 0.87560.8878 ± 0.04780.86
ARcore0.82531.32680.86470.92
Kinect0.51460.99840.89640.96
BAutoencoder3.4656 ± 1.41835.0145 ± 0.93150.7265 ± 0.10180.79
HC1.8425 ± 0.57922.9854 ± 0.74190.8472 ± 0.04780.86
TL-Autoencoder5.2947 ± 1.24159.6542 ± 1.78920.6217 ± 0.14790.59
TL-HC1.9475 ± 0.45314.0154 ± 0.65480.8934 ± 0.12560.89
ARcore0.92052.02680.79080.88
Kinect0.72511.13560.89550.94
CAutoencoder2.1946 ± 0.60263.7146 ± 0.79870.8476 ± 0.17960.82
HC0.9987 ± 1.01472.5892 ± 0.97350.8956 ± 0.01450.90
TL-Autoencoder5.8624 ± 1.975210.2305 ± 1.01740.6204 ± 0.04970.59
TL-HC2.2580 ± 0.03254.0563 ± 0.65200.8942 ± 0.00410.88
ARcore0.95271.1590.88720.96
Kinect0.87451.37460.90040.96
DAutoencoder1.9524 ± 0.84563.2168 ± 0.54280.8128 ± 0.15580.84
HC2.0415 ± 0.54784.0127 ± 0.78120.8872 ± 0.01470.92
TL-Autoencoder4.4527 ± 1.23059.4762 ± 1.98560.5856 ± 0.05860.62
TL-HC2.7586 ± 0.25695.9682 ± 0.43680.9028 ± 0.18680.88
ARcore0.88981.57430.92460.98
Kinect0.99751.71040.91650.96
Table 4. Description of the performance with users to avoid possible collisions in a physical environment.
Table 4. Description of the performance with users to avoid possible collisions in a physical environment.
ExperimentUser 1User 2User 3User 4User 5User 6User 7User 8
146544375
243443143
321211021
401100001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Maldonado-Romo, J.; Maldonado-Romo, A.; Aldape-Pérez, M. Path Generator with Unpaired Samples Employing Generative Adversarial Networks. Sensors 2022, 22, 9411. https://doi.org/10.3390/s22239411

AMA Style

Maldonado-Romo J, Maldonado-Romo A, Aldape-Pérez M. Path Generator with Unpaired Samples Employing Generative Adversarial Networks. Sensors. 2022; 22(23):9411. https://doi.org/10.3390/s22239411

Chicago/Turabian Style

Maldonado-Romo, Javier, Alberto Maldonado-Romo, and Mario Aldape-Pérez. 2022. "Path Generator with Unpaired Samples Employing Generative Adversarial Networks" Sensors 22, no. 23: 9411. https://doi.org/10.3390/s22239411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop