Next Article in Journal
A New Design of the Dual-Mode and Pure Longitudinal EMAT by Using a Radial-Flux-Focusing Magnet
Previous Article in Journal
Using Complementary Ensemble Empirical Mode Decomposition and Gated Recurrent Unit to Predict Landslide Displacements in Dam Reservoir
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-User Collaborative AR System for Industrial Applications

1
State Key Laboratory of Virtual Reality Technology and Systems, School of Computer Science and Engineering, Beihang University, Beijing 100191, China
2
Peng Cheng Laboratory, Shenzhen 518067, China
3
Qingdao Research Institute of Beihang University, Qingdao 266104, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1319; https://doi.org/10.3390/s22041319
Submission received: 6 January 2022 / Revised: 29 January 2022 / Accepted: 31 January 2022 / Published: 9 February 2022
(This article belongs to the Topic Augmented and Mixed Reality)

Abstract

:
Augmented reality (AR) applications are increasingly being used in various fields (e.g., design, maintenance, assembly, repair, training, etc.), as AR techniques help improve efficiency and reduce costs. Moreover, collaborative AR systems extend applicability, allowing for collaborative environments for different roles. In this paper, we propose a multi-user collaborative AR system (aptly called the “multi-user collaborative system”, or MUCSys); it is composed of three ends—MUCStudio, MUCView, and MUCServer. MUCStudio aims to construct industrial content with CAD model transformation, simplification, database update, marker design, scene editing, and exportation, while MUCView contains sensor data analysis, real-time localization, scene loading, annotation editing, and virtual–real rendering. MUCServer—as the bridge between MUCStudio and MUCView—presents collaborative and database services. To achieve this, we implemented the algorithms of local map establishment, global map registration, optimization, and network synchronization. The system provides AR services for diverse industrial processes via three collaborative ways—remote support, collaborative annotation, and editing. According to the system, applications for cutting machines were presented to improve efficiency and reduce costs, covering cutting head designs, production line sales, and cutting machine inspections. Finally, a user study was performed to prove the usage experience of the system.

1. Introduction

Augmented reality (AR) provides a view composed of the real physical world and digital virtual elements, covering a set of technologies, such as localization, rendering, scene understanding, etc. In recent years, AR has trended in various areas, including in the entertainment and medical fields [1,2,3,4,5,6,7,8].
Limitations in devices and algorithms have constrained AR applications up until a few years ago. Nowadays, there is no lack of devices, due to the popularization of mobile devices, such as smartphones and tablets, as well as different types of AR glasses. Moreover, due to technological developments, we have achieved accurate localization and map establishments in general scenes, resulting in stable virtual–real fusion performances. Release libraries, such as ARCore and ARKit, contribute to a variety of applications, further lowering the development threshold.
Pertaining to current research in the AR field—AR industrial systems are mainly applied toward design, maintenance, assembly, repair, training, and inspection processes, with the goal of improving performance and reducing costs and machine losses [9,10]. In the design stage, two application scenarios are the CAD product model and factory layout [11,12,13,14]. Regarding CAD, with AR techniques and mobile devices, the designer can visualize a 3D virtual target object directly superimposed on the real environment. Moreover, with sensor data, the motion of the virtual model is simulated. Based on virtual–real fusion, designers can modify the shape of the CAD model. In the factory layout, the complex manufacturing system always comprises various parts, such as robots, automated guided vehicles, pallet changers, conveyors, etc. To solve this problem, AR systems can arrange these elements in the actual factory, providing the visualization for making decisions.
Similar to the above request, the inspection process can be enhanced by AR technology. Industrial product creation is a complex procedure, it usually includes conception, design, and realization. It is necessary to check whether errors occur after product realization. To achieve this goal, the inspection procedure will involve an organized examination of a particular device. AR is regarded as a promising technology to set up the inspection process [15,16,17,18]. Another inspection application involves the patrol inspection of the running machine. With Internet of Things (IoT) support, the inspector can obtain real-time running information to judge the status of the target machine.
Other AR research fields cover maintenance, repair, and assembly, with the goal of reducing time, costs, and the error rate [19,20,21]. Operators would need to conduct continuous attention switches between the manual and device during some complex tasks, which may involve a high cognitive load. AR technology focuses on this mission by overlaying virtual information onto the real part, to help operators complete their tasks. Virtual information (commonly) comprehends 3D model animations, instruction audio, and textual labels aligned to the machines, to provide detailed operations of the next steps [22]. Moreover, AR applications can enable remote technicians to interactively support maintainers when AR aids are not sufficient. Training—closely associated with the maintenance, repair, and assembly tasks—is another application direction in the industry domain [23,24,25]. Instructors, teachers, and trainers tend to explore new methods to enhance learning efficiency and experience. AR techniques meet these demands and can present novel experiences in the learning process. Compared with traditional approaches, on the one hand, multimedia information enhances user interactions and increases users’ interests. On the other hand, various perilous or equipment-worn procedures could also be directly simulated by AR systems.
In addition to the above, collaborative AR systems further satisfy industrial demands by providing cooperation among operators [26,27,28]. In summary, as shown in Figure 1, we propose a multi-user collaborative AR system, called MUCSys, for industrial applications.
Compared to previous AR systems, our MUCSys has the following peculiarities: (1) the developed system can support various industrial processes, such as design, inspection, assembly, sales, etc.; (2) three collaborative AR modes are presented to bridge interactions among different roles. In detail, remote support provides remote assistance between experts and workers (e.g., salesmen, maintenance people, or after-sales staff) in actual scenes. Collaborative annotation offers asynchronous adding, deleting, or modifying operations for different workers. Meanwhile, multiple users can collectively complete target tasks (e.g., production layout) via collaborative editing in real-time.
The whole system is composed of MUCStudio in PC, and MUCServer and MUCView in mobile devices. In detail, MUCStudio aims for industrial scene construction, while MUCView mainly provides AR experiences for users. Moreover, interlinking MUCStudio and MUCView, MUCServer presents database and collaborative services. To prove the effectiveness of the proposed system, laser cutting machine applications containing designs, sales, and inspections are provided.
In summary, compared to prior works, our work provides the following contributions.
  • We propose a multi-user collaborative AR system for laser cutting machines, containing three collaborative modes—remote support, annotation, and editing. The system comprises MUCStudio for scene generation, MUCView for AR experience, and MUCServer for collaboration.
  • To implement collaborative services, we designed several algorithms. Firstly, the local map was established with additional marker and QR code inputs based on ORB-SLAM3. Secondly, we propose a map registration pipeline by combining ORB match and point cloud registration. Global optimization was performed to promote the relative transformation accuracy. Finally, the network synchronization was exploited to improve the user experience.
  • Based on the system, three applications for laser cutting machines were developed—cutting head design, production line sale, and cutting machine inspection.
  • A user study was conducted to demonstrate the user experience of the system.
The remainder of the paper is organized as follows: related works are summarized in Section 2. Section 3 illustrates the detailed system composition and implementation. Based on the system, Section 4 presents three application scenarios involving laser cutting machines—designs, sales, and inspections. Section 5 presents the results of the user study and Section 6 presents the conclusions.

2. Related Work

2.1. Design

In the design process, AR systems can identify and avoid design errors in the early stages to reduce time, costs, and the number of physical prototypes. The CAD product model and factory layout design are two common application scenarios.
In the context of the CAD product model, Jimeno et al. [11] implemented an AR system with low-cost computational elements that allowed customers to check the quality of the footwear model from an aesthetic perspective. Georgel et al. [29] developed zoom-and-pan tools within mixed views to solve undocumented discrepancies between the CAD model and the final object. Caruso et al. [30], aiming for interactions with virtual objects superposed in a real environment, developed an interactive AR system that integrates stereoscopic visualization and fog screen display technology. Mourtzis et al. [13], to support the customer integration in the design stage, presented a framework consisting of a network design tool with a smart search algorithm, and a mobile application based on AR technology. The BIM-AR system [27], by implementing marker-based AR, provides the ability to view, interact, and collaborate with 3D and 2D BIM data.
Besides the CAD product model, planning a suitable factory layout is also a challenge. Shariatzadeh et al. [14], focusing on rapid factory design and planning, and based on actual industrial demands, took different layouts into consideration to determine the main functionalities. Pentenrieder et al. [31] presented a complete system composed of the requirement analysis, developing process and realization. Based on the system, a concrete usage example for factory planning was declared. Kokkas et al. [12] performed layout planning of machinery with AR tools, aiming to evaluate the suggested layouts through non-measurable factors.
In the laser cutting machine applications of our system, both the CAD product model and layout design can be verified through SMCStudio and SMCView. Through SMCStudio, the user designs the production line layout or the model position relative to the real environment by the marker. Then, SMCView superposes the virtual model on the real environment. Moreover, multiple users can cooperate through the SMCServer. With sensor data supporting, SMCView also simulates the trajectory and drives the virtual part to provide more visualization information for designers.

2.2. Maintenance, Assembly, Repair, and Training

In maintenance, repair, assembly, and training processes, AR technology can efficiently help operators understand the procedure of the tasks by means of overlaying the virtual instruction information onto the real environment [24,32]. The AR-based manual has shown that it can reduce the time and costs, and improve performance.
An early attempt to support technicians in simple maintenance procedures involved the intent-based illustration system proposed by Feiner et al. [33]. Then, various AR applications in manufacturing activities, such as maintenance, repair, and assembly were developed and demonstrated [19,21,34]. The project STAR-MATE [35] realized one of the earliest multi-modal interactions with a virtual pointing device and voice-based commands. In other projects, researchers provided mobile AR experiences in performing procedures in an industrial environment (instead of staying at a stationary workbench) [31,36,37]. Hou et al. [20] configured a prototype animated AR system for assembly tasks that are commonly instructed by text documentation. Then, a series of experiments were conducted to prove the effectiveness of the system. Chen et al. [26] developed a location aware AR collaborative framework for FMM with the interaction between users and facilities.
AR-based industrial training tasks are always related to the repair, assembly, and maintenance processes, with the aim of improving learning efficiency and reducing device costs [24]. Various industrial domains used AR technologies in training and supporting tasks, such as aerospace [23,38], automotive [39], and industrial plants [40]. A flexible training framework for maintenance was proposed by Sanna et al. [25] to address the problem of creating AR content [41], which allowed instructors to generate maintenance procedures conveniently. Wang et al. [28] described a remote collaborative for training in the manufacturing industry, enabling a remote expert to train a local worker in a physical assembly task.
Through our MUCSys, it is convenient to construct maintenance, assembly, repair, and training scenarios. However, in our applications for cutting machines, these kinds of demands are slight, as workers receive meticulous training and need to pass an examination before entering the workplace. Therefore, applications about maintenance, assembly, repair, and training are not provided.

2.3. Inspection

After the design and production phases, the product is realized. On account of the whole process complexity, it is essential to check whether errors and differences occurred during the phases, which can be achieved with AR technology [16].
An AR-based reconfigurable framework for inspection was proposed by Ramakrishna et al. [17], prioritizing the checklist by detecting the parts with deep learning. The framework can be utilized in various applications, such as industrial maintenance, the health sector, and so on. Wasenmuller et al. [18], regarding the discrepancy check task in industrial AR, presented a new approach that consisted of two-step depth mapping, semi-automatic alignment, and a 3D discrepancy check. The approach showed the superior performance compared with the state-of-the-art 3D discrepancy check. Munoz et al. [15] proposed a novel AR-based user interface for quality control inspection of car body surface production lines, to reduce working stress and lift the ergonomics of workers.
The inspection pipeline introduced above mainly focuses on realizing inspections to ensure quality. In our paper, after analyzing the requirements of the laser cutting machine, we employed our system to the patrol inspection process, to judge if the device was under the normal operating conditions and to find the running problem in the early stage.

3. System Design

3.1. Overall Structure

As demonstrated in Figure 2, the whole system is composed of three parts, covering MUCServer, MUCStudio, and MUCView. In detail, the MUCServer mainly provides a database service and multi-user collaboration. Correspondingly, MUCStudio runs on a PC and focuses on database management and industrial content production. Users can update the database that contains 3D industrial models, materials, and user annotations by MUCStudio. Moreover, MUCStudio aims to construct an industrial scene for designs, sales, or other applications. The functions include model transformation, visualization, pose editing, marker design, and exportation. MUCView—accepting the exported scene by MUCStudio—renders virtual models superposed onto the real environment by AR technology. When multiple users perform collaborative works, MUCView provides three collaborative ways from the MUCServer to support cooperation among different workers.

3.2. MUCStudio

MUCStudio looks at “constructing” the industrial scene for the AR experience. On the one hand, an industrial scene always covers various products with different material warehouses, marking machines, transmissions, etc. Through MUCStudio, users can select their wanted products to generate the final scene. On the other hand, the designed virtual scene should be appropriately fused with the real environment. MUCStudio adopted the marker to solve this problem. By designing markers in virtual and real scenes, both coordinate systems are unified, achieving an ideal virtual–real fusion effect.
The interface of MUCStudio is illustrated in Figure 2, covering CAD model processing, database establishment, and scene creation. In detail, the CAD operation module implements a CAD format transformation and simplification. After the process, the lightweight model with common 3D formats (sty, obj, ply, etc.) is obtained to be visualized in MUCStudio. In the actual implementation, we exploited CAD Exchanger SDK to accomplish the task, which is a free library to develop fast and robust 3D applications. According to the demand, users can establish their model database in MUCServer. While creating the scene, users download the arbitrary model from the built database. With the model database, designers can construct their wanted scene to verify the CAD model. For each model, MUCStudio provides the editing function of its position, orientation, and scale, as shown at the left of Figure 2. Furthermore, marker creation and editing help users ensure the relative poses between the real environment and virtual models.

3.3. MUCView

The purpose of MUCView is to provide AR experiences for users. It is mainly composed of camera tracking, collision detection, object editing, data loader, annotation editing, and virtual object rendering.
The camera tracking and object editing parts are implemented by Google ARCore, enabling the mobile device to understand and track its position relative to the world. Based on the obtained feature points and camera pose, the virtual object is rendered in the real environment by Unity. During the collaborative mode, the pose of the virtual object is then transformed by a relative matrix from MUCServer. After rendering the virtual object, users can edit its position, orientation, and scale. For the convenience of moving objects, MUCView presents a grouping function to bind different models. Moreover, to void collision detection during layout design, MUCView gives a reminder when virtual objects have collisions. After determining the model position, the data loader module loads sensor data to drive the model to perform the motion. Annotation editing offers the user addition, modification, and deletion of the virtual annotation, which will be illustrated in the next subsection on patrol inspection.
Through the communication with MUCServer, MUCView achieves multiple collaborative operations covering remote support, collaborative editing, and annotation.

3.4. MUCServer

The MUCServer mainly performs database and collaborative services. The database covers CAD, 3D model, annotation, and scene information that can be updated by MUCStudio and MUCView. The collaborative service provides three ways to satisfy different demands. When users need the help of remote experts, remote support presents virtual–real video transmissions and shared interactions, demonstrated in Figure 3. The expert receives the AR video from the user and gives the guide to the AR video in real-time. Moreover, collaborative editing offers a cooperative environment for multiple users. Using layout adjustment as an example, after dividing the layout work, each user concentrates on his own part, which can improve efficiency. Collaborative annotation—different from the above—is not real-time. The annotation information is bound to the machine. Different operators can achieve cooperation through checking, adding, modifying, or deleting the annotation.
In the implementation, remote support technology relies on the video transmission realized by the FFmpeg library, while collaborative editing and annotation are based on the collaborative localization and map establishment among different devices illustrated in Figure 4. The whole pipeline consists of local map establishment, global map building, search, reuse, update, and global optimization. Based on map registration, we fused local maps to establish the global map. Then we searched the map database to perform map reuse and update. Meanwhile, because of connections between map and annotation, the previous annotations were also obtained. Finally, the evaluated pose and annotations were passed to the MUCView.
Local map. With the gathered image, IMU data, local camera pose by ARCore, feature point, detected marker, QR code, and annotation, the local map is established by ORB-SLAM3 [42], covering sparse point cloud and keyframe dataset. In particular, there are many similar views in the industrial scene, increasing the difficulty of collaborative localization. To solve this problem, we added QR codes and markers into the image match process and gave them more confidence. Similarly, in local optimization, feature points in QR codes and markers are also taken into consideration. In the actual implementation, each machine connected to IoT is always pasted with a QR code, so that user can obtain running information. The markers are always placed in spaces with less textures and corners, which contributes to localization and mapping.
Map registration. To unify the coordinate system among different local maps, we needed to calculate the relative transformation matrices. Global map building, search, reuse, and update are all based on the pipeline. The map registration firstly performs the ORB feature match among keyframes from different local maps. As with ORB-SLAM3, when enough feature correspondences were built, we then exploited global optimization to determine the final relative pose. Because of the device difference and changes of environments, the ORB feature match may fail. In this situation, we leveraged the point cloud registration algorithm instead.
In the registration pipeline, the key points with significant intrinsic shape characteristics were first extracted to reduce subsequent calculations. Then, the FPFH method was used to describe the characteristics of these points. By correspondences between two point clouds, the initial registered matrix was calculated to reduce the range of randomly selected interior points in RANSAC. Meanwhile, the pre-rejection method based on geometric constraints was also exploited to accelerate the coarse registration process. Finally, by using the approximate nearest neighbor ICP method, the fine registration matrix was obtained for map alignment.
Global optimization. A variety of SLAM systems utilize the BA process to reduce accumulative errors [42,43]. Multiple components covering camera poses and map points are jointly optimized by taking advantage of the traditional BA pipeline. In our application scenarios, feature points in makers and QR codes are considered with larger weights. Now, we provide the detailed formula of the whole BA. For convenience, feature points, camera poses, and marker points are denoted as P = P i , C = C i , M = M i ∑ respectively. Then our BA is expressed as the following nonlinear optimization problem,
C * , P * , O * = a r g m i n C i , P i , M i E ( P i , C i ) + α   E ( M i , C i ) ,
where E ( P i , C i ) , E ( M i , C i ) denote the camera-point and camera-maker errors. Both errors are denoted by the 3D point projection error in ORB SLAM3 [42] with the following form,
E ( P i , C i ) = π ( R c P i ) d ,
E ( M i , C i ) = π ( R c M i ) d m ,
where R c is the transformation matrix from world space to camera space, π represent the matrix from camera space to pixel space, d , d m denote pixel coordinates of 3D points.
Network synchronization. Besides the collaborative localization, the network synchronization method is also important in promoting the user experience. Based on the adaptive lockstep algorithm, we set an upper threshold for the synchronization interval on the server to prevent clients with poor networks from delaying other clients. In the network environment of each client, the synchronization interval was calculated in real-time, instead of using a fixed-length synchronization interval. This was to avoid further deterioration of network conditions caused by short intervals and to prevent too long intervals from making clients with good network conditions unable to obtain sufficient fluency. In addition, smooth interpolation of the local client, optimization of local user operation delay, and server lag compensation were also performed.

4. Applications in Cutting Machine

In this section, three applications are presented (by exploiting our proposed system, containing cutting head design, production line sale, and cutting machine inspection). The design process demonstrates the CAD model verification procedure by MUCStudio and MUCView. Sales and inspection depend on AR collaboration. The remote support and collaborative editing establish an environment for salesmen, customers, and experts, while a collaborative annotation provides cooperation among different inspectors.

4.1. Design

The usage in the design stage involves verifying the CAD model through virtual–real fusion and a data-driven simulation. As stated in Figure 5, using cutting head design as an example, after completing one version of the CAD model, the system can help discover the design problem through visualizing the 3D model and driving the model move by sensor data.
To achieve the goal, the designer employs MUCStudio to transform the cutting head CAD model to “obj” format for visualization. Then the relative position between the cutting head and marker is determined. Furthermore, the sensor data are bound to the cutting head model. Then the designer prints the marker and pastes it to the corresponding position as our placing in MUCStudio. Finally, to verify the design of the model, MUCView is exploited to render the virtual CAD model onto the real machine.

4.2. Sale

The demand for laser cutting machine sales is at the layout of the production line. In the actual sale process, based on the cutting demand and field condition of the customer, a salesman should determine the cutting machine type, its corresponding automatic feeder, marking machine, and device information. However, due to differences in the actual field, the selection and layout of the production line are time-consuming. Moreover, no visualization of the production line increases communication difficulty between the salesmen and customers.
The sales process is illustrated in Figure 3, by utilizing our system. Firstly, the salesman gathers user demand that contains field size, cutting materials, capacity, etc. To decide the cutting machine type, it is essential to analyze cutting efficiency and requirement. After ensuring the production line composition, MUCView generates the initial layout of the whole production line. Then, in the actual field, the salesman and customer exploit collaborative layout editing to determine the layout. In particular, the salesman can also apply for remote collaborative assistant by the expert when encountering a difficult case. The remote expert receives a virtual–real fusion video and makes simple marks to communicate with the salesman in real-time.

4.3. Patrol Inspection

The requests for the inspection process come from inspectors and operators who need to ensure machine running status. It is beneficial to detect problems in the early stage to reduce the costs of maintenance by a large margin. From IoT service, sensor data are transferred to the IoT server and checked by preset boundaries. However, the running problem is sometimes neglected by merely checking the sensor data, which need to be inspected by inspectors near the running machine.
As demonstrated in Figure 3, by our system, each running machine can be associated with a unique QR code, which is further related to the ID in the IoT server. After establishing the binding, the inspector can acquire running data from the IoT server in real-time. Combined with the actual machine running status, it is more accurate to detect running problems and then fix them. After inspection, the inspector can upload the annotation to explain the noticeable information. When another inspector checks the same machine, he can browse the annotation and modify it.

5. User Study

In this section, we conducted a user study to investigate the system usability and difficulty through setting up tasks about the design, sale, and inspection for participants.

5.1. Procedure

In our user study, we recruited 20 participants (2 females, 19 to 41 years old, mean = 27.7 , Std. Dev. = 6.58 ). All participants had over 2 years of experience using PCs, smart phones, or tablets (three participants had tried AR applications before).
At the start, each participant read the study information, answered a questionnaire, and signed the consent form. Then, volunteers distributed a prepared system user guide and task information to participants. After completing the reading, we performed a simple training process and Q & A about our system, covering the functions, explanations, usages, and tasks.
The user study was held in an actual factory, which had cutting machines and a large free space. The whole study consisted of three tasks that covered design, sales, and inspection application, introduced in the previous section. The detailed procedures of the three tasks are illustrated in Table 1. The procedures of the tasks were set at a time limit. When the participant did not complete the procedure within the time limit, it was regarded as a failure in this procedure.
To test the usability and difficulty of the study, we exploited the single ease question (SEQ) [44] to measure task difficulty, the subjective mental effort questionnaire (SMEQ) [45] to test the level of mental effort needed during the task, and the system usability scale (SUS) [46] to estimate usability.

5.2. Results

Table 2 reports the mean completion time and failure number of each step. On average, participants spent 12.6 min in the design task, 9.9 min in the sales, and 13.6 min in the inspection. Furthermore, participants only made a small number of errors. Among all participants, twelve persons did not make any errors. In all steps, process 2.3 took the most time and had the highest failure number. The reason is that it was the first time for most participants to operate the virtual object in a real environment. Moreover, operating the virtual model in 3D space with a 2D multi-touch screen also increased the task difficulty.
The detailed SUS [46] results are presented in Figure 6. The mean score of all participants was 75.4 with IQR [ 71.5 , 86.3 ] . From the SUS results, it is obvious that most participants had good user experiences using MUCStudio and MUCView to complete the tasks. For most participants, it was their first time using the AR system. The usability of our system is appreciable.
Regarding task difficulty, the median value of SEQ was 6, with IQR [5–7] (SEQ 1: very difficult; 7: very easy—the higher, the easier), while the median result of SMEQ was 12 with IQR [8–15] (0: not hard at all–150). It is obvious that both SEQ and SMEQ performances show that completing the tasks were not hard. In particular, all participants received a simple training process before completion, and most had not used an AR system before. In actual application, after systemic training, users could have better performance.

6. Conclusions

In this paper, focusing on industrial applications, we proposed a multi-user collaborative AR system composed of MUCStudio for content generation, MUCView for providing AR experiences, and MUCServer for database and collaborative services. On the one hand, the system could be applied to various industrial stages that contain designs, sales, inspections, etc. On the other hand, three collaborative ways are presented among the different roles, covering remote support, collaborative editing, and annotation. To achieve collaborative localization and improve user experience, we implemented the algorithms of local map establishment, global map registration, optimization, and network synchronization. Based on the proposed system, we develop three laser cutting machine applications, which include the cutting head design, production line sale, and patrol inspection of a cutting machine. Furthermore, to evaluate the system, we conducted a user study that covered three tasks through SEQ, SMEQ, and SUS, demonstrating use difficulty and user experience.

Author Contributions

Conceptualization, J.W. and Y.Q.; methodology, J.W. and Y.Q.; software, J.W.; validation, J.W. and Y.Q.; formal analysis, J.W.; investigation, J.W.; resources, Y.Q.; data curation, Y.Q.; writing—original draft preparation, J.W.; writing—review and editing, Y.Q.; visualization, J.W.; supervision, Y.Q.; project administration, Y.Q.; funding acquisition, Y.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the Key-Area Research and Development Program of Guangdong province (no. 2019B010150001), the National Natural Science Foundation of China (no. 62072020), the National Key Research and Development Program of China (no. 2017YFB1002602), and the Leading Talents in Innovation and Entrepreneurship of Qingdao (19-3-2-21-zhc).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank the anonymous reviewers for their valuable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fiorentino, M.; de Amicis, R.; Monno, G.; Stork, A. Spacedesign: A mixed reality workspace for aesthetic industrial design. In Proceedings of the International Symposium on Mixed and Augmented Reality, Darmstadt, Germany, 1 October 2002; pp. 86–318. [Google Scholar]
  2. Fite-Georgel, P. Is there a reality in industrial augmented reality? In Proceedings of the 2011 10th IEEE International Symposium on Mixed and Augmented Reality, Basel, Switzerland, 26–29 October 2011; pp. 201–210. [Google Scholar]
  3. Zhu, Z.; Branzoi, V.; Wolverton, M.; Murray, G.; Vitovitch, N.; Yarnall, L.; Acharya, G.; Samarasekera, S.; Kumar, R. AR-mentor: Augmented reality based mentoring system. In Proceedings of the 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Munich, Germany, 10–12 September 2014; pp. 17–22. [Google Scholar]
  4. Moreta-Martinez, R.; Pose-Diez-de-la Lastra, A.; Calvo-Haro, J.A.; Mediavilla-Santos, L.; Perez-Mananes, R.; Pascau, J. Combining Augmented Reality and 3D Printing to Improve Surgical Workflows in Orthopedic Oncology: Smartphone Application and Clinical Evaluation. Sensors 2021, 21, 1370. [Google Scholar] [CrossRef] [PubMed]
  5. Gao, Y.; Zhao, Y.; Xie, L.; Zheng, G. A Projector-Based Augmented Reality Navigation System for Computer-Assisted Surgery. Sensors 2021, 21, 2931. [Google Scholar] [CrossRef] [PubMed]
  6. Tsai, C.Y.; Lai, Y.C. Design and Validation of an Augmented Reality Teaching System for Primary Logic Programming Education. Sensors 2022, 22, 389. [Google Scholar] [CrossRef] [PubMed]
  7. Morar, A.; Balutoiu, M.A.; Moldoveanu, A.; Moldoveanu, F.; Butean, A. CultReal-A Rapid Development Platform for AR Cultural Spaces, with Fused Localization. Sensors 2021, 21, 6618. [Google Scholar] [CrossRef]
  8. Papakostas, C.; Troussas, C.; Krouska, A.; Sgouropoulou, C. Measuring User Experience, Usability and Interactivity of a Personalized Mobile Augmented Reality Training System. Sensors 2021, 21, 3888. [Google Scholar] [CrossRef]
  9. Schonfelder, R.; Schmalstieg, D. Augmented reality for industrial building acceptance. In Proceedings of the 2008 IEEE Virtual Reality Conference, Reno, NV, USA, 8–12 March 2008; pp. 83–90. [Google Scholar]
  10. Tumler, J.; Doil, F.; Mecke, R.; Paul, G.; Schenk, M.; Pfister, E.A.; Huckauf, A.; Bockelmann, I.; Roggentin, A. Mobile augmented reality in industrial applications: Approaches for solution of user-related issues. In Proceedings of the 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, 15–18 September 2008; pp. 87–90. [Google Scholar]
  11. Jimeno-Morenilla, A.; Sanchez-Romero, J.L.; Salas-Perez, F. Augmented and virtual reality techniques for footwear. Comput. Ind. 2013, 64, 1371–1382. [Google Scholar] [CrossRef] [Green Version]
  12. Kokkas, A.; Vosniakos, G.C. An Augmented Reality approach to factory layout design embedding operation simulation. Int. J. Interact. Des. Manuf. 2019, 13, 1061–1071. [Google Scholar] [CrossRef]
  13. Mourtzis, D.; Vlachou, E.; Giannoulis, C.; Siganakis, E.; Zogopoulos, V. Applications for frugal product customization and design of manufacturing networks. Procedia CIRP 2016, 52, 228–233. [Google Scholar] [CrossRef] [Green Version]
  14. Shariatzadeh, N.; Sivard, G.; Chen, D. Software evaluation criteria for rapid factory layout planning, design and simulation. Procedia CIRP 2012, 3, 299–304. [Google Scholar] [CrossRef] [Green Version]
  15. Munoz, A.; Mahiques, X.; Solanes, J.E.; Marti, A.; Gracia, L.; Tornero, J. Mixed reality-based user interface for quality control inspection of car body surfaces. J. Manuf. Syst. 2019, 53, 75–92. [Google Scholar] [CrossRef]
  16. Perla, R.; Gupta, G.; Hebbalaguppe, R.; Hassan, E. Inspectar: An augmented reality inspection framework for industry. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Merida, Mexico, 19–23 September 2016; pp. 355–356. [Google Scholar]
  17. Ramakrishna, P.; Hassan, E.; Hebbalaguppe, R.; Sharma, M.; Gupta, G.; Vig, L.; Sharma, G.; Shroff, G. An AR inspection framework: Feasibility study with multiple ar devices. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Merida, Mexico, 19–23 September 2016; pp. 221–226. [Google Scholar]
  18. Wasenmuller, O.; Meyer, M.; Stricker, D. Augmented reality 3d discrepancy check in industrial applications. In Proceedings of the 2016 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Merida, Mexico, 19–23 September 2016; pp. 125–134. [Google Scholar]
  19. Harmo, P.; Halme, A.; Virekoski, P.; Halinen, M.; Pitkanen, H. Etala-Virtual Reality Assisted Telepresence System for Remote Maintenance; IFAC: New York, NY, USA, 2000; pp. 1075–1080. [Google Scholar]
  20. Hou, L.; Wang, X.; Bernold, L.; Love, P.E. Using animated augmented reality to cognitively guide assembly. J. Comput. Civ. Eng. 2013, 27, 439–451. [Google Scholar] [CrossRef]
  21. Ong, S.; Yuan, M.; Nee, A. Augmented reality applications in manufacturing: A survey. Int. J. Prod. Res. 2008, 46, 2707–2742. [Google Scholar] [CrossRef]
  22. Papetti, A.; Ciccarelli, M.; Brunzini, A.; Germani, M. Investigating the Application of Augmented Reality to Support Wire Harness Activities. In Design Tools and Methods in Industrial Engineering II; Springer International Publishing: New York, NY, USA, 2022; pp. 116–124. [Google Scholar]
  23. Datcu, D.; Cidota, M.; Lukosch, S.; Oliveira, D.M.; Wolff, M. Virtual co-location to support remote assistance for inflight maintenance in ground training for space missions. In Proceedings of the 15th International Conference on Computer Systems and Technologies, Ruse, Bulgaria, 27–28 June 2014; pp. 134–141. [Google Scholar]
  24. Lorenz, M.; Knopp, S.; Klimant, P. Industrial augmented reality: Requirements for an augmented reality maintenance worker support system. In Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Munich, Germany, 16–20 October 2018; pp. 151–153. [Google Scholar]
  25. Sanna, A.; Manuri, F.; Piumatti, G.; Paravati, G.; Lamberti, F.; Pezzolla, P. A flexible AR-based training system for industrial maintenance. In Proceedings of the International Conference on Augmented and Virtual Reality, Lecce, Italy, 31 August–3 September 2015; pp. 314–331. [Google Scholar]
  26. Chen, K.; Chen, W.; Li, C.T.; Cheng, J.C. A BIM-based location aware AR collaborative framework for facility maintenance management. ITcon 2019, 24, 360–380. [Google Scholar]
  27. Garbett, J.; Hartley, T.; Heesom, D. A multi-user collaborative BIM-AR system to support design and construction. Autom. Constr. 2021, 122, 103487. [Google Scholar] [CrossRef]
  28. Wang, P.; Bai, X.; Billinghurst, M.; Zhang, S.; Han, D.; Lv, H.; He, W.; Yan, Y.; Zhang, X.; Min, H. An MR remote collaborative platform based on 3D CAD models for training in industry. In Proceedings of the 2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Beijing, China, 10–18 October 2019; pp. 91–92. [Google Scholar]
  29. Georgel, P.F.; Schroeder, P.; Navab, N. Navigation tools for viewing augmented cad models. IEEE Comput. Graph. Appl. 2009, 29, 65–73. [Google Scholar] [CrossRef] [PubMed]
  30. Caruso, G.; Carulli, M.; Bordegoni, M. Augmented reality system for the visualization and interaction with 3D digital models in a wide environment. Comput.-Aided Des. Appl. 2015, 12, 86–95. [Google Scholar] [CrossRef] [Green Version]
  31. Pentenrieder, K.; Bade, C.; Doil, F.; Meier, P. Augmented Reality-based factory planning-an application tailored to industrial needs. In Proceedings of the 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; pp. 31–42. [Google Scholar]
  32. Henderson, S.; Feiner, S. Exploring the benefits of augmented reality documentation for maintenance and repair. IEEE Trans. Vis. Comput. Graph. 2010, 17, 1355–1368. [Google Scholar] [CrossRef] [Green Version]
  33. Feiner, S.; MacIntyre, B.; Seligmann, D. Knowledge-based augmented reality. Commun. ACM 1993, 36, 53–62. [Google Scholar] [CrossRef]
  34. Nee, A.Y.; Ong, S.; Chryssolouris, G.; Mourtzis, D. Augmented reality applications in design and manufacturing. CIRP Ann. 2012, 61, 657–679. [Google Scholar] [CrossRef]
  35. Schwald, B.; Figue, J.; Chauvineau, E.; Vu-Hong, F.; Robert, A.; Arbolino, M. STARMATE: Using Augmented Reality technology for computer guided maintenance of complex mechanical elements. E-Work. ECommerce 2001, 1, 196–202. [Google Scholar]
  36. Haritos, T.; Macchiarella, N.D. A mobile application of augmented reality for aerospace maintenance training. In Proceedings of the 24th Digital Avionics Systems Conference, Washington, DC, USA, 30 October–3 November 2005; Volume 1, pp. 5.B.3–5.1. [Google Scholar]
  37. Savioja, P.; Jarvinen, P.; Karhela, T.; Siltanen, P.; Woodward, C. Developing a mobile, service-based augmented reality tool for modern maintenance work. In Proceedings of the International Conference on Virtual Reality, Beijing, China, 22–27 July 2007; pp. 554–563. [Google Scholar]
  38. De Crescenzio, F.; Fantini, M.; Persiani, F.; Di Stefano, L.; Azzari, P.; Salti, S. Augmented reality for aircraft maintenance training and operations support. IEEE Comput. Graph. Appl. 2010, 31, 96–101. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, J.; Feng, Y.; Zeng, C.; Li, S. An augmented reality based system for remote collaborative maintenance instruction of complex products. In Proceedings of the 2014 IEEE International Conference on Automation Science and Engineering (CASE), New Taipei, Taiwan, 18–22 August 2014; pp. 309–314. [Google Scholar]
  40. Goose, S.; Sudarsky, S.; Zhang, X.; Navab, N. Speech-enabled augmented reality supporting mobile industrial maintenance. IEEE Pervasive Comput. 2003, 2, 65–70. [Google Scholar] [CrossRef] [Green Version]
  41. Kerawalla, L.; Luckin, R.; Seljeflot, S.; Woolard, A. Making it real: Exploring the potential of augmented reality for teaching primary school science. Virtual Real. 2006, 10, 163–174. [Google Scholar] [CrossRef]
  42. Campos, C.; Elvira, R.; Rodriguez, J.J.G.; Montiel, J.M.; Tardos, J.D. ORB-SLAM3: An accurate open-source library for visual, visual-inertial and multi-map SLAM. arXiv 2020, arXiv:2007.11898. [Google Scholar] [CrossRef]
  43. Qin, T.; Li, P.; Shen, S. Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef] [Green Version]
  44. Sauro, J.; Dumas, J.S. Comparison of three one-question, post-task usability questionnaires. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Boston, MA, USA, 4–9 April 2009; pp. 1599–1608. [Google Scholar]
  45. Zijlstra, F.R. Efficiency in Work Behaviour: A Design Approach for Modern Tools; Delft University Press: Delft, The Netherlands, 1993. [Google Scholar]
  46. Brooke, J. Sus: A quick and dirty usability. Usability Eval. Ind. 1996, 189, 4–7. [Google Scholar]
Figure 1. A collaborative AR system for multiple users, including designers, inspectors, salesmen, and customers. In each figure, the virtual model is marked with a brown box.
Figure 1. A collaborative AR system for multiple users, including designers, inspectors, salesmen, and customers. In each figure, the virtual model is marked with a brown box.
Sensors 22 01319 g001
Figure 2. Outline architecture of MUCSys.
Figure 2. Outline architecture of MUCSys.
Sensors 22 01319 g002
Figure 3. Application process in production line sale and running cutting machine inspection.
Figure 3. Application process in production line sale and running cutting machine inspection.
Sensors 22 01319 g003
Figure 4. The framework for collaborative remote supporting, editing, and annotation.
Figure 4. The framework for collaborative remote supporting, editing, and annotation.
Sensors 22 01319 g004
Figure 5. Application process in cutting head design. Through MUCStudio, we added a marker and determined the relative pose between the cutting head and marker. Then we adopted MUCView to drive the cutting head to move with the designed trajectory after loading the scene and sensor data.
Figure 5. Application process in cutting head design. Through MUCStudio, we added a marker and determined the relative pose between the cutting head and marker. Then we adopted MUCView to drive the cutting head to move with the designed trajectory after loading the scene and sensor data.
Sensors 22 01319 g005
Figure 6. Detailed results for each question in SUS [46] (1: strongly disagree; 5: strongly agree).
Figure 6. Detailed results for each question in SUS [46] (1: strongly disagree; 5: strongly agree).
Sensors 22 01319 g006
Table 1. Detailed procedures of experimental tasks. The unit of the time limit is “minute”.
Table 1. Detailed procedures of experimental tasks. The unit of the time limit is “minute”.
Task IDDescriptionTime Limit
1. Design
1.1Download the specific CAD model and transform it to obj format in MUCStudio.4
1.2Load both cutting machine and head models.2
1.3Add the marker and set the pose of marker, cutting machine, and head.8
1.4Save the scene and export it to MUCView.2
1.5Print the marker and paste it to the designed position.3
1.6Run MUCView and load the scene file.2
1.7Scan the marker to display the scene model.2
2. Sale
2.1Run MUCView and load the initial model and layout of production line.5
2.2Choose the appropriate position to display the production line.2
2.3Adjust the position and orientation of the production line to the given target layout.10
2.4Save the scene.2
3. Inspection
3.1Generate a QR code and bind it to the given machine ID.2
3.2Print the QR code and paste it to the cutting machine.3
3.3Scan the QR code to display the running information.2
3.4Add a slice of annotation and upload it to the MUCServer.5
3.5Restart the MUCView and load the annotation information.5
3.6Modify the annotation and re-upload it.5
Table 2. Completion time and failure number of each experimental process. The unit time of completion time is “minute”.
Table 2. Completion time and failure number of each experimental process. The unit time of completion time is “minute”.
Task IDCompletion Time/
Failure Number
Task IDCompletion Time/
Failure Number
Task IDCompletion Time/
Failure Number
1. Design
1.12.2 min/01.21.1 min/01.34.7 min/2
1.40.9 min/11.52.1 min/11.60.8 min/0
1.70.8 min/0
Total12.6 min/4
2. Sale
2.12.9 min/02.20.8 min/02.35.7 min/3
2.40.5 min/0
Total9.9 min/3
3. Inspection
3.11.5 min/03.21.9 min/03.30.9 min/0
3.43.1 min/13.52.8 min/03.63.4 min/1
Total13.6 min/2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Qi, Y. A Multi-User Collaborative AR System for Industrial Applications. Sensors 2022, 22, 1319. https://doi.org/10.3390/s22041319

AMA Style

Wang J, Qi Y. A Multi-User Collaborative AR System for Industrial Applications. Sensors. 2022; 22(4):1319. https://doi.org/10.3390/s22041319

Chicago/Turabian Style

Wang, Junyi, and Yue Qi. 2022. "A Multi-User Collaborative AR System for Industrial Applications" Sensors 22, no. 4: 1319. https://doi.org/10.3390/s22041319

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop