Next Article in Journal
Combining Multichannel RSSI and Vision with Artificial Neural Networks to Improve BLE Trilateration
Next Article in Special Issue
Increasing the Generalization of Supervised Fabric Anomaly Detection Methods to Unseen Fabrics
Previous Article in Journal
Industry 4.0 and Precision Livestock Farming (PLF): An up to Date Overview across Animal Productions
Previous Article in Special Issue
Perimeter Intrusion Detection by Video Surveillance: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recent Advances in Video Analytics for Rail Network Surveillance for Security, Trespass and Suicide Prevention—A Survey

1
The University of Sheffield, Sheffield S1 3JD, UK
2
Network Rail, Milton Keynes MK9 1EN, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4324; https://doi.org/10.3390/s22124324
Submission received: 8 April 2022 / Revised: 22 May 2022 / Accepted: 26 May 2022 / Published: 7 June 2022
(This article belongs to the Special Issue Unusual Behavior Detection Based on Machine Learning)

Abstract

:
Railway networks systems are by design open and accessible to people, but this presents challenges in the prevention of events such as terrorism, trespass, and suicide fatalities. With the rapid advancement of machine learning, numerous computer vision methods have been developed in closed-circuit television (CCTV) surveillance systems for the purposes of managing public spaces. These methods are built based on multiple types of sensors and are designed to automatically detect static objects and unexpected events, monitor people, and prevent potential dangers. This survey focuses on recently developed CCTV surveillance methods for rail networks, discusses the challenges they face, their advantages and disadvantages and a vision for future railway surveillance systems. State-of-the-art methods for object detection and behaviour recognition applied to rail network surveillance systems are introduced, and the ethics of handling personal data and the use of automated systems are also considered.

1. Introduction

Rail network systems are, by design, open, accessible and welcoming environments. Control of entry is largely on the basis of revenue protection, not security, and with a few exceptions (e.g., international services), they do not share the strict perimeter security model adopted for air travel. While this is entirely appropriate for efficient transport operation, it does leave the systems vulnerable in several ways. Terrorist activity is often identified as the highest-profile vulnerability, and a range of resilience and design strategies have been adopted to reduce the attractiveness of railways as a target and to enable them to recover rapidly if attacked [1]. However, disruptions are caused by unexpected events such as trespass (often linked with theft of materials or equipment) and suicide attempts [2,3]. This requires that railway employees are trained to look out for and offer support to people who may be considering taking their own life on the railway.
Mitigation measures such as fencing can be used to prevent access at high-risk locations and help address security, trespass and suicide risks. However, access prevention is not a solution for station environments, and it is here that recent advances in video analytics tailored to rail [4,5] offer alternative approaches to reduce these risks in coordination with ‘designing in’ physical solutions such as blast resistant glazing [6], and management solutions such as staff training to identify out-of-the-ordinary situations. The development of digital technology for sensors, sensing, data fusion and machine learning, combined with the advancing communication and cloud technologies have the potential to contribute greatly in these areas. Their application, however, requires careful assessment of ethics concerning public privacy and data handling, for example, as regulated in Europe through General Data Protection Regulation (GDPR) regulations. This survey supports the application of these technologies in rail by bringing together technical and ethical information to protect public security, protect the rail system from trespassing, and reduce railway suicides. This range of applications is useful in correctly balancing the benefits and costs of a potential system, but it leads to difficulty in the choice of language. Where video analytics of a potential attacker may aim to identify a ‘target’ for investigation, the same technology could also be identifying a vulnerable person in distress. For simplicity, in this article, we have adopted the most common language used in each area of technology.
The main contributions of this work are the following: (i) it provides a thorough survey of the latest trends in closed-circuit television (CCTV) systems, with their advantages and disadvantages, (ii) presents the latest trends especially for vision-based behaviour analytics and especially in the light of improving safety and security in rail network stations, (iii) presents a systematic approach encompassing the whole process from the data collection, to storage, communications and video analytics, including ethics and data protection aspects, (iv) traces a vision for the future autonomous vision-based surveillance on several aspects, including the impact of the uncertainties in Artificial Intelligence models.
This paper is organised in the following way. Section 2 introduces an overview of current surveillance systems in rail networks. Section 2.1 introduces the recent research trends in railway video surveillance. Following by the main components of current surveillance systems, including CCTV cameras and non-conventional sensors, and a brief overview of monitoring and recording systems and various algorithms on anomaly detection in surveillance video data. The strengths of multiple modality sensors are discussed to help in sensor selection concerning the level of required security. Section 2.5.1 introduces the modern deep learning methods used in video analytics. Various computer vision technologies built with Artificial Intelligence (AI) models based on deep learning algorithms used in railway network surveillance are introduced in Section 2.1.1.
The impact of the uncertainties in computer vision technologies and state-of-the-art Bayesian deep learning methods for formulating the uncertainties are introduced in Section 3. A broad vision of next-generation systems is given in Section 4, and future data centres are introduced in Section 4.1. Ethical and privacy-related aspects are covered in Section 5, followed by conclusions in Section 6.

2. Overview of Surveillance Systems for Rail Networks

CCTV systems [7] have evolved over the years from simple analogue systems in the 1960s to digital systems developed in the 1980s [8]. The advent of digital systems extended the limited recording and distribution abilities of analogue systems. Object detection and tracking are the basic video analytics technologies in digital systems. Since the 2000s, more and more advanced video analytics technologies have been developed, taking advantage of valuable video data collected by multiple integrated sensors.

2.1. Bibliometric Analysis

Nowadays, the most advanced video analytics for CCTV is capable of providing automated detection and alerts based on information processing, reasonable framework and computer vision technologies such as object detection [9], behaviour analysis and activity recognition [10,11,12], crowd tracking [13], emergency evacuation and deterrence [14,15,16,17]. We have analysed the current trends in CCTV video surveillance for rail network systems and related computer vision technologies. Some of the results since 2010 are shown in Figure 1 and Figure 2. The following keywords are used in the search: “surveillance”, “railway”, “computer vision” and “machine learning”. A tool called Bibliometrix [18] is used here.
During the construction of surveillance systems of railway networks, there are certain rules issued by the government or relevant authorities that must be followed, which are known as requirements documents. Requirements documents related to the standards for surveillance systems in different countries [19,20,21,22,23,24,25,26,27,28,29] are compulsory to follow and are to be prepared based on the outcome of the previous step. It specifies the needs and the justification of the CCTV system [30]. Additionally, these documents cover the legislative and privacy limitations. Other kinds of requirements documents include multiple instructions on building the surveillance systems, for example, how to access data sets on video analysis [31]. An example of legislative limitation is to assume that the recording capacity of the system is one year, whereas the storage time limit defined by the government is actually six months. An example of privacy limitations is blocking the view of a private neighbouring property from the system.

2.1.1. Trends in Computer Vision Technologies in Video Surveillance

These technologies are either currently used in surveillance systems in railway networks or considered essential to be developed in the next generation of smart surveillance systems. Table 1 shows the recent research status of the main computer vision technologies used in railway video surveillance. The result is concluded under the keywords: “surveillance”, “railway”, and “the name of computer vision technology” using the Google Scholar search tool. These statistical results show that the focus is on human activity recognition, anomaly detection and crowd analysis.
Section 2.2 describes the main objectives of CCTV surveillance systems.

2.2. Objectives of CCTV Surveillance Systems

The primary objective of the CCTV systems is to assist in preventing crime and terrorist activities and to provide safety and security to people. Sensors are compulsory elements of the CCTV system. Conventional video-based sensors have limited capability of identifying hazardous materials or weapons used by terrorists or criminals. The main components for a surveillance system are various sensors, monitoring and recording systems, video analytics, communication networks, cloud computing or other storage systems. Communication networks are critical for ensuring the reliable transfer of data among these components. The sensors, video analytics and recording systems are configured from the monitoring system. A block diagram of a CCTV system is shown in Figure 3.
Raw sensor data and the data analytics output can overwhelm security operators if the data are inefficient. In order to detect behaviours of interest correctly, carefully designing an efficient monitoring system is important. A satisfying surveillance system setting relies on the integration of the geographic information system such as the Global Positioning System (GPS) with the CCTV to reference information from a known location. This reduces the training time of the operators as effective monitoring does not solely rely on their knowledge of sensor placement across the site. Data from multiple sensors is combined to provide all the information and intelligence relevant to a specific area in one place. Advanced display software provides an option to view the scene not only from the camera’s point of view but also from multiple viewpoints, which can enhance situational awareness and reduce response time.
Taking the advantages of the surveillance systems built in the railway networks, abnormal behaviour detection methodologies by computer vision methods are widely implemented. Surveillance systems built for railway networks are usually focused on a constrained area, which means there are multiple computer vision technologies that could be applied to specific scenarios. These anomaly detection algorithms based on machine learning methods can autonomously detect unusual behaviours in video surveillance data through the camera sensors without operator intervention. By detecting the change of movement temporally and spatially, the system should be intelligent enough to classify behaviours as regular or abnormal events [8,32,33,34]. State-of-the-art methods for anomaly detection in video surveillance are presented in Section 2.1.1.

2.3. Sensors

The most commonly used sensors for collecting data are optical and infrared (IR) cameras, chemical sensors and motion detectors. CCTV cameras are an important element of security and surveillance systems [7] when deployed as part of a multi-layered security architecture and provide video streams for all areas, both indoor and outdoor.
Conventional sensors, such as video cameras providing behavioural trait analysis, rely on advances in video analysis to detect the behaviour of interest. Non-conventional sensors, e.g., chemical sensors, have their advantages in detection through analysis of vapours, with limitations in tracking and data visualisation. The integration of non-conventional systems is still promising in current and future systems. Other combined conventional and non-conventional sensors, such as health sensors that are useful for detecting and preventing viruses and infections, have also been of benefit during the COVID-19 pandemic. Train stations are major entrances and exits in national and international transport networks. With these kinds of sensors, the surveillance systems could aid in the population-level understanding of epidemic or pandemic spread.
Multimodality (conventional and non-conventional) sensors collaborate to give early warnings and share useful data for sensor management purposes. Data from a single sensor is considered low-level intelligence, whereas automated analysis of the sensor’s data provides higher level intelligence to the security operators. Figure 4 illustrates such a multi-modal sensor framework.
With the continuous advancement of human technology, smart phones or watches carried by passengers have been developed as other types of modern sensors to collect behaviour data for behaviour surveillance [35]. However, due to the privacy and security restrictions, these types of sensors have not been able to be used in surveillance systems.

2.3.1. Cameras

The most important component of a CCTV system is a camera, which is also known as an imaging sensor. In some systems, non-imaging sensors are integrated to enhance operations. The two common imaging sensors in the CCTV system are optical and thermal cameras. The cameras consist of three parts: lens, image sensor and image processor. The lens focuses the light reflected from an object on the image sensor, the image sensor converts the light photons to electrical signals, and the image processor organizes, optimizes and transmits the electrical signals. Some advanced cameras have onboard video analytics as well. The cameras are available in various options. A summary of basic and value-added options is given in Table 2 and Table 3. These tables can be referred to for choosing the camera features as per the risk assessment.

2.3.2. Chemical Sensors

Camera sensors are not capable of detecting potential threats in some cases, such as explosive materials hidden in bags. Chemical sensors are another type of sensor commonly deployed in public transport stations to detect hazardous materials. Chemical sensors detect objects by sensing the vapour concentration leaked from the radioactive and explosive materials. The localisation of the individual carrying hazardous material is a complex problem. The fusion of chemical sensor data with the data from optical cameras can be used for tracking of the suspect [36]. Accurate person localisation is achieved through the fusion of different sensor data in narrow corridor spaces. Once an individual carrying such material is detected and tracked, the sensors will share the information with other sensors and the control centre through the communication system.

2.3.3. Concealed Weapon Sensors

The detection of concealed weapons is a challenge to conventional video and thermal cameras. Weapons are normally characterised by their shape, material and appearance. Millimetre (30–300 GHz) and sub-millimetre (0.3–3 THz) wave sensors are state-of-the-art detectors for concealed weapons. There are active and passive scanners based on the radar principle. Active scanners operate in the range of 350–800 GHz, direct waves on the subject and receive the reflected energy. Passive scanners receive radiation from humans and other objects and operate in the 30–350 GHz range. Compared to active scanners, passive scanners provide a low signal-to-noise ratio (SNR) and low-resolution images, which are more challenging to process. Figure 5 shows an example of using radio waves in public transport stations to perform safety checks.

2.3.4. Health Sensors

Train stations are the entry and exit points of states and countries. The spread of an epidemic or pandemic, such as COVID-19, can be controlled or limited using health sensors. Such health sensors are deployed to classify whether individuals are healthy or not by measuring people’s vital signs, such as fever and heart rate. Infrared thermography (IRT) is currently used at airports for monitoring travellers’ fevers. Accuracy can be improved by using the same IRT sensor on screening based upon all vital signs [38]. The fusion of health sensor data with camera data can provide real-time tracking of a sick individual as shown in Figure 6.

2.3.5. Acoustic and Vibration Sensors

Acoustic or sound sensors are the most commonly used non-conventional sensors. Acoustic and vibration sensors provide detection of humans and other moving objects, which is helpful for object tracking algorithms. Compared to imaging sensors, acoustic and vibration sensors do not require line of sight to provide omnidirectional coverage. These sensors can also trigger an event, such as turning on lights and imaging cameras for video surveillance. In addition, audio analytics helps video analytics detect target events such as screams, glass-breaking sounds, explosions, and gun fires.

2.4. Monitoring and Recording System

A monitoring system consists of a control room, monitor screens, and control devices. The number of monitoring screens depends on the number of cameras and their functionality. Some modern monitoring (IP-based) systems can be deployed on mobile devices such as smartphones and tablets, and even mobile devices provide limited functionality due to the device size. CCTV operations require big screens for optimum operation. There are two modern big screen monitoring technologies, liquid crystal display (LCD) and light-emitting diode (LED) displays, which have superseded the older cathode ray tube (CRT) displays and plasma displays. Table 4 compares these monitors. The following aspects should be kept in mind when choosing a display:
  • Size: Size is usually specified as the length of the diagonal measurement of the monitor. The size is selected based on the number of simultaneous camera displays. As a rule, the operator to screen distance is three to five times the screen size.
  • Resolution (Number of pixels): The number of pixels specifies the display resolution. The pixel number has a one-to-one correspondence to the resolution of the camera output image. For best performance, these should match.
  • Aspect Ratio: Aspect ratio is defined as the ratio of the number of pixels in the horizontal to those in the vertical. For best results, the aspect ratio of the monitor should match that of the camera.
The recording system records digital images and auxiliary data, as well as their copies, on a digital recording device. Some copies are stored on other media platforms, such as Cloud platforms, for safety reasons. The recording system is an important part of the surveillance system since smart surveillance relies on recorded video data.
Traditionally, all the components of a CCTV, from sensors to recording and monitoring stations, are installed at the same location in response to the real-time request. However, the long-term installation and maintenance costs can be high. These costs can be reduced by utilising a distributed computing infrastructure. The users of the CCTV system usually set the regions of interest and the location of monitoring rooms, while the system designers typically set the location of the computing and recording systems. This computing infrastructure can be set up to reduce the costs in four configurations, as shown in Figure 7.

2.5. Video Analytics

Video analytics is defined as an automatic processing of video analysis aimed at assisting the surveillance operations. Some of the main challenges to video analytics for railway surveillance from a computer vision perspective can be subdivided into two big groups depending on whether objects are moving or static: (1) classification, detection, segmentation and decision making for static objects such as left bags or a person laying on the rails of the network; (2) classification, detection, segmentation, tracking and decision making for dynamic events such as crowd monitoring, tracking a terrorist that is running to escape, a thief or the suicide of a passenger.
Both groups of approaches have in common that they need to be able to provide reliable solutions under a wide range of conditions, both naturally occurring (e.g., lighting changes, weather condition changes) and cyber attacks, which include intentional interference during different stages of the imagery data transmission, processing and storage. Hence, there is a demand for a wide range of methods able to meet different requirements and provide trustworthy solutions. One way to provide a level of trustworthiness of machine learning approaches is to evaluate how the different above-mentioned factors, which we call uncertainties, impact the developed solutions. Are the developed solutions reliable, and if yes, under what conditions and circumstances? The answer can be found based on probabilities or other information measures representing different levels of trust in the developed solution.
In addition, there are several other challenges due to occlusions, environmental conditions, scalability—of the considered monitored area, and big heterogeneous sensor data. In terms of network rail stations, we have small, medium and big train stations for railway networks in big geographic areas.
Video analytics requires a combination of several computer vision techniques, which are introduced in Section 2.1.1. There is a wealth of video analytics methods as summarised on Figure 8 out of which machine learning methods are now an active area of research. Classical methodologies [40,41] for tracking the motion of the targets and detecting the change of the motion are implemented by firstly applying object detection techniques such as the Optical Flow [42,43,44,45] techniques to determine objects of interest within the scene and estimate their motion, and then using classification methods such as Gaussian Process Regression [46,47,48] to detect the change point of the behaviour.
In recent years, deep learning methods based on Convolutional Neural Networks (CNNs) [49,50,51,52], which are also widely used in building Artificial Intelligence (AI) have become the most popular computer vision technology. Different from the traditional machine learning methods, deep learning methods developed recently have provided more robust and accurate algorithms available for object detection and pattern recognition.

2.5.1. Deep Learning Methods for Video Analytics

Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Vision Transformer (ViT) are the three most popular deep learning methods used implementing video analytics models for video surveillance.

Convolutional Neural Networks

CNNs have become the most powerful tool for computer vision tasks due to their strong performance on image and video data. CNNs are biologically inspired, and they are built to learn invariant features with millions of neurons. Compared to a regular Neural Network, a Convolutional Neural Network layer consists of weights parameters in three dimensions: width, height, depth, which makes it possible to reduce the image features into a single vector by the end of an entire network’s architecture. In recent years, various deep learning models developed based on CNNs have been demonstrated with strong robustness performance on image datasets such as ImageNet [49], which makes it the most popular deep learning framework in the area of computer vision [52]. One of the most popular applications is face recognition [53,54,55]. A basic CNN structure is shown in Figure 9.

Recurrent Neural Networks

Recurrent Neural Networks (RNNs) [50,56,57,58,59] are another important class of networks for video surveillance. However, learning the image features in certain tasks requires a short-term memory. Hence, deep RNNs [50] are developed to address the problem where traditional neural networks are not able to learn the features of the input data from the previous data in a temporal sequence. RNNs are built based on loops of neural networks. Figure 10 shows the architecture of the loop structure inside RNNs. AI models built with RNNs allow them to understand the current features based on the previous data.
In recent years, RNNs have been widely applied for solving problems with sequential input data such as speech recognition, language modelling and image prediction [58,59]. However, in some cases, not only short-term memories are required, long-term dependencies are also needed. In response to this problem, Long Short-Term Memory networks (LSTMs) [60] are designed as a certain type of RNN, which is especially good at learning the long-term dependencies. Different from RNNs, the repeating module in LSTMs has several neural network layers, which are designed to be able to memorise the previous features in the cell state.
A convolutional LSTM is firstly used in building a spatio-temporal video autoencoder in future frame prediction [61]. The convolutional LSTM cells are used to build the temporal encoder to integrate changes over time. Deep neural networks are normally used as the method for learning the spatial representations. Recent works [62,63] utilise the advantages of LSTM networks with long-term dependencies to build Convolutional LSTM networks used to learn the temporal features of the input sequence video data, which significantly improves the performance of the proposed method of using an LSTM convolutional autoencoder for video anomaly detection.

Vision Transformer (ViT)

In addition to CNNs and RNNs, a novel deep learning structure called a transformer, which is based on a self-attention mechanism [64], has become a popular method used in the area of objective detection and behaviour recognition. A transformer is firstly developed for solving Natural Language Processing (NLP) tasks. Due to its strong representation capabilities and huge success with NLP tasks, researchers are expecting that transformers will be able to achieve huge success performing computer vision tasks as well [65]. Such improved transformers built for computer vision tasks are called Vision Transformers (ViT) [65]. A basic Transformer structure is shown in Figure 11.
When evaluated on computer vision benchmarks, AI models built based on transformers have achieved similar or even better performance than models built based on other types of classic networks such as CNNs and RNNs [65]. Transformers are gaining popularity among the computer vision community due to their strong performance and lack of the necessity for vision-specific inductive bias [66,67]. As a novel deep learning framework, ViT now has gradually become the most powerful method for image classification tasks.

2.6. Motion or Moving Object Detection

Motion or moving object detection (MOD) [68,69,70] is one of the earliest video analytics features made available in CCTV surveillance systems. The aim of MOD is to detect the event of a single object or multiple target objects of interest moving in the region of interest.
There are three traditional MOD approaches, namely frame difference [71,72], background subtraction [73,74] and optical flow. Frame difference relies on the temporal difference of pixel intensity in consecutive frames. The difference of pixels belonging to static objects is close to zero, and the moving objects are detected with high difference values. The moving objects in a static background are detected using a foreground detector. The optical flow approach measures pixel-level velocity vectors to detect the motion.

2.6.1. Static Object Detection

Static object detection (SOD) [75,76,77] alarms a surveillance operator when an instance of an object of interest is detected in the scene captured by the cameras. The methods for static object detection are different from the moving object as there is no relative movement between the object of interest and the background. Most of the other video analytics processes such as tracking, face recognition, person re-identification, and many more rely on SOD as a pre-processing step.
Figure 12 shows an example pipeline of a two-step SOD. The object detection methods generally do not differentiate between objects of one type, e.g., the suitcase and the backpack are classified as one type, i.e., bag. In a single-stage detector, all these steps are performed by a multi-layer neural network. From a user point of view, all objects of interest are mentioned in the requirement document.

2.6.2. Multiple Object Tracking

Multiple object tracking (MOT) [34,79,80] aims at identifying and tracking multiple objects of interest in a video without any prior knowledge of their appearance and disappearance from the scene. The video can be from the same or different cameras. MOT differs from SOD in the sense that each object will be given a unique label, which is maintained in all video frames by the MOT algorithm. The MOT algorithm will identify all objects, such as the example using Gaussian models shown in Figure 13.
Object detection, feature extraction, motion prediction and the association methodology are the key areas of research for static and multiple object tracking. State-of-the-art methods for object detection and feature extraction are deep learning methods [13].
State-of-the-art object detection methods are built based on deep learning models. Generally, there are two types of detectors: two-stage and single-stage detectors [82]. Two-stage models have a separate module to generate region proposals. In the first stage, the models aim to find an arbitrary number of object proposals. Then the objects will be classified and localised in the second stage. Single-stage detectors classify and localize semantic objects in a single shot using dense sampling. Several popular two-stage detectors are: R-CNN [83], Fast R-CNN [84] and R-FCN [85].
Faster-RCNN [86] is one of the advanced CNN-based deep learning objective detection methods with powerful performance. You Only Look Once (YOLO) [87], which is an even more powerful algorithm widely developed in recent years, is also used as the object detector ending up with a significant improvement for objective detection under different kinds of scenarios. These algorithms usually label the objects in the video by bounding box regressions, i.e., the tracking box shown in Figure 14. Then, other object tracking algorithms can be applied to the objects.
Figure 13 shows multiple detected objects, including vehicles passing through the traffic crosses and the pedestrians walking across the roads. In intelligent surveillance systems, Extended Kalman filters and Gaussian Mixture models are commonly used machine learning-based methods to track moving objects [88]. In classical methods, such as optical flow types and background subtraction, the minimum number of frames used to detect traffic flow disruptions is two, although the number of frames used can vary.
After detecting the objects using the object detection algorithms introduced above, then motion prediction is achieved by using Bayesian approaches such as Kalman filters, particle filters and probability hypothesis density filters. The association probability calculated using both the motion and feature information performs better than those which use one of them. Advance methods, such as the Multiple Hypothesis Tracking (MHT) framework, process the data of multiple images to associate labels. These methods are based on data from multiple time steps to give better results.
Figure 14 shows an example pipeline using the topic modelling method to detect the abnormal behaviours under the traffic scenario [81]. The minimum number of required frames using topic modelling to detect the target disruption in traffic flow is two. The disruption in traffic flow is detected when the discriminant function reaches the threshold. In rail network cases, SOD and MOT techniques can be used to detect target objects with abnormal behaviours that could cause potential harm to the safety of the rail system, such as a person leaving a suitcase on the platform. SOD and MOT are the basic techniques for Video Anomaly Detection, which will be introduced in the following section.

2.7. Video Face Recognition

Video face recognition (VFR) [53,54,55] deals with identifying an individual from the video data by comparing a given person’s video with a library of still face images. In rail networks, VFR technologies can be used to automatically raise the alarm when the sensors detect an individual from a specific watchlist. It can also be used to search for a specific person, through an image of the target from the recorded surveillance data. A basic VFR process is shown in Figure 15.
Before the identification stage, the captured video data need pre-processing. The pre-processing step contains face detection, face alignment, and face aggregation. Firstly, the face detector localises the faces in each frame. Then, the face alignment detects facial landmarks such as eyes, nose and other important features. Finally, the face aggregation represents the aligned faces in multiple frames by a comparatively smaller set of frames. As a result of aggregation, noisy and low-quality frames are rejected.
All VFR methods extract a facial feature set corresponding to a separate image before matching them with a library of feature sets. Each feature set in the library will be assigned an identification tag. The matching algorithm identifies the object through the differences between the feature sets extracted from the aggregated frames and stored in the library. Whether a feature library consists of single or multiple entries of the same person, the VFR performs a still-to-video verification process. Some VFR methods that can only identify objects already recorded in the library are called closed-set VFR. A model based on open-set VFR is preferred for such a security system as a railway station. For example, the open-set VFR can be implemented to identify people on a certain watchlist [90]. The VFR methods are categorised into two main streams, conventional and deep learning methods. Most of the modern VFR models that are already deployed at railway stations are built based on CNNs since they are more robust and accurate, and they are able to make decisions by themselves as compared to the conventional methods [9].
VFR helps in crime detection by tracking missing persons and estimating passenger traffic at any time in a rail network. Another application of VFR is that a facial recognition system could be used to replace the need for tickets on trains. Such systems have been tested in the UK [91]. By using two near-infrared lights helping a single camera determine the texture and orientation of each pixel, the smart surveillance system is able to successfully identify passengers’ tickets without the need for them to stop walking and show their tickets.

2.8. Person Re-Identification

Person re-identification (Re-ID) [92,93,94] aims at identifying a person across non-overlapping cameras and different modalities. It is also used for identifying an individual who passes through blind spots and re-appears in the coverage of the same camera. It serves various security purposes such as online tracking of individuals over multiple non-overlapping cameras and searching by a query from several cameras’ recorded video data. This query can be an image, a video or a text description of the person. Depending on the different components involved in the surveillance system, Re-ID can be categorised into closed-world and open-world settings [92]. The closed-world setting requires annotated searched video data and vice-versa for open-world settings. The open-world person Re-ID technology used in video surveillance systems is shown in Figure 16.
As shown in Figure 17, there are five main steps for building a person Re-ID system: raw data collection, bounding box generation, training data annotation, model training and pedestrian retrieval [92]. CCTV firstly detects images containing target individuals. Then, deep learning models provide feature extraction processes. The features extracted from the particular camera images are matched with those obtained from each library image through a matching algorithm based on the similarity. Potential targets and images can thus be filtered out through the matching stage.
The processings are the same when the given image is replaced with an individual’s video or textual description. The feature extraction is the key stage. In recent years, deep learning methods have been deeply developed with their great performance, but the current deep learning methods need further improvements for practical implementations [92].

2.9. Human Activity Recognition

Human activity recognition (HAR) involves classifying an individual’s activities or a small group of individuals. It is categorised into the following three types [11,95]:
  • Gestures. These are elementary motions of the human body, such as stretching of arms or raising a leg.
  • Action. These are the combinations of human gestures, for example, running and punching.
  • Interactions. These involve interactions among multiple individuals and objects. These can be classified into human–object, human–human and human–object–human interactions. The human–object interaction deals with recognising the interaction of a human with an object, for instance, an individual abandoning an object or loitering. An individual interacting with another is human–human, such as a brawl. The interaction between two humans and an object leads to human–object–human activity, such as stealing a bag from another individual.
As shown in Figure 18, the output class is “Abandoned Object”. In addition to the classification, the alert also provides the object’s location by highlighting it on the screen (yellow box in the example). The state-of-the-art methods for HAR are deep learning-based methods that merge the above-mentioned two steps into a single unified framework [12]. The latest architecture of the deep learning methods is based on two streams, one for the spatial and the other for temporal processing [96]. The output of two streams is merged to give the final output. This two-stream architecture is also called the Spatio-Temporal Network (STN). The surveillance system user defines all the classes of gestures, actions, and interactions required to be automatically detected by the HAR algorithm.

2.10. Video Anomaly Detection

Video Anomaly Detection (VAD) aims to detect abnormal behaviour of an individual or a group of people through a camera or multi-camera video. An anomaly or an abnormal behaviour is difficult to define; normally it is a psychological term for actions that fall outside of the realm of what is considered normative in a particular society or culture. In this case of rail networks for safety purposes, it can be defined as an event or behaviour that could lead to a potential threat to public safety. VAD is an essential technology for safety surveillance systems.
In railway systems, VAD can help prevent trespassing and suicides at train stations. Different from HAR, VAD does not rely on a number of manually labelled training data to detect normal and abnormal behaviours. VAR has better performance than HAR in cases where the classifier does not have enough labelled data of abnormal behaviour occurrences. Individual unsupervised anomaly detection (IUAD) is useful for detecting a person attempting suicide [98] or a terrorist attempt [99] in the surveillance region, where it is assumed that the individual carries out the daily routine tasks abnormally with a suicide or terror motive in mind. A basic block diagram of the IUAD algorithm detecting an anomaly of a person falling on the floor is given in Figure 19.
Depending on how the algorithms are built based on training data, there are three main categories of anomaly detection methods [100]: unsupervised methods, weakly supervised methods and supervised methods. Some of the anomaly detection tasks can be performed using supervised or semi-supervised anomaly detection methods. The supervised anomaly detection model is trained on labelled data of normal and abnormal behaviours. The semi-supervised anomaly detection model is trained on labelled data of normal behaviour. The unsupervised anomaly detection model trains on unlabelled data, such as real-time surveillance videos, which is a more practical and dynamic approach. Since the abnormal detection is supposed to be able to recognise the anomaly events with low happening probability, unsupervised algorithms can be considered as novelty detection with self-learning ability, which aims to cluster the outliers from the normal points. These methods provide a better generalisation ability under a single scenario with a fixed surveillance camera of the unknown anomalies.
Some works categorise VAD methods, specifically with respect to their final objectives instead of network architecture or learning strategy, into two categories, namely using reconstruction error or reconstruction-based methods and predicting the future frames or prediction-based methods [101]. Some of the recent deep learning methods [102,103,104] focused on implementing the video anomaly detection by reconstructing the predicted frames based on the features extracted from the previous frames through an autoencoder model. Abnormal events are located by the calculated reconstruction errors. When the reconstruction errors are over the threshold, the event happening in the current frame is considered an abnormal event.
Two methods built upon the autoencoders are proposed to generate a model for computing the irregularity error [102]. An autoencoder [105,106,107] is a feedforward neural network built for denoting a process of encoding the input into a more compact representation and then reconstructing the input with the learned representation as to the output. The irregularity error is implemented through two stages: first is to train a fully connected autoencoder on the spatio-temporal local features. Secondly, build a fully convolutional feedforward autoencoder to learn the classifier. To improve the performance of the Autoencoder, LSTM Convolution Neural Networks are leveraged for memorising the motion features of the past frames [103]. The proposed Convolutional LSTM Autoencoder framework has better performance with respect to encoding the change of the motions of the objectives for normal events.
As the deep Autoencoder has been widely applied for anomaly detection, a new autoencoder model with a memory module called Memory-augmented Autoencoder (MemAE) is implemented [104]. The traditional Autoencoder is improved by memorising the model parameters trained from normal samples at the training stage. Then, the reconstruction error is obtained while updating the rest of the autoencoder model at the test stage. MemAE has been proved to have better generalisation ability and higher effectiveness than the normal autoencoder algorithms. A novel deep Predictive Coding Network (AnoPCN) consisting of two modules: a Predictive Coding Module (PCM) and an Error Refinement Module (ERM) is introduced in [108]. Similar to LSTM, the feedback connections of PCM carry frame prediction, and the feedforward connections are built for computing the prediction error. AnoPCN has been verified with state-of-the-art performance on experimental data sets.
Instead of using an Autoencoder to compute the reconstruction error, there are other deep learning frameworks built to predict future frames. A method was developed that uses Generative Adversarial Networks (GANs) to train a generator, and alternatively, a discriminator is used to achieve the future prediction [100].

2.11. Trajectory Analysis

The MOT localises multiple objects across time. The path of each object is called a trajectory. Analysis of trajectory, also called Trajectory Analysis (TA) [109,110], provides valuable intelligence in the domain of video surveillance. Short-term trajectory analysis detects abnormal events such as walking and fighting. The long-term analysis can detect events such as loitering and interactions. The analysis can also be extended to analyse group activities by analysing multiple relative motions.
Combining TA with prior regional information provides detection of important events. For example, marking out-of-bounds areas in the TA alarms for intrusion detection and integrating the station layout with TA can detect pedestrian falls on the tracks. TA is also used for generating a summary or synopsis of a video. A video synopsis or summarisation aims to shorten a video by selecting critical frames. Fast-forwarding is another way of shortening a long video, but this method may miss essential frames.

2.12. Crowd Analysis

Crowd analysis [111,112] aims at understanding the behaviour of a crowd. A crowd is defined as a group of people behaving coherently. In a train station, travellers often move in groups, which makes it challenging to perform high-level processing such as MOT and anomaly detection. In such scenarios, instead of tracking and analysing each individual, the tracking and analysis of the crowd are feasible. Crowd analysis also aids in crowd management under unforeseen circumstances, such as emergency evacuation. The following types of crowd analysis are essential in video surveillance [16,113]:
Crowd Count (CC) aims at estimating the number of people from a video of a region of interest. CC enhances situational awareness in crowded areas. An increasing number of people in an area requires extra attention for management and safety purposes in a railway station. Moreover, CC can be used for gathering statistical data of visitors in different areas. The conventional methods for crowd counting rely on detection and regression techniques. Detection methods perform worse when there is a dense crowd and high background clutter. Counting by regression extracts low-level features and maps them onto a crowd count without considering spatial information. A block diagram of a typical density-based CC method is shown in Figure 20.
Figure 20 shows an example process of how CC algorithms deal with the input videos consisting of small groups of people. A more intuitive output of the CC algorithm is the congestion level. The crowd count is mapped on a scale ranging from sparse to dense crowd level. An alarm will rise when the crowd density rises beyond a critical threshold, which can cause dangers such as a stampede. One state-of-the-art method for crowd counting in videos is called the deep Spatio-Temporal Network (STN) [17].
Crowd Behaviour Recognition (CBR) identifies specific actions and events of a large gathering or crowd. Identification of different crowd behaviours helps in preventing a difficult situation. It can also be used to manage the flow of pedestrians. For example, bottlenecks and blocking of pedestrian flow are recognised using CBR [114]. A typical block diagram of CBR is similar to that of HAR, shown in Figure 18, where instead of extracting features of individual objects, the crowd features are extracted. Crowd Anomaly Detection (CAD) deals with the detection of abnormal behaviour in a crowd [115]. Similar to VAD, CAD helps automatically detect suspicious activities in crowded scenes and helps prevent a security-related incident from happening.

3. The Impact of Uncertainties in Computer Vision Technologies for Surveillance Systems

In addition to the challenges that exist in the current methods of CCTV surveillance, including the accuracy and efficiency of the results, there is another research area of the uncertainty in deep learning methods, which is the potential challenge of the security of surveillance systems. Investigating whether the results output by AI models are reliable is essential to security-related systems. For surveillance systems in rail networks, a practical and comprehensive AI model with the ability to think can provide trustworthy results for detecting the abnormal behaviour of people in public scenarios.
Most deep learning models introduced in Section 2.1.1 are considered as deterministic functions, which are barely used as solutions for possessing uncertainty information compared to probabilistic models. Probabilistic Graphical Models (PGM) use diagrammatic representations to describe random variables and relationships among them [116], which provide us with a practical tool for investigating the uncertainties inside the neural networks. Different from deterministic deep learning methods relying on estimations of the weights parameters and predictions, probabilistic models instead infer the distributions of the weights to reserve the uncertainties. A Gaussian Process [46,48] is one of the classic methods that define probability distributions over functions. This machine learning probabilistic method provides confidence bounds for data analysis and decision making—reliable information that, for instance, allows a biologist to analyse data or an autonomous car to decide whether to stop or not.

3.1. Uncertainty in Deep Learning

Deep learning models are designed to be able to return prediction results with high confidence when given data under the same category as the training data. However, the data fed to deep learning models can often be different from training data. These kinds of data are called out of distribution test data [117]. Under such circumstances, an ideal model would not only be able to return a prediction but also be capable of returning an analysis with additional information on whether the result lies outside of the data distribution. There are a number of factors that could lead to inaccurate solutions:
(1)
Noisy data/out of distribution data.
(2)
Uncertainty on the deep network parameters that are chosen during the training stage. This is also known as uncertainty in model parameters.
(3)
Model structure difference depending on the chosen algorithm for building the model (also called structure uncertainty). This uncertainty can be reflected in the following ways: an absence of theory and causal models and computational costs.

3.1.1. Types of Uncertainty

Generally, there are two main types of uncertainty: aleatoric uncertainty and epistemic uncertainty [104,117,118]. Aleatoric uncertainty, also known as data uncertainty, refers to the irreducible uncertainty caused by the inconsistency between the data used in training the model and the out of distribution data mentioned above. Compared to an inherent property of the data distribution, aleatoric uncertainty is more like a property of the model, which makes it irreducible. On the other hand, the uncertainty caused by model structure and inadequate knowledge is called epistemic uncertainty [117], which can have a great impact on various uncertainties related to the choice of model structure.
Efficient estimation of these two types of uncertainties can be achieved within the framework of Bayesian neural networks [119]. One way to accomplish this is in the last layer of the Bayesian neural network and by providing the mean and variance, e.g. over the image classification result. The method of Kendall and Gal [119] estimates the aleatoric and epistemic uncertainties by constructing a Bayesian neural network with the last layer before activation consisting of mean and variance of logits. In this method f ω ( x ) = ( μ , σ 2 ) represents the last 2K-dimensional pre-activated linear output of the neural network, where μ and σ 2 are the mean and the variance of K nodes. For the realised weights ω t ^ t = 1 T , with new input x * , predicted mean and variance denoted by hat, with corresponding outputs f ω t ^ ( x * ) = ( μ t ^ , σ t 2 ^ ) [120], an estimator calculates the predicted uncertainty as a sum of the two types of uncertainty
P r e d i c t e d U n c e r t a i n t y = A l e a t o r i c U n c e r t a i n t y + E p i s t e m i c U n c e r t a i n t y .
There are different ways to express the aleatoric and epistemic uncertainty terms from (1) and these aspects are currently a subject of intensive research.
The next subsection discusses common adversarial attacks that are intentionally introduced to disturb deep learning methods.

3.1.2. Adversarial Attacks

In addition to the uncertainties introduced above, another type of uncertainty can be added intentionally into models, which could disrupt significantly security systems such as surveillance systems. These uncertainties can be caused by changes made to the video or image data, which are often called adversarial attacks [121,122,123,124]. These intentional disturbances could have a great impact on how much people trust the results output by the models. As shown in Figure 21, even adding a little change to the original image, the result can be completely different. Many machine learning methods, including state-of-the-art neural networks, are vulnerable to adversarial instances [118].

Adversarial Samples

An example showing that a machine learning model makes an incorrect prediction as a result of tiny, deliberate feature changes is shown on Figure 21. These changes are created by diverse technologies that are especially effective for machine learning models, while they are not comprehensible to human eyes. Adversarial attacks are normally formed by intentionally added small perturbations, which can cause an incorrect result with high confidence [125]. These perturbations are commonly too small to be perceptible for deep learning models. Moreover, the same image perturbations can fool different kinds of deep learning network classifiers [126].
Adversarial attacks can be generated intentionally in surveillance systems in rail networks. This technology is a huge potential hazard to security monitoring systems. For instance, adversarial attacks can be used to turn the abnormal behaviours detected by the camera sensors into normal behaviours. It also can be used to fool the facial recognition systems, turning a dangerous criminal into another totally different person, or even worse, breaking into surveillance system centres to achieve full control of the monitoring systems.
Generally, there are two stages, the training stage and testing stage, where the adversarial attacks can disturb the final results of the deep learning models. White-box attacks and black-box attacks are two types of adversarial attacks in the testing stage [128]. White-box attacks refer to the circumstance where the adversaries have the full acknowledgement of the target models, including its structure, data distribution, and model parameters. However, black-box adversarial attacks can create adversarial examples without access to the gradient and parameters of the underlying deep neural networks. Different defence strategies are more efficient when facing different types of attacks [129].

3.1.3. Impact of COVID-19

The CCTV systems are planned, designed and deployed to keep the objectives and the assessed risks in view simultaneously. Sometimes, unforeseen events occur that are not expected to be handled by the system. The spread of COVID-19 and its impact on the world is a good example of such a rare event. The railway network has been one of the worst-hit industries, as the way we travel, visit or commute through the station was radically changed. During the pandemic, a range of options by which CCTV and related technologies could assist has been developed. Examples from around the world are summarised in this section.
Early in the pandemic, airports began routine walk-through temperature checks for passengers [130], and this is a readily transferable approach for a CCTV system with thermal camera capability (however, an estimated 46% of positive cases would not be detected). If audio recording is available, methods developed by researchers such as at the Carnegie Mellon University [131] or Cambridge University [132] could be deployed based on audio samples of COVID-19 positive and negative patients. The research aimed to detect a COVID-19 patient based on cough, breathing, and voice. The reported accuracy of one such algorithm is above 90% [133]. Face masks and social distancing have become mandatory measurements in many countries to prevent the spread of the virus. Some aspects of video analytics, such as VFR, are affected by the introduction of face masks. The accuracy of the state-of-the-art DNN-based VFR algorithms relies on how well the training data represents real data. These algorithms are trained on faces without masks and may give inaccurate results for masked faces. To erase these uncertainties caused by the inconsistency between training and test data, AI models need to be equipped with advanced deep learning algorithms such as Bayesian Deep Learning methods. Few public datasets of faces with masks are available [134], which can be used to re-train the AI models. A proposed VFR algorithm has improved the accuracy of face recognition with and without masks to 95% [134].
The CCTV system could also automatically detect people with and without masks using video analytics. A DNN-based algorithm, named RetinaMask, can provide greater than 90% accuracy in detecting and localising the face and the mask [135]. Some examples of the detection results are shown in Figure 22. The green and red boxes indicate a face with and without a mask, respectively. The algorithm was shown to perform accurately in confusing situations, e.g., faces covered with hands, hair or even tissues are detected as faces without a mask (red box). Audio analytics has also been used to detect if an individual is speaking through a face mask or not [136].
Other algorithms include crowd analysis to detect whether social distancing is under observation and raise the alarm for areas when it is too crowded for distancing to be effective. However, these are highly dynamic situations. In most cases, observation through surveillance is more useful for understanding trends in behaviour at the population or location level rather than supporting an immediate response. Any response would also need careful consideration of the legal basis of any intervention.

3.2. Uncertainty Quantification with Bayesian Deep Learning Methods

Probabilistic models, also known as Bayesian models, are the main methods for formulating the uncertainty [137,138]. The core idea of Bayesian modelling is, instead of calculating the certain values of weights parameters in neural networks, inferring posterior distribution of weights parameters by Bayesian probability theory. Thus, the certain prediction result output by deep learning models will be replaced by a result with a confidence interval. Combining with a particular rule such as an uncertainty score, a deep learning model with Bayesian methods will be able to tell whether the output result is reliable [139,140,141,142,143,144].
The main challenge of machine learning models such as the Gaussian Process is that they do not scale to big data very well, which makes them not applicable to deep learning models with millions of weights parameters. In order to investigate the uncertainties in computer vision tasks, which normally would have a significantly big image or video data, Bayesian Deep Learning (BDL) methods [116,145,146] are the solutions.
Different from CNNs, Bayesian Neural Networks (BNNs) [138,147] aim to learn the proposed posterior distribution over the weight parameters. The history of BNNs can be traced back to the 1990s [137,138]. These works extended the proposed models to small data sets, providing a robustness to over-fitting problems and uncertainty estimates. Over the decades, a lot of works have been focused on improving the scalability and explainable uncertainties of the existing deep neural networks [47,48,147,148,149].
In Bayesian Neural Networks (BNNs), instead of calculating the certain value of weight parameters ω in neural networks, the goal of the optimising process is to build the prior distributions p ( ω ) by following the Bayesian theory and then using this distribution to estimate which parameters are most likely to have created the data. To capture epistemic uncertainty in a neural network, [119] assumes a prior distribution over the weights, e.g., p ( ω ) can be chosen to be a Gaussian distribution. The posterior distribution over the space of parameters by the Bayesian theorem can be denoted as [118,119]
p ( ω | X , Y ) = p ( Y | X , ω ) p ( ω ) p ( Y | X )
where for classification prediction purposes X = { x 1 , x N } is the training input data and their corresponding classes Y = { y 1 , y N } , p ( Y | X , ω ) is the likelihood function. p ( ω ) is the known prior probability. The aim is using (2) to optimise ω to produce the desired classification output. However, in practice the true posterior distribution p ( ω | X , Y ) given by Equation (2) is hard to evaluate analytically in most real cases. The solution is to define an approximating variational distribution q θ ( ω ) to be as close as possible to the original posterior distribution.
Variational inference (VI) is the basic approximate inference method for approximating the posterior distribution. However, the biggest challenge of VI is that it is difficult to apply to complex deep learning models aiming to solve real-world problems. Based on VI, more research has been focused on applying various advanced approximate inference methods to deep learning models to solve more complex tasks. Some popular researches are Bayes by Backprop [137,150,151,152,153], Monte-Carlo Dropout (MC Dropout) [151,154], Variational Autoencoders (VAE) [105,155], Natural-Parameter Networks (NPN) [156], Bayesian GAN [157,158], and Subnetwork Inference [159].
To avoid incorrect predictions when performing deep learning algorithms, it is important to take uncertainty into account. Bayesian deep learning methods have been applied to solve various computer vision tasks, including but not limited to image/video retrieval [139,160], depth estimation [140,141], object detection [103,161,162], semantic segmentation and scene understanding [142,163,164,165], optical flow estimation and motion prediction [143,166,167], human pose estimation and pedestrian localisation [144,168,169], person re-identification and face recognition [170], camera re-localsation [171], and avoiding adversarial attacks [172,173].
The uncertainty measurement abilities of Bayesian methods have been demonstrated to be useful for accurate object detection [103]. Since large-scale object detection datasets have ambiguities when labelling the bounding boxes, a novel bounding box regression loss function for learning bounding box transformation and localisation variance is proposed to improve the accuracy of localising the objects.
Bayesian Deep Learning [116] offers a promising framework to handle the two types of uncertainties in a general way. Since deep learning networks are mostly over-parameterised, it is a great challenge to handle the uncertainties efficiently in such a large parameter network. By placing a prior over hidden units, which composes the neural networks, Bayesian deep learning can reduce the overfitting problem with advanced regularisation methods, especially when we have insufficient data [154].

4. About the Next Generation of CCTV Surveillance Systems for Railway Stations

4.1. Data Centre

Often, advanced computing and recording systems are located at the periphery of train stations or at remote control centres [174]. The control centres communicate, store and process the big data from sensors. These centres also provide simultaneous storage and software services to multiple CCTV systems. The multiple sensor data are processed, fused, analysed and communicated to the CCTV operator through the monitoring system with appropriate warnings and decision cues in real-time. The concept is shown in Figure 23.
Future CCTV surveillance will have additional capacities, including augmented and virtual reality systems, advanced cloud computing and autonomous recognition and warning technologies. Figure 23 shows that such a CCTV system could be installed in two geographically distant stations communicating with a data centre for data processing and storage. This architecture enables economical use of the resources with the concept of (common) data centres, which are expensive to deploy and maintain. With the fast development of Cloud Computing and 5G networks [175,176] and technologies combined with AI, the function and performance of surveillance systems will be significantly enhanced. Under a deployed 5G network, the data recorded and developed by live sensors could be transformed fast and wirelessly and then stored in a Cloud Centre. A data centre will be an essential part of future surveillance systems.

4.2. Data Analytics

Another important aspect of future CCTV systems is the need of various and efficient methods for analysing the different types of sensor data both in real time and off-line. The data analytics are divided into four intelligence levels, as shown in Figure 24. The intelligence level increases from the bottom to the top, i.e., detection to decision making. The detection is provided by both conventional and non-conventional sensors. The higher-level intelligence is achieved through multisensor data fusion. The video data are important from the presentation point of view as the fused data are presented to the operator on top of the video data. The recent research is focused on autonomous and semi-autonomous detection, tracking, behaviour understanding and activity recognition [33,177,178]. The state-of-the-art data analytics are even smart enough to understand some types of scenes and record the time duration. The future decades will see progress in scene understanding and decision-making. The CCTV systems should be installed while keeping in mind the frequent software and hardware updates of the data analytics system.
The data presentation and control in the monitoring systems is another domain that will see a lot of enhancements in the future. With the increase in the number and modalities of the sensors, the monitoring stations would be upgraded for ease of the security operations. In current systems, the data of all the sensors are presented either on separate screens or in separate windows of the same screen. The operator manually selects different cameras to obtain situational awareness. For this, the operator remembers the building layout and the location of each camera to relate the video to the location within the region of interest. This task requires mental capacity, training and experience.
Future systems may blend real-world data feeds with 3D virtual models of an environment to provide the 3D situational awareness of the sensor data. Such systems are called augmented virtuality approaches [179]. Localisation and tracking of data can be matched between the real and virtual worlds.
The integration of a 3D virtual model of a real environment with sensor feeds from that real environment can draw inspiration from the ideas in a geographic information system (GIS). The sensors are geolocated and produce georeferenced data. On detection of an important event, a geolocated movement sensor, for example, could signal the main system to activate the corresponding CCTV camera for video coverage and recording and also engage any available operator’s attention. Another option would be for an operator to select an area of an environment and for the system to automatically direct all relevant sensors on that area, e.g., PTZ cameras. In addition, event prediction in blind zones could be performed, as illustrated in Figure 25.
Rail networks, especially railway stations, are an important social and economic pillar. While it is important to take state-of-the-art security measures when installing the latest CCTV systems, it is important to keep track of the financial implications. The best solution requires reasonable emerging technologies for the future systems in mind. In addition, cybersecurity measures need to be considered when implementing a surveillance network.

5. Ethics and General Data Protection Regulatory

Computer and information technologies can bring advantages to not only railway networks but also other computer smart cities and public transportation systems, while they will also raise ethical concerns about computer and information security and privacy [181]. Due to the fast development of new networked technologies, such as facial recognition (FR), some research aims to find and raise public awareness of ethical issues lying in urban big data analytics and public transportation systems [182,183,184,185]. Some researchers have been working on providing ethical frameworks for big data and smart cities focusing on contemporary ethical and non-ethical issues in big data analytics applications in smart cities and public transportation systems [183]. There are four emerging issues considered to exist commonly in all smart cities: privacy and surveillance, data integrity, social equity and the ageing population and data collection and analytics in public transportation [183].
The work of video analytics in rail network surveillance relies on the video data collected through CCTV systems. As video surveillance is enhanced due to the advancements in technology, it can produce a vast amount of personalised raw and analysed data. One of the main ethical concerns in the domain of CCTV is privacy. For example, video data used to identify a person using facial and gait features can cause a leak of personal information. Additionally, video analytics could identify an individual’s interactions with other individuals and objects. It helps in automatic monitoring on the one hand, but individual privacy is compromised on the other hand. Personal privacy is an important component of human rights, but maintaining safety and security in justifying an interference is subject to evolving interpretation. The privacy breach is not the only ethical consideration. The sense of being watched by a CCTV system could also constrain individual decisions and behaviour. Moreover, individuals may also be worried and act nervously to avoid being falsely detected as suspicious individuals.
In 2018, many countries enacted a law to protect individual privacy. The European Union General Data Protection Regulation (GDPR) took effect based on the six basic principles [186], as given in Figure 26. The UK implemented GDPR through the Data Protection Act (DPA) 2018, subsequently amending this to UK-GDPR. Under such rules, video data are considered personal data when it can be used to identify the person.
Smart surveillance is an area of evolving interpretation, particularly for automated facial recognition. In the UK, the Information Commissioner gave an opinion about the use of live facial recognition in 2019 [187], and in 2020, the use of this by the South Wales Police was ruled unlawful [188]. Another aspect of ethical concern is ensuring the use is for identifying specific people of concern rather than conducting mass surveillance and demonstrating that the software used does not have a racial or gender bias in its identification processes. Knowledge of the training datasets used in the production of the facial identification software would be crucial in ensuring freedom from these biases, indicating the degree of diligence required of end-users before the legal operation of such systems.

6. Conclusions

This paper provides a survey of the latest trends in CCTV systems and presents a vision for the future of such systems. The ever-evolving nature of security hazards and threats can be countered through analysis and fusion of multi-modal sensor data. The CCTV system produces big data, which can overwhelm the operator if not presented efficiently. Georeferenced monitoring systems with an augmented virtuality display can ease the burden and improve the response of the operator. The document gives a detailed account of the video analytics aspects that aid in the security and management of the station. The data analytics of other sensors, such as audio and thermal sensors, can further enhance the capabilities of the CCTV system. The processing and storage of the big sensor data cannot be met through a centralised computing architecture. The alternative architectures, namely cloud, cloud–fog and cloud–fog–edge, can be implemented for cost-effectiveness. Prior to the operation of automated systems, consideration must be given to ensuring outcomes are free from racial or gender bias imparted by poorly selected training datasets and that personal data are handled according to legal requirements of data privacy.

Author Contributions

T.Z.—conceptualisation, methodology, results, writing, W.A.—conceptualisa- tion, methodology, results, writing, experimentation, L.M.—conceptualisation, methodology, writing, supervision, funding acquisition, C.L.-W.–conceptualisation, writing, review and editing; S.R.—writing—review and editing; supervision; funding acquisition, D.F.—writing—review and editing; supervision; funding acquisition, S.M., writing—review and editing; supervision; funding acquisition, G.B.—discussion of ideas, manuscript editing. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge the support of the EU Horizon 2020 “In2Stempo: Innovative Solutions in Future Stations, Energy Metering and Power Supply” project under grant agreement 777515. We are grateful to EPSRC for funding this work through EP/T013265/1 project NSF-EPSRC: “ShiRAS. Towards Safe and Reliable Autonomy in Sensor Driven Systems” and the support for ShiRAS by the USA National Science Foundation under Grant NSF ECCS 1903466. For the purpose of open access, the authors have applied a Creative Commons Attribution (CC BY) licence to any Author Accepted Manuscript version arising.Sensors 22 04324 i001

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Coaffee, J.; Moore, C.; Fletcher, D.; Bosher, L. Resilient design for community safety and terror-resistant cities. In Proceedings of the Institution of Civil Engineers-Municipal Engineer, London, UK, 1 June 2008; Thomas Telford Ltd.: London, UK, 2015; Volume 161, pp. 103–110. [Google Scholar]
  2. Media Guidelines for Reporting Suicide. 2019. Available online: https://media.samaritans.org/documents/Media_guidelines_-_Rail_suicides_factsheet_UK_Final.pdf (accessed on 7 April 2022).
  3. Suicide Prevention on the Railway—Network Rail. 2021. Available online: https://www.networkrail.co.uk/communities/safety-in-the-community/suicide-prevention-on-the-railway/ (accessed on 7 April 2022).
  4. Kawamura, A.; Yoshimitsu, Y.; Kajitani, K.; Naito, T.; Fujimura, K.; Kamijo, S. Smart camera network system for use in railway stations. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, Anchorage, AL, USA, 9–12 October 2011; pp. 85–90. [Google Scholar]
  5. Li, Y.; Qin, Y.; Xie, Z.; Cao, Z.; Jia, L.; Yu, Z.; Zheng, J.; Zhang, E. Efficient SSD: A Real-Time Intrusion Object Detection Algorithm for Railway Surveillance. In Proceedings of the 2020 International Conference on Sensing, Diagnostics, Prognostics, and Control (SDPC), Beijing, China, 5–7 August 2020; pp. 391–395. [Google Scholar]
  6. Langran-Wheeler, C.; Fletcher, D.; Rigby, S. In2Stempo D9.2. Database: Initial Structural Response Calculations for Glass, Fixings and Station Geometry; Technical Report; The University of Sheffield: Sheffield, UK, 2020. [Google Scholar]
  7. Davies, A.C.; Velastin, S.A. A Progress Review of Intelligent CCTV Surveillance Systems. In Proceedings of the 2005 IEEE Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Sofia, Bulgaria, 5–7 September 2005; pp. 417–423. [Google Scholar]
  8. Tsakanikas, V.; Dagiuklas, T. Video surveillance systems-current status and future trends. Comput. Electr. Eng. 2018, 70, 736–753. [Google Scholar] [CrossRef]
  9. Bashbaghi, S.; Granger, E.; Sabourin, R.; Parchami, M. Deep Learning Architectures for Face Recognition in Video Surveillance. In Deep Learning in Object Detection and Recognition; Springer: Berlin/Heidelberg, Germany, 2019; pp. 133–154. [Google Scholar]
  10. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep Learning for Generic Object Detection: A Survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef] [Green Version]
  11. Aggarwal, J.K.; Ryoo, M.S. Human Activity Analysis: A Review. ACM Comput. Surv. (CSUR) 2011, 43, 1–43. [Google Scholar] [CrossRef]
  12. Kong, Y.; Fu, Y. Human Action Recognition and Prediction: A Survey. arXiv 2018, arXiv:1806.11230. [Google Scholar] [CrossRef]
  13. Ciaparrone, G.; Sánchez, F.L.; Tabik, S.; Troiano, L.; Tagliaferri, R.; Herrera, F. Deep Learning in Video Multi-Object Tracking: A Survey. Neurocomputing 2020, 381, 61–88. [Google Scholar] [CrossRef] [Green Version]
  14. Smeureanu, S.; Ionescu, R.T. Real-Time Deep Learning Method for Abandoned Luggage Detection in Video. In Proceedings of the 26th European Signal Processing Conference (EUSIPCO), Eternal, Italy, 3–7 September 2018; pp. 1775–1779. [Google Scholar]
  15. Chalapathy, R.; Chawla, S. Deep Learning for Anomaly Detection: A Survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
  16. Tripathi, G.; Singh, K.; Vishwakarma, D.K. Convolutional Neural Networks for Crowd Behaviour Analysis: A Survey. Visual Comput. 2019, 35, 753–776. [Google Scholar] [CrossRef]
  17. Gao, G.; Gao, J.; Liu, Q.; Wang, Q.; Wang, Y. CNN-based Density Estimation and Crowd Counting: A Survey. arXiv 2020, arXiv:2003.12783. [Google Scholar]
  18. Aria, M.; Cuccurullo, C. bibliometrix: An R-tool for comprehensive science mapping analysis. J. Inf. 2017, 11, 959–975. [Google Scholar] [CrossRef]
  19. National Security Inspectorate. NSI Code of Practice for Design, Installation and Maintenance of Access Control Systems; Technical Report November; National Security Inspectorate: Maidenhead, UK, 2017. [Google Scholar]
  20. The British Standards Institution. Closed Circuit Television (CCTV)—Management and Operation Code of Practice; Technical Report BS 7958:2009; BSI Standards Limited: London, UK, 2009. [Google Scholar]
  21. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 1–2: System Requirements Performance Requirements for Video Transmission; Technical Report BS EN 62676-1-2-2014; BSI Standards Limited: London, UK, 2016. [Google Scholar]
  22. The British Standards Institution. Installation and Remote Monitoring of Detector-Activated CCTV Systems—Code of Practice; Technical Report BS 8418-2015+A1-2017; BSI Standards Limited: London, UK, 2017. [Google Scholar]
  23. The British Standards Institution. Remote Centres Receiving Signals from Alarm Systems—Code of Practice; Technical Report BS 8591-2014; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  24. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 2–1: Video Transmission Protocols General Requirements; Technical Report BS EN 62676-2-1-2014; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  25. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 2–3: Video Transmission Protocols—IP Interoperability Implementation Based on Web Services; Technical Report BS EN 62676-2-3-2014; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  26. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 3: Analog and Digital Video Interfaces; Technical Report BS EN 62676-3-2015; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  27. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 1–1: System Requirements General; Technical Report BS EN 62676-1-1:2014; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  28. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 2–2: Video Transmission Protocols IP Interoperability Implementation Based on HTTP and REST Services; Technical Report BS EN 62676-2-2-2014; BSI Standards Limited: London, UK, 2014. [Google Scholar]
  29. The British Standards Institution. Video Surveillance Systems for Use in Security Applications Part 4: Application Guidelines; Technical Report BS EN 62676-4:2015; BSI Standards Limited: London, UK, 2015. [Google Scholar]
  30. Ferryman, J. Video Surveillance Standardisation Activities, Process and Roadmap; Technical Report JRC103650; Joint Research Centre (JRC), Science Hub, European Union: Brussels, Belgium, 2016. [Google Scholar]
  31. Marcenaro, L. Access to Data Sets; Technical Report; Erncip Thematic Group on Video Surveillance for Security of Critical Infrastructure: Ispra, Italy, 2016. [Google Scholar]
  32. Medel, J.R.; Savakis, A. Anomaly Detection in Video Using Predictive Convolutional Long Short-Term Memory Networks. arXiv 2016, arXiv:1612.00390. [Google Scholar]
  33. Shidik, G.F.; Noersasongko, E.; Nugraha, A.; Andono, P.N.; Jumanto, J.; Kusuma, E.J. A Systematic Review of Intelligence Video Surveillance: Trends, Techniques, Frameworks, and Datasets. IEEE Access 2019, 7, 170457–170473. [Google Scholar] [CrossRef]
  34. Gautam, A.; Singh, S. Trends in Video Object Tracking in Surveillance: A Survey. In Proceedings of the Third International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Palladam, India, 12–14 December 2019; pp. 729–733. [Google Scholar]
  35. Ma, H.; Tian, J.; Qiu, K.; Lo, D.; Gao, D.; Wu, D.; Jia, C.; Baker, T. Deep-Learning–Based App Sensitive Behavior Surveillance for Android Powered Cyber–Physical Systems. IEEE Trans. Ind. Inform. 2021, 17, 5840–5850. [Google Scholar] [CrossRef]
  36. Koch, W. Towards Cognitive Tools: Systems Engineering Aspects for Public Safety and Security. IEEE Aerosp. Electron. Syst. Mag. 2014, 29, 14–26. [Google Scholar] [CrossRef]
  37. Scanner, L.A. Flickr. 2018. Available online: https://www.flickr.com/photos/36016325@N04/39682091582 (accessed on 7 April 2022).
  38. Sun, G.; Nakayama, Y.; Dagdanpurev, S.; Abe, S.; Nishimura, H.; Kirimoto, T.; Matsui, T. Remote Sensing of Multiple Vital Signs Using a CMOS Camera-Equipped Infrared Thermography System and Its Clinical Application in Rapidly Screening Patients with Suspected Infectious Diseases. Int. J. Infect. Dis. 2017, 55, 113–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Skinner, D. Flickr. 2015. Available online: https://www.flickr.com/photos/28051469@N02/23783501012 (accessed on 7 April 2022).
  40. Yao, R.; Lin, G.; Xia, S.; Zhao, J.; Zhou, Y. Video object segmentation and tracking: A survey. In ACM Transactions on Intelligent Systems and Technology (TIST); ACM: New York, NY, USA, 2020; Volume 11, pp. 1–47. [Google Scholar]
  41. Aafaq, N.; Mian, A.; Liu, W.; Gilani, S.Z.; Shah, M. Video description: A survey of methods, datasets, and evaluation metrics. In AACM Computing Surveys (CSUR); ACM: New York, NY, USA, 2019; Volume 52, pp. 1–37. [Google Scholar]
  42. Horn, B.K.; Schunck, B.G. Determining optical flow. In Proceedings of the Techniques and Applications of Image Understanding, International Society for Optics and Photonics, Washington, DC, USA, 21–23 April 1981; Volume 281, pp. 319–331. [Google Scholar]
  43. Liu, M.; Wu, C.; Zhang, Y. Motion Vehicle Tracking Based on Multi-Resolution Optical Flow and Multi-Scale Harris Corner Detection. In Proceedings of the 2007 IEEE International Conference on Robotics and Biomimetics (ROBIO), Sanya, China, 15–18 December 2007; pp. 2032–2036. [Google Scholar]
  44. Roth, S.; Black, M.J. On the Spatial Statistics of Optical Flow. Int. J. Comput. Vis. 2007, 74, 33–50. [Google Scholar] [CrossRef]
  45. Rasheed, N.; Khan, S.A.; Khalid, A. Tracking and Abnormal Behavior Detection in Video Surveillance Using Optical Flow and Neural Networks. In Proceedings of the 2014 28th International Conference on Advanced Information Networking and Applications Workshops, Victoria, BC, Canada, 13–16 May 2014; pp. 61–66. [Google Scholar]
  46. Bui, T.; Hernández-Lobato, D.; Hernandez-Lobato, J.; Li, Y.; Turner, R. Deep Gaussian processes for regression using approximate expectation propagation. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 1472–1481. [Google Scholar]
  47. Nguyen-Tuong, D.; Peters, J.R.; Seeger, M. Local Gaussian Process Regression for Real Time Online Model Learning. In Proceedings of the Advances in Neural Information Processing Systems, Nice, France, 22–26 September 2008; pp. 1193–1200. [Google Scholar]
  48. Nguyen-Tuong, D.; Seeger, M.; Peters, J. Model Learning with Local Gaussian Process Regression. Adv. Robot. 2009, 23, 2015–2034. [Google Scholar] [CrossRef] [Green Version]
  49. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, CA, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  50. Kalchbrenner, N.; Grefenstette, E.; Blunsom, P. A convolutional neural network for modelling sentences. arXiv 2014, arXiv:1404.2188. [Google Scholar]
  51. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face recognition: A convolutional neural-network approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef] [Green Version]
  52. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional networks and applications in vision. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 17 July 2009; pp. 253–256. [Google Scholar]
  53. Kortli, Y.; Jridi, M.; Al Falou, A.; Atri, M. Face recognition systems: A survey. Sensors 2020, 20, 342. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Chavan, V.D.; Bharate, A.A. A Review Paper on Face Detection and Recognition in Video. Int. J. Innov. Res. Electr. Electron. Instrum. Control Eng. 2016, 4, 97–100. [Google Scholar] [CrossRef]
  55. Samadiani, N.; Huang, G.; Cai, B.; Luo, W.; Chi, C.H.; Xiang, Y.; He, J. A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data. Sensors 2019, 19, 1863. [Google Scholar] [CrossRef] [Green Version]
  56. Giles, C.L.; Kuhn, G.M.; Williams, R.J. Dynamic recurrent neural networks: Theory and applications. IEEE Trans. Neural Netw. 1994, 5, 153–156. [Google Scholar] [CrossRef]
  57. Mandic, D.; Chambers, J. Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability; Wiley: Hoboken, NJ, USA, 2001. [Google Scholar]
  58. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  59. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  60. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  61. Patraucean, V.; Handa, A.; Cipolla, R. Spatio-temporal video autoencoder with differentiable memory. arXiv 2015, arXiv:1511.06309. [Google Scholar]
  62. Chong, Y.S.; Tay, Y.H. Abnormal event detection in videos using spatiotemporal autoencoder. In Lecture Notes in Computer Science, Proceedings of the International Symposium on Neural Networks, Hokkaido, Japan, 21–26 June 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 189–196. [Google Scholar]
  63. Zhao, Y.; Deng, B.; Shen, C.; Liu, Y.; Lu, H.; Hua, X.S. Spatio-temporal autoencoder for video anomaly detection. In Proceedings of the 25th ACM International Conference on Multimedia, New York, NY, USA, 23–27 October 2017; pp. 1933–1941. [Google Scholar]
  64. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 7–9 December 2017; pp. 5998–6008. [Google Scholar]
  65. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. arXiv 2020, arXiv:2012.12556. [Google Scholar] [CrossRef]
  66. Arnab, A.; Dehghani, M.; Heigold, G.; Sun, C.; Lučić, M.; Schmid, C. Vivit: A video vision transformer. arXiv 2021, arXiv:2103.15691. [Google Scholar]
  67. Zhou, D.; Kang, B.; Jin, X.; Yang, L.; Lian, X.; Jiang, Z.; Hou, Q.; Feng, J. Deepvit: Towards deeper vision transformer. arXiv 2021, arXiv:2103.11886. [Google Scholar]
  68. Chapel, M.; Bouwmans, T. Moving Objects Detection with a Moving Camera: A Comprehensive Review. Comput. Sci. Rev. 2020, 38, 100310. [Google Scholar] [CrossRef]
  69. Zhu, H.; Wei, H.; Li, B.; Yuan, X.; Kehtarnavaz, N. Real-time moving object detection in high-resolution video sensing. Sensors 2020, 20, 3591. [Google Scholar] [CrossRef]
  70. Yazdi, M.; Bouwmans, T. New trends on moving object detection in video images captured by a moving camera: A survey. Comput. Sci. Rev. 2018, 28, 157–177. [Google Scholar] [CrossRef]
  71. Liu, H.; Meng, W.; Liu, Z. Key frame extraction of online video based on optimized frame difference. In Proceedings of the 2012 9th International Conference on Fuzzy Systems and Knowledge Discovery, Chongqing, China, 15 January 2012; pp. 1238–1242. [Google Scholar]
  72. Lei, M.; Geng, J. Fusion of Three-frame Difference Method and Background Difference Method to Achieve Infrared Human Target Detection. In Proceedings of the IEEE 1st International Conference on Civil Aviation Safety and Information Technology (ICCASIT), Kunming, China, 17–19 October 2019; pp. 381–384. [Google Scholar]
  73. Srinivasan, K.; Porkumaran, K.; Sainarayanan, G. Improved background subtraction techniques for security in video applications. In Proceedings of the 3rd International Conference on Anti-Counterfeiting, Security, and Identification in Communication, Hong Kong, China, 20–22 August 2009; pp. 114–117. [Google Scholar]
  74. Mohamed, S.S.; Tahir, N.M.; Adnan, R. Background modelling and background subtraction performance for object detection. In Proceedings of the 2010 6th International Colloquium on Signal Processing & Its Applications, Malacca City, Malaysia, 21–23 May 2010; pp. 1–6. [Google Scholar]
  75. Xu, D.; Xie, W.; Zisserman, A. Geometry-Aware Video Object Detection for Static Cameras. arXiv 2019, arXiv:1909.03140. [Google Scholar]
  76. Rehman Butt, W.U.; Servin, M.; Samara, K.; Al Rahman, E.A.; Kouki, S.; Bouchahma, M. Static and Moving Object Detection and Segmentation in Videos. In Proceedings of the 2019 Sixth HCT Information Technology Trends (ITT), Ras Al Khaimah, United Arab Emirates, 20–21 November 2019; pp. 197–201. [Google Scholar]
  77. Ariffa Begum, S.; Askarunisa, A. Performance Analysis of Machine Learning Classification Algorithms in Static Object Detection for Video Surveillance Applications. Wirel. Pers. Commun. 2020, 115, 1291–1307. [Google Scholar] [CrossRef]
  78. CAVIAR: Context Aware Vision Using Image-Based Active Recognition. 2006. Available online: https://homepages.inf.ed.ac.uk/rbf/CAVIAR (accessed on 7 April 2022).
  79. Luo, W.; Xing, J.; Milan, A.; Zhang, X.; Liu, W.; Kim, T.K. Multiple object tracking: A literature review. Artif. Intell. 2021, 293, 103448. [Google Scholar] [CrossRef]
  80. Meyerhoff, H.S.; Papenmeier, F.; Huff, M. Studying visual attention using the multiple object tracking paradigm: A tutorial review. Atten. Percept. Psychophys 2017, 79, 1255–1274. [Google Scholar] [CrossRef]
  81. Isupova, O.; Kuzin, D.; Mihaylova, L. Learning Methods for Dynamic Topic Modeling in Automated Behavior Analysis. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3980–3993. [Google Scholar] [CrossRef] [Green Version]
  82. Zaidi, S.S.A.; Ansari, M.S.; Aslam, A.; Kanwal, N.; Asghar, M.N.; Lee, B.A. A Survey of Modern Deep Learning based Object Detection Models. arXiv 2021, arXiv:2104.11892. [Google Scholar] [CrossRef]
  83. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  84. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  85. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  86. Shah, A.P.; Lamare, J.B.; Nguyen-Anh, T.; Hauptmann, A. CADP: A Novel Dataset for CCTV Traffic Camera Based Accident Analysis. In Proceedings of the 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; pp. 1–9. [Google Scholar]
  87. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 1–26 July 2016; pp. 779–788. [Google Scholar]
  88. Prakash, U.M.; Thamaraiselvi, V.G. Detecting and tracking of multiple moving objects for intelligent video surveillance systems. In Proceedings of the Second International Conference on Current Trends In Engineering and Technology—ICCTET 2014, Coimbatore, India, 8 July 2014; pp. 253–257. [Google Scholar]
  89. Mask Dataset. Available online: https://makeml.app/datasets/mask (accessed on 7 April 2022).
  90. Al-Obaydy, W.N.I.; Suandi, S.A. Open-Set Face Recognition in Video Surveillance: A Survey. In InECCE2019; Springer: Berlin/Heidelberg, Germany, 2020; pp. 425–436. [Google Scholar]
  91. How Facial Recognition Could Replace Train Tickets. 2017. Available online: https://www.bbc.co.uk/news/av/technology-40676084 (accessed on 7 April 2022).
  92. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C. Deep Learning for Person Re-identification: A Survey and Outlook. Image Vis. Comput. 2022, 119, 104394. [Google Scholar] [CrossRef]
  93. Wang, H.; Hu, J.; Zhang, G. Multi-source transfer network for cross domain person re-identification. IEEE Access 2020, 8, 83265–83275. [Google Scholar] [CrossRef]
  94. Sekh, A.A.; Dogra, D.P.; Choi, H.; Chae, S.; Kim, I.J. Person Re-identification in Videos by Analyzing Spatio-temporal Tubes. Multimed. Tools Appl. 2020, 79, 24537–24551. [Google Scholar] [CrossRef]
  95. Wu, L.F.; Wang, Q.; Jian, M.; Qiao, Y.; Zhao, B.X. A Comprehensive Review of Group Activity Recognition in Videos. Int. J. Autom. Comput. 2021, 18, 334–350. [Google Scholar] [CrossRef]
  96. Rodríguez-Moreno, I.; Martínez-Otzeta, J.M.; Sierra, B.; Rodriguez, I.; Jauregi, E. Video Activity Recognition: State-of-the-art. Sensors 2019, 19, 3160. [Google Scholar] [CrossRef] [Green Version]
  97. Hosmer, P. i-LIDS bag and vehicle detection challenge. In Proceedings of the 2007 IEEE International Conference on Advanced Video and Signal based Surveillance (AVSS 2007), London, UK, 5–7 September 2007; Available online: http://www.eecs.qmul.ac.uk/~andrea/avss2007_ss_challenge.html (accessed on 7 April 2022).
  98. Van Beeck, K.; Van Engeland, K.; Vennekens, J.; Goedemé, T. Abnormal Behavior Detection in LWIR Surveillance of Railway Platforms. In Proceedings of the 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar]
  99. Geng, X.; Li, G.; Ye, Y.; Tu, Y.; Dai, H. Abnormal Behavior Detection for Early Warning of Terrorist Attack. In Proceedings of the Australasian Joint Conference on Artificial Intelligence, Hobart, Australia, 4–8 December 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1002–1009. [Google Scholar]
  100. Zhu, S.; Chen, C.; Sultani, W. Video Anomaly Detection for Smart Surveillance. arXiv 2020, arXiv:2004.00222. [Google Scholar]
  101. Mohammadi, B.; Fathy, M.; Sabokrou, M. Image/Video Deep Anomaly Detection: A Survey. arXiv 2021, arXiv:2103.01739. [Google Scholar]
  102. Hasan, M.; Choi, J.; Neumann, J.; Roy-Chowdhury, A.K.; Davis, L.S. Learning Temporal Regularity in Video Sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 1–26 June 2016; pp. 733–742. [Google Scholar]
  103. Luo, W.; Liu, W.; Gao, S. Remembering History with Convolutional Lstm for Anomaly Detection. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 439–444. [Google Scholar]
  104. Gong, D.; Liu, L.; Le, V.; Saha, B.; Mansour, M.R.; Venkatesh, S.; Hengel, A.V.D. Memorizing Normality to Detect Anomaly: Memory-augmented Deep Autoencoder for Unsupervised Anomaly Detection. In Proceedings of the IEEE International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 1705–1714. [Google Scholar]
  105. Mescheder, L.; Nowozin, S.; Geiger, A. Adversarial variational bayes: Unifying variational autoencoders and generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, International Convention Centre, Sydney, Australia, 6–11 August 2017; pp. 2391–2400. [Google Scholar]
  106. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  107. Nishino, K.; Inaba, M. Bayesian AutoEncoder: Generation of Bayesian networks with hidden nodes for features. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  108. Ye, M.; Peng, X.; Gan, W.; Wu, W.; Qiao, Y. Anopcn: Video Anomaly Detection via Deep Predictive Coding Network. In Proceedings of the 27th ACM International Conference on Multimedia, New York, NY, USA, 21–25 October 2019; pp. 1805–1813. [Google Scholar]
  109. Anjum, N.; Cavallaro, A. Unsupervised Fuzzy Clustering for Trajectory Analysis. In Proceedings of the 2007 IEEE International Conference on Image Processing, San Antonio, TX, USA, 16–19 September 2007; Volume 3, pp. III-213–III-216. [Google Scholar]
  110. Zhang, H.; Ma, L.; Qian, L. Trajectory optimization by Gauss pseudospectral method and it’s realization in flight. In Proceedings of the 33rd Chinese Control Conference, Nanjing, China, 28–30 July 2014; pp. 1007–1012. [Google Scholar]
  111. Jin, Z.; Bhanu, B. Analysis-by-synthesis: Pedestrian tracking with crowd simulation models in a multi-camera video network. Comput. Vis. Image Underst. 2015, 134, 48–63. [Google Scholar] [CrossRef] [Green Version]
  112. Haghani, M.; Sarvi, M. Crowd behaviour and motion: Empirical methods. Transp. Res. Part Methodol. 2018, 107, 253–294. [Google Scholar] [CrossRef]
  113. Ahmed, S.A.; Dogra, D.P.; Kar, S.; Roy, P.P. Surveillance Scene Representation and Trajectory Abnormality Detection Using Aggregation of Multiple Concepts. Expert Syst. Appl. 2018, 101, 43–55. [Google Scholar] [CrossRef]
  114. Solmaz, B.; Moore, B.E.; Shah, M. Identifying Behaviors in Crowd Scenes Using Stability Analysis for Dynamical Systems. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2064–2070. [Google Scholar] [CrossRef]
  115. Yuan, Y.; Feng, Y.; Lu, X. Structured Dictionary Learning for Abnormal Event Detection in Crowded Scenes. Pattern Recognit. 2018, 73, 99–110. [Google Scholar] [CrossRef]
  116. Wang, H.; Yeung, D.Y. A survey on Bayesian deep learning. ACM Comput. Surv. (CSUR) 2020, 53, 1–37. [Google Scholar] [CrossRef]
  117. Hüllermeier, E.; Waegeman, W. Aleatoric and epistemic uncertainty in machine learning: An introduction to concepts and methods. Mach. Learn. 2021, 110, 457–506. [Google Scholar] [CrossRef]
  118. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamzadeh, M.; Fieguth, P.; Khosravi, A.; Acharya, U.R.; Makarenkov, V.; et al. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf. Fusion 2021, 75, 243–297. [Google Scholar] [CrossRef]
  119. Kendall, A.; Gal, Y. What uncertainties do we need in bayesian deep learning for computer vision? arXiv 2017, arXiv:1703.04977. [Google Scholar]
  120. Kwon, Y.; Won, J.-H.; Kim, B.J.; Paik, M.C. Uncertainty quantification using Bayesian neural networks in classification: Application to biomedical image segmentation. Comput. Stat. Data Anal. 2020, 142, 106816. [Google Scholar] [CrossRef]
  121. Shukla, S.N.; Sahu, A.K.; Willmott, D.; Kolter, J.Z. Black-box adversarial attacks with bayesian optimization. arXiv 2019, arXiv:1909.13857. [Google Scholar]
  122. Su, L.; Vaidya, N.H. Defending non-Bayesian learning against adversarial attacks. Distrib. Comput. 2019, 32, 277–289. [Google Scholar] [CrossRef] [Green Version]
  123. Wan, X.; Kenlay, H.; Ru, R.; Blaas, A.; Osborne, M.; Dong, X. Adversarial Attacks on Graph Classifiers via Bayesian Optimisation. Adv. Neural Inf. Process. Syst. 2021, 34, 6983–6996. [Google Scholar]
  124. Ye, N.; Zhu, Z. Bayesian adversarial learning. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; pp. 6892–6901. [Google Scholar]
  125. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–11. [Google Scholar]
  126. Akhtar, N.; Mian, A. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
  127. Google AI Blog: Introducing the Unrestricted Adversarial Examples Challenge. Available online: https://ai.googleblog.com/2018/09/introducing-unrestricted-adversarial.html (accessed on 7 April 2022).
  128. Qiu, S.; Liu, Q.; Zhou, S.; Wu, C. Review of artificial intelligence adversarial attack and defense technologies. Appl. Sci. 2019, 9, 909. [Google Scholar] [CrossRef] [Green Version]
  129. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018—Conference Track Proceedings, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–28. [Google Scholar]
  130. Quilty, B.J.; Clifford, S.; CMMID nCoV Working Group; Flasche, S.; Eggo, R.M. Effectiveness of airport screening at detecting travellers infected with novel coronavirus (2019-nCoV). Euro Surveill. Bull. Eur. Sur Les Mal. Transm. 2020, 25, 2000080. [Google Scholar] [CrossRef]
  131. Index Page. 2022. Available online: https://cvd.lti.cmu.edu (accessed on 7 April 2022).
  132. Covid-19 Sounds App—University of Cambridge. 2022. Available online: https://www.covid-19-sounds.org/en (accessed on 7 April 2022).
  133. Deshpande, G.; Schuller, B. An Overview on Audio, Signal, Speech, & Language Processing for COVID-19. arXiv 2020, arXiv:2005.08579. [Google Scholar]
  134. Wang, Z.; Wang, G.; Huang, B.; Xiong, Z.; Hong, Q.; Wu, H.; Yi, P.; Jiang, K.; Wang, N.; Pei, Y.; et al. Masked Face Recognition Dataset and Application. arXiv 2020, arXiv:2003.09093. [Google Scholar]
  135. Jiang, M.; Fan, X. RetinaMask: A Face Mask detector. arXiv 2020, arXiv:2005.03950. [Google Scholar]
  136. Schuller, B.W.; Batliner, A.; Bergler, C.; Messner, E.M.; Hamilton, A.; Amiriparian, S.; Baird, A.; Rizos, G.; Schmitt, M.; Stappen, L.; et al. The Interspeech 2020 Computational Paralinguistics Challenge: Elderly Emotion, Breathing & Masks. In Proceedings of the Interspeech, Shanghai, China, 25–29 October 2020. [Google Scholar]
  137. David, J. A practical Bayesian framework for backprop networks. Neural Comput. 1992, 4, 448–472. [Google Scholar]
  138. Bai, S.; He, Z.; Lei, Y.; Wu, W.; Zhu, C.; Sun, M.; Yan, J. Traffic Anomaly Detection via Perspective Map based on Spatial-temporal Information Matrix. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 117–124. [Google Scholar]
  139. Taha, A.; Chen, Y.T.; Misu, T.; Shrivastava, A.; Davis, L. Unsupervised data uncertainty learning in visual retrieval systems. arXiv 2019, arXiv:1902.02586. [Google Scholar]
  140. Asai, A.; Ikami, D.; Aizawa, K. Multi-Task Learning based on Separable Formulation of Depth Estimation and its Uncertainty. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 21–24. [Google Scholar]
  141. Liu, C.; Gu, J.; Kim, K.; Narasimhan, S.G.; Kautz, J. Neural rgb (r) d sensing: Depth and uncertainty from a video camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10986–10995. [Google Scholar]
  142. Postels, J.; Ferroni, F.; Coskun, H.; Navab, N.; Tombari, F. Sampling-free epistemic uncertainty estimation using approximated variance propagation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 2931–2940. [Google Scholar]
  143. Gast, J.; Roth, S. Lightweight probabilistic deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3369–3378. [Google Scholar]
  144. Gundavarapu, N.B.; Srivastava, D.; Mitra, R.; Sharma, A.; Jain, A. Structured Aleatoric Uncertainty in Human Pose Estimation. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; Volume 2, p. 2. [Google Scholar]
  145. Ul Abideen, Z.; Ghafoor, M.; Munir, K.; Saqib, M.; Ullah, A.; Zia, T.; Tariq, S.A.; Ahmed, G.; Zahra, A. Uncertainty Assisted Robust Tuberculosis Identification With Bayesian Convolutional Neural Networks. IEEE Access 2020, 8, 22812–22825. [Google Scholar] [CrossRef]
  146. Park, J.G.; Jo, S. Bayesian Weight Decay on Bounded Approximation for Deep Convolutional Neural Networks. IEEE Trans. Neural. Netw. Learn Syst. 2019, 30, 2866–2875. [Google Scholar] [CrossRef]
  147. Zafar, U.; Ghafoor, M.; Zia, T.; Ahmed, G.; Latif, A.; Malik, K.R.; Sharif, A.M. Face Recognition with Bayesian Convolutional Networks for Robust Surveillance Systems. EURASIP J. Image Video Process. 2019, 2019, 10. [Google Scholar] [CrossRef] [Green Version]
  148. Gal, Y.; Ghahramani, Z. Bayesian convolutional neural networks with Bernoulli approximate variational inference. arXiv 2015, arXiv:1506.02158. [Google Scholar]
  149. Murray, S.M. An Exploratory Analysis of Multi-Class Uncertainty Approximation in Bayesian Convolutional Neural Networks. Master’s Thesis, The University of Bergen, Bergen, Norway, 2018. [Google Scholar]
  150. Blundell, C.; Cornebise, J.; Kavukcuoglu, K.; Wierstra, D. Weight Uncertainty in Neural Networks. arXiv 2015, arXiv:1505.05424. [Google Scholar]
  151. Kingma, D.P.; Salimans, T.; Welling, M. Variational Dropout and the Local Reparameterization Trick. arXiv 2015, arXiv:1506.02557. [Google Scholar]
  152. Shridhar, K.; Laumann, F.; Liwicki, M. A comprehensive guide to bayesian convolutional neural network with variational inference. arXiv 2019, arXiv:1901.02731. [Google Scholar]
  153. Shridhar, K.; Laumann, F.; Liwicki, M. Uncertainty estimations by softplus normalization in bayesian convolutional neural networks with variational inference. arXiv 2018, arXiv:1806.05978. [Google Scholar]
  154. Gal, Y.; Ghahramani, Z. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. arXiv 2015, arXiv:1506.02142. [Google Scholar]
  155. Nalisnick, E.; Smyth, P. Stick-breaking variational autoencoders. arXiv 2016, arXiv:1605.06197. [Google Scholar]
  156. Wang, H.; Shi, X.; Yeung, D.Y. Natural-parameter networks: A class of probabilistic neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  157. Saatci, Y.; Wilson, A.G. Bayesian GAN. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  158. He, H.; Wang, H.; Lee, G.H.; Tian, Y. Probgan: Towards probabilistic gan with theoretical guarantees. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 Apri–3 May 2018. [Google Scholar]
  159. Daxberger, E.; Nalisnick, E.; Allingham, J.U.; Antorán, J.; Hernández-Lobato, J.M. Bayesian deep learning via subnetwork inference. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 18–24 July 2021; pp. 2510–2521. [Google Scholar]
  160. Dorta, G.; Vicente, S.; Agapito, L.; Campbell, N.D.; Simpson, I. Structured uncertainty prediction networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18—22 June 2018; pp. 5477–5485. [Google Scholar]
  161. Harakeh, A.; Smart, M.; Waslander, S.L. Bayesod: A bayesian approach for uncertainty estimation in deep object detectors. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 87–93. [Google Scholar]
  162. Le, M.T.; Diehl, F.; Brunner, T.; Knol, A. Uncertainty Estimation for Deep Neural Object Detectors in Safety-Critical Applications. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 3873–3878. [Google Scholar] [CrossRef] [Green Version]
  163. Huang, P.Y.; Hsu, W.T.; Chiu, C.Y.; Wu, T.F.; Sun, M. Efficient uncertainty estimation for semantic segmentation in videos. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 520–535. [Google Scholar]
  164. Pascual, G.; Seguí, S.; Vitria, J. Uncertainty gated network for land cover segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 276–279. [Google Scholar]
  165. Martinez, C.; Potter, K.M.; Smith, M.D.; Donahue, E.A.; Collins, L.; Korbin, J.P.; Roberts, S.A. Segmentation certainty through uncertainty: Uncertainty-refined binary volumetric segmentation under multifactor domain shift. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  166. Ilg, E.; Çiçek, Ö.; Galesso, S.; Klein, A.; Makansi, O.; Hutter, F.; Brox, T. Uncertainty estimates for optical flow with multi-hypotheses networks. arXiv 2018, arXiv:1802.07095. [Google Scholar]
  167. Loquercio, A.; Segu, M.; Scaramuzza, D. A general framework for uncertainty estimation in deep learning. IEEE Robot. Autom. Lett. 2020, 5, 3153–3160. [Google Scholar] [CrossRef] [Green Version]
  168. Bertoni, L.; Kreiss, S.; Alahi, A. Monoloco: Monocular 3d pedestrian localization and uncertainty estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6861–6871. [Google Scholar]
  169. Prokudin, S.; Gehler, P.; Nowozin, S. Deep directional statistics: Pose estimation with uncertainty quantification. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 534–551. [Google Scholar]
  170. Yu, T.; Li, D.; Yang, Y.; Hospedales, T.M.; Xiang, T. Robust person re-identification by modelling feature uncertainty. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 552–561. [Google Scholar]
  171. Zheng, J. Augmented Deep Representations for Unconstrained Still/Video-Based Face Recognition. Ph.D. Thesis, University of Maryland, College Park, MD, USA, 2019. [Google Scholar]
  172. Peterson, J.C.; Battleday, R.M.; Griffiths, T.L.; Russakovsky, O. Human uncertainty makes classification more robust. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9617–9626. [Google Scholar]
  173. Carbone, G.; Wicker, M.; Laurenti, L.; Patane, A.; Bortolussi, L.; Sanguinetti, G. Robustness of bayesian neural networks to gradient-based attacks. arXiv 2020, arXiv:2002.04359. [Google Scholar]
  174. RDG’s Policing & Security Implementation Group. National Rail & Underground Closed Circuit Television (CCTV) Guidance Document; Technical Report; Rail Delivery Group: London, UK, 2015; Available online: https://www.raildeliverygroup.com/media-centre-docman/archive/384-2015-10-national-rail-underground-cctv-guidance/file.html (accessed on 7 April 2022).
  175. hAnnaidh, B.O.; Fitzgerald, P.; Berney, H.; Lakshmanan, R.; Coburn, N.; Geary, S.; Mulvey, B. Devices and Sensors Applicable to 5G System Implementations. In Proceedings of the Proc. of the IEEE MTT-S International Microwave Workshop Series on 5G Hardware and System Technologies (IMWS-5G), Dublin, Ireland, 30–31 August 2018; pp. 1–3. [Google Scholar]
  176. Kim, H.; Cha, Y.; Kim, T.; Kim, P. A Study on the Security Threats and Privacy Policy of Intelligent Video Surveillance System Considering 5G Network Architecture. In Proceedings of the 2020 International Conference on Electronics, Information, and Communication (ICEIC), Barcelona, Spain, 19–22 January 2020; pp. 1–4. [Google Scholar]
  177. Ben Mabrouk, A.; Zagrouba, E. Abnormal Behavior Recognition for Intelligent Video Surveillance Systems: A Review. Expert Syst. Appl. 2018, 91, 480–491. [Google Scholar] [CrossRef]
  178. Ribeiro, M.; Lazzaretti, A.E.; Lopes, H.S. A Study of Deep Convolutional Auto-Encoders for Anomaly Detection in Videos. Pattern Recognit. Lett. 2018, 105, 13–22. [Google Scholar] [CrossRef]
  179. Milgram, P.; Takemura, H.; Utsumi, A.; Kishino, F. Augmented reality: A class of displays on the reality-virtuality continuum. In Proceedings of the Telemanipulator and Telepresence Technologies, International Society for Optics and Photonics, Bellingham, WA, USA, 21 December 1995; Volume 2351, pp. 282–292. [Google Scholar]
  180. Xie, Y.; Wang, M.; Liu, X.; Mao, B.; Wang, F. Integration of Multi-Camera Video Moving Objects and GIS. ISPRS Int. J. Geo-Inf. 2019, 8, 561. [Google Scholar] [CrossRef] [Green Version]
  181. Brey, P. Ethical Aspects of Information Security and Privacy. In Security, Privacy, and Trust in Modern Data Management; Petković, M., Jonker, W., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 21–36. [Google Scholar] [CrossRef]
  182. Sharon, T.; Koops, B.J. The ethics of inattention: Revitalising civil inattention as a privacy-protecting mechanism in public spaces. Ethics Inf. Technol. 2021, 23, 331–343. [Google Scholar] [CrossRef]
  183. Chang, V. An ethical framework for big data and smart cities. Technol. Forecast. Soc. Chang. 2021, 165, 120559. [Google Scholar] [CrossRef]
  184. Boyer, M.; Veigl, S. Privacy preserving video surveillance infrastructure with particular regard to modular video analytics. In Proceedings of the 6th International Conference on Imaging for Crime Prevention and Detection (ICDP-15), London, UK, 15–17 July 2015. [Google Scholar]
  185. McStay, A. Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy. Big Data Soc. 2020, 7, 2053951720904386. [Google Scholar] [CrossRef] [Green Version]
  186. Šidlauskas, A. Video Surveillance and the GDPR. Soc. Transform. Contemp. Soc. Open J. 2019, 7, 55–65. [Google Scholar]
  187. Denham, E. Information Commissioner’s Opinion: The Use of Live Facial Recognition Technology by Law Enforcement in Public Places; Technical Report 2019/01; Information Commissioner’s Office: Wilmslow, UK, 2019. [Google Scholar]
  188. Steel, A.; Olive, A. R Bridges v Chief Constable South Wales Police and Information Commissioner; Technical Report Citation Number: [2020] EWCA Civ 1058; Royal Courts of Justice: London, UK, 2020. [Google Scholar]
Figure 1. Keyword co-occurrence network of the published papers under the topic of video surveillance analytics published since 2010.
Figure 1. Keyword co-occurrence network of the published papers under the topic of video surveillance analytics published since 2010.
Sensors 22 04324 g001
Figure 2. Conceptual structure map using the Correspondence Analysis (CA) method. The clusters represent how ideas are connected; the closer they are, the stronger the association.
Figure 2. Conceptual structure map using the Correspondence Analysis (CA) method. The clusters represent how ideas are connected; the closer they are, the stronger the association.
Sensors 22 04324 g002
Figure 3. CCTV modules. In railway surveillance, cameras are the main sensors, which are responsible for collecting video data. Different types of sensors (introduced in Section 2.3), communication systems and other computing facilities (introduced in Section 2.4) make up the monitoring and recording systems. The video analytics module receives the data passed from monitoring and recording systems and then makes decisions using computer vision technologies.
Figure 3. CCTV modules. In railway surveillance, cameras are the main sensors, which are responsible for collecting video data. Different types of sensors (introduced in Section 2.3), communication systems and other computing facilities (introduced in Section 2.4) make up the monitoring and recording systems. The video analytics module receives the data passed from monitoring and recording systems and then makes decisions using computer vision technologies.
Sensors 22 04324 g003
Figure 4. An example of a sensor system composed of conventional sensors, such as electro-optical and thermal imaging sensors, and non-conventional sensors, which enhance the capability in other frequency domains. Acoustic sensors provide omnidirectional detection and tracking based upon trilateration from temporal differences in the sound waves. The health sensor provides vital signs of the individuals for detecting sick people passing through a station for further screening. A chemical sensor detects specific chemicals for the identification of possible explosive devices. The wave scanner detects concealed objects, including weapons.
Figure 4. An example of a sensor system composed of conventional sensors, such as electro-optical and thermal imaging sensors, and non-conventional sensors, which enhance the capability in other frequency domains. Acoustic sensors provide omnidirectional detection and tracking based upon trilateration from temporal differences in the sound waves. The health sensor provides vital signs of the individuals for detecting sick people passing through a station for further screening. A chemical sensor detects specific chemicals for the identification of possible explosive devices. The wave scanner detects concealed objects, including weapons.
Sensors 22 04324 g004
Figure 5. Using radio waves in the millimeter spectrum to safely penetrate clothing and reflect off body-worn concealed threats. Images were taken from the “L-3 Airport Scanner” by the Pacific Northwest National Laboratory [37].
Figure 5. Using radio waves in the millimeter spectrum to safely penetrate clothing and reflect off body-worn concealed threats. Images were taken from the “L-3 Airport Scanner” by the Pacific Northwest National Laboratory [37].
Sensors 22 04324 g005
Figure 6. Infrared thermography used in a train station. Images taken from “Waterloo station” by branestawm2002 [39].
Figure 6. Infrared thermography used in a train station. Images taken from “Waterloo station” by branestawm2002 [39].
Sensors 22 04324 g006
Figure 7. Computing architectures.
Figure 7. Computing architectures.
Sensors 22 04324 g007
Figure 8. Block diagram of sensor-based computer vision methods for object detection and behaviour analysis tasks. The data collected by sensors will be transformed as input data to the train models. Models built based on various structures and methods will handle the process of extracting the features from the training data set. Then, well-trained models can be used as a detector or classifier to recognise the activities of the detected objects in surveillance systems.
Figure 8. Block diagram of sensor-based computer vision methods for object detection and behaviour analysis tasks. The data collected by sensors will be transformed as input data to the train models. Models built based on various structures and methods will handle the process of extracting the features from the training data set. Then, well-trained models can be used as a detector or classifier to recognise the activities of the detected objects in surveillance systems.
Sensors 22 04324 g008
Figure 9. A basic CNN structure used for classification tasks. Generally, there are three types of layers used to form a full CNNs architecture: Convolutional layer, Pooling layer and Fully-connected (FC) layer. A convolution layer is a linear transformation that preserves spatial information in the input image. It will compute the output of neurons that are connected to local regions in the input, each computing a dot product between their weights and a small region they are connected to in the input volume. Pooling layers take the output of a convolution layer and reduce its dimensionality under certain rules. FC layers are used to connect the neurons between two different layers. It consists of the weights and biases along with the neurons. The FC layers are usually deployed before the output layer to work on the last few layers of a CNN architecture.
Figure 9. A basic CNN structure used for classification tasks. Generally, there are three types of layers used to form a full CNNs architecture: Convolutional layer, Pooling layer and Fully-connected (FC) layer. A convolution layer is a linear transformation that preserves spatial information in the input image. It will compute the output of neurons that are connected to local regions in the input, each computing a dot product between their weights and a small region they are connected to in the input volume. Pooling layers take the output of a convolution layer and reduce its dimensionality under certain rules. FC layers are used to connect the neurons between two different layers. It consists of the weights and biases along with the neurons. The FC layers are usually deployed before the output layer to work on the last few layers of a CNN architecture.
Sensors 22 04324 g009
Figure 10. A diagram for a one-unit RNN. From bottom to top: input state, hidden state, output state. U , V , W are the weights of the networks. The compressed diagram is on the left, and the unfolded version is on the right.
Figure 10. A diagram for a one-unit RNN. From bottom to top: input state, hidden state, output state. U , V , W are the weights of the networks. The compressed diagram is on the left, and the unfolded version is on the right.
Sensors 22 04324 g010
Figure 11. Pipeline of a Transformer [64]. A transformer is made up of an encoder module and a decoder module with multiple identical encoders and decoders. Each encoder and decoder contains a self-attention layer and a feedforward neural network. Each decoder has extra an encoder–decoder attention layer.
Figure 11. Pipeline of a Transformer [64]. A transformer is made up of an encoder module and a decoder module with multiple identical encoders and decoders. Each encoder and decoder contains a self-attention layer and a feedforward neural network. Each decoder has extra an encoder–decoder attention layer.
Sensors 22 04324 g011
Figure 12. A two-step SOD pipeline using Background Subtraction. The images are from the CAVIAR dataset [78]. By comparing two different frames in the video data provided by the same camera sensor, the input image is divided into multiple region proposals based on the objects of interest. Each region is processed separately, and the object features in each region are extracted. The background subtraction detects all moving objects from the background. Then, by continuing to compare other frames in the video, the static object in the frame apart from the moving objects, such as people and trains, will be separated from other objects. Therefore, the static objective—in this case, the luggage—will be successfully detected.
Figure 12. A two-step SOD pipeline using Background Subtraction. The images are from the CAVIAR dataset [78]. By comparing two different frames in the video data provided by the same camera sensor, the input image is divided into multiple region proposals based on the objects of interest. Each region is processed separately, and the object features in each region are extracted. The background subtraction detects all moving objects from the background. Then, by continuing to compare other frames in the video, the static object in the frame apart from the moving objects, such as people and trains, will be separated from other objects. Therefore, the static objective—in this case, the luggage—will be successfully detected.
Sensors 22 04324 g012
Figure 13. Multiple object detection examples [81]. The targets are detected and marked in red boxes.
Figure 13. Multiple object detection examples [81]. The targets are detected and marked in red boxes.
Sensors 22 04324 g013
Figure 14. Multiple Object Detection with Topic Modelling [81]. The movement of the Fire engine, which is considered an abnormal event, in this case, is detected and marked with red boxes.
Figure 14. Multiple Object Detection with Topic Modelling [81]. The movement of the Fire engine, which is considered an abnormal event, in this case, is detected and marked with red boxes.
Sensors 22 04324 g014
Figure 15. Block diagram of VFR. Images come from a public dataset Face Mask Detection [89].
Figure 15. Block diagram of VFR. Images come from a public dataset Face Mask Detection [89].
Sensors 22 04324 g015
Figure 16. Person Re-Identification in video surveillance systems. This figure shows a top view of the region of interest comprising an entrance hall, a waiting area and a shop, monitored by three cameras (red is optical and black is thermal) with non-overlapping coverages (orange triangles). The locations of three individuals (blue, cyan and green dots) are shown at different timestamps in the region of interest. The person Re-ID aims to associate paths of the individuals denoted by blue and green dots detected by different modalities (optical and thermal) cameras.
Figure 16. Person Re-Identification in video surveillance systems. This figure shows a top view of the region of interest comprising an entrance hall, a waiting area and a shop, monitored by three cameras (red is optical and black is thermal) with non-overlapping coverages (orange triangles). The locations of three individuals (blue, cyan and green dots) are shown at different timestamps in the region of interest. The person Re-ID aims to associate paths of the individuals denoted by blue and green dots detected by different modalities (optical and thermal) cameras.
Sensors 22 04324 g016
Figure 17. Block diagram of Person Re-ID. The images are from the CAVIAR dataset [78]. Five main steps for designing a Person Re-ID system: (1) raw data collection, (2) training data annotation, (3) model training, (4) bounding box generation and (5) pedestrian retrieval.
Figure 17. Block diagram of Person Re-ID. The images are from the CAVIAR dataset [78]. Five main steps for designing a Person Re-ID system: (1) raw data collection, (2) training data annotation, (3) model training, (4) bounding box generation and (5) pedestrian retrieval.
Sensors 22 04324 g017
Figure 18. A block diagram of an HAR algorithm using an example of detecting an abandoned object. The images come from the i-LIDS bag and vehicle detection challenge of AVSS 2007 conference [97]. Four video frames with labelled objects are input to the HAR algorithm. The first step is feature extraction. The spatial and temporal information contained in the video helps in the recognition. Hence, both spatial (green) and temporal (blue) features are extracted. The video classifier classifies the extracted features. The surveillance operator is alerted if the video is classified as belonging to any of the pre-defined classes.
Figure 18. A block diagram of an HAR algorithm using an example of detecting an abandoned object. The images come from the i-LIDS bag and vehicle detection challenge of AVSS 2007 conference [97]. Four video frames with labelled objects are input to the HAR algorithm. The first step is feature extraction. The spatial and temporal information contained in the video helps in the recognition. Hence, both spatial (green) and temporal (blue) features are extracted. The video classifier classifies the extracted features. The surveillance operator is alerted if the video is classified as belonging to any of the pre-defined classes.
Sensors 22 04324 g018
Figure 19. Block diagram of IUAD. The images come from the CAVIAR dataset [78]. Four labelled video frames are input to the algorithm. The spatial and temporal features are extracted first, followed by the anomaly detector. The anomaly detector uses parameters from a model trained on historical video data, which are all classified as normal. The output of anomaly detectors is either score or binary, where a binary output is used in this case.
Figure 19. Block diagram of IUAD. The images come from the CAVIAR dataset [78]. Four labelled video frames are input to the algorithm. The spatial and temporal features are extracted first, followed by the anomaly detector. The anomaly detector uses parameters from a model trained on historical video data, which are all classified as normal. The output of anomaly detectors is either score or binary, where a binary output is used in this case.
Sensors 22 04324 g019
Figure 20. Block diagram of Crowd Count. The input frames are processed by two streams, spatial and a temporal, to extract features. The spatial features are represented by a yellow colour, and the temporal using a red colour indicates the direction of motion. The extracted features are processed to generate an estimated density map using different colours. The blue colour represents minimum or zero capacity, where yellow and red represent the bigger and maximum density. Finally, a well-trained density model is able to provide a prediction of the crowd number, which is 888 in this case.
Figure 20. Block diagram of Crowd Count. The input frames are processed by two streams, spatial and a temporal, to extract features. The spatial features are represented by a yellow colour, and the temporal using a red colour indicates the direction of motion. The extracted features are processed to generate an estimated density map using different colours. The blue colour represents minimum or zero capacity, where yellow and red represent the bigger and maximum density. Finally, a well-trained density model is able to provide a prediction of the crowd number, which is 888 in this case.
Sensors 22 04324 g020
Figure 21. Adversarial attack examples. Image was taken from the Google AI Blog [127].
Figure 21. Adversarial attack examples. Image was taken from the Google AI Blog [127].
Sensors 22 04324 g021
Figure 22. Facial recognition with masks. Images are from a public dataset Face Mask Detection [89].
Figure 22. Facial recognition with masks. Images are from a public dataset Face Mask Detection [89].
Sensors 22 04324 g022
Figure 23. Data processing in the next generation of CCTV surveillance systems.
Figure 23. Data processing in the next generation of CCTV surveillance systems.
Sensors 22 04324 g023
Figure 24. Intelligence levels of data analytics for CCTV.
Figure 24. Intelligence levels of data analytics for CCTV.
Sensors 22 04324 g024
Figure 25. Event prediction in a blind zone [180]. This figure shows an example of two persons leaving the coverage of two different cameras (orange and green) and entering a blind zone. The video analytics system would predict and display the most probable trajectory of these individuals. The same concept could be used for the prediction of complex events such as criminal activity, possibly giving an operator extra time to respond.
Figure 25. Event prediction in a blind zone [180]. This figure shows an example of two persons leaving the coverage of two different cameras (orange and green) and entering a blind zone. The video analytics system would predict and display the most probable trajectory of these individuals. The same concept could be used for the prediction of complex events such as criminal activity, possibly giving an operator extra time to respond.
Sensors 22 04324 g025
Figure 26. Six basic principles of GDPR.
Figure 26. Six basic principles of GDPR.
Sensors 22 04324 g026
Table 1. Number of publications of computer vision techniques used in rail networks surveillance by Google Scholar since 2010.
Table 1. Number of publications of computer vision techniques used in rail networks surveillance by Google Scholar since 2010.
Computer Vision Techniques Used
in Rail Networks Surveillance
Number of Publications
Shown in Google Scholar
Motion or Moving Object Detection83
Video Face Recognition137
Person Re-Identification6250
Human Activity Recognition6160
Video Anomaly Detection1060
Trajectory Analysis1440
Crowd Analysis1500
Table 2. Camera options (Basic).
Table 2. Camera options (Basic).
FeatureOptionsPros/Cons
Connectivity
Type
AnalogueAnalogue point to point connection.
Low-cost solution when upgrading conventional (analogue) CCTV system.
Secure against cyber security threats.
DigitalDigital (IP) interface.
Flexible point to multipoint connection.
High-cost solution when upgrading conventional CCTV system.
Vulnerable to cyber security threats.
Connectivity
Medium
WiredSecure against cybersecurity intrusion.
Less noise interference.
WirelessFlexible in terms of installation.
Field of view
(FOV)
FixedStationary mount.
Monitor one small area of interest.
Pan-tilt-zoom (PTZ) cameraMonitor large area using automatic/remote controlled pan (left/right),
tilt (up/down) and zoom (in/out).
Lighting
conditions
Day/nightAutomatically adjust to light conditions. Coloured video during day
and black and white at night.
Used for conditions such as glare, direct sunlight, reflections and
strong backlight.
Low lightUsed in low light, e.g., indoor restaurants, streetlights, etc.
Cannot be used in completely dark conditions.
Night vision/
Fog/Smoke
Used in complete darkness.
Detect through obstructions such as fog and smoke.
These are either active infrared (IR) or thermal cameras. The IR
cameras use built-in IR illuminators and near IR (NIR) and IR cameras for
monitoring. The thermal cameras are passive and might not provide video
through glass or water.
Thermal cameras are more costly than IR cameras.
Image
sensor
ColourUsed in daylight or well-lit conditions.
Provide accurate colours at the monitoring and recording systems.
MonochromeUsed in near dark situations and give more details than a naked human
eye can perceive.
Might give low contrast during the day.
HousingDomeSpherical in shape to reduce wind and vibration effects.
Protect and conceal camera direction.
Some advanced units, called speed domes, rotate the camera and give an all
around view.
SealedUsed in hostile situations.
All electrical components are sealed to avoid explosion hazards.
Impact resistantMilitary grade.
Used in high crime areas.
Tamper resistantSimilar to impact-resistant but are additionally protected against tool vandalism.
Usually resistant to cutting, hammering and prying.
Bullet resistantSimilar to impact-resistant with additional safety for the glass.
MountIndoorWall. A bracket supports housing, and camera FOV is adjustable.
Pendant. Suspend the camera from the ceiling.
Corner. Used where two walls meet at the right angle (both interior and exterior).
Dome. These are installed on the ceiling or other surfaces for dome type housings.
These are susceptible to vibrations.
OutdoorPole. These are used for unobstructed FOV.
Corner. These are installed where two walls meet at the right angle.
The best FOV is achieved by installing it close to the roof.
Image
sensor
CCDStands for charged coupled devices.
Used in daylight, lowlight and NIR cameras.
Generate less heat.
Susceptible to blooming (blooming is when a bright light source in the FOV
hides some of the image details).
Used for high resolution and quality images.
CMOSStands for Complementary Metal Oxide Semiconductor.
Low power and suitable for mobile devices or power constraint environments.
Cheaper.
Better video quality in outdoor areas on a bright sunny day.
Focal length
(Lens)
FixedUsed in fixed cameras for focusing on one area only.
VarifocalFocal length can be varied in a range, and focus is manually adjusted.
Capture close-ups of activities at longer distances.
ZoomUsed in PTZ cameras.
Focal length can be varied in a range, but the focus is automatically adjusted.
Optical zoom
factor 
Optical zoom factor is calculated from the focal length range of the
varifocal and zoom lens. For example, a range of 2–10 mm focal length
represents a (10/2 =) 5× zoom factor. It does not mean 5 times enlargement
of the image.
Aperture
control
FixedSuitable for constant lighting conditions.
ManualUsed for fixed cameras with controlled lighting.
Less expensive as compared to the automatic.
Requires a technician to operate.
AutomaticUsed in outdoor situations or where extreme changes in lighting
conditions are expected.
Aperture
size
LargeUsed in dim light conditions.
Image fore- and background are out of focus and blurred in daylight.
SmallComplete scene will be in focus.
FilterNeutral Density (ND)Control the level of visible light and reduce it when it is too high.
PolarisingOrient the light in a specific direction.
Used to eliminate reflected light and glare.
Improve image contrast.
IR-cutTo control the NIR light sources such as the sun.
Most image sensors are sensitive to NIR light. The NIR sources during the
day can degrade the performance of image sensors. Without the IR-cut filter,
the image will have unpredictable colours and poor contrast.
Table 3. Camera options (Value Added).
Table 3. Camera options (Value Added).
OptionsPros/Cons
External synchronizationThe internal clock is synchronised to an external clock. It is
important for any automation system, including CCTV, to have
a synchronised clock.
Remote configurationThe cameras, especially the image properties, are
remotely configurable.
Privacy maskingSelective blocking of private areas.
Covert camerasThese are hidden cameras and are preferably battery operated
and wireless.
Dummy or drone camerasThese are non-functioning cameras and are installed for deterrence.
Auto scan of PTZCamera can perform automatic scanning of an area of interest.
Pre-set of PTZCamera orientation and lens setting can be programmed to scan/focus
specific areas.
Slip ring PTZThis allows the PTZ camera to rotate without twisting the cable.
These are prone to contamination and temperature changes.
Motion detectionCamera can detect and provide alarm using built-in motion detection.
Backlight compensation (BLC)The camera with BLC provides high-contrast images with a
bright background.
Digital noise reduction (DNR)The camera with DNR reduces the noise in the video. This noise is
prominent in low-light or dark environments.
Mobile compatibilityThe camera video can be viewed on a mobile device. The preferred
choice is viewing without the installation of an additional application on the
remote device.
Sun shields on housingThis protects the camera from direct sunlight, which can dramatically
reduce the life of the camera.
Wipers on housingThese are similar to car windshield wipers. These are not recommended as
they might erode the optical surface of the glass.
Heaters and ventilators for housingTemperature differences between the interior and exterior of the
housing can cause fog or icing on the glass. The heaters/ventilators help in
maintaining the temperature inside the housing.
Smart cameraA smart camera is a small version of the CCTV system. It consists of a sensor,
processor, video analytics and communication interface with a remote
monitoring device.
Table 4. Comparison of displays.
Table 4. Comparison of displays.
Type of DisplayPros/Cons
MonochromeProvides better details in some scenarios.
PlasmaWider viewing angles.
Higher contrast ratio.
Higher black levels.
LCDCompact and lightweight.
Less power requirement.
Low heat dissipation.
Less prone to burn-in.
No flicker.
Long life.
Low image contrast.
True black colour is not produced.
Restricted viewing angle.
LEDAll benefits of the LCD.
Better contrast.
Wide colour range.
Shorter life span.
More expensive.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, T.; Aftab, W.; Mihaylova, L.; Langran-Wheeler, C.; Rigby, S.; Fletcher, D.; Maddock, S.; Bosworth, G. Recent Advances in Video Analytics for Rail Network Surveillance for Security, Trespass and Suicide Prevention—A Survey. Sensors 2022, 22, 4324. https://doi.org/10.3390/s22124324

AMA Style

Zhang T, Aftab W, Mihaylova L, Langran-Wheeler C, Rigby S, Fletcher D, Maddock S, Bosworth G. Recent Advances in Video Analytics for Rail Network Surveillance for Security, Trespass and Suicide Prevention—A Survey. Sensors. 2022; 22(12):4324. https://doi.org/10.3390/s22124324

Chicago/Turabian Style

Zhang, Tianhao, Waqas Aftab, Lyudmila Mihaylova, Christian Langran-Wheeler, Samuel Rigby, David Fletcher, Steve Maddock, and Garry Bosworth. 2022. "Recent Advances in Video Analytics for Rail Network Surveillance for Security, Trespass and Suicide Prevention—A Survey" Sensors 22, no. 12: 4324. https://doi.org/10.3390/s22124324

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop