Next Article in Journal
Design of Optimal SPR-Based Multimode Waveguide Sensor for a Wide Range of Liquid Analytes
Next Article in Special Issue
Indoor Visible-Light 3D Positioning System Based on GRU Neural Network
Previous Article in Journal
Development of Multiple Fano-Resonance-Based All-Dielectric Metastructure for High-Contrast Biomedical Applications
Previous Article in Special Issue
Recent Advances in Optical Injection Locking for Visible Light Communication Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review on Image Sensor Communication and Its Applications to Vehicles

1
Department of Information and Communication Engineering, Nagoya University, Nagoya 464-8603, Japan
2
Institute of Liberal Arts and Sciences, Nagoya University, Nagoya 464-8603, Japan
*
Author to whom correspondence should be addressed.
Photonics 2023, 10(6), 617; https://doi.org/10.3390/photonics10060617
Submission received: 25 March 2023 / Revised: 25 April 2023 / Accepted: 18 May 2023 / Published: 26 May 2023
(This article belongs to the Special Issue Advances in Visible Light Communication)

Abstract

:
Image sensor communication (ISC), also known as optical camera communication, is a form of visible light communication that utilizes image sensors rather than a single photodiode, for data reception. ISC offers spatial separation properties and robustness to ambient noise, making it suitable for outdoor applications such as intelligent transportation systems (ITSs). This review analyzes the research trends in ISC, specifically concerning its application in ITSs. Our focus is on various ISC receivers, including rolling shutter cameras, global shutter high-speed cameras, optical communication image sensors, and event cameras. We analyze how each of these receivers is being utilized in ISC vehicular applications. In addition, we highlight the use of ISC in range estimation techniques and the ability to achieve simultaneous communication and range estimation. By examining these topics, we aim to provide a comprehensive overview of the role of ISC technology in ITSs and its potential for future development.

1. Introduction

Visible light communication (VLC) is a wireless communication method that utilizes visible light for information transmission [1,2,3,4]. The luminance of the VLC light sources can be modulated extremely fast to make the switching of light imperceptible to the human eye to achieve high-speed data transmission [4]. In contrast to traditional wireless communications that use electromagnetic waves, wireless communication using visible light has a much wider wavelength range and has been considered a technique to address the issues caused by the growing spectrum need for wireless communication [5].
The origins of VLC can be traced back to the 7th century BC, when people in ancient China used smoke and fire to transmit information about their enemies across long distances. Over two millennia later, in the 1880s, Alexander Graham Bell made his groundbreaking attempt to use machinery to implement VLC by transmitting voice information through sunlight, carried by electricity [6]. Unfortunately, however, this technology did not reach early commercialization due to its limitations in communication distance and the inevitable effects of weather and obstacles such as rain. It was not until the 1990s that the invention of a high-brightness light-emitting diode (LED) [7] revolutionized the lighting industry, opening up new possibilities for VLC. The high response and fast switching speed of LEDs allow using visible light to convey high-speed data [4,8]. In 2004, T. Komine and M. Nagakawa proposed an indoor VLC system using white LED lights for both illumination and optical wireless communication [4]. They considered the interference and reflection of the multiple light sources installed in a room and demonstrated its potential for high-speed data transmission of around 10 Gbps. This paper has been highly influential in the field of VLC, leading to widespread research interest and study in VLC. VLC technology is also applied outdoors, for instance, in traffic lights and streetlights. In 1999, G. Pang et al. showed the potential of using a traffic light as a communication device [9]. In [9], high-brightness LED lights were modulated and encoded with audio messages. The receiver was designed to demodulate the optically transmitted audio messages using a photodiode (PD) to convert the light to electrical signals through direct detection. However, there were many constraints to using a PD in outdoor environments due to the strong ambient noise. In 2001, an image sensor was introduced as a new type of receiver for VLC, which was more robust to outdoor ambient noises [10]. An image sensor consists of an enormous array of PDs. This structure of an image sensor gives it the ability to spatially separate the light sources, also resulting in being robust to outdoor ambient noise and being able to easily track moving vehicles. Nowadays, we have coined the term image sensor communication (ISC) for the VLC that uses image sensors as receivers [11,12,13]. ISC can also be referred to as optical camera communication (OCC) [14,15]. The primary distinction between ISC and OCC is that the term “camera” in OCC encompasses both the lens and circuitry for an image sensor. Additionally, these two terms are named by different institutions. In this review, we will discuss ISC and its outdoor applications, such as intelligent transportation systems (ITSs), which are also known as ITS-ISC.
For ITSs, low latency and accurate localization are crucial factors [16,17,18]. Low latency is necessary for real-time communication between vehicles and their surrounding transportation elements to efficiently coordinate traffic flow and avoid accidents timely. Accurate localization is important for car navigation and collision avoidance, as well as for providing location-based services to passengers. ISC can be an applicable technology to realize low latency and accurate localization simultaneously. On the one hand, ISC can achieve low latency by exploiting the high-speed capabilities of digital image sensors, for instance, using high-speed cameras as the ISC receiver [19] or utilizing the rolling shutter effect [20]. The high-speed transmission allows real-time data transfer between vehicles, road infrastructure, and other elements in the ITS. On the other hand, ISC systems can achieve high localization accuracy by exploiting the high-brightness properties of visible light. Even in complex traffic environments, such as tunnels or intersections, LED transmitters can be recognized from complicated backgrounds using algorithms in ISC. In contrast, localization using conventional wireless technologies can be inaccurate due to signal reflection, multi-path propagation, or environmental interference [21]. Furthermore, many existing automotive applications are now equipped with image sensors, such as driver recorders and autonomous driving sensors [22]. Meanwhile, the road and vehicles are equipped with LED lights. Therefore, ISC that utilizes both image sensors and LEDs can be a promising technology for ITS.
We will provide detailed discussions of the ISC studies in the remainder of this review paper. In Section 2, we present the research trend of VLC based on the statistical data collected on Scopus. We analyze two receiver types of VLC: PDs and image sensors. Then, we discuss the advantages of using image sensors in outdoor environments. It then leads to the research status of ITS-ISC. We also analyze the constitutions of transmitter and receiver types of ISC. In Section 3, we explain the concepts necessary for ITS-ISC, including vehicle-to-vehicle (V2V), vehicle-to-infrastructure (V2I), infrastructure-to-vehicle (I2V), and vehicle-to-everything (V2X) communications. We then describe the architecture of the ITS-ISC system. In addition, the advantages and limitations of image sensor receivers are discussed. In Section 4, a comprehensive review of the receiver types of ISC is provided, including rolling shutter cameras, global shutter high-speed cameras, optical communication image sensors (OCIs), and event cameras (dynamic vision sensors). In Section 5, we illustrate the basic mechanism, challenges, and solutions of range estimation using LED transmitters and image sensor receivers. Finally, Section 6 presents the conclusion.

2. Research Trend

2.1. Research Trend of VLC

Figure 1a shows the annual publication numbers of VLC literature published before 2023 searched in Scopus. We found a total of 9345 publications. From Figure 1a, it can be seen that the number of VLC publications has greatly increased since around 2010 but declined after 2019. Figure 1b shows the results of classifying these research publications by document type. According to Figure 1b, conference papers constitute 50.3% of the total. This might be the reason behind the decline in publication numbers from 2019 when the COVID-19 pandemic started making it challenging for people to attend international conferences.
Figure 1c describes the percentage of publications using PD or image sensor receivers in VLC. 88% of the publications used PDs, and the number was 8194 publications. Only 12% of the publications used image sensors, with 1151 publications. Therefore, it can be concluded that most works utilize PDs as the receiver. There are many advantages to using PDs, such as rapid response and low cost. Due to the high-speed response, the method of using a PD as a receiver device has achieved communication at the Gbps level [5,23,24]. In contrast, image sensors communicate much slower than PDs due to the frame rate limitation. Based on our survey, research on PDs is being conducted mainly for indoor use purposes. For example, illumination optical communication including Light-Fidelity (LiFi) [5] and indoor visible light positioning [25]. In terms of image sensor research, its proportion in indoor use is lower compared to the research on PDs.
Image sensors are more suitable for outdoor applications than PDs. The main reasons are that image sensors are more robust to solar noise than PDs and are simpler to track moving vehicles. When used outdoors, especially in an ITS environment, the effect of background light noise, such as sunlight and other light sources, cannot be ignored. Therefore, for a PD, the viewing angle must be narrowed to ensure an adequate signal-to-noise ratio. At the same time, mechanical manipulation is required to orient the PD toward the direction of the transmitter. However, when the car is driving at high speed and changing its direction, it is difficult for a PD with a narrowed viewing angle to swiftly adjust the direction of its optical axis. On the contrary, an image sensor can track moving vehicles with little adjustments to its direction since it has a large number of pixels. Moreover, an image sensor is capable of spatially separating light sources and selecting only the pixels that contain the desired transmitted signal while filtering out other pixels that may contain background light noise.
Additionally, there are ongoing efforts to establish international standards for VLC. In 2011, the IEEE 802.15.7 standard for short-range wireless optical communication was established [26]. It was revised later to include ISC and new PHY modes in IEEE 802.15.7a task group [13]. In July 2020, a new amendment to IEEE 802.15.7a was proposed, which adds a deep learning mechanism for ISC. This amendment is expected to be published in September 2023. Furthermore, in December 2017, a proposal for a new ISO standard for localized communication using ISC in ITSs was approved. The proposal suggested a new communication interface called “ITS-OCC”, which adopted ISC profile, communication adaptation layer, and management adaptation entity from IEEE 802.15.7:2018, ISO 14296:2016, and ISO 22738:2020, respectively [27,28,29]. The ISO 22738:2020 standard was published in July 2020 and is expected to facilitate the development of OCC-based ITS applications [29]. Finally, regarding the performance comparison between IEEE 802.15.7 for optical wireless communication and IEEE 802.15.7a for OCC, the addition of image sensor receivers and deep learning mechanisms in IEEE 802.15.7a OCC is expected to improve the accuracy and robustness of communication performance and range estimation in challenging environments. It may also allow for the development of new applications that were previously not possible with traditional range estimation methods. Further research and development in this area may lead to improved performance and new use cases for ISC.

2.2. Research Status of ISC

Figure 2 shows the growth in the number of ISC literature published before 2023 searched in Scopus. We found a total of 1151 papers. There is a significant increase in the quantity of literature starting in 2013; the growth in the number of papers levels off after 2019. Roughly 24% of this literature is applied to vehicles.
We selected papers with the literature type of “article” and “conference paper”, and classified these publications by applications, transmitter types, and receiver types. Note that the publications were categorized manually by the authors; subjective interpretation and potential errors could be introduced into the categorization process.
Figure 3 shows the proportions of publication numbers of different ISC applications. Communication application accounts for 78.7% of the whole. Here, communication refers to being able to transmit information such as text, image, voice, and video through optical signals. Papers with communication applications cover a wide range of subjects, such as bit-error-rate (BER) measurement, communication model design, and other related areas. In the context of an ITS, communication enables safety and efficiency applications such as collision avoidance, traffic flow optimization, and emergency vehicle notifications. Ranging, ranging application, also considered as localization or positioning, accounts for 13.9% of the whole. Ranging relies on computer vision techniques and can be used for applications such as obstacle avoidance and navigation. By using ISC, high-accuracy positioning can be achieved within two milliseconds, which is faster than light detection and ranging (LiDAR) or global positioning system (GPS) [30]. Additionally, detection and tracking comprise 3.6% of all ISC papers. It refers to extracting LEDs from the received image streams. Inaccurate detection or tracking can lead to incorrect data decoding because it is conducted prior to communication and ranging. The advantage of ISC-based communication and ranging technology is that the use of LEDs provides communication and navigation services along with illumination, thus minimizing the need for additional power. It reduces front-end costs, eliminating the need for additional installation and configuration of signal access points, and reducing the cost and complexity of the communication and navigation system.
Figure 4 shows the ratio of the different transmitters and receivers in the ISC. However, Figure 4b may have some level of inaccuracy. This is because some authors did not specify the type of image sensor used in their articles, and these kinds of articles are classified as “others”. Additionally, due to time constraints, the “others” category was not fully categorized. We conclude that since the rolling shutter camera is the majority of the receivers, some instances of rolling shutter cameras may have been categorized as “others”.
As shown in Figure 4a, the transmitters can be divided into more than two types, including LEDs and screens. It can be seen that LEDs are used in a much higher proportion than screens. LEDs are also the most suitable transmitters for ITS applications. Not only are existing traffic lights equipped with LED arrays, but their switching speed is also the fastest. LED transmitters can also be divided into different categories, for example, single LED, LED arrays, and rotating LED arrays. The number and size of LEDs in the transmitter determine to some extent how much data can be transmitted over how many distances. Because the number of pixels taken up by the transmitter decreases as the distance between the transmitter and the image sensor increases. With regard to the ISC receiver, the receivers can be divided into more than four types as shown in Figure 4b, including rolling shutter cameras, high-speed cameras (global shutter), event cameras, and OCI. They will be discussed in Section 4.

2.3. Image-Sensor-Communication-Based Intelligent Transportation System (ITS-ISC)

Figure 5 shows the proportions of different application scenarios in ISC, where outdoor applications comprise 24% and the number is 251 papers. Despite having a lower proportion than indoor applications, the potential of ISC outdoor applications in the research field of ITSs cannot be overlooked.
ITS refers to a system that connects people, roads, and vehicles through the use of communication technologies, aiming to improve road traffic safety and smooth traffic flow by reducing congestion, and achieving comfort dedicated short-range communications (DSRC) [16,38]. A familiar example of an ITS is the electronic toll collection (ETC) system. The latest report shows that the global intelligent transportation market reached a size of 110.53 billion USD in 2022 and is projected to increase at a compound annual growth rate of 13.0% during the forecast period from 2023 to 2030 [39]. This growth trend in the market highlights the need for vehicles to use intelligent transportation services.
The technological development of autonomous vehicles is currently attracting worldwide attention, especially in the field of ITSs. The recognition of the surrounding road conditions using various onboard sensors is important for driving safety, as well as driving support for autonomous vehicles. Cameras are one of the most commonly used onboard sensors in self-driving cars and advanced driver assistance systems [22,40]. These onboard cameras are mainly used for applications such as visual aids, object detection, and driving recording. If these cameras are used as receivers for ISC, they can receive the optical signal from LEDs on the road to provide communication and localization services. Information sent from the vehicle is expected to include location information such as the vehicle’s longitude and latitude, vehicle signals indicating driving conditions such as speed, steering, acceleration, braking, or vehicle front view, and vehicle-specific information such as vehicle model, width, and height. Information sent from the side of the road is expected to include traffic signal statuses, such as color and transition time, and traffic information, such as weather conditions and nearby traffic congestion. These kinds of information transmitted through the ISC system can provide efficient and reliable communication for a variety of applications, including traffic management, road safety, and autonomous driving.
In addition, image sensors can help to minimize or manage latency in vehicular networks, which is crucial for safety-critical applications, such as collision avoidance and cooperative driving. ISC can offer better spatial and directional accuracy than conventional radio frequency-based communications, which can enhance the reliability of achieving low latency. Hence, the integration of ISC in vehicular networks can considerably enhance the performance and safety of upcoming ITSs.

3. Basic Concept and Architecture of an ITS-ISC

3.1. Vehicle-to-Everything (V2X) Communications Using Image Sensors and LEDs

ISC can be implemented in a variety of contexts in an ITS, including V2V, V2I, and I2V communications. These means of communication are involved in vehicle-to-everything (V2X) communications, which refers to any type of communication between a vehicle and its surrounding traffic environment, as illustrated in Figure 6.
In V2X communication, vehicles can communicate with each other and share information about their speed, location, and other important data, which helps improve safety and traffic flow. For example, suppose a vehicle detects an obstacle on the road ahead. In that case, the vehicle can transmit this information to the rear vehicles to alert them and allow them to take appropriate action. This type of communication between vehicles is referred to as V2V communication. Moreover, V2X communications enable vehicles to send and receive information about the road environment, such as traffic conditions or road hazards, to improve safety and efficiency. For example, a traffic signal can transmit data to a vehicle about the time remaining before the signal changes, allowing the vehicle to adjust its speed accordingly. Furthermore, the information transmitted via V2X communications may include information about traffic patterns, weather conditions, and road closures. For example, sensors embedded in the road can detect the presence of ice or snow and transmit this information to vehicles to help them navigate safely. These types of communication between infrastructures and vehicles are referred to as I2V and V2I.
However, ISC is restricted to light-of-sight (LOS) situations due to the reliance on visible light. If an obstacle interrupts the light, it will interrupt the information transmission, which is likely to be the case in practical transportation applications such as lane changing. There are solutions to handle such non-light-of-sight (NLOS) situations based on reflections [41]. Alternatively, the vehicle can receive information indirectly through the obstacle if it is an ISC transmitter. In addition, DSRC can be used for NLOS situations [16]. It transmits by wireless radio waves and can penetrate obstacles. However, DSRC is limited by high levels of noise and multipath effects, which ISC can mitigate. Therefore, in certain situations, a hybrid approach that considers both methods may be appropriate.

3.2. Optical Channel Characteristic and Modulation Schemes

The optical channel can be expressed by
R ( t ) = I ( t ) X h ( t ) + A ( t )
where  I ( t )  is the transmitted waveform,  R ( t )  is the received waveform,  h ( t )  is an impulse response,  A ( t )  represents the noise, ⊗ symbol denotes convolution, and X is the detector responsivity. In addition, the average transmitted optical power  P t  in the time period  2 T  is given by
P t = 1 2 T T T I ( t ) d t
The average received optical power P can be presented by
P = H ( 0 ) P t
where  H ( 0 )  is the channel DC gain, given by  H ( 0 ) = h ( t ) d t . In addition, the digital signal-to-noise ratio (SNR) is given by
S N R = X 2 P 2 X b N 0
where  X b  is the bit rate, and  N 0  is the double-sided power-spectral density [42].
There are several modulation schemes available for ISC, each of which can affect the optical power  P t . The most commonly used modulation scheme is on-off keying (OOK), although others such as pulse-width modulation (PWM), pulse-position modulation (PPM), pulse-amplitude modulation (PAM), and orthogonal frequency division multiplexing (OFDM) are also utilized.

3.3. Advantages and Limitations in Image Sensor Receivers

An important feature of the image sensor is spatial separation. In ISC, spatial separation refers to spatially separating signals coming from different transmitters, such as traffic signals, street lights, and brake lights of the lead vehicle. A visual explanation of spatial separation is shown in Figure 7.
Spatial separation enables ISC to perform parallel transmissions using multiple light sources and separating them by image processing. The interference between each light source is significantly reduced by spatial separation, while it is hard to separate the light sources using a single PD due to the interference. An LED array, consisting of multiple LED elements arranged in a grid pattern, can be used to achieve spatial separation [43,44]. Each element in the LED array can be individually controlled to emit light at different intensities. By modulating the light emitted by each LED element, we can encode data into the light signal and transmit it to a receiver. This approach allows for parallel transmissions and can increase the overall data transmission rate. By using an image sensor receiver, we can reduce the impact of interference between light streams from different LEDs. This is particularly useful when dealing with high-speed data transmission, where even small amounts of interference can lead to significant errors. Furthermore, the spatial separation of the image sensor is particularly beneficial in distance measurement. An image sensor captures a wide field of view, allowing obtaining depth information through monocular vision or stereo vision. This can help vehicles make more accurate and safe driving decisions.
However, the effectiveness of spatial separation depends on three factors of the image sensor: resolution, frame rate, and dynamic range. First, regarding sensor resolution, using a low-resolution image sensor will lead to lower data rates and increased vulnerability to errors. A low-resolution image sensor captures images with few pixels, so the amount of information that can be transmitted through ISC is limited to its pixel amount. In addition, if the resolution of the image sensor is low, the interference of the optical signal between adjacent pixels can be significant and can affect the accuracy of demodulation. The low-resolution issue may be solved by image processing algorithms such as Kalman filter and resolution compensation [45]. Second, in terms of frame rate, low frame rates result in low data rates. A low frame rate commonly occurs if an image sensor has a large number of pixels, because a large number of pixels requires high processing ability, resulting in a slow processing speed for each frame. The frame rate of a conventional complementary metal–oxide–semiconductor (CMOS) image sensor is usually around 30 frames per second (fps) [46]. While some commercial cameras or industrial cameras can exceed 30 fps, such as iPhone 14, which has slo-mo video support at 120 fps or 240 fps [47], they sacrifice the video quality. Finally, another problem with ISC is the limited dynamic range of the image sensors. When light irradiates onto the image sensor, the pixels absorb photons and convert them into electrons. These charges then accumulate in the pixel potential well, and when the accumulation limit is reached, no more photoelectric conversion can take place, even if the light is brighter (more photons). In other words, the output digital signal reaches saturation. A low dynamic range leads to a low saturation threshold. If the pixels that contain LEDs are saturated, the demodulation error will be larger, especially with luminance-based modulation such as pulse-width modulation. To solve the issue caused by saturation, we can use a pre-coding method [48] or image processing algorithms, such as phase-only correlation (POC) [49].

4. Isc Receivers

To be used effectively in a moving vehicle environment, the ISC signal must be transmitted at high speed. While CMOS image sensors are the most commonly used sensors for ISC, conventional CMOS sensors often have low frame rates that make it challenging to transmit signals at high speeds. As a result, the ISC receivers, namely the image sensors, are required to utilize encoding techniques and image processing algorithms to achieve high-speed transmission. This section analyzes various ISC receivers, including rolling shutter cameras, high-speed cameras, OCI, and event cameras.

4.1. Rolling-Shutter Camera

According to our survey in Section 2.2, rolling shutter cameras are the most widely used receivers in ISC. The reasons are that rolling shutter cameras are cheap and are utilized in most commercial cameras, including dashcams, closed-circuit television, and those on mobile phones.
In the rolling shutter mechanism, each row of the image sensor is exposed sequentially. This feature enables high-speed data transmission in ISC by allocating data to each row of the rolling shutter sensor [20,50]. Figure 8 provides detailed descriptions of the rolling shutter camera. Figure 8a shows a conventional four-transistor rolling shutter circuit. The row scanner selects the rows from top to bottom individually, and then the column scanner controls reading out the data of an entire row. A pixel in the rolling shutter sensor has no ability to store the accumulated charges due to the lack of memory nodes. Therefore, the exposure process should be followed by an immediate readout to ensure consistent exposure time for each row. This results in the rolling shutter mechanism shown in Figure 8b. The row pixels are read out sequentially thus the reset time starts sequentially. This rolling shutter effect causes the object to appear skewed in the image when the object is moving at high speed. Despite the image skewing drawback, by switching the transmitter faster than the frame rate of the rolling shutter camera, a data rate higher than the frame rate can be achieved due to the row-by-row exposure mechanism. An example is shown in Figure 8c. If a traffic light embedded an array of LEDs blinks at the same speed as the readout rate of each row of the image sensor in this way, the image sensor can receive data at high speed by decoding each row of the image sensor. Each row appears as an independent pattern in the horizontal direction.
In ISC systems that employ rolling shutter effects, the LED blinking rate is higher than the frame rate. Consequently, the images capture stripes due to the LED blinking pattern, such as shown in Figure 8c. In order to decode the stripe image, it is necessary to first extract the LED area in the image, as it is often present in a complex background. Once the LED area has been identified, the stripes can be decoded based on their modulation methods.
The mobility effect of rolling shutter cameras has been investigated in [51,52]. For rolling shutter cameras, which capture the image line-by-line, the motion of the camera during the exposure time can cause distortion or blur in the captured image. This effect can be a significant challenge in ISC systems, where high-quality image capture is essential for reliable communication.
Typically, region-of-interest (RoI) detection methods are used for LED detection. RoI detection is commonly used in computer vision and we can use algorithms such as image thresholding [53] or cam-shift tracking algorithm [54] to extract the LED transmitters. S. Kamiya, et al. have succeeded in error-free detection and communication using a thresholding detecting method when the vehicle is moving at a speed of 15 km/h [55]. Moreover, there are studies conducted to track LEDs by deep learning [52,56].
After completing the LED detection process, we can proceed to the demodulation process. However, the choice of modulation method can significantly impact system performance. In most works the modulation method of OOK is utilized. Other conventional modulation methods include OFDM [52], PWM [57], color-shift keying [58], etc. Furthermore, there are under-sampled modulation methods, including under-sampled frequency shift OOK [59], under-sampled phase shift OOK, spatial-2 phase shift keying (S2-PSK) [60], spatial multiplexing [61], etc.

4.2. Global-Shutter High-Speed Camera

Another type of CMOS image sensor used in ISC is the global shutter image sensor [19,62,63]. Global shutter functions differently from the rolling shutter sensors. Instead of exposing rows sequentially, the global shutter exposes each row simultaneously, making it ideal for capturing high-speed motion without skewing. Cameras that use global shutters to capture high-speed motions are called high-speed cameras, and their frame rates can achieve thousands of fps. According to Section 2.2, 13.8% of the ISC publications used high-speed cameras, which is lower than that of rolling shutter cameras. The main reason is that high-speed cameras are much more expensive than rolling shutter cameras.
Figure 9 provides detailed descriptions of how a global shutter camera works for ISC. Figure 9a shows a conventional five-transistor global shutter sensor circuit. Its scanning mechanism is the same as the rolling shutter that reads out the data row by row. However, the pixel in the global shutter sensor has a memory node, so it can store the accumulated charges. Therefore, the readout process is not required to be conducted immediately after the exposure. In other words, all rows can be exposed simultaneously, and wait a period of time before reading out the charges, as shown in Figure 9b. An example of a global shutter camera capturing a traffic light is shown in Figure 9c. Assume that all the LEDs in the traffic light blink at the same speed as the readout rate of each row of the image sensor, in this way, the image sensor will receive an image of a traffic light with all LEDs illuminated. The output image is different from that of the rolling shutter, even the LED light blinking status is the same. In practice, the LED and the image sensor are usually asynchronous, so the switching speed of the LED must be equal to or less than half of the camera frame rate, according to the Nyquist criterion. As the global shutter high-speed camera captures generally at over 1000 Hz and the data rate is at least half of the frame rate, the data rate can exceed 500 bps.
Since the frame rate of a high-speed camera is over 1000 fps, the position of the transmitter on the image differs only slightly from the last consecutive frame. For example, suppose a vehicle has a speed of 50 km/h and the camera has a frame rate of 1000 fps. Then the vehicle moves only 0.01 m in the time period of each frame, and this distance is usually reflected within one pixel on the image plane. As a result, we can take advantage of this feature that high-speed cameras have less displacement in the position of adjacent frames for vehicle tracking, which rolling shutter cameras cannot perform. In [64], T. Nagura et al. proposed an LED array tracking method using inverted signals. The method involves transmitting an inverted pattern immediately after the original LED array pattern, where the entire LED array is obtained by adding these two consecutive patterns. Since the displacement of the LEDs is very small between the two consecutive frames, this method can detect the position of the LED array in the images pretty accurately. However, the data rate is reduced by half. In [65], S. Usui et al. proposed an LED array detection method using spatial and temporal gradients. Spatial refers to the horizontal and vertical gradients of the image of the current frame for which LEDs need to be detected, calculated using Sobel operator; temporal refers to the gradients of the current frame with respect to the previous and next frames, also calculated using Sobel operator. This method identified LED arrays with low spatial-gradient values and high temporal-gradient values. The experiment results showed that error-free LED tracking was achieved when the vehicle was driving at 30 km/h. In [65], the direction of vehicle motion was perpendicular to the LED array plane, while it could also be used when the direction of vehicle motion was parallel to the LED plane [66].
The mobility effect of the vehicle motion has less influence on high-speed image sensors, as shown in [31]. A pinhole camera model was introduced to project world coordinates to image coordinates in [31]. Three types of ISC systems: I2V-ISC, V2I-ISC, and V2V-ISC, were discussed. For I2V-ISC, the camera moved with the vehicle, and the transmitter was static, while for V2I-ISC, the camera was static, and the transmitter moved with the vehicle. Ref. [31] also compared the vehicle motion models of I2V-ISC and V2I-ISC, and the effects of camera posture on these models. Additionally, Ref. [31] discussed how the relative distance between the transmitter and receiver affects the apparent size and position of objects in the image.
Although the high-speed camera can capture fast motions, it is easily subject to the effect of distance. For a rolling shutter camera, the capturing pattern will not change according to the distance between the vehicle and the transmitter. However, for a global shutter high-speed camera, the size of the transmitter is changing according to the communication distances. Of course, the size of the transmitter on the image is also determined by the image sensor resolution. Hierarchical coding scheme has been proposed in [67] to solve the problem caused by signal degrading at long distance. It divides the LED patterns into different priorities according to the communication distance and assigns different frequencies to their respective priorities by wavelet transform. However, the hierarchical coding scheme has limitations on the number and arrangement of LEDs, which may not match the design of practical transmitters such as traffic lights. To overcome this issue, S. Nishimoto et al. proposed a method of overlay coding in [68,69]. The overlay coding is a more flexible way to design LED applications depending on the transmitter. It distributes long-range or short-range data into large-scale or small-scale of the LED array and encodes them by overlaying each patterns. The experimental results showed that the error-free communication distance of the overlay coding could reach 70 m.
Other limitations of a high-speed camera receiver are the high cost and low resolutions. Due to the advanced technology and components used in high-speed cameras, they can be quite expensive to manufacture and purchase. This can make them cost-prohibitive for smaller organizations that do not have significant budgets for equipment. Another limitation of high-speed camera receivers is their relatively low resolutions compared to other types of cameras. While they are designed to capture images at incredibly high frame rates, the resulting images may not have the same level of detail or clarity as images captured by other types of cameras. This is because high-speed cameras often use smaller sensors and less advanced optics in order to capture images at such high speeds.

4.3. Optical Communication Image Sensor (OCI)

In the previous sections, we discussed two types of image sensors used in ISC: rolling shutter and global shutter. Both of these sensors were originally designed to capture images of the entire scene. However, ISC often requires only pixels that contain transmitters. If these pixels can be extracted and attached to a receiver that receives optical signals at high speed, the data rate will be significantly improved. This kind of receiver is proposed in [19,70,71,72,73,74,75], and known as the OCI. Compared to conventional CMOS image sensors, OCI contains two different kinds of pixels, image pixels (IPx) and communication pixels (CPx), and arranges them side by side as shown in Figure 10. The IPx is a conventional CMOS pixel structure, similar to that of rolling shutter, with an output rate of a few tens of fps. On the other hand, the CPx reduces the capacitance so that it can output the communication signal faster with an output rate of approximately a thousand times that of the IPx [72]. The OCI detects an LED at the IPx and outputs the corresponding CPx in that area. The IPx enables tracking even in a moving vehicle condition. The CPx output is analog, functioning similarly to a PD.
In [72], I. Takai et al. presented the design, fabrication, and capabilities of the OCI, and the experiment results had a 20 Mbps per pixel data rate without LED detection and a 15 Mbps per pixel data rate with real-time LED detection. In [73], I. Takai et al. discussed the capabilities of the OCI used in a V2V communication system, and successfully transmitted vehicle internal data and 13-fps front-view image data of the lead vehicle. In [75], Y. Goto et al. designed optical OFDM into OCI vehicular system, considering frequency response characteristics and circuit noise of the OCI. The system had a performance of 45 Mb/s without bit errors and 55 Mb/s with a BER of  10 5 .
Apparently, the advantage of OCI is the high data rates. It is difficult for conventional CMOS image sensors to achieve data rate at Mbps level. However, the disadvantage of OCI is its low resolution. The current maximum resolution of OCI is 642 × 480 pixels [75], which is much lower than nowadays image sensors. The low resolution may limit its application in long-range communication.

4.4. Event Camera (Dynamic Vision Sensor)

Event cameras, also known as dynamic vision sensors or neuromorphic vision sensors, can be used as a receiver of ISC as well. Unlike rolling shutter or global shutter cameras, an event camera does not acquire images with a shutter or frame, but rather record changes in luminance in a pixel and outputs them as asynchronous events [76]. Figure 11 shows the luminance change of a pixel in an event camera and the event output. “ON” refers to positive luminance change and “OFF” refers to negative luminance change.
Since an event camera only outputs positive or negative values, it is simpler to process than a conventional CMOS image sensor, which gives it microsecond-level temporal resolution. Likewise, LEDs can blink at microsecond-level frequencies, so there is an opportunity to achieve ultra-high data rate in ISC with an event camera. In addition, the dynamic range of the event camera is higher than any other conventional CMOS cameras. Its high dynamic range allows the event camera to capture subtle changes in brightness and improve decoding accuracy.
The existing works on the use of event cameras in ISC have only reached a preliminary level. In [77], W. Shen et al. proposed a pulse waveform for event camera-based ISC systems. The experiment compared the inverse pulse position modulation waveform to the proposed pulse waveform, and the communication distance was 3 m and 8 m, respectively. In [78], G. Chen et al. proposed a positioning method using an event camera as a receiver. The proposed system achieved a positioning accuracy of 3 cm when the height between LEDs and the event camera was within 1 m. The low latency and microsecond-level temporal resolution of the event camera made it possible to identify multiple high-frequency flickering LEDs simultaneously without traditional image processing methods. In [79], Z. Tang et al. investigated a new communication scheme for event camera-based ISC systems using a propeller-type rotary LED transmitter and showed the potential of event-based ISC in moving environment.
These above works demonstrated the potential of event camera-based ISC systems for low-latency data communication. However, the ability to quickly detect motion also leads to problems with the use of event cameras for ISC. Since event cameras work by detecting changes in luminance, they may have difficulty distinguishing between events generated by motion of LEDs and events generated by changes in luminance. Event cameras may encounter difficulties when the camera or LED is moving because the motion causes changes in luminance, which may be interpreted as LED events, especially if the LED is flashing rapidly or in complex patterns. In addition, as with any imaging system, the event camera is subject to a variety of noise sources, including electronic noise from the sensor and shot noise from the light source. This can make it more difficult to distinguish a true LED event from background noise or other sources of interference. Overall, while event cameras offer some promising advantages for VLC systems, there are still some technical challenges that need to be addressed to realize their full potential.

4.5. Summary of ISC Receivers

Table 1 summarizes the ISC receivers discussed in this chapter: rolling shutter camera, high-speed camera, OCI, and event camera. Considering the practical application scenario is important when choosing the appropriate receiver. However, sensor fusion is a possible option that can be applied in many situations.

5. Range Estimation Using LEDs and Image Sensors

In an ITS, accurate range estimation between a vehicle and its surrounding objects is crucial for ensuring safety. By enabling positioning, the vehicle can determine the distance to its surroundings and take measures to avoid collisions. In an ITS-ISC, image sensors have the ability to obtain depth information through the process of triangulation. It takes advantage of the spatial separation between the viewpoints to calculate the relative positions of the LEDs in the scene, which can then be used to determine their distances to the camera. In this section, we introduce ISC ranging methods based on stereo and monocular vision, covering their principles, challenges, and solutions.

5.1. Stereo Vision-Based Range Estimation

The stereo vision-based ISC ranging scheme uses two camera receivers. The distance between the LEDs and the cameras can be determined according to the three-dimensional geometry. Figure 12 illustrates the principle of obtaining range using two cameras and one LED based on the pinhole camera model. The two cameras are considered identical and prior calibration of the two cameras is necessary. Only a single LED is possible to estimate the range between the transmitter and receiver. The cameras are regarded as onboard cameras, and the LED is assumed to be on the lead vehicle or traffic light.
The LED is projected into  ( x l , v )  and  ( x r , v )  on the image planes of left and right cameras, respectively. According to the triangle similarity criteria, the range Z between the LED and the cameras can be given by
Z = f b ρ ( x l x r ) ,
where f is the focal length of two cameras, b is the distance between the two cameras, and  ρ  is the actual size per pixel. Thus, if we can find the corresponding feature point on the image plane of the left and right views, the position of the LED can be determined accurately.
The precision of range estimation is significantly affected by the accuracy of computing  ( x l x r ) , known as the disparity [80]. In [49], the disparity is calculated using phase-only correlation and the sinc function matching to estimate the disparity in subpixel accuracy. Similarly, Ref. [81] estimated the disparity at the subpixel level using equiangular line fitting, of which the processing speed is faster than phase-only correlation. Furthermore, a positioning algorithm based on neural networks is proposed in [82], and a technique for compensating the rolling shutter effect is proposed in [83].

5.2. Range Estimation Based on Monocular Vision

The range between the LED transmitter and the vehicle can also be estimated using a single camera. Figure 13 shows a schematic of using one camera to estimate the distance to the LED.
We can use a pinhole camera model to express the transmitter position on both three-dimensional world space and the image plane of the camera. The range from the LED to the camera can be calculated by triangulation, given by
Z m = f b m ρ ( x 2 x 1 ) ,
where  Z m  is the range between LED and camera,  b m  is the distance between LED1 and LED2, f is the focal length of the camera, and  ρ  is the size per pixel.
However, it is also possible to estimate the range using more than two LEDs based on monocular vision. Reference [84] employed three LEDs to determine the three-dimensional positions of the vehicle, conducting under tunnel scenarios. Reference [85] focused on the multiple-input and multiple-output (MIMO) cases, where more than two LEDs are required, and utilized the S2-PSK modulation method. Furthermore, when it comes to vehicular applications, it is important to address issues related to vehicle vibration. In [86], a Kalman filter is used to reduce the random errors caused by vehicle moving. In [30], two specialized patterns are introduced to tackle the noise generated by the vibrations of a moving vehicle. Since POC is used for estimating  ( x 2 x 1 ) , two LEDs are separately placed on the opposite edge of an LED array in two consecutive frames. The displacement of the upper-left LED on the image plane between the two frames is used to calculate the vehicle vibration. The monocular ranging in [30] is only possible using a high-speed camera, because its time interval must be short enough to compensate a variation caused by moving.

5.3. Range Estimation Using Machine Learning

Machine learning or deep learning techniques have been increasingly used for range estimation in vehicular ISC systems [56,82,87]. In [82], a back-propagation (BP) neural-network learning method has been used for positioning and range estimation. In [87], a coding approach is proposed to provide short-distance and long-distance communication. The authors used an artificial neural network (ANN) to forecast the vehicle’s location. Long-range communication and high-precision positioning can be conducted simultaneously. The results in [87] showed that the average ranging error was 19.8 mm within a distance of 30 m. Machine learning algorithms can also be combined with traditional signal processing techniques to improve range estimation accuracy. For example, machine learning models can be used to denoise the received signal, compensate for distortion, or enhance the features relevant to range estimation, thereby improving the overall accuracy of the range estimation.

5.4. Simultaneous Ranging and Communication

ISC ranging also allows simultaneous communication, making it a more efficient system compared to traditional stand-alone communication or ranging systems. Simultaneous communication saves time, reduces latency, and enhances the overall performance of both communication and ranging. In addition, the hardware complexity of the ISC system can be reduced and the reliability can be improved compared to a separate system because fewer components are required. Furthermore, simultaneous ranging and communication can improve the robustness of the system by reducing the effects of noise and other sources of interference, as well as improving the accuracy of ranging estimations.
Simultaneous ranging and communication can be completed using various approaches. One approach is to use multiple image sensor receivers to receive data signal and estimate the range by stereo vision [56,88]. Another technique is to encode data and ranging information in multiple ISC transmitters, which is then received and decoded by one image sensor receiver [73]. The specific method used depends on the requirements and limitations of the system and the ideal trade-off between communication efficiency and ranging accuracy. Table 2 summarizes the ranging performance in the literature that has been discussed in this review.

6. Conclusions

This review paper analyzed the research trend and basics of ISC, especially its applications to vehicles. ISC is characterized by the use of a two-dimensional image sensor and, therefore, has spatial separation characteristics. It also has the ability to simultaneously perform communication and ranging. We have made a comprehensive review of various types of ISC receivers (rolling shutter cameras, high-speed cameras, event cameras, and optical communication image sensor) and ISC range estimation techniques. It also highlights the challenges and expected future developments in the field of ISC.

Author Contributions

Conceptualization, R.H. and T.Y.; writing—original draft preparation, R.H.; writing—review and editing, T.Y. and R.H.; visualization, R.H.; supervision, T.Y.; project administration, T.Y.; funding acquisition, R.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Nagoya University Interdisciplinary Frontier Fellowship”, supported by Nagoya University and JST, the establishment of university fellowships towards the creation of science technology innovation, Grant Number JPMJFS2120.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
4T-APSfour-transistor active pixel sensor
BERBit-error-rate
CCDCharge-coupled device
CMOSComplementary metal–oxide–semiconductor
CPxCommunication pixel
DSRCDedicated short-range communication
ETCElectronic toll collection
FDfloating diffusion
fpsframe per second
GPSGlobal positioning system
I2VInfrastructure to vehicle
ISCImage sensor communication
IPxImage pixel
ITSIntelligent transportation system
LEDLight-emitting diode
LiFiLight fidelity
LiDARLight detection and ranging
LOSLine-of-sight
MEMmemory
NLOSNon-line-of-sight
OCCOptical camera communication
OCIOptical communication image sensor
OFDMOrthogonal frequency division multiplexing
OOKOn-off keying
PDPhotodiode
POCPhase-only correlation
S2-PSKspatial-2 phase shift keying
V2VVehicle to vehicle
V2IVehicle to infrastructure
V2XVehicle to everything
VLCVisible light communication

References

  1. Pathak, P.; Feng, X.; Hu, P.; Mohapatra, P. Visible Light Communication, Networking, and Sensing: A Survey, Potential and Challenges. IEEE Commun. Surv. Tutor. 2015, 17, 2047–2077. [Google Scholar] [CrossRef]
  2. Jovicic, A.; Li, J.; Richardson, T. Visible light communication: Opportunities, challenges and the path to market. IEEE Commun. Mag. 2013, 51, 26–32. [Google Scholar] [CrossRef]
  3. Karunatilaka, D.; Zafar, F.; Kalavally, V.; Parthiban, R. LED based indoor visible light communications: State of the art. IEEE Commun. Surv. Tutor. 2015, 17, 1649–1678. [Google Scholar] [CrossRef]
  4. Komine, T.; Nakagawa, M. Fundamental analysis for visible-light communication system using LED lights. IEEE Trans. Consum. Electron. 2004, 50, 100–107. [Google Scholar] [CrossRef]
  5. Haas, H.; Yin, L.; Wang, Y.; Chen, C. What is LiFi? J. Light. Technol. 2016, 34, 1533–1544. [Google Scholar] [CrossRef]
  6. Bell, A.G. On the production and reproduction of sound by light. Am. J. Sci. 1880, s3-20, 305–324. [Google Scholar] [CrossRef]
  7. Vanderwater, D.; Tan, I.H.; Hofler, G.; Defevere, D.; Kish, F. High-brightness AlGaInP light emitting diodes. Proc. IEEE 1997, 85, 1752–1764. [Google Scholar] [CrossRef]
  8. Tanaka, Y.; Haruyama, S.; Nakagawa, M. Wireless optical transmissions with white colored LED for wireless home links. In Proceedings of the 11th IEEE International Symposium on Personal Indoor and Mobile Radio Communications, PIMRC 2000. Proceedings (Cat. No. 00TH8525), London, UK, 18–21 September 2000; Volume 2, pp. 1325–1329. [Google Scholar] [CrossRef]
  9. Pang, G.; Kwan, T.; Chan, C.H.; Liu, H. LED traffic light as a communications device. In Proceedings of the 199 IEEE/IEEJ/JSAI International Conference on Intelligent Transportation Systems (Cat. No. 99TH8383), Tokyo, Japan, 5–8 October 1999; pp. 788–793. [Google Scholar] [CrossRef]
  10. Pang, G.K.H.; Liu, H.H.S. LED location beacon system based on processing of digital images. IEEE Trans. Intell. Transp. Syst. 2001, 2, 135–150. [Google Scholar] [CrossRef]
  11. Haruyama, S.; Yamazato, T. Image sensor based visible light communication. In Visible Light Communication; Cambridge University Press: Cambridge, UK, 2015; pp. 181–205. [Google Scholar] [CrossRef]
  12. Kamakura, K. Image Sensors Meet LEDs. IEICE Trans. Commun. 2017, E100.B, 917–925. [Google Scholar] [CrossRef]
  13. Nguyen, T.; Islam, A.; Yamazato, T.; Jang, Y.M. Technical Issues on IEEE 802.15.7m Image Sensor Communication Standardization. IEEE Commun. Mag. 2018, 56, 213–218. [Google Scholar] [CrossRef]
  14. Nguyen, T.; Islam, A.; Hossan, T.; Jang, Y.M. Current Status and Performance Analysis of Optical Camera Communication Technologies for 5G Networks. IEEE Access 2017, 5, 4574–4594. [Google Scholar] [CrossRef]
  15. Hasan, M.K.; Ali, M.O.; Rahman, M.H.; Chowdhury, M.Z.; Jang, Y.M. Optical Camera Communication in Vehicular Applications: A Review. IEEE Trans. Intell. Transp. Syst. 2022, 23, 6260–6281. [Google Scholar] [CrossRef]
  16. Abboud, K.; Omar, H.A.; Zhuang, W. Interworking of DSRC and Cellular Network Technologies for V2X Communications: A Survey. IEEE Trans. Veh. Technol. 2016, 65, 9457–9470. [Google Scholar] [CrossRef]
  17. Lei, A.; Cruickshank, H.; Cao, Y.; Asuquo, P.; Ogah, C.P.A.; Sun, Z. Blockchain-Based Dynamic Key Management for Heterogeneous Intelligent Transportation Systems. IEEE Internet Things J. 2017, 4, 1832–1843. [Google Scholar] [CrossRef]
  18. ETSI TR 102 863 V1.1.1 (2011-06); Intelligent Transport Systems (ITS); Vehicular Communications; Basic Set of Applications; Local Dynamic Map (LDM); Rationale for and Guidance on Standardization. European Telecommunications Standards Institute: Sophia Antipolis, France, 2011. Available online: https://www.etsi.org/deliver/etsi_tr/102800_102899/102863/01.01.01_60/tr_102863v010101p.pdf (accessed on 25 May 2023).
  19. Yamazato, T.; Takai, I.; Okada, H.; Fujii, T.; Yendo, T.; Arai, S.; Andoh, M.; Harada, T.; Yasutomi, K.; Kagawa, K.; et al. Image-sensor-based visible light communication for automotive applications. IEEE Commun. Mag. 2014, 52, 88–97. [Google Scholar] [CrossRef]
  20. Danakis, C.; Afgani, M.; Povey, G.; Underwood, I.; Haas, H. Using a CMOS camera sensor for visible light communication. In Proceedings of the 2012 IEEE Globecom Workshops, Anaheim, CA, USA, 3–7 December 2012; pp. 1244–1248. [Google Scholar] [CrossRef]
  21. Wymeersch, H.; Lien, J.; Win, M.Z. Cooperative Localization in Wireless Networks. Proc. IEEE 2009, 97, 427–450. [Google Scholar] [CrossRef]
  22. Yurtsever, E.; Lambert, J.; Carballo, A.; Takeda, K. A Survey of Autonomous Driving: Common Practices and Emerging Technologies. IEEE Access 2020, 8, 58443–58469. [Google Scholar] [CrossRef]
  23. Bian, R.; Tavakkolnia, I.; Haas, H. 15.73 Gb/s Visible Light Communication With Off-the-Shelf LEDs. J. Light. Technol. 2019, 37, 2418–2424. [Google Scholar] [CrossRef]
  24. Tsonev, D.; Videv, S.; Haas, H. Towards a 100 Gb/s visible light wireless access network. Opt. Express 2015, 23, 1627–1637. [Google Scholar] [CrossRef]
  25. Yang, S.H.; Kim, H.S.; Son, Y.H.; Han, S.K. Three-Dimensional Visible Light Indoor Localization Using AOA and RSS With Multiple Optical Receivers. J. Light. Technol. 2014, 32, 2480–2485. [Google Scholar] [CrossRef]
  26. IEEE Std 802.15.7-2011; IEEE Standard for Local and Metropolitan Area Networks–Part 15.7: Short-Range Wireless Optical Communication Using Visible Light. IEEE: Piscataway, NJ, USA, 2011; pp. 1–309. [CrossRef]
  27. IEEE Std 802.15.7-2018 (Revision of IEEE Std 802.15.7-2011); IEEE Standard for Local and metropolitan area networks–Part 15.7: Short-Range Optical Wireless Communications. IEEE: Piscataway, NJ, USA, 2019. [CrossRef]
  28. ISO 14296:2016; Intelligent Transport Systems—Extension of Map Database Specifications for Applications of Cooperative ITS. ISO: Geneva, Switzerland, 2016. Available online: https://www.iso.org/standard/54587.html (accessed on 25 May 2023).
  29. ISO 22738:2020; Intelligent Transport Systems—Localized Communications— Optical Camera Communication. ISO: Geneva, Switzerland, 2020. Available online: https://www.iso.org/standard/73769.html (accessed on 25 May 2023).
  30. Yamazato, T.; Ohmura, A.; Okada, H.; Fujii, T.; Yendo, T.; Arai, S.; Kamakura, K. Range estimation scheme for integrated I2V-VLC using a high-speed image sensor. In Proceedings of the 2016 IEEE International Conference on Communications Workshops (ICC), Kuala Lumpur, Malaysia, 23–27 May 2016; pp. 326–330. [Google Scholar] [CrossRef]
  31. Yamazato, T.; Kinoshita, M.; Arai, S.; Souke, E.; Yendo, T.; Fujii, T.; Kamakura, K.; Okada, H. Vehicle Motion and Pixel Illumination Modeling for Image Sensor Based Visible Light Communication. IEEE J. Sel. Areas Commun. 2015, 33, 1793–1805. [Google Scholar] [CrossRef]
  32. Kinoshita, M.; Yamazato, T.; Okada, H.; Fujii, T.; Arai, S.; Yendo, T.; Kamakura, K. Motion modeling of mobile transmitter for image sensor based I2V-VLC, V2I-VLC, and V2V-VLC. In Proceedings of the Globecom Workshops (GC Wkshps), Austin, TX, USA, 8–12 December 2014; pp. 450–455. [Google Scholar] [CrossRef]
  33. Griffiths, A.D.; Herrnsdorf, J.; Strain, M.J.; Dawson, M.D. Scalable visible light communications with a micro-LED array projector and high-speed smartphone camera. Opt. Express 2019, 27, 15585–15594. [Google Scholar] [CrossRef] [PubMed]
  34. Chavez-Burbano, P.; Vitek, S.; Teli, S.; Guerra, V.; Rabadan, J.; Perez-Jimenez, R.; Zvanovec, S. Optical camera communication system for Internet of Things based on organic light emitting diodes. Electron. Lett. 2019, 55, 334–336. [Google Scholar] [CrossRef]
  35. Arisue, T.; Yamazato, T.; Okada, H.; Fujii, T.; Kinoshita, M.; Kamakura, K.; Arai, S.; Yendo, T. BER Measurement for Transmission Pattern Design of ITS Image Sensor Communication Using DMD Projector. In Proceedings of the 2020 IEEE 17th Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA, 10–13 January 2020; pp. 1–6. [Google Scholar] [CrossRef]
  36. Ahmed, F.; Conde, M.H.; Martinez, P.L.; Kerstein, T.; Buxbaum, B. Pseudo-Passive Time-of-Flight Imaging: Simultaneous Illumination, Communication, and 3D Sensing. IEEE Sens. J. 2022, 22, 21218–21231. [Google Scholar] [CrossRef]
  37. Zhong, S.; Zhu, Y.; Chi, X.; Shi, H.; Sun, H.; Wang, S. Optical Lensless-Camera Communications Aided by Neural Network. Appl. Sci. 2019, 9, 3238. [Google Scholar] [CrossRef]
  38. Kenney, J.B. Dedicated Short-Range Communications (DSRC) Standards in the United States. Proc. IEEE 2011, 99, 1162–1182. [Google Scholar] [CrossRef]
  39. Grand View Research. Smart Transportation Market Size, Share & Trends Analysis Report By Solution (Ticketing Management System, Parking Management System, Integrated Supervision System), by Service, by Region, and Segment Forecasts, 2023–2030. 2023. Available online: https://www.grandviewresearch.com/industry-analysis/smart-transportation-market (accessed on 25 May 2023).
  40. Hu, J.W.; Zheng, B.Y.; Wang, C.; Zhao, C.H.; Hou, X.L.; Pan, Q.; Xu, Z. A survey on multi-sensor fusion based obstacle detection for intelligent ground vehicles in off-road environments. Front. Inf. Technol. Electron. Eng. 2020, 21, 675–692. [Google Scholar] [CrossRef]
  41. Hassan, N.B.; Ghassemlooy, Z.; Zvánovec, S.; Biagi, M.; Vegni, A.M.; Zhang, M.; Huang, Y. Interference cancellation in MIMO NLOS optical-camera-communication-based intelligent transport systems. Appl. Opt. 2019, 58, 9384–9391. [Google Scholar] [CrossRef]
  42. Kahn, J.; Barry, J. Wireless infrared communications. Proc. IEEE 1997, 85, 265–298. [Google Scholar] [CrossRef]
  43. Nagura, T.; Yamazato, T.; Katayama, M.; Yendo, T.; Fujii, T.; Okada, H. Improved Decoding Methods of Visible Light Communication System for ITS Using LED Array and High-Speed Camera. In Proceedings of the 2010 IEEE 71st Vehicular Technology Conference, Taipei, Taiwan, 16–19 May 2010; pp. 1–5. [Google Scholar] [CrossRef]
  44. Arai, S.; Shiraki, Y.; Yamazato, T.; Okada, H.; Fujii, T.; Yendo, T. Multiple LED arrays acquisition for image-sensor-based I2V-VLC using block matching. In Proceedings of the 2014 IEEE 11th Consumer Communications and Networking Conference (CCNC), Las Vegas, NV, USA, 10–13 January 2014; pp. 605–610. [Google Scholar] [CrossRef]
  45. Tram, V.T.B.; Yoo, M. Vehicle-to-Vehicle Distance Estimation Using a Low-Resolution Camera Based on Visible Light Communications. IEEE Access 2018, 6, 4521–4527. [Google Scholar] [CrossRef]
  46. Ohta, J. Smart CMOS Image Sensors and Applications; CRC Press: Boca Raton, FL, USA, 2020; p. 157. [Google Scholar] [CrossRef]
  47. Apple. iPhone 14 Specifications. 2023. Available online: https://www.apple.com/iphone-14/specs/ (accessed on 25 May 2023).
  48. Kamegawa, S.; Kinoshita, M.; Yamazato, T.; Okada, H.; Fujii, T.; Kamakura, K.; Yendo, T.; Arai, S. Performance Evaluation of Precoded Pulse Width Modulation for Image Sensor Communication. In Proceedings of the 2018 IEEE Globecom Workshops (GC Wkshps), Abu Dhabi, United Arab Emirates, 9–13 December 2018. [Google Scholar]
  49. Huang, R.; Kinoshita, M.; Yamazato, T.; Okada, H.; Kamakura, K.; Arai, S.; Yendo, T.; Fujii, T. Performance Evaluation of Range Estimation for Image Sensor Communication Using Phase-only Correlation. In Proceedings of the 2020 IEEE Globecom Workshops (GC Wkshps), Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  50. Chow, C.W.; Chen, C.Y.; Chen, S.H. Visible light communication using mobile-phone camera with data rate higher than frame rate. Opt. Express 2015, 23, 26080–26085. [Google Scholar] [CrossRef]
  51. Ziehn, J.R.; Roschani, M.; Ruf, M.; Bruestle, D.; Beyerer, J.; Helmer, M. Imaging vehicle-to-vehicle communication using visible light. Adv. Opt. Technol. 2020, 9, 339–348. [Google Scholar] [CrossRef]
  52. Nguyen, H.; Nguyen, V.L.; Tran, D.H.; Jang, Y.M. Rolling Shutter OFDM Scheme for Optical Camera Communication Considering Mobility Environment Based on Deep Learning. Appl. Sci. 2022, 12, 8269. [Google Scholar] [CrossRef]
  53. Liu, Y.; He, W. Signal Detection and Identification in an Optical Camera Communication System in Moving State. J. Phys. Conf. Ser. 2021, 1873, 012015. [Google Scholar] [CrossRef]
  54. Liu, Z.; Guan, W.; Wen, S. Improved Target Signal Source Tracking and Extraction Method Based on Outdoor Visible Light Communication Using an Improved Particle Filter Algorithm Based on Cam-Shift Algorithm. IEEE Photonics J. 2019, 11, 7907520. [Google Scholar] [CrossRef]
  55. Kamiya, S.; Tang, Z.; Yamazato, T. Visible Light Communication System Using Rolling Shutter Image Sensor for ITS. In Proceedings of the 2022 IEEE International Conference on Communications Workshops (ICC Workshops), Seoul, Republic of Korea, 16–20 May 2022; pp. 640–645. [Google Scholar] [CrossRef]
  56. Islam, A.; Hossan, M.T.; Jang, Y.M. Convolutional neural networkscheme–based optical camera communication system for intelligent Internet of vehicles. Int. J. Distrib. Sens. Netw. 2018, 14, 1550147718770153. [Google Scholar] [CrossRef]
  57. Roberts, R. A MIMO protocol for camera communications (CamCom) using undersampled frequency shift ON-OFF keying (UFSOOK). In Proceedings of the 2013 IEEE Globecom Workshops (GC Wkshps), Atlanta, GA, USA, 9–13 December 2013; pp. 1052–1057. [Google Scholar] [CrossRef]
  58. Halawi, S.; Yaacoub, E.; Kassir, S.; Dawy, Z. Performance Analysis of Circular Color Shift Keying in VLC Systems With Camera-Based Receivers. IEEE Trans. Commun. 2019, 67, 4252–4266. [Google Scholar] [CrossRef]
  59. Ji, P.; Tsai, H.M.; Wang, C.; Liu, F. Vehicular Visible Light Communications with LED Taillight and Rolling Shutter Camera. In Proceedings of the 2014 IEEE 79th Vehicular Technology Conference (VTC Spring), Seoul, Republic of Korea, 18–21 May 2014; pp. 1–6. [Google Scholar] [CrossRef]
  60. Nguyen, T.; Islam, A.; Jang, Y.M. Region-of-Interest Signaling Vehicular System Using Optical Camera Communications. IEEE Photonics J. 2017, 9, 7900720. [Google Scholar] [CrossRef]
  61. Tsai, T.T.; Chow, C.W.; Chang, Y.H.; Jian, Y.H.; Liu, Y.; Yeh, C.H. 130-m Image sensor based Visible Light Communication (VLC) using under-sample modulation and spatial modulation. Opt. Commun. 2022, 519, 128405. [Google Scholar] [CrossRef]
  62. Takahashi, K.; Kamakura, K.; Kinoshita, M.; Yamazato, T. Nonlinear Transform for Parallel Transmission for Image-Sensor-based Visible Light Communication. In Proceedings of the ICC 2020—2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–5. [Google Scholar] [CrossRef]
  63. Kibe, S.; Kamakura, K.; Yamazato, T. Parallel Transmission Using M-Point DFT for Image-Sensor-Based Visible Light Communication. In Proceedings of the 2018 IEEE Globecom Workshops (GC Wkshps), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  64. Nagura, T.; Yamazato, T.; Katayama, M.; Yendo, T.; Fujii, T.; Okada, H. Tracking an LED array transmitter for visible light communications in the driving situation. In Proceedings of the 2010 7th International Symposium on Wireless Communication Systems (ISWCS), York, UK, 19–22 September 2010; pp. 765–769. [Google Scholar] [CrossRef]
  65. Usui, S.; Yamazato, T.; Arai, S.; Yendo, T.; Fujii, T.; Okada, H. Utilization of Spatio-temporal image for LED array acquisition in Road to Vehicle Visible Light Communication. In Proceedings of the 20th World Congress on Intelligent Transport Systems, Tokyo, Japan, 14–18 October 2013. [Google Scholar]
  66. Nakamura, K.; Huang, R.; Yamazato, T.; Kinoshita, M.; Kamakura, K.; Arai, S.; Yendo, T.; Fujii, T. Roadside LED array acquisition for road-to-vehicle visible light communication using spatial-temporal gradient values. IEICE Commun. Express 2022, 11, 462–467. [Google Scholar] [CrossRef]
  67. Arai, S.; Mase, S.; Yamazato, T.; Endo, T.; Fujii, T.; Tanimoto, M.; Kidono, K.; Kimura, Y.; Ninomiya, Y. Experimental on Hierarchical Transmission Scheme for Visible Light Communication using LED Traffic Light and High-Speed Camera. In Proceedings of the 2007 IEEE 66th Vehicular Technology Conference, Baltimore, MD, USA, 30 September–3 October 2007; pp. 2174–2178. [Google Scholar] [CrossRef]
  68. Nishimoto, S.; Nagura, T.; Yamazato, T.; Yendo, T.; Fujii, T.; Okada, H.; Arai, S. Overlay coding for road-to-vehicle visible light communication using LED array and high-speed camera. In Proceedings of the 14th International IEEE Conference on Intelligent Transportation Systems (ITSC 2011), Washington, DC, USA, 5–7 October 2011; pp. 1704–1709. [Google Scholar] [CrossRef]
  69. Nishimoto, S.; Yamazato, T.; Okada, H.; Fujii, T.; Yendo, T.; Arai, S. High-speed transmission of overlay coding for road-to-vehicle visible light communication using LED array and high-speed camera. In Proceedings of the IEEE Globecom Workshops (OWC 2012), Anaheim, CA, USA, 3–7 December 2012; pp. 1234–1238. [Google Scholar] [CrossRef]
  70. Oike, Y.; Ikeda, M.; Asada, K. A smart image sensor with high-speed feeble ID-beacon detection for augmented reality system. In Proceedings of the ESSCIRC 2004—29th European Solid-State Circuits Conference (IEEE Cat. No. 03EX705), Estoril, Portugal, 16–18 September 2003; pp. 125–128. [Google Scholar] [CrossRef]
  71. Itoh, S.; Takai, I.; Sarker, M.; Hamai, M.; Yasutomi, K.; Andoh, M.; Kawahito, S. A CMOS image sensor for 10Mb/s 70m-range LED-based spatial optical communication. In Proceedings of the IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC’10), San Francisco, CA, USA, 7–11 February 2010; pp. 402–403. [Google Scholar] [CrossRef]
  72. Takai, I.; Ito, S.; Yasutomi, K.; Kagawa, K.; Andoh, M.; Kawahito, S. LED and CMOS Image Sensor Based Optical Wireless Communication System for Automotive Applications. IEEE Photonics J. 2013, 5, 6801418. [Google Scholar] [CrossRef]
  73. Takai, I.; Harada, T.; Andoh, M.; Yasutomi, K.; Kagawa, K.; Kawahito, S. Optical Vehicle-to-Vehicle Communication System Using LED Transmitter and Camera Receiver. IEEE Photonics J. 2014, 6, 7902513. [Google Scholar] [CrossRef]
  74. Goto, Y.; Takai, I.; Yamazato, T.; Okada, H.; Fujii, A.; Kawahito, S.; Arai, S.; Yendo, T.; Kamakura, K. BER characteristic of optical-OFDM using OCI. In Proceedings of the 2014 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), Ishigaki, Japan, 17–20 November 2014; pp. 328–331. [Google Scholar] [CrossRef]
  75. Goto, Y.; Takai, I.; Yamazato, T.; Okada, H.; Fujii, T.; Kawahito, S.; Arai, S.; Yendo, T.; Kamakura, K. A New Automotive VLC System Using Optical Communication Image Sensor. IEEE Photonics J. 2016, 8, 6802716. [Google Scholar] [CrossRef]
  76. Gallego, G.; Delbrück, T.; Orchard, G.; Bartolozzi, C.; Taba, B.; Censi, A.; Leutenegger, S.; Davison, A.J.; Conradt, J.; Daniilidis, K.; et al. Event-Based Vision: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 154–180. [Google Scholar] [CrossRef]
  77. Shen, W.H.; Chen, P.W.; Tsai, H.M. Vehicular Visible Light Communication with Dynamic Vision Sensor: A Preliminary Study. In Proceedings of the 2018 IEEE Vehicular Networking Conference (VNC), Taipei, Taiwan, 5–7 December 2018; pp. 1–8. [Google Scholar] [CrossRef]
  78. Chen, G.; Chen, W.; Yang, Q.; Xu, Z.; Yang, L.; Conradt, J.; Knoll, A. A Novel Visible Light Positioning System With Event-Based Neuromorphic Vision Sensor. IEEE Sensors J. 2020, 20, 10211–10219. [Google Scholar] [CrossRef]
  79. Tang, Z.; Yamazato, T.; Arai, S. A Preliminary Investigation For Event Camera-Based Visible Light Communication Using The Propeller-type Rotary LED Transmitter. In Proceedings of the 2022 IEEE International Conference on Communications Workshops (ICC Workshops), Seoul, Republic of Korea, 16–20 May 2022; pp. 646–650. [Google Scholar] [CrossRef]
  80. Kaehler, A.; Bradski, G. Learning OpenCV 3: Computer Vision in C++ with the OpenCV Library, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2016. [Google Scholar]
  81. Kinoshita, M.; Kamakura, K.; Yamazato, T.; Okada, H.; Fujii, T.; Arai, S.; Yendo, T. Stereo Ranging Method Using LED Transmitter for Visible Light Communication. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  82. Ifthekhar, M.S.; Saha, N.; Jang, Y.M. Stereo-vision-based cooperative-vehicle positioning using OCC and neural networks. Opt. Commun. 2015, 352, 166–180. [Google Scholar] [CrossRef]
  83. Do, T.; Yoo, M. Visible light communication based vehicle positioning using LED street light and rolling shutter CMOS sensors. Opt. Commun. 2018, 407, 112–126. [Google Scholar] [CrossRef]
  84. Kim, B.W.; Jung, S.Y. Vehicle Positioning Scheme Using V2V and V2I Visible Light Communications. In Proceedings of the 2016 IEEE 83rd Vehicular Technology Conference (VTC Spring), Nanjing, China, 15–18 May 2016; pp. 1–5. [Google Scholar] [CrossRef]
  85. Hossan, M.T.; Chowdhury, M.Z.; Hasan, M.K.; Shahjalal, M.; Nguyen, T.; Le, N.T.; Jang, Y.M. A New Vehicle Localization Scheme Based on Combined Optical Camera Communication and Photogrammetry. Mob. Inf. Syst. 2018, 2018, 8501898. [Google Scholar] [CrossRef]
  86. He, J.; Tang, K.; He, J.; Shi, J. Effective vehicle-to-vehicle positioning method using monocular camera based on VLC. Opt. Express 2020, 28, 4433–4443. [Google Scholar] [CrossRef]
  87. He, J.; Zhou, B. A Deep Learning-Assisted Visible Light Positioning Scheme for Vehicles With Image Sensor. IEEE Photonics J. 2022, 14, 1–7. [Google Scholar] [CrossRef]
  88. Huang, R.; Kinoshita, M.; Yamazato, T.; Okada, H.; Kamakura, K.; Arai, S.; Yendo, T.; Fujii, T. Simultaneous Visible Light Communication and Ranging Using High-Speed Stereo Cameras Based on Bicubic Interpolation Considering Multi-Level Pulse-Width Modulation (Advance Publication). IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2023, E106-A. [Google Scholar] [CrossRef]
Figure 1. Literaturesurveys on visible light communications (VLC). The survey was performed by searching the Scopus database for documents with “visible light communication” in the title, keywords, and abstracts, and publication year before 2023. These data are based on search records in February 2023. (a) Annual growth in the number of research publications related to VLC [4,9]. (b) Analysis of document types in research publications of VLC. (c) Categorizing receiver types in research publications of VLC.
Figure 1. Literaturesurveys on visible light communications (VLC). The survey was performed by searching the Scopus database for documents with “visible light communication” in the title, keywords, and abstracts, and publication year before 2023. These data are based on search records in February 2023. (a) Annual growth in the number of research publications related to VLC [4,9]. (b) Analysis of document types in research publications of VLC. (c) Categorizing receiver types in research publications of VLC.
Photonics 10 00617 g001
Figure 2. A literature survey of image sensor communication (ISC) [10,19,20]. The survey was performed by searching the Scopus database for publications with “image sensor communication” or “optical camera communication” or “screen camera communication” or “display camera communication” or {“visible light communication” and “camera”} or {“visible light communication” and “image sensor”} in the title, keywords, and abstracts, and publication year before 2023. These data are based on search records in February 2023.
Figure 2. A literature survey of image sensor communication (ISC) [10,19,20]. The survey was performed by searching the Scopus database for publications with “image sensor communication” or “optical camera communication” or “screen camera communication” or “display camera communication” or {“visible light communication” and “camera”} or {“visible light communication” and “image sensor”} in the title, keywords, and abstracts, and publication year before 2023. These data are based on search records in February 2023.
Photonics 10 00617 g002
Figure 3. Categorizing image sensor communication research publications by applications of communication, ranging, simultaneous ranging and communication, detection and tracking, or others. Others include channel characterizing [31], vehicle vibration modeling [32], etc.
Figure 3. Categorizing image sensor communication research publications by applications of communication, ranging, simultaneous ranging and communication, detection and tracking, or others. Others include channel characterizing [31], vehicle vibration modeling [32], etc.
Photonics 10 00617 g003
Figure 4. Categorizing image sensor communication (ISC) research publications by transmitter types and receiver types. The number of publications is 1021. (a) Categorizing ISC publications by transmitter type. Others includes micro-LED [33], organic-LED [34], projector [35], laser diodes, etc. (b) Categorizing ISC publications by receiver type. Others includes charge-coupled device (CCD) camera, time-of-flight sensor [36], lensless-camera [37], etc.
Figure 4. Categorizing image sensor communication (ISC) research publications by transmitter types and receiver types. The number of publications is 1021. (a) Categorizing ISC publications by transmitter type. Others includes micro-LED [33], organic-LED [34], projector [35], laser diodes, etc. (b) Categorizing ISC publications by receiver type. Others includes charge-coupled device (CCD) camera, time-of-flight sensor [36], lensless-camera [37], etc.
Photonics 10 00617 g004
Figure 5. Categorizing image sensor communication research publications by application scenarios including indoor, outdoor, underwater, or others.
Figure 5. Categorizing image sensor communication research publications by application scenarios including indoor, outdoor, underwater, or others.
Photonics 10 00617 g005
Figure 6. The basic idea of image sensor communication for vehicular applications including vehicle-to-vehicle (V2V), vehicle-to-infrastructure (V2I), and infrastructure-to-vehicle (I2V) communications. Data signals are emitted from traffic lights and vehicle headlights and taillights. Image sensors are mounted on the vehicle to receive visible light signals. The image sensor converts the received visible light signal into an electrical signal and decodes the signal using image processing technology.
Figure 6. The basic idea of image sensor communication for vehicular applications including vehicle-to-vehicle (V2V), vehicle-to-infrastructure (V2I), and infrastructure-to-vehicle (I2V) communications. Data signals are emitted from traffic lights and vehicle headlights and taillights. Image sensors are mounted on the vehicle to receive visible light signals. The image sensor converts the received visible light signal into an electrical signal and decodes the signal using image processing technology.
Photonics 10 00617 g006
Figure 7. A visual explanation of spatial separation. Two cars are driving on the road and the camera of the following vehicle receives two light streams from the traffic light and the backlight of the lead vehicle. An image sensor receiver can separate these two light sources spatially.
Figure 7. A visual explanation of spatial separation. Two cars are driving on the road and the camera of the following vehicle receives two light streams from the traffic light and the backlight of the lead vehicle. An image sensor receiver can separate these two light sources spatially.
Photonics 10 00617 g007
Figure 8. The circuit (a), exposure mechanism (b), and an example (c) of a rolling shutter image sensor. (a) A simplified rolling shutter image sensor circuit with 7 × 7 pixels. The image on the top right shows the basic circuit for a traditional rolling shutter pixel which is a four-transistor active pixel sensor (4T-APS) [46]. A 4T-APS contains a photodiode (PD), a floating diffusion (FD), and four transistors,  M R S M T G M S F , and  M S E L . (b) The rolling shutter exposure schematic shown over time. The rolling shutter method exposes rows sequentially, followed by an immediate readout. The specific mechanism is decided by the pixel circuit. (c) An example of a 7 × 7 rolling shutter image sensor capturing a blinking traffic light. We assume that the LED switching rate is equal to the readout rate. The reset is included in the exposure.
Figure 8. The circuit (a), exposure mechanism (b), and an example (c) of a rolling shutter image sensor. (a) A simplified rolling shutter image sensor circuit with 7 × 7 pixels. The image on the top right shows the basic circuit for a traditional rolling shutter pixel which is a four-transistor active pixel sensor (4T-APS) [46]. A 4T-APS contains a photodiode (PD), a floating diffusion (FD), and four transistors,  M R S M T G M S F , and  M S E L . (b) The rolling shutter exposure schematic shown over time. The rolling shutter method exposes rows sequentially, followed by an immediate readout. The specific mechanism is decided by the pixel circuit. (c) An example of a 7 × 7 rolling shutter image sensor capturing a blinking traffic light. We assume that the LED switching rate is equal to the readout rate. The reset is included in the exposure.
Photonics 10 00617 g008aPhotonics 10 00617 g008b
Figure 9. A simplified sensor circuit (a), exposure mechanism (b), and an example of capturing a blinking traffic light (c) with a global shutter camera. (a) A simplified global shutter image sensor circuit with 7 × 7 pixels. The image on the upper-right shows the basic circuit of a traditional five-transistor global shutter active pixel sensor (5T-global shutter APS) [46]. A 5T-global shutter APS contains a photodiode (PD), a memory (MEM) node, and five transistors,  M P D M G S M R S T M S F , and  M S E L  [46]. The main difference to the rolling shutter pixel is that a global shutter pixel has a memory node for keeping the charge. (b) The global shutter exposure schematic shown over time. Global shutter first exposes all the rows simultaneously, and then each row waits for the readout. (c) An example of a 7 × 7 global shutter image sensor capturing a blinking traffic light. We assume that the LED switching rate is equal to the readout rate. The global shutter only captures the first blinking pattern of the traffic light.
Figure 9. A simplified sensor circuit (a), exposure mechanism (b), and an example of capturing a blinking traffic light (c) with a global shutter camera. (a) A simplified global shutter image sensor circuit with 7 × 7 pixels. The image on the upper-right shows the basic circuit of a traditional five-transistor global shutter active pixel sensor (5T-global shutter APS) [46]. A 5T-global shutter APS contains a photodiode (PD), a memory (MEM) node, and five transistors,  M P D M G S M R S T M S F , and  M S E L  [46]. The main difference to the rolling shutter pixel is that a global shutter pixel has a memory node for keeping the charge. (b) The global shutter exposure schematic shown over time. Global shutter first exposes all the rows simultaneously, and then each row waits for the readout. (c) An example of a 7 × 7 global shutter image sensor capturing a blinking traffic light. We assume that the LED switching rate is equal to the readout rate. The global shutter only captures the first blinking pattern of the traffic light.
Photonics 10 00617 g009aPhotonics 10 00617 g009b
Figure 10. A simplified circuit diagram for an optical communication image sensor with 3 × 7 image pixels (IPx) and 3 × 7 communication pixels (CPx). The CPx and IPx are arranged side by side. The two figures on the right show the structure of an IPx and a CPx, adapted from [72]. CDS refers to correlated double sampling, M1 is a transistor used for readout amplifier, M2 is a transistor used for CPx selection, VDD stands for power supply voltage, and Vcs stands for charge-sensing node. Compared to an IPx, the CPx reduce two transistors, making the response faster. Details can be found in [72].
Figure 10. A simplified circuit diagram for an optical communication image sensor with 3 × 7 image pixels (IPx) and 3 × 7 communication pixels (CPx). The CPx and IPx are arranged side by side. The two figures on the right show the structure of an IPx and a CPx, adapted from [72]. CDS refers to correlated double sampling, M1 is a transistor used for readout amplifier, M2 is a transistor used for CPx selection, VDD stands for power supply voltage, and Vcs stands for charge-sensing node. Compared to an IPx, the CPx reduce two transistors, making the response faster. Details can be found in [72].
Photonics 10 00617 g010
Figure 11. Analyzing generated events by correlating with actual LED blinking status and their signal intensity. The  I m a x  intensity indicates when the LED is illuminated, while the  I m i n  intensity indicates when the LED is off. When a negative luminance change is detected, the event camera generates an “OFF” event, and when a positive luminance change is detected, the event camera generates an “ON” event.
Figure 11. Analyzing generated events by correlating with actual LED blinking status and their signal intensity. The  I m a x  intensity indicates when the LED is illuminated, while the  I m i n  intensity indicates when the LED is off. When a negative luminance change is detected, the event camera generates an “OFF” event, and when a positive luminance change is detected, the event camera generates an “ON” event.
Photonics 10 00617 g011
Figure 12. A schematic of two cameras capturing a single LED. The two cameras are supposed to be perfectly calibrated. The top view is shown on the right. P refers to the LED, b is the distance between the left and right camera, f is the camera focal length,  ρ  is the size for one pixel, and Z is the distance we need to estimate.
Figure 12. A schematic of two cameras capturing a single LED. The two cameras are supposed to be perfectly calibrated. The top view is shown on the right. P refers to the LED, b is the distance between the left and right camera, f is the camera focal length,  ρ  is the size for one pixel, and Z is the distance we need to estimate.
Photonics 10 00617 g012
Figure 13. The schematic of range estimation using two LEDs and one camera. The top view is shown on the right.  b m  is the distance between LED1 and LED2, f is the camera focal length,  ρ  is the size for one pixel, and  Z m  is the distance we need to estimate.
Figure 13. The schematic of range estimation using two LEDs and one camera. The top view is shown on the right.  b m  is the distance between LED1 and LED2, f is the camera focal length,  ρ  is the size for one pixel, and  Z m  is the distance we need to estimate.
Photonics 10 00617 g013
Table 1. A summary of communication performance in existing literature categorizing by image sensor communication (ISC) receivers.
Table 1. A summary of communication performance in existing literature categorizing by image sensor communication (ISC) receivers.
Receiver TypeReferenceData RateCommunication DistanceVehicle Speed
Rolling shutter[55]600∼1000 bps5∼70 m15∼20 km/h
[61]720 bps100 mN/A
High-speed camera (global shutter)[19]32 kbps30∼65 m30 km/h
[43]128 kbps10∼120 mN/A
[69]40 kbps20∼70 mN/A
Optical communication image sensor (OCI)[19]10 Mbps20 m25 km/h
[72]20 MbpsN/AN/A
[75]55 Mbps1.5 mN/A
Event camera[77]16 kbps8 mN/A
Table 2. A summary of ranging performance in existing literature.
Table 2. A summary of ranging performance in existing literature.
Ranging MethodReferenceRanging ErrorCommunication DistanceReceiverVehicle SpeedSimultaneous Communication
Monocular Ranging[30]0.3 m30∼60 mhigh-speed camera30 km/hNo
[73]N/Aaround 8 mOCI12.6∼14.0 km/hYes
[84]1 m0∼60 mN/AN/ANo
Stereo Ranging[88]0.5 m20∼60 mhigh-speed cameraN/AYes
[83]0.1∼1.5 m0∼100 mrolling shutter camera0∼100 km/hYes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, R.; Yamazato, T. A Review on Image Sensor Communication and Its Applications to Vehicles. Photonics 2023, 10, 617. https://doi.org/10.3390/photonics10060617

AMA Style

Huang R, Yamazato T. A Review on Image Sensor Communication and Its Applications to Vehicles. Photonics. 2023; 10(6):617. https://doi.org/10.3390/photonics10060617

Chicago/Turabian Style

Huang, Ruiyi, and Takaya Yamazato. 2023. "A Review on Image Sensor Communication and Its Applications to Vehicles" Photonics 10, no. 6: 617. https://doi.org/10.3390/photonics10060617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop