Back to articles
JIST-first
Volume: 35 | Article ID: ISS-337
Image
Self-Supervised Intensity-Event Stereo Matching
  DOI :  10.2352/J.ImagingSci.Technol.2022.66.6.060402  Published OnlineNovember 2022
Abstract
Abstract

Event cameras are novel bio-inspired vision sensors that output pixel-level intensity changes in microsecond accuracy with high dynamic range and low power consumption. Despite these advantages, event cameras cannot be directly applied to computational imaging tasks due to the inability to obtain high-quality intensity and events simultaneously. This paper aims to connect a standalone event camera and a modern intensity camera so that applications can take advantage of both sensors. We establish this connection through a multi-modal stereo matching task. We first convert events to a reconstructed image and extend the existing stereo networks to this multi-modality condition. We propose a self-supervised method to train the multi-modal stereo network without using ground truth disparity data. The structure loss calculated on image gradients is used to enable self-supervised learning on such multi-modal data. Exploiting the internal stereo constraint between views with different modalities, we introduce general stereo loss functions, including disparity cross-consistency loss and internal disparity loss, leading to improved performance and robustness compared to existing approaches. Our experiments demonstrate the effectiveness of the proposed method, especially the proposed general stereo loss functions, on both synthetic and real datasets. Finally, we shed light on employing the aligned events and intensity images in downstream tasks, e.g., video interpolation application.

Subject Areas :
Views 9
Downloads 1
 articleview.views 9
 articleview.downloads 1
  Cite this article 

Jinjin Gu, Jinan Zhou, Ringo Sai Wo Chu, Yan Chen, Jiawei Zhang, Xuanye Cheng, Song Zhang, Jimmy S. Ren, "Self-Supervised Intensity-Event Stereo Matchingin Electronic Imaging,  2022,  pp 060402-1 - 060402-16,  https://doi.org/10.2352/J.ImagingSci.Technol.2022.66.6.060402

 Copy citation
  Copyright statement 
Copyright © Society for Imaging Science and Technology 2022
  Article timeline 
  • received July 2022
  • accepted October 2022
  • PublishedNovember 2022

Preprint submitted to:
jist
JIMTE6
Journal of Imaging Science and Technology
J. Imaging Sci. Technol.
J. Imaging Sci. Technol.
1062-3701
1943-3522
Society for Imaging Science and Technology
1.
Introduction
Event cameras measure changes in brightness at each pixel independently instead of reporting pixel activations. Event cameras have attracted increasing attention for their high temporal resolution, high dynamic range and low power consumption features, and have been applied to various computer vision tasks [15]. However, existing event cameras are either unable to obtain high-quality image pixel intensities (DVS [6] sensors only output events) or suffer low spatial resolution and lack of color information (dynamic and active-pixel vision sensors [7]). These limitations make it difficult for event cameras to assist computational imaging tasks, as we cannot obtain high-resolution intensity images and events simultaneously.
In this paper, we aim to connect a standalone event camera and a separate modern intensity camera so that applications could exploit the advantages of both sensors (see Figure 1). Such application scenarios are not uncommon for most consumer-level imaging devices, simply because acquiring colorful visual contents with high resolution, high speed, and low power consumption is without the scope of any individual image sensors. We establish a connection between these two sensors through a computational stereo matching model and estimate their disparity. This disparity describes the relationship between these two sensors and allows the sensors to be combined to complete the task that one sensor cannot achieve, e.g., obtaining both high-resolution images and events simultaneously for downstream tasks.
Figure 1.
The proposed intensity-event stereo setting, in which we use an event camera and an intensity camera. With the proposed self-supervised stereo matching model, we can not only obtain the disparity used to calculate the depth map, but also build a connection between these two sensors. (a) The signals are from these two displaced sensors and are unaligned. (b) We align these two signals by the proposed method.
However, studying this problem is NOT a naive extension of the existing stereo matching methods on a new sensor setting, owing to the following technical barriers. First, the current stereo networks are not optimal for multi-modal problems. They assume that left and right view images have the same modality and use the weights shared feature extraction model for these images. Second, it is challenging to obtain multi-modal data for training as the acquisition of ground truth disparity for each different setting is expensive. To practically apply a multi-modal stereo framework, we need a robust training strategy and get rid of the shackles of data annotation. On this front, self-supervised learning provides a promising perspective to use the inherent constraints and characteristics of the data to learn the desired stereo matching without ground truth disparity indirectly. However, the prerequisite for the success of the existing self-supervised stereo matching framework is to establish the photometric consistency relationship between the projected images from two views. This brings up the third problem. The left and right view signals have different physical meanings and data structures in our setting. This causes the failure of the self-supervised learning framework as the previous photometric constraint does not hold.
In this work, we propose a self-supervised method for learning the multi-modal stereo matching without any ground truth disparity (see Figure 2). To facilitate the existing outstanding image stereo models on the proposed intensity-event setting, we first convert the event stream to roughly reconstructed images through the off-the-shelf models [8, 9]. The roughly reconstructed images are still in a different modality from the images of the other view as the color and detail information cannot be well restored. We improve the existing stereo networks and make images with different modalities to use modality-specific feature extraction sub-modules. In the proposed self-supervised method, we introduce a gradient structure consistency loss for the geometry constraints between the intensity and the reconstructed images after projection, which mainly leverages the edge information provided by events. Last but not least, only using the structure consistency may result in poor quality disparity maps as the supervision is sparse and vague. To overcome this issue, we propose a novel loss based on the cross-consistency between the disparities calculated across different views using different modality images. We also constrain our training according to the fact that the disparity of the same view should be zero. The proposed loss functions lead to improved stereo matching performance and robustness.
The calculated disparity maps can be used in many computational photography tasks, with depth estimation first. Projecting events to intensity camera view also allows many applications that could not be realized in the past due to hardware limitations. We can now obtain high-resolution events and intensity images simultaneously. At last, we experimentally demonstrate the potential of the proposed framework using the warped event to facilitate event-based video frame interpolation task.
2.
Related Work
2.1
Event Cameras
Event camera is a kind of sensor that records signals when the scene exhibits illumination changes [6, 7]. An event camera reports signals (events) asynchronously when the log intensity change exceeds a preset threshold τ. We have witnessed the rise of event cameras due to their distinctive advantages over conventional active pixel cameras, e.g., higher frame rate, higher dynamic range and lower power consumption. These properties attracted the use of event cameras in many computer vision tasks, e.g., tracking [1, 10], deblurring [2, 11], optical flow estimation [3, 12], SLAM [4, 13, 14], video frame interpolation [5, 11]. However, the unique data structure of event cameras renders the existing computer vision tools and algorithms unusable, which places a major obstacle against the application of event cameras. Many works have been focusing on bridging events and conventional cameras by reconstructing intensity frames from events [1520], thus allowing modern vision algorithm to take place. Rebecq et al. [8] proposed E2VID, a recurrent network to reconstruct videos from a stream of events and trained it on a large amount of simulated event data. Scheerlinck et al. [9] proposed FireNet, which simplified the neural architecture in Ref. [8] with a smaller number of parameters while maintaining similar quantitative results. Although many studies have attempted to reconstruct the intensity image from the event, none of these methods can recover the intensity and color information well. Therefore, the absence of color information in the reconstructed image degrades the performance for downstream tasks. In our application, the color mismatch makes the existing self-supervised stereo matching algorithm based on photometric consistency invalid.
Figure 2.
This figure shows the overall framework of the proposed method. We first obtain rough reconstruction results using the existing methods from the right view events. The multi-modal stereo network predicted a disparity map based on the right reconstructed results and left intensity images. The whole system uses three kinds of loss functions for self-supervised training.
2.2
Stereo Matching
Stereo matching is the process of linking the pixels in different views that correspond to the same point of the scene. It follows a long line of research works. Early works involve searching and matching corresponding pixels on the epipolar line [21, 22]. Recently, deep learning based methods have dominated the field of stereo matching due to their superior performance and usability. Zbontar and LeCun [23] are among the first to use a convolutional network for computing stereo matching cost in image pairs. Following this, a number of studies were proposed to improve the performance, e.g., inner product layer [24], encoder-decoder architecture [25], 3D convolution cost-volume module [26], spatial pyramid pooling and 3D hourglass convolution [27], guided attention cost-column [28], PatchMatch module for sparse cost volume representation [29], intra-scaling cost aggregation [30]. With the development of various sensors, multi-model and cross-spectral stereo matching has become an emerging topic [3136]. But none of them is suitable for calculating the correspondence between intensity images and events or event reconstruction images. Concurrent with our work, Mostafavi et al. [37] investigated stereo matching with event-intensity cameras on both views and proposed an event-intensity network that refines image details using events. Our work is essentially different in purpose and method; we use only one intensity and one event camera and train our model self-supervised.
2.3
Self-Supervised Learning
Learning-based stereo methods are data-hungry. They often require a lot of ground truth data for training. Over the past few years, self-supervised models have been developed to learn stereo matching without ground truth annotations. They are usually built on the principles of disparity smoothness prior and re-projection photometric consistency constraints. Garg et al. [38] tackled monocular depth estimation by minimizing the loss between the source image and backwards-warping from the subsidiary stereo image. Similarly, Godard et al. [39] included a left-right consistency to enforce disparity prediction. They further proposed a new minimum re-projection loss and auto-masking loss to improve the performance [40]. Zhou et al. [41] adopted left-right check to guide the training and pick suitable matching as training data. Zhi et al. [32] proposed a self-supervised learning framework for cross-spectral stereo matching. They introduced a material-aware loss function to handle regions with unreliable matching. However, their method involves the translation between intensity and near-infrared images and is thus unsuitable for our setting.
3.
Method
In this section, we describe our self-supervised intensity-event stereo matching framework. We first introduce the problem formulation and overall framework design in Section 3.1. We then describe the modified stereo network for multi-modal problem in Section 3.2. The loss functions are introduced in Section 3.3, featuring a gradient structure consistency loss and general losses for multi-modal stereo matching.
Figure 3.
The visualization of different loss functions. On the left of the figure is the event, the reconstructed image using E2VID [8], the intensity image and their gradient visualization. The heat maps show the loss values with different displaced pixel numbers. In the centre of the heat maps, there is no displacement between the two images being compared, and the loss value should be the smallest.
3.1
Overall Framework
The multi-modal intensity-event stereo matching problem is first formulated as follows. As shown in Fig. 1, an intensity camera and an auxiliary horizontally displaced event camera are used in our setting. In this work, we assume that the camera on the left is the image camera, and the one on the right is the event camera. Let Il be the left view intensity image and {Emr}mN be the event stream obtained by the right event camera within a short amount of time before the intensity image is captured. The underlying problem can be considered a data association problem, that is, to find correspondences between the points in the left image and right events. The correspondences are formed as the final disparity map, which is also the output of the stereo matching problem.
However, the right signal (event) is not in the same modality as the left signal (intensity images), failing the existing methods with the given problem setting. A reconstruction network is employed at first to convert the event stream {Emr}mN to a roughly reconstructed image Ir. In our work, we use two popular event reconstruction models, E2VID [8, 42] and FireNet [9]. Note that these models are replaceable. Given these two images, we can adapt the existing image stereo models to predict disparity between images, which is also the disparity between the left image and the right events. However, the event camera does not record the value of the pixels and only records the pixels changing. Thus the rough reconstruction Ir only contains usable edge information but are unreliable in color and detail and still in different modalities with Il (see the reconstruction result in Figure 3).
3.2
Multi-Modal Stereo Network
As stated above, we need a multi-modal stereo network to handle inputs with different modalities (either the right view is event voxel or reconstructed images). Although we can apply the previous convolutional stereo matching networks theoretically, the difference in modalities still poses challenges. Most stereo networks are composed of feature extraction, correlation and aggregation sub-models, and the feature extraction model usually share weights for both two views. This weight sharing strategy is effective originally but poses limitations for images with different modalities. We make the minor changes to these networks to make images with different modalities using modality-specific feature extraction sub-modules. This design has two advantages. First, the feature extraction models dedicated to different modalities avoid confusion between different images. Second, the new design allows us to swap the modalities of the left and right views and predict disparity of the other view, as long as we swap them together with the feature extraction branches. The second property is essential for the cross-consistency loss, which we will describe in Section 3.3. Note that this structure also allows the event voxel to be used directly as input.
Figure 4.
Using projection method and the multi-modal stereo model, we can obtain all four combinations of two views and two modalities. The proposed general multi-modal stereo losses are derived from the geometry constraints between these images, including disparity cross-consistency loss and internal disparity loss.
Figure 5.
Results of different methods on synthetic dataset. Monocular model does not work well as it cannot be fine-tuned under this setting. Stereo matching between intensity and reconstruction fails because the color discrepancy prevents the network to relate corresponding pixels. After self-supervised training, our predictions are on par with the ground truth. The reconstruction network is FireNet and the stereo network is the modified AANet.
Figure 6.
Comparison of different methods on the MVSEC dataset [49]. Notice that the direct event reconstruction quality on real data is far inferior to that of synthetics data, yet our framework can still achieve decent results, which shows its robustness. The reconstruction network is E2VID, and the stereo network is the modified AANet.
3.3
Loss Function
To achieve the goal of self-supervised learning, we define the loss function as the combination of the following four parts:
(1)
L=λgdLgd+λsmLsm+λccLcc+λitnLitn,
where the λs denote the hyper-parameters that control the loss weights. Next, we describe each component of the loss function.
3.3.1
Gradient Structure Consistency Loss
One core guarantee for the success of self-supervised stereo matching is that the output disparities indicate the epipolar geometry relationship between the left and right views. Usually, we can achieve this goal by comparing the projected left image and right images with image similarity metrics, e.g., pixel-wise loss and perceptual loss [43, 44]. However, the quality of the reconstruction result is usually poor, which fails the previous loss functions. Fig. 3 shows an example of reconstruction. As one can observe, the reconstruction network fails to recover any color information. However, its gradient reserves the structure information of the scene. We propose to use image structure loss [45] calculated on image gradient to constrain the stereo training. Let Gl = ∇xyIl be the gradient of Il and Gr = ρxyIr be the adjusted gradient of Ir with scaling factor ρ. The used gradient structure consistency loss is formulated as
(2)
Lgd=12μGlμGr+c1μGl2+μGr2+c1×2σGlGr+c2σGl2+σGr2+c2,
where c1, c2 are constants and μGl, μGr, σGl, σGr, σGlGr represent means, standard deviations and cross-covariance of the gradient pair. In practice, Eq. (2) is calculated on the local patch pairs and then summed up as the final loss.
We compare different losses in Fig. 3. We set the reconstructed image unmoved and shift the intensity images so that they are unaligned to simulate the situation after projection during self-supervised training. We then visualize the distribution of the loss values. We examine commonly used pixel-wise l1-norm loss [39], pixel-wise l1-norm loss on the image gradient, structure loss [39] on the image, and the proposed structure loss on the image gradient. As one can observe, the pixel-wise losses cannot indicate the optimal point. The structure loss calculated on images can indicate the optimal point but has an unsmooth loss landscape. Only the proposed loss has a relatively smooth loss landscape while successfully indicating the optimal point.
Directly calculating photometric consistency between warped images may introduce blurring to the predicted disparity map because there are occlusion areas between left and right view scenes, where warping cannot fill. We introduce the occlusion mask M to mask out these occlusion pixels. We first perform left-right consistency check by projecting the right disparity using the left disparity map and calculate their coherence. The inconsistent region, which is likely to be the occlusion region, is marked as the occlusion mask M, which can be formulated as
(3)
M=0,DlP(Dr;Dl)1<t1,DlP(Dr;Dl)1t,
where P(Dr; Dl) represents projecting the right disparity using the left disparity and t is the threshold parameter.
3.3.2
Disparity Smoothness Loss
After obtaining the disparity maps for both views Dl and Dr, we follow the previous methods for estimating dense flow or disparity [39] and employ an edge-ware smoothness loss. This loss is symmetrical for both left and right views. Thus we omit the superscript. We encourage disparities to be locally smooth with a penalty on the disparity gradients ∇xD and ∇yD. As depth discontinuities often occur at image edges, we weight this cost with an edge-aware term using the image gradients ∇xI and ∇yI, which is formulated as
(4)
Lsm=1Ni,j|xDij|e|xIij|+|yDij|e|yIij|,
where D denoted the disparity map corresponded with I, the subscripts i and j indicates pixel coordinates, N is the total number of pixels.
3.3.3
General Multi-Modal Stereo Losses
Although gradient structure consistency loss can guide stereo training, the provided supervision is sparse and not that specific as pixel-level losses. We cannot obtain accurate disparity only with the above losses. Exploiting the internal stereo relationship between different views and different modalities, we propose general multi-modal stereo losses. An simple illustration of the proposed losses is shown in Figure 4. With the Il and Ir at one hand, we calculate the disparities Dl and Dr that correspond to Il and Ir, respectively. By projecting Il and Ir according to Dl and Dr, we obtain Iwr and Iwl, which represent different views and are in different modalities: Iwl is with the same modality with Ir but with the same view with Il; and Iwr is with the same modality with Il but with the same view with Ir. Using the same multi-modal stereo network, we can obtain Dwl and Dwr – the disparities calculated on two projected images. The proposed loss functions are built based on two facts. According to the fact that the disparity between Iwl and Iwr should be the same disparity between Il and Ir, we build the disparity cross-consistency loss to make them as close as possible:
(5)
Lcc=1Ni,j|Dl||Dwr|+|Dr||Dwl|,
where we take the absolute value for disparities as the projection directions may be opposite, and we only need their shapes. According to another fact that there should be no disparities within the same view, we build the internal disparity loss:
(6)
Litn=1Ni,j|Ditnr|+|Ditnl|,
where Ditnr is the calculated between Ir and Iwr and Ditnl is the calculated between Il and Iwl.
4.
Experiments
4.1
Implementation Details
In this section, we experimentally evaluate the stereo matching performance of the proposed method. We use both synthetic and real data in our experiments. For the experiments based on synthetic data, we employ the Stereo Blur Dataset proposed by Zhou et al. [46], which contains 20,637 blurry-sharp stereo image pairs from 126 diverse sequences and their corresponding bidirectional disparities. To reliably synthesize events, we first increase the sequence frame rate from 60 fps to 2,400 fps via a high-quality frame interpolation algorithm [47] and then applying the V2E event simulator [48] to the high frame rate sequences. We use the officially split method, where 89 sequences are used for self-supervised training, and 37 sequences are used for testing. For real sensor data, we use the MVSEC [49] dataset, which contains the stereo intensity images and events captured by DAVIS 240C. MVSEC also provides ground truth depth captured by LiDAR. We use the official split method for the MVSEC dataset. Our method is implemented using Pytorch [50] framework and trained using NVIDIA V100 GPUs. For the stereo network design, we build our multi-modal networks by modifying DeepPruner [29] and AANet [30] according to Sec. 3.2. Note that our framework is compatible with the most alternative architectures of the reconstruction and the stereo networks. For optimization, we use Adam [51] with β1 = 0.9, β2 = 0.999 and learning rate 1 × 10−4. We set the weighting of the different loss components to λgd = 1, λsm = 0.1, λcc = 0.025 and λitn = 0.005. The settings of ρ and t are experimental. The final value of ρ is 1 on the synthetic dataset and 1.5 on the real world data. The value of t is 2 for all datasets. The overall self-supervised training costs about 2 days.
Figure 7.
The convergence curves of using different losses. The left figure shows that the proposed gradient structure loss can provide meaningful self-supervision. The right figure indicates that the proposed general stereo loss could produce robust training effects. Some methods failed in the middle of training.
4.2
Stereo Matching Results
We first quantitatively demonstrate the effectiveness of the proposed method in Tables I and II. The metrics are averaged end-point error (EPE) and >1-pixel, >3-pixel and >5-pixel error. Under the proposed setting, only limited methods can be used to obtain disparity maps as there is no multi-modal stereo matching model for event-intensity setting. We included the monocular depth model [40] for comparison, which was not fine-tuned on the target data, as we cannot obtain intensity images from both left and right views at the same time in this setting theoretically. We also consider the monocular event depth model [52], but the advantage of the event model is to combat high-speed motion and low-light situations that the intensity camera cannot handle, and its effect cannot be compared with the results predicted using the intensity images. We can have the following observations. Firstly, the proposed method can achieve much better results compared with monocular disparity estimation results using monodepthv2 [40] that is only based on the left image. It indicates that the information of another view plays an essential role in stereo matching. The unsatisfactory effect prevents us from using this disparity map to align events with the intensity image. We then show the results of directly performing stereo matching using the reconstructed right image and left image (marked as the “baseline experiments”). As can be seen, since the reconstructed images have a huge color difference against the left view image, it isn’t easy to obtain a good result by employing the existing stereo vision models. The visualization results in Figure 5 also speak to similar conclusion. However, introducing the modified stereo network and self-supervised learning using only gradient structure loss can improve the stereo estimation results. Even if all the loss functions are used, the network architecture is still a significant obstacle to improving performance. Third, the proposed self-supervised learning method unleashes the full potential of the overall framework’s effect, which proves the advantage of our self-supervised learning strategy, that is, learning from unlabeled data. We visualize the results of the proposed method in Fig. 5 and Figure 6. As one can observe, we can only get poor matching results in these scenarios based on the pre-trained stereo network. Our full model produces accurate object boundaries and better preserves the overall scene structures. We also provide the upper bound performance obtained by a fine-tuned DeepPruner network using both sides’ intensity images as a reference. It can be seen that the information lost by the event is detrimental to the final matching result. But the purpose of our method is not to rely on events to obtain better matching results but to make it possible to calculate reasonable disparity under the proposed intensity-event setting.
Table I.
Quantitative comparison of different approaches on stereo matching using our synthetic stereo event dataset [46]. means the higher the better while means the lower the better. “*” indicates using the modified stereo network.
ModelEPE Bad Pixels
δ > 1δ > 3δ > 5
Monodepth28.8490.9530.7810.648
DeepPruner (upper bound)0.7120.1230.0270.015
FireNet+AANet (baseline)4.8110.6490.4190.336
E2VID+AANet (baseline)5.1540.6730.4400.379
FireNet+DeepPruner (baseline)10.290.4170.2260.181
E2VID+DeepPruner (baseline)6.3860.3810.1840.140
FireNet+AANet* (Lgd and Lsm)1.5910.3660.1390.088
E2VID+AANet* (Lgd and Lsm)1.4960.3510.1230.075
FireNet+DeepPruner* (Lgd and Lsm)1.3360.3550.1230.068
E2VID+DeepPruner* (Lgd and Lsm)1.3210.3550.1160.068
FireNet+AANet (all losses)1.9880.4090.1890.134
E2VID+AANet (all losses)1.7750.3780.1660.117
FireNet+DeepPruner (all losses)1.6260.3770.1470.097
E2VID+DeepPruner (all losses)1.570.3680.1430.094
FireNet+AANet* (all losses)1.2010.3060.1100.065
E2VID+AANet* (all losses)1.1010.2870.0940.057
FireNet+DeepPruner* (all losses)0.9710.3170.0870.049
E2VID+DeepPruner* (all losses)0.9130.2890.0740.042
Table II.
Quantitative comparison of different approaches on stereo matching using real-world dataset MVSEC [49]. means the higher the better while means the lower the better. - means the method completely fails.
ModelEPE Bad Pixels
δ > 1δ > 3δ > 5
Monodepth210.2350.9140.8440.768
E2VID+AANet (baseline)11.3320.9540.8640.776
E2VID+AANet (all losses)5.8300.7360.6600.434
E2VID+DeepPruner (all losses)4.9790.6730.5810.384
E2VID+AANet* (all losses)2.7340.6530.3300.197
E2VID+DeepPruner* (all losses)2.3970.6010.2680.164
4.3
Ablation Study
To study the effects of each component in the proposed method, we conduct several ablation studies. All the experiments are conducted using FireNet reconstruction and a modified AANet stereo model. We first examine the use of gradient structure loss function. We train our model using the four alternative loss functions described in Section 3.3 and their convergence curves are shown in the left figure of Figure 7. As can be seen, all pixel-wise loss functions fail to converge. They can only provide very limited information, and continuous optimization of these losses will bring adverse effects and make training fail. The structure loss directly calculated on images performs well initially, but it could no longer provide adequate supervision as the training progressed. The proposed gradient structural loss can produce a relatively reliable convergence curve, although it is still tricky to improve performance with training steadily. We next involve the proposed cross-consistency loss and internal disparity loss in training. The results are shown in right figure of Fig. 7. It is surprising that although it cannot be compared with competitors initially, the proposed losses make the training process more stable. The proposed method can steadily and continuously improve performance without any ground truth data. We visually explain why the proposed cross-consistency loss is so effective. Figure 8 shows some comparisons between models trained with and without the cross-consistency loss. The results speak to the fact that the inclusion of this loss improves the quality of the result. This loss term helps outline the precise edge and generate sharp, accurate shapes in the disparity map.
Figure 8.
Comparison between our method with and without the cross-consistency loss. Our consistency term helps outline the precise edge in the disparity map.
4.4
Connecting Events and Intensity
We next demonstrate the use of a disparity map to establish a connection between the intensity camera and the event camera. Due to the displacement between these two cameras, even if we can obtain the left view intensity image and the right view events, the disparity between them makes it difficult to make full use of the advantages of the two sensors, as shown in Figure 9(b). In this case, many algorithms and applications that require alignment of events and images cannot be implemented, e.g., [11, 53]. We first obtain the disparity map using the proposed multi-modal stereo method. Each value in the disparity map indicates the number of pixels that need to be shifted horizontally. We warp events by changing the x coordinate in each event tuple (xm, ym, tm, pm), where xm, ym, tm denote the spatial-temporal coordinates, and pm ∈{− 1, +1} denotes the polarity of the event. The warped event are visualized in Fig. 9(c). It can be seen that the warped right view events and the left view image are well aligned both spatially and temporally. Obtaining such a connection between two sensors allows many downstream tasks. Here we show the application potential by the event-based video interpolation task. Motivated by the physical model of event that the residuals between a blurry image and sharp frames are the integrals of events, Lin et al. [11] propose to estimate the residuals for the sharp frame restoration based on events. Our reconstruction result shows good fidelity performance, which further proves the application value of the proposed problem setting.
4.5
Limitations
Finally, we show some failure cases and analyze the potential limitations. We show two failure cases in Figure 10 and both of them are from MVSEC. The direct reconstruction using E2VID contains only very limited information, resulting in the failure of stereo matching. This shows a possible flaw of the proposed method, that is, it still faces significant challenges when reconstruction results are very vague. A possible solution is to design a stereo network to perform stereo matching between event streams and intensity images directly. In that case, the loss functions described in this paper can still provide good self-supervised learning results.
Figure 9.
Connecting events and intensity. a, the calculated disparity map using the proposed method. b, before warping, events and intensity image are not aligned spatially. c, after warping using disparity map, the events and intensity image are well aligned. d, we employ the warped events and intensity image to perform temporal frame interpolation using [11].
Figure 10.
Failure cases visualization. The proposed method faces challenges when the reconstruction only contains very limited information.
5.
Conclusion
This paper presents a novel camera setting with an intensity camera and an event camera and establishes a connection between them with a multi-modal stereo matching task. Based on the proposed self-supervised method, we can obtain fine disparity maps under this novel setting and not collect any ground truth disparities. Experiments demonstrate the effectiveness and the application value of the proposed method.
Appendix A.
More Stereo Matching Results
We first provide more stereo matching results using different methods. In Figure C4, we provide the comparison of synthetic events data. The synthesis method is described in Sec 4.1. In Figure C5, we provide the comparison results on the MVSEC [49] dataset. In these experiments, we use E2VID as the reconstruction network and AANet [30] as the stereo matching network. It can be seen that our method provides the best results and performs well in these cases. In Fig. C5, we also compare our method with another commonly used method when encountering multi-modal problems, e.g., Zhi et al. [32] proposed to use a spectral translation network to facilitate cross-spectral stereo matching. We develop a similar adaptation network to translate the rough reconstructed right view image to add color information. The network structure is ResNet image translation structure, similar to SRResNet [54] but without an upsampling layer. We use a supervised training method to train this network, and the ground truth of the intensity image is provided in MVSEC. As can be seen, simply using the modality alignment method cannot bring better results. There are two main reasons for this. First, the color information has been lost in the events and reconstruction results and cannot be simply recovered by an adaptation network. Second, the pre-trained stereo model cannot generalize well in the MVSEC dataset. This also provides the necessities of learning using target data in a self-supervised manner.
Appendix B.
Event Warping and Video Frame Interpolation
In this section, we show more results of warping events to intensity image. We first obtain the disparity between these two sensors using the proposed method. The warping provides us with events and intensity images that are aligned both spatially and temporally. We can use the obtained events and images to support downstream applications. In this supplementary material, we show more video frame interpolation [11] results in Figure C6.
The proposed method has additional value in this respect. Due to the hardware limitations of the event camera, high-resolution events and high-resolution intensity images are not available simultaneously. But with the proposed method, we can obtain high-resolution events and images simultaneously through two sensors. This makes a series of applications, such as video frame interpolation possible.
Appendix C.
Effect of the Loss functions
We provide more results for the proposed general multi-modal stereo loss functions. We first show the convergence curves with different loss functions and metrics in Figure C1. It can be seen that the proposed loss functions enable the model to improve its performance through self-supervised learning continuously. The model without the proposed losses only provides a good guide initially, but when optimizing continuously, it does not match the purpose we want to achieve. Since the gradient structure loss will be numerically unstable when the difference between the two images is large, some methods will fail halfway. Although the internal disparity loss only provides simple, naive supervision, it has successfully improved the performance. We will further understand these two loss functions through visualization results. Figure C2 and Figure C3 show some error map visualization results. One can first observe from Fig. C3 that the proposed cross-consistency loss helps outline the edge and shape of the disparity. The proposed cross-consistency loss promotes the consistency of shapes between different views and provides additional information for training. We can also see from the cross-consistency error maps and internal disparity error maps that the introduction of these losses reduces the degree of these inconsistencies, especially for the internal disparity loss. Fig. C2 shows a failure case and also show how the internal disparity loss works. In this case, the disparity of the front object exceeds the upper limit of the network (we set the max disparity to be 41 pixels). The internal disparity loss reveals the failurearea.
Figure C1.
The convergence curves of using different losses. The proposed general stereo loss could produce robust training effects. The method of “no general stereo losses” failed in the middle of training.
Figure C2.
Visualization results of the proposed loss functions. This is a failure case. The disparity of the front object exceeds the upper limit of the network (in this case, the upper limit is 41 pixels). One can see that the proposed internal disparity loss points out where the error occurred.
Figure C3.
Visualization results of the proposed loss functions.
Figure C4.
Stereo matching results of different methods on synthetic dataset.
Figure C5.
Stereo matching comparison of different methods on the MVSEC [49] real-world dataset.
Figure C6.
Connecting events and intensity. The event warping results and the temporal frame interpolation results using the warped events and intensity images using [11].
Appendix D.
Alternative Framework
We also present an alternative framework where the stereo network takes the right event voxel and the left intensity image as input directly. The framework is shown in Figure D1. In this framework, the network can also be the convolutional multi-modal stereo network. However, using convolution to process event voxel directly tends to bring poor results. We did not use this alternative since existing convolutional networks would be significantly better at processing images. But this alternative shows that our general multi-modal stereo consistency loss can be generalized to a wider range of application scenarios.
Figure D1.
An alternative framework to the proposed self-supervised learning method. The stereo network takes the right event voxel and the left intensity image as input directly.
References
1RebecqH.HorstschäferT.GallegoG.ScaramuzzaD.2016Evo: A geometric approach to event-based 6-dof parallel tracking and mapping in real timeIEEE Robot. Autom. Lett.2593600593–60010.1109/LRA.2016.2645143
2JiangZ.ZhangY.ZouD.RenJ.LvJ.LiuY.Learning event-based motion deblurringCVPR2020IEEEPiscataway, NJ332033293320–910.1109/CVPR42600.2020.00338
3ZhuA. Z.YuanL.2018Ev-flownet: Self-supervised optical flow estimation for event-based camerasRobotics: Science and Systems10.15607/RSS.2018.XIV.062
4VidalA. R.RebecqH.HorstschaeferT.ScaramuzzaD.2018Ultimate slam? combining events, images, and imu for robust visual slam in hdr and high-speed scenariosIEEE Roboti. Autom. Lett.39941001994–100110.1109/LRA.2018.2793357
5TulyakovS.GehrigD.GeorgoulisS.ErbachJ.GehrigM.LiY.ScaramuzzaD.Time lens: Event-based video frame interpolationCVPR2021IEEEPiscataway, NJ161551616416155–6410.1109/CVPR46437.2021.01589
6PatrickL.PoschC.DelbruckT.2008A 128 × 128 120 db 15 μs latency asynchronous temporal contrast vision sensorIEEE J. Solid-State Circuits43566576566–7610.1109/JSSC.2007.914337
7BrandliC.BernerR.YangM.LiuS.-C.DelbruckT.2014A 240×180 130 db 3 μs latency global shutter spatiotemporal vision sensorIEEE J. Solid-State Circuits49233323412333–4110.1109/JSSC.2014.2342715
8RebecqH.RanftlR.KoltunV.ScaramuzzaD.Events-to-video: Bringing modern computer vision to event camerasCVPR2019IEEEPiscataway, NJ385738663857–6610.1109/CVPR.2019.00398
9ScheerlinckC.RebecqH.GehrigD.BarnesN.MahonyR.ScaramuzzaD.Fast image reconstruction with an event cameraWACV2020IEEEPiscataway, NJ156163156–6310.1109/WACV45572.2020.9093366
10GehrigD.RebecqH.GallegoG.ScaramuzzaD.Asynchronous, photometric feature tracking using events and framesECCV2018SpringerCham750765750–6510.1007/978-3-030-01258-8_46
11LinS.ZhangJ.PanJ.JiangZ.ZouD.WangY.ChenJ.RenJ.Learning event-driven video deblurring and interpolationECCV2020SpringerCham695710695–71010.1007/978-3-030-58598-3_41
12ZhuA. Z.YuanL.ChaneyK.DaniilidisK.Unsupervised event-based learning of optical flow, depth, and egomotionCVPR2019IEEEPiscataway, NJ10.1109/CVPR.2019.00108
13KimH.LeuteneggerS.DavisonA. J.Real-time 3d reconstruction and 6-dof tracking with an event cameraECCV2016SpringerCham349364349–6410.1007/978-3-319-46466-4_21
14KuengB.MuegglerE.GallegoG.ScaramuzzaD.Low-latency visual odometry using event-based feature tracksIROS2016IEEEPiscataway, NJ162316–2310.1109/IROS.2016.7758089
15BaruaS.MiyataniY.VeeraraghavanA.Direct face detection and video reconstruction from event camerasWACV2016IEEEPiscataway, NJ10.1109/WACV.2016.7477561
16BardowP.DavisonA. J.LeuteneggerS.Simultaneous optical flow and intensity estimation from an event cameraCVPR2016IEEEPiscataway, NJ10.1109/CVPR.2016.102
17WangL.Mohammad MostafaviI. S.HoY.-S.YoonK.-J.Event-based high dynamic range image and very high frame rate video generation using conditional generative adversarial networksCVPR2019IEEEPiscataway, NJ10.1109/CVPR.2019.01032
18Mohammad MostafaviI. S.ChoiJ.YoonK.-J.Learning to super resolve intensity images from eventsCVPR2020IEEEPiscataway, NJ10.1109/CVPR42600.2020.00284
19WangL.KimT.-K.YoonK.-J.EventSR: From asynchronous events to image reconstruction, restoration, and super-resolution via end-to-end adversarial learningCVPR2020IEEEPiscataway, NJ10.1109/CVPR42600.2020.00834
20StoffregenT.ScheerlinckC.ScaramuzzaD.DrummondT.BarnesN.KleemanL.MahonyR.Reducing the sim-to-real gap for event camerasECCV2020SpringerCham10.1007/978-3-030-58583-9_32
21YoonK.-J.KweonI. S.Adaptive support-weight approach for correspondence searchIEEE Transactions on Pattern Analysis and Machine Intelligence2006Vol. 28IEEEPiscataway, NJ650656650–610.1109/TPAMI.2006.70
22HosniA.RhemannC.BleyerM.RotherC.GelautzM.Fast cost-volume filtering for visual correspondence and beyondIEEE TPAMI2013IEEEPiscataway, NJ10.1109/TPAMI.2012.156
23ZbontarJ.LeCunY.Computing the stereo matching cost with a convolutional neural networkCVPR2015IEEEPiscataway, NJ10.1109/CVPR.2015.7298767
24LuoW.SchwingA. G.UrtasunR.Efficient deep learning for stereo matchingCVPR2016IEEEPiscataway, NJ569557035695–70310.1109/CVPR.2016.614
25MayerN.IlgE.HausserP.FischerP.CremersD.DosovitskiyA.BroxT.A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimationCVPR2016IEEEPiscataway, NJ404040484040–810.1109/CVPR.2016.438
26KendallA.MartirosyanH.DasguptaS.HenryP.KennedyR.BachrachA.BryA.End-to-end learning of geometry and context for deep stereo regressionICCV2017IEEEPiscataway, NJ10.1109/ICCV.2017.17
27ChangJ.-R.ChenY.-S.Pyramid stereo matching networkCVPR2018IEEEPiscataway, NJ10.1109/CVPR.2018.00567
28ZhangF.PrisacariuV.YangR.TorrP. H. S.Ga-net: Guided aggregation net for end-to-end stereo matchingCVPR2019IEEEPiscataway, NJ10.1109/CVPR.2019.00027
29DuggalS.WangS.MaW.-C.HuR.UrtasunR.DeepPruner: Learning efficient stereo matching via differentiable patchmatchICCV2019IEEEPiscataway, NJ438443934384–9310.1109/ICCV.2019.00448
30XuH.ZhangJ.AANET: Adaptive aggregation network for efficient stereo matchingCVPR2020IEEEPiscataway, NJ195919681959–6810.1109/CVPR42600.2020.00203
31ChiuW. W.-C.BlankeU.FritzM.Improving the kinect by cross-modal stereoBMVC2011Dundee116110116–0
32ZhiT.PiresB. R.HebertM.NarasimhanS. G.Deep material-aware cross-spectral stereo matchingProc. IEEE Conf. on Computer Vision and Pattern Recognition2018IEEEPiscataway, NJ191619251916–2510.1109/CVPR.2018.00205
33ShenX.XuL.ZhangQ.JiaJ.Multi-modal and multi-spectral registration for natural imagesECCV2014SpringerCham309324309–2410.1007/978-3-319-10593-2_21
34JeonH.-G.LeeJ.-Y.ImS.HaH.KweonI. S.Stereo matching with color and monochrome cameras in low-light conditionsCVPR2016IEEEPiscataway, NJ408640944086–9410.1109/CVPR.2016.443
35KimS.MinD.HamB.RyuS.DoM. N.SohnK.DASC: Dense adaptive self-correlation descriptor for multi-modal and multi-spectral correspondenceCVPR2015IEEEPiscataway, NJ210321122103–1210.1109/CVPR.2015.7298822
36KimS.MinD.LinS.SohnK.Deep self-correlation descriptor for dense cross-modal correspondenceEuropean Conf. on Computer Vision2016SpringerCham679695679–9510.1007/978-3-319-46484-8_41
37Mohammad MostafaviI. S.YoonK.- J.ChoiJ.Event-intensity stereo: Estimating depth by the best of both worldsICCV2021IEEEPiscataway, NJ425842674258–6710.1109/ICCV48922.2021.00422
38GargR.KumarB. V.CarneiroG.ReidI.Unsupervised CNN for single view depth estimation: Geometry to the rescueECCV2016SpringerCham740756740–5610.1007/978-3-319-46484-8_45
39GodardC.AodhaO. M.J. BrostowG.Unsupervised monocular depth estimation with left-right consistencyCVPR2017IEEEPiscataway, NJ10.1109/CVPR.2017.699
40GodardC.AodhaO. M.FirmanM.J. BrostowG.Digging into self-supervised monocular depth estimationICCV2019IEEEPiscataway, NJ382838383828–3810.1109/ICCV.2019.00393
41ZhouC.ZhangH.ShenX.JiaJ.Unsupervised learning of stereo matchingICCV2017IEEEPiscataway, NJ156715751567–7510.1109/ICCV.2017.174
42RebecqH.RanftlR.KoltunV.ScaramuzzaD.High speed and high dynamic range video with an event cameraIEEE Transactions on Pattern Analysis and Machine Intelligence2021Vol. 43IEEEPiscataway, NJ196419801964–8010.1109/TPAMI.2019.2963386
43ZhangR.IsolaP.EfrosA. A.ShechtmanE.WangO.The unreasonable effectiveness of deep features as a perceptual metricCVPR2018IEEEPiscataway, NJ586595586–9510.1109/CVPR.2018.00068
44JinjinG.HaomingC.HaoyuC.XiaoxingY.RenJ. S.ChaoD.Pipal: a large-scale image quality assessment dataset for perceptual image restorationECCV2020SpringerCham633651633–5110.1007/978-3-030-58621-8_37
45WangZ.BovikA. C.SheikhH. R.SimoncelliE. P.2004Image quality assessment: from error visibility to structural similarityIEEE Trans. Image Process.13600612600–1210.1109/TIP.2003.819861
46ZhouS.ZhangJ.ZuoW.XieH.PanJ.RenJ. S.DAVANet: Stereo deblurring with view aggregationCVPR2019IEEEPiscataway, NJ109961100510996–100510.1109/CVPR.2019.01125
47JiangH.SunD.JampaniV.YangM.-H.Learned-MillerE.KautzJ.Super SloMo: High quality estimation of multiple intermediate frames for video interpolationCVPR2018IEEEPiscataway, NJ900090089000–810.1109/CVPR.2018.00938
48DelbruckT.HuY.HeZ.V2E: From video frames to realistic DVS event camera streamsCVPR2021IEEEPiscataway, NJ
49ZhuA. Z.ThakurD.ÖzaslanT.PfrommerB.KumarV.DaniilidisK.2018The multivehicle stereo event camera dataset: An event camera dataset for 3d perceptionIEEE Robot. Autom. Lett.3203220392032–910.1109/LRA.2018.2800793
50PaszkeA.GrossS.MassaF.LererA.BradburyJ.ChananG.KilleenT.LinZ.GimelsheinN.AntigaL.DesmaisonA.Pytorch: An imperative style, high-performance deep learning libraryNeurIPS2019Curran AssociatesRed Hook, NY802680378026–37
51KingmaD. P.BaJ.Adam: A method for stochastic optimizationICLR2015ICLRNew Orleans, LA
52GehrigD.RüeggM.GehrigM.Hidalgo-CarrióJ.ScaramuzzaD.2021Combining events and frames using recurrent asynchronous multimodal networks for monocular depth predictionIEEE Robot. Autom. Lett.6282228292822–910.1109/LRA.2021.3060707
53DuanP.WangZ. W.ShiB.CossairtO.HuangT.KatsaggelosA.2021Guided event filtering: Synergy between intensity images and neuromorphic events for high performance imagingIEEE Trans. Pattern Anal. Mach. Intell.44826182758261–7510.1109/TPAMI.2021.3113344
54WangX.YuK.WuS.GuJ.LiuY.DongC.QiaoY.LoyC. C.ESRGAN: Enhanced super-resolution generative adversarial networksECCV2018SpringerCham637963–7910.1007/978-3-030-11021-5_5