# hrdepth_high_resolution_selfsupervised_monocular_depth_estimation__f57121bf.pdf HR-Depth: High Resolution Self-Supervised Monocular Depth Estimation Xiaoyang Lyu 1, Liang Liu 1, Mengmeng Wang 1, Xin Kong 1, Lina Liu 1, Yong Liu* 1, Xinxin Chen 1, Yi Yuan 2 1 Zhejiang University 2 Fuxi AI Lab, Net Ease {shawlyu, leonliuz, mengmengwang, xinkong, linaliu}@zju.edu.cn, yongliu@iipc.zju.edu.cn, cxx-compiler@zju.edu.cn, yuanyi@corp.netease.com Self-supervised learning shows great potential in monocular depth estimation, using image sequences as the only source of supervision. Although people try to use high-resolution image for depth estimation, the accuracy of prediction has not been significantly improved. In this work, we find the core reason comes from the inaccurate depth estimation in large gradient regions, making the bilinear interpolation error gradually disappear as the resolution increases. To obtain more accurate depth estimation in large gradient regions, it is necessary to obtain high-resolution features with spatial and semantic information. Therefore, we present an improved Depth Net, HR-Depth, with two effective strategies: (1) redesign the skip-connection in Depth Net to get better highresolution features and (2) propose feature fusion Squeezeand-Excitation(f SE) module to fuse feature more efficiently. Using Resnet-18 as the encoder, HR-Depth surpasses all previous state-of-the-art(So TA) methods with the least parameters at both high and low resolution. Moreover, previous So TA methods are based on fairly complex and deep networks with many parameters which limits their real applications. Thus we also construct a lightweight network which uses Mobile Net V3 as encoder. Experiments show that the lightweight network can perform on par with many large models like Monodepth2 at high-resolution with only 20% parameters. All codes and models will be available at https: //github.com/shaw Lyu/HR-Depth. Introduction Accurate depth estimation from single image is an active research filed to help computer reconstruct and understand the real scenes. It also has a large range of applications in diverse fields such as autonomous vehicles, robotics, augmented reality, etc. While supervised monocular depth estimation has been successful, it is suffering from the expensive access to ground truth. Self-supervised methods use geometrical constraints on image sequences or stereo images as the sole source of supervision. Recent works (Zhou et al. 2017; Godard et al. 2017) in self-supervised depth estimation are limited to training in low-resolution inputs due to the large memory requirements *Corresponding author Copyright 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Depth prediction from single image on KITTI. We compared our method with Monodepth2(Godard et al. 2019), our method can predict higher quality and sharper depth map with fewer parameters. The input resolution for all three models is 1024 320. of the model. However, with the improvement of computing and storage capacity, high-resolution images are used by more and more computer vision tasks. In depth estimation, (Pillai et al. 2019) introduce sub-pixel-convolutional (Shi et al. 2016) layers replacing the deconvolution and resizeconvolution layers to improve the effect of up-sampling. And (Godard et al. 2019) directly leverage high-resolution images for depth estimation. Although the above works explore high resolution depth estimation, but the performance has not improved significantly on KITTI. Inspired by this observation, we perform an analysis of existing method and find the core reason comes from the inaccurate depth estimation at object boundaries. In depth estimation, object boundaries are mainly determined by two parts: semantic and spatial information. Semantic information obtains clear boundaries by constraining the categories of pixels, while spatial information uses geometric constraints to describe the outline of objects. In the previous work, the depth estimation network is based on the U-Net(Ronneberger et al. 2015) architecture, which mainly uses skip-connection to fuse semantic information and spatial information. However, the semantic gap between the encoder and decoder feature maps is too large, which leads to a poor integration of spatial and semantic information. So the previous work is difficult to get an accurate depth estimation at object boundaries. In order to reduce the semantic gap, we redesigned the skip connection to better fuse feature maps. Besides, we also found that the basic convolution can not integrate spatial information and semantic information The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) well, so we propose to replace it by a feature fusion squeeze and excitation (f SE) block. This block not only improves the feature fusion effect but also reduces the number of parameters. We evaluate our results on KITTI benchmark, and the experiments demonstrate that the our well-designed network can predict sharper edges and achieve (So TA). As a side effect, higher resolution input brings extra computational costs. As a consequence, lightweight is one of the key principals for high-resolution model design. However, the previous So TA has a huge amount of parameters like Packnet-Sf M with 127M parameters, and the performance of lightweight networks is not appropriate for actual application like (Poggi et al. 2018). Therefore, in order to maintain high performance of lightweight network, we introduce a simple yet effective designing strategy in this paper. We successfully train a lightweight network based on this strategy, which can achieve the accuracy of Monodepth2 with only 3.1M parameters. To summarize, the main contributions of this work are listed below in fourfold: We provide a deep analysis of high-resolution monocular depth estimation and prove that predicting more accurate boundaries can improve performance. We redesign the skip connection to get high-resolution semantic feature maps, which can help network predict sharper edges. We propose feature fusion squeeze and excitation block to improve the efficiency and effect of feature fusion. We present a simple yet effective lightweight design strategy to train a lightweight depth estimation network that can achieve the performance of complex network. Related Work Supervised Monocular Depth Estimation Depth estimation from a single image is an ill-posed problem as the same input image can be projected to multiple plausible depths. Therefore, many works begin with supervised learning. (Eigen et al. 2015) is the first to propose learning-based depth estimation structure trained on RGBD dataset. Their work considers depth estimation as a regression problem, using a coarse-to-fine network to derive pixel-by-pixel depth value. With the rise of fully convolution neural network, (Laina et al. 2016) uses convolution layer to replace fully connected layer, and uses pre-trained encoder for feature extraction. This work enables the depth estimation task to be trained with a deeper network and has the accuracy comparable to the depth sensor. In order to predict sharp and accurate occlusion boundaries, (Ramamonjisoa et al. 2020) introduces refine net predicting an additive residual depth map to refine the first estimation results. However, supervised methods fall into bottleneck on account of the poor generalization performance and the difficulty of obtaining ground truth depth value. So people began to explore self-supervised monocular depth estimation. Self-Supervised Monocular Depth Estimation Using stereo images to train depth network is one intuitive of self-supervision. (Garg et al. 2016) proposes one of the earliest work in self-supervised depth estimation using stereo pairs and (Godard et al. 2017) produces results superior to contemporary supervised method by introducing a left-right depth consistency loss. In order to reduce the limitation of stereo camera, (Zhou et al. 2017) firstly proposes to use Pose CNN to estimate relative pose between adjacent frames. This work enables the network to be trained completely depending on monocular image sequences. (Godard et al. 2019) introduces auto-masking and min re-projection loss to solve the problems of moving objects and occlusion and makes Monodepth2 become the most widely used baseline. For the sake of further improving the network performance, (Guizilini et al. 2020b) introduces the pre-trained semantic segmentation network and pixel-adaptive convolution to guide depth network to further utilize semantic information. But there are two disadvantages to (Guizilini et al. 2020b). First of all, we hope to relieve the pressure of pixel level annotation by self-supervising, but we need expensive semantic label in this work. Secondly, semantic segmentation and depth estimation network should run simultaneously. It will increase the cost of depth estimation. In addition, (Guizilini et al. 2020a) utilizes packing and unpacking block to preserve the spatial information in image and low level feature. They thought standard convolutional and pooling operation can not preserve sufficient details, so they proposed 3D packing and unpacking blocks to replace standard down-sample and up-sample operation. Packing and unpacking block are invertible, so they can recover important spatial information for depth estimation. But these two blocks depend on 3D convolution, so the number of network parameters is greatly increased and it is difficult to deploy to mobile devices. But these two works show that abundant semantic and spatial information can get sharper edges, so as to improve the accuracy of depth estimation. In our work, we show that, by simply fusing the information extracted by the encoder, we can obtain the ideal features with spatial and semantic information. Through our insightful design, without introducing much more parameters, these features significantly improve the overall performance, obtaining sharper edges. Lightweight Network for Depth Estimation Depth estimation from a single image is also a very attractive technique with several implications in robotic, autonomous navigation. Therefore, it is necessary to leverage depth prediction network to quickly infer an accurate depth map on a CPU. (Wofk et al. 2019) proposed a lightweight supervised depth estimation network, who use encoder-decoder architecture and include 1.34M parameters after pruning. In unsupervised filed, (Poggi et al. 2018; Aleotti et al. 2020) proposed Py DNet with 1.9M parameters. Although the above two works have less parameters, their performance also decreases a lot. In this paper, we propose a novel yet simple strategy for network designing, which can make the performance of lightweight network comparable to or even surpass the complex network. Figure 2: Analysis of High Resolution Depth Estimation. Abs Rel is an evaluation index in depth estimation, and lower is better. HR means high resolution and LR means low resolution. All interpolation results are caculated by Open CV library. HR-Depth Network Problem Formulation In self-supervised monocular depth estimation task, the goal is to use depth network f D to learn depth information D from RGB image I. Due to the lack of ground truth depth value, we need an additional network f P to predict relative pose p = [R|t] between source image Is and target image It. As a common setting, the target image is It, and the source images set are consist of adjacent images It 1, It+1. The depth network is optimized by minimizing the photometric re-projection error: r(It, It ) = α 2 (1 SSIM(It, It ))+(1 α)||It It ||1 (1) where It is the warped result from Is to It, SSIM is the operator of structural similarity to measure the patch similarity. We follow the form of per pixel minimum loss in (Godard et al. 2019) to handle occlusion. The photometric loss is denoted as Lre = min t r(It, It ). (2) Furthermore, in order to regularize the disparities in textureless low-image gradient regions, edges aware smooth regularization term is used: Lsmooth = |δx Dt|e |δx It| + |δy Dt|e |δy It|. (3) where δx and δy symbols for partial differentiation of depth and RGB images, and the exponential operation of a matrix is an element-wise operation. So the final loss function is the summation of the re-projected losses and the smooth loss on multi scale images: i (Li re + λLi smooth) (4) where s is the number of scales, and λ is the weight for smooth term. Resolution Abs rel δ < 1.25 Monodepth2 640 192 0.115 0.877 Monodepth2 1024 320 0.115 0.879 Table 1: Quantitative results from Monodepth2 with different resolution setting training and testing. Analysis on High Resolution Performance As a consensus of dense prediction tasks, higher resolution brings more accurate results instinctively (Sun et al. 2019; Zhou et al. 2019). Especially in the depth estimation task, pixel level disparity is more important since it is inversely proportional to the square of depth error(You et al. 2019). However, we noticed that most of the previous work use low-resolution inputs and interpolate low resolution prediction to high-resolution one. The low-resolution experimental setup essentially makes these works cannot benefit from high-resolution images. Some methods also conduct highresolution experiments where they train their models with larger images (Godard et al. 2019; Pillai et al. 2019). However, the performance improvements upon the small inputs are very limited, i.e. (Pillai et al. 2019), their models cannot take advantage of high resolution. For instance, we evaluate a recent well-known work named Monodepth2(Godard et al. 2019) with higher resolution setting. As shown in Table 1, the depth errors are almost the same with high and lowresolution settings. Therefore, we argue that their method can not make full use of the information from higher input resolution. Furthermore, we make a deep analysis on the actual reason that existing methods cannot improve the depth estimation by high-resolution inputs. We find out that the most vital point is un-negligible error from the bilinear interpolation process for up-sampling the low resolution prediction to high ones. In detail, taking Monodepth2 as an example, when feeding a model with small inputs, we need to up- Figure 3: Illustration of our proposed framework. The network is mainly composed of three different types of nodes. Xe i denotes a feature extraction node, which is mainly composed of residual blocks. Xi,j denotes a feature fusion node which only has 3 3 convolution operation. Xd i is feature fusion node which is mainly composed of our proposed f SE module. Disparity is decoded by Disp Conv block, which contains 3 3 convolution and sigmoid activation function. sample the outputs to obtain a required high-resolution one with interpolation. As shown in Figure 2, it can be seen that local prediction error will severely harm global accuracy in regions with large depth gradient such as instance edges, while for the region with small depth gradient. Interestingly, when the prediction of low-resolution outputs is poor like the upper right part in Figure 2, it will unexpectedly compensate the local prediction error of bilinear interpolation. The results show that the performance of low-resolution can be on par of the high one. In other words, the actual reason that most methods cannot benefit from larger inputs is that the gap between these two settings is remedied by the above interesting phenomenon. Therefore, only more precise prediction of regions with large depth gradient can make highresolution prediction more accurate. We can also summarize that the performance of high-resolution can be improved by predicting more accurate depth in large gradient regions and sharper edges. Redesign Skip Connection Based on the above analyses, we try to enhance them from both the spatial and semantic features to predict more accurate boundaries. We believe (1) semantic information can produce boundaries between different categories, which can reduce depth estimation error caused by misclassification, (2) spatial information can help the network to know the location of the boundaries to estimate them better. Here, we will first discuss Depth Net with U-Net architecture. Skip-connection is one of core components of U-Net, whose purpose is to recover information lost in downsampling. However, we argue that it could be less effective to directly combine features from different layers since there are gaps among them in semantic levels and spatial resolution. As previous research(Zhang et al. 2018) indicates, deep neural networks represent more semantic features as the layers going deeper. Thus, if low-level features could include more semantic information, like relatively clearer semantic boundaries, then the fusion becomes easier and we can obtain sharper depth estimation by decoding these features. In order to decrease semantic and resolution gap, we propose dense skip connection inspired by (Zhou et al. 2018). As shown in Figure 3, in addition to the nodes in the original encoder and decoder, we also add a lot of intermediate nodes to aggregate features. Let xe i denotes the output of encoder node Xe i , xd i denotes the output of node Xd i , xi,j denotes the output of node Xi,j where i indexes the downsampling layer along the encoder and j indexes the aggregation layer along the skip connection. Consider a single image I is passed through Depth Net. The stack of feature maps is computed as xe i = E(I), i = 1 E(xe i 1), i > 1 (5) F( h xe i, [xi,k]j 1 0 , U(xi+1,j 1) i ), j = 1 F( h xe i, [xi,k]j 1 0 , U(xe i+1) i ), j > 1 ( D(U(xd i+1)]), j = 0 D( h xe i, [xi,k]j 1 0 , U(xd i+1) i ), j > 0 (7) where E( ) is a feature extraction block like residual block, F( ) is a feature fusion block consist of convolution operation followed by an activation function, D( ) is a feature fusion operation composed by feature fusion block, U( ) is an upsampling block with convolution and bilinear interpolation operation, and [ ] denotes the concatenation layer. The details of dense skip-connection and each node are shown in Figure 3. With dense skip connection, each node in the decoder is presented with the final aggregated feature maps, the intermediate aggregated feature maps and the original feature maps from the encoder. Then, the decoder can use highresolution features with richer semantic information to predict more sharper depth maps. Feature Fusion SE Block The U-Net based Depth Net uses a 3 3 convolution to fuse upsampling feature maps and original feature maps from the encoder, and the parameters of convolution is calculated as Cin Cout k2 + Cout, (8) where Cin is the number of input channels, Cout is the number of output channels and k denotes as convolution kernel size. Just like (Huang et al. 2017), the dense skipconnections make the decoder nodes have dramatically increased input feature maps, so this operation will inevitably decrease the efficiency of the network. Inspired by (Hu et al. 2019), we propose a lightweight module, feature fusion Squeeze-Excitation(f SE), to improve feature fusion accuracy and efficiency. The f SE module squeezes the feature maps by global average pooling to represent channel information and uses two fully-connected(FC) layers followed by a sigmoid function to measure the importance of each feature and re-weight them in the meantime. Then a 1 1 convolution is used to fuse channels to obtain high-quality feature maps. The parameter amount of this module is 2 r C2 in + (Cin + 1) Cout, (9) where r denotes as reduction ratio and is always set to 4 in all experiments of this paper. When using Resnet-18 as encoder, the f SE module can reduce the parameters of HRDepth from 16.06M to 14.62M, even less than Monodepth2 with 14.84M parameters. Meanwhile, since the f SE module will focus more on feature fusion, the network performance will also be improved. Lite-HR-Depth Most previous So TA self-supervised monocular depth estimation algorithms are based on fairly complex deep neural networks with a mass of parameters which limit their applications on practical platforms like embedded devices. However, for existing lightweight networks, such as Py DNet(Poggi et al. 2018), the accuracy is greatly decreased when they reduce the amount of parameters. In order to make the depth estimation network get rid of the limitation of GPU but keep its great performance, we propose a simple yet effective lightweight network, named Lite-HR-Depth. Our Lite-HR-Depth employs Mobile Net V3 as an encoder and downsizes the feature fusion and decoder nodes with only 3.1M parameters, of which the 2.82M parameters comes from the encoder. In addition, we further improve the accuracy of our Lite-HR-Depth with knowledge distillation(Hinton et al. 2015). For self-supervised learning, due to lack of ground truth, we have to use view syntheses as the supervisory signal, which increases the difficultly of training small network. Therefore, we propose to source a direct form of supervision from the learned large model. By training a large network in a self-supervised manner, we obtain a high-performance network instance T . Then we train a second instance of lightweight model, namely S, to minimize Lsup = ||d T d S||, (10) where d T means disparity from network T and d S means disparity from network S. With this strategy, we obtain a network S which even more accurate than T . It is worth noting that the parameters of Lite-HR-Depth is only 20% of Monodepth2, but it can even perform better than Monodepth2 at high resolution. More analyses of this lite network will be shown in our experiment. Experiments In this section, we validate that (1) our redesigned skipconnection can improve the results, especially predicting sharper edges, (2) the f SE module can significantly reduce parameters and improve accuracy, and (3) the design method we propose can easily obtain high precision lightweight network. We evaluate our models on the KITTI dataset(Geiger et al. 2012), to allow comparison with previous published monocular methods. Datasets KITTI. The KITTI benchmark (Geiger et al. 2012) is most widely used for depth evaluation. We adopt the data split of (Eigen et al. 2015), and removed the static frames followed by (Zhou et al. 2017). Ultimately, we used 39810 images for training, 4424 for validation and 697 for evaluation. Furthermore, we use the same intrinsic for all images, setting the principal point of the camera to the image center and the focal length to the average of all the focal lengths in KITTI. For stereo training, we set the transformation between the two frames to be a pure horizontal translation of fixed length. City Scapes. City Scape (Cordts et al. 2016) is another large automatic driving dataset. So we also experiment with pretraining our structure on City Scape and then we finetuned and evaluated it on KITTI dataset. The left Img8bit Sequence were considered as training split for the City Scapes dataset, using the same training parameters as KITTI for 20 epochs. Implementation Details We implement our models on Py Torch(Paszke et al. 2017) and train them on one Telsa V100 GPU. We use the Adam Optimizer(Kingma et al. 2014) with β1 = 0.9, β2 = 0.999. The Depth Net and Pose Net are trained for 20 epochs, with a batch size of 12. The initial learning rates for both network are 1 10 4 and decayed after 15 epochs by factor of 10. The training sequences are consist of three consecutive images. We set the SSIM weight to α = 0.85 and smooth loss weight to λ = 1 10 3. Depth Net. We implement our HR-Depth with Res Net-18(He et al. 2016) as encoder, and use Mobile Net V3(Howard et al. 2019) as encoder for Lite-HRDepth. The details of our proposed architecture will be described at supplemental material. All four disparity maps are used in loss calculation during training. For evaluation, we only use the maximum output scale, after being Supervision Resolution Dataset Abs rel Sq Rel RMSE RMSElog δ < 1.25 δ < 1.252 δ < 1.253 Sf MLearner M 416 128 CS + K 0.198 1.836 6.565 0.275 0.718 0.901 0.960 Vid2Depth M 416 128 CS + K 0.159 1.231 5.912 0.243 0.784 0.923 0.970 Struct2Depth M 416 128 K 0.141 1.026 5.291 0.215 0.816 0.945 0.979 Monodepth2 M 640 192 K 0.115 0.903 4.863 0.193 0.877 0.959 0.981 Pack Net-Sf M M 640 192 K 0.111 0.785 4.601 0.189 0.878 0.960 0.982 HR-Depth(Ours) M 640 192 K 0.109 0.792 4.632 0.185 0.884 0.962 0.983 HR-Depth(Ours) M 640 192 CS + K 0.108 0.955 4.800 0.190 0.887 0.961 0.981 Monodepth2 MS 640 192 K 0.106 0.818 4.750 0.196 0.874 0.957 0.979 HR-Depth(Ours) MS 640 192 K 0.107 0.785 4.612 0.185 0.887 0.962 0.982 HR-Depth(Ours) MS 640 192 CS + K 0.104 0.786 4.544 0.182 0.893 0.964 0.983 Zhou et al. M 1248 384 K 0.121 0.837 4.945 0.197 0.853 0.955 0.982 Monodepth2 M 1024 320 K 0.115 0.882 4.701 0.190 0.879 0.961 0.982 Pack Net-Sf M M 1280 384 K 0.107 0.802 4.538 0.186 0.889 0.962 0.981 HR-Depth(Ours) M 1024 320 K 0.106 0.755 4.472 0.181 0.892 0.966 0.984 HR-Depth(Ours) M 1280 384 K 0.104 0.727 4.410 0.179 0.894 0.966 0.984 Monodepth2 MS 1024 320 K 0.106 0.818 4.750 0.196 0.874 0.957 0.979 HR-Depth(Ours) MS 1024 320 K 0.101 0.716 4.395 0.179 0.899 0.966 0.983 Table 2: Quantitative results of depth estimation on KITTI dataset for distance up to 80m. For error evaluating indexes, Abs Rel, Sq Rel, RMSE and RMSElog, lower is better, and for accuracy evaluating indexes, δ < 1.25, δ < 1.252, δ < 1.253, higher is better. In the dataset column, CS + K refers to pre-training on City Scapes(CS) and fine-tuning on KITTI(K). M refers to Depth Net that is supervised by monocular(M) image sequence and MS refers to Depth Net that is supervised by monocular and stereo (MS) images. At test time, we scale outputs of Depth Net with median ground-truth Li DAR information. Input image Ours(14.62M) Pack Net-Sf M(128.29M) Monodepth2(14.84M) Ours(Lite)(3.1M) Figure 4: Qualitative monocular depth estimation performance comparing HR-Depth and Lite-HR-Depth with previous SOTA. on frames from the KITTI dataset. Our network is able to predict more sharper edges than Monodepth2(Godard et al. 2019), and its performance is comparable to Pack Net-Sf M while with much fewer parameters. resized to the ground-truth depth resolution using bilinear interpolation. Pose Net. The architecture of Pose Net is proposed by (Godard et al. 2019). The Pose Net is built on Resnet-18 and the first-level convolution channel is changed from 3 to 6, which allows the adjacent frames to feed into the network. And the outputs of Pose Net is the relative pose which is parameterized with 6-DOF vector. The first three dimensions represent translation vectors and the last three represent Euler angles. Depth Estimation Performance We evaluate depth prediction on KITTI using the metrics described in (Eigen et al. 2015). The evaluation results are summarized in Table 2 and then illustrate their performance qualitatively in Figure 4. We show that our proposed architecture outperforms all existing So TA self-supervised approaches. We also outperform recent model (Guizilini et al. 2020a) with 120M parameters. Furthermore, we also introduce an additional source of unlabeled videos, City Scapes dataset(CS+K), and we can further improve the Depth Net performance. We also show that at higher resolution out model s performance significantly increases. Our best results are achieved So TA when processing higher resolution input images with minimum parameters. As shown in Table 3, when training with low-resolution images, Lite-HR-Depth can perform better than Monodepth2 with teacher (T) network supervising. However, when training with high-resolution images, Lite-HR-Depth can perform better than Monodepth2 without additional supervision signal (M). Ablation Studies To further explore the performance improvements that our network provides, we perform an ablative analysis on the Supervision Resolution #Para Abs rel Sq Rel RMSE RMSElog δ < 1.25 δ < 1.252 δ < 1.253 Py D-Net M 512 256 1.9M 0.146 1.291 5.907 0.245 0.801 0.926 0.967 Monodepth2 M 640 192 14.84M 0.115 0.903 4.863 0.193 0.877 0.959 0.981 Lite-HR-Depth(Ours) M 640 192 3.1M 0.116 0.845 4.841 0.190 0.866 0.957 0.982 Lite-HR-Depth(Ours) T 640 192 3.1M 0.112 0.767 4.557 0.184 0.876 0.963 0.984 Monodepth2 M 1024 320 14.84M 0.115 0.882 4.701 0.190 0.879 0.961 0.982 Lite-HR-Depth(Ours) M 1024 320 3.1M 0.111 0.799 4.612 0.184 0.878 0.963 0.984 Lite-HR-Depth(Ours) T 1024 320 3.1M 0.105 0.704 4.374 0.178 0.891 0.967 0.985 Lite-HR-Depth(Ours) T 1280 384 3.1M 0.104 0.697 4.349 0.176 0.893 0.967 0.985 Table 3: Quantitative performance of Lite-Network on KITTI dataset for distance up to 80m. Evaluation matrices and methods are the same as Table 2. In supervision column, T refers to using teacher network to guide lite network to train. Figure 5: Visualization of feature maps at first level Left side shows feature maps in HR-Depth and right side shows feature maps in Monodepth2. different architectural components introduced. We choose Monodepth2 as our baseline model, and we can see all our contributions can lead a significant improvement. Benefits of dense skip-connection. As shown in the Figure 5, the semantic gap in skip connection is too large to be fused well, but the dense skip-connection can leverage intermediate features to effectively reduces the semantic gap between encoder and decoder. Therefore, we can get high-resolution feature maps with more semantic information which can significantly improve the performance. f SE Block. The f SE block is designed to improve feature fusion and computation efficiency. To validate f SE, we also apply the SE(Hu et al. 2019) to dense skip connection and compare it with the proposed f SE block. As shown in Table 4, the SE block will introduce additional parameters but our f SE block can greatly reduce the parameters introduced by dense skip-connection and can further improve the performance of the network, even better than SE block. Feature Map Visualization As explained before, the purpose of redesigning skipconnection is to decrease semantic and spatial gap, so as to obtain high-resolution feature maps with rich semantic information. Therefore, in order to illustrate the effects of feature fusion, we visualize the intermediate feature maps and we also plot the outputs of f SE to explore the influence of #Para Abs rel RMSE δ < 1.25 baseline 14.84M 0.115 4.863 0.877 +dense SC 16.06M 0.112 4.706 0.881 +f SE 13.48M 0.114 4.821 0.875 +dense SC and SE 16.23M 0.110 4.655 0.882 +dense SC and f SE 14.62M 0.109 4.632 0.884 Table 4: Ablation Studies. Results for different variants of our model with monocular training on KITTI at low resolution on Eigen split. The baseline model is Monodepth2. We introduce the dense skip-connection(dense SC) to original structure and compare effect of SE block and f SE block. intermediate feature maps. Figure 4 shows that the low-level node like Xe 1 undergoes slight transformation and obtains simple spatial information whereas the output of decoder node like Xd 2 gets the rich semantic information. Hence, there is large gap between the representation capability of Xe 1 and Xd 2. The dense skip connection can gradually add semantic information to the intermediate feature, thus reducing the gap between node Xe 1 and Xd 2. Conclusion In this paper, we show theoretical and empirical evidence that how to improve high-resolution estimation performance. And based on analysis, we present a new convolutional network architecture, referred as HR-Depth, for high-resolution self-supervised monocular depth estimation. It leverages novel dense skip-connection and f SE block to reduce the gap between resolution and semantic. Although purely trained on image sequences, out approach outperforms other existing self and semi-supervised methods and is even competitive with supervised method. Furthermore, we propose a simple yet efficient strategy to design the lightweight network. The experiments demonstrate that Lite-HR-Depth can perform on par with large model with fewer parameters. Acknowledgments This work is supported in part by the Key Research and Development Program of Guangdong Province of China (2019B010120001) and the National Natural Science Foundation of China under Grant 61836015. References Aleotti, F.; Zaccaroni, G.; Bartolomei, L.; Poggi, M.; Tosi, F.; and Mattoccia, S. 2020. Real-time single image depth perception in the wild with handheld devices. ar Xiv preprint ar Xiv:2006.05724 . Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; and Schiele, B. 2016. The Cityscapes Dataset for Semantic Urban Scene Understanding. In CVPR. Eigen, D.; and Fergus, R. 2015. Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In ICCV, 2650 2658. Garg, R.; BG, V. K.; Carneiro, G.; and Reid, I. 2016. Unsupervised cnn for single view depth estimation: Geometry to the rescue. In ECCV, 740 756. Springer. Geiger, A.; Lenz, P.; and Urtasun, R. 2012. Are we ready for autonomous driving? the kitti vision benchmark suite. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 3354 3361. IEEE. Godard, C.; Mac Aodha, O.; and Brostow, G. J. 2017. Unsupervised monocular depth estimation with left-right consistency. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 270 279. Godard, C.; Mac Aodha, O.; Firman, M.; and Brostow, G. J. 2019. Digging into self-supervised monocular depth estimation. In ICCV, 3828 3838. Guizilini, V.; Ambrus, R.; Pillai, S.; Raventos, A.; and Gaidon, A. 2020a. 3D Packing for Self-Supervised Monocular Depth Estimation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Guizilini, V.; Hou, R.; Li, J.; Ambrus, R.; and Gaidon, A. 2020b. Semantically-Guided Representation Learning for Self-Supervised Monocular Depth. In International Conference on Learning Representations (ICLR). He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. Hinton, G.; Vinyals, O.; and Dean, J. 2015. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531 . Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. 2019. Searching for mobilenetv3. In Proceedings of the IEEE International Conference on Computer Vision, 1314 1324. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; and Wu, E. 2019. Squeeze-and-Excitation Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence 1 1. Huang, G.; Liu, Z.; Van Der Maaten, L.; and Weinberger, K. Q. 2017. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, 4700 4708. Kingma, D. P.; and Ba, J. 2014. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980 . Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; and Navab, N. 2016. Deeper depth prediction with fully convolutional residual networks. In 2016 Fourth international conference on 3D vision (3DV), 239 248. IEEE. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; De Vito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; and Lerer, A. 2017. Automatic differentiation in pytorch . Pillai, S.; Ambrus, R.; and Gaidon, A. 2019. Super Depth: Self-Supervised, Super-Resolved Monocular Depth Estimation. In IEEE International Conference on Robotics and Automation. Poggi, M.; Aleotti, F.; Tosi, F.; and Mattoccia, S. 2018. Towards real-time unsupervised monocular depth estimation on cpu. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, 5848 5854. IEEE. Ramamonjisoa, M.; Du, Y.; and Lepetit, V. 2020. Predicting Sharp and Accurate Occlusion Boundaries in Monocular Depth Estimation Using Displacement Fields. ar Xiv: Computer Vision and Pattern Recognition . Ronneberger, O.; Fischer, P.; and Brox, T. 2015. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, 234 241. Springer. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A. P.; Bishop, R.; Rueckert, D.; and Wang, Z. 2016. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network 1874 1883. Sun, K.; Xiao, B.; Liu, D.; and Wang, J. 2019. Deep High Resolution Representation Learning for Human Pose Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Wofk, D.; Ma, F.; Yang, T.-J.; Karaman, S.; and Sze, V. 2019. Fastdepth: Fast monocular depth estimation on embedded systems. In 2019 International Conference on Robotics and Automation, 6101 6108. IEEE. You, Y.; Wang, Y.; Chao, W.-L.; Garg, D.; Pleiss, G.; Hariharan, B.; Campbell, M.; and Weinberger, K. Q. 2019. Pseudo Li DAR++: Accurate Depth for 3D Object Detection in Autonomous Driving. In ICLR. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; and Sun, J. 2018. Ex Fuse: Enhancing Feature Fusion for Semantic Segmentation. In ECCV. Zhou, J.; Wang, Y.; Qin, K.; and Zeng, W. 2019. Unsupervised high-resolution depth learning from videos with dual networks. In Proceedings of the IEEE International Conference on Computer Vision, 6872 6881. Zhou, T.; Brown, M.; Snavely, N.; and Lowe, D. G. 2017. Unsupervised learning of depth and ego-motion from video. In CVPR, 1851 1858. Zhou, Z.; Siddiquee, M. M. R.; Tajbakhsh, N.; and Liang, J. 2018. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, 3 11. Springer.