INTRODUCTION
Depth estimation is of paramount importance in minimally invasive surgery (MIS) facilitating, soft tissue 3D reconstruction, surgical robot navigation and augmented reality. A popular method of perceiving depth is with a stereo camera by estimating the horizontal displacement (disparity) from the left image pixels to the corresponding right. End-to-end deep learning stereo reconstruction methods have become state-of-the-art. They are typically categorized as supervised and self-supervised. Supervised learning methods directly predict disparity (regression) by training on ground truth disparity data []. They usually undergo training using synthetic data [], followed by fine-tuning on real-world scenes to overcome the limited availability of ground truth. Despite their popularity, they face several challenges in surgical applications. Discrepancy between training and inference data, limited ground truth data, lower image quality, restricted space, and a multitude of scene variations, are factors that lead to reduced performance.
Self-supervised methods don't require ground truth disparity, making images-only training data abundant. These methods optimize photometric re-projection error through novel view synthesis []. Self-supervised methods show potential, but are not as effective as supervised methods. This is because the photometric error can be optimized for a large range of disparity values, resulting in inconsistent geometry. Furthermore, the relative depth information between stereo images is inherently scale ambiguous. Making it challenging to learn robust representation of depth via self-supervision, especially in scenes with occlusions, textureless regions, or repetitive structures, which is prevalent in surgery. Supervised learning is key for robust stereo depth, since training on ground truth provides a strong, unambiguous signal for learning accurate representations. A question emerges though; can we have the best of both worlds, i.e. unambiguous depth that is also generalizable to different scenes?
For achieving generalizability, visual representation learning is key. Learning robust feature representations also allows improved downstream performance as shown in []. These techniques utilize image reconstruction as a preliminary task, relying on the principle that by learning to predict patches in a masked image, useful representations about the context of the scene can be obtained for further tasks. This is proven by the enhanced data label efficiency on various standard benchmark datasets. Hence, in this work we experiment with this masked image modelling (MIM) approach, to train an encoder model to learn stereo feature representations, that can be fine-tuned downstream to encode robust depth feature representations. This endows our model with both generalizable features via MIM and unambigous sharp depth via supervised-learning for disparity estimation.
In this paper we build on [] and propose StereoMAE a two-stage training process which entails (1) training an encoder via MIM, to generate robust feature representations for left and right views, followed by (2) supervised training for disparity estimation. Furthermore, we enhance MIM in (1) by using perceptual similarity learning [], leading to learned representations that effectively capture the intricacies of the scene and object boundaries without explicit guidance or manually designed inductive biases. Our contributions are the following:
- A novel approach for training stereo depth models by combining self-supervised MIM and supervised stereo depth estimation. To the best of our knowledge, this is the first work to apply MIM for stereo depth estimation;
- We propose a new approach to boost MIM by incorporating perceptual similarity loss term for learning generalizable visual semantic concepts;
- We present a modular model architecture for combining any pre-trained MIM encoder model with any off-the-shelf decoder to enhance depth estimation;
- Our joint MIM-supervised approach enhances performance, yielding a generalizable model with sub-millimetre accuracy in surgical depth estimation
Specific to the field of surgery the lack of datasets with ground truth, restricts state-of-the-art stereo models from reaching sub-mm precision on surgical scenes despite their generalizability. This is due to the need for deep understanding of scene semantics, a task beyond the capabilities of supervised learning alone. Hence, our method pairs MIM with StereoMAE model architecture, and perceptual learning to surmount the limited training data issue, achieving high generalizability and sub-mm accuracy in surgery. This blend of generalizability and precision offers a significant advantage, expanding its potential use across diverse medical imaging tasks, not just specific surgical scenarios like laparoscopic surgery where depth estimation at sub-millimetre accuracy is critical as it helps the localization and perception of surgical instruments, tumour and surrounding healthy tissue (essential in the development of autonomous task execution in robotic MIS). The performance of StereoMAE has been evaluated on (3) MIS datasets; SCARED [], Hamlyn [], and SERV-CT [], and on (2) non-surgical scenes data; ETH3D [] and Middlebury [].
METHODS
The method is divided into two phases; (1) Pre-training via MIM for enhanced representation learning followed by (2) supervised downstream fine-tuning for stereo depth estimation. The learning framework for (1) is inspired by MAE []. Supervised training in (2) is based on RAFT-Stereo []. However for (2) any training methodology from the supervised stereo depth literature can be used.
Stereo masked image modelling
MIM involves randomly masking pixels of an input image and training a model to predict the invisible content []. The intuition is that the model learns the representation of the masked pixels by inferring details from the surrounding valid pixels. In the end, if a model is capable of reconstructing the missing content with valid pixels, the model has learnt the context within the image by encoding its semantic features.
We adapt the method in [], which utilizes an auto-encoder model for stereo images as shown in Figure . In particular, we construct the auto-encoder model, using vision transformer (ViT) components []. The encoder consists of several layers of ViT-Base (ViT-B), a 12 layer transformer model. The custom decoder is composed of 8 transformer layers. Given the left image as input, it is first resized to 224448, divided into patches of size 1632 and 75% of them are randomly masked. The encoder is fed only the un-masked patches as input, and it generates corresponding features representing the visible scene parts. These features are then concatenated with the patches that were not fed into the encoder and inputted into the decoder to reconstruct the full left image as output. The model is trained via pixel image reconstruction (photometric error). The same is done for the right input image. For the left and right input views, the weights of the encoder and decoder are shared, allowing them to learn joint global stereo representations for reconstructing the scene.
[IMAGE OMITTED. SEE PDF]
In our work, we do not train the aforementioned model using the mean square error (MSE) loss from the original work []. MSE, which compares pixel intensities, is a low-level metric and insufficient for capturing complex structures in an image as it assumes pixel-wise independence. Hence a model's capacity is wasted since high frequency global feature distribution will not be captured. Perceptual similarity, which mimics human visual perception [], is a better metric for evaluating image similarity and capturing high-level semantic features, essential for model generalization. Hence, we hypothesize, that training MIM with perceptual similarity will enhance the models' output quality and coherence. The focus is on learning high-level relationships between elements in the image, not just low-level details. Thereby reducing the model's vulnerability to masked regions and improving overall semantic content and structure capture.
Therefore, to train StereoMAE we utilize a perceptual loss for MIM for learning the required high-level semantic features that represent the scene context (thereby improving its downstream generalizability). Inspired by [] our perceptual loss comprises of three loss terms: (i) L1 (also known as absolute error loss, measures the absolute difference between the predicted and actual pixel values, calculated as the sum of the absolute differences between the two values). (ii) Feature matching (compares intermediate features from a pre-trained model like VGG, to encourage learning similar features for the predicted and actual pixel values). Lastly, (iii) style transfer loss (computes correlations between feature maps extracted from a model like VGG, encouraging the predicted and actual image distributions match). L1 loss encourages sparse representations by optimizing absolute intensity values. The feature matching loss compares the feature representations of the reconstructed output and target image, while the style loss measures the difference in statistical distribution of high-level feature activations, capturing image texture and aesthetics. To calculate (ii) and (iii) we utilize a pre-trained feature extractor model, i.e. VGG16 (though any arbitrary model can be used).
Supervised downstream fine-tuning
Once the encoders have been pre-trained via MIM in phase (1) we finetune them for disparity estimation. The full model architecture for downstream training is displayed in Figure . The architecture is designed to be modular such that the pre-trained encoders can be combined with any off-the-shelf decoder of the user's choice. Thereby enabling a smooth transition from MIM to disparity estimation without further modifications. The pre-trained ViT encoders generate features for the (un-masked) left and right input images, and output a 3D tensor. The features have to be reshaped into a 4D tensor for it to be processed by the decoders which are typically composed of convolution layers. However, once reshaped, the positional encodings of the features change. To ensure the decoder focuses on the necessary elements of the feature map, we parse the feature map through a feature converter block; its purpose is to learn the transfer of MIM-trained features to stereo-disparity features and select the relevant elements needed for the decoder. It comprises of two 55 2D convolution layers and a bilinear interpolation module that resizes the tensor to a scale of choice. The scale depends on the type of decoder used. In our experiments we utilize the RAFT-Stereo decoder []. The feature converter block essentially plays a crucial role in ensuring the appropriate transfer and transformation of features from the pre-trained ViT encoder to the disparity decoder. In our experiments this is vital as the encoder is trained only for the MIM task, hence the features must be adapted for disparity downstream.
[IMAGE OMITTED. SEE PDF]
To train the final model we utilized a supervised-learning loss comprised of the L1 error between the predicted and ground truth disparity. The RAFT-Stereo decoder generates a sequence of predictions (from its multi-level GRU networks), hence our L1 error is computed over the full sequence of predictions, [], with exponentially increasing weights. If a different decoder that outputs a single disparity map is generated then, only the final single output would have been used for the loss calculation. Considering ground truth disparity is defined as , the loss is defined as
Implementation details
For MIM training, we use the ViT-B architecture [], trained for 150 epochs on a combination of the following datasets [] (a total of 411,942 stereo image pairs). The input patch size is fixed to 1632 and we mask 75% of input patches during training. The ViT-B encoder, comprises of 12 transformer layers (), each with 12 self-attention heads and the final hidden dimension output is of size 768. The decoder architecture comprises of 8 transformer layers () each with 16 self-attention heads and the hidden dimension output of size 512. The data augmentation strategies include (1) resizing / cropping image to 224448, (2) adding random distracting shapes like rectangles of varying size and (3) random colour jittering. We train with a batch size of 8 on 4-Nvidia Tesla GPUs. Adam optimizer with a learning rate of 0.00015, and weight decay of 0.05 (cosine strategy), 40 warm-up epochs and the momentum parameters and are 0.9 and 0.95 respectively were used. For downstream training, the pretrained ViT-B model is used as feature extractor with the RAFT-stereo decoder architecture, for disparity estimation. The same training strategy as [] was utilized, though since our model is modular, it can be combined with any downstream stereo-decoder model and training method. For the downstream training of disparity estimation, only the Sceneflow training split [] was used (a combination of FlyingThings, Monkaa and Driving). Hence, only synthetic data was used for training. Note, synthetic data is used for perfect, unambiguous ground truth disparity. Synthetic surgical scenes can also be used for training, provided high quality disparity maps are available. Once, trained inference operates at 14 frames per second (fps) on a single Nvidia Tesla GPU (where frames here includes both left and right pairs).
EXPERIMENTAL RESULTS AND DISCUSSION
We evaluated the performance of our model on both surgical and natural scenes, despite not training on them. The following datasets were used for testing; SCARED, Hamlyn, SERV-CT, ETH3D, and Middlebury datasets. To compare our results with other methods in the literature, we calculate the standard metrics for stereo disparity evaluation, i.e. average end-point-error (EPE) and the percent of pixels with EPE greater than a specified threshold (0.5, 1.0, and 2.0). Specifically for SCARED and SERV-CT, we also calculate the mean absolute depth error in mm on test sets 1 and 2. We also exhibit qualitative performance of our MIM pre-training and downstream disparity estimation in Figures and . Furthermore, due to lack of ground truth available in Hamlyn MIS data we show qualitative downstream performance comparison with other SOTA methods in Figure .
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
Figure compares the reconstructed outputs of StereoMAE trained via the MSE loss from [] and the proposed perceptual loss, on unseen datasets. It can be observed that StereoMAE trained with perceptual loss outperforms MSE in all scenes. When using perceptual loss, we observe significant increase in the fidelity of the reconstructed patches with finer textural and structural details. Exhibiting perceptual loss aids the model in learning higher-level feature representations that can generalize to any scene type despite never having observed surgical scenes. When finetuned for disparity estimation, StereoMAE visibly generates sharper depth maps despite only being trained on synthetic data, as shown in Figure . Specifically, in sample A, StereoMAE generates the fine details on the robotic tool, whereas other methods either fail to achieve the correct disparity range or miss fine details around the edges. Similarly in B and C, StereoMAE generates less holes around the edges, as can be seen in the wheels/handle of the bike in B and the structures in the background in C. Hence, by learning a generalizable feature distribution for stereo image representations and disparity estimation, without any training on datasets from Figure , StereoMAE creates smooth yet fine disparities. We also conducted an ablation study by creating a new StereoMAE variant. It pairs the StereoMAE encoder with HITNet's decoder [], fine-tuned on the Scene Flow dataset. It outperforms the original HITNet model in ETH3D, Middlebury and SCARED evaluations, confirming our method's decoder-independence and effectiveness in enhancing any decoder for disparity estimation. Original HITNet achieves EPE of 0.31, 1.62 on ETH3D and Middlebury, respectively and 4.12 mm depth error on SCARED. Whereas our variant achieves 0.25, 1.27, and 2.98 mm error, respectively. Our approach, as demonstrated in Figure , is independent of baseline, lighting, image type (like ETH3D's greyscale), or specific scene types. Despite training solely on simulations, it attains sub-mm accuracy (shown in Tables and ) on surgical scenes, delivering SOTA performance on simulated, natural, and surgical scenes alike. Affirming the significance of learning deep semantic features via MIM in pre-training encoders. This is further reflected in Figure where it is clear samples A, B, and E StereoMAE generates sharper details around the background anatomical structures and the edges of the robotic tools. Furthermore StereoMAE in sample D resolves disparity for the anatomical structures and the left-robotic tool even in the presence of smoke. Despite not training on surgical scenes, StereoMAE consistently outperformes competitors by generating sharper disparity maps under challenging lighting conditions and image resolution.
The quantitative evaluation in Tables also show our model outperforming previous state-of-the-art stereo depth estimation models in Stereo-benchmark datasets. StereoMAE achieves lower EPE across the board on all datasets, despite not being trained on them. In Table , SCARED benchmark only shows EPE and 3px error as they are the most commonly reported by previous methods on this dataset []. Furthermore, as shown in Tables and , StereoMAE also sets a new record on surgical scenes, achieving sub-millimetre accuracy in depth estimation. This exhibits the importance of visual representation learning for developing generalizable feature distributions, and still retaining the fine details of disparity estimation from supervised learning. Enabling generalizability to surgical images, without training on them.
TABLE 1 Results of generalization without fine-tuning on ETH3D, Middlebury, and SCARED datasets. Lower is better for all metrics. Method highlighted in bold is best.
ETH3D | Middlebury 2014 (H) | SCARED Small | ||||||||
EPE | 0.5 px | 1 px | 2 px | EPE | 0.5 px | 1 px | 2 px | EPE | 3 px | |
Method | (pixel) | (%) | (%) | (%) | (pixel) | (%) | (%) | (%) | (pixel) | (%) |
LEAStereo [] | 0.63 | 22.50 | 8.54 | 3.94 | 3.20 | 49.44 | 28.51 | 16.03 | 22.62 | 27.07 |
STTR [] | 1.29 | 30.37 | 11.33 | 4.85 | 2.22 | 40.30 | 22.00 | 12.21 | 6.03 | 9.52 |
RAFT-Stereo [] | 0.27 | 8.10 | 2.33 | 0.96 | 1.12 | 24.28 | 14.02 | 8.43 | 1.16 | 4.59 |
StereoMAE | 0.22 | 6.77 | 2.10 | 0.78 | 0.98 | 22.14 | 11.33 | 7.39 | 0.96 | 3.94 |
TABLE 2 Mean absolute depth error (mm) on SCARED. Best results are highlighted in bold. None of the models were fine-tuned on SCARED, inference only.
SCARED 2019 | SCARED Small | ||
Testdata 1 | Testdata 2 | Small 19 [] | |
Methods | MAE () | MAE () | MAE () |
LEAStereo [] | 3.82 | 4.51 | 1.48 |
STTR [] | 4.14 | 5.91 | 11.31 |
RAFT-Stereo [] | 3.74 | 4.28 | 1.01 |
StereoMAE | 1.66 | 2.01 | 0.92 |
TABLE 3 Mean absolute depth error (mm) and EPE on SERV-CT data. Best results are highlighted in bold. None of the models were fine-tuned on SERV-CT, inference only.
SERV-CT | ||||
Experiment 1 | Experiment 2 | |||
Methods | MAE () | EPE | MAE () | EPE |
LEAStereo [] | 3.14 | 3.05 | 3.66 | 5.39 |
STTR [] | 16.60 | 7.54 | 25.69 | 14.52 |
RAFT-Stereo [] | 2.03 | 1.32 | 2.61 | 2.27 |
StereoMAE | 0.99 | 1.06 | 1.19 | 1.68 |
CONCLUSIONS
Accurate and generalizable stereo depth estimation is crucial for surgical applications. Supervised learning is effective but limited by the scarcity of ground truth data in surgical settings, restricting their generalizability. Self-supervised methods have issues with scale ambiguity and inaccurate disparity prediction. Our StereoMAE approach demonstrates that by integrating MIM for feature representation learning with supervised depth estimation, achieves the benefits of both methods. Despite training only on synthetic data, StereoMAE shows generalizability to surgical and natural scenes, achieving sub-millimetre accuracy in surgical scenes. This is attributed to robust feature representations learned during MIM pre-training, which enhances the model's performance for subsequent tasks. Future work will explore alternative designs and loss terms to enhance MIM and improve stereo depth estimation.
NOMENCLATURE
StereoMAE
Stereo masked autoencoder
MIM
Masked image modelling
MIS
Minimally invasive surgery
AUTHOR CONTRIBUTIONS
All authors contributed to the conceptualization and design of the study. Samyakh Tukra wrote the basline script for training StereoMAE both pre-training and downstream finetuning. Haozheng Xu and Chi Xu performed the analysis, Haozheng Xu further aided with the interpretation of the results. All authors contributed to manuscript revision, and read and approved the submitted version.
ACKNOWLEDGMENTS
All authors are supported by the Royal Society (UF140290 and RGF EA 180084) and the NIHR Imperial Biomedical Research Centre (BRC).
CONFLICT OF INTEREST STATEMENT
The authors declare no conflicts of interest.
DATA AVAILABILITY STATEMENT
All data used in the study is open sourced. Hence, no data are available.
Lipson, L., Teed, Z., Deng, J.: RAFT‐stereo: multilevel recurrent field transforms for stereo matching. arXiv:210907547 (2021)
Guo, X., Yang, K., Yang, W., Wang, X., Li, H.: Group‐wise correlation stereo network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3273–3282. IEEE, Piscataway, NJ (2019)
Zhang, F., Prisacariu, V., Yang, R., Torr, P.H.: GA‐Net: guided aggregation net for end‐to‐end stereo matching. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 185–194. IEEE, Piscataway, NJ (2019)
Tankovich, V., Häne, C., Fanello, S., Zhang, Y., Izadi, S., Bouaziz, S.: Hitnet: Hierarchical iterative tile refinement network for real‐time stereo matching. In: 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14357–14367. IEEE, Piscataway, NJ (2021)
Mayer, N., Ilg, E., Häusser, P., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14357–14367. IEEE, Piscataway, NJ (2016)
Tukra, S., Giannarou, S.: Stereo depth estimation via self‐supervised contrastive representation learning. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, pp. 604–614. Springer, Cham (2022)
Pilzer, A., Lathuilière, S., Xu, D., Puscas, M.M., Ricci, E., Sebe, N.: Progressive fusion for unsupervised binocular depth estimation using cycled networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2380–2395 (2019)
Wang, Y., Wang, P., Yang, Z., Luo, C., Yang, Y., Xu, W.: UnOS: unified unsupervised optical‐flow and stereo‐depth estimation by watching videos. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8063–8073. IEEE, Piscataway, NJ (2019)
Zhong, Y., Dai, Y., Li, H.: Self‐supervised learning for stereo matching with self‐improving ability. arXiv:1709.00930 (2017)
Tonioni, A., Tosi, F., Poggi, M., Mattoccia, S., Di Stefano, L.: Real‐time self‐adaptive deep stereo. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 195–204. IEEE, Piscataway, NJ (2019)
He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16000–16009. IEEE, Piscataway, NJ (2022)
Bao, H., Dong, L., Piao, S., Wei, F.: BEiT: BERT pre‐training of image transformers. arXiv:2106.08254 (2022)
Chen, M., Radford, A., Child, R., et al.: Generative pretraining from pixels. In: H.D. III, A. Singh, (eds.) ICML. vol. 119 of Proceedings of Machine Learning Research, pp. 1691–1703. Microtome Publishing, Brookline, MA (2020)
Johnson, J., Alahi, A., Fei‐Fei, L.: Perceptual losses for real‐time style transfer and super‐resolution. In: European Conference on Computer Vision, pp. 694–711. Springer, Cham (2016)
Allan, M., McLeod, A.J., Wang, C.C., et. al, J.R.: Stereo correspondence and reconstruction of endoscopic data challenge. arXiv:2101.01133 (2021)
Giannarou, S., Visentini‐Scarzanella, M., Yang, G.Z.: Probabilistic tracking of affine‐invariant anisotropic regions. IEEE Trans. Pattern Anal. Machine Intell. 35(1), 130–143 (2012)
Edwards, P., Psychogyios, D., Speidel, S., Maier‐Hein, L., Stoyanov, D.: SERV‐CT: a disparity dataset from CT for validation of endoscopic 3D reconstruction. arXiv:201211779 (2020)
Schöps, T., Schönberger, J.L., Galliani, S., et al.: A multi‐view stereo benchmark with high‐resolution images and multi‐camera videos. In: Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2538–2547. IEEE, Piscataway, NJ (2017)
Hirschmüller, H., Scharstein, D.: Evaluation of cost functions for stereo matching. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8. IEEE, Piscataway, NJ (2007)
Kolesnikov, A., Dosovitskiy, A., Weissenborn, D., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: (2021)
Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. arXiv:2010.11929 (2018)
Tukra, S., Marcus, H.J., Giannarou, S.: See‐through vision with unsupervised scene occlusion reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3779–3790 (2022). [DOI: https://dx.doi.org/10.1109/TPAMI.2021.3058410]
Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High‐resolution image synthesis and semantic manipulation with conditional GANs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798–8807. IEEE, Piscataway, NJ (2018)
Cordts, M., Omran, M., Ramos, S., et al.: The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3213–3223. IEEE, Piscataway, NJ (2016)
Mayer, N., Ilg, E., Häusser, P., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4040–4048. IEEE, Piscataway, NJ (2016)
Tremblay, J., To, T., Birchfield, S.: Falling things: a synthetic dataset for 3D object detection and pose estimation. In: CVPR Workshop on Real World Challenges and New Benchmarks for Deep Learning in Robotic Vision, pp. 2119–21193. IEEE, Piscataway, NJ (2018)
Wang, W., Zhu, D., Wang, X., et al.: TartanAir: a dataset to push the limits of visual SLAM. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4909–4916. IEEE, Piscataway, NJ (2020)
Butler, D.J., Wulff, J., Stanley, G.B., Black, M.J.: A naturalistic open source movie for optical flow evaluation. In: A. Fitzgibbon et al. (eds.) European Conf. on Computer Vision (ECCV). Part IV, LNCS 7577, pp. 611–625. Springer, Berlin, Heidelberg (2012)
Cheng, X., Zhong, Y., Harandi, M., et al.: Hierarchical neural architecture search for deep stereo matching. In: NIPS'20: Proceedings of the 34th International Conference on Neural Information Processing Systems, pp. 22158–22169. ACM, New York (2020)
Li, Z., Liu, X., Drenkow, N., et al.: Revisiting stereo depth estimation from a sequence‐to‐sequence perspective with transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 6197–6206. IEEE, Piscataway, NJ (2021)
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Generalizable and accurate stereo depth estimation is vital for 3D reconstruction, especially in surgery. Supervised learning methods obtain best performance however, limited ground truth data for surgical scenes limits generalizability. Self‐supervised methods don't need ground truth, but suffer from scale ambiguity and incorrect disparity prediction due to inconsistency of photometric loss. This work proposes a two‐phase training procedure that is generalizable and retains the high performance of supervised methods. It entails: (1) performing self‐supervised representation learning of left and right views via masked image modelling (MIM) to learn generalizable semantic stereo features (2) utilizing the MIM pre‐trained model to learn robust depth representation via supervised learning for disparity estimation on synthetic data only. To improve stereo representations learnt via MIM, perceptual loss terms are introduced, which improve the model's stereo representations learnt by explicitly encouraging the learning of higher scene‐level features. Qualitative and quantitative performance evaluation on surgical and natural scenes shows that the approach achieves sub‐millimetre accuracy and lowest errors respectively, setting a new state‐of‐the‐art. Despite not training on surgical nor natural scene data for disparity estimation.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer