skip to main content
research-article

Enhanced 3D Shape Reconstruction With Knowledge Graph of Category Concept

Published:04 March 2022Publication History
Skip Abstract Section

Abstract

Reconstructing three-dimensional (3D) objects from images has attracted increasing attention due to its wide applications in computer vision and robotic tasks. Despite the promising progress of recent deep learning–based approaches, which directly reconstruct the full 3D shape without considering the conceptual knowledge of the object categories, existing models have limited usage and usually create unrealistic shapes. 3D objects have multiple forms of representation, such as 3D volume, conceptual knowledge, and so on. In this work, we show that the conceptual knowledge for a category of objects, which represents objects as prototype volumes and is structured by graph, can enhance the 3D reconstruction pipeline. We propose a novel multimodal framework that explicitly combines graph-based conceptual knowledge with deep neural networks for 3D shape reconstruction from a single RGB image. Our approach represents conceptual knowledge of a specific category as a structure-based knowledge graph. Specifically, conceptual knowledge acts as visual priors and spatial relationships to assist the 3D reconstruction framework to create realistic 3D shapes with enhanced details. Our 3D reconstruction framework takes an image as input. It first predicts the conceptual knowledge of the object in the image, then generates a 3D object based on the input image and the predicted conceptual knowledge. The generated 3D object satisfies the following requirements: (1) it is consistent with the predicted graph in concept, and (2) consistent with the input image in geometry. Extensive experiments on public datasets (i.e.,  ShapeNet, Pix3D, and Pascal3D+) with 13 object categories show that (1) our method outperforms the state-of-the-art methods, (2) our prototype volume-based conceptual knowledge representation is more effective, and (3) our pipeline-agnostic approach can enhance the reconstruction quality of various 3D shape reconstruction pipelines.

REFERENCES

  1. [1] Abadi Martín, Barham Paul, Chen Jianmin, Chen Zhifeng, Davis Andy, Dean Jeffrey, Devin Matthieu, Ghemawat Sanjay, Irving Geoffrey, Isard Michael, et al. 2016. tensorflow: A system for large-scale machine learning. In USENIX Symposium on Operating Systems Design and Implementation. 265283.Google ScholarGoogle Scholar
  2. [2] Afzal Hassan, Aouada Djamila, Mirbach Bruno, and Ottersten Björn. 2018. Full 3D reconstruction of non-rigidly deforming objects. ACM Transactions on Multimedia Computing, Communications, and Applications 14, 1s (2018), 123.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. [3] Soltani Amir Arsalan, Huang Haibin, Wu Jiajun, Kulkarni Tejas D., and Tenenbaum Joshua B.. 2017. Synthesizing 3D shapes via modeling multi-view depth maps and silhouettes with deep generative networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu. IEEE. 15111519.Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Cadena Cesar, Carlone Luca, Carrillo Henry, Latif Yasir, Scaramuzza Davide, Neira José, Reid Ian, and Leonard John J.. 2016. Past, present, and future of simultaneous localization and mapping: Toward the robust-perception age. IEEE Transactions on Robotics 32, 6 (2016), 13091332.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Chang Angel X., Funkhouser Thomas, Guibas Leonidas, Hanrahan Pat, Huang Qixing, Li Zimo, Savarese Silvio, Savva Manolis, Song Shuran, Su Hao, et al. 2015. shapenet: An information-rich 3D model repository. arXiv:1512.03012.Google ScholarGoogle Scholar
  6. [6] Chen Hao, Sun Kunyang, Tian Zhi, Shen Chunhua, Huang Yongming, and Yan Youliang. 2020. Blendmask: Top-down meets bottom-up for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual. IEEE. 85738581.Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Chen Zhiqin and Zhang Hao. 2019. Learning implicit fields for generative shape modeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Long Beach. IEEE. 59395948.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Cheung K. M. G., Baker Simon, and Kanade Takeo. 2003. Shape-from-silhouette of articulated objects and its use for human body kinematics estimation and motion capture. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vol. 1, Wisconsin. IEEE. 11.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Choi Seonghwa, Nguyen Anh-Duc, Kim Jinwoo, Ahn Sewoong, and Lee Sanghoon. 2019. Point cloud deformation for single image 3d reconstruction. In IEEE International Conference on Image Processing (ICIP’19). Taipei. IEEE. 23792383.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Choy Christopher B., Xu Danfei, Gwak JunYoung, Chen Kevin, and Savarese Silvio. 2016. 3D-R2N2: A unified approach for single and multi-view 3D object reconstruction. In Proceedings of the European Conference on Computer Vision(Lecture Notes in Computer Science), Vol. 9912, Springer. Amsterdam. Springer. 628644.Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Dai Angela, Qi Charles Ruizhongtai, and Nießner Matthias. 2017. Shape completion using 3D-encoder-predictor CNNs and shape synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu. IEEE. 58685877.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Durrant-Whyte Hugh and Bailey Tim. 2006. Simultaneous localization and mapping. IEEE Robotics & Automation Magazine 13, 2 (2006), 99110.Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Fan Haoqiang, Su Hao, and Guibas Leonidas J.. 2017. A point set generation network for 3D object reconstruction from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu. IEEE. 605613.Google ScholarGoogle ScholarCross RefCross Ref
  14. [14] Gkioxari Georgia, Malik Jitendra, and Johnson Justin. 2019. Mesh R-CNN. In Proceedings of the IEEE International Conference on Computer Vision. Long Beach. IEEE. 97859795.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Groueix Thibault, Fisher Matthew, Kim Vladimir G., Russell Bryan C., and Aubry Mathieu. 2018. A papier-mâché approach to learning 3D surface generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City. IEEE. 216224.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Han Xiaoguang, Li Zhen, Huang Haibin, Kalogerakis Evangelos, and Yu Yizhou. 2017. High-resolution shape completion using deep neural networks for global structure and local geometry inference. In Proceedings of the IEEE International Conference on Computer Vision. Honolulu. IEEE. 8593.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Han Xian-Feng, Laga Hamid, and Bennamoun Mohammed. 2019. Image-based 3D object reconstruction: State-of-the-art and trends in the deep learning era. IEEE Transactions on Pattern Analysis and Machine Intelligence 43, 5 (2019), 15781604.Google ScholarGoogle Scholar
  18. [18] Häne Christian, Tulsiani Shubham, and Malik Jitendra. 2017. Hierarchical surface prediction for 3D object reconstruction. In International Conference on 3D Vision. Qingdao. IEEE. 412420.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Hartley Richard and Zisserman Andrew. 2003. Multiple View Geometry in Computer Vision. Cambridge University Press.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Insafutdinov Eldar and Dosovitskiy Alexey. 2018. Unsupervised learning of shape and pose with differentiable point clouds. In Advances in Neural Information Processing Systems. Montréal. MIT Press. 28022812.Google ScholarGoogle Scholar
  21. [21] Izadi Shahram, Kim David, Hilliges Otmar, Molyneaux David, Newcombe Richard, Kohli Pushmeet, Shotton Jamie, Hodges Steve, Freeman Dustin, Davison Andrew, et al. 2011. KinectFusion: Real-time 3D reconstruction and interaction using a moving depth camera. In Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology. Santa Barbara. ACM. 559568.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] Johnston Adrian, Garg Ravi, Carneiro Gustavo, Reid Ian, and Hengel Anton van den. 2017. Scaling CNNs for high resolution volumetric reconstruction from a single image. In Proceedings of the IEEE International Conference on Computer Vision Workshops. Venice. IEEE. 939948.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Kanazawa Angjoo, Tulsiani Shubham, Efros Alexei A., and Malik Jitendra. 2018. Learning category-specific mesh reconstruction from image collections. In Proceedings of the European Conference on Computer Vision (Lecture Notes in Computer Science), Vol. 11219, Munich. Springer. 386402.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Kar Abhishek, Häne Christian, and Malik Jitendra. 2017. Learning a multi-view stereo machine. In Advances in Neural Information Processing Systems. Long Beach. MIT Press, 365376.Google ScholarGoogle Scholar
  25. [25] Kar Abhishek, Tulsiani Shubham, Carreira Joao, and Malik Jitendra. 2015. Category-specific object reconstruction from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston. IEEE. 19661974.Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Kato Hiroharu and Harada Tatsuya. 2019. Learning view priors for single-view 3D reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Long Beach. IEEE. 97789787.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Kingma Diederik P. and Ba Jimmy. 2014. Adam: A method for stochastic optimization. arXiv:1412.6980 (2014).Google ScholarGoogle Scholar
  28. [28] Kurenkov Andrey, Ji Jingwei, Garg Animesh, Mehta Viraj, Gwak Junyoung, Choy Christopher, and Savarese Silvio. 2018. Deformnet: Free-form deformation network for 3D shape reconstruction from a single image. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision. Lake Tahoe. IEEE. 858866.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] LeCun Yann, Bengio Yoshua, and Hinton Geoffrey. 2015. Deep learning. Nature 521, 7553 (2015), 436444.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Li Jie, Han Kai, Wang Peng, Liu Yu, and Yuan Xia. 2020. Anisotropic convolutional networks for 3D semantic scene completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual. IEEE. 33513359.Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Liao Yi-Lun, Yang Yao-Cheng, Lin Yuan-Fang, Chen Pin-Jung, Kuo Chia-Wen, Chiu Wei-Chen, and Wang Yu-Chiang Frank. 2019. Learning pose-aware 3D reconstruction via 2D-3D self-consistency. In IEEE International Conference on Acoustics, Speech and Signal Processing. Brighton. IEEE. 38573861.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Lin Chen-Hsuan, Kong Chen, and Lucey Simon. 2018. Learning efficient point cloud generation for dense 3D object reconstruction. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 32, New Orleans. AAAI, 1–1.Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Meagher Donald J. R.. 1980. Octree Encoding: A New Technique for the Representation, Manipulation and Display of Arbitrary 3D Objects by Computer. Electrical and Systems Engineering Department, Rensseiaer Polytechnic.Google ScholarGoogle Scholar
  34. [34] Mo Kaichun, Zhu Shilin, Chang Angel X., Yi Li, Tripathi Subarna, Guibas Leonidas J., and Su Hao. 2019. partnet: A large-scale benchmark for fine-grained and hierarchical part-level 3D object understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Long Beach. IEEE. 909918.Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Navaneet K. L., Mandikal Priyanka, Agarwal Mayank, and Babu R. Venkatesh. 2019. capnet: Continuous approximation projection for 3D point cloud reconstruction using 2D supervision. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33, Honolulu. AAAI. 88198826.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Newcombe Richard A., Izadi Shahram, Hilliges Otmar, Molyneaux David, Kim David, Davison Andrew J., Kohi Pushmeet, Shotton Jamie, Hodges Steve, and Fitzgibbon Andrew. 2011. KinectFusion: Real-time dense surface mapping and tracking. In IEEE International Symposium on Mixed and Augmented Reality. Basel. IEEE. 127136.Google ScholarGoogle Scholar
  37. [37] Nguyen Anh-Duc, Choi Seonghwa, Kim Woojae, and Lee Sanghoon. 2019. Graphx-convolution for point cloud deformation in 2D-to-3D conversion. In Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul. IEEE. 86288637.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Pala Pietro and Berretti Stefano. 2019. Reconstructing 3D face models by incremental aggregation and refinement of depth frames. ACM Transactions on Multimedia Computing, Communications, and Applications 15, 1 (2019), 124.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. [39] Pan Junyi, Han Xiaoguang, Chen Weikai, Tang Jiapeng, and Jia Kui. 2019. Deep mesh reconstruction from single RGB images via topology modification networks. In Proceedings of the IEEE International Conference on Computer Vision. Seoul. IEEE. 99649973.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Pan Junyi, Li Jun, Han Xiaoguang, and Jia Kui. 2018. Residual meshnet: Learning to deform meshes for single-view 3D reconstruction. In International Conference on 3D Vision. Verona. IEEE. 719727.Google ScholarGoogle ScholarCross RefCross Ref
  41. [41] Pan Yun-he. 2019. On visual knowledge. Frontiers of Information Technology & Electronic Engineering 20, 8 (2019), 10211025.Google ScholarGoogle Scholar
  42. [42] Pinheiro Pedro O., Rostamzadeh Negar, and Ahn Sungjin. 2019. Domain-adaptive single-view 3D reconstruction. In Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul. IEEE. 76387647.Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Richter Stephan R. and Roth Stefan. 2015. Discriminative shape from shading in uncalibrated illumination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston. IEEE. 11281136.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Richter Stephan R. and Roth Stefan. 2018. Matryoshka networks: Predicting 3D geometry via nested shape layers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City. IEEE. 19361944.Google ScholarGoogle Scholar
  45. [45] Shi Dingfeng, Zhao Yifan, and Li Jia. 2020. Reconstructing part-level 3D models from a single image. In 2020 IEEE International Conference on Multimedia and Expo. Virtual. IEEE. 16.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Smith Edward, Fujimoto Scott, Romero Adriana, and Meger David. 2019. GEOMetrics: Exploiting geometric structure for graph-encoded objects. In Proceedings of the International Conference on Machine Learning, Vol. 97, Long Beach. ACM. 58665876.Google ScholarGoogle Scholar
  47. [47] Stutz David and Geiger Andreas. 2020. Learning 3D shape completion under weak supervision. International Journal of Computer Vision 128, 5 (2020), 11621181.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. [48] Sun Xingyuan, Wu Jiajun, Zhang Xiuming, Zhang Zhoutong, Zhang Chengkai, Xue Tianfan, Tenenbaum Joshua B., and Freeman William T.. 2018. Pix3D: Dataset and methods for single-image 3D shape modeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City. IEEE. 29742983.Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Tang Jiapeng, Han Xiaoguang, Pan Junyi, Jia Kui, and Tong Xin. 2019. A skeleton-bridged deep learning approach for generating meshes of complex topologies from single RGB images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Long Beach. IEEE. 45414550.Google ScholarGoogle ScholarCross RefCross Ref
  50. [50] Tatarchenko Maxim, Dosovitskiy Alexey, and Brox Thomas. 2017. Octree generating networks: Efficient convolutional architectures for high-resolution 3D outputs. In Proceedings of the IEEE International Conference on Computer Vision. Venice. IEEE. 20882096.Google ScholarGoogle ScholarCross RefCross Ref
  51. [51] Tulsiani Shubham, Efros Alexei A., and Malik Jitendra. 2018. Multi-view consistency as supervisory signal for learning shape and pose prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City. IEEE. 28972905.Google ScholarGoogle ScholarCross RefCross Ref
  52. [52] Tulsiani Shubham, Zhou Tinghui, Efros Alexei A., and Malik Jitendra. 2017. Multi-view supervision for single-view reconstruction via differentiable ray consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 26262634.Google ScholarGoogle ScholarCross RefCross Ref
  53. [53] Wang Hanqing, Yang Jiaolong, Liang Wei, and Tong Xin. 2019. Deep single-view 3D object reconstruction with visual hull embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33, Honolulu. AAAI, 89418948.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. [54] Wang Jianren and Fang Zhaoyuan. 2020. GSIR: Generalizable 3D shape interpretation and reconstruction. In European Conference on Computer Vision. Virtual. Springer. 498514.Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. [55] Wang Jinglu, Sun Bo, and Lu Yan. 2019. MVPNet: Multi-view point regression networks for 3D object reconstruction from a single image. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33, Honolulu. AAAI, 89498956.Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. [56] Wang Meng, Wang Lingjing, and Fang Yi. 2017. 3DensiNet: A robust neural network architecture towards 3D volumetric object prediction from 2D image. In Proceedings of the ACM International Conference on Multimedia. Mountain View. ACM. 961969.Google ScholarGoogle ScholarDigital LibraryDigital Library
  57. [57] Wang Nanyang, Zhang Yinda, Li Zhuwen, Fu Yanwei, Yu Hang, Liu Wei, Xue Xiangyang, and Jiang Yu-Gang. 2020. Pixel2Mesh: 3D mesh model generation via image guided deformation. IEEE Transactions on Pattern Analysis and Machine Intelligence (2020).Google ScholarGoogle Scholar
  58. [58] Wang Weiyue, Huang Qiangui, You Suya, Yang Chao, and Neumann Ulrich. 2017. Shape inpainting using 3D generative adversarial network and recurrent convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision. Venice. IEEE. 22982306.Google ScholarGoogle ScholarCross RefCross Ref
  59. [59] Westoby Matthew J., Brasington James, Glasser Niel F., Hambrey Michael J., and Reynolds Jennifer M.. 2012. ‘structure-from-motion’ photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology 179 (2012), 300314.Google ScholarGoogle ScholarCross RefCross Ref
  60. [60] Witkin Andrew P.. 1981. Recovering surface shape and orientation from texture. Artificial Intelligence 17, 1-3 (1981), 1745.Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. [61] Wu Jiajun, Wang Yifan, Xue Tianfan, Sun Xingyuan, Freeman Bill, and Tenenbaum Josh. 2017. MarrNet: 3D shape reconstruction via 2.5D sketches. In Advances in Neural Information Processing Systems, Long Beach. MIT Press, 540550.Google ScholarGoogle Scholar
  62. [62] Wu Jiajun, Zhang Chengkai, Xue Tianfan, Freeman Bill, and Tenenbaum Josh. 2016. Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. In Advances in Neural Information Processing Systems. 8290.Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. [63] Wu Jiajun, Zhang Chengkai, Zhang Xiuming, Zhang Zhoutong, Freeman William T., and Tenenbaum Joshua B.. 2018. Learning shape priors for single-view 3D completion and reconstruction. In Proceedings of the European Conference on Computer Vision(Lecture Notes in Computer Science), Vol. 11215, Springer. 673691.Google ScholarGoogle ScholarCross RefCross Ref
  64. [64] Wu Zhirong, Song Shuran, Khosla Aditya, Yu Fisher, Zhang Linguang, Tang Xiaoou, and Xiao Jianxiong. 2015. 3D ShapeNets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston. IEEE. 19121920.Google ScholarGoogle Scholar
  65. [65] Xiang Nan, Wang Li, Jiang Tao, Li Yanran, Yang Xiaosong, and Zhang Jianjun. 2019. Single-image mesh reconstruction and pose estimation via generative normal map. In Proceedings of the International Conference on Computer Animation and Social Agents. Paris. ACM. 7984.Google ScholarGoogle ScholarDigital LibraryDigital Library
  66. [66] Xiang Yu, Mottaghi Roozbeh, and Savarese Silvio. 2014. Beyond Pascal: A benchmark for 3D object detection in the wild. In IEEE Winter Conference on Applications of Computer Vision. Steamboat Springs. IEEE. 7582.Google ScholarGoogle ScholarCross RefCross Ref
  67. [67] Xie Haozhe, Yao Hongxun, Sun Xiaoshuai, Zhou Shangchen, and Zhang Shengping. 2019. Pix2Vox: Context-aware 3D reconstruction from single and multi-view images. In Proceedings of the IEEE International Conference on Computer Vision. Seoul. IEEE. 26902698.Google ScholarGoogle ScholarCross RefCross Ref
  68. [68] Xie Haozhe, Yao Hongxun, Zhang Shengping, Zhou Shangchen, and Sun Wenxiu. 2020. Pix2Vox++: Multi-scale context-aware 3D object reconstruction from single and multiple images. International Journal of Computer Vision 128, 12 (2020), 29192935.Google ScholarGoogle ScholarDigital LibraryDigital Library
  69. [69] Yan Xinchen, Yang Jimei, Yumer Ersin, Guo Yijie, and Lee Honglak. 2016. Perspective transformer nets: Learning single-view 3D object reconstruction without 3D supervision. In Advances in Neural Information Processing Systems, Barcelona. MIT Press, 16961704.Google ScholarGoogle Scholar
  70. [70] Yang Bo, Rosa Stefano, Markham Andrew, Trigoni Niki, and Wen Hongkai. 2018. Dense 3D object reconstruction from a single depth view. IEEE Transactions on Pattern Analysis and Machine Intelligence 41, 12 (2018), 28202834.Google ScholarGoogle ScholarCross RefCross Ref
  71. [71] Yang Guandao, Cui Yin, Belongie Serge, and Hariharan Bharath. 2018. Learning single-view 3D reconstruction with limited pose supervision. In Proceedings of the European Conference on Computer Vision(Lecture Notes in Computer Science), Vol. 11219, Munich. Springer, 90105.Google ScholarGoogle ScholarCross RefCross Ref
  72. [72] Yang Shuo, Xu Min, Xie Haozhe, Perry Stuart, and Xia Jiahao. 2021. Single-view 3D object reconstruction from shape priors in memory. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE. 31523161.Google ScholarGoogle ScholarDigital LibraryDigital Library
  73. [73] Yao Chun-Han, Hung Wei-Chih, Jampani Varun, and Yang Ming-Hsuan. 2021. Discovering 3D parts from image collections. In Proceedings of the IEEE International Conference on Computer Vision. IEEE. 12981–12990.Google ScholarGoogle ScholarCross RefCross Ref
  74. [74] Yao Yuan, Schertler Nico, Rosales Enrique, Rhodin Helge, Sigal Leonid, and Sheffer Alla. 2020. start hereFront2Back: Single view 3D shape reconstruction via front to back prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Virtual. IEEE. 531540.Google ScholarGoogle ScholarCross RefCross Ref
  75. [75] Zhang Chen. 2019. CuFusion2: Accurate and denoised volumetric 3D object reconstruction using depth cameras. IEEE Access 7 (2019), 4988249893.Google ScholarGoogle ScholarCross RefCross Ref
  76. [76] Zhang Chen and Hu Yu. 2017. CuFusion: Accurate real-time camera tracking and volumetric scene reconstruction with a cuboid. Sensors 17, 10 (2017), 2260.Google ScholarGoogle ScholarCross RefCross Ref
  77. [77] Zhou Qian-Yi and Koltun Vladlen. 2015. Depth camera tracking with contour cues. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Boston. IEEE. 632638.Google ScholarGoogle Scholar

Index Terms

  1. Enhanced 3D Shape Reconstruction With Knowledge Graph of Category Concept

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Multimedia Computing, Communications, and Applications
        ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 18, Issue 3
        August 2022
        478 pages
        ISSN:1551-6857
        EISSN:1551-6865
        DOI:10.1145/3505208
        Issue’s Table of Contents

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 4 March 2022
        • Received: 1 December 2021
        • Accepted: 1 September 2021
        • Revised: 1 August 2021
        Published in tomm Volume 18, Issue 3

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      View Full Text

      HTML Format

      View this article in HTML Format .

      View HTML Format
      About Cookies On This Site

      We use cookies to ensure that we give you the best experience on our website.

      Learn more

      Got it!