A Rapid, End-to-end, Generative Model for Gaseous Phenomena from Limited Views
Sheng Qiu
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorChen Li
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorCorresponding Author
Changbo Wang
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorHong Qin
Department of Computer Science, Stony Brook University, USA
Search for more papers by this authorSheng Qiu
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorChen Li
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorCorresponding Author
Changbo Wang
School of Computer Science and Technology, East China Normal University, Shanghai, China
Search for more papers by this authorHong Qin
Department of Computer Science, Stony Brook University, USA
Search for more papers by this authorS. Qiu and C. Li contribute equally to the article.
Abstract
Despite the rapid development and proliferation of computer graphics hardware devices for scene capture in the most recent decade, the high-resolution 3D/4D acquisition of gaseous scenes (e.g., smokes) in real time remains technically challenging in graphics research nowadays. In this paper, we explore a hybrid approach to simultaneously taking advantage of both the model-centric method and the data-driven method. Specifically, this paper develops a novel conditional generative model to rapidly reconstruct the temporal density and velocity fields of gaseous phenomena based on the sequence of two projection views. With the data-driven method, we can achieve the strong coupling of density update and the estimation of flow motion, as a result, we can greatly improve the reconstruction performance for smoke scenes. First, we employ a conditional generative network to generate the initial density field from input projection views and estimate the flow motion based on the adjacent frames. Second, we utilize the differentiable advection layer and design a velocity estimation network with the long-term mechanism to help achieve the end-to-end training and more stable graphics effects. Third, we can re-simulate the input scene with flexible coupling effects based on the estimated velocity field subject to artists' guidance or user interaction. Moreover, our generative model could accommodate single projection view as input. In practice, more input projection views are enabling and facilitating the high-fidelity reconstruction with more realistic and finer details. We have conducted extensive experiments to confirm the effectiveness, efficiency, and robustness of our new method compared with the previous state-of-the-art techniques.
Supporting Information
Filename | Description |
---|---|
cgf14270-sup-0001-VideoS1.avi13.2 MB | Video S1 |
Please note: The publisher is not responsible for the content or functionality of any supporting information supplied by the authors. Any queries (other than missing content) should be directed to the corresponding author for the article.
References
- [AIH*08] Atcheson B., Ihrke I., Heidrich W., Tevs A., Seidel H. P.: Time-resolved 3d capture of non-stationary gas flows. ACM Transactions on Graphics 27, 5 (2008), 132.
- [CT17] Chu M., Thuerey N.: Data-driven synthesis of smoke flows with cnn-based feature descriptors. ACM Transactions on Graphics 36, 4 (2017), 69.
- [CXM*20] Chu M., Xie Y., Mayer J., Leal-Taixé L., Thuerey N.: Learning temporal coherence via self-supervision for gan-based video generation. ACM Transactions on Graphics 39, 4 (2020), 75–1.
- [DFI*15] Dosovitskiy A., Fischer P., Ilg E., Hausser P., Hazirbas C., Golkov V., Van Der Smagt P., Cremers D., Brox T.: Flownet: Learning optical flow with convolutional networks. In IEEE International Conference on Computer Vision (Santiago, Chile, 2015), pp. 2758–2766.
10.1109/ICCV.2015.316 Google Scholar
- [EHT18] Eckert M.-L., Heidrich W., Thuerey N.: Coupled fluid density and motion from single views. Computer Graphics Forum 37, 8 (2018), 47–58.
- [EUT19] Eckert M., Um K., Thuerey N.: Scalarflow: a large-scale volumetric data set of real-world scalar transport flows for computer animation and machine learning. ACM Transactions on Graphics 38, 6 (2019), 239:1–239:16.
- [FWKH17] Fong J., Wrenninge M., Kulla C., Habel R.: Production volume rendering: Siggraph 2017 course. In ACM SIGGRAPH 2017 Courses (Los Angeles, CA, USA, 2017), ACM, pp. 1–79.
10.1145/3084873.3084907 Google Scholar
- [GGT*18] Gaume J., Gast T. F., Teran J., van Herwijnen A., Jiang C.: Unified modeling of the release and flow of snow avalanches using the material point method. In Geophysical Research Abstracts (2018), vol. 20, pp. 1–5.
- [GITH14] Gregson J., Ihrke I., Thuerey N., Heidrich W.: From capture to simulation - connecting forward and inverse problems in fluids. ACM Transactions on Graphics 33, 4 (2014), 139.
- [GKHH12] Gregson J., Krimerman M., Hullin M. B., Heidrich W.: Stochastic tomography and its applications in 3d imaging of mixing fluids. ACM Transactions on Graphics 31, 4 (2012), 52.
- [GLI16] Guo X., Li W., Iorio F.: Convolutional neural networks for steady flow approximation. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (San Francisco, CA, USA, 2016), ACM, pp. 481–490.
10.1145/2939672.2939738 Google Scholar
- [GNG*13] Gu J., Nayar S. K., Grinspun E., Belhumeur P. N., Ramamoorthi R.: Compressive structured light for recovering inhomogeneous participating media. IEEE Transactions on Pattern Analysis and Machine Intelligence 35, 3 (2013), 555–567.
- [GPB*19] Gissler C., Peer A., Band S., Bender J., Teschner M.: Interlinked sph pressure solvers for strong fluid-rigid coupling. ACM Transactions on Graphics 38, 1 (2019), 5.
- [GPM*14] Goodfellow I. J., Pouget-Abadie J., Mirza M., Xu B., Warde-Farley D., Ozair S., Courville A. C., Bengio Y.: Generative adversarial nets. In Advances in Neural Information Processing Systems (Montreal, Quebec, Canada, 2014), pp. 2672–2680.
- [GZL*20] Gao Y., Zhang Q., Li S., Hao A., Qin H.: Accelerating liquid simulation with an improved data-driven method. Computer Graphics Forum 39, 6 (2020), 180–191.
- [HK07] Hasinoff S. W., Kutulakos K. N.: Photo-consistent reconstruction of semitransparent scenes by density-sheet decomposition. IEEE Transactions on Pattern Analysis and Machine Intelligence 29, 5 (2007), 870–885.
- [HTCL18] Hui T.-W., Tang X., Change Loy C.: Liteflownet: A lightweight convolutional neural network for optical flow estimation. In IEEE Conference on Computer Vision and Pattern Recognition (Salt Lake City, UT, USA, 2018), pp. 8981–8989.
10.1109/CVPR.2018.00936 Google Scholar
- [HW65] Harlow F. H., Welch J. E.: Numerical calculation of time-dependent viscous incompressible flow of fluid with free surface. The Physics of Fluids 8, 12 (1965), 2182–2189.
- [HZRS16] He K., Zhang X., Ren S., Sun J.: Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV, USA, 2016), pp. 770–778.
10.1109/CVPR.2016.90 Google Scholar
- [IMS*17] Ilg E., Mayer N., Saikia T., Keuper M., Dosovitskiy A., Brox T.: Flownet 2.0: Evolution of optical flow estimation with deep networks. In IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI, USA, 2017), pp. 2462–2470.
10.1109/CVPR.2017.179 Google Scholar
- [IOS*14] Ihmsen M., Orthmann J., Solenthaler B., Kolb A., Teschner M.: SPH fluids in computer graphics. In Eurographics (State of the Art Reports) (Strasbourg, France, 2014), pp. 21–42.
- [Jak10] Jakob W.: Mitsuba renderer, 2010. http://www.mitsuba-renderer.org.
- [KAGS19] Kim B., Azevedo V. C., Gross M., Solenthaler B.: Transport-based neural style transfer for smoke simulations. ACM Transactions on Graphics 38, 6 (2019), 1–11.
- [KAT*19] Kim B., Azevedo V. C., Thuerey N., Kim T., Gross M., Solenthaler B.: Deep fluids: A generative network for parameterized fluid simulations. Computer Graphics Forum 38, 2 (2019), 59–70.
- [KTJG08] Kim T., Thürey N., James D., Gross M.: Wavelet turbulence for fluid simulation. ACM Transactions on Graphics 27, 3 (2008), 50.
- [LCZ*15] Li F.-M., Chen X.-W., Zhou B., Lu F.-X., Guo K., Fu Q.: Monocular video guided garment simulation. Journal of Computer Science and Technology 30, 3 (2015), 528–539.
- [LHM*18] Liang S., Huang X., Meng X., Chen K., Shapiro L. G., Kemelmacher-Shlizerman I.: Video to fully automatic 3d hair model. ACM Transactions on Graphics 37, 6 (2018), 206.
- [LQWQ20] Li C., Qiu S., Wang C., Qin H.: Learning physical parameters and detail enhancement for gaseous scene design based on data guidance. IEEE Transactions on Visualization and Computer Graphics (2020), 1–1.
- [LSW*20] Luo R., Shao T., Wang H., Xu W., Chen X., Zhou K., Yang Y.: Nnwarp: Neural network-based nonlinear deformation. IEEE Transactions on Visualization and Computer Graphics 26, 4 (2020), 1745–1759.
- [LTH*17] Ledig C., Theis L., Huszar F., Caballero J., Cunningham A., Acosta A., Aitken A. P., Tejani A., Totz J., Wang Z., Shi W.: Photo-realistic single image super-resolution using a generative adversarial network. In IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI, USA, 2017), pp. 105–114.
10.1109/CVPR.2017.19 Google Scholar
- [LTLH19] Liu Z., Tang H., Lin Y., Han S.: Point-voxel cnn for efficient 3d deep learning. In Advances in Neural Information Processing Systems (Vancouver, BC, Canada, 2019), pp. 965–975.
- [LWLH17] Li Y., Wang N., Liu J., Hou X.: Demystifying neural style transfer. In Proceedings of the 26th International Joint Conference on Artificial Intelligence (Melbourne, Australia, 2017), pp. 2230–2236.
10.24963/ijcai.2017/310 Google Scholar
- [MGMHGV*19] Martínez-González A., Moreno-Hernández D., Guerrero-Viramontes J. A., León-Rodríguez M., Zamarripa-Ramírez J. C. I., Carrillo-Delgado C.: Temperature measurement of fluid flows by using a focusing schlieren method. Sensors 19, 1 (2019), 12.
- [MSY*17] Meng C., Sun M., Yang J., Qiu M., Gu Y.: Training deeper models by gpu memory optimization on tensorflow. In Proceedings of ML Systems Workshop in NIPS (Long Beach, CA, USA, 2017), vol. 7.
- [NB16] Nielsen M. B., Bridson R.: Spatially adaptive flip fluid simulations in bifrost. In ACM SIGGRAPH 2016 Talks (Anaheim, CA, USA, 2016), ACM, p. 41.
10.1145/2897839.2927399 Google Scholar
- [ODAO15] Okabe M., Dobashi Y., Anjyo K., Onai R.: Fluid volume modeling from sparse multi-view images by appearance transfer. ACM Transactions on Graphics 34, 4 (2015), 93:1–93:10.
- [OLL18] Oh Y., Lee T. M., Lee I.: Hierarchical cloth simulation using deep neural networks. In Computer Graphics International (Bintan Island, Indonesia, 2018), pp. 139–146.
10.1145/3208159.3208162 Google Scholar
- [RB17] Ranjan A., Black M. J.: Optical flow estimation using a spatial pyramid network. In IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI, USA, 2017), pp. 4161–4170.
10.1109/CVPR.2017.291 Google Scholar
- [RFB15] Ronneberger O., Fischer P., Brox T.: U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention (Munich, Germany, 2015), vol. 9351 of Lecture Notes in Computer Science, Springer, pp. 234–241.
10.1007/978-3-319-24574-4_28 Google Scholar
- [RGJ*15] Ram D., Gast T., Jiang C., Schroeder C., Stomakhin A., Teran J., Kavehpour P.: A material point method for viscoelastic fluids, foams and sponges. In Proceedings of the 14th ACM SIGGRAPH/Eurographics Symposium on Computer Animation (Los Angeles, CA, USA, 2015), ACM, pp. 157–163.
10.1145/2786784.2786798 Google Scholar
- [SFK*08] Selle A., Fedkiw R., Kim B., Liu Y., Rossignac J.: An unconditionally stable maccormack method. Journal of Scientific Computing 35, 2-3 (2008), 350–371.
- [Sta99] Stam J.: Stable fluids. In SIGGRAPH (Los Angeles, CA, USA, 1999), ACM, pp. 121–128.
10.1145/311535.311548 Google Scholar
- [STD11] Sportouche H., Tupin F., Denise L.: Extraction and three-dimensional reconstruction of isolated buildings in urban scenes from high-resolution optical and sar spaceborne images. IEEE Transactions on Geoscience and Remote Sensing 49, 10 (2011), 3932–3946.
- [TLK16] Teng Y., Levin D. I. W., Kim T.: Eulerian solid-fluid coupling. ACM Transactions on Graphics 35, 6 (2016), 200.
- [TSSP17] Tompson J., Schlachter K., Sprechmann P., Perlin K.: Accelerating eulerian fluid simulation with convolutional networks. In Proceedings of the 34th International Conference on Machine Learning (Sydney, NSW, Australia, 2017), vol. 70 of Proceedings of Machine Learning Research, PMLR, pp. 3424–3433.
- [TYL17] Tran L., Yin X., Liu X.: Disentangled representation learning gan for pose-invariant face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI, USA, 2017), pp. 1415–1424.
10.1109/CVPR.2017.141 Google Scholar
- [UHT18] Um K., Hu X., Thuerey N.: Liquid splash modeling with neural networks. Computer Graphics Forum 37, 8 (2018), 171–182.
- [WKA*20] Wiewel S., Kim B., Azevedo V. C., Solenthaler B., Thuerey N.: Latent space subdivision: stable and controllable time predictions for fluid flow. Computer Graphics Forum 39, 8 (2020), 15–25.
- [WLZ*09] Wang H., Liao M., Zhang Q., Yang R., Turk G.: Physically guided liquid surface modeling from videos. ACM Transactions on Graphics 28, 3 (2009), 90.
- [WXCT19] Werhahn M., Xie Y., Chu M., Thuerey N.: A multi-pass GAN for fluid flow super-resolution. Proceedings of the ACM on Computer Graphics and Interactive Techniques 2, 2 (2019), 10:1–10:21.
10.1145/3340251 Google Scholar
- [XFCT18] Xie Y., Franz E., Chu M., Thuerey N.: Tempogan: A temporally coherent, volumetric gan for super-resolution fluid flow. ACM Transactions on Graphics 37, 4 (2018), 95.
- [XZH*19] Xie X., Zhai X., Hou F., Hao A., Qin H.: Multitask learning on monocular water images: Surface reconstruction and image synthesis. Computer Animation and Virtual Worlds 30, 3-4 (2019), e1896.
- [YLX*15] Yang L., Li S., Xia Q., Qin H., Hao A.: A novel integrated analysis-and-simulation approach for detail enhancement in flip fluid interaction. In Virtual Reality Software and Technology (Beijing, China, 2015), ACM, pp. 103–112.
- [YZX*19] Yuan F., Zhang L., Xia X., Huang Q., Li X.: A wave-shaped deep neural network for smoke density estimation. IEEE Transactions on Image Processing 29 (2019), 2301–2313.
- [ZBG15] Zhang X., Bridson R., Greif C.: Restoring the missing vorticity in advection-projection fluid solvers. ACM Transactions on Graphics 34, 4 (2015), 52.
- [ZLC*13] Zhu B., Lu W., Cong M., Kim B., Fedkiw R.: A new grid structure for domain extension. ACM Transactions on Graphics 32, 4 (2013), 63.
- [ZLP*18] Zheng Y., Liu Y., Pitre J. J., Bull J. L., Gurm H. S., Shih A. J.: Computational fluid dynamics modeling of the burr orbital motion in rotational atherectomy with particle image velocimetry validation. Annals of biomedical engineering 46, 4 (2018), 567–578.
- [ZXL17] Zhang H., Xu T., Li H.: Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In IEEE International Conference on Computer Vision (Venice, Italy, 2017), pp. 5908–5916.
10.1109/ICCV.2017.629 Google Scholar
- [ZYWL15] Zhang S., Yang X., Wu Z., Liu H.: Position-based fluid control. In Interactive 3D Graphics and Games (San Francisco, CA, USA, 2015), ACM, pp. 61–68.
10.1145/2699276.2699287 Google Scholar
- [ZZL*16] Zhou X., Zhu M., Leonardos S., Derpanis K. G., Daniilidis K.: Sparseness meets deepness: 3d human pose estimation from monocular video. In IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV, USA, 2016), pp. 4966–4975.
10.1109/CVPR.2016.537 Google Scholar