Skip to Main Content
Quarterly of Applied Mathematics

Quarterly of Applied Mathematics

Online ISSN 1552-4485; Print ISSN 0033-569X

   
 
 

 

A tale of three probabilistic families: Discriminative, descriptive, and generative models


Authors: Ying Nian Wu, Ruiqi Gao, Tian Han and Song-Chun Zhu
Journal: Quart. Appl. Math. 77 (2019), 423-465
MSC (2010): Primary 62M40
DOI: https://doi.org/10.1090/qam/1528
Published electronically: December 31, 2018
MathSciNet review: 3932965
Full-text PDF

Abstract | References | Similar Articles | Additional Information

Abstract: The pattern theory of Grenander is a mathematical framework where patterns are represented by probability models on random variables of algebraic structures. In this paper, we review three families of probability models, namely, the discriminative models, the descriptive models, and the generative models. A discriminative model is in the form of a classifier. It specifies the conditional probability of the class label given the input signal. A descriptive model specifies the probability distribution of the signal, based on an energy function defined on the signal. A generative model assumes that the signal is generated by some latent variables via a transformation. We shall review these models within a common framework and explore their connections. We shall also review the recent developments that take advantage of the high approximation capacities of deep neural networks.


References [Enhancements On Off] (What's this?)

References
  • Guillaume Alain and Yoshua Bengio, What regularized auto-encoders learn from the data-generating distribution, J. Mach. Learn. Res. 15 (2014), 3563–3593. MR 3291406
  • Shun-ichi Amari and Hiroshi Nagaoka, Methods of information geometry, Translations of Mathematical Monographs, vol. 191, American Mathematical Society, Providence, RI; Oxford University Press, Oxford, 2000. Translated from the 1993 Japanese original by Daishi Harada. MR 1800071
  • Martin Arjovsky, Soumith Chintala, and Léon Bottou, Wasserstein gan, arXiv preprint arXiv:1701.07875 (2017).
  • Ian Goodfellow, Yoshua Bengio, and Aaron Courville, Deep learning, Adaptive Computation and Machine Learning, MIT Press, Cambridge, MA, 2016. MR 3617773
  • Julian Besag, Spatial interaction and the statistical analysis of lattice systems, J. Roy. Statist. Soc. Ser. B 36 (1974), 192–236. MR 373208
  • Leo Breiman, Jerome H. Friedman, Richard A. Olshen, and Charles J. Stone, Classification and regression trees, Wadsworth Statistics/Probability Series, Wadsworth Advanced Books and Software, Belmont, CA, 1984. MR 726392
  • Le Chang and Doris Y Tsao, The code for facial identity in the primate brain, Cell 169 (2017), no. 6, 1013–1028.
  • Scott Shaobing Chen, David L. Donoho, and Michael A. Saunders, Atomic decomposition by basis pursuit, SIAM J. Sci. Comput. 20 (1998), no. 1, 33–61. MR 1639094, DOI https://doi.org/10.1137/S1064827596304010
  • Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel, Infogan: Interpretable representation learning by information maximizing generative adversarial nets, Advances in Neural Information Processing Systems, 2016, pp. 2172–2180.
  • Anna Choromanska, Mikael Henaff, Michael Mathieu, Gérard Ben Arous, and Yann LeCun, The loss surface of multilayer networks, arXiv preprint arXiv:1412.0233 (2014).
  • Timothy F Cootes, Gareth J Edwards, and Christopher J Taylor, Active appearance models, IEEE Transactions on Pattern Analysis and Machine Intelligence (2001), no. 6, 681–685.
  • Corinna Cortes and Vladimir Vapnik, Support-vector networks, Machine learning 20 (1995), no. 3, 273–297.
  • Jifeng Dai, Yang Lu, and Ying Nian Wu, Generative modeling of convolutional neural networks, Stat. Interface 9 (2016), no. 4, 485–496. MR 3553376, DOI https://doi.org/10.4310/SII.2016.v9.n4.a8
  • Zihang Dai, Amjad Almahairi, Philip Bachman, Eduard Hovy, and Aaron Courville, Calibrating energy-based generative adversarial networks, arXiv preprint arXiv:1702.01691 (2017).
  • John G Daugman, Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters, JOSA A 2 (1985), no. 7, 1160–1169.
  • Stephen Della Pietra, Vincent Della Pietra, and John Lafferty, Inducing features of random fields, IEEE Transactions on Pattern Analysis and Machine Intelligence (1997), no. 4, 380–393.
  • A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, J. Roy. Statist. Soc. Ser. B 39 (1977), no. 1, 1–38. With discussion. MR 501537
  • Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei, Imagenet: A large-scale hierarchical image database, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2009, pp. 248–255.
  • Laurent Dinh, David Krueger, and Yoshua Bengio, Nice: Non-linear independent components estimation, arXiv preprint arXiv:1410.8516 (2014).
  • Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio, Density estimation using real nvp, arXiv preprint arXiv:1605.08803 (2016).
  • Alexey Dosovitskiy, Jost Tobias Springenberg, and Thomas Brox, Learning to generate chairs with convolutional neural networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 1538–1546.
  • Yoav Freund and Robert E. Schapire, A decision-theoretic generalization of on-line learning and an application to boosting, J. Comput. System Sci. 55 (1997), no. 1, 119–139. Second Annual European Conference on Computational Learning Theory (EuroCOLT ’95) (Barcelona, 1995). MR 1473055, DOI https://doi.org/10.1006/jcss.1997.1504
  • Jerome H. Friedman, Multivariate adaptive regression splines, Ann. Statist. 19 (1991), no. 1, 1–141. With discussion and a rejoinder by the author. MR 1091842, DOI https://doi.org/10.1214/aos/1176347963
  • Ruiqi Gao, Yang Lu, Junpei Zhou, Song-Chun Zhu, and Ying Nian Wu, Learning generative convnets via multi-grid modeling and sampling, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 9155–9164.
  • Stuart Geman and Donald Geman, Stochastic relaxation, gibbs distributions, and the bayesian restoration of images, IEEE Transactions on Pattern Analysis and Machine Intelligence (1997), no. 4, 380–393.
  • Fu Qing Gao, Transition laws, Markov specifications and Markov random fields, J. Math. (Wuhan) 8 (1988), no. 3, 297–306 (Chinese, with English summary). MR 1015198
  • Stuart Geman, Daniel F. Potter, and Zhiyi Chi, Composition systems, Quart. Appl. Math. 60 (2002), no. 4, 707–736. MR 1939008, DOI https://doi.org/10.1090/qam/1939008
  • J. Willard Gibbs, Elementary principles in statistical mechanics: developed with especial reference to the rational foundation of thermodynamics, Dover publications, Inc., New York, 1960. MR 0116523
  • Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio, Generative adversarial nets, Advances in Neural Information Processing Systems, 2014, pp. 2672–2680.
  • Ulf Grenander, A unified approach to pattern analysis, Advances in Computers 10 (1970), 175–216.
  • Ulf Grenander and Michael I. Miller, Pattern theory: from representation to inference, Oxford University Press, Oxford, 2007. MR 2285439
  • Ralph Gross, Iain Matthews, Jeffrey Cohn, Takeo Kanade, and Simon Baker, Multi-pie, Image Vision Comput. 28 (2010), no. 5, 807–813.
  • Cheng-En Guo, Song-Chun Zhu, and Ying Nian Wu, Modeling visual patterns by integrating descriptive and generative methods, International Journal of Computer Vision 53 (2003), no. 1, 5–29.
  • Cheng-en Guo, Song-Chun Zhu, and Ying Nian Wu, Primal sketch: Integrating structure and texture, Computer Vision and Image Understanding 106 (2007), no. 1, 5–19.
  • Tian Han, Yang Lu, Song-Chun Zhu, and Ying Nian Wu, Alternating back-propagation for generator network., AAAI, vol. 3, 2017, p. 13.
  • Tian Han, Erik Nijkamp, Xiaolin Fang, Song-Chun Zhu, and Ying Nian Wu, Divergence triangle for joint training of energy-based model, generator model and inference model, (2018).
  • Tian Han, Jiawen Wu, and Ying Nian Wu, Replicating active appearance model by generator network, International Joint Conferences on Artificial Intelligence, 2018.
  • Tian Han, Xianglei Xing, and Ying Nian Wu, Learning multi-view generator network for shared representation, International Conference on Pattern Recognition, 2018.
  • Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • Geoffrey E. Hinton, Training products of experts by minimizing contrastive divergence., Neural Computation 14 (2002), no. 8, 1771–1800.
  • Geoffrey E. Hinton, Peter Dayan, Brendan J Frey, and Radford M Neal, The "wake-sleep" algorithm for unsupervised neural networks, Science 268 (1995), no. 5214, 1158–1161.
  • Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh, A fast learning algorithm for deep belief nets, Neural Comput. 18 (2006), no. 7, 1527–1554. MR 2224485, DOI https://doi.org/10.1162/neco.2006.18.7.1527
  • Sepp Hochreiter and Jürgen Schmidhuber, Long short-term memory, Neural computation 9 (1997), no. 8, 1735–1780.
  • Yi Hong, Zhangzhang Si, Wenze Hu, Song-Chun Zhu, and Ying Nian Wu, Unsupervised learning of compositional sparse code for natural image representation, Quart. Appl. Math. 72 (2014), no. 2, 373–406. MR 3186243, DOI https://doi.org/10.1090/S0033-569X-2013-01361-5
  • J. J. Hopfield, Neural networks and physical systems with emergent collective computational abilities, Proc. Nat. Acad. Sci. U.S.A. 79 (1982), no. 8, 2554–2558. MR 652033, DOI https://doi.org/10.1073/pnas.79.8.2554
  • Aapo Hyvärinen, Estimation of non-normalized statistical models by score matching, J. Mach. Learn. Res. 6 (2005), 695–709. MR 2249836
  • Aapo Hyvarinen, Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables, IEEE Transactions on neural networks 18 (2007), no. 5, 1529–1531.
  • Aapo Hyvärinen, Juha Karhunen, and Erkki Oja, Independent component analysis, vol. 46, John Wiley & Sons, 2004.
  • Sergey Ioffe and Christian Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, arXiv preprint arXiv:1502.03167 (2015).
  • Long Jin, Justin Lazarow, and Zhuowen Tu, Introspective learning for discriminative classification, Advances in Neural Information Processing Systems, 2017.
  • Bela Julesz, Textons, the elements of texture perception, and their interactions, Nature 290 (1981), no. 5802, 91–97.
  • Taesup Kim and Yoshua Bengio, Deep directed generative models with energy-based probability estimation, arXiv preprint arXiv:1606.03439 (2016).
  • Diederik P. Kingma and Max Welling, Auto-encoding variational bayes, International Conference for Learning Representations (2014).
  • Yehuda Koren, Robert Bell, and Chris Volinsky, Matrix factorization techniques for recommender systems, Computer (2009), no. 8, 30–37.
  • Alex Krizhevsky and Geoffrey E Hinton, Learning multiple layers of features from tiny images, (2009).
  • Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton, Imagenet classification with deep convolutional neural networks, Advances in Neural Information Processing Systems, 2012, pp. 1097–1105.
  • Justin Lazarow, Long Jin, and Zhuowen Tu, Introspective neural networks for generative modeling, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2774–2783.
  • Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE 86 (1998), no. 11, 2278–2324.
  • Yann LeCun, Sumit Chopra, Rata Hadsell, Mare’Aurelio Ranzato, and Fu Jie Huang, A tutorial on energy-based learning, Predicting Structured Data, MIT Press, 2006.
  • Daniel D Lee and H Sebastian Seung, Algorithms for non-negative matrix factorization, Advances in Neural Information Processing Systems, 2001, pp. 556–562.
  • Honglak Lee, Roger Grosse, Rajesh Ranganath, and Andrew Y Ng, Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations, International Conference on Machine Learning, 2009, pp. 609–616.
  • Kwonjoon Lee, Weijian Xu, Fan Fan, and Zhuowen Tu, Wasserstein introspective neural networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018.
  • Min Lin, Qiang Chen, and Shuicheng Yan, Network in network, arXiv preprint arXiv:1312.4400 (2013).
  • Ce Liu, Song-Chun Zhu, and Heung-Yeung Shum, Learning inhomogeneous gibbs model of faces by minimax entropy, International Conference on Computer Vision, 2001, pp. 281–287.
  • Jun S. Liu, Monte Carlo strategies in scientific computing, Springer Series in Statistics, Springer, New York, 2008. MR 2401592
  • Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang, Deep learning face attributes in the wild, International Conference on Computer Vision, 2015, pp. 3730–3738.
  • Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean, Distributed representations of words and phrases and their compositionality, Advances in Neural Information Processing Systems, 2013, pp. 3111–3119.
  • Andriy Mnih and Karol Gregor, Neural variational inference and learning in belief networks, International Conference on Machine Learning, 2014, pp. 1791–1799.
  • Guido F Montufar, Razvan Pascanu, Kyunghyun Cho, and Yoshua Bengio, On the number of linear regions of deep neural networks, Advances in Neural Information Processing Systems, 2014, pp. 2924–2932.
  • David Mumford and Agnès Desolneux, Pattern theory, Applying Mathematics, A K Peters, Ltd., Natick, MA, 2010. The stochastic analysis of real-world signals. MR 2723182
  • Radford M. Neal, MCMC using Hamiltonian dynamics, Handbook of Markov chain Monte Carlo, Chapman & Hall/CRC Handb. Mod. Stat. Methods, CRC Press, Boca Raton, FL, 2011, pp. 113–162. MR 2858447
  • Jiquan Ngiam, Zhenghao Chen, Pang W Koh, and Andrew Y Ng, Learning deep energy models, International Conference on Machine Learning, 2011, pp. 1105–1112.
  • Bruno A Olshausen and David J Field, Sparse coding with an overcomplete basis set: A strategy employed by v1?, Vision Research 37 (1997), no. 23, 3311–3325.
  • Aaron van den Oord, Nal Kalchbrenner, and Koray Kavukcuoglu, Pixel recurrent neural networks, arXiv preprint arXiv:1601.06759 (2016).
  • Pentti Paatero and Unto Tapper, Positive matrix factorization: A non-negative factor model with optimal utilization of error estimates of data values, Environmetrics 5 (1994), no. 2, 111–126.
  • Razvan Pascanu, Guido Montufar, and Yoshua Bengio, On the number of response regions of deep feed forward networks with piece-wise linear activations, arXiv preprint arXiv:1312.6098 (2013).
  • Alec Radford, Luke Metz, and Soumith Chintala, Unsupervised representation learning with deep convolutional generative adversarial networks, arXiv preprint arXiv:1511.06434 (2015).
  • Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra, Stochastic backpropagation and approximate inference in deep generative models, International Conference on Machine Learning, 2014, pp. 1278–1286.
  • Herbert Robbins and Sutton Monro, A stochastic approximation method, Ann. Math. Statistics 22 (1951), 400–407. MR 42668, DOI https://doi.org/10.1214/aoms/1177729586
  • Stefan Roth and Michael J Black, Fields of experts: A framework for learning image priors, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, vol. 2, 2005, pp. 860–867.
  • Sam T Roweis and Lawrence K Saul, Nonlinear dimensionality reduction by locally linear embedding, Science 290 (2000), no. 5500, 2323–2326.
  • Donald B. Rubin, Multiple imputation for nonresponse in surveys, Wiley Classics Library, Wiley-Interscience [John Wiley & Sons], Hoboken, NJ, 2004. Reprint of the 1987 edition [John Wiley & Sons, Inc., New York; MR899519]. MR 2117498
  • Donald B. Rubin and Dorothy T. Thayer, EM algorithms for ML factor analysis, Psychometrika 47 (1982), no. 1, 69–76. MR 668505, DOI https://doi.org/10.1007/BF02293851
  • Ruslan Salakhutdinov and Geoffrey E Hinton, Deep boltzmann machines, International Conference on Artificial Intelligence and Statistics, 2009.
  • Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen, Improved techniques for training gans, Advances in Neural Information Processing Systems, 2016, pp. 2226–2234.
  • H. Sebastian Seung, Learning continuous attractors in recurrent networks, Advances in Neural Information Processing Systems, 1998, pp. 654–660.
  • Zhangzhang Si and Song-Chun Zhu, Learning hybrid image templates (hit) by information projection, IEEE Transactions on Pattern Analysis and Machine Intelligence 99 (2011), no. 7, 1354–1367.
  • Zhangzhang Si and Song-Chun Zhu, Learning and-or templates for object recognition and detection, IEEE Transactions on Pattern Analysis and Machine Intelligence 35 (2013), no. 9, 2189–2205.
  • David Silver, Julian Schrittwieser, Karen Simonyan, Ioannis Antonoglou, Aja Huang, Arthur Guez, Thomas Hubert, Lucas Baker, Matthew Lai, Adrian Bolton, et al., Mastering the game of go without human knowledge, Nature 550 (2017), no. 7676, 354–359.
  • Karen Simonyan and Andrew Zisserman, Very deep convolutional networks for large-scale image recognition, ICLR (2015).
  • Kevin Swersky, Marc’Aurelio Ranzato, David Buchman, Benjamin Marlin, and Nando Freitas, On autoencoders and score matching for energy based models, ICML, ACM, 2011, pp. 1201–1208.
  • Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna, Rethinking the inception architecture for computer vision, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2818–2826.
  • Yee Whye Teh, Max Welling, Simon Osindero, and Geoffrey E. Hinton, Energy-based models for sparse overcomplete representations, J. Mach. Learn. Res. 4 (2004), no. 7-8, 1235–1260. MR 2103628, DOI https://doi.org/10.1162/jmlr.2003.4.7-8.1235
  • Robert Tibshirani, Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc. Ser. B 58 (1996), no. 1, 267–288. MR 1379242
  • Tijmen Tieleman, Training restricted boltzmann machines using approximations to the likelihood gradient, International Conference on Machine Learning, 2008, pp. 1064–1071.
  • Zhuowen Tu, Learning generative models via discriminative approaches, 2007 IEEE Conference on Computer Vision and Pattern Recognition, 2007, pp. 1–8.
  • Zhuowen Tu and Song-Chun Zhu, Image segmentation by data-driven markov chain monte carlo, IEEE Transactions on pattern analysis and machine intelligence 24 (2002), no. 5, 657–673.
  • Zhuowen Tu and Song-Chun Zhu, Parsing images into regions, curves and curve groups, International Journal of Computer Vision 69 (2006), no. 2, 223–249.
  • Pascal Vincent, A connection between score matching and denoising autoencoders, Neural Comput. 23 (2011), no. 7, 1661–1674. MR 2839543, DOI https://doi.org/10.1162/NECO_a_00142
  • Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol, Extracting and composing robust features with denoising autoencoders, International Conference on Machine Learning, 2008, pp. 1096–1103.
  • Max Welling, Herding dynamical weights to learn, International Conference on Machine Learning, 2009, pp. 1121–1128.
  • Ying Nian Wu, Zhangzhang Si, Haifeng Gong, and Song-Chun Zhu, Learning active basis model for object detection and recognition, Int. J. Comput. Vis. 90 (2010), no. 2, 198–235. MR 2719010, DOI https://doi.org/10.1007/s11263-009-0287-0
  • Ying Nian Wu, Song Chun Zhu, and Xiuwen Liu, Equivalence of julesz ensembles and frame models, International Journal of Computer Vision 38 (2000), no. 3, 247–265.
  • Jianwen Xie, Yang Lu, Ruiqi Gao, and Ying Nian Wu, Cooperative learning of energy-based model and latent variable model via mcmc teaching, AAAI, 2018.
  • Jianwen Xie, Yang Lu, Ruiqi Gao, Song-Chun Zhu, and Ying Nian Wu, Cooperative training of descriptor and generator networks, IEEE Transactions on Pattern Analysis and Machine Intelligence (2018), no. preprints.
  • Jianwen Xie, Yang Lu, Song-Chun Zhu, and Ying Nian Wu, A theory of generative convnet, International Conference on Machine Learning, 2016, pp. 2635–2644.
  • Jianwen Xie, Song-Chun Zhu, and Ying Nian Wu, Synthesizing dynamic patterns by spatial-temporal generative convnet, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 7093–7101.
  • Laurent Younes, On the convergence of Markovian stochastic algorithms with rapidly decreasing ergodicity rates, Stochastics Stochastics Rep. 65 (1999), no. 3-4, 177–228. MR 1687636, DOI https://doi.org/10.1080/17442509908834179
  • Fisher Yu, Yinda Zhang, Shuran Song, Ari Seff, and Jianxiong Xiao, LSUN: construction of a large-scale image dataset using deep learning with humans in the loop, CoRR abs/1506.03365 (2015).
  • Matthew D Zeiler, Graham W Taylor, and Rob Fergus, Adaptive deconvolutional networks for mid and high level feature learning, International Conference on Computer Vision, 2011, pp. 2018–2025.
  • Junbo Zhao, Michael Mathieu, and Yann LeCun, Energy-based generative adversarial network, arXiv preprint arXiv:1609.03126 (2016).
  • Bolei Zhou, Agata Lapedriza, Jianxiong Xiao, Antonio Torralba, and Aude Oliva, Learning deep features for scene recognition using places database, Advances in Neural Information Processing Systems, 2014, pp. 487–495.
  • Song-Chun Zhu, Statistical modeling and conceptualization of visual patterns, IEEE Transactions on Pattern Analysis and Machine Intelligence 25 (2003), no. 6, 691–712.
  • Song-Chun Zhu, Cheng-En Guo, Yizhou Wang, and Zijian Xu, What are textons?, International Journal of Computer Vision 62 (2005), no. 1-2, 121–143.
  • Song-Chun Zhu and David Mumford, Grade: Gibbs reaction and diffusion equations., International Conference on Computer Vision, 1998, pp. 847–854.
  • Song-Chun Zhu, David Mumford, et al., A stochastic grammar of images, Foundations and Trends® in Computer Graphics and Vision 2 (2007), no. 4, 259–362.
  • Song-Chun Zhu, Ying Nian Wu, and David Mumford, Minimax entropy principle and its application to texture modeling, Neural Computation 9 (1997), no. 8, 1627–1660.

Similar Articles

Retrieve articles in Quarterly of Applied Mathematics with MSC (2010): 62M40

Retrieve articles in all journals with MSC (2010): 62M40


Additional Information

Ying Nian Wu
Affiliation: Department of Statistics, University of California, Los Angeles, California 90095-1554
MR Author ID: 360276
Email: ywu@stat.ucla.edu

Ruiqi Gao
Affiliation: Department of Statistics, University of California, Los Angeles, California 90095-1554
Email: ruiqigao@ucla.edu

Tian Han
Affiliation: Department of Statistics, University of California, Los Angeles, California 90095-1554
Email: hthth0801@gmail.com

Song-Chun Zhu
Affiliation: Department of Statistics, University of California, Los Angeles, California 90095-1554
MR Author ID: 712282
Email: sczhu@stat.ucla.edu

Received by editor(s): February 21, 2018
Received by editor(s) in revised form: October 9, 2018
Published electronically: December 31, 2018
Additional Notes: The work was supported by NSF DMS 1310391, DARPA SIMPLEX N66001-15-C-4035, ONR MURI N00014-16-1-2007, and DARPA ARO W911NF-16-1-0579.
Article copyright: © Copyright 2018 Brown University