CIFAR

The CIFAR-10 and CIFAR-100 are labeled subsets of the 80 million tiny images dataset. They were collected by Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. The CIFAR-10 dataset consists of 60000 32x32 colour images in 10 classes, with 6000 images per class. There are 50000 training images and 10000 test images. The dataset is divided into five training batches and one test batch, each with 10000 images. The test batch contains exactly 1000 randomly-selected images from each class. The training batches contain the remaining images in random order, but some training batches may contain more images from one class than another. Between them, the training batches contain exactly 5000 images from each class. The CIFAR-100 dataset: This dataset is just like the CIFAR-10, except it has 100 classes containing 600 images each. There are 500 training images and 100 testing images per class. The 100 classes in the CIFAR-100 are grouped into 20 superclasses. Each image comes with a ”fine” label (the class to which it belongs) and a ”coarse” label (the superclass to which it belongs).


References in zbMATH (referenced in 151 articles )

Showing results 81 to 100 of 151.
Sorted by year (citations)
  1. Song, Jingkuan; He, Tao; Gao, Lianli; Xu, Xing; Hanjalic, Alan; Shen, Heng Tao: Unified binary generative adversarial network for image retrieval and compression (2020)
  2. Tan, Hao; He, Cheng; Tang, Dexuan; Cheng, Ran: Efficient evolutionary neural architecture search (NAS) by modular inheritable crossover (2020)
  3. Tonolini, Francesco; Radford, Jack; Turpin, Alex; Faccio, Daniele; Murray-Smith, Roderick: Variational inference for computational imaging inverse problems (2020)
  4. Valle, Marcos Eduardo; Lobo, Rodolfo Anibal: Quaternion-valued recurrent projection neural networks on unit quaternions (2020)
  5. van Engelen, Jesper E.; Hoos, Holger H.: A survey on semi-supervised learning (2020)
  6. Vignotto, Edoardo; Engelke, Sebastian: Extreme value theory for anomaly detection -- the GPD classifier (2020)
  7. Ward, Rachel; Wu, Xiaoxia; Bottou, Leon: AdaGrad stepsizes: sharp convergence over nonconvex landscapes (2020)
  8. Wu, Bijiao; Wang, Dingheng; Zhao, Guangshe; Deng, Lei; Li, Guoqi: Hybrid tensor decomposition in neural network compression (2020)
  9. Wu, Jiasong; Wu, Fuzhi; Yang, Qihan; Zhang, Yan; Liu, Xilin; Kong, Youyong; Senhadji, Lotfi; Shu, Huazhong: Fractional spectral graph wavelets and their applications (2020)
  10. Wu, Min; Wicker, Matthew; Ruan, Wenjie; Huang, Xiaowei; Kwiatkowska, Marta: A game-based approximate verification of deep neural networks with provable guarantees (2020)
  11. Xu, Jian; Liu, Heng; Wu, Dexin; Zhou, Fucai; Gao, Chong-zhi; Jiang, Linzhi: Generating universal adversarial perturbation with ResNet (2020)
  12. Zheng, Qinghe; Tian, Xinyu; Yang, Mingqiang; Wu, Yulin; Su, Huake: PAC-Bayesian framework based drop-path method for 2D discriminative convolutional network pruning (2020)
  13. Zheng, Qinghe; Yang, Mingqiang; Tian, Xinyu; Jiang, Nan; Wang, Deqiang: A full stage data augmentation method in deep convolutional neural network for natural image classification (2020)
  14. Zhou, Dongruo; Xu, Pan; Gu, Quanquan: Stochastic nested variance reduction for nonconvex optimization (2020)
  15. Zou, Difan; Cao, Yuan; Zhou, Dongruo; Gu, Quanquan: Gradient descent optimizes over-parameterized deep ReLU networks (2020)
  16. Altalhi, A. H.; Forcén, J. I.; Pagola, M.; Barrenechea, E.; Bustince, H.; Takáč, Zdenko: Moderate deviation and restricted equivalence functions for measuring similarity between data (2019)
  17. Bo Chang, Minmin Chen, Eldad Haber, Ed H. Chi: AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks (2019) arXiv
  18. Chaudhari, Pratik; Choromanska, Anna; Soatto, Stefano; LeCun, Yann; Baldassi, Carlo; Borgs, Christian; Chayes, Jennifer; Sagun, Levent; Zecchina, Riccardo: Entropy-SGD: biasing gradient descent into wide valleys (2019)
  19. Cherubin, Giovanni: Majority vote ensembles of conformal predictors (2019)
  20. Conor Durkan, Artur Bekasov, Iain Murray, George Papamakarios: Neural Spline Flows (2019) arXiv