c-aero commited on
Commit
374785e
·
1 Parent(s): ed0960a

Upload train.jsonl

Browse files
Files changed (2) hide show
  1. .gitattributes +1 -0
  2. train.jsonl +3 -1
.gitattributes CHANGED
@@ -53,3 +53,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
 
 
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
56
+ train.jsonl filter=lfs diff=lfs merge=lfs -text
train.jsonl CHANGED
@@ -1 +1,3 @@
1
- {"doi": "1102.0183", "chunk-id": "0", "chunk": "High-Performance Neural Networks\nfor Visual Object Classi\fcation\nDan C. Cire\u0018 san, Ueli Meier, Jonathan Masci,\nLuca M. Gambardella and J\u007f urgen Schmidhuber\nTechnical Report No. IDSIA-01-11\nJanuary 2011\nIDSIA / USI-SUPSI\nDalle Molle Institute for Arti\fcial Intelligence\nGalleria 2, 6928 Manno, Switzerland\nIDSIA is a joint institute of both University of Lugano (USI) and University of Applied Sciences of Southern Switzerland (SUPSI),\nand was founded in 1988 by the Dalle Molle Foundation which promoted quality of life.\nThis work was partially supported by the Swiss Commission for Technology and Innovation (CTI), Project n. 9688.1 IFF:\nIntelligent Fill in Form.arXiv:1102.0183v1 [cs.AI] 1 Feb 2011\nTechnical Report No. IDSIA-01-11 1\nHigh-Performance Neural Networks\nfor Visual Object Classi\fcation\nDan C. Cire\u0018 san, Ueli Meier, Jonathan Masci,\nLuca M. Gambardella and J\u007f urgen Schmidhuber\nJanuary 2011\nAbstract\nWe present a fast, fully parameterizable GPU implementation of Convolutional Neural\nNetwork variants. Our feature extractors are neither carefully designed nor pre-wired, but", "id": "1102.0183", "title": "High-Performance Neural Networks for Visual Object Classification", "summary": "We present a fast, fully parameterizable GPU implementation of Convolutional\nNeural Network variants. Our feature extractors are neither carefully designed\nnor pre-wired, but rather learned in a supervised way. Our deep hierarchical\narchitectures achieve the best published results on benchmarks for object\nclassification (NORB, CIFAR10) and handwritten digit recognition (MNIST), with\nerror rates of 2.53%, 19.51%, 0.35%, respectively. Deep nets trained by simple\nback-propagation perform better than more shallow ones. Learning is\nsurprisingly rapid. NORB is completely trained within five epochs. Test error\nrates on MNIST drop to 2.42%, 0.97% and 0.48% after 1, 3 and 17 epochs,\nrespectively.", "source": "http://arxiv.org/pdf/1102.0183", "authors": ["Dan C. Cire\u015fan", "Ueli Meier", "Jonathan Masci", "Luca M. Gambardella", "J\u00fcrgen Schmidhuber"], "categories": ["cs.AI", "cs.NE"], "comment": "12 pages, 2 figures, 5 tables", "journal_ref": null, "primary_category": "cs.AI", "published": "20110201", "updated": "20110201", "references": []}
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98df176df2a3ed854545c6ef14524aca83991ce694e064ab23f0cfd73f4dc263
3
+ size 14417793