In the following table, the top-x error value in () indicates the result of the project, and - indicates no test.
Model
Dataset
Top-1 error (val)
Top-5 error (val)
AlexNet
ImageNet_1K
36.7%(43.8%)
15.4%(21.3%)
# Download `AlexNet-ImageNet_1K-9df8cd0f.pth.tar` weights to `./results/pretrained_models`
# More detail see `README.md<Download weights>`
python3 ./inference.py
If you find a bug, create a GitHub issue, or even better, submit a pull request. Similarly, if you have questions, simply post them as GitHub issues.
I look forward to seeing what the community does with these models!
Credit
ImageNet Classification with Deep Convolutional Neural Networks
Alex Krizhevsky,Ilya Sutskever,Geoffrey E. Hinton
Abstract
We trained a large, deep convolutional neural network to classify the 1.2 million
high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5%
and 17.0% which is considerably better than the previous state-of-the-art. The
neural network, which has 60 million parameters and 650,000 neurons, consists
of five convolutional layers, some of which are followed by max-pooling layers,
and three fully-connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully-connected
layers we employed a recently-developed regularization method called “dropout”
that proved to be very effective. We also entered a variant of this model in the
ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%,
compared to 26.2% achieved by the second-best entry.
@article{AlexNet,
title = {ImageNet Classification with Deep Convolutional Neural Networks},
author = {Alex Krizhevsky,Ilya Sutskever,Geoffrey E. Hinton},
journal = {nips},
year = {2012}
}
AlexNet-PyTorch
Overview
This repository contains an op-for-op PyTorch reimplementation of ImageNet Classification with Deep Convolutional Neural Networks.
Table of contents
Download weights
Download datasets
Contains MNIST, CIFAR10&CIFAR100, TinyImageNet_200, MiniImageNet_1K, ImageNet_1K, Caltech101&Caltech256 and more etc.
Please refer to
README.md
in thedata
directory for the method of making a dataset.How Test and Train
Both training and testing only need to modify the
config.py
file.Test
model_num_classes
change to1000
.mode
change totest
.model_path
change to./results/pretrained_models/AlexNet-ImageNet_1K-9df8cd0f.pth.tar
.Train model
model_num_classes
change to1000
.mode
change totrain
.exp_name
change toAlexNet-ImageNet_1K
.pretrained_model_path
change to./results/pretrained_models/AlexNet-ImageNet_1K-9df8cd0f.pth.tar
.Resume train model
model_num_classes
change to1000
.mode
change totrain
.exp_name
change toAlexNet-ImageNet_1K
.resume
change to./samples/AlexNet-ImageNet_1K/epoch_xxx.pth.tar
.Result
Source of original paper results: https://proceedings.neurips.cc/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf
In the following table, the top-x error value in
()
indicates the result of the project, and-
indicates no test.Input:
Output:
Contributing
If you find a bug, create a GitHub issue, or even better, submit a pull request. Similarly, if you have questions, simply post them as GitHub issues.
I look forward to seeing what the community does with these models!
Credit
ImageNet Classification with Deep Convolutional Neural Networks
Alex Krizhevsky,Ilya Sutskever,Geoffrey E. Hinton
Abstract
We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0% which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully-connected layers we employed a recently-developed regularization method called “dropout” that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.
[Paper]