[1] The original paper's primary result was that the depth of the model was essential for its high performance, which was computationally expensive, but made feasible due to the utilization of graphics processing units (GPUs) during training.
[1] The three formed team SuperVision and submitted AlexNet in the ImageNet Large Scale Visual Recognition Challenge on September 30, 2012.
The architecture influenced a large number of subsequent work in deep learning, especially in applying neural networks to computer vision.
[4] Based on these values, the GPUs were theoretically capable of performing approximately 395 forward passes per second, assuming ideal conditions.
The LeNet-5 (Yann LeCun et al., 1989)[7][8] was trained by supervised learning with backpropagation algorithm, with an architecture that is essentially the same as AlexNet on a small scale.
[9][10] During the 2000s, as GPU hardware improved, some researchers adapted these for general-purpose computing, including neural network training.
", and Jitendra Malik, a sceptic of neural networks, recommended the PASCAL Visual Object Classes challenge.
[17] AlexNet is highly influential, resulting in much subsequent work in using CNNs for computer vision and using GPUs to accelerate deep learning.
The codebase for AlexNet was released under a BSD license, and had been commonly used in neural network research for several subsequent years.