Skip to content

Latest commit

 

History

History
15 lines (10 loc) · 914 Bytes

README.md

File metadata and controls

15 lines (10 loc) · 914 Bytes

Visual transformer for MNIST handwriting digit classification

  • Here is a short explanation of how we can use a neural network based on the transformer architecture trained on a handwritten digit dataset to classify handwritten digits:
  1. We first break down the handwritten digit image into patches.
  2. We then pass the embedded patches to the transformer architecture.
  3. The transformer architecture learns long-range dependencies between the patches.
  4. The output of the transformer architecture is a vector that represents the handwritten digit image.
  5. We then use a classifier to predict the digit that the handwritten digit image represents.

image

Prediction output

image