Difference between revisions of "Attention-based Neural Networks for Handwriting Recognition"
(Added documentation video link) |
(Added summer notes) |
||
Line 18: | Line 18: | ||
- Distill knowledge from currently cited papers in proposal | - Distill knowledge from currently cited papers in proposal | ||
− | == | + | == Summer 2021 Notes == |
+ | |||
+ | Pretraining run notes: | ||
+ | |||
+ | model_2021-06-18_11_53_47 is first successful run, using ReLU on fully connected layers. 4x image subsampling on 32x64 basic unit means that letters don't fill much of the space. | ||
+ | |||
+ | model_2021-06-22_16_58_03 implements random pixel sampling to handle images that don't fit in memory. 2x image subsampling makes letters bigger, but decreases random sampling fraction. |
Revision as of 12:49, 23 June 2021
Documentation
Recording of working system: https://smith.zoom.us/rec/share/841_Ne3snhwP3mSduKZu63ctTFzYvdDdCrwsdPvCQWOAFDxka9tsdDTwGGZM3fWw.n5T9sxD24vdlCBzQ Passcode: iZQc4=5s
Fall 2020
This honors thesis aims to improve current handwriting recognition through refining the use of attention mechanisms in sequence-to-sequence and Transformer models in HTR systems.
Week 1: 09/04 - 09/10
Goals:
- Install PyTorch and get something running
- Find a good starting point - Review more literature (particularly for sequence-to-sequence models) - Distill knowledge from currently cited papers in proposal
Summer 2021 Notes
Pretraining run notes:
model_2021-06-18_11_53_47 is first successful run, using ReLU on fully connected layers. 4x image subsampling on 32x64 basic unit means that letters don't fill much of the space.
model_2021-06-22_16_58_03 implements random pixel sampling to handle images that don't fit in memory. 2x image subsampling makes letters bigger, but decreases random sampling fraction.