Skip to content

Dawars/DocMAE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DocMAE

Unofficial implementation of DocMAE: Document Image Rectification via Self-supervised Representation Learning

https://arxiv.org/abs/2304.10341

TODO

  • Document background segmentation network using U2 net
  • Synthetic data generation for self-supervised pre-training
  • Pre-training
  • Fine-tuning for document rectification (In progress)
  • Evaluation
  • Code clean up and documentation
  • Model release

Demo

Find a jupyter notebook at demo/background_segmentation.ipynb

Data

Pre-training

Pretraining on 200k documents:

MAE

Run training via:

python pretrain.py -c config/config.json Visualize trained model using https://github.com/NielsRogge/Transformers-Tutorials/blob/master/ViTMAE/ViT_MAE_visualization_demo.ipynb

Acknowledgement

Test documents come from DIR300 dataset https://github.com/fh2019ustc/DocGeoNet

About

Unofficial implementation of DocMAE (WIP): Document Image Rectification via Self-supervised Representation Learning

Topics

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages