var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/show?bib=https%3A%2F%2Fbibbase.org%2Fnetwork%2Ffiles%2F5DneMFM8NjZGPWGqX&jsonp=1&noBootstrap=1&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/show?bib=https%3A%2F%2Fbibbase.org%2Fnetwork%2Ffiles%2F5DneMFM8NjZGPWGqX&jsonp=1&noBootstrap=1\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/show?bib=https%3A%2F%2Fbibbase.org%2Fnetwork%2Ffiles%2F5DneMFM8NjZGPWGqX&jsonp=1&noBootstrap=1\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2026\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Automatic scoring system for digitized handwritten answers.\n \n \n \n\n\n \n Asakura, T.; Nguyen, H. T; Truong, T.; Tsuchida, Y.; Yamamoto, T.; Miyazawa, H.; Ly, N. T; Nemoto, M.; Ito, M.; Horie, T.; and others\n\n\n \n\n\n\n IEEE Access. 2026.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{asakura2026automatic,\n  title={Automatic scoring system for digitized handwritten answers},\n  author={Asakura, Tomo and Nguyen, Hung T and Truong, Thanh-Nghia and Tsuchida, Yoichi and Yamamoto, Takahiro and Miyazawa, Hiroshi and Ly, Nam T and Nemoto, Masato and Ito, Masamitsu and Horie, Toshihiko and others},\n  journal={IEEE Access},\n  year={2026},\n  publisher={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Siamese Network-Based Handwritten Pattern Similarity for Few-Shot Automatic Scoring of Very Short Answers.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, H. T.; and Nakagawa, M.\n\n\n \n\n\n\n In In Proceedings of the 15th International Conference on Pattern Recognition Applications and Methods - ICPRAM, pages 197–206, 2026. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2026siamese,\n  title={Siamese Network-Based Handwritten Pattern Similarity for Few-Shot Automatic Scoring of Very Short Answers},\n  author={Ly, Nam Tuan and Nguyen, Hung Tuan and Nakagawa, Masaki},\n  booktitle={In Proceedings of the 15th International Conference on Pattern Recognition Applications and Methods - ICPRAM},\n  pages={197--206},\n  year={2026}\n}\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2025\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n TRH2TQA: table recognition with hierarchical relationships to table question-answering on business table images.\n \n \n \n\n\n \n Jirachanchaisiri, P.; Ly, N. T.; and Takasu, A.\n\n\n \n\n\n\n In 2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 8844–8852, 2025. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{jirachanchaisiri2025trh2tqa,\n  title={TRH2TQA: table recognition with hierarchical relationships to table question-answering on business table images},\n  author={Jirachanchaisiri, Pongsakorn and Ly, Nam Tuan and Takasu, Atsuhiro},\n  booktitle={2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)},\n  pages={8844--8852},\n  year={2025},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Automated Recognition and Scoring of Handwritten Short Answer: Insights from Japanese Elementary and Junior High Schools.\n \n \n \n\n\n \n Nguyen, H. T.; Truong, T.; Ly, N. T.; Nakagawa, M.; and Horie, T.\n\n\n \n\n\n\n In International Conference on Document Analysis and Recognition, pages 281–293, 2025. Springer Nature Switzerland Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{nguyen2025automated,\n  title={Automated Recognition and Scoring of Handwritten Short Answer: Insights from Japanese Elementary and Junior High Schools},\n  author={Nguyen, Hung Tuan and Truong, Thanh-Nghia and Ly, Nam Tuan and Nakagawa, Masaki and Horie, Toshihiko},\n  booktitle={International Conference on Document Analysis and Recognition},\n  pages={281--293},\n  year={2025},\n  organization={Springer Nature Switzerland Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2024\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Local Attention-Based Multitask Learning Network for Table Image Recognition.\n \n \n \n\n\n \n Ly, N. T.; Takasu, A.; and Nakagawa, M.\n\n\n \n\n\n\n Available at SSRN 4826329. 2024.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ly2024local,\n  title={Local Attention-Based Multitask Learning Network for Table Image Recognition},\n  author={Ly, Nam Tuan and Takasu, Atsuhiro and Nakagawa, Masaki},\n  journal={Available at SSRN 4826329},\n  year={2024}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Content-Based Similarity for Automatic Scoring of Handwritten Descriptive Answers.\n \n \n \n\n\n \n Truong, N. T.; Nguyen, H. T.; Ly, N. T.; Horie, T.; and Nakagawa, M.\n\n\n \n\n\n\n In International Conference on Document Analysis and Recognition, pages 268–281, 2024. Springer Nature Switzerland Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{truong2024content,\n  title={Content-Based Similarity for Automatic Scoring of Handwritten Descriptive Answers},\n  author={Truong, Nghia Thanh and Nguyen, Hung Tuan and Ly, Nam Tuan and Horie, Toshihiko and Nakagawa, Masaki},\n  booktitle={International Conference on Document Analysis and Recognition},\n  pages={268--281},\n  year={2024},\n  organization={Springer Nature Switzerland Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Two Experiments for automatic scoring of handwritten descriptive answers.\n \n \n \n\n\n \n Nakagawa, M.; Nguyen, H. T.; Truong, N. T.; Ly, N. T.; Nguyen, C. T.; Oka, H.; Ishioka, T.; Asakura, T.; Miyazawa, H.; Yamamoto, T.; and others\n\n\n \n\n\n\n In International Workshop on Document Analysis Systems, pages 3–19, 2024. Springer Nature Switzerland Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{nakagawa2024two,\n  title={Two Experiments for automatic scoring of handwritten descriptive answers},\n  author={Nakagawa, Masaki and Nguyen, Hung Tuan and Truong, Nghia Thanh and Ly, Nam Tuan and Nguyen, Cuong Tuan and Oka, Haruki and Ishioka, Tsunenori and Asakura, Tomo and Miyazawa, Hiroshi and Yamamoto, Takahiro and others},\n  booktitle={International Workshop on Document Analysis Systems},\n  pages={3--19},\n  year={2024},\n  organization={Springer Nature Switzerland Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Two Experiments for Automatic Scoring of Handwritten Descriptive Answers.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, C. T.; Oka, H.; Ishioka, T.; Asakura, T.; Miyazawa, H.; Yamamoto, T.; Horie, T.; and Yasuno, F.\n\n\n \n\n\n\n In Document Analysis Systems: 16th IAPR International Workshop, DAS 2024, Athens, Greece, August 30–31, 2024, Proceedings, pages 3, 2024. Springer Nature\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2024two,\n  title={Two Experiments for Automatic Scoring of Handwritten Descriptive Answers},\n  author={Ly, Nam Tuan and Nguyen, Cuong Tuan and Oka, Haruki and Ishioka, Tsunenori and Asakura, Tomo and Miyazawa, Hiroshi and Yamamoto, Takahiro and Horie, Toshihiko and Yasuno, Fumiko},\n  booktitle={Document Analysis Systems: 16th IAPR International Workshop, DAS 2024, Athens, Greece, August 30--31, 2024, Proceedings},\n  pages={3},\n  year={2024},\n  organization={Springer Nature}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2023\n \n \n (6)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Towards a biosignatures image detection system for planetary exploration with UAVs.\n \n \n \n\n\n \n Galvez-Serna, J.; Ly, P. N.; Furlan, F.; Zepeda, V.; Vanegas, F.; Flannery, D. T.; and Gonzalez, F.\n\n\n \n\n\n\n In 2023 IEEE Aerospace Conference, pages 1–14, 2023. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{galvez2023towards,\n  title={Towards a biosignatures image detection system for planetary exploration with UAVs},\n  author={Galvez-Serna, Julian and Ly, Phuong Nam and Furlan, Federico and Zepeda, Vanessa and Vanegas, Fernando and Flannery, David Timothy and Gonzalez, Felipe},\n  booktitle={2023 IEEE Aerospace Conference},\n  pages={1--14},\n  year={2023},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Rethinking image-based table recognition using weakly supervised methods.\n \n \n \n\n\n \n Ly, N. T.; Takasu, A.; Nguyen, P.; and Takeda, H.\n\n\n \n\n\n\n arXiv preprint arXiv:2303.07641. 2023.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ly2023rethinking,\n  title={Rethinking image-based table recognition using weakly supervised methods},\n  author={Ly, Nam Tuan and Takasu, Atsuhiro and Nguyen, Phuc and Takeda, Hideaki},\n  journal={arXiv preprint arXiv:2303.07641},\n  year={2023}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n An end-to-end multi-task learning model for image-based table recognition.\n \n \n \n\n\n \n Ly, N. T.; and Takasu, A.\n\n\n \n\n\n\n arXiv preprint arXiv:2303.08648. 2023.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ly2023end,\n  title={An end-to-end multi-task learning model for image-based table recognition},\n  author={Ly, Nam Tuan and Takasu, Atsuhiro},\n  journal={arXiv preprint arXiv:2303.08648},\n  year={2023}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n TabIQA: Table Questions Answering on Business Document Images.\n \n \n \n\n\n \n Nguyen, P.; Ly, N. T.; Takeda, H.; and Takasu, A.\n\n\n \n\n\n\n arXiv preprint arXiv:2303.14935. 2023.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{nguyen2023tabiqa,\n  title={TabIQA: Table Questions Answering on Business Document Images},\n  author={Nguyen, Phuc and Ly, Nam Tuan and Takeda, Hideaki and Takasu, Atsuhiro},\n  journal={arXiv preprint arXiv:2303.14935},\n  year={2023}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Digitalizing educational workbooks and collecting handwritten answers for automatic scoring.\n \n \n \n\n\n \n Asakura, T.; Nguyen, H.; Truong, N.; Ly, N. T.; Nguyen, C. T.; Miyazawa, H.; Tsuchida, Y.; Yamamoto, T.; Ito, M.; Horie, T.; and others\n\n\n \n\n\n\n In iTextbooks@ AIED, pages 78–87, 2023. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{asakura2023digitalizing,\n  title={Digitalizing educational workbooks and collecting handwritten answers for automatic scoring.},\n  author={Asakura, Tomo and Nguyen, Hung and Truong, Nghia and Ly, Nam Tuan and Nguyen, Cuong Tuan and Miyazawa, Hiroshi and Tsuchida, Yoichi and Yamamoto, Takahiro and Ito, Masamitsu and Horie, Toshihiko and others},\n  booktitle={iTextbooks@ AIED},\n  pages={78--87},\n  year={2023}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n An end-to-end local attention based model for table recognition.\n \n \n \n\n\n \n Ly, N. T.; and Takasu, A.\n\n\n \n\n\n\n In International Conference on Document Analysis and Recognition, pages 20–36, 2023. Springer Nature Switzerland Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2023end,\n  title={An end-to-end local attention based model for table recognition},\n  author={Ly, Nam Tuan and Takasu, Atsuhiro},\n  booktitle={International Conference on Document Analysis and Recognition},\n  pages={20--36},\n  year={2023},\n  organization={Springer Nature Switzerland Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2022\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Real-time segmentation of desiccation cracks onboard UAVs for planetary exploration.\n \n \n \n\n\n \n Galvez-Serna, J.; Mandel, N.; Sandino, J.; Vanegas, F.; Ly, N.; Flannery, D. T.; and Gonzalez, F.\n\n\n \n\n\n\n In 2022 IEEE Aerospace Conference (AERO), pages 1–12, 2022. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{galvez2022real,\n  title={Real-time segmentation of desiccation cracks onboard UAVs for planetary exploration},\n  author={Galvez-Serna, Julian and Mandel, Nicolas and Sandino, Juan and Vanegas, Fernando and Ly, Nam and Flannery, David Timothy and Gonzalez, Felipe},\n  booktitle={2022 IEEE Aerospace Conference (AERO)},\n  pages={1--12},\n  year={2022},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2021\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Recurrent neural network transducer for Japanese and Chinese offline handwritten text recognition.\n \n \n \n\n\n \n Ngo, T. T.; Nguyen, H. T.; Ly, N. T.; and Nakagawa, M.\n\n\n \n\n\n\n In International Conference on Document Analysis and Recognition, pages 364–376, 2021. Springer International Publishing Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ngo2021recurrent,\n  title={Recurrent neural network transducer for Japanese and Chinese offline handwritten text recognition},\n  author={Ngo, Trung Tan and Nguyen, Hung Tuan and Ly, Nam Tuan and Nakagawa, Masaki},\n  booktitle={International Conference on Document Analysis and Recognition},\n  pages={364--376},\n  year={2021},\n  organization={Springer International Publishing Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n 2D self-attention convolutional recurrent network for offline handwritten text recognition.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, H. T.; and Nakagawa, M.\n\n\n \n\n\n\n In International conference on document analysis and recognition, pages 191–204, 2021. Springer International Publishing Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly20212d,\n  title={2D self-attention convolutional recurrent network for offline handwritten text recognition},\n  author={Ly, Nam Tuan and Nguyen, Hung Tuan and Nakagawa, Masaki},\n  booktitle={International conference on document analysis and recognition},\n  pages={191--204},\n  year={2021},\n  organization={Springer International Publishing Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n A self-attention based model for offline handwritten text recognition.\n \n \n \n\n\n \n Ly, N. T.; Ngo, T. T.; and Nakagawa, M.\n\n\n \n\n\n\n In Asian Conference on Pattern Recognition, pages 356–369, 2021. Springer International Publishing Cham\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2021self,\n  title={A self-attention based model for offline handwritten text recognition},\n  author={Ly, Nam Tuan and Ngo, Trung Tan and Nakagawa, Masaki},\n  booktitle={Asian Conference on Pattern Recognition},\n  pages={356--369},\n  year={2021},\n  organization={Springer International Publishing Cham}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Handwritten text recognition by deep neural networks.\n \n \n \n\n\n \n Ly, N. T.\n\n\n \n\n\n\n Ph.D. Thesis, 東京農工大学, 2021.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@phdthesis{ly2021handwritten,\n  title={Handwritten text recognition by deep neural networks},\n  author={Ly, Nam Tuan},\n  year={2021},\n  school={東京農工大学}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2020\n \n \n (3)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n An attention-based row-column encoder-decoder model for text recognition in Japanese historical documents.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n Pattern Recognition Letters, 136: 134–141. 2020.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ly2020attention,\n  title={An attention-based row-column encoder-decoder model for text recognition in Japanese historical documents},\n  author={Ly, Nam Tuan and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  journal={Pattern Recognition Letters},\n  volume={136},\n  pages={134--141},\n  year={2020},\n  publisher={North-Holland}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Hand-Drawn Object Detection for Scoring Wartegg Zeichen Test.\n \n \n \n\n\n \n Ly, N. T.; Liu, L.; Suen, C. Y.; and Nakagawa, M.\n\n\n \n\n\n\n In ICPRAI 2020 - Second International Conference on Pattern Recognition and Artificial Intelligence, pages 109–114, 2020. Springer\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2020hand,\n  title={Hand-Drawn Object Detection for Scoring Wartegg Zeichen Test},\n  author={Ly, Nam Tuan and Liu, Lili and Suen, Ching Y. and Nakagawa, Masaki},\n  booktitle={ICPRAI 2020 - Second International Conference on Pattern Recognition and Artificial Intelligence},\n  pages={109--114},\n  year={2020},\n  organization={Springer}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Attention augmented convolutional recurrent network for handwritten Japanese text recognition.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n In 2020 17th International Conference on Frontiers in Handwriting Recognition (ICFHR), pages 163–168, 2020. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2020attention,\n  title={Attention augmented convolutional recurrent network for handwritten Japanese text recognition},\n  author={Ly, Nam Tuan and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  booktitle={2020 17th International Conference on Frontiers in Handwriting Recognition (ICFHR)},\n  pages={163--168},\n  year={2020},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2019\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Recognition of anomalously deformed kana sequences in Japanese historical documents.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, K. C.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n IEICE TRANSACTIONS on Information and Systems, 102(8): 1554–1564. 2019.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{ly2019recognition,\n  title={Recognition of anomalously deformed kana sequences in Japanese historical documents},\n  author={Ly, Nam Tuan and Nguyen, Kha Cong and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  journal={IEICE TRANSACTIONS on Information and Systems},\n  volume={102},\n  number={8},\n  pages={1554--1564},\n  year={2019},\n  publisher={The Institute of Electronics, Information and Communication Engineers}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Recognition of Japanese historical text lines by an attention-based encoder-decoder and text line generation.\n \n \n \n\n\n \n Le, A. D.; Mochihashi, D.; Masuda, K.; Mima, H.; and Ly, N. T.\n\n\n \n\n\n\n In Proceedings of the 5th International Workshop on Historical Document Imaging and Processing, pages 37–41, 2019. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{le2019recognition,\n  title={Recognition of Japanese historical text lines by an attention-based encoder-decoder and text line generation},\n  author={Le, Anh Duc and Mochihashi, Daichi and Masuda, Katsuya and Mima, Hideki and Ly, Nam Tuan},\n  booktitle={Proceedings of the 5th International Workshop on Historical Document Imaging and Processing},\n  pages={37--41},\n  year={2019}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n An attention-based end-to-end model for multiple text lines recognition in japanese historical documents.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n In 2019 International Conference on Document Analysis and Recognition (ICDAR), pages 629–634, 2019. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2019attention,\n  title={An attention-based end-to-end model for multiple text lines recognition in japanese historical documents},\n  author={Ly, Nam Tuan and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  booktitle={2019 International Conference on Document Analysis and Recognition (ICDAR)},\n  pages={629--634},\n  year={2019},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n 013 Ⅱ 論考編 研究成果の活用と関連研究 変体仮名の認識.\n \n \n \n\n\n \n Hung Tuan Nguyen, N. T. L.; and Kha Cong Nguyen, C. T. N.\n\n\n \n\n\n\n 木簡など出土文字資料の資源化のための機能的情報集約と知の結集,199–202. 2019.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{hung2019013,\n  title={013 Ⅱ 論考編 研究成果の活用と関連研究 変体仮名の認識},\n  author={Hung Tuan Nguyen, Nam Tuan Ly, Kha Cong Nguyen, Cuong Tuan Nguyen, 中川 正樹},\n  journal={木簡など出土文字資料の資源化のための機能的情報集約と知の結集},\n  pages={199--202},\n  year={2019},\n  publisher={独立行政法人国立文化財機構奈良文化財研究所}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2018\n \n \n (1)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Training an end-to-end model for offline handwritten Japanese text recognition by generated synthetic patterns.\n \n \n \n\n\n \n Ly, N. T.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n In 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), pages 74–79, 2018. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2018training,\n  title={Training an end-to-end model for offline handwritten Japanese text recognition by generated synthetic patterns},\n  author={Ly, Nam Tuan and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  booktitle={2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR)},\n  pages={74--79},\n  year={2018},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2017\n \n \n (4)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n Deep convolutional recurrent network for segmentation-free offline handwritten Japanese text recognition.\n \n \n \n\n\n \n Ly, N.; Nguyen, C.; Nguyen, K.; and Nakagawa, M.\n\n\n \n\n\n\n In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), volume 7, pages 5–9, 2017. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ly2017deep,\n  title={Deep convolutional recurrent network for segmentation-free offline handwritten Japanese text recognition},\n  author={Ly, Nam-Tuan and Nguyen, Cuong-Tuan and Nguyen, Kha-Cong and Nakagawa, Masaki},\n  booktitle={2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR)},\n  volume={7},\n  pages={5--9},\n  year={2017},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Attempts to recognize anomalously deformed Kana in Japanese historical documents.\n \n \n \n\n\n \n Nguyen, H. T.; Ly, N. T.; Nguyen, K. C.; Nguyen, C. T.; and Nakagawa, M.\n\n\n \n\n\n\n In Proceedings of the 4th International Workshop on Historical Document Imaging and Processing, pages 31–36, 2017. \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{nguyen2017attempts,\n  title={Attempts to recognize anomalously deformed Kana in Japanese historical documents},\n  author={Nguyen, Hung Tuan and Ly, Nam Tuan and Nguyen, Kha Cong and Nguyen, Cuong Tuan and Nakagawa, Masaki},\n  booktitle={Proceedings of the 4th International Workshop on Historical Document Imaging and Processing},\n  pages={31--36},\n  year={2017}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n Multi-column CNNs for skeleton based human gesture recognition.\n \n \n \n\n\n \n Nguyen, D.; Ly, T.; Truong, T.; and Nguyen, D.\n\n\n \n\n\n\n In 2017 9th International Conference on Knowledge and Systems Engineering (KSE), pages 179–184, 2017. IEEE\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{nguyen2017multi,\n  title={Multi-column CNNs for skeleton based human gesture recognition},\n  author={Nguyen, Dai-Hai and Ly, Tuan-Nam and Truong, Thi-Huong and Nguyen, Duc-Dung},\n  booktitle={2017 9th International Conference on Knowledge and Systems Engineering (KSE)},\n  pages={179--184},\n  year={2017},\n  organization={IEEE}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n AI recognizes anomalously deformed Kana.\n \n \n \n\n\n \n Nam Tuan Ly, K. C. N.\n\n\n \n\n\n\n 2017.\n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@misc{nam2017ai,\n  title={AI recognizes anomalously deformed Kana},\n  author={Nam Tuan Ly, Kha Cong Nguyen, Masaki Nakagawa},\n  journal={https://www.tuat.ac.jp/en/NEWS/research/20171217\\_01.html},\n  year={2017}\n}\n\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n undefined\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n .\n \n \n \n\n\n \n \n\n\n \n\n\n\n In . \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
\n
\n\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n .\n \n \n \n\n\n \n \n\n\n \n\n\n\n In . \n \n\n\n\n
\n\n\n\n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);