PengBoXiangShang / multigraph_transformer

IEEE TNNLS 2021, transformer, multi-graph transformer, graph, graph classification, sketch recognition, sketch classification, free-hand sketch, official code of the paper "Multi-Graph Transformer for Free-Hand Sketch Recognition"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Transformer training problem

lonelygoatherd opened this issue · comments

Hi, is there any other tricks on training Transformer used in you work? I have used the model to train another graph task, but the "Transformer + Batchnorm +Softmax" seems never convergent?

Thank you for your interest. All details for training are exactly the same as in the paper associated with this repository; there are no additional tricks.

That being said, if you are working on some other graph task which is not based on Sketch, I think it is normal that the exact architecture we proposed may not give the best performance. You will have to play around with the architecture design such as choice of hidden dimension, layers, normalization scheme, etc.