Replies: 2 comments 1 reply
-
Glad that the code is useful! Today, I would go for the implementation in P.S. FYI, this NeurIPS 2022 paper may also be relevant. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Awesome! Thanks for the quick reply - and I have seen the paper, great work! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Thanks for the great repos and for making this code open-source!
I noticed the disclaimer in the README:
It is clear that there are differences in implementation between the Transformer class in this repo and the FTTransformer class in rtdl. I am wondering: what is the practical consequence of these differences?
If we are not interested in reproducing the exact results of the NeurIPS 2021 paper (but instead using FT-Transformer as a baseline for other tasks, where its architecture and hyperparameters will be tuned anyway), is there any reason not to use the (more mature) implementation in
rtdl
?Beta Was this translation helpful? Give feedback.
All reactions