We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
大佬们江湖救急,我是新手想问一下ResT中的EMSA自注意力的泛化能力怎么样?能不能在别的transformer网络中替换MSA?EMSA的代码是rest.py中48-67行定义的Attention这个类吗?求告知,谢谢!
The text was updated successfully, but these errors were encountered:
应该不是,除了Attention这个类,还有下边Block中的deforward()中的第一行,Attention中我好像没发现有Residual connect,如果加上Block才是完整的EMSA,还有这个EMSA(v1)和CMT中的LMSHA结构非常像,那个图画得更容易理解一些, 如有不对,请指出。
Sorry, something went wrong.
No branches or pull requests
大佬们江湖救急,我是新手想问一下ResT中的EMSA自注意力的泛化能力怎么样?能不能在别的transformer网络中替换MSA?EMSA的代码是rest.py中48-67行定义的Attention这个类吗?求告知,谢谢!
The text was updated successfully, but these errors were encountered: