You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I see the hyper-parameter table in Llama3 paper, which shows as below.
However, I tried to calculate the number of parameter of models. They were totally different. Today I found when 405B used ffn_hidden_size=53248, the result matched.
Is there sth. wrong with the report or I missed some information that matters.
The text was updated successfully, but these errors were encountered:
I see the hyper-parameter table in Llama3 paper, which shows as below.
However, I tried to calculate the number of parameter of models. They were totally different. Today I found when 405B used ffn_hidden_size=53248, the result matched.
Is there sth. wrong with the report or I missed some information that matters.
The text was updated successfully, but these errors were encountered: