forked from pytorch/ao
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Llama3.1 with torchtune (pytorch#1123)
* added model source and type for torchtune flamingo support * added model source and type for torchtune flamingo support * grab missing enum * fix ModelArgs init * create init func for ModelArgs for BC * update pipeline for ModleSource and ModelType * revert lintrunner update on ET * introduce flamingo modules form torchtune * back up to move to linux * mitigate building issue * pass local test * structual model builder * update torchtune address * update install requirement * support new torchtune flamingo component * specific version for vision and ao * unify text-only model generation pipeline * convert installation back and bypass torchtune * restructual model definition * update exportation variable name * remove redunctant function * 1/n torchtune 3.1 8b * installation update * torchtune 3.1 8b / 30b * bring torchchat llama3.1 back * bring tok vali back to torchchat model + revert install_requirements.sh * solve bugs related to tt model support * bypass torchtune import issue * solve Jack's wonderful comments * remveo extra dot * add type.Callable * fix torchchat typos * solve bug when args.model is None * support builder_args.params_table is None * remove all .DS_Store * bring gguf back * remove reduntant updates * bring checkpoint back * debug * debug * debug * new factory func to produce Model from modelargs * solve comments
- Loading branch information
Showing
7 changed files
with
173 additions
and
56 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,15 @@ | ||
{ | ||
"model_type": "llama3_1", | ||
"text": { | ||
"vocab_size": 128256, | ||
"num_layers": 80, | ||
"num_heads": 64, | ||
"num_kv_heads": 8, | ||
"embed_dim": 8192, | ||
"max_seq_len": 8192, | ||
"intermediate_dim": 28672, | ||
"attn_dropout": 0.0, | ||
"norm_eps": 1e-5, | ||
"rope_base": 500000.0 | ||
} | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,15 @@ | ||
{ | ||
"model_type": "llama3_1", | ||
"text": { | ||
"vocab_size": 128256, | ||
"num_layers": 32, | ||
"num_heads": 32, | ||
"num_kv_heads": 8, | ||
"embed_dim": 4096, | ||
"max_seq_len": 8192, | ||
"intermediate_dim": 14336, | ||
"attn_dropout": 0.0, | ||
"norm_eps": 1e-5, | ||
"rope_base": 500000.0 | ||
} | ||
} |
Oops, something went wrong.