mjbuehler commited on
Commit
3b747f6
1 Parent(s): f759ba1

Upload Phi3VForCausalLMMoE

Browse files
config.json CHANGED
@@ -148,5 +148,6 @@
148
  "torch_dtype": "bfloat16",
149
  "transformers_version": "4.41.1",
150
  "use_cache": true,
 
151
  "vocab_size": 32064
152
  }
 
148
  "torch_dtype": "bfloat16",
149
  "transformers_version": "4.41.1",
150
  "use_cache": true,
151
+ "use_embeddings_in_router": false,
152
  "vocab_size": 32064
153
  }
pytorch_model-00001-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13395a89696b690cb4cdffd5e8bb225e64181a33535417761e830a5ce4672c18
3
  size 4925359194
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7279e01f405abef92c815c034cb6aa2f78a9c8fc4f674b82139309babe7ac68
3
  size 4925359194
pytorch_model-00002-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ec54a2ec76df6fd71f158726797b1744d9dc94f6bfbd3682a0753a260f26d0a
3
  size 4983183042
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c27d331f2bb51f80d4f3115011ec4b50b7c24192fa8c2b1fcc6259a8d40beb5
3
  size 4983183042
pytorch_model-00003-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7dc0f73ed03e0e85ccdfd31c2cdbd01a73bd28cc55978e7d06778a50db952c49
3
  size 4907652664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe8de5bf199c4203b98f99af7e97bc03b7f6f4ffb69db2bde22bc6d88759e51
3
  size 4907652664
pytorch_model-00004-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d15760408a807be3102337cc34d3bd25406bf0d4eec9f6517757bd59b60a6694
3
- size 3535621544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:674033b724ef5248a0ff360fe15b1e2a61ff1414c1b5448aa4e5fab950b87fbf
3
+ size 3338620328
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 18351511744
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00004-of-00004.bin",
 
1
  {
2
  "metadata": {
3
+ "total_size": 18154510528
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00004-of-00004.bin",