Fill-Mask
Transformers
PyTorch
Safetensors
English
perceiver
nielsr HF staff commited on
Commit
260b471
1 Parent(s): 1609e2e

Update PyTorch model

Browse files
Files changed (2) hide show
  1. config.json +5 -1
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -3,22 +3,26 @@
3
  "PerceiverForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
6
  "cross_attention_shape_for_attention": "kv",
7
  "cross_attention_widening_factor": 1,
8
  "d_latents": 1280,
9
  "d_model": 768,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
 
12
  "initializer_range": 0.02,
13
  "layer_norm_eps": 1e-12,
14
  "model_type": "perceiver",
15
  "num_blocks": 1,
16
  "num_cross_attention_heads": 8,
 
17
  "num_latents": 256,
18
  "num_self_attends_per_block": 26,
19
  "num_self_attention_heads": 8,
20
  "position_embedding_init_scale": 0.02,
21
  "qk_channels": 256,
 
22
  "self_attention_widening_factor": 1,
23
  "seq_len": 2048,
24
  "torch_dtype": "float32",
@@ -26,7 +30,7 @@
26
  368,
27
  496
28
  ],
29
- "transformers_version": "4.11.0.dev0",
30
  "use_query_residual": true,
31
  "v_channels": 1280,
32
  "vocab_size": 262
 
3
  "PerceiverForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
+ "audio_samples_per_frame": 1920,
7
  "cross_attention_shape_for_attention": "kv",
8
  "cross_attention_widening_factor": 1,
9
  "d_latents": 1280,
10
  "d_model": 768,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
+ "image_size": 56,
14
  "initializer_range": 0.02,
15
  "layer_norm_eps": 1e-12,
16
  "model_type": "perceiver",
17
  "num_blocks": 1,
18
  "num_cross_attention_heads": 8,
19
+ "num_frames": 16,
20
  "num_latents": 256,
21
  "num_self_attends_per_block": 26,
22
  "num_self_attention_heads": 8,
23
  "position_embedding_init_scale": 0.02,
24
  "qk_channels": 256,
25
+ "samples_per_patch": 16,
26
  "self_attention_widening_factor": 1,
27
  "seq_len": 2048,
28
  "torch_dtype": "float32",
 
30
  368,
31
  496
32
  ],
33
+ "transformers_version": "4.12.0.dev0",
34
  "use_query_residual": true,
35
  "v_channels": 1280,
36
  "vocab_size": 262
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:405418b61079daa8780df29d31fa88cc8edb2e36ad2b53543c65fe0675df965b
3
- size 804615663
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:941759e1c7489c5ffc7a922551a29e93581868972007aed48ae40e4a525ecba7
3
+ size 804615599