MambaVision-B-1K / configuration_mambavision.py
ahatamiz's picture
Upload model
54acb42 verified
raw
history blame contribute delete
768 Bytes
from transformers import PretrainedConfig
class MambaVisionConfig(PretrainedConfig):
model_type = "mambavision"
def __init__(
self,
depths=[3, 3, 10, 5],
num_heads=[2, 4, 8, 16],
window_size=[8, 8, 14, 7],
dim=128,
in_dim=64,
mlp_ratio=4,
drop_path_rate=0.3,
layer_scale=1e-5,
layer_scale_conv=None,
**kwargs,
):
self.depths = depths
self.num_heads = num_heads
self.window_size = window_size
self.dim = dim
self.in_dim = in_dim
self.mlp_ratio = mlp_ratio
self.drop_path_rate = drop_path_rate
self.layer_scale=layer_scale
self.layer_scale_conv=layer_scale_conv
super().__init__(**kwargs)