{ "act_dim": 6, "action_tanh": true, "activation_function": "relu", "adv_act_dim": 6, "architectures": [ "TrainableDT" ], "attn_pdrop": 0.1, "bos_token_id": 50256, "context_size": 20, "embd_pdrop": 0.1, "eos_token_id": 50256, "hidden_size": 128, "initializer_range": 0.02, "lambda1": 0.01, "lambda2": 1.0, "layer_norm_epsilon": 1e-05, "max_ep_len": 1000, "max_ep_return": 15000, "max_obs_len": 1000, "max_obs_return": 11252.035989716649, "model_type": "decision_transformer", "n_head": 1, "n_inner": null, "n_layer": 3, "n_positions": 1024, "pr_act_dim": 6, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "returns_scale": 1000, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "state_dim": 17, "state_mean": [ -0.04489211615979194, 0.03232611992451431, 0.06034820791204714, -0.1708161758114941, -0.1947702336464969, -0.057516805782271266, 0.09701419538351468, 0.032391781978295245, 11.047338665669947, -0.07997213109949118, -0.3236324481697188, 0.3629688906533557, 0.4232352315326554, 0.4083653650984332, 1.10850102882522, -0.48743752095133835, -0.07375080715086205 ], "state_std": [ 0.04003438547801023, 0.41146983308045787, 0.542171076653844, 0.4154335823272121, 0.23797831126847824, 0.6205308852047869, 0.301055996310825, 0.21742004892182604, 2.2114253628640745, 0.5726970124299755, 1.7259367582375336, 11.845533609872767, 12.067533717608823, 7.052666254859624, 13.50640651853262, 7.197609616290412, 5.027523940529379 ], "torch_dtype": "float32", "transformers_version": "4.29.2", "use_cache": true, "vocab_size": 1, "warmup_steps": 1000 }