{ "_name_or_path": "/mnt/bn/ridger1/flame/configs/RetNet/retnet_340M_hybrid_6.json", "architectures": [ "RetNetForCausalLM" ], "attn": { "layers": [ 5, 11, 17, 23 ], "num_heads": 16, "num_kv_heads": 8, "qkv_bias": false, "rope_theta": 10000.0, "window_size": null }, "attn_mode": "chunk", "bos_token_id": 1, "clamp_min": null, "conv_size": 4, "elementwise_affine": true, "eos_token_id": 2, "expand_k": 0.5, "expand_v": 1, "feature_map": null, "fuse_cross_entropy": true, "fuse_norm": true, "fuse_swiglu": true, "hidden_act": "swish", "hidden_ratio": 4, "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": null, "max_position_embeddings": 2048, "model_type": "retnet", "norm_eps": 1e-06, "num_heads": 4, "num_hidden_layers": 24, "num_kv_heads": null, "tie_word_embeddings": true, "torch_dtype": "float32", "transformers_version": "4.49.0", "use_cache": true, "use_output_gate": true, "use_short_conv": false, "vocab_size": 32000 }