lirannoc commited on
Commit
fc594cb
·
verified ·
1 Parent(s): ee8c45f

Delete zero_shot

Browse files
Files changed (1) hide show
  1. zero_shot/config.json +0 -39
zero_shot/config.json DELETED
@@ -1,39 +0,0 @@
1
- {
2
- "_name_or_path": "super_linear",
3
- "architectures": [
4
- "SuperLinearForCausalLM"
5
- ],
6
- "auto_map": {
7
- "AutoConfig": "configuration_super_linear.SuperLinearConfig",
8
- "AutoModelForCausalLM": "modeling_super_linear.SuperLinearForCausalLM"
9
- },
10
- "auto_regressive": 1,
11
- "d_model": 128,
12
- "dropout": 0.0,
13
- "fft_len": 5000,
14
- "freeze_experts": 1,
15
- "freq_experts": "mean_naive_1/4_1/6_1/7_1/8_1/12_1/14_1/16_1/21_1/24_1/28_1/30_1/32_1/36_1/42_1/48_1/52_1/56_1/60_1/72_1/84_1/90_1/96_1/120_1/144_1/168_1/180_1/224_1/252_1/288_1/336_1/365_1/504_1/672_1/1008_1/1440_1/2016_1/3600",
16
- "inf_pred_len": 96,
17
- "layer_type": "RLinear",
18
- "linear_checkpoints_dir": "checkpoints5",
19
- "linear_checkpoints_path": "/cs/azencot_fsas/MoE/",
20
- "load_linear": 0,
21
- "load_weights" :0,
22
- "max_horizon": 96,
23
- "misc_moe": 10,
24
- "mlp_gating": 0,
25
- "model_type": "super_linear",
26
- "moe": 1,
27
- "moe_n_experts": 12,
28
- "moe_temp": 1,
29
- "noisy_gating_std": 0.1,
30
- "noisy_gating_std_decay": 1,
31
- "pred_len": 96,
32
- "seq_len": 512,
33
- "moe_norm": 0,
34
- "top_k_experts": 12,
35
- "torch_dtype": "float32",
36
- "transformers_version": "4.40.1",
37
- "use_fft": 1,
38
- "misc_moe" :12,
39
- }