AbstractPhil commited on
Commit
38f56f8
·
verified ·
1 Parent(s): 3c07ac0

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +19 -10
config.json CHANGED
@@ -1,35 +1,44 @@
1
  {
2
- "adapter_id": "003",
3
- "name": "DualShuntAdapter-G",
 
4
 
 
5
  "t5": {
6
  "model": "google/flan-t5-base",
7
  "hidden_size": 768,
8
  },
9
  "clip": {
10
- "model": "laion/CLIP-ViT-bigG-14-laion2B-39B-b160k",
11
- "hidden_size": 1280,
12
  },
13
 
14
- "bottleneck": 640,
15
- "heads": 20,
 
16
 
 
17
  "tau_init": 0.1,
18
  "max_guidance": 10.0,
19
 
20
- "proj_layers": 2,
21
- "layer_norm": true,
 
22
  "dropout": 0.1,
23
  "use_dropout": true,
24
  "use_proj_stack": true,
 
 
25
  "assert_input_dims": true,
26
 
 
27
  "routing": {
28
  "type": "cross_attention",
29
  "enable_causal_mask": false,
30
  "bidirectional": true
31
  },
32
 
33
- "version": "v0.3.2",
34
- "description": "Final Dual Shunt Adapter with projection stack, dropout, and stacked residual refinement pocket."
 
35
  }
 
1
  {
2
+ # Model Integration IDs
3
+ "adapter_id": "002",
4
+ "name": "TwoStreamShuntAdapter",
5
 
6
+ # Backbone Model Dimensions
7
  "t5": {
8
  "model": "google/flan-t5-base",
9
  "hidden_size": 768,
10
  },
11
  "clip": {
12
+ "model": "openai/clip-vit-large-patch14",
13
+ "hidden_size": 768,
14
  },
15
 
16
+ # Adapter Dimensions
17
+ "bottleneck": 384,
18
+ "heads": 12,
19
 
20
+ # Guidance Parameters
21
  "tau_init": 0.1,
22
  "max_guidance": 10.0,
23
 
24
+ # Projection Configuration
25
+ "proj_layers": 2, # number of linear+GELU layers
26
+ "layer_norm": true, # apply LayerNorm before stack
27
  "dropout": 0.1,
28
  "use_dropout": true,
29
  "use_proj_stack": true,
30
+
31
+ # Runtime Safeguards
32
  "assert_input_dims": true,
33
 
34
+ # Routing Logic
35
  "routing": {
36
  "type": "cross_attention",
37
  "enable_causal_mask": false,
38
  "bidirectional": true
39
  },
40
 
41
+ # Version & Metadata
42
+ "version": "v0.3.1",
43
+ "description": "Upgraded FLAN-T5 ↔ CLIP-L token shunt with projection stack, dropout, and field-consistent architecture."
44
  }