Any-to-Any
Safetensors
ml-4m
4M-7-SR_L_CC12M / config.json
roman-bachmann's picture
Init
0c6245e
raw
history blame contribute delete
568 Bytes
{"domains_in": ["rgb@224", "caption", "det", "tok_rgb@224", "tok_depth@224", "tok_normal@224", "tok_semseg@224", "tok_clip@224", "rgb@448", "tok_rgb@448", "tok_depth@448", "tok_normal@448", "tok_semseg@448", "tok_clip@448"], "domains_out": ["tok_rgb@448", "tok_depth@448", "tok_normal@448", "tok_semseg@448", "tok_clip@448"], "encoder_depth": 24, "decoder_depth": 24, "dim": 1024, "num_heads": 16, "mlp_ratio": 4, "qkv_bias": false, "proj_bias": false, "mlp_bias": false, "norm_bias": false, "act_layer": "SiLU", "gated_mlp": true, "image_size": 224, "patch_size": 16}