prithivMLmods commited on
Commit
daf4683
·
verified ·
1 Parent(s): eeeea5e

Upload folder using huggingface_hub

Browse files
transformer/config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "QwenImageTransformer2DModel",
3
+ "_diffusers_version": "0.37.0.dev0",
4
+ "_name_or_path": "/home/user/.cache/huggingface/hub/models--FireRedTeam--FireRed-Image-Edit-1.0/snapshots/0aea9d520c801c9f4b691cc92f736a7a8628e6a2/transformer",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "guidance_embeds": false,
12
+ "in_channels": 64,
13
+ "joint_attention_dim": 3584,
14
+ "num_attention_heads": 24,
15
+ "num_layers": 60,
16
+ "out_channels": 16,
17
+ "patch_size": 2,
18
+ "quantization_config": {
19
+ "config_groups": {
20
+ "group_0": {
21
+ "format": "float-quantized",
22
+ "input_activations": {
23
+ "actorder": null,
24
+ "block_structure": null,
25
+ "dynamic": true,
26
+ "group_size": null,
27
+ "num_bits": 8,
28
+ "observer": null,
29
+ "observer_kwargs": {},
30
+ "scale_dtype": null,
31
+ "strategy": "token",
32
+ "symmetric": true,
33
+ "type": "float",
34
+ "zp_dtype": null
35
+ },
36
+ "output_activations": null,
37
+ "targets": [
38
+ "Linear"
39
+ ],
40
+ "weights": {
41
+ "actorder": null,
42
+ "block_structure": null,
43
+ "dynamic": false,
44
+ "group_size": null,
45
+ "num_bits": 8,
46
+ "observer": "memoryless_minmax",
47
+ "observer_kwargs": {},
48
+ "scale_dtype": null,
49
+ "strategy": "channel",
50
+ "symmetric": true,
51
+ "type": "float",
52
+ "zp_dtype": null
53
+ }
54
+ }
55
+ },
56
+ "format": "float-quantized",
57
+ "global_compression_ratio": null,
58
+ "ignore": [
59
+ "proj_out"
60
+ ],
61
+ "kv_cache_scheme": null,
62
+ "quant_method": "compressed-tensors",
63
+ "quantization_status": "compressed",
64
+ "sparsity_config": {},
65
+ "transform_config": {},
66
+ "version": "0.13.1.a20260219"
67
+ },
68
+ "use_additional_t_cond": false,
69
+ "use_layer3d_rope": false,
70
+ "zero_cond_t": false
71
+ }
transformer/diffusion_pytorch_model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a8c11dcb524f889c7b13b00eaf4bf04c66b9406f8359a662326377ac421770c
3
+ size 9998179576
transformer/diffusion_pytorch_model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5e6d22ef6b4939ef01b102dfcd92beade6e1cc98f47f5be11a4b87d615d602d
3
+ size 9957755328
transformer/diffusion_pytorch_model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59758d51e4491ce5131285daa8bd73257bd424b27666a9697e8f3650cc532c77
3
+ size 491663088
transformer/diffusion_pytorch_model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
transformer/recipe.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ default_stage:
2
+ default_modifiers:
3
+ QuantizationModifier:
4
+ targets: [Linear]
5
+ ignore: [proj_out, lm_head]
6
+ scheme: FP8_DYNAMIC
7
+ bypass_divisibility_checks: false