Libidrave commited on
Commit
877b8b7
1 Parent(s): 33dd6ff

commit files to HF hub

Browse files
README.md ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - image-classification
4
+ - pytorch
5
+ - huggingpics
6
+ metrics:
7
+ - accuracy
8
+
9
+ model-index:
10
+ - name: BlurorBokehv1
11
+ results:
12
+ - task:
13
+ name: Image Classification
14
+ type: image-classification
15
+ metrics:
16
+ - name: Accuracy
17
+ type: accuracy
18
+ value: 0.9755244851112366
19
+ ---
20
+
21
+ # BlurorBokehv1
22
+
23
+
24
+ Blur Or Bokeh Model
config.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "OpenGVLab/pvt_v2_b5",
3
+ "architectures": [
4
+ "PvtV2ForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "depths": [
8
+ 3,
9
+ 6,
10
+ 40,
11
+ 3
12
+ ],
13
+ "drop_path_rate": 0.3,
14
+ "hidden_act": "gelu",
15
+ "hidden_dropout_prob": 0.0,
16
+ "hidden_sizes": [
17
+ 64,
18
+ 128,
19
+ 320,
20
+ 512
21
+ ],
22
+ "id2label": {
23
+ "0": "Blur",
24
+ "1": "Bokeh",
25
+ "2": "Normal"
26
+ },
27
+ "image_size": {
28
+ "height": 224,
29
+ "width": 224
30
+ },
31
+ "initializer_range": 0.02,
32
+ "label2id": {
33
+ "Blur": "0",
34
+ "Bokeh": "1",
35
+ "Normal": "2"
36
+ },
37
+ "layer_norm_eps": 1e-06,
38
+ "linear_attention": false,
39
+ "mlp_ratios": [
40
+ 4,
41
+ 4,
42
+ 4,
43
+ 4
44
+ ],
45
+ "model_type": "pvt_v2",
46
+ "num_attention_heads": [
47
+ 1,
48
+ 2,
49
+ 5,
50
+ 8
51
+ ],
52
+ "num_channels": 3,
53
+ "num_encoder_blocks": 4,
54
+ "out_features": [
55
+ "stage1",
56
+ "stage2",
57
+ "stage3",
58
+ "stage4"
59
+ ],
60
+ "out_indices": [
61
+ 0,
62
+ 1,
63
+ 2,
64
+ 3
65
+ ],
66
+ "patch_sizes": [
67
+ 7,
68
+ 3,
69
+ 3,
70
+ 3
71
+ ],
72
+ "problem_type": "single_label_classification",
73
+ "qkv_bias": true,
74
+ "reshape_last_stage": true,
75
+ "sr_ratios": [
76
+ 8,
77
+ 4,
78
+ 2,
79
+ 1
80
+ ],
81
+ "stage_names": [
82
+ "stage1",
83
+ "stage2",
84
+ "stage3",
85
+ "stage4"
86
+ ],
87
+ "strides": [
88
+ 4,
89
+ 2,
90
+ 2,
91
+ 2
92
+ ],
93
+ "torch_dtype": "float32",
94
+ "transformers_version": "4.39.3"
95
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da4681fbaf147d3f17aac0ce927cbc205385a2d0197e755f50718984dac89137
3
+ size 325922948
preprocessor_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_rescale",
8
+ "rescale_factor",
9
+ "do_normalize",
10
+ "image_mean",
11
+ "image_std",
12
+ "return_tensors",
13
+ "data_format",
14
+ "input_data_format"
15
+ ],
16
+ "do_normalize": true,
17
+ "do_rescale": true,
18
+ "do_resize": true,
19
+ "image_mean": [
20
+ 0.485,
21
+ 0.456,
22
+ 0.406
23
+ ],
24
+ "image_processor_type": "PvtImageProcessor",
25
+ "image_std": [
26
+ 0.229,
27
+ 0.224,
28
+ 0.225
29
+ ],
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "size": {
33
+ "height": 224,
34
+ "width": 224
35
+ }
36
+ }
runs/events.out.tfevents.1712417923.f5a2f39c8870.200.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1819b20d2194603db4b857166287f6f11cec1764ef06506a2bc24bb8bac100a1
3
+ size 4433