kleinay commited on
Commit
8b99d80
·
1 Parent(s): 1af26a8
Files changed (2) hide show
  1. config.json +52 -16
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,47 +1,85 @@
1
  {
2
- "_name_or_path": "trained_models/t5_10ep-joint-qanom_15.12.21",
 
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
6
- "batch_size": 12,
7
  "d_ff": 2048,
8
  "d_kv": 64,
9
  "d_model": 512,
 
10
  "decoder_start_token_id": 0,
11
- "description": "duplicating qanom training data ",
 
 
12
  "dropout_rate": 0.1,
13
  "eos_token_id": 1,
 
 
14
  "feed_forward_proj": "relu",
15
  "fp16": true,
16
- "gradient_checkpointing": false,
17
  "initializer_factor": 1.0,
18
  "is_encoder_decoder": true,
19
  "layer_norm_epsilon": 1e-06,
 
 
 
 
 
20
  "model_type": "t5",
21
  "n_positions": 512,
 
22
  "num_decoder_layers": 6,
23
  "num_heads": 8,
24
  "num_layers": 6,
25
  "output_past": true,
26
  "overwrite_output_dir": true,
27
  "pad_token_id": 0,
 
 
 
 
28
  "preprocess_input_func": "input_predicate_marker",
29
  "preprocessing_kwargs": {
30
- "batch_size": 12,
31
- "description": "duplicating qanom training data ",
 
 
 
 
 
 
32
  "fp16": true,
 
 
 
 
 
 
33
  "model_type": "t5",
 
34
  "overwrite_output_dir": true,
 
 
 
 
35
  "preprocess_input_func": "input_predicate_marker",
36
- "source_prefix": "<predicate-type>",
37
- "test_dataset": "qanom",
 
 
 
38
  "train_dataset": "joint_qanom",
39
- "train_epochs": 5,
40
- "use_bilateral_predicate_marker": true,
41
- "wandb_run_name": "2021-12-16--00:13:35_5ep_t5_joint_qanom_factor=14"
42
  },
 
43
  "relative_attention_num_buckets": 32,
44
- "source_prefix": "<predicate-type>",
 
 
 
45
  "task_specific_params": {
46
  "summarization": {
47
  "early_stopping": true,
@@ -71,13 +109,11 @@
71
  "prefix": "translate English to Romanian: "
72
  }
73
  },
74
- "test_dataset": "qanom",
75
  "torch_dtype": "float32",
76
  "train_dataset": "joint_qanom",
77
- "train_epochs": 5,
78
  "transformers_version": "4.14.1",
79
  "use_bilateral_predicate_marker": true,
80
  "use_cache": true,
81
- "vocab_size": 32101,
82
- "wandb_run_name": "2021-12-16--00:13:35_5ep_t5_joint_qanom_factor=14"
83
  }
 
1
  {
2
+ "_name_or_path": "trained_models/t5_qanom-joint-23.03.22",
3
+ "append_verb_form": true,
4
  "architectures": [
5
  "T5ForConditionalGeneration"
6
  ],
 
7
  "d_ff": 2048,
8
  "d_kv": 64,
9
  "d_model": 512,
10
+ "debug_mode": false,
11
  "decoder_start_token_id": 0,
12
+ "description": "optimal joint config from sweep, mainly for qanom",
13
+ "dir_switch": "joint_optimal",
14
+ "do_eval_on": "validation",
15
  "dropout_rate": 0.1,
16
  "eos_token_id": 1,
17
+ "eval_steps": 500,
18
+ "evaluation_strategy": "steps",
19
  "feed_forward_proj": "relu",
20
  "fp16": true,
21
+ "gradient_accumulation_steps": 14,
22
  "initializer_factor": 1.0,
23
  "is_encoder_decoder": true,
24
  "layer_norm_epsilon": 1e-06,
25
+ "learning_rate": 0.001,
26
+ "load_best_model_at_end": true,
27
+ "logging_steps": 500,
28
+ "logging_strategy": "steps",
29
+ "metric_for_best_model": "eval_loss",
30
  "model_type": "t5",
31
  "n_positions": 512,
32
+ "num_beams": 5,
33
  "num_decoder_layers": 6,
34
  "num_heads": 8,
35
  "num_layers": 6,
36
  "output_past": true,
37
  "overwrite_output_dir": true,
38
  "pad_token_id": 0,
39
+ "per_device_eval_batch_size": 12,
40
+ "per_device_train_batch_size": 12,
41
+ "predicate_marker_type": "generic",
42
+ "predict_with_generate": true,
43
  "preprocess_input_func": "input_predicate_marker",
44
  "preprocessing_kwargs": {
45
+ "append_verb_form": true,
46
+ "debug_mode": false,
47
+ "description": "optimal joint config from sweep, mainly for qanom",
48
+ "dir_switch": "joint_optimal",
49
+ "do_eval_on": "validation",
50
+ "dropout_rate": 0.1,
51
+ "eval_steps": 500,
52
+ "evaluation_strategy": "steps",
53
  "fp16": true,
54
+ "gradient_accumulation_steps": 14,
55
+ "learning_rate": 0.001,
56
+ "load_best_model_at_end": true,
57
+ "logging_steps": 500,
58
+ "logging_strategy": "steps",
59
+ "metric_for_best_model": "eval_loss",
60
  "model_type": "t5",
61
+ "num_beams": 5,
62
  "overwrite_output_dir": true,
63
+ "per_device_eval_batch_size": 12,
64
+ "per_device_train_batch_size": 12,
65
+ "predicate_marker_type": "generic",
66
+ "predict_with_generate": true,
67
  "preprocess_input_func": "input_predicate_marker",
68
+ "qanom_joint_factor": 14,
69
+ "save_steps": 500,
70
+ "save_strategy": "steps",
71
+ "seed": 44,
72
+ "source_prefix": "parse: ",
73
  "train_dataset": "joint_qanom",
74
+ "train_epochs": 20,
75
+ "use_bilateral_predicate_marker": true
 
76
  },
77
+ "qanom_joint_factor": 14,
78
  "relative_attention_num_buckets": 32,
79
+ "save_steps": 500,
80
+ "save_strategy": "steps",
81
+ "seed": 44,
82
+ "source_prefix": "parse: ",
83
  "task_specific_params": {
84
  "summarization": {
85
  "early_stopping": true,
 
109
  "prefix": "translate English to Romanian: "
110
  }
111
  },
 
112
  "torch_dtype": "float32",
113
  "train_dataset": "joint_qanom",
114
+ "train_epochs": 20,
115
  "transformers_version": "4.14.1",
116
  "use_bilateral_predicate_marker": true,
117
  "use_cache": true,
118
+ "vocab_size": 32101
 
119
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2272dd02b62d084677d2333038457d6865bbd751d962f945fd9209fe652fc7d0
3
- size 242030331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67987af0aa8b7d2584e73cfaa33571bae84c52e4bc3534c5e896e1e7824343bd
3
+ size 242028475