adbaral commited on
Commit
bdcc381
·
verified ·
1 Parent(s): 7d6c725

Training in progress, step 20000

Browse files
config.json CHANGED
@@ -4,7 +4,7 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 50281,
8
  "classifier_activation": "gelu",
9
  "classifier_bias": false,
10
  "classifier_dropout": 0.0,
@@ -14,7 +14,7 @@
14
  "deterministic_flash_attn": false,
15
  "dtype": "bfloat16",
16
  "embedding_dropout": 0.0,
17
- "eos_token_id": 50282,
18
  "global_attn_every_n_layers": 3,
19
  "global_rope_theta": 160000.0,
20
  "gradient_checkpointing": false,
@@ -40,7 +40,7 @@
40
  "norm_eps": 1e-05,
41
  "num_attention_heads": 12,
42
  "num_hidden_layers": 22,
43
- "pad_token_id": 50283,
44
  "position_embedding_type": "absolute",
45
  "repad_logits_with_grad": false,
46
  "sentence_transformers": {
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
  "classifier_activation": "gelu",
9
  "classifier_bias": false,
10
  "classifier_dropout": 0.0,
 
14
  "deterministic_flash_attn": false,
15
  "dtype": "bfloat16",
16
  "embedding_dropout": 0.0,
17
+ "eos_token_id": null,
18
  "global_attn_every_n_layers": 3,
19
  "global_rope_theta": 160000.0,
20
  "gradient_checkpointing": false,
 
40
  "norm_eps": 1e-05,
41
  "num_attention_heads": 12,
42
  "num_hidden_layers": 22,
43
+ "pad_token_id": 50284,
44
  "position_embedding_type": "absolute",
45
  "repad_logits_with_grad": false,
46
  "sentence_transformers": {
eval/CrossEncoderClassificationEvaluator_test_cls_results.csv ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,Accuracy,Accuracy_Threshold,F1,F1_Threshold,Precision,Recall,Average_Precision
2
+ 0,0,0.5526425637918265,0.9589844,0.6182535546787762,0.6035156,0.4474616451825811,0.9999097011106763,0.4262552623354809
3
+ 0,0,0.5526425637918265,0.9589844,0.6182535546787762,0.6035156,0.4474616451825811,0.9999097011106763,0.4262552623354809
4
+ 0,0,0.5526425637918265,0.9589844,0.6182535546787762,0.6035156,0.4474616451825811,0.9999097011106763,0.4262552623354809
5
+ 0,0,0.5526425637918265,0.9589844,0.6182535546787762,0.6035156,0.4474616451825811,0.9999097011106763,0.4262552623354809
6
+ 0.09596330363269086,20000,0.6076078906618192,0.1875,0.6569540025519276,0.07910156,0.5033899120065394,0.9453390723294104,0.5956208960759726
7
+ 0.09596330363269086,20000,0.6040126573756144,0.20800781,0.6606378016199482,0.083984375,0.5099261877874339,0.9378141648857719,0.590106659536392
8
+ 0.09596330363269086,20000,0.6046724567427456,0.18457031,0.6571979174430118,0.07714844,0.5084514003294893,0.928964873732053,0.5918929196664637
9
+ 0.09596330363269086,20000,0.6017100922372585,0.19921875,0.6539650060550382,0.07470703,0.5006153600255734,0.9427505041687987,0.5839840201608288
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57f77ebd6a51dda0900d118d06b23e5a84f8ccadceab69fc303efa78783a0617
3
  size 299228626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:160aa613bec4156768917281d3dad813fafd9298631e755cf64516c28d208134
3
  size 299228626
tokenizer.json CHANGED
@@ -1,7 +1,19 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 299,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": "BatchLongest",
11
+ "direction": "Right",
12
+ "pad_to_multiple_of": null,
13
+ "pad_id": 50284,
14
+ "pad_type_id": 0,
15
+ "pad_token": "[MASK]"
16
+ },
17
  "added_tokens": [
18
  {
19
  "id": 0,
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75f4396753950ebe568629d637dab7058b2af2533948250b5f7114489c99dca7
3
+ size 6353