scottgdaniel
commited on
Commit
•
324b6c2
1
Parent(s):
8b13ee2
Model save
Browse files- README.md +70 -0
- added_tokens.json +3 -0
- config.json +35 -0
- model.safetensors +3 -0
- runs/events.out.tfevents.1718056601.delilah.39183.0 +3 -0
- runs/events.out.tfevents.1718056997.delilah.39183.1 +3 -0
- special_tokens_map.json +51 -0
- spm.model +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +65 -0
- training_args.bin +3 -0
- training_progress.csv +82 -0
README.md
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: scales-okn/docket-language-model
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
+
metrics:
|
6 |
+
- accuracy
|
7 |
+
- f1
|
8 |
+
- precision
|
9 |
+
- recall
|
10 |
+
model-index:
|
11 |
+
- name: ontology-answer-test
|
12 |
+
results: []
|
13 |
+
---
|
14 |
+
|
15 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
+
|
18 |
+
# ontology-answer-test
|
19 |
+
|
20 |
+
This model is a fine-tuned version of [scales-okn/docket-language-model](https://huggingface.co/scales-okn/docket-language-model) on the None dataset.
|
21 |
+
It achieves the following results on the evaluation set:
|
22 |
+
- Loss: 0.0009
|
23 |
+
- Accuracy: 1.0
|
24 |
+
- F1: 1.0
|
25 |
+
- Precision: 1.0
|
26 |
+
- Recall: 1.0
|
27 |
+
|
28 |
+
## Model description
|
29 |
+
|
30 |
+
More information needed
|
31 |
+
|
32 |
+
## Intended uses & limitations
|
33 |
+
|
34 |
+
More information needed
|
35 |
+
|
36 |
+
## Training and evaluation data
|
37 |
+
|
38 |
+
More information needed
|
39 |
+
|
40 |
+
## Training procedure
|
41 |
+
|
42 |
+
### Training hyperparameters
|
43 |
+
|
44 |
+
The following hyperparameters were used during training:
|
45 |
+
- learning_rate: 3e-05
|
46 |
+
- train_batch_size: 8
|
47 |
+
- eval_batch_size: 16
|
48 |
+
- seed: 42
|
49 |
+
- gradient_accumulation_steps: 2
|
50 |
+
- total_train_batch_size: 16
|
51 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
52 |
+
- lr_scheduler_type: linear
|
53 |
+
- lr_scheduler_warmup_ratio: 0.06
|
54 |
+
- num_epochs: 5
|
55 |
+
|
56 |
+
### Training results
|
57 |
+
|
58 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
|
59 |
+
|:-------------:|:------:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
|
60 |
+
| 0.0233 | 1.2903 | 100 | 0.0434 | 0.9862 | 0.9720 | 0.9630 | 0.9811 |
|
61 |
+
| 0.0007 | 2.5806 | 200 | 0.0072 | 0.9954 | 0.9905 | 1.0 | 0.9811 |
|
62 |
+
| 0.0003 | 3.8710 | 300 | 0.0009 | 1.0 | 1.0 | 1.0 | 1.0 |
|
63 |
+
|
64 |
+
|
65 |
+
### Framework versions
|
66 |
+
|
67 |
+
- Transformers 4.41.2
|
68 |
+
- Pytorch 2.1.2+cu121
|
69 |
+
- Datasets 2.19.2
|
70 |
+
- Tokenizers 0.19.1
|
added_tokens.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"[MASK]": 128000
|
3 |
+
}
|
config.json
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "scales-okn/docket-language-model",
|
3 |
+
"architectures": [
|
4 |
+
"DebertaV2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"hidden_act": "gelu",
|
8 |
+
"hidden_dropout_prob": 0.1,
|
9 |
+
"hidden_size": 1024,
|
10 |
+
"initializer_range": 0.02,
|
11 |
+
"intermediate_size": 4096,
|
12 |
+
"layer_norm_eps": 1e-07,
|
13 |
+
"max_position_embeddings": 512,
|
14 |
+
"max_relative_positions": -1,
|
15 |
+
"model_type": "deberta-v2",
|
16 |
+
"norm_rel_ebd": "layer_norm",
|
17 |
+
"num_attention_heads": 16,
|
18 |
+
"num_hidden_layers": 24,
|
19 |
+
"pad_token_id": 0,
|
20 |
+
"pooler_dropout": 0,
|
21 |
+
"pooler_hidden_act": "gelu",
|
22 |
+
"pooler_hidden_size": 1024,
|
23 |
+
"pos_att_type": [
|
24 |
+
"p2c",
|
25 |
+
"c2p"
|
26 |
+
],
|
27 |
+
"position_biased_input": false,
|
28 |
+
"position_buckets": 256,
|
29 |
+
"relative_attention": true,
|
30 |
+
"share_att_key": true,
|
31 |
+
"torch_dtype": "float32",
|
32 |
+
"transformers_version": "4.41.2",
|
33 |
+
"type_vocab_size": 0,
|
34 |
+
"vocab_size": 128001
|
35 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d73de8e9e8fbb84901b725ece1bb92cc851d64037a8f591f3086c2ba392b3832
|
3 |
+
size 1739898936
|
runs/events.out.tfevents.1718056601.delilah.39183.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca6c0e09317e41c2e5971315213a2997489e3fac06515359bf55e27a4da2dce5
|
3 |
+
size 22886
|
runs/events.out.tfevents.1718056997.delilah.39183.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d0e18ae640e58a4e9e1661a6906177d73b5cc5438d72d9ee6eb786eda2f0db50
|
3 |
+
size 512
|
special_tokens_map.json
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"cls_token": {
|
10 |
+
"content": "[CLS]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"eos_token": {
|
17 |
+
"content": "[SEP]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"mask_token": {
|
24 |
+
"content": "[MASK]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"pad_token": {
|
31 |
+
"content": "[PAD]",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
},
|
37 |
+
"sep_token": {
|
38 |
+
"content": "[SEP]",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false
|
43 |
+
},
|
44 |
+
"unk_token": {
|
45 |
+
"content": "[UNK]",
|
46 |
+
"lstrip": false,
|
47 |
+
"normalized": false,
|
48 |
+
"rstrip": false,
|
49 |
+
"single_word": false
|
50 |
+
}
|
51 |
+
}
|
spm.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
3 |
+
size 2464616
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "[CLS]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "[SEP]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "[UNK]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"128000": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"bos_token": "[CLS]",
|
45 |
+
"clean_up_tokenization_spaces": true,
|
46 |
+
"cls_token": "[CLS]",
|
47 |
+
"do_lower_case": false,
|
48 |
+
"eos_token": "[SEP]",
|
49 |
+
"mask_token": "[MASK]",
|
50 |
+
"max_length": 256,
|
51 |
+
"model_max_length": 1000000000000000019884624838656,
|
52 |
+
"pad_to_multiple_of": null,
|
53 |
+
"pad_token": "[PAD]",
|
54 |
+
"pad_token_type_id": 0,
|
55 |
+
"padding_side": "right",
|
56 |
+
"sep_token": "[SEP]",
|
57 |
+
"sp_model_kwargs": {},
|
58 |
+
"split_by_punct": false,
|
59 |
+
"stride": 0,
|
60 |
+
"tokenizer_class": "DebertaV2Tokenizer",
|
61 |
+
"truncation_side": "right",
|
62 |
+
"truncation_strategy": "longest_first",
|
63 |
+
"unk_token": "[UNK]",
|
64 |
+
"vocab_type": "spm"
|
65 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a58e6b4fe46438892548857b39aae1d27703c6e32b0b6bc13989226a177399c4
|
3 |
+
size 5112
|
training_progress.csv
ADDED
@@ -0,0 +1,82 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
loss,grad_norm,learning_rate,epoch,steps,eval_loss,eval_accuracy,eval_f1,eval_precision,eval_recall,eval_runtime,eval_samples_per_second,eval_steps_per_second,train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss
|
2 |
+
0.6909,4.090293884277344,6.25e-06,0.06451612903225806,5,,,,,,,,,,,,,
|
3 |
+
0.6278,5.705702781677246,1.25e-05,0.12903225806451613,10,,,,,,,,,,,,,
|
4 |
+
0.549,2.5465235710144043,1.8750000000000002e-05,0.1935483870967742,15,,,,,,,,,,,,,
|
5 |
+
0.4981,3.545476198196411,2.5e-05,0.25806451612903225,20,,,,,,,,,,,,,
|
6 |
+
0.3127,1.7096995115280151,2.991689750692521e-05,0.3225806451612903,25,,,,,,,,,,,,,
|
7 |
+
0.2114,3.098984479904175,2.950138504155125e-05,0.3870967741935484,30,,,,,,,,,,,,,
|
8 |
+
0.1548,2.043656826019287,2.9085872576177285e-05,0.45161290322580644,35,,,,,,,,,,,,,
|
9 |
+
0.0987,0.37046054005622864,2.8670360110803324e-05,0.5161290322580645,40,,,,,,,,,,,,,
|
10 |
+
0.1391,2.608640670776367,2.8254847645429363e-05,0.5806451612903226,45,,,,,,,,,,,,,
|
11 |
+
0.0606,0.4053419530391693,2.78393351800554e-05,0.6451612903225806,50,,,,,,,,,,,,,
|
12 |
+
0.0275,4.057948112487793,2.7423822714681444e-05,0.7096774193548387,55,,,,,,,,,,,,,
|
13 |
+
0.0928,0.23072686791419983,2.700831024930748e-05,0.7741935483870968,60,,,,,,,,,,,,,
|
14 |
+
0.0733,0.05615957826375961,2.6592797783933518e-05,0.8387096774193549,65,,,,,,,,,,,,,
|
15 |
+
0.1347,2.636467456817627,2.6177285318559557e-05,0.9032258064516129,70,,,,,,,,,,,,,
|
16 |
+
0.0766,0.11293677985668182,2.5761772853185596e-05,0.967741935483871,75,,,,,,,,,,,,,
|
17 |
+
0.0055,0.1519118696451187,2.5346260387811635e-05,1.032258064516129,80,,,,,,,,,,,,,
|
18 |
+
0.0052,0.2113182097673416,2.4930747922437674e-05,1.096774193548387,85,,,,,,,,,,,,,
|
19 |
+
0.0395,2.404287099838257,2.4515235457063713e-05,1.1612903225806452,90,,,,,,,,,,,,,
|
20 |
+
0.0084,0.08060836046934128,2.409972299168975e-05,1.2258064516129032,95,,,,,,,,,,,,,
|
21 |
+
0.0233,0.7740814685821533,2.368421052631579e-05,1.2903225806451613,100,,,,,,,,,,,,,
|
22 |
+
,,,1.2903225806451613,100,0.04341869056224823,0.9862385321100917,0.9719626168224299,0.9629629629629629,0.9811320754716981,4.3861,49.702,3.192,,,,,
|
23 |
+
0.0204,1.7034190893173218,2.326869806094183e-05,1.3548387096774195,105,,,,,,,,,,,,,
|
24 |
+
0.0664,0.06643515080213547,2.2853185595567868e-05,1.4193548387096775,110,,,,,,,,,,,,,
|
25 |
+
0.0126,0.048984918743371964,2.2437673130193904e-05,1.4838709677419355,115,,,,,,,,,,,,,
|
26 |
+
0.0027,0.42834651470184326,2.2022160664819946e-05,1.5483870967741935,120,,,,,,,,,,,,,
|
27 |
+
0.0291,0.06453526765108109,2.1606648199445985e-05,1.6129032258064515,125,,,,,,,,,,,,,
|
28 |
+
0.2114,7.842092990875244,2.1191135734072024e-05,1.6774193548387095,130,,,,,,,,,,,,,
|
29 |
+
0.0494,0.04162972792983055,2.0775623268698063e-05,1.7419354838709677,135,,,,,,,,,,,,,
|
30 |
+
0.0017,0.06891004741191864,2.0360110803324098e-05,1.8064516129032258,140,,,,,,,,,,,,,
|
31 |
+
0.0021,0.040090594440698624,1.9944598337950137e-05,1.870967741935484,145,,,,,,,,,,,,,
|
32 |
+
0.0017,0.07595518976449966,1.952908587257618e-05,1.935483870967742,150,,,,,,,,,,,,,
|
33 |
+
0.0311,10.209275245666504,1.9113573407202218e-05,2.0,155,,,,,,,,,,,,,
|
34 |
+
0.0027,0.14497926831245422,1.8698060941828257e-05,2.064516129032258,160,,,,,,,,,,,,,
|
35 |
+
0.0024,0.40518391132354736,1.8282548476454293e-05,2.129032258064516,165,,,,,,,,,,,,,
|
36 |
+
0.0403,0.01545148529112339,1.786703601108033e-05,2.193548387096774,170,,,,,,,,,,,,,
|
37 |
+
0.0008,0.020430177450180054,1.745152354570637e-05,2.258064516129032,175,,,,,,,,,,,,,
|
38 |
+
0.0008,0.01945902779698372,1.7036011080332413e-05,2.3225806451612905,180,,,,,,,,,,,,,
|
39 |
+
0.0007,0.01845143549144268,1.662049861495845e-05,2.3870967741935485,185,,,,,,,,,,,,,
|
40 |
+
0.0007,0.018286943435668945,1.6204986149584487e-05,2.4516129032258065,190,,,,,,,,,,,,,
|
41 |
+
0.0024,0.014586242847144604,1.5789473684210526e-05,2.5161290322580645,195,,,,,,,,,,,,,
|
42 |
+
0.0007,0.012451103888452053,1.5373961218836565e-05,2.5806451612903225,200,,,,,,,,,,,,,
|
43 |
+
,,,2.5806451612903225,200,0.007189903408288956,0.9954128440366973,0.9904761904761905,1.0,0.9811320754716981,4.502,48.422,3.11,,,,,
|
44 |
+
0.0009,0.011877791956067085,1.4958448753462605e-05,2.6451612903225805,205,,,,,,,,,,,,,
|
45 |
+
0.0005,0.012689828872680664,1.4542936288088642e-05,2.709677419354839,210,,,,,,,,,,,,,
|
46 |
+
0.0007,0.017396604642271996,1.4127423822714681e-05,2.774193548387097,215,,,,,,,,,,,,,
|
47 |
+
0.0006,0.01691826991736889,1.3711911357340722e-05,2.838709677419355,220,,,,,,,,,,,,,
|
48 |
+
0.0007,0.026222320273518562,1.3296398891966759e-05,2.903225806451613,225,,,,,,,,,,,,,
|
49 |
+
0.0017,0.011283783242106438,1.2880886426592798e-05,2.967741935483871,230,,,,,,,,,,,,,
|
50 |
+
0.0006,0.016978178173303604,1.2465373961218837e-05,3.032258064516129,235,,,,,,,,,,,,,
|
51 |
+
0.0007,0.008052507415413857,1.2049861495844876e-05,3.096774193548387,240,,,,,,,,,,,,,
|
52 |
+
0.0004,0.007332273758947849,1.1634349030470915e-05,3.161290322580645,245,,,,,,,,,,,,,
|
53 |
+
0.0005,0.0079089580103755,1.1218836565096952e-05,3.225806451612903,250,,,,,,,,,,,,,
|
54 |
+
0.0005,0.009510311298072338,1.0803324099722992e-05,3.2903225806451615,255,,,,,,,,,,,,,
|
55 |
+
0.0005,0.024516815319657326,1.0387811634349031e-05,3.3548387096774195,260,,,,,,,,,,,,,
|
56 |
+
0.0005,0.021235521882772446,9.972299168975068e-06,3.4193548387096775,265,,,,,,,,,,,,,
|
57 |
+
0.0004,0.008695265278220177,9.556786703601109e-06,3.4838709677419355,270,,,,,,,,,,,,,
|
58 |
+
0.0004,0.008880405686795712,9.141274238227146e-06,3.5483870967741935,275,,,,,,,,,,,,,
|
59 |
+
0.0003,0.007136228494346142,8.725761772853185e-06,3.6129032258064515,280,,,,,,,,,,,,,
|
60 |
+
0.0004,0.008296678774058819,8.310249307479226e-06,3.6774193548387095,285,,,,,,,,,,,,,
|
61 |
+
0.0003,0.008309051394462585,7.894736842105263e-06,3.741935483870968,290,,,,,,,,,,,,,
|
62 |
+
0.0003,0.006046871189028025,7.479224376731303e-06,3.806451612903226,295,,,,,,,,,,,,,
|
63 |
+
0.0003,0.009278281591832638,7.063711911357341e-06,3.870967741935484,300,,,,,,,,,,,,,
|
64 |
+
,,,3.870967741935484,300,0.0009307524305768311,1.0,1.0,1.0,1.0,4.3119,50.557,3.247,,,,,
|
65 |
+
0.0036,0.008976064622402191,6.6481994459833796e-06,3.935483870967742,305,,,,,,,,,,,,,
|
66 |
+
0.0006,0.009819263592362404,6.2326869806094184e-06,4.0,310,,,,,,,,,,,,,
|
67 |
+
0.0004,0.007393244653940201,5.817174515235457e-06,4.064516129032258,315,,,,,,,,,,,,,
|
68 |
+
0.0004,0.006700852420181036,5.401662049861496e-06,4.129032258064516,320,,,,,,,,,,,,,
|
69 |
+
0.0004,0.009492475539445877,4.986149584487534e-06,4.193548387096774,325,,,,,,,,,,,,,
|
70 |
+
0.0004,0.014097731560468674,4.570637119113573e-06,4.258064516129032,330,,,,,,,,,,,,,
|
71 |
+
0.0003,0.0077942414209246635,4.155124653739613e-06,4.32258064516129,335,,,,,,,,,,,,,
|
72 |
+
0.0003,0.007693646941334009,3.7396121883656513e-06,4.387096774193548,340,,,,,,,,,,,,,
|
73 |
+
0.0003,0.009858954697847366,3.3240997229916898e-06,4.451612903225806,345,,,,,,,,,,,,,
|
74 |
+
0.0003,0.005932302679866552,2.9085872576177287e-06,4.516129032258064,350,,,,,,,,,,,,,
|
75 |
+
0.0003,0.010720741003751755,2.493074792243767e-06,4.580645161290323,355,,,,,,,,,,,,,
|
76 |
+
0.0004,0.007814132608473301,2.0775623268698064e-06,4.645161290322581,360,,,,,,,,,,,,,
|
77 |
+
0.0003,0.007145343814045191,1.6620498614958449e-06,4.709677419354839,365,,,,,,,,,,,,,
|
78 |
+
0.0003,0.0064388057217001915,1.2465373961218836e-06,4.774193548387097,370,,,,,,,,,,,,,
|
79 |
+
0.0003,0.008449292741715908,8.310249307479224e-07,4.838709677419355,375,,,,,,,,,,,,,
|
80 |
+
0.0004,0.006520669441670179,4.155124653739612e-07,4.903225806451613,380,,,,,,,,,,,,,
|
81 |
+
0.0004,0.008102640509605408,0.0,4.967741935483871,385,,,,,,,,,,,,,
|
82 |
+
,,,4.967741935483871,385,,,,,,,,,391.5134,15.823,0.983,2868504101425152.0,0.056238630435477895
|