Add new SentenceTransformer model
Browse files- .gitattributes +1 -0
- 1_Pooling/config.json +10 -0
- README.md +1412 -0
- config.json +28 -0
- config_sentence_transformers.json +10 -0
- model.safetensors +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +51 -0
- tokenizer.json +3 -0
- tokenizer_config.json +63 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
1_Pooling/config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 1024,
|
3 |
+
"pooling_mode_cls_token": true,
|
4 |
+
"pooling_mode_mean_tokens": false,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
7 |
+
"pooling_mode_weightedmean_tokens": false,
|
8 |
+
"pooling_mode_lasttoken": false,
|
9 |
+
"include_prompt": true
|
10 |
+
}
|
README.md
ADDED
@@ -0,0 +1,1412 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- sentence-transformers
|
4 |
+
- sentence-similarity
|
5 |
+
- feature-extraction
|
6 |
+
- generated_from_trainer
|
7 |
+
- dataset_size:1138596
|
8 |
+
- loss:CachedGISTEmbedLoss
|
9 |
+
base_model: seongil-dn/unsupervised_20m_3800
|
10 |
+
widget:
|
11 |
+
- source_sentence: How many people were reported to have died in the Great Fire of
|
12 |
+
London in 1666?
|
13 |
+
sentences:
|
14 |
+
- City of London 1666. Both of these fires were referred to as "the" Great Fire.
|
15 |
+
After the fire of 1666, a number of plans were drawn up to remodel the City and
|
16 |
+
its street pattern into a renaissance-style city with planned urban blocks, squares
|
17 |
+
and boulevards. These plans were almost entirely not taken up, and the medieval
|
18 |
+
street pattern re-emerged almost intact. By the late 16th century, London increasingly
|
19 |
+
became a major centre for banking, international trade and commerce. The Royal
|
20 |
+
Exchange was founded in 1565 by Sir Thomas Gresham as a centre of commerce for
|
21 |
+
London's merchants, and gained Royal patronage in
|
22 |
+
- Great Atlanta fire of 1917 Great Atlanta fire of 1917 The Great Atlanta Fire of
|
23 |
+
1917 began just after noon on 21 May 1917 in the Old Fourth Ward of Atlanta, Georgia.
|
24 |
+
It is unclear just how the fire started, but it was fueled by hot temperatures
|
25 |
+
and strong winds which propelled the fire. The fire, which burned for nearly 10
|
26 |
+
hours, destroyed and 1,900 structures displacing over 10,000 people. Damages were
|
27 |
+
estimated at $5 million, ($ million when adjusted for inflation). It was a clear,
|
28 |
+
warm and sunny day with a brisk breeze from the south. This was not the only fire
|
29 |
+
of the
|
30 |
+
- Great Plague of London they had ever been seen ...". Plague cases continued to
|
31 |
+
occur sporadically at a modest rate until the summer of 1666. On the second and
|
32 |
+
third of September that year, the Great Fire of London destroyed much of the City
|
33 |
+
of London, and some people believed that the fire put an end to the epidemic.
|
34 |
+
However, it is now thought that the plague had largely subsided before the fire
|
35 |
+
took place. In fact, most of the later cases of plague were found in the suburbs,
|
36 |
+
and it was the City of London itself that was destroyed by the Fire. According
|
37 |
+
- Monument to the Great Fire of London Monument to the Great Fire of London The
|
38 |
+
Monument to the Great Fire of London, more commonly known simply as the Monument,
|
39 |
+
is a Doric column in London, United Kingdom, situated near the northern end of
|
40 |
+
London Bridge. Commemorating the Great Fire of London, it stands at the junction
|
41 |
+
of Monument Street and Fish Street Hill, in height and 202 feet west of the spot
|
42 |
+
in Pudding Lane where the Great Fire started on 2 September 1666. Constructed
|
43 |
+
between 1671 and 1677, it was built on the site of St. Margaret's, Fish Street,
|
44 |
+
the first church to be destroyed by
|
45 |
+
- 'How to Have Sex in an Epidemic New York City government and organizations within
|
46 |
+
the LGBT community. The Gay Men''s Health Crisis offered to buy all 5,000 pamphlets
|
47 |
+
and promote them, with the condition that any mentions of the multifactorial model
|
48 |
+
be removed from the writing. The authors refused. Berkowitz recounts in an interview
|
49 |
+
it being "infuriating" that in 1985, the city still hadn''t adopted any standard
|
50 |
+
safe sex education. The advent of safe sex in urban gay male populations came
|
51 |
+
too late for many people: by 1983, more than 1,476 people had died from AIDS and
|
52 |
+
David France estimated that as much as half of all'
|
53 |
+
- 'Monument to the Great Fire of London six years to complete the 202 ft column.
|
54 |
+
It was two more years before the inscription (which had been left to Wren — or
|
55 |
+
to Wren''s choice — to decide upon) was set in place. "Commemorating — with a
|
56 |
+
brazen disregard for the truth — the fact that ''London rises again...three short
|
57 |
+
years complete that which was considered the work of ages.''" Hooke''s surviving
|
58 |
+
drawings show that several versions of the monument were submitted for consideration:
|
59 |
+
a plain obelisk, a column garnished with tongues of fire, and the fluted Doric
|
60 |
+
column that was eventually chosen. The real contention came with'
|
61 |
+
- source_sentence: '"The Claude Francois song ""Comme d''habitude"" (translation ""as
|
62 |
+
usual"") was a hit in English for Frank Sinatra under what title?"'
|
63 |
+
sentences:
|
64 |
+
- Young at Heart (Frank Sinatra song) young, Dick Van Dyke recorded a duet with
|
65 |
+
his wife, Arlene, at Capital Records Studio in Los Angeles, filmed for the HBO
|
66 |
+
Special on aging "If I'm not in the Obituary, I'll have Breakfast" starring Carl
|
67 |
+
Reiner, and featuring other young at heart +90 treasures, Mel Brooks, Norman Lear,
|
68 |
+
Stan Lee & Betty White among others. Van Dyke was recorded using Frank Sinatra's
|
69 |
+
microphone. Young at Heart (Frank Sinatra song) "Young at Heart" is a pop standard,
|
70 |
+
a ballad with music by Johnny Richards and lyrics by Carolyn Leigh. The song was
|
71 |
+
written and published in 1953, with Leigh contributing
|
72 |
+
- 'Comme d''habitude a relationship that is falling out of love, while the English
|
73 |
+
language version is set at the end of a lifetime, approaching death, and looking
|
74 |
+
back without regret – expressing feelings that are more related to Piaf''s song
|
75 |
+
"Non, je ne regrette rien". Many artists sang "Comme d''Habitude" in French after
|
76 |
+
Claude François''s success (and international success through ''"My Way"), notably:
|
77 |
+
David Bowie has said that in 1968 – the year before Paul Anka acquired the French
|
78 |
+
song – his manager, Kenneth Pitt, asked him to write English lyrics for "Comme
|
79 |
+
d''habitude" but that his version, titled "Even a Fool'
|
80 |
+
- Frank Sinatra Me" with Billy May, designed as a musical world tour. It reached
|
81 |
+
the top spot on the Billboard album chart in its second week, remaining at the
|
82 |
+
top for five weeks, and was nominated for the Grammy Award for Album of the Year
|
83 |
+
at the inaugural Grammy Awards. The title song, "Come Fly With Me", written especially
|
84 |
+
for him, would become one of his best known standards. On May 29 he recorded seven
|
85 |
+
songs in a single session, more than double the usual yield of a recording session,
|
86 |
+
and an eighth was planned, "Lush Life", but Sinatra found it too
|
87 |
+
- Frank Sinatra Original Song. Sinatra released "Softly, as I Leave You", and collaborated
|
88 |
+
with Bing Crosby and Fred Waring on "America, I Hear You Singing", a collection
|
89 |
+
of patriotic songs recorded as a tribute to the assassinated President John F.
|
90 |
+
Kennedy. Sinatra increasingly became involved in charitable pursuits in this period.
|
91 |
+
In 1961 and 1962 he went to Mexico, with the sole purpose of putting on performances
|
92 |
+
for Mexican charities, and in July 1964 he was present for the dedication of the
|
93 |
+
Frank Sinatra International Youth Center for Arab and Jewish children in Nazareth.
|
94 |
+
Sinatra's phenomenal success in 1965, coinciding with his
|
95 |
+
- Comme ci comme ça (Basim song) to the charm of it all. Working both Danish and
|
96 |
+
Moroccan Arabic, Basim sings about a girl he is ready to commit to. It doesn’t
|
97 |
+
mater what she wants to do — it’s comme ci comme ça — and he just wants her."
|
98 |
+
An official music video to accompany the release of "Comme ci comme ça" was first
|
99 |
+
released onto YouTube on 20 September 2017 at a total length of three minutes
|
100 |
+
and twelve seconds. Comme ci comme ça (Basim song) "Comme ci comme ça" is a song
|
101 |
+
performed by Danish pop singer and songwriter Basim, featuring vocals from Gilli.
|
102 |
+
- Personal life of Frank Sinatra A third child, Christina Sinatra, known as "Tina",
|
103 |
+
was born on June 20, 1948. Nancy Barbato Sinatra and Frank Sinatra announced their
|
104 |
+
separation on Valentine's Day, February 14, 1950, with Frank's additional extra-marital
|
105 |
+
affair with Ava Gardner compounding his transgressions and becoming public knowledge
|
106 |
+
once again. After originally just seeking a legal separation, Frank and Nancy
|
107 |
+
Sinatra decided some months later to file for divorce, and this divorce became
|
108 |
+
legally final on October 29, 1951. Frank Sinatra's affair and relationship with
|
109 |
+
Gardner had become more and more serious, and she later became his second wife.
|
110 |
+
What was perhaps less widely
|
111 |
+
- source_sentence: What was the name of the first Indiana Jones movie?
|
112 |
+
sentences:
|
113 |
+
- Indiana Jones and the Temple of Doom point. Old-time, 15-part movie serials didn't
|
114 |
+
have shape. They just went on and on and on, which is what "Temple of Doom" does
|
115 |
+
with humor and technical invention." Neal Gabler commented that "I think in some
|
116 |
+
ways, "Indiana Jones and the Temple of Doom" was better than "Raiders of the Lost
|
117 |
+
Ark". In some ways it was less. In sum total, I'd have to say I enjoyed it more.
|
118 |
+
That doesn't mean it's better necessarily, but I got more enjoyment out of it."
|
119 |
+
Colin Covert of the "Star Tribune" called the film "sillier, darkly violent and
|
120 |
+
a bit dumbed down,
|
121 |
+
- Indiana Jones and the Temple of Doom (1985 video game) Theme music plays in the
|
122 |
+
background which is the best part of the game. Most of the sound effects are not
|
123 |
+
sharp and not enough of them exist. "Indiana Jones and the Temple of Doom" is
|
124 |
+
a bad game all the way around. It looks bad, has bad controls, and is way too
|
125 |
+
short." Indiana Jones and the Temple of Doom (1985 video game) Indiana Jones and
|
126 |
+
The Temple of Doom is a 1985 action arcade game developed and published by Atari
|
127 |
+
Games, based on the 1984 film of the same name, the second film in the "Indiana
|
128 |
+
Jones" franchise.
|
129 |
+
- Indiana Jones and the Spear of Destiny Indiana Jones and the Spear of Destiny
|
130 |
+
Indiana Jones and The Spear of Destiny is a four-issue comic book mini-series
|
131 |
+
published by Dark Horse Comics from April to July 1995. It was their seventh series
|
132 |
+
about the adult Indiana Jones. Indiana Jones reached for the Holy Grail, perched
|
133 |
+
in a crack in the Temple of the Sun. Hanging onto him, his father, Professor Henry
|
134 |
+
Jones urged him to let it go, and Indy turned back and let his father help him
|
135 |
+
up. As the Joneses ride out into the Canyon of the Crescent Moon with Marcus Brody
|
136 |
+
and Sallah, they
|
137 |
+
- Lego Indiana Jones sets" The line was discontinued in 2010, but since Lucas plans
|
138 |
+
to make a fifth installment to the franchise, the sets may be re-released along
|
139 |
+
with new sets of the possible fifth Indiana Jones film. Due to the fact Disney
|
140 |
+
bought Lucasfilm and will be making a new Indiana Jones movie, chances of new
|
141 |
+
sets are high. The Indiana Jones sets proved to be one of the most popular Lego
|
142 |
+
themes, and by the end of 2008 were credited, along with Lego Star Wars, of boosting
|
143 |
+
the Lego Group's profits within a stagnant toy market. The product line was said
|
144 |
+
- Indiana Jones and the Staff of Kings point-and-click adventure "Indiana Jones
|
145 |
+
and the Fate of Atlantis". GameSpot criticized its "terribly laid-out checkpoints",
|
146 |
+
"out-of-date" visuals, and "atrocious, annoying motion controls". Indiana Jones
|
147 |
+
and the Staff of Kings The game was initially developed for the higher-end PlayStation
|
148 |
+
3 and Xbox 360 systems, before switching to the aforementioned lower-end platforms.
|
149 |
+
As a result, both systems never saw a proper "Indiana Jones" video game being
|
150 |
+
released besides the "" duology. The plot centers around Indy's search for the
|
151 |
+
Staff of Moses. The Wii version of the game includes an exclusive co-op story
|
152 |
+
mode (with Indy and Henry Jones Sr.) and unlockable
|
153 |
+
- 'Indiana Jones and the Last Crusade: The Graphic Adventure Indiana Jones and the
|
154 |
+
Last Crusade: The Graphic Adventure Indiana Jones and the Last Crusade: The Graphic
|
155 |
+
Adventure is a graphic adventure game, released in 1989 (to coincide with the
|
156 |
+
release of the film of the same name), published by Lucasfilm Games (now LucasArts).
|
157 |
+
It was the third game to use the SCUMM engine. "Last Crusade" was one of the most
|
158 |
+
innovative of the LucasArts adventures. It expanded on LucasArts'' traditional
|
159 |
+
adventure game structure by including a flexible point system—the IQ score, or
|
160 |
+
"Indy Quotient"—and by allowing the game to be completed in several different
|
161 |
+
ways. The point system was'
|
162 |
+
- source_sentence: '"Who was the Anglo-Irish scientist who, in the 17th century, discovered
|
163 |
+
that ""the volume of a given mass of gas at a given temperature is inversely proportional
|
164 |
+
to its pressure""?"'
|
165 |
+
sentences:
|
166 |
+
- 'Gay-Lussac''s law Gay-Lussac''s law Gay-Lussac''s law can refer to several discoveries
|
167 |
+
made by French chemist Joseph Louis Gay-Lussac (1778–1850) and other scientists
|
168 |
+
in the late 18th and early 19th centuries pertaining to thermal expansion of gases
|
169 |
+
and the relationship between temperature, volume, and pressure. It states that
|
170 |
+
the pressure of a given mass of gas varies directly with the absolute temperature
|
171 |
+
of the gas, when the volume is kept constant. Mathematically, it can be written
|
172 |
+
as: P/T=constant, Gay-Lussac is most often recognized for the Pressure Law which
|
173 |
+
established that the pressure of an enclosed gas is directly proportional to its
|
174 |
+
temperature and'
|
175 |
+
- 'Gas constant "V" is the volume of gas (SI unit cubic metres), "n" is the amount
|
176 |
+
of gas (SI unit moles), "m" is the mass (SI unit kilograms) contained in "V",
|
177 |
+
and "T" is the thermodynamic temperature (SI unit kelvins). "R" is the molar-weight-specific
|
178 |
+
gas constant, discussed below. The gas constant is expressed in the same physical
|
179 |
+
units as molar entropy and molar heat capacity. From the general equation "PV"
|
180 |
+
= "nRT" we get: where "P" is pressure, "V" is volume, "n" is number of moles of
|
181 |
+
a given substance, and "T" is temperature. As pressure is defined as force per
|
182 |
+
unit'
|
183 |
+
- The Boy Who Was a King term. The film presents not only the life of the former
|
184 |
+
Tsar, but also intertwines within the story vignettes of various Bulgarians, who
|
185 |
+
were supporting him, sending him gifts, or merely tattooing his face on their
|
186 |
+
body. The story is told through personal footage and vast amounts of archive material.
|
187 |
+
The film received praise for its editing and use of archives with Variety's Robert
|
188 |
+
Koehler writing that "Pic’s terrific use of archival footage includes an exiled
|
189 |
+
Simeon interviewed in the early ’60s, disputing his playboy rep." and "Editing
|
190 |
+
is aces." The Boy Who Was a King The Boy Who Was
|
191 |
+
- Francis Hauksbee In 1708, Hauksbee independently discovered Charles's law of gases,
|
192 |
+
which states that, for a given mass of gas at a constant pressure, the volume
|
193 |
+
of the gas is proportional to its temperature. Hauksbee published accounts of
|
194 |
+
his experiments in the Royal Society's journal "Philosophical Transactions". In
|
195 |
+
1709 he self-published "Physico-Mechanical Experiments on Various Subjects" which
|
196 |
+
collected together many of these experiments along with discussion that summarized
|
197 |
+
much of his scientific work. An Italian translation was published in 1716. A second
|
198 |
+
edition was published posthumously in 1719. There were also translations to Dutch
|
199 |
+
(1735) and French (1754). The Royal Society Hauksbee
|
200 |
+
- 'Boyle''s law air moves from high to low pressure. Related phenomena: Other gas
|
201 |
+
laws: Boyle''s law Boyle''s law, sometimes referred to as the Boyle–Mariotte law,
|
202 |
+
or Mariotte''s law (especially in France), is an experimental gas law that describes
|
203 |
+
how the pressure of a gas tends to increase as the volume of the container decreases.
|
204 |
+
A modern statement of Boyle''s law is The absolute pressure exerted by a given
|
205 |
+
mass of an ideal gas is inversely proportional to the volume it occupies if the
|
206 |
+
temperature and amount of gas remain unchanged within a closed system. Mathematically,
|
207 |
+
Boyle''s law can be stated as or'
|
208 |
+
- Boyle's law of the gas, and "k" is a constant. The equation states that the product
|
209 |
+
of pressure and volume is a constant for a given mass of confined gas and this
|
210 |
+
holds as long as the temperature is constant. For comparing the same substance
|
211 |
+
under two different sets of conditions, the law can be usefully expressed as The
|
212 |
+
equation shows that, as volume increases, the pressure of the gas decreases in
|
213 |
+
proportion. Similarly, as volume decreases, the pressure of the gas increases.
|
214 |
+
The law was named after chemist and physicist Robert Boyle, who published the
|
215 |
+
original law in 1662. This relationship
|
216 |
+
- source_sentence: Peter Stuyvesant, born in Holland, became Governor of which American
|
217 |
+
city in 1647?
|
218 |
+
sentences:
|
219 |
+
- Peter Stuyvesant at the corner of Thirteenth Street and Third Avenue until 1867
|
220 |
+
when it was destroyed by a storm, bearing fruit almost to the last. The house
|
221 |
+
was destroyed by fire in 1777. He also built an executive mansion of stone called
|
222 |
+
Whitehall. In 1645, Stuyvesant married Judith Bayard (–1687) of the Bayard family.
|
223 |
+
Her brother, Samuel Bayard, was the husband of Stuyvesant's sister, Anna Stuyvesant.
|
224 |
+
Petrus and Judith had two sons together. He died in August 1672 and his body was
|
225 |
+
entombed in the east wall of St. Mark's Church in-the-Bowery, which sits on the
|
226 |
+
site of Stuyvesant’s family chapel.
|
227 |
+
- 'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal
|
228 |
+
prosecution - if companies wilfully break the laws. However last year, when questioned
|
229 |
+
on why no such action was being pursued against Imperial Tobacco a spokeswoman
|
230 |
+
for Federal Health said: ""No instances of non-compliance with the Act have been
|
231 |
+
identified by the Department that warrant the initiation of Court proceedings
|
232 |
+
in the first instance, and without attempting alternative dispute resolution to
|
233 |
+
achieve compliance"". Peter Stuyvesant is or was sold in the following countries:
|
234 |
+
Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany,
|
235 |
+
France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,'
|
236 |
+
- Jochem Pietersen Kuyter September 25, 1647, until the city was incorporated, in
|
237 |
+
1653, when he was made schout (sheriff). Kuyter twice came in conflict with the
|
238 |
+
Director of New Netherland. Kuyter was a man of good education, what is evident
|
239 |
+
by his dealings with Willem Kieft., who he believed damaged the colony with his
|
240 |
+
policies and the start of Kieft's War in 1643. In 1647, when Peter Stuyvesant
|
241 |
+
arrived in New Amsterdam to replace Kieft, Kuyter and Cornelis Melyn acting in
|
242 |
+
name of the citizens of New Amsterdam, brought charges against the outgoing governor,
|
243 |
+
demanding an investigation of his conduct while in office.
|
244 |
+
- Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging
|
245 |
+
changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported
|
246 |
+
that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers
|
247 |
+
which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came
|
248 |
+
with 26 cigarettes. The material included images of a young woman with pink hair
|
249 |
+
putting on lipstick and men on the streets of New York and also included a calendar
|
250 |
+
and small poster that were clearly intended to glamorise smoking. Anti-smoking
|
251 |
+
campaigner Mike Daube said although the material did not break the law because
|
252 |
+
- 'Peter Stuyvesant but the order was soon revoked under pressure from the States
|
253 |
+
of Holland and the city of Amsterdam. Stuyvesant prepared against an attack by
|
254 |
+
ordering the citizens to dig a ditch from the North River to the East River and
|
255 |
+
to erect a fortification. In 1653, a convention of two deputies from each village
|
256 |
+
in New Netherland demanded reforms, and Stuyvesant commanded that assembly to
|
257 |
+
disperse, saying: "We derive our authority from God and the company, not from
|
258 |
+
a few ignorant subjects." In the summer of 1655, he sailed down the Delaware River
|
259 |
+
with a fleet of seven vessels and'
|
260 |
+
- Peter Stuyvesant Dutch Reformed church, a Calvinist denomination, holding to the
|
261 |
+
Three Forms of Unity (Belgic Confession, Heidelberg Catechism, Canons of Dordt).
|
262 |
+
The English were Anglicans, holding to the 39 Articles, a Protestant confession,
|
263 |
+
with bishops. In 1665, Stuyvesant went to the Netherlands to report on his term
|
264 |
+
as governor. On his return to the colony, he spent the remainder of his life on
|
265 |
+
his farm of sixty-two acres outside the city, called the Great Bouwerie, beyond
|
266 |
+
which stretched the woods and swamps of the village of Nieuw Haarlem. A pear tree
|
267 |
+
that he reputedly brought from the Netherlands in 1647 remained
|
268 |
+
pipeline_tag: sentence-similarity
|
269 |
+
library_name: sentence-transformers
|
270 |
+
---
|
271 |
+
|
272 |
+
# SentenceTransformer based on seongil-dn/unsupervised_20m_3800
|
273 |
+
|
274 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
275 |
+
|
276 |
+
## Model Details
|
277 |
+
|
278 |
+
### Model Description
|
279 |
+
- **Model Type:** Sentence Transformer
|
280 |
+
- **Base model:** [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800) <!-- at revision 1cda749f242e2b5c9e4f3c1122a61e76fec1fee5 -->
|
281 |
+
- **Maximum Sequence Length:** 1024 tokens
|
282 |
+
- **Output Dimensionality:** 1024 dimensions
|
283 |
+
- **Similarity Function:** Cosine Similarity
|
284 |
+
<!-- - **Training Dataset:** Unknown -->
|
285 |
+
<!-- - **Language:** Unknown -->
|
286 |
+
<!-- - **License:** Unknown -->
|
287 |
+
|
288 |
+
### Model Sources
|
289 |
+
|
290 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
291 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
292 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
293 |
+
|
294 |
+
### Full Model Architecture
|
295 |
+
|
296 |
+
```
|
297 |
+
SentenceTransformer(
|
298 |
+
(0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
|
299 |
+
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
300 |
+
(2): Normalize()
|
301 |
+
)
|
302 |
+
```
|
303 |
+
|
304 |
+
## Usage
|
305 |
+
|
306 |
+
### Direct Usage (Sentence Transformers)
|
307 |
+
|
308 |
+
First install the Sentence Transformers library:
|
309 |
+
|
310 |
+
```bash
|
311 |
+
pip install -U sentence-transformers
|
312 |
+
```
|
313 |
+
|
314 |
+
Then you can load this model and run inference.
|
315 |
+
```python
|
316 |
+
from sentence_transformers import SentenceTransformer
|
317 |
+
|
318 |
+
# Download from the 🤗 Hub
|
319 |
+
model = SentenceTransformer("seongil-dn/bge-m3-831")
|
320 |
+
# Run inference
|
321 |
+
sentences = [
|
322 |
+
'Peter Stuyvesant, born in Holland, became Governor of which American city in 1647?',
|
323 |
+
'Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came with 26 cigarettes. The material included images of a young woman with pink hair putting on lipstick and men on the streets of New York and also included a calendar and small poster that were clearly intended to glamorise smoking. Anti-smoking campaigner Mike Daube said although the material did not break the law because',
|
324 |
+
'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal prosecution - if companies wilfully break the laws. However last year, when questioned on why no such action was being pursued against Imperial Tobacco a spokeswoman for Federal Health said: ""No instances of non-compliance with the Act have been identified by the Department that warrant the initiation of Court proceedings in the first instance, and without attempting alternative dispute resolution to achieve compliance"". Peter Stuyvesant is or was sold in the following countries: Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany, France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,',
|
325 |
+
]
|
326 |
+
embeddings = model.encode(sentences)
|
327 |
+
print(embeddings.shape)
|
328 |
+
# [3, 1024]
|
329 |
+
|
330 |
+
# Get the similarity scores for the embeddings
|
331 |
+
similarities = model.similarity(embeddings, embeddings)
|
332 |
+
print(similarities.shape)
|
333 |
+
# [3, 3]
|
334 |
+
```
|
335 |
+
|
336 |
+
<!--
|
337 |
+
### Direct Usage (Transformers)
|
338 |
+
|
339 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
340 |
+
|
341 |
+
</details>
|
342 |
+
-->
|
343 |
+
|
344 |
+
<!--
|
345 |
+
### Downstream Usage (Sentence Transformers)
|
346 |
+
|
347 |
+
You can finetune this model on your own dataset.
|
348 |
+
|
349 |
+
<details><summary>Click to expand</summary>
|
350 |
+
|
351 |
+
</details>
|
352 |
+
-->
|
353 |
+
|
354 |
+
<!--
|
355 |
+
### Out-of-Scope Use
|
356 |
+
|
357 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
358 |
+
-->
|
359 |
+
|
360 |
+
<!--
|
361 |
+
## Bias, Risks and Limitations
|
362 |
+
|
363 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
364 |
+
-->
|
365 |
+
|
366 |
+
<!--
|
367 |
+
### Recommendations
|
368 |
+
|
369 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
370 |
+
-->
|
371 |
+
|
372 |
+
## Training Details
|
373 |
+
|
374 |
+
### Training Dataset
|
375 |
+
|
376 |
+
#### Unnamed Dataset
|
377 |
+
|
378 |
+
* Size: 1,138,596 training samples
|
379 |
+
* Columns: <code>anchor</code>, <code>positive</code>, <code>negative</code>, <code>negative_2</code>, <code>negative_3</code>, <code>negative_4</code>, and <code>negative_5</code>
|
380 |
+
* Approximate statistics based on the first 1000 samples:
|
381 |
+
| | anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 |
|
382 |
+
|:--------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|
|
383 |
+
| type | string | string | string | string | string | string | string |
|
384 |
+
| details | <ul><li>min: 9 tokens</li><li>mean: 22.32 tokens</li><li>max: 119 tokens</li></ul> | <ul><li>min: 127 tokens</li><li>mean: 157.45 tokens</li><li>max: 420 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 154.65 tokens</li><li>max: 212 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 155.52 tokens</li><li>max: 218 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 156.04 tokens</li><li>max: 284 tokens</li></ul> | <ul><li>min: 124 tokens</li><li>mean: 156.3 tokens</li><li>max: 268 tokens</li></ul> | <ul><li>min: 121 tokens</li><li>mean: 156.15 tokens</li><li>max: 249 tokens</li></ul> |
|
385 |
+
* Samples:
|
386 |
+
| anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 |
|
387 |
+
|:---------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
388 |
+
| <code>What African country is projected to pass the United States in population by the year 2055?</code> | <code>African immigration to the United States officially 40,000 African immigrants, although it has been estimated that the population is actually four times this number when considering undocumented immigrants. The majority of these immigrants were born in Ethiopia, Egypt, Nigeria, and South Africa. African immigrants like many other immigrant groups are likely to establish and find success in small businesses. Many Africans that have seen the social and economic stability that comes from ethnic enclaves such as Chinatowns have recently been establishing ethnic enclaves of their own at much higher rates to reap the benefits of such communities. Such examples include Little Ethiopia in Los Angeles and</code> | <code>What Will Happen to the Gang Next Year? watching television at the time of the broadcast. This made it the lowest-rated episode in "30 Rock"<nowiki>'</nowiki>s history. and a decrease from the previous episode "The Return of Avery Jessup" (2.92 million) What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States</code> | <code>Christianity in the United States Christ is the fifth-largest denomination, the largest Pentecostal church, and the largest traditionally African-American denomination in the nation. Among Eastern Christian denominations, there are several Eastern Orthodox and Oriental Orthodox churches, with just below 1 million adherents in the US, or 0.4% of the total population. Christianity was introduced to the Americas as it was first colonized by Europeans beginning in the 16th and 17th centuries. Going forward from its foundation, the United States has been called a Protestant nation by a variety of sources. Immigration further increased Christian numbers. Today most Christian churches in the United States are either</code> | <code>What Will Happen to the Gang Next Year? What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States on May 17, 2012. In the episode, Jack (Alec Baldwin) and Avery (Elizabeth Banks) seek to renew their vows; Criss (James Marsden) sets out to show Liz (Tina Fey) he can pay</code> | <code>History of the Jews in the United States Representatives by Rep. Samuel Dickstein (D; New York). This also failed to pass. During the Holocaust, fewer than 30,000 Jews a year reached the United States, and some were turned away due to immigration policies. The U.S. did not change its immigration policies until 1948. Currently, laws requiring teaching of the Holocaust are on the books in five states. The Holocaust had a profound impact on the community in the United States, especially after 1960, as Jews tried to comprehend what had happened, and especially to commemorate and grapple with it when looking to the future. Abraham Joshua Heschel summarized</code> | <code>Public holidays in the United States will have very few customers that day. The labor force in the United States comprises about 62% (as of 2014) of the general population. In the United States, 97% of the private sector businesses determine what days this sector of the population gets paid time off, according to a study by the Society for Human Resource Management. The following holidays are observed by the majority of US businesses with paid time off: This list of holidays is based off the official list of federal holidays by year from the US Government. The holidays however are at the discretion of employers</code> |
|
389 |
+
| <code>Which is the largest species of the turtle family?</code> | <code>Loggerhead sea turtle turtle is debated, but most authors consider it a single polymorphic species. Molecular genetics has confirmed hybridization of the loggerhead sea turtle with the Kemp's ridley sea turtle, hawksbill sea turtle, and green sea turtles. The extent of natural hybridization is not yet determined; however, second-generation hybrids have been reported, suggesting some hybrids are fertile. Although evidence is lacking, modern sea turtles probably descended from a single common ancestor during the Cretaceous period. Like all other sea turtles except the leatherback, loggerheads are members of the ancient family Cheloniidae, and appeared about 40 million years ago. Of the six species</code> | <code>Convention on the Conservation of Migratory Species of Wild Animals take joint action. At May 2018, there were 126 Parties to the Convention. The CMS Family covers a great diversity of migratory species. The Appendices of CMS include many mammals, including land mammals, marine mammals and bats; birds; fish; reptiles and one insect. Among the instruments, AEWA covers 254 species of birds that are ecologically dependent on wetlands for at least part of their annual cycle. EUROBATS covers 52 species of bat, the Memorandum of Understanding on the Conservation of Migratory Sharks seven species of shark, the IOSEA Marine Turtle MOU six species of marine turtle and the Raptors MoU</code> | <code>Razor-backed musk turtle Razor-backed musk turtle The razor-backed musk turtle ("Sternotherus carinatus") is a species of turtle in the family Kinosternidae. The species is native to the southern United States. There are no subspecies that are recognized as being valid. "S. carinatus" is found in the states of Alabama, Arkansas, Louisiana, Mississippi, Oklahoma, and Texas. The razor-backed musk turtle grows to a straight carapace length of about . It has a brown-colored carapace, with black markings at the edges of each scute. The carapace has a distinct, sharp keel down the center of its length, giving the species its common name. The body</code> | <code>African helmeted turtle African helmeted turtle The African helmeted turtle ("Pelomedusa subrufa"), also known commonly as the marsh terrapin, the crocodile turtle, or in the pet trade as the African side-necked turtle, is a species of omnivorous side-necked terrapin in the family Pelomedusidae. The species naturally occurs in fresh and stagnant water bodies throughout much of Sub-Saharan Africa, and in southern Yemen. The marsh terrapin is typically a rather small turtle, with most individuals being less than in straight carapace length, but one has been recorded with a length of . It has a black or brown carapace. The top of the tail</code> | <code>Box turtle Box turtle Box turtles are North American turtles of the genus Terrapene. Although box turtles are superficially similar to tortoises in terrestrial habits and overall appearance, they are actually members of the American pond turtle family (Emydidae). The twelve taxa which are distinguished in the genus are distributed over four species. They are largely characterized by having a domed shell, which is hinged at the bottom, allowing the animal to close its shell tightly to escape predators. The genus name "Terrapene" was coined by Merrem in 1820 as a genus separate from "Emys" for those species which had a sternum</code> | <code>Vallarta mud turtle Vallarta mud turtle The Vallarta mud turtle ("Kinosternon vogti") is a recently identified species of mud turtle in the family Kinosternidae. While formerly considered conspecific with the Jalisco mud turtle, further studies indicated that it was a separate species. It can be identified by a combination of the number of plastron and carapace scutes, body size, and the distinctive yellow rostral shield in males. It is endemic to Mexican state of Jalisco. It is only known from a few human-created or human-affected habitats (such as small streams and ponds) found around Puerto Vallarta. It is one of only 3 species</code> |
|
390 |
+
| <code>How many gallons of beer are in an English barrel?</code> | <code>Low-alcohol beer Prohibition in the United States. Near beer could not legally be labeled as "beer" and was officially classified as a "cereal beverage". The public, however, almost universally called it "near beer". The most popular "near beer" was Bevo, brewed by the Anheuser-Busch company. The Pabst company brewed "Pablo", Miller brewed "Vivo", and Schlitz brewed "Famo". Many local and regional breweries stayed in business by marketing their own near-beers. By 1921 production of near beer had reached over 300 million US gallons (1 billion L) a year (36 L/s). A popular illegal practice was to add alcohol to near beer. The</code> | <code>Keg terms "half-barrel" and "quarter-barrel" are derived from the U.S. beer barrel, legally defined as being equal to 31 U.S. gallons (this is not the same volume as some other units also known as "barrels"). A 15.5 U.S. gallon keg is also equal to: However, beer kegs can come in many sizes: In European countries the most common keg size is 50 liters. This includes the UK, which uses a non-metric standard keg of 11 imperial gallons, which is coincidentally equal to . The German DIN 6647-1 and DIN 6647-2 have also defined kegs in the sizes of 30 and 20</code> | <code>Beer in Chile craft beers. They are generally low or very low volume producers. In Chile there are more than 150 craft beer producers distributed along the 15 Chilean Regions. The list below includes: Beer in Chile The primary beer brewed and consumed in Chile is pale lager, though the country also has a tradition of brewing corn beer, known as chicha. Chile’s beer history has a strong German influence – some of the bigger beer producers are from the country’s southern lake district, a region populated by a great number of German immigrants during the 19th century. Chile also produces English ale-style</code> | <code>Barrel variation. In modern times, produce barrels for all dry goods, excepting cranberries, contain 7,056 cubic inches, about 115.627 L. Barrel A barrel, cask, or tun is a hollow cylindrical container, traditionally made of wooden staves bound by wooden or metal hoops. Traditionally, the barrel was a standard size of measure referring to a set capacity or weight of a given commodity. For example, in the UK a barrel of beer refers to a quantity of . Wine was shipped in barrels of . Modern wooden barrels for wine-making are either made of French common oak ("Quercus robur") and white oak</code> | <code>The Rare Barrel The Rare Barrel The Rare Barrel is a brewery and brewpub in Berkeley, California, United States, that exclusively produces sour beers. Founders Jay Goodwin and Alex Wallash met while attending UCSB. They started home-brewing in their apartment and decided that they would one day start a brewery together. Goodwin started working at The Bruery, where he worked his way from a production assistant to brewer, eventually becoming the head of their barrel aging program. The Rare Barrel brewed its first batch of beer in February 2013, and opened its tasting room on December 27, 2013. The Rare Barrel was named</code> | <code>Barrel (unit) Barrel (unit) A barrel is one of several units of volume applied in various contexts; there are dry barrels, fluid barrels (such as the UK beer barrel and US beer barrel), oil barrels and so on. For historical reasons the volumes of some barrel units are roughly double the volumes of others; volumes in common usage range from about . In many connections the term "drum" is used almost interchangeably with "barrel". Since medieval times the term barrel as a unit of measure has had various meanings throughout Europe, ranging from about 100 litres to 1000 litres. The name was</code> |
|
391 |
+
* Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
|
392 |
+
```json
|
393 |
+
{'guide': SentenceTransformer(
|
394 |
+
(0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
|
395 |
+
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
396 |
+
(2): Normalize()
|
397 |
+
), 'temperature': 0.01}
|
398 |
+
```
|
399 |
+
|
400 |
+
### Training Hyperparameters
|
401 |
+
#### Non-Default Hyperparameters
|
402 |
+
|
403 |
+
- `per_device_train_batch_size`: 1024
|
404 |
+
- `learning_rate`: 3e-05
|
405 |
+
- `weight_decay`: 0.01
|
406 |
+
- `warmup_ratio`: 0.05
|
407 |
+
- `bf16`: True
|
408 |
+
- `batch_sampler`: no_duplicates
|
409 |
+
|
410 |
+
#### All Hyperparameters
|
411 |
+
<details><summary>Click to expand</summary>
|
412 |
+
|
413 |
+
- `overwrite_output_dir`: False
|
414 |
+
- `do_predict`: False
|
415 |
+
- `eval_strategy`: no
|
416 |
+
- `prediction_loss_only`: True
|
417 |
+
- `per_device_train_batch_size`: 1024
|
418 |
+
- `per_device_eval_batch_size`: 8
|
419 |
+
- `per_gpu_train_batch_size`: None
|
420 |
+
- `per_gpu_eval_batch_size`: None
|
421 |
+
- `gradient_accumulation_steps`: 1
|
422 |
+
- `eval_accumulation_steps`: None
|
423 |
+
- `torch_empty_cache_steps`: None
|
424 |
+
- `learning_rate`: 3e-05
|
425 |
+
- `weight_decay`: 0.01
|
426 |
+
- `adam_beta1`: 0.9
|
427 |
+
- `adam_beta2`: 0.999
|
428 |
+
- `adam_epsilon`: 1e-08
|
429 |
+
- `max_grad_norm`: 1.0
|
430 |
+
- `num_train_epochs`: 3
|
431 |
+
- `max_steps`: -1
|
432 |
+
- `lr_scheduler_type`: linear
|
433 |
+
- `lr_scheduler_kwargs`: {}
|
434 |
+
- `warmup_ratio`: 0.05
|
435 |
+
- `warmup_steps`: 0
|
436 |
+
- `log_level`: passive
|
437 |
+
- `log_level_replica`: warning
|
438 |
+
- `log_on_each_node`: True
|
439 |
+
- `logging_nan_inf_filter`: True
|
440 |
+
- `save_safetensors`: True
|
441 |
+
- `save_on_each_node`: False
|
442 |
+
- `save_only_model`: False
|
443 |
+
- `restore_callback_states_from_checkpoint`: False
|
444 |
+
- `no_cuda`: False
|
445 |
+
- `use_cpu`: False
|
446 |
+
- `use_mps_device`: False
|
447 |
+
- `seed`: 42
|
448 |
+
- `data_seed`: None
|
449 |
+
- `jit_mode_eval`: False
|
450 |
+
- `use_ipex`: False
|
451 |
+
- `bf16`: True
|
452 |
+
- `fp16`: False
|
453 |
+
- `fp16_opt_level`: O1
|
454 |
+
- `half_precision_backend`: auto
|
455 |
+
- `bf16_full_eval`: False
|
456 |
+
- `fp16_full_eval`: False
|
457 |
+
- `tf32`: None
|
458 |
+
- `local_rank`: 0
|
459 |
+
- `ddp_backend`: None
|
460 |
+
- `tpu_num_cores`: None
|
461 |
+
- `tpu_metrics_debug`: False
|
462 |
+
- `debug`: []
|
463 |
+
- `dataloader_drop_last`: True
|
464 |
+
- `dataloader_num_workers`: 0
|
465 |
+
- `dataloader_prefetch_factor`: None
|
466 |
+
- `past_index`: -1
|
467 |
+
- `disable_tqdm`: False
|
468 |
+
- `remove_unused_columns`: True
|
469 |
+
- `label_names`: None
|
470 |
+
- `load_best_model_at_end`: False
|
471 |
+
- `ignore_data_skip`: False
|
472 |
+
- `fsdp`: []
|
473 |
+
- `fsdp_min_num_params`: 0
|
474 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
475 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
476 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
477 |
+
- `deepspeed`: None
|
478 |
+
- `label_smoothing_factor`: 0.0
|
479 |
+
- `optim`: adamw_torch
|
480 |
+
- `optim_args`: None
|
481 |
+
- `adafactor`: False
|
482 |
+
- `group_by_length`: False
|
483 |
+
- `length_column_name`: length
|
484 |
+
- `ddp_find_unused_parameters`: None
|
485 |
+
- `ddp_bucket_cap_mb`: None
|
486 |
+
- `ddp_broadcast_buffers`: False
|
487 |
+
- `dataloader_pin_memory`: True
|
488 |
+
- `dataloader_persistent_workers`: False
|
489 |
+
- `skip_memory_metrics`: True
|
490 |
+
- `use_legacy_prediction_loop`: False
|
491 |
+
- `push_to_hub`: False
|
492 |
+
- `resume_from_checkpoint`: None
|
493 |
+
- `hub_model_id`: None
|
494 |
+
- `hub_strategy`: every_save
|
495 |
+
- `hub_private_repo`: None
|
496 |
+
- `hub_always_push`: False
|
497 |
+
- `gradient_checkpointing`: False
|
498 |
+
- `gradient_checkpointing_kwargs`: None
|
499 |
+
- `include_inputs_for_metrics`: False
|
500 |
+
- `include_for_metrics`: []
|
501 |
+
- `eval_do_concat_batches`: True
|
502 |
+
- `fp16_backend`: auto
|
503 |
+
- `push_to_hub_model_id`: None
|
504 |
+
- `push_to_hub_organization`: None
|
505 |
+
- `mp_parameters`:
|
506 |
+
- `auto_find_batch_size`: False
|
507 |
+
- `full_determinism`: False
|
508 |
+
- `torchdynamo`: None
|
509 |
+
- `ray_scope`: last
|
510 |
+
- `ddp_timeout`: 1800
|
511 |
+
- `torch_compile`: False
|
512 |
+
- `torch_compile_backend`: None
|
513 |
+
- `torch_compile_mode`: None
|
514 |
+
- `dispatch_batches`: None
|
515 |
+
- `split_batches`: None
|
516 |
+
- `include_tokens_per_second`: False
|
517 |
+
- `include_num_input_tokens_seen`: False
|
518 |
+
- `neftune_noise_alpha`: None
|
519 |
+
- `optim_target_modules`: None
|
520 |
+
- `batch_eval_metrics`: False
|
521 |
+
- `eval_on_start`: False
|
522 |
+
- `use_liger_kernel`: False
|
523 |
+
- `eval_use_gather_object`: False
|
524 |
+
- `average_tokens_across_devices`: False
|
525 |
+
- `prompts`: None
|
526 |
+
- `batch_sampler`: no_duplicates
|
527 |
+
- `multi_dataset_batch_sampler`: proportional
|
528 |
+
|
529 |
+
</details>
|
530 |
+
|
531 |
+
### Training Logs
|
532 |
+
<details><summary>Click to expand</summary>
|
533 |
+
|
534 |
+
| Epoch | Step | Training Loss |
|
535 |
+
|:------:|:----:|:-------------:|
|
536 |
+
| 0.0036 | 1 | 1.0283 |
|
537 |
+
| 0.0072 | 2 | 1.0155 |
|
538 |
+
| 0.0108 | 3 | 0.9858 |
|
539 |
+
| 0.0144 | 4 | 0.9519 |
|
540 |
+
| 0.0181 | 5 | 0.9434 |
|
541 |
+
| 0.0217 | 6 | 0.898 |
|
542 |
+
| 0.0253 | 7 | 0.8798 |
|
543 |
+
| 0.0289 | 8 | 0.7976 |
|
544 |
+
| 0.0325 | 9 | 0.7797 |
|
545 |
+
| 0.0361 | 10 | 0.7464 |
|
546 |
+
| 0.0397 | 11 | 0.743 |
|
547 |
+
| 0.0433 | 12 | 0.716 |
|
548 |
+
| 0.0469 | 13 | 0.7076 |
|
549 |
+
| 0.0505 | 14 | 0.666 |
|
550 |
+
| 0.0542 | 15 | 0.631 |
|
551 |
+
| 0.0578 | 16 | 0.5905 |
|
552 |
+
| 0.0614 | 17 | 0.6537 |
|
553 |
+
| 0.0650 | 18 | 0.5755 |
|
554 |
+
| 0.0686 | 19 | 0.5422 |
|
555 |
+
| 0.0722 | 20 | 0.5393 |
|
556 |
+
| 0.0758 | 21 | 0.5741 |
|
557 |
+
| 0.0794 | 22 | 0.498 |
|
558 |
+
| 0.0830 | 23 | 0.5522 |
|
559 |
+
| 0.0866 | 24 | 0.5592 |
|
560 |
+
| 0.0903 | 25 | 0.4797 |
|
561 |
+
| 0.0939 | 26 | 0.4684 |
|
562 |
+
| 0.0975 | 27 | 0.5207 |
|
563 |
+
| 0.1011 | 28 | 0.4692 |
|
564 |
+
| 0.1047 | 29 | 0.4459 |
|
565 |
+
| 0.1083 | 30 | 0.4439 |
|
566 |
+
| 0.1119 | 31 | 0.4656 |
|
567 |
+
| 0.1155 | 32 | 0.4737 |
|
568 |
+
| 0.1191 | 33 | 0.4391 |
|
569 |
+
| 0.1227 | 34 | 0.4386 |
|
570 |
+
| 0.1264 | 35 | 0.4107 |
|
571 |
+
| 0.1300 | 36 | 0.4513 |
|
572 |
+
| 0.1336 | 37 | 0.3789 |
|
573 |
+
| 0.1372 | 38 | 0.4103 |
|
574 |
+
| 0.1408 | 39 | 0.3929 |
|
575 |
+
| 0.1444 | 40 | 0.4226 |
|
576 |
+
| 0.1480 | 41 | 0.391 |
|
577 |
+
| 0.1516 | 42 | 0.3674 |
|
578 |
+
| 0.1552 | 43 | 0.3607 |
|
579 |
+
| 0.1588 | 44 | 0.3738 |
|
580 |
+
| 0.1625 | 45 | 0.3842 |
|
581 |
+
| 0.1661 | 46 | 0.3498 |
|
582 |
+
| 0.1697 | 47 | 0.3586 |
|
583 |
+
| 0.1733 | 48 | 0.3538 |
|
584 |
+
| 0.1769 | 49 | 0.3572 |
|
585 |
+
| 0.1805 | 50 | 0.3547 |
|
586 |
+
| 0.1841 | 51 | 0.3179 |
|
587 |
+
| 0.1877 | 52 | 0.3436 |
|
588 |
+
| 0.1913 | 53 | 0.3502 |
|
589 |
+
| 0.1949 | 54 | 0.3381 |
|
590 |
+
| 0.1986 | 55 | 0.3547 |
|
591 |
+
| 0.2022 | 56 | 0.3362 |
|
592 |
+
| 0.2058 | 57 | 0.3407 |
|
593 |
+
| 0.2094 | 58 | 0.31 |
|
594 |
+
| 0.2130 | 59 | 0.3039 |
|
595 |
+
| 0.2166 | 60 | 0.3362 |
|
596 |
+
| 0.2202 | 61 | 0.2948 |
|
597 |
+
| 0.2238 | 62 | 0.3429 |
|
598 |
+
| 0.2274 | 63 | 0.3096 |
|
599 |
+
| 0.2310 | 64 | 0.35 |
|
600 |
+
| 0.2347 | 65 | 0.2997 |
|
601 |
+
| 0.2383 | 66 | 0.3258 |
|
602 |
+
| 0.2419 | 67 | 0.3376 |
|
603 |
+
| 0.2455 | 68 | 0.3213 |
|
604 |
+
| 0.2491 | 69 | 0.3185 |
|
605 |
+
| 0.2527 | 70 | 0.3282 |
|
606 |
+
| 0.2563 | 71 | 0.2988 |
|
607 |
+
| 0.2599 | 72 | 0.33 |
|
608 |
+
| 0.2635 | 73 | 0.3066 |
|
609 |
+
| 0.2671 | 74 | 0.3303 |
|
610 |
+
| 0.2708 | 75 | 0.3067 |
|
611 |
+
| 0.2744 | 76 | 0.2996 |
|
612 |
+
| 0.2780 | 77 | 0.3063 |
|
613 |
+
| 0.2816 | 78 | 0.3235 |
|
614 |
+
| 0.2852 | 79 | 0.2902 |
|
615 |
+
| 0.2888 | 80 | 0.302 |
|
616 |
+
| 0.2924 | 81 | 0.3223 |
|
617 |
+
| 0.2960 | 82 | 0.297 |
|
618 |
+
| 0.2996 | 83 | 0.2936 |
|
619 |
+
| 0.3032 | 84 | 0.3279 |
|
620 |
+
| 0.3069 | 85 | 0.2973 |
|
621 |
+
| 0.3105 | 86 | 0.2881 |
|
622 |
+
| 0.3141 | 87 | 0.3014 |
|
623 |
+
| 0.3177 | 88 | 0.2986 |
|
624 |
+
| 0.3213 | 89 | 0.3057 |
|
625 |
+
| 0.3249 | 90 | 0.2887 |
|
626 |
+
| 0.3285 | 91 | 0.2765 |
|
627 |
+
| 0.3321 | 92 | 0.2818 |
|
628 |
+
| 0.3357 | 93 | 0.2904 |
|
629 |
+
| 0.3394 | 94 | 0.267 |
|
630 |
+
| 0.3430 | 95 | 0.2948 |
|
631 |
+
| 0.3466 | 96 | 0.2766 |
|
632 |
+
| 0.3502 | 97 | 0.2782 |
|
633 |
+
| 0.3538 | 98 | 0.3082 |
|
634 |
+
| 0.3574 | 99 | 0.2697 |
|
635 |
+
| 0.3610 | 100 | 0.3006 |
|
636 |
+
| 0.3646 | 101 | 0.2986 |
|
637 |
+
| 0.3682 | 102 | 0.2789 |
|
638 |
+
| 0.3718 | 103 | 0.2756 |
|
639 |
+
| 0.3755 | 104 | 0.2884 |
|
640 |
+
| 0.3791 | 105 | 0.273 |
|
641 |
+
| 0.3827 | 106 | 0.2687 |
|
642 |
+
| 0.3863 | 107 | 0.2808 |
|
643 |
+
| 0.3899 | 108 | 0.2763 |
|
644 |
+
| 0.3935 | 109 | 0.2738 |
|
645 |
+
| 0.3971 | 110 | 0.2642 |
|
646 |
+
| 0.4007 | 111 | 0.2612 |
|
647 |
+
| 0.4043 | 112 | 0.2859 |
|
648 |
+
| 0.4079 | 113 | 0.2558 |
|
649 |
+
| 0.4116 | 114 | 0.2565 |
|
650 |
+
| 0.4152 | 115 | 0.2747 |
|
651 |
+
| 0.4188 | 116 | 0.2684 |
|
652 |
+
| 0.4224 | 117 | 0.2643 |
|
653 |
+
| 0.4260 | 118 | 0.241 |
|
654 |
+
| 0.4296 | 119 | 0.2563 |
|
655 |
+
| 0.4332 | 120 | 0.2754 |
|
656 |
+
| 0.4368 | 121 | 0.2503 |
|
657 |
+
| 0.4404 | 122 | 0.2544 |
|
658 |
+
| 0.4440 | 123 | 0.2729 |
|
659 |
+
| 0.4477 | 124 | 0.2589 |
|
660 |
+
| 0.4513 | 125 | 0.2626 |
|
661 |
+
| 0.4549 | 126 | 0.2693 |
|
662 |
+
| 0.4585 | 127 | 0.2687 |
|
663 |
+
| 0.4621 | 128 | 0.2903 |
|
664 |
+
| 0.4657 | 129 | 0.2663 |
|
665 |
+
| 0.4693 | 130 | 0.2604 |
|
666 |
+
| 0.4729 | 131 | 0.2601 |
|
667 |
+
| 0.4765 | 132 | 0.2649 |
|
668 |
+
| 0.4801 | 133 | 0.2597 |
|
669 |
+
| 0.4838 | 134 | 0.2608 |
|
670 |
+
| 0.4874 | 135 | 0.245 |
|
671 |
+
| 0.4910 | 136 | 0.2587 |
|
672 |
+
| 0.4946 | 137 | 0.2618 |
|
673 |
+
| 0.4982 | 138 | 0.2599 |
|
674 |
+
| 0.5018 | 139 | 0.265 |
|
675 |
+
| 0.5054 | 140 | 0.2427 |
|
676 |
+
| 0.5090 | 141 | 0.2448 |
|
677 |
+
| 0.5126 | 142 | 0.2608 |
|
678 |
+
| 0.5162 | 143 | 0.2188 |
|
679 |
+
| 0.5199 | 144 | 0.2471 |
|
680 |
+
| 0.5235 | 145 | 0.2604 |
|
681 |
+
| 0.5271 | 146 | 0.2571 |
|
682 |
+
| 0.5307 | 147 | 0.2684 |
|
683 |
+
| 0.5343 | 148 | 0.2319 |
|
684 |
+
| 0.5379 | 149 | 0.2572 |
|
685 |
+
| 0.5415 | 150 | 0.2243 |
|
686 |
+
| 0.5451 | 151 | 0.2562 |
|
687 |
+
| 0.5487 | 152 | 0.2457 |
|
688 |
+
| 0.5523 | 153 | 0.255 |
|
689 |
+
| 0.5560 | 154 | 0.2664 |
|
690 |
+
| 0.5596 | 155 | 0.24 |
|
691 |
+
| 0.5632 | 156 | 0.2612 |
|
692 |
+
| 0.5668 | 157 | 0.243 |
|
693 |
+
| 0.5704 | 158 | 0.2345 |
|
694 |
+
| 0.5740 | 159 | 0.2359 |
|
695 |
+
| 0.5776 | 160 | 0.2384 |
|
696 |
+
| 0.5812 | 161 | 0.2541 |
|
697 |
+
| 0.5848 | 162 | 0.2496 |
|
698 |
+
| 0.5884 | 163 | 0.2429 |
|
699 |
+
| 0.5921 | 164 | 0.2411 |
|
700 |
+
| 0.5957 | 165 | 0.2261 |
|
701 |
+
| 0.5993 | 166 | 0.2164 |
|
702 |
+
| 0.6029 | 167 | 0.2251 |
|
703 |
+
| 0.6065 | 168 | 0.2417 |
|
704 |
+
| 0.6101 | 169 | 0.2494 |
|
705 |
+
| 0.6137 | 170 | 0.2359 |
|
706 |
+
| 0.6173 | 171 | 0.2489 |
|
707 |
+
| 0.6209 | 172 | 0.2261 |
|
708 |
+
| 0.6245 | 173 | 0.2367 |
|
709 |
+
| 0.6282 | 174 | 0.2355 |
|
710 |
+
| 0.6318 | 175 | 0.2423 |
|
711 |
+
| 0.6354 | 176 | 0.2454 |
|
712 |
+
| 0.6390 | 177 | 0.2438 |
|
713 |
+
| 0.6426 | 178 | 0.2415 |
|
714 |
+
| 0.6462 | 179 | 0.2237 |
|
715 |
+
| 0.6498 | 180 | 0.2419 |
|
716 |
+
| 0.6534 | 181 | 0.2373 |
|
717 |
+
| 0.6570 | 182 | 0.2659 |
|
718 |
+
| 0.6606 | 183 | 0.2201 |
|
719 |
+
| 0.6643 | 184 | 0.2342 |
|
720 |
+
| 0.6679 | 185 | 0.2149 |
|
721 |
+
| 0.6715 | 186 | 0.2241 |
|
722 |
+
| 0.6751 | 187 | 0.2443 |
|
723 |
+
| 0.6787 | 188 | 0.2489 |
|
724 |
+
| 0.6823 | 189 | 0.2354 |
|
725 |
+
| 0.6859 | 190 | 0.2483 |
|
726 |
+
| 0.6895 | 191 | 0.2193 |
|
727 |
+
| 0.6931 | 192 | 0.229 |
|
728 |
+
| 0.6968 | 193 | 0.2335 |
|
729 |
+
| 0.7004 | 194 | 0.2484 |
|
730 |
+
| 0.7040 | 195 | 0.2317 |
|
731 |
+
| 0.7076 | 196 | 0.2203 |
|
732 |
+
| 0.7112 | 197 | 0.2329 |
|
733 |
+
| 0.7148 | 198 | 0.2084 |
|
734 |
+
| 0.7184 | 199 | 0.2341 |
|
735 |
+
| 0.7220 | 200 | 0.2369 |
|
736 |
+
| 0.7256 | 201 | 0.2364 |
|
737 |
+
| 0.7292 | 202 | 0.2276 |
|
738 |
+
| 0.7329 | 203 | 0.215 |
|
739 |
+
| 0.7365 | 204 | 0.2486 |
|
740 |
+
| 0.7401 | 205 | 0.2237 |
|
741 |
+
| 0.7437 | 206 | 0.218 |
|
742 |
+
| 0.7473 | 207 | 0.2444 |
|
743 |
+
| 0.7509 | 208 | 0.2276 |
|
744 |
+
| 0.7545 | 209 | 0.2127 |
|
745 |
+
| 0.7581 | 210 | 0.2283 |
|
746 |
+
| 0.7617 | 211 | 0.2234 |
|
747 |
+
| 0.7653 | 212 | 0.207 |
|
748 |
+
| 0.7690 | 213 | 0.24 |
|
749 |
+
| 0.7726 | 214 | 0.2317 |
|
750 |
+
| 0.7762 | 215 | 0.2056 |
|
751 |
+
| 0.7798 | 216 | 0.2149 |
|
752 |
+
| 0.7834 | 217 | 0.2211 |
|
753 |
+
| 0.7870 | 218 | 0.2232 |
|
754 |
+
| 0.7906 | 219 | 0.2222 |
|
755 |
+
| 0.7942 | 220 | 0.2481 |
|
756 |
+
| 0.7978 | 221 | 0.227 |
|
757 |
+
| 0.8014 | 222 | 0.2305 |
|
758 |
+
| 0.8051 | 223 | 0.2091 |
|
759 |
+
| 0.8087 | 224 | 0.2278 |
|
760 |
+
| 0.8123 | 225 | 0.2123 |
|
761 |
+
| 0.8159 | 226 | 0.2233 |
|
762 |
+
| 0.8195 | 227 | 0.2365 |
|
763 |
+
| 0.8231 | 228 | 0.2165 |
|
764 |
+
| 0.8267 | 229 | 0.2192 |
|
765 |
+
| 0.8303 | 230 | 0.2145 |
|
766 |
+
| 0.8339 | 231 | 0.2382 |
|
767 |
+
| 0.8375 | 232 | 0.2232 |
|
768 |
+
| 0.8412 | 233 | 0.2273 |
|
769 |
+
| 0.8448 | 234 | 0.2296 |
|
770 |
+
| 0.8484 | 235 | 0.2229 |
|
771 |
+
| 0.8520 | 236 | 0.2213 |
|
772 |
+
| 0.8556 | 237 | 0.2343 |
|
773 |
+
| 0.8592 | 238 | 0.2208 |
|
774 |
+
| 0.8628 | 239 | 0.2315 |
|
775 |
+
| 0.8664 | 240 | 0.2137 |
|
776 |
+
| 0.8700 | 241 | 0.2201 |
|
777 |
+
| 0.8736 | 242 | 0.2185 |
|
778 |
+
| 0.8773 | 243 | 0.2337 |
|
779 |
+
| 0.8809 | 244 | 0.2153 |
|
780 |
+
| 0.8845 | 245 | 0.2369 |
|
781 |
+
| 0.8881 | 246 | 0.2216 |
|
782 |
+
| 0.8917 | 247 | 0.2338 |
|
783 |
+
| 0.8953 | 248 | 0.2241 |
|
784 |
+
| 0.8989 | 249 | 0.213 |
|
785 |
+
| 0.9025 | 250 | 0.2245 |
|
786 |
+
| 0.9061 | 251 | 0.2074 |
|
787 |
+
| 0.9097 | 252 | 0.2283 |
|
788 |
+
| 0.9134 | 253 | 0.2003 |
|
789 |
+
| 0.9170 | 254 | 0.2099 |
|
790 |
+
| 0.9206 | 255 | 0.2288 |
|
791 |
+
| 0.9242 | 256 | 0.2168 |
|
792 |
+
| 0.9278 | 257 | 0.215 |
|
793 |
+
| 0.9314 | 258 | 0.2146 |
|
794 |
+
| 0.9350 | 259 | 0.2126 |
|
795 |
+
| 0.9386 | 260 | 0.2178 |
|
796 |
+
| 0.9422 | 261 | 0.2065 |
|
797 |
+
| 0.9458 | 262 | 0.2327 |
|
798 |
+
| 0.9495 | 263 | 0.2116 |
|
799 |
+
| 0.9531 | 264 | 0.2324 |
|
800 |
+
| 0.9567 | 265 | 0.2235 |
|
801 |
+
| 0.9603 | 266 | 0.2189 |
|
802 |
+
| 0.9639 | 267 | 0.2175 |
|
803 |
+
| 0.9675 | 268 | 0.2171 |
|
804 |
+
| 0.9711 | 269 | 0.1925 |
|
805 |
+
| 0.9747 | 270 | 0.225 |
|
806 |
+
| 0.9783 | 271 | 0.2149 |
|
807 |
+
| 0.9819 | 272 | 0.204 |
|
808 |
+
| 0.9856 | 273 | 0.2004 |
|
809 |
+
| 0.9892 | 274 | 0.2055 |
|
810 |
+
| 0.9928 | 275 | 0.2045 |
|
811 |
+
| 0.9964 | 276 | 0.2186 |
|
812 |
+
| 1.0 | 277 | 0.2215 |
|
813 |
+
| 1.0036 | 278 | 0.1545 |
|
814 |
+
| 1.0072 | 279 | 0.169 |
|
815 |
+
| 1.0108 | 280 | 0.152 |
|
816 |
+
| 1.0144 | 281 | 0.1597 |
|
817 |
+
| 1.0181 | 282 | 0.1626 |
|
818 |
+
| 1.0217 | 283 | 0.1692 |
|
819 |
+
| 1.0253 | 284 | 0.1639 |
|
820 |
+
| 1.0289 | 285 | 0.1638 |
|
821 |
+
| 1.0325 | 286 | 0.1507 |
|
822 |
+
| 1.0361 | 287 | 0.1594 |
|
823 |
+
| 1.0397 | 288 | 0.1621 |
|
824 |
+
| 1.0433 | 289 | 0.1565 |
|
825 |
+
| 1.0469 | 290 | 0.1549 |
|
826 |
+
| 1.0505 | 291 | 0.1731 |
|
827 |
+
| 1.0542 | 292 | 0.152 |
|
828 |
+
| 1.0578 | 293 | 0.1586 |
|
829 |
+
| 1.0614 | 294 | 0.1593 |
|
830 |
+
| 1.0650 | 295 | 0.1406 |
|
831 |
+
| 1.0686 | 296 | 0.1524 |
|
832 |
+
| 1.0722 | 297 | 0.1474 |
|
833 |
+
| 1.0758 | 298 | 0.158 |
|
834 |
+
| 1.0794 | 299 | 0.1743 |
|
835 |
+
| 1.0830 | 300 | 0.1485 |
|
836 |
+
| 1.0866 | 301 | 0.1648 |
|
837 |
+
| 1.0903 | 302 | 0.1337 |
|
838 |
+
| 1.0939 | 303 | 0.1554 |
|
839 |
+
| 1.0975 | 304 | 0.1434 |
|
840 |
+
| 1.1011 | 305 | 0.1642 |
|
841 |
+
| 1.1047 | 306 | 0.159 |
|
842 |
+
| 1.1083 | 307 | 0.1658 |
|
843 |
+
| 1.1119 | 308 | 0.1554 |
|
844 |
+
| 1.1155 | 309 | 0.1425 |
|
845 |
+
| 1.1191 | 310 | 0.1432 |
|
846 |
+
| 1.1227 | 311 | 0.1517 |
|
847 |
+
| 1.1264 | 312 | 0.148 |
|
848 |
+
| 1.1300 | 313 | 0.1636 |
|
849 |
+
| 1.1336 | 314 | 0.1735 |
|
850 |
+
| 1.1372 | 315 | 0.151 |
|
851 |
+
| 1.1408 | 316 | 0.1423 |
|
852 |
+
| 1.1444 | 317 | 0.1501 |
|
853 |
+
| 1.1480 | 318 | 0.1537 |
|
854 |
+
| 1.1516 | 319 | 0.1554 |
|
855 |
+
| 1.1552 | 320 | 0.1553 |
|
856 |
+
| 1.1588 | 321 | 0.149 |
|
857 |
+
| 1.1625 | 322 | 0.1605 |
|
858 |
+
| 1.1661 | 323 | 0.1551 |
|
859 |
+
| 1.1697 | 324 | 0.1555 |
|
860 |
+
| 1.1733 | 325 | 0.1443 |
|
861 |
+
| 1.1769 | 326 | 0.1533 |
|
862 |
+
| 1.1805 | 327 | 0.1658 |
|
863 |
+
| 1.1841 | 328 | 0.15 |
|
864 |
+
| 1.1877 | 329 | 0.1626 |
|
865 |
+
| 1.1913 | 330 | 0.172 |
|
866 |
+
| 1.1949 | 331 | 0.1542 |
|
867 |
+
| 1.1986 | 332 | 0.166 |
|
868 |
+
| 1.2022 | 333 | 0.1513 |
|
869 |
+
| 1.2058 | 334 | 0.1612 |
|
870 |
+
| 1.2094 | 335 | 0.1521 |
|
871 |
+
| 1.2130 | 336 | 0.1552 |
|
872 |
+
| 1.2166 | 337 | 0.1503 |
|
873 |
+
| 1.2202 | 338 | 0.1613 |
|
874 |
+
| 1.2238 | 339 | 0.1563 |
|
875 |
+
| 1.2274 | 340 | 0.1429 |
|
876 |
+
| 1.2310 | 341 | 0.1587 |
|
877 |
+
| 1.2347 | 342 | 0.1477 |
|
878 |
+
| 1.2383 | 343 | 0.1561 |
|
879 |
+
| 1.2419 | 344 | 0.1418 |
|
880 |
+
| 1.2455 | 345 | 0.1495 |
|
881 |
+
| 1.2491 | 346 | 0.1533 |
|
882 |
+
| 1.2527 | 347 | 0.1521 |
|
883 |
+
| 1.2563 | 348 | 0.1422 |
|
884 |
+
| 1.2599 | 349 | 0.1446 |
|
885 |
+
| 1.2635 | 350 | 0.146 |
|
886 |
+
| 1.2671 | 351 | 0.1473 |
|
887 |
+
| 1.2708 | 352 | 0.1566 |
|
888 |
+
| 1.2744 | 353 | 0.1411 |
|
889 |
+
| 1.2780 | 354 | 0.1502 |
|
890 |
+
| 1.2816 | 355 | 0.1383 |
|
891 |
+
| 1.2852 | 356 | 0.1622 |
|
892 |
+
| 1.2888 | 357 | 0.1391 |
|
893 |
+
| 1.2924 | 358 | 0.1455 |
|
894 |
+
| 1.2960 | 359 | 0.1541 |
|
895 |
+
| 1.2996 | 360 | 0.1476 |
|
896 |
+
| 1.3032 | 361 | 0.1662 |
|
897 |
+
| 1.3069 | 362 | 0.1476 |
|
898 |
+
| 1.3105 | 363 | 0.1452 |
|
899 |
+
| 1.3141 | 364 | 0.1372 |
|
900 |
+
| 1.3177 | 365 | 0.1542 |
|
901 |
+
| 1.3213 | 366 | 0.1531 |
|
902 |
+
| 1.3249 | 367 | 0.1623 |
|
903 |
+
| 1.3285 | 368 | 0.1544 |
|
904 |
+
| 1.3321 | 369 | 0.1625 |
|
905 |
+
| 1.3357 | 370 | 0.1459 |
|
906 |
+
| 1.3394 | 371 | 0.1474 |
|
907 |
+
| 1.3430 | 372 | 0.1499 |
|
908 |
+
| 1.3466 | 373 | 0.1495 |
|
909 |
+
| 1.3502 | 374 | 0.1361 |
|
910 |
+
| 1.3538 | 375 | 0.1444 |
|
911 |
+
| 1.3574 | 376 | 0.1495 |
|
912 |
+
| 1.3610 | 377 | 0.1583 |
|
913 |
+
| 1.3646 | 378 | 0.1642 |
|
914 |
+
| 1.3682 | 379 | 0.1646 |
|
915 |
+
| 1.3718 | 380 | 0.1595 |
|
916 |
+
| 1.3755 | 381 | 0.149 |
|
917 |
+
| 1.3791 | 382 | 0.1448 |
|
918 |
+
| 1.3827 | 383 | 0.1603 |
|
919 |
+
| 1.3863 | 384 | 0.1269 |
|
920 |
+
| 1.3899 | 385 | 0.1491 |
|
921 |
+
| 1.3935 | 386 | 0.1367 |
|
922 |
+
| 1.3971 | 387 | 0.1501 |
|
923 |
+
| 1.4007 | 388 | 0.1414 |
|
924 |
+
| 1.4043 | 389 | 0.156 |
|
925 |
+
| 1.4079 | 390 | 0.1428 |
|
926 |
+
| 1.4116 | 391 | 0.1559 |
|
927 |
+
| 1.4152 | 392 | 0.1452 |
|
928 |
+
| 1.4188 | 393 | 0.1547 |
|
929 |
+
| 1.4224 | 394 | 0.1432 |
|
930 |
+
| 1.4260 | 395 | 0.1648 |
|
931 |
+
| 1.4296 | 396 | 0.166 |
|
932 |
+
| 1.4332 | 397 | 0.1485 |
|
933 |
+
| 1.4368 | 398 | 0.1494 |
|
934 |
+
| 1.4404 | 399 | 0.1635 |
|
935 |
+
| 1.4440 | 400 | 0.1498 |
|
936 |
+
| 1.4477 | 401 | 0.1509 |
|
937 |
+
| 1.4513 | 402 | 0.1431 |
|
938 |
+
| 1.4549 | 403 | 0.1547 |
|
939 |
+
| 1.4585 | 404 | 0.1576 |
|
940 |
+
| 1.4621 | 405 | 0.1426 |
|
941 |
+
| 1.4657 | 406 | 0.132 |
|
942 |
+
| 1.4693 | 407 | 0.1511 |
|
943 |
+
| 1.4729 | 408 | 0.1551 |
|
944 |
+
| 1.4765 | 409 | 0.16 |
|
945 |
+
| 1.4801 | 410 | 0.1507 |
|
946 |
+
| 1.4838 | 411 | 0.1591 |
|
947 |
+
| 1.4874 | 412 | 0.1536 |
|
948 |
+
| 1.4910 | 413 | 0.1507 |
|
949 |
+
| 1.4946 | 414 | 0.1564 |
|
950 |
+
| 1.4982 | 415 | 0.153 |
|
951 |
+
| 1.5018 | 416 | 0.1404 |
|
952 |
+
| 1.5054 | 417 | 0.1627 |
|
953 |
+
| 1.5090 | 418 | 0.1432 |
|
954 |
+
| 1.5126 | 419 | 0.1456 |
|
955 |
+
| 1.5162 | 420 | 0.1369 |
|
956 |
+
| 1.5199 | 421 | 0.1554 |
|
957 |
+
| 1.5235 | 422 | 0.1412 |
|
958 |
+
| 1.5271 | 423 | 0.1547 |
|
959 |
+
| 1.5307 | 424 | 0.1555 |
|
960 |
+
| 1.5343 | 425 | 0.1575 |
|
961 |
+
| 1.5379 | 426 | 0.1595 |
|
962 |
+
| 1.5415 | 427 | 0.1464 |
|
963 |
+
| 1.5451 | 428 | 0.1738 |
|
964 |
+
| 1.5487 | 429 | 0.1692 |
|
965 |
+
| 1.5523 | 430 | 0.1566 |
|
966 |
+
| 1.5560 | 431 | 0.1452 |
|
967 |
+
| 1.5596 | 432 | 0.1433 |
|
968 |
+
| 1.5632 | 433 | 0.1584 |
|
969 |
+
| 1.5668 | 434 | 0.1579 |
|
970 |
+
| 1.5704 | 435 | 0.157 |
|
971 |
+
| 1.5740 | 436 | 0.1533 |
|
972 |
+
| 1.5776 | 437 | 0.148 |
|
973 |
+
| 1.5812 | 438 | 0.1381 |
|
974 |
+
| 1.5848 | 439 | 0.1605 |
|
975 |
+
| 1.5884 | 440 | 0.163 |
|
976 |
+
| 1.5921 | 441 | 0.1492 |
|
977 |
+
| 1.5957 | 442 | 0.1601 |
|
978 |
+
| 1.5993 | 443 | 0.1456 |
|
979 |
+
| 1.6029 | 444 | 0.1439 |
|
980 |
+
| 1.6065 | 445 | 0.1553 |
|
981 |
+
| 1.6101 | 446 | 0.1371 |
|
982 |
+
| 1.6137 | 447 | 0.1382 |
|
983 |
+
| 1.6173 | 448 | 0.1458 |
|
984 |
+
| 1.6209 | 449 | 0.14 |
|
985 |
+
| 1.6245 | 450 | 0.1463 |
|
986 |
+
| 1.6282 | 451 | 0.1433 |
|
987 |
+
| 1.6318 | 452 | 0.1472 |
|
988 |
+
| 1.6354 | 453 | 0.1481 |
|
989 |
+
| 1.6390 | 454 | 0.1408 |
|
990 |
+
| 1.6426 | 455 | 0.1525 |
|
991 |
+
| 1.6462 | 456 | 0.1223 |
|
992 |
+
| 1.6498 | 457 | 0.1452 |
|
993 |
+
| 1.6534 | 458 | 0.159 |
|
994 |
+
| 1.6570 | 459 | 0.1389 |
|
995 |
+
| 1.6606 | 460 | 0.1479 |
|
996 |
+
| 1.6643 | 461 | 0.1451 |
|
997 |
+
| 1.6679 | 462 | 0.1651 |
|
998 |
+
| 1.6715 | 463 | 0.1336 |
|
999 |
+
| 1.6751 | 464 | 0.1496 |
|
1000 |
+
| 1.6787 | 465 | 0.1384 |
|
1001 |
+
| 1.6823 | 466 | 0.143 |
|
1002 |
+
| 1.6859 | 467 | 0.1423 |
|
1003 |
+
| 1.6895 | 468 | 0.1403 |
|
1004 |
+
| 1.6931 | 469 | 0.1577 |
|
1005 |
+
| 1.6968 | 470 | 0.1511 |
|
1006 |
+
| 1.7004 | 471 | 0.1429 |
|
1007 |
+
| 1.7040 | 472 | 0.1445 |
|
1008 |
+
| 1.7076 | 473 | 0.1431 |
|
1009 |
+
| 1.7112 | 474 | 0.1326 |
|
1010 |
+
| 1.7148 | 475 | 0.1554 |
|
1011 |
+
| 1.7184 | 476 | 0.1406 |
|
1012 |
+
| 1.7220 | 477 | 0.1479 |
|
1013 |
+
| 1.7256 | 478 | 0.1521 |
|
1014 |
+
| 1.7292 | 479 | 0.1475 |
|
1015 |
+
| 1.7329 | 480 | 0.1584 |
|
1016 |
+
| 1.7365 | 481 | 0.1393 |
|
1017 |
+
| 1.7401 | 482 | 0.1291 |
|
1018 |
+
| 1.7437 | 483 | 0.1373 |
|
1019 |
+
| 1.7473 | 484 | 0.1555 |
|
1020 |
+
| 1.7509 | 485 | 0.1473 |
|
1021 |
+
| 1.7545 | 486 | 0.1654 |
|
1022 |
+
| 1.7581 | 487 | 0.1568 |
|
1023 |
+
| 1.7617 | 488 | 0.1557 |
|
1024 |
+
| 1.7653 | 489 | 0.1531 |
|
1025 |
+
| 1.7690 | 490 | 0.1385 |
|
1026 |
+
| 1.7726 | 491 | 0.1381 |
|
1027 |
+
| 1.7762 | 492 | 0.1375 |
|
1028 |
+
| 1.7798 | 493 | 0.1472 |
|
1029 |
+
| 1.7834 | 494 | 0.1581 |
|
1030 |
+
| 1.7870 | 495 | 0.1448 |
|
1031 |
+
| 1.7906 | 496 | 0.1443 |
|
1032 |
+
| 1.7942 | 497 | 0.1422 |
|
1033 |
+
| 1.7978 | 498 | 0.1295 |
|
1034 |
+
| 1.8014 | 499 | 0.1463 |
|
1035 |
+
| 1.8051 | 500 | 0.1346 |
|
1036 |
+
| 1.8087 | 501 | 0.1387 |
|
1037 |
+
| 1.8123 | 502 | 0.1463 |
|
1038 |
+
| 1.8159 | 503 | 0.1439 |
|
1039 |
+
| 1.8195 | 504 | 0.1404 |
|
1040 |
+
| 1.8231 | 505 | 0.1433 |
|
1041 |
+
| 1.8267 | 506 | 0.136 |
|
1042 |
+
| 1.8303 | 507 | 0.14 |
|
1043 |
+
| 1.8339 | 508 | 0.1355 |
|
1044 |
+
| 1.8375 | 509 | 0.1446 |
|
1045 |
+
| 1.8412 | 510 | 0.1564 |
|
1046 |
+
| 1.8448 | 511 | 0.1413 |
|
1047 |
+
| 1.8484 | 512 | 0.1451 |
|
1048 |
+
| 1.8520 | 513 | 0.1453 |
|
1049 |
+
| 1.8556 | 514 | 0.1484 |
|
1050 |
+
| 1.8592 | 515 | 0.1403 |
|
1051 |
+
| 1.8628 | 516 | 0.1568 |
|
1052 |
+
| 1.8664 | 517 | 0.1566 |
|
1053 |
+
| 1.8700 | 518 | 0.1318 |
|
1054 |
+
| 1.8736 | 519 | 0.1483 |
|
1055 |
+
| 1.8773 | 520 | 0.1339 |
|
1056 |
+
| 1.8809 | 521 | 0.1423 |
|
1057 |
+
| 1.8845 | 522 | 0.1349 |
|
1058 |
+
| 1.8881 | 523 | 0.1302 |
|
1059 |
+
| 1.8917 | 524 | 0.1341 |
|
1060 |
+
| 1.8953 | 525 | 0.1456 |
|
1061 |
+
| 1.8989 | 526 | 0.1334 |
|
1062 |
+
| 1.9025 | 527 | 0.1382 |
|
1063 |
+
| 1.9061 | 528 | 0.1462 |
|
1064 |
+
| 1.9097 | 529 | 0.1315 |
|
1065 |
+
| 1.9134 | 530 | 0.1606 |
|
1066 |
+
| 1.9170 | 531 | 0.1308 |
|
1067 |
+
| 1.9206 | 532 | 0.1319 |
|
1068 |
+
| 1.9242 | 533 | 0.1407 |
|
1069 |
+
| 1.9278 | 534 | 0.1385 |
|
1070 |
+
| 1.9314 | 535 | 0.1471 |
|
1071 |
+
| 1.9350 | 536 | 0.1621 |
|
1072 |
+
| 1.9386 | 537 | 0.1436 |
|
1073 |
+
| 1.9422 | 538 | 0.151 |
|
1074 |
+
| 1.9458 | 539 | 0.1423 |
|
1075 |
+
| 1.9495 | 540 | 0.1411 |
|
1076 |
+
| 1.9531 | 541 | 0.1535 |
|
1077 |
+
| 1.9567 | 542 | 0.143 |
|
1078 |
+
| 1.9603 | 543 | 0.149 |
|
1079 |
+
| 1.9639 | 544 | 0.1384 |
|
1080 |
+
| 1.9675 | 545 | 0.1479 |
|
1081 |
+
| 1.9711 | 546 | 0.1452 |
|
1082 |
+
| 1.9747 | 547 | 0.1372 |
|
1083 |
+
| 1.9783 | 548 | 0.1418 |
|
1084 |
+
| 1.9819 | 549 | 0.1443 |
|
1085 |
+
| 1.9856 | 550 | 0.1344 |
|
1086 |
+
| 1.9892 | 551 | 0.1278 |
|
1087 |
+
| 1.9928 | 552 | 0.1447 |
|
1088 |
+
| 1.9964 | 553 | 0.1366 |
|
1089 |
+
| 2.0 | 554 | 0.141 |
|
1090 |
+
| 2.0036 | 555 | 0.1161 |
|
1091 |
+
| 2.0072 | 556 | 0.1099 |
|
1092 |
+
| 2.0108 | 557 | 0.126 |
|
1093 |
+
| 2.0144 | 558 | 0.1163 |
|
1094 |
+
| 2.0181 | 559 | 0.1234 |
|
1095 |
+
| 2.0217 | 560 | 0.1171 |
|
1096 |
+
| 2.0253 | 561 | 0.1073 |
|
1097 |
+
| 2.0289 | 562 | 0.1126 |
|
1098 |
+
| 2.0325 | 563 | 0.1175 |
|
1099 |
+
| 2.0361 | 564 | 0.1086 |
|
1100 |
+
| 2.0397 | 565 | 0.1038 |
|
1101 |
+
| 2.0433 | 566 | 0.1121 |
|
1102 |
+
| 2.0469 | 567 | 0.1154 |
|
1103 |
+
| 2.0505 | 568 | 0.0973 |
|
1104 |
+
| 2.0542 | 569 | 0.1208 |
|
1105 |
+
| 2.0578 | 570 | 0.1064 |
|
1106 |
+
| 2.0614 | 571 | 0.1159 |
|
1107 |
+
| 2.0650 | 572 | 0.1093 |
|
1108 |
+
| 2.0686 | 573 | 0.113 |
|
1109 |
+
| 2.0722 | 574 | 0.1033 |
|
1110 |
+
| 2.0758 | 575 | 0.1152 |
|
1111 |
+
| 2.0794 | 576 | 0.1029 |
|
1112 |
+
| 2.0830 | 577 | 0.1204 |
|
1113 |
+
| 2.0866 | 578 | 0.1079 |
|
1114 |
+
| 2.0903 | 579 | 0.1288 |
|
1115 |
+
| 2.0939 | 580 | 0.0998 |
|
1116 |
+
| 2.0975 | 581 | 0.1058 |
|
1117 |
+
| 2.1011 | 582 | 0.1235 |
|
1118 |
+
| 2.1047 | 583 | 0.1059 |
|
1119 |
+
| 2.1083 | 584 | 0.0998 |
|
1120 |
+
| 2.1119 | 585 | 0.1142 |
|
1121 |
+
| 2.1155 | 586 | 0.1082 |
|
1122 |
+
| 2.1191 | 587 | 0.0973 |
|
1123 |
+
| 2.1227 | 588 | 0.1017 |
|
1124 |
+
| 2.1264 | 589 | 0.1045 |
|
1125 |
+
| 2.1300 | 590 | 0.123 |
|
1126 |
+
| 2.1336 | 591 | 0.1065 |
|
1127 |
+
| 2.1372 | 592 | 0.1135 |
|
1128 |
+
| 2.1408 | 593 | 0.1027 |
|
1129 |
+
| 2.1444 | 594 | 0.1166 |
|
1130 |
+
| 2.1480 | 595 | 0.1082 |
|
1131 |
+
| 2.1516 | 596 | 0.1113 |
|
1132 |
+
| 2.1552 | 597 | 0.1108 |
|
1133 |
+
| 2.1588 | 598 | 0.114 |
|
1134 |
+
| 2.1625 | 599 | 0.1064 |
|
1135 |
+
| 2.1661 | 600 | 0.0955 |
|
1136 |
+
| 2.1697 | 601 | 0.113 |
|
1137 |
+
| 2.1733 | 602 | 0.1136 |
|
1138 |
+
| 2.1769 | 603 | 0.1125 |
|
1139 |
+
| 2.1805 | 604 | 0.1146 |
|
1140 |
+
| 2.1841 | 605 | 0.1054 |
|
1141 |
+
| 2.1877 | 606 | 0.1144 |
|
1142 |
+
| 2.1913 | 607 | 0.1038 |
|
1143 |
+
| 2.1949 | 608 | 0.1113 |
|
1144 |
+
| 2.1986 | 609 | 0.1187 |
|
1145 |
+
| 2.2022 | 610 | 0.1166 |
|
1146 |
+
| 2.2058 | 611 | 0.1035 |
|
1147 |
+
| 2.2094 | 612 | 0.1054 |
|
1148 |
+
| 2.2130 | 613 | 0.118 |
|
1149 |
+
| 2.2166 | 614 | 0.125 |
|
1150 |
+
| 2.2202 | 615 | 0.1142 |
|
1151 |
+
| 2.2238 | 616 | 0.1119 |
|
1152 |
+
| 2.2274 | 617 | 0.1173 |
|
1153 |
+
| 2.2310 | 618 | 0.1024 |
|
1154 |
+
| 2.2347 | 619 | 0.105 |
|
1155 |
+
| 2.2383 | 620 | 0.1025 |
|
1156 |
+
| 2.2419 | 621 | 0.1022 |
|
1157 |
+
| 2.2455 | 622 | 0.0995 |
|
1158 |
+
| 2.2491 | 623 | 0.1022 |
|
1159 |
+
| 2.2527 | 624 | 0.1198 |
|
1160 |
+
| 2.2563 | 625 | 0.0995 |
|
1161 |
+
| 2.2599 | 626 | 0.1162 |
|
1162 |
+
| 2.2635 | 627 | 0.1172 |
|
1163 |
+
| 2.2671 | 628 | 0.1037 |
|
1164 |
+
| 2.2708 | 629 | 0.1093 |
|
1165 |
+
| 2.2744 | 630 | 0.1018 |
|
1166 |
+
| 2.2780 | 631 | 0.1168 |
|
1167 |
+
| 2.2816 | 632 | 0.1015 |
|
1168 |
+
| 2.2852 | 633 | 0.101 |
|
1169 |
+
| 2.2888 | 634 | 0.1064 |
|
1170 |
+
| 2.2924 | 635 | 0.1185 |
|
1171 |
+
| 2.2960 | 636 | 0.1055 |
|
1172 |
+
| 2.2996 | 637 | 0.1142 |
|
1173 |
+
| 2.3032 | 638 | 0.0966 |
|
1174 |
+
| 2.3069 | 639 | 0.1039 |
|
1175 |
+
| 2.3105 | 640 | 0.1139 |
|
1176 |
+
| 2.3141 | 641 | 0.1181 |
|
1177 |
+
| 2.3177 | 642 | 0.1168 |
|
1178 |
+
| 2.3213 | 643 | 0.1201 |
|
1179 |
+
| 2.3249 | 644 | 0.0984 |
|
1180 |
+
| 2.3285 | 645 | 0.1068 |
|
1181 |
+
| 2.3321 | 646 | 0.1007 |
|
1182 |
+
| 2.3357 | 647 | 0.1179 |
|
1183 |
+
| 2.3394 | 648 | 0.1043 |
|
1184 |
+
| 2.3430 | 649 | 0.1213 |
|
1185 |
+
| 2.3466 | 650 | 0.1027 |
|
1186 |
+
| 2.3502 | 651 | 0.1119 |
|
1187 |
+
| 2.3538 | 652 | 0.1077 |
|
1188 |
+
| 2.3574 | 653 | 0.1061 |
|
1189 |
+
| 2.3610 | 654 | 0.1054 |
|
1190 |
+
| 2.3646 | 655 | 0.1135 |
|
1191 |
+
| 2.3682 | 656 | 0.1136 |
|
1192 |
+
| 2.3718 | 657 | 0.1062 |
|
1193 |
+
| 2.3755 | 658 | 0.1105 |
|
1194 |
+
| 2.3791 | 659 | 0.1157 |
|
1195 |
+
| 2.3827 | 660 | 0.1036 |
|
1196 |
+
| 2.3863 | 661 | 0.1098 |
|
1197 |
+
| 2.3899 | 662 | 0.1195 |
|
1198 |
+
| 2.3935 | 663 | 0.1151 |
|
1199 |
+
| 2.3971 | 664 | 0.1116 |
|
1200 |
+
| 2.4007 | 665 | 0.1086 |
|
1201 |
+
| 2.4043 | 666 | 0.1151 |
|
1202 |
+
| 2.4079 | 667 | 0.1156 |
|
1203 |
+
| 2.4116 | 668 | 0.116 |
|
1204 |
+
| 2.4152 | 669 | 0.1055 |
|
1205 |
+
| 2.4188 | 670 | 0.1051 |
|
1206 |
+
| 2.4224 | 671 | 0.0952 |
|
1207 |
+
| 2.4260 | 672 | 0.1012 |
|
1208 |
+
| 2.4296 | 673 | 0.1042 |
|
1209 |
+
| 2.4332 | 674 | 0.1069 |
|
1210 |
+
| 2.4368 | 675 | 0.1148 |
|
1211 |
+
| 2.4404 | 676 | 0.0981 |
|
1212 |
+
| 2.4440 | 677 | 0.1131 |
|
1213 |
+
| 2.4477 | 678 | 0.1026 |
|
1214 |
+
| 2.4513 | 679 | 0.1014 |
|
1215 |
+
| 2.4549 | 680 | 0.1071 |
|
1216 |
+
| 2.4585 | 681 | 0.1171 |
|
1217 |
+
| 2.4621 | 682 | 0.1009 |
|
1218 |
+
| 2.4657 | 683 | 0.1056 |
|
1219 |
+
| 2.4693 | 684 | 0.1107 |
|
1220 |
+
| 2.4729 | 685 | 0.1114 |
|
1221 |
+
| 2.4765 | 686 | 0.1118 |
|
1222 |
+
| 2.4801 | 687 | 0.1166 |
|
1223 |
+
| 2.4838 | 688 | 0.1023 |
|
1224 |
+
| 2.4874 | 689 | 0.1154 |
|
1225 |
+
| 2.4910 | 690 | 0.0968 |
|
1226 |
+
| 2.4946 | 691 | 0.1164 |
|
1227 |
+
| 2.4982 | 692 | 0.1221 |
|
1228 |
+
| 2.5018 | 693 | 0.1131 |
|
1229 |
+
| 2.5054 | 694 | 0.1039 |
|
1230 |
+
| 2.5090 | 695 | 0.1022 |
|
1231 |
+
| 2.5126 | 696 | 0.1052 |
|
1232 |
+
| 2.5162 | 697 | 0.1072 |
|
1233 |
+
| 2.5199 | 698 | 0.1062 |
|
1234 |
+
| 2.5235 | 699 | 0.1035 |
|
1235 |
+
| 2.5271 | 700 | 0.107 |
|
1236 |
+
| 2.5307 | 701 | 0.1152 |
|
1237 |
+
| 2.5343 | 702 | 0.0991 |
|
1238 |
+
| 2.5379 | 703 | 0.1139 |
|
1239 |
+
| 2.5415 | 704 | 0.1148 |
|
1240 |
+
| 2.5451 | 705 | 0.1099 |
|
1241 |
+
| 2.5487 | 706 | 0.1064 |
|
1242 |
+
| 2.5523 | 707 | 0.1069 |
|
1243 |
+
| 2.5560 | 708 | 0.1104 |
|
1244 |
+
| 2.5596 | 709 | 0.1157 |
|
1245 |
+
| 2.5632 | 710 | 0.1109 |
|
1246 |
+
| 2.5668 | 711 | 0.0991 |
|
1247 |
+
| 2.5704 | 712 | 0.105 |
|
1248 |
+
| 2.5740 | 713 | 0.1104 |
|
1249 |
+
| 2.5776 | 714 | 0.1134 |
|
1250 |
+
| 2.5812 | 715 | 0.1252 |
|
1251 |
+
| 2.5848 | 716 | 0.1205 |
|
1252 |
+
| 2.5884 | 717 | 0.112 |
|
1253 |
+
| 2.5921 | 718 | 0.1109 |
|
1254 |
+
| 2.5957 | 719 | 0.1151 |
|
1255 |
+
| 2.5993 | 720 | 0.097 |
|
1256 |
+
| 2.6029 | 721 | 0.1018 |
|
1257 |
+
| 2.6065 | 722 | 0.1205 |
|
1258 |
+
| 2.6101 | 723 | 0.107 |
|
1259 |
+
| 2.6137 | 724 | 0.102 |
|
1260 |
+
| 2.6173 | 725 | 0.1106 |
|
1261 |
+
| 2.6209 | 726 | 0.1068 |
|
1262 |
+
| 2.6245 | 727 | 0.1024 |
|
1263 |
+
| 2.6282 | 728 | 0.1153 |
|
1264 |
+
| 2.6318 | 729 | 0.0984 |
|
1265 |
+
| 2.6354 | 730 | 0.1019 |
|
1266 |
+
| 2.6390 | 731 | 0.1029 |
|
1267 |
+
| 2.6426 | 732 | 0.1147 |
|
1268 |
+
| 2.6462 | 733 | 0.1081 |
|
1269 |
+
| 2.6498 | 734 | 0.0996 |
|
1270 |
+
| 2.6534 | 735 | 0.1133 |
|
1271 |
+
| 2.6570 | 736 | 0.1102 |
|
1272 |
+
| 2.6606 | 737 | 0.1063 |
|
1273 |
+
| 2.6643 | 738 | 0.1119 |
|
1274 |
+
| 2.6679 | 739 | 0.1062 |
|
1275 |
+
| 2.6715 | 740 | 0.1021 |
|
1276 |
+
| 2.6751 | 741 | 0.1058 |
|
1277 |
+
| 2.6787 | 742 | 0.1026 |
|
1278 |
+
| 2.6823 | 743 | 0.1049 |
|
1279 |
+
| 2.6859 | 744 | 0.0894 |
|
1280 |
+
| 2.6895 | 745 | 0.1127 |
|
1281 |
+
| 2.6931 | 746 | 0.1107 |
|
1282 |
+
| 2.6968 | 747 | 0.1134 |
|
1283 |
+
| 2.7004 | 748 | 0.103 |
|
1284 |
+
| 2.7040 | 749 | 0.1081 |
|
1285 |
+
| 2.7076 | 750 | 0.1156 |
|
1286 |
+
| 2.7112 | 751 | 0.1092 |
|
1287 |
+
| 2.7148 | 752 | 0.1182 |
|
1288 |
+
| 2.7184 | 753 | 0.1092 |
|
1289 |
+
| 2.7220 | 754 | 0.1077 |
|
1290 |
+
| 2.7256 | 755 | 0.1165 |
|
1291 |
+
| 2.7292 | 756 | 0.1109 |
|
1292 |
+
| 2.7329 | 757 | 0.1061 |
|
1293 |
+
| 2.7365 | 758 | 0.1141 |
|
1294 |
+
| 2.7401 | 759 | 0.1073 |
|
1295 |
+
| 2.7437 | 760 | 0.1074 |
|
1296 |
+
| 2.7473 | 761 | 0.1042 |
|
1297 |
+
| 2.7509 | 762 | 0.1083 |
|
1298 |
+
| 2.7545 | 763 | 0.1011 |
|
1299 |
+
| 2.7581 | 764 | 0.1083 |
|
1300 |
+
| 2.7617 | 765 | 0.1078 |
|
1301 |
+
| 2.7653 | 766 | 0.1333 |
|
1302 |
+
| 2.7690 | 767 | 0.107 |
|
1303 |
+
| 2.7726 | 768 | 0.1114 |
|
1304 |
+
| 2.7762 | 769 | 0.1027 |
|
1305 |
+
| 2.7798 | 770 | 0.0976 |
|
1306 |
+
| 2.7834 | 771 | 0.1175 |
|
1307 |
+
| 2.7870 | 772 | 0.1099 |
|
1308 |
+
| 2.7906 | 773 | 0.102 |
|
1309 |
+
| 2.7942 | 774 | 0.1083 |
|
1310 |
+
| 2.7978 | 775 | 0.0999 |
|
1311 |
+
| 2.8014 | 776 | 0.1221 |
|
1312 |
+
| 2.8051 | 777 | 0.0996 |
|
1313 |
+
| 2.8087 | 778 | 0.0995 |
|
1314 |
+
| 2.8123 | 779 | 0.0974 |
|
1315 |
+
| 2.8159 | 780 | 0.1098 |
|
1316 |
+
| 2.8195 | 781 | 0.1012 |
|
1317 |
+
| 2.8231 | 782 | 0.1006 |
|
1318 |
+
| 2.8267 | 783 | 0.1047 |
|
1319 |
+
| 2.8303 | 784 | 0.1214 |
|
1320 |
+
| 2.8339 | 785 | 0.105 |
|
1321 |
+
| 2.8375 | 786 | 0.1152 |
|
1322 |
+
| 2.8412 | 787 | 0.0897 |
|
1323 |
+
| 2.8448 | 788 | 0.1189 |
|
1324 |
+
| 2.8484 | 789 | 0.1058 |
|
1325 |
+
| 2.8520 | 790 | 0.1051 |
|
1326 |
+
| 2.8556 | 791 | 0.1134 |
|
1327 |
+
| 2.8592 | 792 | 0.1016 |
|
1328 |
+
| 2.8628 | 793 | 0.1015 |
|
1329 |
+
| 2.8664 | 794 | 0.1119 |
|
1330 |
+
| 2.8700 | 795 | 0.1093 |
|
1331 |
+
| 2.8736 | 796 | 0.108 |
|
1332 |
+
| 2.8773 | 797 | 0.109 |
|
1333 |
+
| 2.8809 | 798 | 0.0922 |
|
1334 |
+
| 2.8845 | 799 | 0.1123 |
|
1335 |
+
| 2.8881 | 800 | 0.1094 |
|
1336 |
+
| 2.8917 | 801 | 0.1075 |
|
1337 |
+
| 2.8953 | 802 | 0.1138 |
|
1338 |
+
| 2.8989 | 803 | 0.1071 |
|
1339 |
+
| 2.9025 | 804 | 0.105 |
|
1340 |
+
| 2.9061 | 805 | 0.1153 |
|
1341 |
+
| 2.9097 | 806 | 0.1015 |
|
1342 |
+
| 2.9134 | 807 | 0.1199 |
|
1343 |
+
| 2.9170 | 808 | 0.1025 |
|
1344 |
+
| 2.9206 | 809 | 0.1082 |
|
1345 |
+
| 2.9242 | 810 | 0.1025 |
|
1346 |
+
| 2.9278 | 811 | 0.1073 |
|
1347 |
+
| 2.9314 | 812 | 0.1053 |
|
1348 |
+
| 2.9350 | 813 | 0.1088 |
|
1349 |
+
| 2.9386 | 814 | 0.1064 |
|
1350 |
+
| 2.9422 | 815 | 0.119 |
|
1351 |
+
| 2.9458 | 816 | 0.1099 |
|
1352 |
+
| 2.9495 | 817 | 0.1096 |
|
1353 |
+
| 2.9531 | 818 | 0.1101 |
|
1354 |
+
| 2.9567 | 819 | 0.086 |
|
1355 |
+
| 2.9603 | 820 | 0.1121 |
|
1356 |
+
| 2.9639 | 821 | 0.103 |
|
1357 |
+
| 2.9675 | 822 | 0.1108 |
|
1358 |
+
| 2.9711 | 823 | 0.1132 |
|
1359 |
+
| 2.9747 | 824 | 0.1034 |
|
1360 |
+
| 2.9783 | 825 | 0.0995 |
|
1361 |
+
| 2.9819 | 826 | 0.115 |
|
1362 |
+
| 2.9856 | 827 | 0.0966 |
|
1363 |
+
| 2.9892 | 828 | 0.1037 |
|
1364 |
+
| 2.9928 | 829 | 0.1124 |
|
1365 |
+
| 2.9964 | 830 | 0.1071 |
|
1366 |
+
| 3.0 | 831 | 0.1017 |
|
1367 |
+
|
1368 |
+
</details>
|
1369 |
+
|
1370 |
+
### Framework Versions
|
1371 |
+
- Python: 3.10.12
|
1372 |
+
- Sentence Transformers: 3.4.1
|
1373 |
+
- Transformers: 4.49.0
|
1374 |
+
- PyTorch: 2.5.1+cu124
|
1375 |
+
- Accelerate: 1.4.0
|
1376 |
+
- Datasets: 3.3.2
|
1377 |
+
- Tokenizers: 0.21.0
|
1378 |
+
|
1379 |
+
## Citation
|
1380 |
+
|
1381 |
+
### BibTeX
|
1382 |
+
|
1383 |
+
#### Sentence Transformers
|
1384 |
+
```bibtex
|
1385 |
+
@inproceedings{reimers-2019-sentence-bert,
|
1386 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
1387 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
1388 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
1389 |
+
month = "11",
|
1390 |
+
year = "2019",
|
1391 |
+
publisher = "Association for Computational Linguistics",
|
1392 |
+
url = "https://arxiv.org/abs/1908.10084",
|
1393 |
+
}
|
1394 |
+
```
|
1395 |
+
|
1396 |
+
<!--
|
1397 |
+
## Glossary
|
1398 |
+
|
1399 |
+
*Clearly define terms in order to be accessible across audiences.*
|
1400 |
+
-->
|
1401 |
+
|
1402 |
+
<!--
|
1403 |
+
## Model Card Authors
|
1404 |
+
|
1405 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
1406 |
+
-->
|
1407 |
+
|
1408 |
+
<!--
|
1409 |
+
## Model Card Contact
|
1410 |
+
|
1411 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
1412 |
+
-->
|
config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "output/checkpoint-831",
|
3 |
+
"architectures": [
|
4 |
+
"XLMRobertaModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "gelu",
|
11 |
+
"hidden_dropout_prob": 0.1,
|
12 |
+
"hidden_size": 1024,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 4096,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 8194,
|
17 |
+
"model_type": "xlm-roberta",
|
18 |
+
"num_attention_heads": 16,
|
19 |
+
"num_hidden_layers": 24,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 1,
|
22 |
+
"position_embedding_type": "absolute",
|
23 |
+
"torch_dtype": "float32",
|
24 |
+
"transformers_version": "4.49.0",
|
25 |
+
"type_vocab_size": 1,
|
26 |
+
"use_cache": true,
|
27 |
+
"vocab_size": 250002
|
28 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "3.4.1",
|
4 |
+
"transformers": "4.49.0",
|
5 |
+
"pytorch": "2.5.1+cu124"
|
6 |
+
},
|
7 |
+
"prompts": {},
|
8 |
+
"default_prompt_name": null,
|
9 |
+
"similarity_fn_name": "cosine"
|
10 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38aae0daf9ef0dc0d26588426536492b90667d00992c30f1c2761329f29d382c
|
3 |
+
size 2271064456
|
modules.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
},
|
14 |
+
{
|
15 |
+
"idx": 2,
|
16 |
+
"name": "2",
|
17 |
+
"path": "2_Normalize",
|
18 |
+
"type": "sentence_transformers.models.Normalize"
|
19 |
+
}
|
20 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 1024,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"cls_token": {
|
10 |
+
"content": "<s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"eos_token": {
|
17 |
+
"content": "</s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"mask_token": {
|
24 |
+
"content": "<mask>",
|
25 |
+
"lstrip": true,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"pad_token": {
|
31 |
+
"content": "<pad>",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
},
|
37 |
+
"sep_token": {
|
38 |
+
"content": "</s>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false
|
43 |
+
},
|
44 |
+
"unk_token": {
|
45 |
+
"content": "<unk>",
|
46 |
+
"lstrip": false,
|
47 |
+
"normalized": false,
|
48 |
+
"rstrip": false,
|
49 |
+
"single_word": false
|
50 |
+
}
|
51 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e3b8957de04e3a4ed42b1a11381556f9adad8d0d502b9dd071c75f626b28f40
|
3 |
+
size 17083053
|
tokenizer_config.json
ADDED
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<s>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<pad>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "<unk>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"250001": {
|
36 |
+
"content": "<mask>",
|
37 |
+
"lstrip": true,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"bos_token": "<s>",
|
45 |
+
"clean_up_tokenization_spaces": true,
|
46 |
+
"cls_token": "<s>",
|
47 |
+
"eos_token": "</s>",
|
48 |
+
"extra_special_tokens": {},
|
49 |
+
"mask_token": "<mask>",
|
50 |
+
"max_length": 1024,
|
51 |
+
"model_max_length": 1024,
|
52 |
+
"pad_to_multiple_of": null,
|
53 |
+
"pad_token": "<pad>",
|
54 |
+
"pad_token_type_id": 0,
|
55 |
+
"padding_side": "right",
|
56 |
+
"sep_token": "</s>",
|
57 |
+
"sp_model_kwargs": {},
|
58 |
+
"stride": 0,
|
59 |
+
"tokenizer_class": "XLMRobertaTokenizerFast",
|
60 |
+
"truncation_side": "right",
|
61 |
+
"truncation_strategy": "longest_first",
|
62 |
+
"unk_token": "<unk>"
|
63 |
+
}
|