seongil-dn commited on
Commit
a3e313f
·
verified ·
1 Parent(s): 6d849fd

Add new SentenceTransformer model

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 1024,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,1253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:1138596
8
+ - loss:CachedGISTEmbedLoss
9
+ base_model: seongil-dn/unsupervised_20m_3800
10
+ widget:
11
+ - source_sentence: How many people were reported to have died in the Great Fire of
12
+ London in 1666?
13
+ sentences:
14
+ - City of London 1666. Both of these fires were referred to as "the" Great Fire.
15
+ After the fire of 1666, a number of plans were drawn up to remodel the City and
16
+ its street pattern into a renaissance-style city with planned urban blocks, squares
17
+ and boulevards. These plans were almost entirely not taken up, and the medieval
18
+ street pattern re-emerged almost intact. By the late 16th century, London increasingly
19
+ became a major centre for banking, international trade and commerce. The Royal
20
+ Exchange was founded in 1565 by Sir Thomas Gresham as a centre of commerce for
21
+ London's merchants, and gained Royal patronage in
22
+ - Great Atlanta fire of 1917 Great Atlanta fire of 1917 The Great Atlanta Fire of
23
+ 1917 began just after noon on 21 May 1917 in the Old Fourth Ward of Atlanta, Georgia.
24
+ It is unclear just how the fire started, but it was fueled by hot temperatures
25
+ and strong winds which propelled the fire. The fire, which burned for nearly 10
26
+ hours, destroyed and 1,900 structures displacing over 10,000 people. Damages were
27
+ estimated at $5 million, ($ million when adjusted for inflation). It was a clear,
28
+ warm and sunny day with a brisk breeze from the south. This was not the only fire
29
+ of the
30
+ - Great Plague of London they had ever been seen ...". Plague cases continued to
31
+ occur sporadically at a modest rate until the summer of 1666. On the second and
32
+ third of September that year, the Great Fire of London destroyed much of the City
33
+ of London, and some people believed that the fire put an end to the epidemic.
34
+ However, it is now thought that the plague had largely subsided before the fire
35
+ took place. In fact, most of the later cases of plague were found in the suburbs,
36
+ and it was the City of London itself that was destroyed by the Fire. According
37
+ - Monument to the Great Fire of London Monument to the Great Fire of London The
38
+ Monument to the Great Fire of London, more commonly known simply as the Monument,
39
+ is a Doric column in London, United Kingdom, situated near the northern end of
40
+ London Bridge. Commemorating the Great Fire of London, it stands at the junction
41
+ of Monument Street and Fish Street Hill, in height and 202 feet west of the spot
42
+ in Pudding Lane where the Great Fire started on 2 September 1666. Constructed
43
+ between 1671 and 1677, it was built on the site of St. Margaret's, Fish Street,
44
+ the first church to be destroyed by
45
+ - 'How to Have Sex in an Epidemic New York City government and organizations within
46
+ the LGBT community. The Gay Men''s Health Crisis offered to buy all 5,000 pamphlets
47
+ and promote them, with the condition that any mentions of the multifactorial model
48
+ be removed from the writing. The authors refused. Berkowitz recounts in an interview
49
+ it being "infuriating" that in 1985, the city still hadn''t adopted any standard
50
+ safe sex education. The advent of safe sex in urban gay male populations came
51
+ too late for many people: by 1983, more than 1,476 people had died from AIDS and
52
+ David France estimated that as much as half of all'
53
+ - 'Monument to the Great Fire of London six years to complete the 202 ft column.
54
+ It was two more years before the inscription (which had been left to Wren — or
55
+ to Wren''s choice — to decide upon) was set in place. "Commemorating — with a
56
+ brazen disregard for the truth — the fact that ''London rises again...three short
57
+ years complete that which was considered the work of ages.''" Hooke''s surviving
58
+ drawings show that several versions of the monument were submitted for consideration:
59
+ a plain obelisk, a column garnished with tongues of fire, and the fluted Doric
60
+ column that was eventually chosen. The real contention came with'
61
+ - source_sentence: '"The Claude Francois song ""Comme d''habitude"" (translation ""as
62
+ usual"") was a hit in English for Frank Sinatra under what title?"'
63
+ sentences:
64
+ - Young at Heart (Frank Sinatra song) young, Dick Van Dyke recorded a duet with
65
+ his wife, Arlene, at Capital Records Studio in Los Angeles, filmed for the HBO
66
+ Special on aging "If I'm not in the Obituary, I'll have Breakfast" starring Carl
67
+ Reiner, and featuring other young at heart +90 treasures, Mel Brooks, Norman Lear,
68
+ Stan Lee & Betty White among others. Van Dyke was recorded using Frank Sinatra's
69
+ microphone. Young at Heart (Frank Sinatra song) "Young at Heart" is a pop standard,
70
+ a ballad with music by Johnny Richards and lyrics by Carolyn Leigh. The song was
71
+ written and published in 1953, with Leigh contributing
72
+ - 'Comme d''habitude a relationship that is falling out of love, while the English
73
+ language version is set at the end of a lifetime, approaching death, and looking
74
+ back without regret – expressing feelings that are more related to Piaf''s song
75
+ "Non, je ne regrette rien". Many artists sang "Comme d''Habitude" in French after
76
+ Claude François''s success (and international success through ''"My Way"), notably:
77
+ David Bowie has said that in 1968 – the year before Paul Anka acquired the French
78
+ song – his manager, Kenneth Pitt, asked him to write English lyrics for "Comme
79
+ d''habitude" but that his version, titled "Even a Fool'
80
+ - Frank Sinatra Me" with Billy May, designed as a musical world tour. It reached
81
+ the top spot on the Billboard album chart in its second week, remaining at the
82
+ top for five weeks, and was nominated for the Grammy Award for Album of the Year
83
+ at the inaugural Grammy Awards. The title song, "Come Fly With Me", written especially
84
+ for him, would become one of his best known standards. On May 29 he recorded seven
85
+ songs in a single session, more than double the usual yield of a recording session,
86
+ and an eighth was planned, "Lush Life", but Sinatra found it too
87
+ - Frank Sinatra Original Song. Sinatra released "Softly, as I Leave You", and collaborated
88
+ with Bing Crosby and Fred Waring on "America, I Hear You Singing", a collection
89
+ of patriotic songs recorded as a tribute to the assassinated President John F.
90
+ Kennedy. Sinatra increasingly became involved in charitable pursuits in this period.
91
+ In 1961 and 1962 he went to Mexico, with the sole purpose of putting on performances
92
+ for Mexican charities, and in July 1964 he was present for the dedication of the
93
+ Frank Sinatra International Youth Center for Arab and Jewish children in Nazareth.
94
+ Sinatra's phenomenal success in 1965, coinciding with his
95
+ - Comme ci comme ça (Basim song) to the charm of it all. Working both Danish and
96
+ Moroccan Arabic, Basim sings about a girl he is ready to commit to. It doesn’t
97
+ mater what she wants to do — it’s comme ci comme ça — and he just wants her."
98
+ An official music video to accompany the release of "Comme ci comme ça" was first
99
+ released onto YouTube on 20 September 2017 at a total length of three minutes
100
+ and twelve seconds. Comme ci comme ça (Basim song) "Comme ci comme ça" is a song
101
+ performed by Danish pop singer and songwriter Basim, featuring vocals from Gilli.
102
+ - Personal life of Frank Sinatra A third child, Christina Sinatra, known as "Tina",
103
+ was born on June 20, 1948. Nancy Barbato Sinatra and Frank Sinatra announced their
104
+ separation on Valentine's Day, February 14, 1950, with Frank's additional extra-marital
105
+ affair with Ava Gardner compounding his transgressions and becoming public knowledge
106
+ once again. After originally just seeking a legal separation, Frank and Nancy
107
+ Sinatra decided some months later to file for divorce, and this divorce became
108
+ legally final on October 29, 1951. Frank Sinatra's affair and relationship with
109
+ Gardner had become more and more serious, and she later became his second wife.
110
+ What was perhaps less widely
111
+ - source_sentence: What was the name of the first Indiana Jones movie?
112
+ sentences:
113
+ - Indiana Jones and the Temple of Doom point. Old-time, 15-part movie serials didn't
114
+ have shape. They just went on and on and on, which is what "Temple of Doom" does
115
+ with humor and technical invention." Neal Gabler commented that "I think in some
116
+ ways, "Indiana Jones and the Temple of Doom" was better than "Raiders of the Lost
117
+ Ark". In some ways it was less. In sum total, I'd have to say I enjoyed it more.
118
+ That doesn't mean it's better necessarily, but I got more enjoyment out of it."
119
+ Colin Covert of the "Star Tribune" called the film "sillier, darkly violent and
120
+ a bit dumbed down,
121
+ - Indiana Jones and the Temple of Doom (1985 video game) Theme music plays in the
122
+ background which is the best part of the game. Most of the sound effects are not
123
+ sharp and not enough of them exist. "Indiana Jones and the Temple of Doom" is
124
+ a bad game all the way around. It looks bad, has bad controls, and is way too
125
+ short." Indiana Jones and the Temple of Doom (1985 video game) Indiana Jones and
126
+ The Temple of Doom is a 1985 action arcade game developed and published by Atari
127
+ Games, based on the 1984 film of the same name, the second film in the "Indiana
128
+ Jones" franchise.
129
+ - Indiana Jones and the Spear of Destiny Indiana Jones and the Spear of Destiny
130
+ Indiana Jones and The Spear of Destiny is a four-issue comic book mini-series
131
+ published by Dark Horse Comics from April to July 1995. It was their seventh series
132
+ about the adult Indiana Jones. Indiana Jones reached for the Holy Grail, perched
133
+ in a crack in the Temple of the Sun. Hanging onto him, his father, Professor Henry
134
+ Jones urged him to let it go, and Indy turned back and let his father help him
135
+ up. As the Joneses ride out into the Canyon of the Crescent Moon with Marcus Brody
136
+ and Sallah, they
137
+ - Lego Indiana Jones sets" The line was discontinued in 2010, but since Lucas plans
138
+ to make a fifth installment to the franchise, the sets may be re-released along
139
+ with new sets of the possible fifth Indiana Jones film. Due to the fact Disney
140
+ bought Lucasfilm and will be making a new Indiana Jones movie, chances of new
141
+ sets are high. The Indiana Jones sets proved to be one of the most popular Lego
142
+ themes, and by the end of 2008 were credited, along with Lego Star Wars, of boosting
143
+ the Lego Group's profits within a stagnant toy market. The product line was said
144
+ - Indiana Jones and the Staff of Kings point-and-click adventure "Indiana Jones
145
+ and the Fate of Atlantis". GameSpot criticized its "terribly laid-out checkpoints",
146
+ "out-of-date" visuals, and "atrocious, annoying motion controls". Indiana Jones
147
+ and the Staff of Kings The game was initially developed for the higher-end PlayStation
148
+ 3 and Xbox 360 systems, before switching to the aforementioned lower-end platforms.
149
+ As a result, both systems never saw a proper "Indiana Jones" video game being
150
+ released besides the "" duology. The plot centers around Indy's search for the
151
+ Staff of Moses. The Wii version of the game includes an exclusive co-op story
152
+ mode (with Indy and Henry Jones Sr.) and unlockable
153
+ - 'Indiana Jones and the Last Crusade: The Graphic Adventure Indiana Jones and the
154
+ Last Crusade: The Graphic Adventure Indiana Jones and the Last Crusade: The Graphic
155
+ Adventure is a graphic adventure game, released in 1989 (to coincide with the
156
+ release of the film of the same name), published by Lucasfilm Games (now LucasArts).
157
+ It was the third game to use the SCUMM engine. "Last Crusade" was one of the most
158
+ innovative of the LucasArts adventures. It expanded on LucasArts'' traditional
159
+ adventure game structure by including a flexible point system—the IQ score, or
160
+ "Indy Quotient"—and by allowing the game to be completed in several different
161
+ ways. The point system was'
162
+ - source_sentence: '"Who was the Anglo-Irish scientist who, in the 17th century, discovered
163
+ that ""the volume of a given mass of gas at a given temperature is inversely proportional
164
+ to its pressure""?"'
165
+ sentences:
166
+ - 'Gay-Lussac''s law Gay-Lussac''s law Gay-Lussac''s law can refer to several discoveries
167
+ made by French chemist Joseph Louis Gay-Lussac (1778–1850) and other scientists
168
+ in the late 18th and early 19th centuries pertaining to thermal expansion of gases
169
+ and the relationship between temperature, volume, and pressure. It states that
170
+ the pressure of a given mass of gas varies directly with the absolute temperature
171
+ of the gas, when the volume is kept constant. Mathematically, it can be written
172
+ as: P/T=constant, Gay-Lussac is most often recognized for the Pressure Law which
173
+ established that the pressure of an enclosed gas is directly proportional to its
174
+ temperature and'
175
+ - 'Gas constant "V" is the volume of gas (SI unit cubic metres), "n" is the amount
176
+ of gas (SI unit moles), "m" is the mass (SI unit kilograms) contained in "V",
177
+ and "T" is the thermodynamic temperature (SI unit kelvins). "R" is the molar-weight-specific
178
+ gas constant, discussed below. The gas constant is expressed in the same physical
179
+ units as molar entropy and molar heat capacity. From the general equation "PV"
180
+ = "nRT" we get: where "P" is pressure, "V" is volume, "n" is number of moles of
181
+ a given substance, and "T" is temperature. As pressure is defined as force per
182
+ unit'
183
+ - The Boy Who Was a King term. The film presents not only the life of the former
184
+ Tsar, but also intertwines within the story vignettes of various Bulgarians, who
185
+ were supporting him, sending him gifts, or merely tattooing his face on their
186
+ body. The story is told through personal footage and vast amounts of archive material.
187
+ The film received praise for its editing and use of archives with Variety's Robert
188
+ Koehler writing that "Pic’s terrific use of archival footage includes an exiled
189
+ Simeon interviewed in the early ’60s, disputing his playboy rep." and "Editing
190
+ is aces." The Boy Who Was a King The Boy Who Was
191
+ - Francis Hauksbee In 1708, Hauksbee independently discovered Charles's law of gases,
192
+ which states that, for a given mass of gas at a constant pressure, the volume
193
+ of the gas is proportional to its temperature. Hauksbee published accounts of
194
+ his experiments in the Royal Society's journal "Philosophical Transactions". In
195
+ 1709 he self-published "Physico-Mechanical Experiments on Various Subjects" which
196
+ collected together many of these experiments along with discussion that summarized
197
+ much of his scientific work. An Italian translation was published in 1716. A second
198
+ edition was published posthumously in 1719. There were also translations to Dutch
199
+ (1735) and French (1754). The Royal Society Hauksbee
200
+ - 'Boyle''s law air moves from high to low pressure. Related phenomena: Other gas
201
+ laws: Boyle''s law Boyle''s law, sometimes referred to as the Boyle–Mariotte law,
202
+ or Mariotte''s law (especially in France), is an experimental gas law that describes
203
+ how the pressure of a gas tends to increase as the volume of the container decreases.
204
+ A modern statement of Boyle''s law is The absolute pressure exerted by a given
205
+ mass of an ideal gas is inversely proportional to the volume it occupies if the
206
+ temperature and amount of gas remain unchanged within a closed system. Mathematically,
207
+ Boyle''s law can be stated as or'
208
+ - Boyle's law of the gas, and "k" is a constant. The equation states that the product
209
+ of pressure and volume is a constant for a given mass of confined gas and this
210
+ holds as long as the temperature is constant. For comparing the same substance
211
+ under two different sets of conditions, the law can be usefully expressed as The
212
+ equation shows that, as volume increases, the pressure of the gas decreases in
213
+ proportion. Similarly, as volume decreases, the pressure of the gas increases.
214
+ The law was named after chemist and physicist Robert Boyle, who published the
215
+ original law in 1662. This relationship
216
+ - source_sentence: Peter Stuyvesant, born in Holland, became Governor of which American
217
+ city in 1647?
218
+ sentences:
219
+ - Peter Stuyvesant at the corner of Thirteenth Street and Third Avenue until 1867
220
+ when it was destroyed by a storm, bearing fruit almost to the last. The house
221
+ was destroyed by fire in 1777. He also built an executive mansion of stone called
222
+ Whitehall. In 1645, Stuyvesant married Judith Bayard (–1687) of the Bayard family.
223
+ Her brother, Samuel Bayard, was the husband of Stuyvesant's sister, Anna Stuyvesant.
224
+ Petrus and Judith had two sons together. He died in August 1672 and his body was
225
+ entombed in the east wall of St. Mark's Church in-the-Bowery, which sits on the
226
+ site of Stuyvesant’s family chapel.
227
+ - 'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal
228
+ prosecution - if companies wilfully break the laws. However last year, when questioned
229
+ on why no such action was being pursued against Imperial Tobacco a spokeswoman
230
+ for Federal Health said: ""No instances of non-compliance with the Act have been
231
+ identified by the Department that warrant the initiation of Court proceedings
232
+ in the first instance, and without attempting alternative dispute resolution to
233
+ achieve compliance"". Peter Stuyvesant is or was sold in the following countries:
234
+ Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany,
235
+ France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,'
236
+ - Jochem Pietersen Kuyter September 25, 1647, until the city was incorporated, in
237
+ 1653, when he was made schout (sheriff). Kuyter twice came in conflict with the
238
+ Director of New Netherland. Kuyter was a man of good education, what is evident
239
+ by his dealings with Willem Kieft., who he believed damaged the colony with his
240
+ policies and the start of Kieft's War in 1643. In 1647, when Peter Stuyvesant
241
+ arrived in New Amsterdam to replace Kieft, Kuyter and Cornelis Melyn acting in
242
+ name of the citizens of New Amsterdam, brought charges against the outgoing governor,
243
+ demanding an investigation of his conduct while in office.
244
+ - Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging
245
+ changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported
246
+ that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers
247
+ which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came
248
+ with 26 cigarettes. The material included images of a young woman with pink hair
249
+ putting on lipstick and men on the streets of New York and also included a calendar
250
+ and small poster that were clearly intended to glamorise smoking. Anti-smoking
251
+ campaigner Mike Daube said although the material did not break the law because
252
+ - 'Peter Stuyvesant but the order was soon revoked under pressure from the States
253
+ of Holland and the city of Amsterdam. Stuyvesant prepared against an attack by
254
+ ordering the citizens to dig a ditch from the North River to the East River and
255
+ to erect a fortification. In 1653, a convention of two deputies from each village
256
+ in New Netherland demanded reforms, and Stuyvesant commanded that assembly to
257
+ disperse, saying: "We derive our authority from God and the company, not from
258
+ a few ignorant subjects." In the summer of 1655, he sailed down the Delaware River
259
+ with a fleet of seven vessels and'
260
+ - Peter Stuyvesant Dutch Reformed church, a Calvinist denomination, holding to the
261
+ Three Forms of Unity (Belgic Confession, Heidelberg Catechism, Canons of Dordt).
262
+ The English were Anglicans, holding to the 39 Articles, a Protestant confession,
263
+ with bishops. In 1665, Stuyvesant went to the Netherlands to report on his term
264
+ as governor. On his return to the colony, he spent the remainder of his life on
265
+ his farm of sixty-two acres outside the city, called the Great Bouwerie, beyond
266
+ which stretched the woods and swamps of the village of Nieuw Haarlem. A pear tree
267
+ that he reputedly brought from the Netherlands in 1647 remained
268
+ pipeline_tag: sentence-similarity
269
+ library_name: sentence-transformers
270
+ ---
271
+
272
+ # SentenceTransformer based on seongil-dn/unsupervised_20m_3800
273
+
274
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
275
+
276
+ ## Model Details
277
+
278
+ ### Model Description
279
+ - **Model Type:** Sentence Transformer
280
+ - **Base model:** [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800) <!-- at revision 1cda749f242e2b5c9e4f3c1122a61e76fec1fee5 -->
281
+ - **Maximum Sequence Length:** 1024 tokens
282
+ - **Output Dimensionality:** 1024 dimensions
283
+ - **Similarity Function:** Cosine Similarity
284
+ <!-- - **Training Dataset:** Unknown -->
285
+ <!-- - **Language:** Unknown -->
286
+ <!-- - **License:** Unknown -->
287
+
288
+ ### Model Sources
289
+
290
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
291
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
292
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
293
+
294
+ ### Full Model Architecture
295
+
296
+ ```
297
+ SentenceTransformer(
298
+ (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
299
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
300
+ (2): Normalize()
301
+ )
302
+ ```
303
+
304
+ ## Usage
305
+
306
+ ### Direct Usage (Sentence Transformers)
307
+
308
+ First install the Sentence Transformers library:
309
+
310
+ ```bash
311
+ pip install -U sentence-transformers
312
+ ```
313
+
314
+ Then you can load this model and run inference.
315
+ ```python
316
+ from sentence_transformers import SentenceTransformer
317
+
318
+ # Download from the 🤗 Hub
319
+ model = SentenceTransformer("seongil-dn/bge-m3-672")
320
+ # Run inference
321
+ sentences = [
322
+ 'Peter Stuyvesant, born in Holland, became Governor of which American city in 1647?',
323
+ 'Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came with 26 cigarettes. The material included images of a young woman with pink hair putting on lipstick and men on the streets of New York and also included a calendar and small poster that were clearly intended to glamorise smoking. Anti-smoking campaigner Mike Daube said although the material did not break the law because',
324
+ 'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal prosecution - if companies wilfully break the laws. However last year, when questioned on why no such action was being pursued against Imperial Tobacco a spokeswoman for Federal Health said: ""No instances of non-compliance with the Act have been identified by the Department that warrant the initiation of Court proceedings in the first instance, and without attempting alternative dispute resolution to achieve compliance"". Peter Stuyvesant is or was sold in the following countries: Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany, France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,',
325
+ ]
326
+ embeddings = model.encode(sentences)
327
+ print(embeddings.shape)
328
+ # [3, 1024]
329
+
330
+ # Get the similarity scores for the embeddings
331
+ similarities = model.similarity(embeddings, embeddings)
332
+ print(similarities.shape)
333
+ # [3, 3]
334
+ ```
335
+
336
+ <!--
337
+ ### Direct Usage (Transformers)
338
+
339
+ <details><summary>Click to see the direct usage in Transformers</summary>
340
+
341
+ </details>
342
+ -->
343
+
344
+ <!--
345
+ ### Downstream Usage (Sentence Transformers)
346
+
347
+ You can finetune this model on your own dataset.
348
+
349
+ <details><summary>Click to expand</summary>
350
+
351
+ </details>
352
+ -->
353
+
354
+ <!--
355
+ ### Out-of-Scope Use
356
+
357
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
358
+ -->
359
+
360
+ <!--
361
+ ## Bias, Risks and Limitations
362
+
363
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
364
+ -->
365
+
366
+ <!--
367
+ ### Recommendations
368
+
369
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
370
+ -->
371
+
372
+ ## Training Details
373
+
374
+ ### Training Dataset
375
+
376
+ #### Unnamed Dataset
377
+
378
+ * Size: 1,138,596 training samples
379
+ * Columns: <code>anchor</code>, <code>positive</code>, <code>negative</code>, <code>negative_2</code>, <code>negative_3</code>, <code>negative_4</code>, and <code>negative_5</code>
380
+ * Approximate statistics based on the first 1000 samples:
381
+ | | anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 |
382
+ |:--------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|
383
+ | type | string | string | string | string | string | string | string |
384
+ | details | <ul><li>min: 9 tokens</li><li>mean: 22.32 tokens</li><li>max: 119 tokens</li></ul> | <ul><li>min: 127 tokens</li><li>mean: 157.45 tokens</li><li>max: 420 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 154.65 tokens</li><li>max: 212 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 155.52 tokens</li><li>max: 218 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 156.04 tokens</li><li>max: 284 tokens</li></ul> | <ul><li>min: 124 tokens</li><li>mean: 156.3 tokens</li><li>max: 268 tokens</li></ul> | <ul><li>min: 121 tokens</li><li>mean: 156.15 tokens</li><li>max: 249 tokens</li></ul> |
385
+ * Samples:
386
+ | anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 |
387
+ |:---------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
388
+ | <code>What African country is projected to pass the United States in population by the year 2055?</code> | <code>African immigration to the United States officially 40,000 African immigrants, although it has been estimated that the population is actually four times this number when considering undocumented immigrants. The majority of these immigrants were born in Ethiopia, Egypt, Nigeria, and South Africa. African immigrants like many other immigrant groups are likely to establish and find success in small businesses. Many Africans that have seen the social and economic stability that comes from ethnic enclaves such as Chinatowns have recently been establishing ethnic enclaves of their own at much higher rates to reap the benefits of such communities. Such examples include Little Ethiopia in Los Angeles and</code> | <code>What Will Happen to the Gang Next Year? watching television at the time of the broadcast. This made it the lowest-rated episode in "30 Rock"<nowiki>'</nowiki>s history. and a decrease from the previous episode "The Return of Avery Jessup" (2.92 million) What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States</code> | <code>Christianity in the United States Christ is the fifth-largest denomination, the largest Pentecostal church, and the largest traditionally African-American denomination in the nation. Among Eastern Christian denominations, there are several Eastern Orthodox and Oriental Orthodox churches, with just below 1 million adherents in the US, or 0.4% of the total population. Christianity was introduced to the Americas as it was first colonized by Europeans beginning in the 16th and 17th centuries. Going forward from its foundation, the United States has been called a Protestant nation by a variety of sources. Immigration further increased Christian numbers. Today most Christian churches in the United States are either</code> | <code>What Will Happen to the Gang Next Year? What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States on May 17, 2012. In the episode, Jack (Alec Baldwin) and Avery (Elizabeth Banks) seek to renew their vows; Criss (James Marsden) sets out to show Liz (Tina Fey) he can pay</code> | <code>History of the Jews in the United States Representatives by Rep. Samuel Dickstein (D; New York). This also failed to pass. During the Holocaust, fewer than 30,000 Jews a year reached the United States, and some were turned away due to immigration policies. The U.S. did not change its immigration policies until 1948. Currently, laws requiring teaching of the Holocaust are on the books in five states. The Holocaust had a profound impact on the community in the United States, especially after 1960, as Jews tried to comprehend what had happened, and especially to commemorate and grapple with it when looking to the future. Abraham Joshua Heschel summarized</code> | <code>Public holidays in the United States will have very few customers that day. The labor force in the United States comprises about 62% (as of 2014) of the general population. In the United States, 97% of the private sector businesses determine what days this sector of the population gets paid time off, according to a study by the Society for Human Resource Management. The following holidays are observed by the majority of US businesses with paid time off: This list of holidays is based off the official list of federal holidays by year from the US Government. The holidays however are at the discretion of employers</code> |
389
+ | <code>Which is the largest species of the turtle family?</code> | <code>Loggerhead sea turtle turtle is debated, but most authors consider it a single polymorphic species. Molecular genetics has confirmed hybridization of the loggerhead sea turtle with the Kemp's ridley sea turtle, hawksbill sea turtle, and green sea turtles. The extent of natural hybridization is not yet determined; however, second-generation hybrids have been reported, suggesting some hybrids are fertile. Although evidence is lacking, modern sea turtles probably descended from a single common ancestor during the Cretaceous period. Like all other sea turtles except the leatherback, loggerheads are members of the ancient family Cheloniidae, and appeared about 40 million years ago. Of the six species</code> | <code>Convention on the Conservation of Migratory Species of Wild Animals take joint action. At May 2018, there were 126 Parties to the Convention. The CMS Family covers a great diversity of migratory species. The Appendices of CMS include many mammals, including land mammals, marine mammals and bats; birds; fish; reptiles and one insect. Among the instruments, AEWA covers 254 species of birds that are ecologically dependent on wetlands for at least part of their annual cycle. EUROBATS covers 52 species of bat, the Memorandum of Understanding on the Conservation of Migratory Sharks seven species of shark, the IOSEA Marine Turtle MOU six species of marine turtle and the Raptors MoU</code> | <code>Razor-backed musk turtle Razor-backed musk turtle The razor-backed musk turtle ("Sternotherus carinatus") is a species of turtle in the family Kinosternidae. The species is native to the southern United States. There are no subspecies that are recognized as being valid. "S. carinatus" is found in the states of Alabama, Arkansas, Louisiana, Mississippi, Oklahoma, and Texas. The razor-backed musk turtle grows to a straight carapace length of about . It has a brown-colored carapace, with black markings at the edges of each scute. The carapace has a distinct, sharp keel down the center of its length, giving the species its common name. The body</code> | <code>African helmeted turtle African helmeted turtle The African helmeted turtle ("Pelomedusa subrufa"), also known commonly as the marsh terrapin, the crocodile turtle, or in the pet trade as the African side-necked turtle, is a species of omnivorous side-necked terrapin in the family Pelomedusidae. The species naturally occurs in fresh and stagnant water bodies throughout much of Sub-Saharan Africa, and in southern Yemen. The marsh terrapin is typically a rather small turtle, with most individuals being less than in straight carapace length, but one has been recorded with a length of . It has a black or brown carapace. The top of the tail</code> | <code>Box turtle Box turtle Box turtles are North American turtles of the genus Terrapene. Although box turtles are superficially similar to tortoises in terrestrial habits and overall appearance, they are actually members of the American pond turtle family (Emydidae). The twelve taxa which are distinguished in the genus are distributed over four species. They are largely characterized by having a domed shell, which is hinged at the bottom, allowing the animal to close its shell tightly to escape predators. The genus name "Terrapene" was coined by Merrem in 1820 as a genus separate from "Emys" for those species which had a sternum</code> | <code>Vallarta mud turtle Vallarta mud turtle The Vallarta mud turtle ("Kinosternon vogti") is a recently identified species of mud turtle in the family Kinosternidae. While formerly considered conspecific with the Jalisco mud turtle, further studies indicated that it was a separate species. It can be identified by a combination of the number of plastron and carapace scutes, body size, and the distinctive yellow rostral shield in males. It is endemic to Mexican state of Jalisco. It is only known from a few human-created or human-affected habitats (such as small streams and ponds) found around Puerto Vallarta. It is one of only 3 species</code> |
390
+ | <code>How many gallons of beer are in an English barrel?</code> | <code>Low-alcohol beer Prohibition in the United States. Near beer could not legally be labeled as "beer" and was officially classified as a "cereal beverage". The public, however, almost universally called it "near beer". The most popular "near beer" was Bevo, brewed by the Anheuser-Busch company. The Pabst company brewed "Pablo", Miller brewed "Vivo", and Schlitz brewed "Famo". Many local and regional breweries stayed in business by marketing their own near-beers. By 1921 production of near beer had reached over 300 million US gallons (1 billion L) a year (36 L/s). A popular illegal practice was to add alcohol to near beer. The</code> | <code>Keg terms "half-barrel" and "quarter-barrel" are derived from the U.S. beer barrel, legally defined as being equal to 31 U.S. gallons (this is not the same volume as some other units also known as "barrels"). A 15.5 U.S. gallon keg is also equal to: However, beer kegs can come in many sizes: In European countries the most common keg size is 50 liters. This includes the UK, which uses a non-metric standard keg of 11 imperial gallons, which is coincidentally equal to . The German DIN 6647-1 and DIN 6647-2 have also defined kegs in the sizes of 30 and 20</code> | <code>Beer in Chile craft beers. They are generally low or very low volume producers. In Chile there are more than 150 craft beer producers distributed along the 15 Chilean Regions. The list below includes: Beer in Chile The primary beer brewed and consumed in Chile is pale lager, though the country also has a tradition of brewing corn beer, known as chicha. Chile’s beer history has a strong German influence – some of the bigger beer producers are from the country’s southern lake district, a region populated by a great number of German immigrants during the 19th century. Chile also produces English ale-style</code> | <code>Barrel variation. In modern times, produce barrels for all dry goods, excepting cranberries, contain 7,056 cubic inches, about 115.627 L. Barrel A barrel, cask, or tun is a hollow cylindrical container, traditionally made of wooden staves bound by wooden or metal hoops. Traditionally, the barrel was a standard size of measure referring to a set capacity or weight of a given commodity. For example, in the UK a barrel of beer refers to a quantity of . Wine was shipped in barrels of . Modern wooden barrels for wine-making are either made of French common oak ("Quercus robur") and white oak</code> | <code>The Rare Barrel The Rare Barrel The Rare Barrel is a brewery and brewpub in Berkeley, California, United States, that exclusively produces sour beers. Founders Jay Goodwin and Alex Wallash met while attending UCSB. They started home-brewing in their apartment and decided that they would one day start a brewery together. Goodwin started working at The Bruery, where he worked his way from a production assistant to brewer, eventually becoming the head of their barrel aging program. The Rare Barrel brewed its first batch of beer in February 2013, and opened its tasting room on December 27, 2013. The Rare Barrel was named</code> | <code>Barrel (unit) Barrel (unit) A barrel is one of several units of volume applied in various contexts; there are dry barrels, fluid barrels (such as the UK beer barrel and US beer barrel), oil barrels and so on. For historical reasons the volumes of some barrel units are roughly double the volumes of others; volumes in common usage range from about . In many connections the term "drum" is used almost interchangeably with "barrel". Since medieval times the term barrel as a unit of measure has had various meanings throughout Europe, ranging from about 100 litres to 1000 litres. The name was</code> |
391
+ * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters:
392
+ ```json
393
+ {'guide': SentenceTransformer(
394
+ (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
395
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
396
+ (2): Normalize()
397
+ ), 'temperature': 0.01}
398
+ ```
399
+
400
+ ### Training Hyperparameters
401
+ #### Non-Default Hyperparameters
402
+
403
+ - `per_device_train_batch_size`: 1024
404
+ - `learning_rate`: 3e-05
405
+ - `weight_decay`: 0.01
406
+ - `warmup_ratio`: 0.05
407
+ - `bf16`: True
408
+ - `batch_sampler`: no_duplicates
409
+
410
+ #### All Hyperparameters
411
+ <details><summary>Click to expand</summary>
412
+
413
+ - `overwrite_output_dir`: False
414
+ - `do_predict`: False
415
+ - `eval_strategy`: no
416
+ - `prediction_loss_only`: True
417
+ - `per_device_train_batch_size`: 1024
418
+ - `per_device_eval_batch_size`: 8
419
+ - `per_gpu_train_batch_size`: None
420
+ - `per_gpu_eval_batch_size`: None
421
+ - `gradient_accumulation_steps`: 1
422
+ - `eval_accumulation_steps`: None
423
+ - `torch_empty_cache_steps`: None
424
+ - `learning_rate`: 3e-05
425
+ - `weight_decay`: 0.01
426
+ - `adam_beta1`: 0.9
427
+ - `adam_beta2`: 0.999
428
+ - `adam_epsilon`: 1e-08
429
+ - `max_grad_norm`: 1.0
430
+ - `num_train_epochs`: 3
431
+ - `max_steps`: -1
432
+ - `lr_scheduler_type`: linear
433
+ - `lr_scheduler_kwargs`: {}
434
+ - `warmup_ratio`: 0.05
435
+ - `warmup_steps`: 0
436
+ - `log_level`: passive
437
+ - `log_level_replica`: warning
438
+ - `log_on_each_node`: True
439
+ - `logging_nan_inf_filter`: True
440
+ - `save_safetensors`: True
441
+ - `save_on_each_node`: False
442
+ - `save_only_model`: False
443
+ - `restore_callback_states_from_checkpoint`: False
444
+ - `no_cuda`: False
445
+ - `use_cpu`: False
446
+ - `use_mps_device`: False
447
+ - `seed`: 42
448
+ - `data_seed`: None
449
+ - `jit_mode_eval`: False
450
+ - `use_ipex`: False
451
+ - `bf16`: True
452
+ - `fp16`: False
453
+ - `fp16_opt_level`: O1
454
+ - `half_precision_backend`: auto
455
+ - `bf16_full_eval`: False
456
+ - `fp16_full_eval`: False
457
+ - `tf32`: None
458
+ - `local_rank`: 0
459
+ - `ddp_backend`: None
460
+ - `tpu_num_cores`: None
461
+ - `tpu_metrics_debug`: False
462
+ - `debug`: []
463
+ - `dataloader_drop_last`: True
464
+ - `dataloader_num_workers`: 0
465
+ - `dataloader_prefetch_factor`: None
466
+ - `past_index`: -1
467
+ - `disable_tqdm`: False
468
+ - `remove_unused_columns`: True
469
+ - `label_names`: None
470
+ - `load_best_model_at_end`: False
471
+ - `ignore_data_skip`: False
472
+ - `fsdp`: []
473
+ - `fsdp_min_num_params`: 0
474
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
475
+ - `fsdp_transformer_layer_cls_to_wrap`: None
476
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
477
+ - `deepspeed`: None
478
+ - `label_smoothing_factor`: 0.0
479
+ - `optim`: adamw_torch
480
+ - `optim_args`: None
481
+ - `adafactor`: False
482
+ - `group_by_length`: False
483
+ - `length_column_name`: length
484
+ - `ddp_find_unused_parameters`: None
485
+ - `ddp_bucket_cap_mb`: None
486
+ - `ddp_broadcast_buffers`: False
487
+ - `dataloader_pin_memory`: True
488
+ - `dataloader_persistent_workers`: False
489
+ - `skip_memory_metrics`: True
490
+ - `use_legacy_prediction_loop`: False
491
+ - `push_to_hub`: False
492
+ - `resume_from_checkpoint`: None
493
+ - `hub_model_id`: None
494
+ - `hub_strategy`: every_save
495
+ - `hub_private_repo`: None
496
+ - `hub_always_push`: False
497
+ - `gradient_checkpointing`: False
498
+ - `gradient_checkpointing_kwargs`: None
499
+ - `include_inputs_for_metrics`: False
500
+ - `include_for_metrics`: []
501
+ - `eval_do_concat_batches`: True
502
+ - `fp16_backend`: auto
503
+ - `push_to_hub_model_id`: None
504
+ - `push_to_hub_organization`: None
505
+ - `mp_parameters`:
506
+ - `auto_find_batch_size`: False
507
+ - `full_determinism`: False
508
+ - `torchdynamo`: None
509
+ - `ray_scope`: last
510
+ - `ddp_timeout`: 1800
511
+ - `torch_compile`: False
512
+ - `torch_compile_backend`: None
513
+ - `torch_compile_mode`: None
514
+ - `dispatch_batches`: None
515
+ - `split_batches`: None
516
+ - `include_tokens_per_second`: False
517
+ - `include_num_input_tokens_seen`: False
518
+ - `neftune_noise_alpha`: None
519
+ - `optim_target_modules`: None
520
+ - `batch_eval_metrics`: False
521
+ - `eval_on_start`: False
522
+ - `use_liger_kernel`: False
523
+ - `eval_use_gather_object`: False
524
+ - `average_tokens_across_devices`: False
525
+ - `prompts`: None
526
+ - `batch_sampler`: no_duplicates
527
+ - `multi_dataset_batch_sampler`: proportional
528
+
529
+ </details>
530
+
531
+ ### Training Logs
532
+ <details><summary>Click to expand</summary>
533
+
534
+ | Epoch | Step | Training Loss |
535
+ |:------:|:----:|:-------------:|
536
+ | 0.0036 | 1 | 1.0283 |
537
+ | 0.0072 | 2 | 1.0155 |
538
+ | 0.0108 | 3 | 0.9858 |
539
+ | 0.0144 | 4 | 0.9519 |
540
+ | 0.0181 | 5 | 0.9434 |
541
+ | 0.0217 | 6 | 0.898 |
542
+ | 0.0253 | 7 | 0.8798 |
543
+ | 0.0289 | 8 | 0.7976 |
544
+ | 0.0325 | 9 | 0.7797 |
545
+ | 0.0361 | 10 | 0.7464 |
546
+ | 0.0397 | 11 | 0.743 |
547
+ | 0.0433 | 12 | 0.716 |
548
+ | 0.0469 | 13 | 0.7076 |
549
+ | 0.0505 | 14 | 0.666 |
550
+ | 0.0542 | 15 | 0.631 |
551
+ | 0.0578 | 16 | 0.5905 |
552
+ | 0.0614 | 17 | 0.6537 |
553
+ | 0.0650 | 18 | 0.5755 |
554
+ | 0.0686 | 19 | 0.5422 |
555
+ | 0.0722 | 20 | 0.5393 |
556
+ | 0.0758 | 21 | 0.5741 |
557
+ | 0.0794 | 22 | 0.498 |
558
+ | 0.0830 | 23 | 0.5522 |
559
+ | 0.0866 | 24 | 0.5592 |
560
+ | 0.0903 | 25 | 0.4797 |
561
+ | 0.0939 | 26 | 0.4684 |
562
+ | 0.0975 | 27 | 0.5207 |
563
+ | 0.1011 | 28 | 0.4692 |
564
+ | 0.1047 | 29 | 0.4459 |
565
+ | 0.1083 | 30 | 0.4439 |
566
+ | 0.1119 | 31 | 0.4656 |
567
+ | 0.1155 | 32 | 0.4737 |
568
+ | 0.1191 | 33 | 0.4391 |
569
+ | 0.1227 | 34 | 0.4386 |
570
+ | 0.1264 | 35 | 0.4107 |
571
+ | 0.1300 | 36 | 0.4513 |
572
+ | 0.1336 | 37 | 0.3789 |
573
+ | 0.1372 | 38 | 0.4103 |
574
+ | 0.1408 | 39 | 0.3929 |
575
+ | 0.1444 | 40 | 0.4226 |
576
+ | 0.1480 | 41 | 0.391 |
577
+ | 0.1516 | 42 | 0.3674 |
578
+ | 0.1552 | 43 | 0.3607 |
579
+ | 0.1588 | 44 | 0.3738 |
580
+ | 0.1625 | 45 | 0.3842 |
581
+ | 0.1661 | 46 | 0.3498 |
582
+ | 0.1697 | 47 | 0.3586 |
583
+ | 0.1733 | 48 | 0.3538 |
584
+ | 0.1769 | 49 | 0.3572 |
585
+ | 0.1805 | 50 | 0.3547 |
586
+ | 0.1841 | 51 | 0.3179 |
587
+ | 0.1877 | 52 | 0.3436 |
588
+ | 0.1913 | 53 | 0.3502 |
589
+ | 0.1949 | 54 | 0.3381 |
590
+ | 0.1986 | 55 | 0.3547 |
591
+ | 0.2022 | 56 | 0.3362 |
592
+ | 0.2058 | 57 | 0.3407 |
593
+ | 0.2094 | 58 | 0.31 |
594
+ | 0.2130 | 59 | 0.3039 |
595
+ | 0.2166 | 60 | 0.3362 |
596
+ | 0.2202 | 61 | 0.2948 |
597
+ | 0.2238 | 62 | 0.3429 |
598
+ | 0.2274 | 63 | 0.3096 |
599
+ | 0.2310 | 64 | 0.35 |
600
+ | 0.2347 | 65 | 0.2997 |
601
+ | 0.2383 | 66 | 0.3258 |
602
+ | 0.2419 | 67 | 0.3376 |
603
+ | 0.2455 | 68 | 0.3213 |
604
+ | 0.2491 | 69 | 0.3185 |
605
+ | 0.2527 | 70 | 0.3282 |
606
+ | 0.2563 | 71 | 0.2988 |
607
+ | 0.2599 | 72 | 0.33 |
608
+ | 0.2635 | 73 | 0.3066 |
609
+ | 0.2671 | 74 | 0.3303 |
610
+ | 0.2708 | 75 | 0.3067 |
611
+ | 0.2744 | 76 | 0.2996 |
612
+ | 0.2780 | 77 | 0.3063 |
613
+ | 0.2816 | 78 | 0.3235 |
614
+ | 0.2852 | 79 | 0.2902 |
615
+ | 0.2888 | 80 | 0.302 |
616
+ | 0.2924 | 81 | 0.3223 |
617
+ | 0.2960 | 82 | 0.297 |
618
+ | 0.2996 | 83 | 0.2936 |
619
+ | 0.3032 | 84 | 0.3279 |
620
+ | 0.3069 | 85 | 0.2973 |
621
+ | 0.3105 | 86 | 0.2881 |
622
+ | 0.3141 | 87 | 0.3014 |
623
+ | 0.3177 | 88 | 0.2986 |
624
+ | 0.3213 | 89 | 0.3057 |
625
+ | 0.3249 | 90 | 0.2887 |
626
+ | 0.3285 | 91 | 0.2765 |
627
+ | 0.3321 | 92 | 0.2818 |
628
+ | 0.3357 | 93 | 0.2904 |
629
+ | 0.3394 | 94 | 0.267 |
630
+ | 0.3430 | 95 | 0.2948 |
631
+ | 0.3466 | 96 | 0.2766 |
632
+ | 0.3502 | 97 | 0.2782 |
633
+ | 0.3538 | 98 | 0.3082 |
634
+ | 0.3574 | 99 | 0.2697 |
635
+ | 0.3610 | 100 | 0.3006 |
636
+ | 0.3646 | 101 | 0.2986 |
637
+ | 0.3682 | 102 | 0.2789 |
638
+ | 0.3718 | 103 | 0.2756 |
639
+ | 0.3755 | 104 | 0.2884 |
640
+ | 0.3791 | 105 | 0.273 |
641
+ | 0.3827 | 106 | 0.2687 |
642
+ | 0.3863 | 107 | 0.2808 |
643
+ | 0.3899 | 108 | 0.2763 |
644
+ | 0.3935 | 109 | 0.2738 |
645
+ | 0.3971 | 110 | 0.2642 |
646
+ | 0.4007 | 111 | 0.2612 |
647
+ | 0.4043 | 112 | 0.2859 |
648
+ | 0.4079 | 113 | 0.2558 |
649
+ | 0.4116 | 114 | 0.2565 |
650
+ | 0.4152 | 115 | 0.2747 |
651
+ | 0.4188 | 116 | 0.2684 |
652
+ | 0.4224 | 117 | 0.2643 |
653
+ | 0.4260 | 118 | 0.241 |
654
+ | 0.4296 | 119 | 0.2563 |
655
+ | 0.4332 | 120 | 0.2754 |
656
+ | 0.4368 | 121 | 0.2503 |
657
+ | 0.4404 | 122 | 0.2544 |
658
+ | 0.4440 | 123 | 0.2729 |
659
+ | 0.4477 | 124 | 0.2589 |
660
+ | 0.4513 | 125 | 0.2626 |
661
+ | 0.4549 | 126 | 0.2693 |
662
+ | 0.4585 | 127 | 0.2687 |
663
+ | 0.4621 | 128 | 0.2903 |
664
+ | 0.4657 | 129 | 0.2663 |
665
+ | 0.4693 | 130 | 0.2604 |
666
+ | 0.4729 | 131 | 0.2601 |
667
+ | 0.4765 | 132 | 0.2649 |
668
+ | 0.4801 | 133 | 0.2597 |
669
+ | 0.4838 | 134 | 0.2608 |
670
+ | 0.4874 | 135 | 0.245 |
671
+ | 0.4910 | 136 | 0.2587 |
672
+ | 0.4946 | 137 | 0.2618 |
673
+ | 0.4982 | 138 | 0.2599 |
674
+ | 0.5018 | 139 | 0.265 |
675
+ | 0.5054 | 140 | 0.2427 |
676
+ | 0.5090 | 141 | 0.2448 |
677
+ | 0.5126 | 142 | 0.2608 |
678
+ | 0.5162 | 143 | 0.2188 |
679
+ | 0.5199 | 144 | 0.2471 |
680
+ | 0.5235 | 145 | 0.2604 |
681
+ | 0.5271 | 146 | 0.2571 |
682
+ | 0.5307 | 147 | 0.2684 |
683
+ | 0.5343 | 148 | 0.2319 |
684
+ | 0.5379 | 149 | 0.2572 |
685
+ | 0.5415 | 150 | 0.2243 |
686
+ | 0.5451 | 151 | 0.2562 |
687
+ | 0.5487 | 152 | 0.2457 |
688
+ | 0.5523 | 153 | 0.255 |
689
+ | 0.5560 | 154 | 0.2664 |
690
+ | 0.5596 | 155 | 0.24 |
691
+ | 0.5632 | 156 | 0.2612 |
692
+ | 0.5668 | 157 | 0.243 |
693
+ | 0.5704 | 158 | 0.2345 |
694
+ | 0.5740 | 159 | 0.2359 |
695
+ | 0.5776 | 160 | 0.2384 |
696
+ | 0.5812 | 161 | 0.2541 |
697
+ | 0.5848 | 162 | 0.2496 |
698
+ | 0.5884 | 163 | 0.2429 |
699
+ | 0.5921 | 164 | 0.2411 |
700
+ | 0.5957 | 165 | 0.2261 |
701
+ | 0.5993 | 166 | 0.2164 |
702
+ | 0.6029 | 167 | 0.2251 |
703
+ | 0.6065 | 168 | 0.2417 |
704
+ | 0.6101 | 169 | 0.2494 |
705
+ | 0.6137 | 170 | 0.2359 |
706
+ | 0.6173 | 171 | 0.2489 |
707
+ | 0.6209 | 172 | 0.2261 |
708
+ | 0.6245 | 173 | 0.2367 |
709
+ | 0.6282 | 174 | 0.2355 |
710
+ | 0.6318 | 175 | 0.2423 |
711
+ | 0.6354 | 176 | 0.2454 |
712
+ | 0.6390 | 177 | 0.2438 |
713
+ | 0.6426 | 178 | 0.2415 |
714
+ | 0.6462 | 179 | 0.2237 |
715
+ | 0.6498 | 180 | 0.2419 |
716
+ | 0.6534 | 181 | 0.2373 |
717
+ | 0.6570 | 182 | 0.2659 |
718
+ | 0.6606 | 183 | 0.2201 |
719
+ | 0.6643 | 184 | 0.2342 |
720
+ | 0.6679 | 185 | 0.2149 |
721
+ | 0.6715 | 186 | 0.2241 |
722
+ | 0.6751 | 187 | 0.2443 |
723
+ | 0.6787 | 188 | 0.2489 |
724
+ | 0.6823 | 189 | 0.2354 |
725
+ | 0.6859 | 190 | 0.2483 |
726
+ | 0.6895 | 191 | 0.2193 |
727
+ | 0.6931 | 192 | 0.229 |
728
+ | 0.6968 | 193 | 0.2335 |
729
+ | 0.7004 | 194 | 0.2484 |
730
+ | 0.7040 | 195 | 0.2317 |
731
+ | 0.7076 | 196 | 0.2203 |
732
+ | 0.7112 | 197 | 0.2329 |
733
+ | 0.7148 | 198 | 0.2084 |
734
+ | 0.7184 | 199 | 0.2341 |
735
+ | 0.7220 | 200 | 0.2369 |
736
+ | 0.7256 | 201 | 0.2364 |
737
+ | 0.7292 | 202 | 0.2276 |
738
+ | 0.7329 | 203 | 0.215 |
739
+ | 0.7365 | 204 | 0.2486 |
740
+ | 0.7401 | 205 | 0.2237 |
741
+ | 0.7437 | 206 | 0.218 |
742
+ | 0.7473 | 207 | 0.2444 |
743
+ | 0.7509 | 208 | 0.2276 |
744
+ | 0.7545 | 209 | 0.2127 |
745
+ | 0.7581 | 210 | 0.2283 |
746
+ | 0.7617 | 211 | 0.2234 |
747
+ | 0.7653 | 212 | 0.207 |
748
+ | 0.7690 | 213 | 0.24 |
749
+ | 0.7726 | 214 | 0.2317 |
750
+ | 0.7762 | 215 | 0.2056 |
751
+ | 0.7798 | 216 | 0.2149 |
752
+ | 0.7834 | 217 | 0.2211 |
753
+ | 0.7870 | 218 | 0.2232 |
754
+ | 0.7906 | 219 | 0.2222 |
755
+ | 0.7942 | 220 | 0.2481 |
756
+ | 0.7978 | 221 | 0.227 |
757
+ | 0.8014 | 222 | 0.2305 |
758
+ | 0.8051 | 223 | 0.2091 |
759
+ | 0.8087 | 224 | 0.2278 |
760
+ | 0.8123 | 225 | 0.2123 |
761
+ | 0.8159 | 226 | 0.2233 |
762
+ | 0.8195 | 227 | 0.2365 |
763
+ | 0.8231 | 228 | 0.2165 |
764
+ | 0.8267 | 229 | 0.2192 |
765
+ | 0.8303 | 230 | 0.2145 |
766
+ | 0.8339 | 231 | 0.2382 |
767
+ | 0.8375 | 232 | 0.2232 |
768
+ | 0.8412 | 233 | 0.2273 |
769
+ | 0.8448 | 234 | 0.2296 |
770
+ | 0.8484 | 235 | 0.2229 |
771
+ | 0.8520 | 236 | 0.2213 |
772
+ | 0.8556 | 237 | 0.2343 |
773
+ | 0.8592 | 238 | 0.2208 |
774
+ | 0.8628 | 239 | 0.2315 |
775
+ | 0.8664 | 240 | 0.2137 |
776
+ | 0.8700 | 241 | 0.2201 |
777
+ | 0.8736 | 242 | 0.2185 |
778
+ | 0.8773 | 243 | 0.2337 |
779
+ | 0.8809 | 244 | 0.2153 |
780
+ | 0.8845 | 245 | 0.2369 |
781
+ | 0.8881 | 246 | 0.2216 |
782
+ | 0.8917 | 247 | 0.2338 |
783
+ | 0.8953 | 248 | 0.2241 |
784
+ | 0.8989 | 249 | 0.213 |
785
+ | 0.9025 | 250 | 0.2245 |
786
+ | 0.9061 | 251 | 0.2074 |
787
+ | 0.9097 | 252 | 0.2283 |
788
+ | 0.9134 | 253 | 0.2003 |
789
+ | 0.9170 | 254 | 0.2099 |
790
+ | 0.9206 | 255 | 0.2288 |
791
+ | 0.9242 | 256 | 0.2168 |
792
+ | 0.9278 | 257 | 0.215 |
793
+ | 0.9314 | 258 | 0.2146 |
794
+ | 0.9350 | 259 | 0.2126 |
795
+ | 0.9386 | 260 | 0.2178 |
796
+ | 0.9422 | 261 | 0.2065 |
797
+ | 0.9458 | 262 | 0.2327 |
798
+ | 0.9495 | 263 | 0.2116 |
799
+ | 0.9531 | 264 | 0.2324 |
800
+ | 0.9567 | 265 | 0.2235 |
801
+ | 0.9603 | 266 | 0.2189 |
802
+ | 0.9639 | 267 | 0.2175 |
803
+ | 0.9675 | 268 | 0.2171 |
804
+ | 0.9711 | 269 | 0.1925 |
805
+ | 0.9747 | 270 | 0.225 |
806
+ | 0.9783 | 271 | 0.2149 |
807
+ | 0.9819 | 272 | 0.204 |
808
+ | 0.9856 | 273 | 0.2004 |
809
+ | 0.9892 | 274 | 0.2055 |
810
+ | 0.9928 | 275 | 0.2045 |
811
+ | 0.9964 | 276 | 0.2186 |
812
+ | 1.0 | 277 | 0.2215 |
813
+ | 1.0036 | 278 | 0.1545 |
814
+ | 1.0072 | 279 | 0.169 |
815
+ | 1.0108 | 280 | 0.152 |
816
+ | 1.0144 | 281 | 0.1597 |
817
+ | 1.0181 | 282 | 0.1626 |
818
+ | 1.0217 | 283 | 0.1692 |
819
+ | 1.0253 | 284 | 0.1639 |
820
+ | 1.0289 | 285 | 0.1638 |
821
+ | 1.0325 | 286 | 0.1507 |
822
+ | 1.0361 | 287 | 0.1594 |
823
+ | 1.0397 | 288 | 0.1621 |
824
+ | 1.0433 | 289 | 0.1565 |
825
+ | 1.0469 | 290 | 0.1549 |
826
+ | 1.0505 | 291 | 0.1731 |
827
+ | 1.0542 | 292 | 0.152 |
828
+ | 1.0578 | 293 | 0.1586 |
829
+ | 1.0614 | 294 | 0.1593 |
830
+ | 1.0650 | 295 | 0.1406 |
831
+ | 1.0686 | 296 | 0.1524 |
832
+ | 1.0722 | 297 | 0.1474 |
833
+ | 1.0758 | 298 | 0.158 |
834
+ | 1.0794 | 299 | 0.1743 |
835
+ | 1.0830 | 300 | 0.1485 |
836
+ | 1.0866 | 301 | 0.1648 |
837
+ | 1.0903 | 302 | 0.1337 |
838
+ | 1.0939 | 303 | 0.1554 |
839
+ | 1.0975 | 304 | 0.1434 |
840
+ | 1.1011 | 305 | 0.1642 |
841
+ | 1.1047 | 306 | 0.159 |
842
+ | 1.1083 | 307 | 0.1658 |
843
+ | 1.1119 | 308 | 0.1554 |
844
+ | 1.1155 | 309 | 0.1425 |
845
+ | 1.1191 | 310 | 0.1432 |
846
+ | 1.1227 | 311 | 0.1517 |
847
+ | 1.1264 | 312 | 0.148 |
848
+ | 1.1300 | 313 | 0.1636 |
849
+ | 1.1336 | 314 | 0.1735 |
850
+ | 1.1372 | 315 | 0.151 |
851
+ | 1.1408 | 316 | 0.1423 |
852
+ | 1.1444 | 317 | 0.1501 |
853
+ | 1.1480 | 318 | 0.1537 |
854
+ | 1.1516 | 319 | 0.1554 |
855
+ | 1.1552 | 320 | 0.1553 |
856
+ | 1.1588 | 321 | 0.149 |
857
+ | 1.1625 | 322 | 0.1605 |
858
+ | 1.1661 | 323 | 0.1551 |
859
+ | 1.1697 | 324 | 0.1555 |
860
+ | 1.1733 | 325 | 0.1443 |
861
+ | 1.1769 | 326 | 0.1533 |
862
+ | 1.1805 | 327 | 0.1658 |
863
+ | 1.1841 | 328 | 0.15 |
864
+ | 1.1877 | 329 | 0.1626 |
865
+ | 1.1913 | 330 | 0.172 |
866
+ | 1.1949 | 331 | 0.1542 |
867
+ | 1.1986 | 332 | 0.166 |
868
+ | 1.2022 | 333 | 0.1513 |
869
+ | 1.2058 | 334 | 0.1612 |
870
+ | 1.2094 | 335 | 0.1521 |
871
+ | 1.2130 | 336 | 0.1552 |
872
+ | 1.2166 | 337 | 0.1503 |
873
+ | 1.2202 | 338 | 0.1613 |
874
+ | 1.2238 | 339 | 0.1563 |
875
+ | 1.2274 | 340 | 0.1429 |
876
+ | 1.2310 | 341 | 0.1587 |
877
+ | 1.2347 | 342 | 0.1477 |
878
+ | 1.2383 | 343 | 0.1561 |
879
+ | 1.2419 | 344 | 0.1418 |
880
+ | 1.2455 | 345 | 0.1495 |
881
+ | 1.2491 | 346 | 0.1533 |
882
+ | 1.2527 | 347 | 0.1521 |
883
+ | 1.2563 | 348 | 0.1422 |
884
+ | 1.2599 | 349 | 0.1446 |
885
+ | 1.2635 | 350 | 0.146 |
886
+ | 1.2671 | 351 | 0.1473 |
887
+ | 1.2708 | 352 | 0.1566 |
888
+ | 1.2744 | 353 | 0.1411 |
889
+ | 1.2780 | 354 | 0.1502 |
890
+ | 1.2816 | 355 | 0.1383 |
891
+ | 1.2852 | 356 | 0.1622 |
892
+ | 1.2888 | 357 | 0.1391 |
893
+ | 1.2924 | 358 | 0.1455 |
894
+ | 1.2960 | 359 | 0.1541 |
895
+ | 1.2996 | 360 | 0.1476 |
896
+ | 1.3032 | 361 | 0.1662 |
897
+ | 1.3069 | 362 | 0.1476 |
898
+ | 1.3105 | 363 | 0.1452 |
899
+ | 1.3141 | 364 | 0.1372 |
900
+ | 1.3177 | 365 | 0.1542 |
901
+ | 1.3213 | 366 | 0.1531 |
902
+ | 1.3249 | 367 | 0.1623 |
903
+ | 1.3285 | 368 | 0.1544 |
904
+ | 1.3321 | 369 | 0.1625 |
905
+ | 1.3357 | 370 | 0.1459 |
906
+ | 1.3394 | 371 | 0.1474 |
907
+ | 1.3430 | 372 | 0.1499 |
908
+ | 1.3466 | 373 | 0.1495 |
909
+ | 1.3502 | 374 | 0.1361 |
910
+ | 1.3538 | 375 | 0.1444 |
911
+ | 1.3574 | 376 | 0.1495 |
912
+ | 1.3610 | 377 | 0.1583 |
913
+ | 1.3646 | 378 | 0.1642 |
914
+ | 1.3682 | 379 | 0.1646 |
915
+ | 1.3718 | 380 | 0.1595 |
916
+ | 1.3755 | 381 | 0.149 |
917
+ | 1.3791 | 382 | 0.1448 |
918
+ | 1.3827 | 383 | 0.1603 |
919
+ | 1.3863 | 384 | 0.1269 |
920
+ | 1.3899 | 385 | 0.1491 |
921
+ | 1.3935 | 386 | 0.1367 |
922
+ | 1.3971 | 387 | 0.1501 |
923
+ | 1.4007 | 388 | 0.1414 |
924
+ | 1.4043 | 389 | 0.156 |
925
+ | 1.4079 | 390 | 0.1428 |
926
+ | 1.4116 | 391 | 0.1559 |
927
+ | 1.4152 | 392 | 0.1452 |
928
+ | 1.4188 | 393 | 0.1547 |
929
+ | 1.4224 | 394 | 0.1432 |
930
+ | 1.4260 | 395 | 0.1648 |
931
+ | 1.4296 | 396 | 0.166 |
932
+ | 1.4332 | 397 | 0.1485 |
933
+ | 1.4368 | 398 | 0.1494 |
934
+ | 1.4404 | 399 | 0.1635 |
935
+ | 1.4440 | 400 | 0.1498 |
936
+ | 1.4477 | 401 | 0.1509 |
937
+ | 1.4513 | 402 | 0.1431 |
938
+ | 1.4549 | 403 | 0.1547 |
939
+ | 1.4585 | 404 | 0.1576 |
940
+ | 1.4621 | 405 | 0.1426 |
941
+ | 1.4657 | 406 | 0.132 |
942
+ | 1.4693 | 407 | 0.1511 |
943
+ | 1.4729 | 408 | 0.1551 |
944
+ | 1.4765 | 409 | 0.16 |
945
+ | 1.4801 | 410 | 0.1507 |
946
+ | 1.4838 | 411 | 0.1591 |
947
+ | 1.4874 | 412 | 0.1536 |
948
+ | 1.4910 | 413 | 0.1507 |
949
+ | 1.4946 | 414 | 0.1564 |
950
+ | 1.4982 | 415 | 0.153 |
951
+ | 1.5018 | 416 | 0.1404 |
952
+ | 1.5054 | 417 | 0.1627 |
953
+ | 1.5090 | 418 | 0.1432 |
954
+ | 1.5126 | 419 | 0.1456 |
955
+ | 1.5162 | 420 | 0.1369 |
956
+ | 1.5199 | 421 | 0.1554 |
957
+ | 1.5235 | 422 | 0.1412 |
958
+ | 1.5271 | 423 | 0.1547 |
959
+ | 1.5307 | 424 | 0.1555 |
960
+ | 1.5343 | 425 | 0.1575 |
961
+ | 1.5379 | 426 | 0.1595 |
962
+ | 1.5415 | 427 | 0.1464 |
963
+ | 1.5451 | 428 | 0.1738 |
964
+ | 1.5487 | 429 | 0.1692 |
965
+ | 1.5523 | 430 | 0.1566 |
966
+ | 1.5560 | 431 | 0.1452 |
967
+ | 1.5596 | 432 | 0.1433 |
968
+ | 1.5632 | 433 | 0.1584 |
969
+ | 1.5668 | 434 | 0.1579 |
970
+ | 1.5704 | 435 | 0.157 |
971
+ | 1.5740 | 436 | 0.1533 |
972
+ | 1.5776 | 437 | 0.148 |
973
+ | 1.5812 | 438 | 0.1381 |
974
+ | 1.5848 | 439 | 0.1605 |
975
+ | 1.5884 | 440 | 0.163 |
976
+ | 1.5921 | 441 | 0.1492 |
977
+ | 1.5957 | 442 | 0.1601 |
978
+ | 1.5993 | 443 | 0.1456 |
979
+ | 1.6029 | 444 | 0.1439 |
980
+ | 1.6065 | 445 | 0.1553 |
981
+ | 1.6101 | 446 | 0.1371 |
982
+ | 1.6137 | 447 | 0.1382 |
983
+ | 1.6173 | 448 | 0.1458 |
984
+ | 1.6209 | 449 | 0.14 |
985
+ | 1.6245 | 450 | 0.1463 |
986
+ | 1.6282 | 451 | 0.1433 |
987
+ | 1.6318 | 452 | 0.1472 |
988
+ | 1.6354 | 453 | 0.1481 |
989
+ | 1.6390 | 454 | 0.1408 |
990
+ | 1.6426 | 455 | 0.1525 |
991
+ | 1.6462 | 456 | 0.1223 |
992
+ | 1.6498 | 457 | 0.1452 |
993
+ | 1.6534 | 458 | 0.159 |
994
+ | 1.6570 | 459 | 0.1389 |
995
+ | 1.6606 | 460 | 0.1479 |
996
+ | 1.6643 | 461 | 0.1451 |
997
+ | 1.6679 | 462 | 0.1651 |
998
+ | 1.6715 | 463 | 0.1336 |
999
+ | 1.6751 | 464 | 0.1496 |
1000
+ | 1.6787 | 465 | 0.1384 |
1001
+ | 1.6823 | 466 | 0.143 |
1002
+ | 1.6859 | 467 | 0.1423 |
1003
+ | 1.6895 | 468 | 0.1403 |
1004
+ | 1.6931 | 469 | 0.1577 |
1005
+ | 1.6968 | 470 | 0.1511 |
1006
+ | 1.7004 | 471 | 0.1429 |
1007
+ | 1.7040 | 472 | 0.1445 |
1008
+ | 1.7076 | 473 | 0.1431 |
1009
+ | 1.7112 | 474 | 0.1326 |
1010
+ | 1.7148 | 475 | 0.1554 |
1011
+ | 1.7184 | 476 | 0.1406 |
1012
+ | 1.7220 | 477 | 0.1479 |
1013
+ | 1.7256 | 478 | 0.1521 |
1014
+ | 1.7292 | 479 | 0.1475 |
1015
+ | 1.7329 | 480 | 0.1584 |
1016
+ | 1.7365 | 481 | 0.1393 |
1017
+ | 1.7401 | 482 | 0.1291 |
1018
+ | 1.7437 | 483 | 0.1373 |
1019
+ | 1.7473 | 484 | 0.1555 |
1020
+ | 1.7509 | 485 | 0.1473 |
1021
+ | 1.7545 | 486 | 0.1654 |
1022
+ | 1.7581 | 487 | 0.1568 |
1023
+ | 1.7617 | 488 | 0.1557 |
1024
+ | 1.7653 | 489 | 0.1531 |
1025
+ | 1.7690 | 490 | 0.1385 |
1026
+ | 1.7726 | 491 | 0.1381 |
1027
+ | 1.7762 | 492 | 0.1375 |
1028
+ | 1.7798 | 493 | 0.1472 |
1029
+ | 1.7834 | 494 | 0.1581 |
1030
+ | 1.7870 | 495 | 0.1448 |
1031
+ | 1.7906 | 496 | 0.1443 |
1032
+ | 1.7942 | 497 | 0.1422 |
1033
+ | 1.7978 | 498 | 0.1295 |
1034
+ | 1.8014 | 499 | 0.1463 |
1035
+ | 1.8051 | 500 | 0.1346 |
1036
+ | 1.8087 | 501 | 0.1387 |
1037
+ | 1.8123 | 502 | 0.1463 |
1038
+ | 1.8159 | 503 | 0.1439 |
1039
+ | 1.8195 | 504 | 0.1404 |
1040
+ | 1.8231 | 505 | 0.1433 |
1041
+ | 1.8267 | 506 | 0.136 |
1042
+ | 1.8303 | 507 | 0.14 |
1043
+ | 1.8339 | 508 | 0.1355 |
1044
+ | 1.8375 | 509 | 0.1446 |
1045
+ | 1.8412 | 510 | 0.1564 |
1046
+ | 1.8448 | 511 | 0.1413 |
1047
+ | 1.8484 | 512 | 0.1451 |
1048
+ | 1.8520 | 513 | 0.1453 |
1049
+ | 1.8556 | 514 | 0.1484 |
1050
+ | 1.8592 | 515 | 0.1403 |
1051
+ | 1.8628 | 516 | 0.1568 |
1052
+ | 1.8664 | 517 | 0.1566 |
1053
+ | 1.8700 | 518 | 0.1318 |
1054
+ | 1.8736 | 519 | 0.1483 |
1055
+ | 1.8773 | 520 | 0.1339 |
1056
+ | 1.8809 | 521 | 0.1423 |
1057
+ | 1.8845 | 522 | 0.1349 |
1058
+ | 1.8881 | 523 | 0.1302 |
1059
+ | 1.8917 | 524 | 0.1341 |
1060
+ | 1.8953 | 525 | 0.1456 |
1061
+ | 1.8989 | 526 | 0.1334 |
1062
+ | 1.9025 | 527 | 0.1382 |
1063
+ | 1.9061 | 528 | 0.1462 |
1064
+ | 1.9097 | 529 | 0.1315 |
1065
+ | 1.9134 | 530 | 0.1606 |
1066
+ | 1.9170 | 531 | 0.1308 |
1067
+ | 1.9206 | 532 | 0.1319 |
1068
+ | 1.9242 | 533 | 0.1407 |
1069
+ | 1.9278 | 534 | 0.1385 |
1070
+ | 1.9314 | 535 | 0.1471 |
1071
+ | 1.9350 | 536 | 0.1621 |
1072
+ | 1.9386 | 537 | 0.1436 |
1073
+ | 1.9422 | 538 | 0.151 |
1074
+ | 1.9458 | 539 | 0.1423 |
1075
+ | 1.9495 | 540 | 0.1411 |
1076
+ | 1.9531 | 541 | 0.1535 |
1077
+ | 1.9567 | 542 | 0.143 |
1078
+ | 1.9603 | 543 | 0.149 |
1079
+ | 1.9639 | 544 | 0.1384 |
1080
+ | 1.9675 | 545 | 0.1479 |
1081
+ | 1.9711 | 546 | 0.1452 |
1082
+ | 1.9747 | 547 | 0.1372 |
1083
+ | 1.9783 | 548 | 0.1418 |
1084
+ | 1.9819 | 549 | 0.1443 |
1085
+ | 1.9856 | 550 | 0.1344 |
1086
+ | 1.9892 | 551 | 0.1278 |
1087
+ | 1.9928 | 552 | 0.1447 |
1088
+ | 1.9964 | 553 | 0.1366 |
1089
+ | 2.0 | 554 | 0.141 |
1090
+ | 2.0036 | 555 | 0.1161 |
1091
+ | 2.0072 | 556 | 0.1099 |
1092
+ | 2.0108 | 557 | 0.126 |
1093
+ | 2.0144 | 558 | 0.1163 |
1094
+ | 2.0181 | 559 | 0.1234 |
1095
+ | 2.0217 | 560 | 0.1171 |
1096
+ | 2.0253 | 561 | 0.1073 |
1097
+ | 2.0289 | 562 | 0.1126 |
1098
+ | 2.0325 | 563 | 0.1175 |
1099
+ | 2.0361 | 564 | 0.1086 |
1100
+ | 2.0397 | 565 | 0.1038 |
1101
+ | 2.0433 | 566 | 0.1121 |
1102
+ | 2.0469 | 567 | 0.1154 |
1103
+ | 2.0505 | 568 | 0.0973 |
1104
+ | 2.0542 | 569 | 0.1208 |
1105
+ | 2.0578 | 570 | 0.1064 |
1106
+ | 2.0614 | 571 | 0.1159 |
1107
+ | 2.0650 | 572 | 0.1093 |
1108
+ | 2.0686 | 573 | 0.113 |
1109
+ | 2.0722 | 574 | 0.1033 |
1110
+ | 2.0758 | 575 | 0.1152 |
1111
+ | 2.0794 | 576 | 0.1029 |
1112
+ | 2.0830 | 577 | 0.1204 |
1113
+ | 2.0866 | 578 | 0.1079 |
1114
+ | 2.0903 | 579 | 0.1288 |
1115
+ | 2.0939 | 580 | 0.0998 |
1116
+ | 2.0975 | 581 | 0.1058 |
1117
+ | 2.1011 | 582 | 0.1235 |
1118
+ | 2.1047 | 583 | 0.1059 |
1119
+ | 2.1083 | 584 | 0.0998 |
1120
+ | 2.1119 | 585 | 0.1142 |
1121
+ | 2.1155 | 586 | 0.1082 |
1122
+ | 2.1191 | 587 | 0.0973 |
1123
+ | 2.1227 | 588 | 0.1017 |
1124
+ | 2.1264 | 589 | 0.1045 |
1125
+ | 2.1300 | 590 | 0.123 |
1126
+ | 2.1336 | 591 | 0.1065 |
1127
+ | 2.1372 | 592 | 0.1135 |
1128
+ | 2.1408 | 593 | 0.1027 |
1129
+ | 2.1444 | 594 | 0.1166 |
1130
+ | 2.1480 | 595 | 0.1082 |
1131
+ | 2.1516 | 596 | 0.1113 |
1132
+ | 2.1552 | 597 | 0.1108 |
1133
+ | 2.1588 | 598 | 0.114 |
1134
+ | 2.1625 | 599 | 0.1064 |
1135
+ | 2.1661 | 600 | 0.0955 |
1136
+ | 2.1697 | 601 | 0.113 |
1137
+ | 2.1733 | 602 | 0.1136 |
1138
+ | 2.1769 | 603 | 0.1125 |
1139
+ | 2.1805 | 604 | 0.1146 |
1140
+ | 2.1841 | 605 | 0.1054 |
1141
+ | 2.1877 | 606 | 0.1144 |
1142
+ | 2.1913 | 607 | 0.1038 |
1143
+ | 2.1949 | 608 | 0.1113 |
1144
+ | 2.1986 | 609 | 0.1187 |
1145
+ | 2.2022 | 610 | 0.1166 |
1146
+ | 2.2058 | 611 | 0.1035 |
1147
+ | 2.2094 | 612 | 0.1054 |
1148
+ | 2.2130 | 613 | 0.118 |
1149
+ | 2.2166 | 614 | 0.125 |
1150
+ | 2.2202 | 615 | 0.1142 |
1151
+ | 2.2238 | 616 | 0.1119 |
1152
+ | 2.2274 | 617 | 0.1173 |
1153
+ | 2.2310 | 618 | 0.1024 |
1154
+ | 2.2347 | 619 | 0.105 |
1155
+ | 2.2383 | 620 | 0.1025 |
1156
+ | 2.2419 | 621 | 0.1022 |
1157
+ | 2.2455 | 622 | 0.0995 |
1158
+ | 2.2491 | 623 | 0.1022 |
1159
+ | 2.2527 | 624 | 0.1198 |
1160
+ | 2.2563 | 625 | 0.0995 |
1161
+ | 2.2599 | 626 | 0.1162 |
1162
+ | 2.2635 | 627 | 0.1172 |
1163
+ | 2.2671 | 628 | 0.1037 |
1164
+ | 2.2708 | 629 | 0.1093 |
1165
+ | 2.2744 | 630 | 0.1018 |
1166
+ | 2.2780 | 631 | 0.1168 |
1167
+ | 2.2816 | 632 | 0.1015 |
1168
+ | 2.2852 | 633 | 0.101 |
1169
+ | 2.2888 | 634 | 0.1064 |
1170
+ | 2.2924 | 635 | 0.1185 |
1171
+ | 2.2960 | 636 | 0.1055 |
1172
+ | 2.2996 | 637 | 0.1142 |
1173
+ | 2.3032 | 638 | 0.0966 |
1174
+ | 2.3069 | 639 | 0.1039 |
1175
+ | 2.3105 | 640 | 0.1139 |
1176
+ | 2.3141 | 641 | 0.1181 |
1177
+ | 2.3177 | 642 | 0.1168 |
1178
+ | 2.3213 | 643 | 0.1201 |
1179
+ | 2.3249 | 644 | 0.0984 |
1180
+ | 2.3285 | 645 | 0.1068 |
1181
+ | 2.3321 | 646 | 0.1007 |
1182
+ | 2.3357 | 647 | 0.1179 |
1183
+ | 2.3394 | 648 | 0.1043 |
1184
+ | 2.3430 | 649 | 0.1213 |
1185
+ | 2.3466 | 650 | 0.1027 |
1186
+ | 2.3502 | 651 | 0.1119 |
1187
+ | 2.3538 | 652 | 0.1077 |
1188
+ | 2.3574 | 653 | 0.1061 |
1189
+ | 2.3610 | 654 | 0.1054 |
1190
+ | 2.3646 | 655 | 0.1135 |
1191
+ | 2.3682 | 656 | 0.1136 |
1192
+ | 2.3718 | 657 | 0.1062 |
1193
+ | 2.3755 | 658 | 0.1105 |
1194
+ | 2.3791 | 659 | 0.1157 |
1195
+ | 2.3827 | 660 | 0.1036 |
1196
+ | 2.3863 | 661 | 0.1098 |
1197
+ | 2.3899 | 662 | 0.1195 |
1198
+ | 2.3935 | 663 | 0.1151 |
1199
+ | 2.3971 | 664 | 0.1116 |
1200
+ | 2.4007 | 665 | 0.1086 |
1201
+ | 2.4043 | 666 | 0.1151 |
1202
+ | 2.4079 | 667 | 0.1156 |
1203
+ | 2.4116 | 668 | 0.116 |
1204
+ | 2.4152 | 669 | 0.1055 |
1205
+ | 2.4188 | 670 | 0.1051 |
1206
+ | 2.4224 | 671 | 0.0952 |
1207
+ | 2.4260 | 672 | 0.1012 |
1208
+
1209
+ </details>
1210
+
1211
+ ### Framework Versions
1212
+ - Python: 3.10.12
1213
+ - Sentence Transformers: 3.4.1
1214
+ - Transformers: 4.49.0
1215
+ - PyTorch: 2.5.1+cu124
1216
+ - Accelerate: 1.4.0
1217
+ - Datasets: 3.3.2
1218
+ - Tokenizers: 0.21.0
1219
+
1220
+ ## Citation
1221
+
1222
+ ### BibTeX
1223
+
1224
+ #### Sentence Transformers
1225
+ ```bibtex
1226
+ @inproceedings{reimers-2019-sentence-bert,
1227
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
1228
+ author = "Reimers, Nils and Gurevych, Iryna",
1229
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
1230
+ month = "11",
1231
+ year = "2019",
1232
+ publisher = "Association for Computational Linguistics",
1233
+ url = "https://arxiv.org/abs/1908.10084",
1234
+ }
1235
+ ```
1236
+
1237
+ <!--
1238
+ ## Glossary
1239
+
1240
+ *Clearly define terms in order to be accessible across audiences.*
1241
+ -->
1242
+
1243
+ <!--
1244
+ ## Model Card Authors
1245
+
1246
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
1247
+ -->
1248
+
1249
+ <!--
1250
+ ## Model Card Contact
1251
+
1252
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
1253
+ -->
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "output/checkpoint-672",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 8194,
17
+ "model_type": "xlm-roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
+ "output_past": true,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.49.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 250002
28
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.4.1",
4
+ "transformers": "4.49.0",
5
+ "pytorch": "2.5.1+cu124"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54f537ab0ceb96cc972b2c2982bba5a67632bc4c89d01b4c63496854f616698c
3
+ size 2271064456
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 1024,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e3b8957de04e3a4ed42b1a11381556f9adad8d0d502b9dd071c75f626b28f40
3
+ size 17083053
tokenizer_config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "extra_special_tokens": {},
49
+ "mask_token": "<mask>",
50
+ "max_length": 1024,
51
+ "model_max_length": 1024,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "<pad>",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "</s>",
57
+ "sp_model_kwargs": {},
58
+ "stride": 0,
59
+ "tokenizer_class": "XLMRobertaTokenizerFast",
60
+ "truncation_side": "right",
61
+ "truncation_strategy": "longest_first",
62
+ "unk_token": "<unk>"
63
+ }