iqrakiran commited on
Commit
5f50c1d
1 Parent(s): b6e4c87

Add new SentenceTransformer model.

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,432 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilbert/distilroberta-base
3
+ datasets: []
4
+ language: []
5
+ library_name: sentence-transformers
6
+ pipeline_tag: sentence-similarity
7
+ tags:
8
+ - sentence-transformers
9
+ - sentence-similarity
10
+ - feature-extraction
11
+ - generated_from_trainer
12
+ - dataset_size:4093
13
+ - loss:TripletLoss
14
+ widget:
15
+ - source_sentence: I am angry about an incident at the supermarket, and it has been
16
+ bothering me for the last three or four days. The person involved was flirting
17
+ with me, and I think they should apologize. I didn't ask for an apology or indicate
18
+ that I was upset during the conversation because I felt it would be inappropriate.
19
+ I believe they should call me and set aside an appropriate time to apologize.
20
+ sentences:
21
+ - If you are a counsellor, please answer the questions based on the description
22
+ of the patient.
23
+ - It seems like this incident really impacted you and you believe that an apology
24
+ is necessary for closure. It's understandable that you didn't address it on the
25
+ spot if you felt it was inappropriate. However, people sometimes don't realize
26
+ they've upset someone unless it's brought to their attention. Since you feel so
27
+ strongly about this, have you considered reaching out to the person to express
28
+ your feelings and provide them an opportunity to apologize?
29
+ - It sounds like you have built a strong support system in San Diego with meaningful
30
+ relationships and emotional investments. It's natural to feel apprehensive about
31
+ moving away from that. Maybe you could consider the potential opportunities that
32
+ LA has to offer and weigh them against what you would be leaving behind in San
33
+ Diego. Additionally, think about the ways you can stay connected with your loved
34
+ ones after moving. Just remember that it's a difficult decision, and it's important
35
+ to give yourself time to make the best choice for you.
36
+ - source_sentence: Lately, I've been feeling a bit overwhelmed with everything going
37
+ on in my life. I've been hearing things that aren't really there, like whispers
38
+ and someone calling my name, but it's not happening all the time, just occasionally.
39
+ I also find myself getting a bit disorganized, but I wouldn't say I'm completely
40
+ disorganized. Sometimes, I just feel disconnected with what is happening around
41
+ me. I've been like this for a couple of weeks now. My parents said that I've been
42
+ acting differently, but I don't know, maybe I'm just overreacting? Can you help
43
+ me with this?
44
+ sentences:
45
+ - If you are a counsellor, please answer the questions based on the description
46
+ of the patient.
47
+ - Thank you for sharing your experience with me. I can understand how hearing things
48
+ and feeling disorganized might be confusing and concerning for you. It's important
49
+ to discuss these symptoms and feelings in order to provide proper support. Can
50
+ you tell me if you've experienced any changes in your mood, daily activities,
51
+ or sleep habits during this time? Also, have you noticed any patterns or triggers
52
+ for when you hear the whispers or feel disconnected from what's happening around
53
+ you?
54
+ - I appreciate you sharing your feelings and experiences with me. It sounds like
55
+ you're going through a challenging period in your life, dealing with stress and
56
+ worry. It's important to understand the factors contributing to your stress and
57
+ how they might be influencing your thoughts and emotions. Could you provide some
58
+ more specific examples of the situations that have caused you to feel stressed
59
+ and worried? Additionally, have you noticed any patterns or triggers that seem
60
+ to make your worries more intense?
61
+ - source_sentence: I've been experiencing anxiety due to the difficult transition
62
+ from being at home, taking care of my son, to moving back into the work world.
63
+ I've been trying to manage this anxiety by paying attention to nutrition, exercise,
64
+ and practicing gratitude. Although I feel like I'm managing better now than when
65
+ I first started, I'm still struggling.
66
+ sentences:
67
+ - It's great to hear that you've made some progress in managing your anxiety through
68
+ healthy practices like focusing on nutrition, exercise, and gratitude. It's common
69
+ for individuals to face challenges while transitioning between roles, and it might
70
+ take more time for you to adjust fully. If you still find yourself struggling,
71
+ consider seeking additional support through therapy or support groups. Is there
72
+ a specific aspect of this transition that continues to cause you worry?
73
+ - If you are a counsellor, please answer the questions based on the description
74
+ of the patient.
75
+ - It's great to hear that you have a strong drive for self-improvement and setting
76
+ ambitious goals for yourself. Acknowledging this inner motivation is an essential
77
+ step towards achieving what you desire. Remember, it's important to embrace the
78
+ process of learning and growing, and not to put too much pressure on yourself
79
+ to reach an end goal quickly. Take it one step at a time, and don't forget to
80
+ celebrate your milestones along the way.
81
+ - source_sentence: Recently, I have been experiencing difficulties while learning.
82
+ Sometimes it's quite challenging, and I feel upset with myself to the point of
83
+ crying. However, at other times, I find it easy and feel a sense of accomplishment
84
+ when I finally get things right.
85
+ sentences:
86
+ - If you are a counsellor, please answer the questions based on the description
87
+ of the patient.
88
+ - Thank you for sharing that with me. I'm glad you're feeling better now that you're
89
+ here. Sometimes waiting can cause anxiety or stress, so it's important to find
90
+ ways to manage these feelings. If you'd like, we can discuss some coping strategies
91
+ for dealing with stress while waiting or in similar situations.
92
+ - It seems like you've been having ups and downs with your learning experience.
93
+ It's great that you sometimes feel accomplished when you finally understand something.
94
+ When you're feeling upset or frustrated, it might be helpful to take a break and
95
+ come back to it later, as you've mentioned. This can give you a fresh perspective
96
+ and may help reduce stress. It's also important to remember that learning is a
97
+ process, and it's okay to experience difficulties along the way.
98
+ - source_sentence: I remember generating alternative predictions. We said that I might
99
+ get nervous but if I stay focused on what I want to do, I can probably do it.
100
+ It's just hard to believe when I'm in this situation.
101
+ sentences:
102
+ - It seems like you've been dealing with unexpected expenses and some impulsive
103
+ spending habits. By planning for a stricter budget and understanding the difference
104
+ between essential and non-essential expenses, you can start to develop healthy
105
+ financial habits. Have you considered creating a budget plan and setting aside
106
+ a specific amount for unexpected expenses to help manage your spending?
107
+ - It's normal to feel nervous in difficult situations. As a counsellor, I suggest
108
+ continuing to test these alternative predictions directly in real-life experiences.
109
+ As you encounter situations where you manage to handle things even while nervous,
110
+ your belief in your abilities and the predictions will strengthen. Remember to
111
+ stay focused and trust yourself.
112
+ - If you are a counsellor, please answer the questions based on the description
113
+ of the patient.
114
+ ---
115
+
116
+ # SentenceTransformer based on distilbert/distilroberta-base
117
+
118
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
119
+
120
+ ## Model Details
121
+
122
+ ### Model Description
123
+ - **Model Type:** Sentence Transformer
124
+ - **Base model:** [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base) <!-- at revision fb53ab8802853c8e4fbdbcd0529f21fc6f459b2b -->
125
+ - **Maximum Sequence Length:** 512 tokens
126
+ - **Output Dimensionality:** 768 tokens
127
+ - **Similarity Function:** Cosine Similarity
128
+ <!-- - **Training Dataset:** Unknown -->
129
+ <!-- - **Language:** Unknown -->
130
+ <!-- - **License:** Unknown -->
131
+
132
+ ### Model Sources
133
+
134
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
135
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
136
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
137
+
138
+ ### Full Model Architecture
139
+
140
+ ```
141
+ SentenceTransformer(
142
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
143
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
144
+ )
145
+ ```
146
+
147
+ ## Usage
148
+
149
+ ### Direct Usage (Sentence Transformers)
150
+
151
+ First install the Sentence Transformers library:
152
+
153
+ ```bash
154
+ pip install -U sentence-transformers
155
+ ```
156
+
157
+ Then you can load this model and run inference.
158
+ ```python
159
+ from sentence_transformers import SentenceTransformer
160
+
161
+ # Download from the 🤗 Hub
162
+ model = SentenceTransformer("iqrakiran/my-sentence-transformer-model")
163
+ # Run inference
164
+ sentences = [
165
+ "I remember generating alternative predictions. We said that I might get nervous but if I stay focused on what I want to do, I can probably do it. It's just hard to believe when I'm in this situation.",
166
+ "It's normal to feel nervous in difficult situations. As a counsellor, I suggest continuing to test these alternative predictions directly in real-life experiences. As you encounter situations where you manage to handle things even while nervous, your belief in your abilities and the predictions will strengthen. Remember to stay focused and trust yourself.",
167
+ 'If you are a counsellor, please answer the questions based on the description of the patient.',
168
+ ]
169
+ embeddings = model.encode(sentences)
170
+ print(embeddings.shape)
171
+ # [3, 768]
172
+
173
+ # Get the similarity scores for the embeddings
174
+ similarities = model.similarity(embeddings, embeddings)
175
+ print(similarities.shape)
176
+ # [3, 3]
177
+ ```
178
+
179
+ <!--
180
+ ### Direct Usage (Transformers)
181
+
182
+ <details><summary>Click to see the direct usage in Transformers</summary>
183
+
184
+ </details>
185
+ -->
186
+
187
+ <!--
188
+ ### Downstream Usage (Sentence Transformers)
189
+
190
+ You can finetune this model on your own dataset.
191
+
192
+ <details><summary>Click to expand</summary>
193
+
194
+ </details>
195
+ -->
196
+
197
+ <!--
198
+ ### Out-of-Scope Use
199
+
200
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
201
+ -->
202
+
203
+ <!--
204
+ ## Bias, Risks and Limitations
205
+
206
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
207
+ -->
208
+
209
+ <!--
210
+ ### Recommendations
211
+
212
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
213
+ -->
214
+
215
+ ## Training Details
216
+
217
+ ### Training Dataset
218
+
219
+ #### Unnamed Dataset
220
+
221
+
222
+ * Size: 4,093 training samples
223
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
224
+ * Approximate statistics based on the first 1000 samples:
225
+ | | sentence_0 | sentence_1 | sentence_2 |
226
+ |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
227
+ | type | string | string | string |
228
+ | details | <ul><li>min: 11 tokens</li><li>mean: 57.82 tokens</li><li>max: 197 tokens</li></ul> | <ul><li>min: 30 tokens</li><li>mean: 76.95 tokens</li><li>max: 149 tokens</li></ul> | <ul><li>min: 22 tokens</li><li>mean: 22.0 tokens</li><li>max: 22 tokens</li></ul> |
229
+ * Samples:
230
+ | sentence_0 | sentence_1 | sentence_2 |
231
+ |:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------|
232
+ | <code>Lately, I've been feeling really stressed out and I keep having this thought that my partner might be cheating on me. It started when I noticed changes in their behavior a couple of months ago. At times, I feel like my partner acts distant, but when I confront them about it, they assure me that everything is fine between us. I feel like maybe it's not a strong enough thought to be classified as a delusion, but it's really bothering me and affecting our relationship.</code> | <code>Thank you for sharing your concerns and feelings with me. It's important to know that it's normal for people to experience doubts and insecurities in relationships sometimes. I understand that feeling stressed and having negative thoughts is difficult to deal with. If you're open to it, perhaps we could explore these thoughts and feelings further to better understand their origin and find ways to cope with them. Additionally, involving your partner in a constructive conversation about your feelings and concerns might help to improve your relationship and provide reassurance.</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
233
+ | <code>I think counting would mean that somebody cares about me and I matter to them. In my life, for example, I think Ms. Jones is upset that I am not as involved or doing the things that she saw me do last year. So I guess I matter to her.</code> | <code>It's important to recognize that feeling valued and cared for by others can contribute to our overall well-being. It's great that you have someone like Ms. Jones in your life who notices your changes and seems to care about you. Can you tell me more about your relationship with Ms. Jones or the changes that have occurred in the past year?</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
234
+ | <code>Lately, I've been feeling quite down, and it's been affecting my daily life. I've also been having a hard time concentrating and making decisions, and often find myself lost in thought. On top of that, I've started hearing voices that aren't there, but it only happens occasionally. It all started when I moved to a new city a few months ago. I feel so isolated, and the stress of adjusting to a new environment might be causing these issues.</code> | <code>Thank you for sharing your feelings and experiences with me. It sounds like you are struggling with some challenging emotions and experiences since your move to a new city. Adjusting to a new environment can be a significant source of stress. It's essential to acknowledge what you're going through and understand that seeking help is healthy. Can you provide further details about the voices you've been hearing? How often do they occur, and how exactly do they affect your daily life? Also, can you share your previous mental health history, if any? Your additional information will help me get a better understanding of your situation and provide the best support possible.</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
235
+ * Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
236
+ ```json
237
+ {
238
+ "distance_metric": "TripletDistanceMetric.EUCLIDEAN",
239
+ "triplet_margin": 5
240
+ }
241
+ ```
242
+
243
+ ### Training Hyperparameters
244
+ #### Non-Default Hyperparameters
245
+
246
+ - `per_device_train_batch_size`: 16
247
+ - `per_device_eval_batch_size`: 16
248
+ - `num_train_epochs`: 10
249
+ - `multi_dataset_batch_sampler`: round_robin
250
+
251
+ #### All Hyperparameters
252
+ <details><summary>Click to expand</summary>
253
+
254
+ - `overwrite_output_dir`: False
255
+ - `do_predict`: False
256
+ - `eval_strategy`: no
257
+ - `prediction_loss_only`: True
258
+ - `per_device_train_batch_size`: 16
259
+ - `per_device_eval_batch_size`: 16
260
+ - `per_gpu_train_batch_size`: None
261
+ - `per_gpu_eval_batch_size`: None
262
+ - `gradient_accumulation_steps`: 1
263
+ - `eval_accumulation_steps`: None
264
+ - `torch_empty_cache_steps`: None
265
+ - `learning_rate`: 5e-05
266
+ - `weight_decay`: 0.0
267
+ - `adam_beta1`: 0.9
268
+ - `adam_beta2`: 0.999
269
+ - `adam_epsilon`: 1e-08
270
+ - `max_grad_norm`: 1
271
+ - `num_train_epochs`: 10
272
+ - `max_steps`: -1
273
+ - `lr_scheduler_type`: linear
274
+ - `lr_scheduler_kwargs`: {}
275
+ - `warmup_ratio`: 0.0
276
+ - `warmup_steps`: 0
277
+ - `log_level`: passive
278
+ - `log_level_replica`: warning
279
+ - `log_on_each_node`: True
280
+ - `logging_nan_inf_filter`: True
281
+ - `save_safetensors`: True
282
+ - `save_on_each_node`: False
283
+ - `save_only_model`: False
284
+ - `restore_callback_states_from_checkpoint`: False
285
+ - `no_cuda`: False
286
+ - `use_cpu`: False
287
+ - `use_mps_device`: False
288
+ - `seed`: 42
289
+ - `data_seed`: None
290
+ - `jit_mode_eval`: False
291
+ - `use_ipex`: False
292
+ - `bf16`: False
293
+ - `fp16`: False
294
+ - `fp16_opt_level`: O1
295
+ - `half_precision_backend`: auto
296
+ - `bf16_full_eval`: False
297
+ - `fp16_full_eval`: False
298
+ - `tf32`: None
299
+ - `local_rank`: 0
300
+ - `ddp_backend`: None
301
+ - `tpu_num_cores`: None
302
+ - `tpu_metrics_debug`: False
303
+ - `debug`: []
304
+ - `dataloader_drop_last`: False
305
+ - `dataloader_num_workers`: 0
306
+ - `dataloader_prefetch_factor`: None
307
+ - `past_index`: -1
308
+ - `disable_tqdm`: False
309
+ - `remove_unused_columns`: True
310
+ - `label_names`: None
311
+ - `load_best_model_at_end`: False
312
+ - `ignore_data_skip`: False
313
+ - `fsdp`: []
314
+ - `fsdp_min_num_params`: 0
315
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
316
+ - `fsdp_transformer_layer_cls_to_wrap`: None
317
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
318
+ - `deepspeed`: None
319
+ - `label_smoothing_factor`: 0.0
320
+ - `optim`: adamw_torch
321
+ - `optim_args`: None
322
+ - `adafactor`: False
323
+ - `group_by_length`: False
324
+ - `length_column_name`: length
325
+ - `ddp_find_unused_parameters`: None
326
+ - `ddp_bucket_cap_mb`: None
327
+ - `ddp_broadcast_buffers`: False
328
+ - `dataloader_pin_memory`: True
329
+ - `dataloader_persistent_workers`: False
330
+ - `skip_memory_metrics`: True
331
+ - `use_legacy_prediction_loop`: False
332
+ - `push_to_hub`: False
333
+ - `resume_from_checkpoint`: None
334
+ - `hub_model_id`: None
335
+ - `hub_strategy`: every_save
336
+ - `hub_private_repo`: False
337
+ - `hub_always_push`: False
338
+ - `gradient_checkpointing`: False
339
+ - `gradient_checkpointing_kwargs`: None
340
+ - `include_inputs_for_metrics`: False
341
+ - `eval_do_concat_batches`: True
342
+ - `fp16_backend`: auto
343
+ - `push_to_hub_model_id`: None
344
+ - `push_to_hub_organization`: None
345
+ - `mp_parameters`:
346
+ - `auto_find_batch_size`: False
347
+ - `full_determinism`: False
348
+ - `torchdynamo`: None
349
+ - `ray_scope`: last
350
+ - `ddp_timeout`: 1800
351
+ - `torch_compile`: False
352
+ - `torch_compile_backend`: None
353
+ - `torch_compile_mode`: None
354
+ - `dispatch_batches`: None
355
+ - `split_batches`: None
356
+ - `include_tokens_per_second`: False
357
+ - `include_num_input_tokens_seen`: False
358
+ - `neftune_noise_alpha`: None
359
+ - `optim_target_modules`: None
360
+ - `batch_eval_metrics`: False
361
+ - `eval_on_start`: False
362
+ - `eval_use_gather_object`: False
363
+ - `batch_sampler`: batch_sampler
364
+ - `multi_dataset_batch_sampler`: round_robin
365
+
366
+ </details>
367
+
368
+ ### Training Logs
369
+ | Epoch | Step | Training Loss |
370
+ |:------:|:----:|:-------------:|
371
+ | 1.9531 | 500 | 0.2321 |
372
+ | 3.9062 | 1000 | 0.0 |
373
+ | 5.8594 | 1500 | 0.0001 |
374
+ | 7.8125 | 2000 | 0.0 |
375
+ | 9.7656 | 2500 | 0.0 |
376
+
377
+
378
+ ### Framework Versions
379
+ - Python: 3.10.12
380
+ - Sentence Transformers: 3.0.1
381
+ - Transformers: 4.44.2
382
+ - PyTorch: 2.4.0+cu121
383
+ - Accelerate: 0.33.0
384
+ - Datasets: 2.21.0
385
+ - Tokenizers: 0.19.1
386
+
387
+ ## Citation
388
+
389
+ ### BibTeX
390
+
391
+ #### Sentence Transformers
392
+ ```bibtex
393
+ @inproceedings{reimers-2019-sentence-bert,
394
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
395
+ author = "Reimers, Nils and Gurevych, Iryna",
396
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
397
+ month = "11",
398
+ year = "2019",
399
+ publisher = "Association for Computational Linguistics",
400
+ url = "https://arxiv.org/abs/1908.10084",
401
+ }
402
+ ```
403
+
404
+ #### TripletLoss
405
+ ```bibtex
406
+ @misc{hermans2017defense,
407
+ title={In Defense of the Triplet Loss for Person Re-Identification},
408
+ author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
409
+ year={2017},
410
+ eprint={1703.07737},
411
+ archivePrefix={arXiv},
412
+ primaryClass={cs.CV}
413
+ }
414
+ ```
415
+
416
+ <!--
417
+ ## Glossary
418
+
419
+ *Clearly define terms in order to be accessible across audiences.*
420
+ -->
421
+
422
+ <!--
423
+ ## Model Card Authors
424
+
425
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
426
+ -->
427
+
428
+ <!--
429
+ ## Model Card Contact
430
+
431
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
432
+ -->
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilroberta-base",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 6,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.44.2",
24
+ "type_vocab_size": 1,
25
+ "use_cache": true,
26
+ "vocab_size": 50265
27
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.0.1",
4
+ "transformers": "4.44.2",
5
+ "pytorch": "2.4.0+cu121"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": null
10
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91b0eef6b288b62d98c15ce75d0da0cb94736f830ebb0612ac611e22bfb96bb
3
+ size 328485128
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "mask_token": "<mask>",
51
+ "model_max_length": 512,
52
+ "pad_token": "<pad>",
53
+ "sep_token": "</s>",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
+ "unk_token": "<unk>"
57
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff