Add new SentenceTransformer model.
Browse files- 1_Pooling/config.json +10 -0
- README.md +432 -0
- config.json +27 -0
- config_sentence_transformers.json +10 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- modules.json +14 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +15 -0
- tokenizer.json +0 -0
- tokenizer_config.json +57 -0
- vocab.json +0 -0
1_Pooling/config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 768,
|
3 |
+
"pooling_mode_cls_token": false,
|
4 |
+
"pooling_mode_mean_tokens": true,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
7 |
+
"pooling_mode_weightedmean_tokens": false,
|
8 |
+
"pooling_mode_lasttoken": false,
|
9 |
+
"include_prompt": true
|
10 |
+
}
|
README.md
ADDED
@@ -0,0 +1,432 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: distilbert/distilroberta-base
|
3 |
+
datasets: []
|
4 |
+
language: []
|
5 |
+
library_name: sentence-transformers
|
6 |
+
pipeline_tag: sentence-similarity
|
7 |
+
tags:
|
8 |
+
- sentence-transformers
|
9 |
+
- sentence-similarity
|
10 |
+
- feature-extraction
|
11 |
+
- generated_from_trainer
|
12 |
+
- dataset_size:4093
|
13 |
+
- loss:TripletLoss
|
14 |
+
widget:
|
15 |
+
- source_sentence: I am angry about an incident at the supermarket, and it has been
|
16 |
+
bothering me for the last three or four days. The person involved was flirting
|
17 |
+
with me, and I think they should apologize. I didn't ask for an apology or indicate
|
18 |
+
that I was upset during the conversation because I felt it would be inappropriate.
|
19 |
+
I believe they should call me and set aside an appropriate time to apologize.
|
20 |
+
sentences:
|
21 |
+
- If you are a counsellor, please answer the questions based on the description
|
22 |
+
of the patient.
|
23 |
+
- It seems like this incident really impacted you and you believe that an apology
|
24 |
+
is necessary for closure. It's understandable that you didn't address it on the
|
25 |
+
spot if you felt it was inappropriate. However, people sometimes don't realize
|
26 |
+
they've upset someone unless it's brought to their attention. Since you feel so
|
27 |
+
strongly about this, have you considered reaching out to the person to express
|
28 |
+
your feelings and provide them an opportunity to apologize?
|
29 |
+
- It sounds like you have built a strong support system in San Diego with meaningful
|
30 |
+
relationships and emotional investments. It's natural to feel apprehensive about
|
31 |
+
moving away from that. Maybe you could consider the potential opportunities that
|
32 |
+
LA has to offer and weigh them against what you would be leaving behind in San
|
33 |
+
Diego. Additionally, think about the ways you can stay connected with your loved
|
34 |
+
ones after moving. Just remember that it's a difficult decision, and it's important
|
35 |
+
to give yourself time to make the best choice for you.
|
36 |
+
- source_sentence: Lately, I've been feeling a bit overwhelmed with everything going
|
37 |
+
on in my life. I've been hearing things that aren't really there, like whispers
|
38 |
+
and someone calling my name, but it's not happening all the time, just occasionally.
|
39 |
+
I also find myself getting a bit disorganized, but I wouldn't say I'm completely
|
40 |
+
disorganized. Sometimes, I just feel disconnected with what is happening around
|
41 |
+
me. I've been like this for a couple of weeks now. My parents said that I've been
|
42 |
+
acting differently, but I don't know, maybe I'm just overreacting? Can you help
|
43 |
+
me with this?
|
44 |
+
sentences:
|
45 |
+
- If you are a counsellor, please answer the questions based on the description
|
46 |
+
of the patient.
|
47 |
+
- Thank you for sharing your experience with me. I can understand how hearing things
|
48 |
+
and feeling disorganized might be confusing and concerning for you. It's important
|
49 |
+
to discuss these symptoms and feelings in order to provide proper support. Can
|
50 |
+
you tell me if you've experienced any changes in your mood, daily activities,
|
51 |
+
or sleep habits during this time? Also, have you noticed any patterns or triggers
|
52 |
+
for when you hear the whispers or feel disconnected from what's happening around
|
53 |
+
you?
|
54 |
+
- I appreciate you sharing your feelings and experiences with me. It sounds like
|
55 |
+
you're going through a challenging period in your life, dealing with stress and
|
56 |
+
worry. It's important to understand the factors contributing to your stress and
|
57 |
+
how they might be influencing your thoughts and emotions. Could you provide some
|
58 |
+
more specific examples of the situations that have caused you to feel stressed
|
59 |
+
and worried? Additionally, have you noticed any patterns or triggers that seem
|
60 |
+
to make your worries more intense?
|
61 |
+
- source_sentence: I've been experiencing anxiety due to the difficult transition
|
62 |
+
from being at home, taking care of my son, to moving back into the work world.
|
63 |
+
I've been trying to manage this anxiety by paying attention to nutrition, exercise,
|
64 |
+
and practicing gratitude. Although I feel like I'm managing better now than when
|
65 |
+
I first started, I'm still struggling.
|
66 |
+
sentences:
|
67 |
+
- It's great to hear that you've made some progress in managing your anxiety through
|
68 |
+
healthy practices like focusing on nutrition, exercise, and gratitude. It's common
|
69 |
+
for individuals to face challenges while transitioning between roles, and it might
|
70 |
+
take more time for you to adjust fully. If you still find yourself struggling,
|
71 |
+
consider seeking additional support through therapy or support groups. Is there
|
72 |
+
a specific aspect of this transition that continues to cause you worry?
|
73 |
+
- If you are a counsellor, please answer the questions based on the description
|
74 |
+
of the patient.
|
75 |
+
- It's great to hear that you have a strong drive for self-improvement and setting
|
76 |
+
ambitious goals for yourself. Acknowledging this inner motivation is an essential
|
77 |
+
step towards achieving what you desire. Remember, it's important to embrace the
|
78 |
+
process of learning and growing, and not to put too much pressure on yourself
|
79 |
+
to reach an end goal quickly. Take it one step at a time, and don't forget to
|
80 |
+
celebrate your milestones along the way.
|
81 |
+
- source_sentence: Recently, I have been experiencing difficulties while learning.
|
82 |
+
Sometimes it's quite challenging, and I feel upset with myself to the point of
|
83 |
+
crying. However, at other times, I find it easy and feel a sense of accomplishment
|
84 |
+
when I finally get things right.
|
85 |
+
sentences:
|
86 |
+
- If you are a counsellor, please answer the questions based on the description
|
87 |
+
of the patient.
|
88 |
+
- Thank you for sharing that with me. I'm glad you're feeling better now that you're
|
89 |
+
here. Sometimes waiting can cause anxiety or stress, so it's important to find
|
90 |
+
ways to manage these feelings. If you'd like, we can discuss some coping strategies
|
91 |
+
for dealing with stress while waiting or in similar situations.
|
92 |
+
- It seems like you've been having ups and downs with your learning experience.
|
93 |
+
It's great that you sometimes feel accomplished when you finally understand something.
|
94 |
+
When you're feeling upset or frustrated, it might be helpful to take a break and
|
95 |
+
come back to it later, as you've mentioned. This can give you a fresh perspective
|
96 |
+
and may help reduce stress. It's also important to remember that learning is a
|
97 |
+
process, and it's okay to experience difficulties along the way.
|
98 |
+
- source_sentence: I remember generating alternative predictions. We said that I might
|
99 |
+
get nervous but if I stay focused on what I want to do, I can probably do it.
|
100 |
+
It's just hard to believe when I'm in this situation.
|
101 |
+
sentences:
|
102 |
+
- It seems like you've been dealing with unexpected expenses and some impulsive
|
103 |
+
spending habits. By planning for a stricter budget and understanding the difference
|
104 |
+
between essential and non-essential expenses, you can start to develop healthy
|
105 |
+
financial habits. Have you considered creating a budget plan and setting aside
|
106 |
+
a specific amount for unexpected expenses to help manage your spending?
|
107 |
+
- It's normal to feel nervous in difficult situations. As a counsellor, I suggest
|
108 |
+
continuing to test these alternative predictions directly in real-life experiences.
|
109 |
+
As you encounter situations where you manage to handle things even while nervous,
|
110 |
+
your belief in your abilities and the predictions will strengthen. Remember to
|
111 |
+
stay focused and trust yourself.
|
112 |
+
- If you are a counsellor, please answer the questions based on the description
|
113 |
+
of the patient.
|
114 |
+
---
|
115 |
+
|
116 |
+
# SentenceTransformer based on distilbert/distilroberta-base
|
117 |
+
|
118 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
119 |
+
|
120 |
+
## Model Details
|
121 |
+
|
122 |
+
### Model Description
|
123 |
+
- **Model Type:** Sentence Transformer
|
124 |
+
- **Base model:** [distilbert/distilroberta-base](https://huggingface.co/distilbert/distilroberta-base) <!-- at revision fb53ab8802853c8e4fbdbcd0529f21fc6f459b2b -->
|
125 |
+
- **Maximum Sequence Length:** 512 tokens
|
126 |
+
- **Output Dimensionality:** 768 tokens
|
127 |
+
- **Similarity Function:** Cosine Similarity
|
128 |
+
<!-- - **Training Dataset:** Unknown -->
|
129 |
+
<!-- - **Language:** Unknown -->
|
130 |
+
<!-- - **License:** Unknown -->
|
131 |
+
|
132 |
+
### Model Sources
|
133 |
+
|
134 |
+
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
135 |
+
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
136 |
+
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
137 |
+
|
138 |
+
### Full Model Architecture
|
139 |
+
|
140 |
+
```
|
141 |
+
SentenceTransformer(
|
142 |
+
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
|
143 |
+
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
|
144 |
+
)
|
145 |
+
```
|
146 |
+
|
147 |
+
## Usage
|
148 |
+
|
149 |
+
### Direct Usage (Sentence Transformers)
|
150 |
+
|
151 |
+
First install the Sentence Transformers library:
|
152 |
+
|
153 |
+
```bash
|
154 |
+
pip install -U sentence-transformers
|
155 |
+
```
|
156 |
+
|
157 |
+
Then you can load this model and run inference.
|
158 |
+
```python
|
159 |
+
from sentence_transformers import SentenceTransformer
|
160 |
+
|
161 |
+
# Download from the 🤗 Hub
|
162 |
+
model = SentenceTransformer("iqrakiran/my-sentence-transformer-model")
|
163 |
+
# Run inference
|
164 |
+
sentences = [
|
165 |
+
"I remember generating alternative predictions. We said that I might get nervous but if I stay focused on what I want to do, I can probably do it. It's just hard to believe when I'm in this situation.",
|
166 |
+
"It's normal to feel nervous in difficult situations. As a counsellor, I suggest continuing to test these alternative predictions directly in real-life experiences. As you encounter situations where you manage to handle things even while nervous, your belief in your abilities and the predictions will strengthen. Remember to stay focused and trust yourself.",
|
167 |
+
'If you are a counsellor, please answer the questions based on the description of the patient.',
|
168 |
+
]
|
169 |
+
embeddings = model.encode(sentences)
|
170 |
+
print(embeddings.shape)
|
171 |
+
# [3, 768]
|
172 |
+
|
173 |
+
# Get the similarity scores for the embeddings
|
174 |
+
similarities = model.similarity(embeddings, embeddings)
|
175 |
+
print(similarities.shape)
|
176 |
+
# [3, 3]
|
177 |
+
```
|
178 |
+
|
179 |
+
<!--
|
180 |
+
### Direct Usage (Transformers)
|
181 |
+
|
182 |
+
<details><summary>Click to see the direct usage in Transformers</summary>
|
183 |
+
|
184 |
+
</details>
|
185 |
+
-->
|
186 |
+
|
187 |
+
<!--
|
188 |
+
### Downstream Usage (Sentence Transformers)
|
189 |
+
|
190 |
+
You can finetune this model on your own dataset.
|
191 |
+
|
192 |
+
<details><summary>Click to expand</summary>
|
193 |
+
|
194 |
+
</details>
|
195 |
+
-->
|
196 |
+
|
197 |
+
<!--
|
198 |
+
### Out-of-Scope Use
|
199 |
+
|
200 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
201 |
+
-->
|
202 |
+
|
203 |
+
<!--
|
204 |
+
## Bias, Risks and Limitations
|
205 |
+
|
206 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
207 |
+
-->
|
208 |
+
|
209 |
+
<!--
|
210 |
+
### Recommendations
|
211 |
+
|
212 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
213 |
+
-->
|
214 |
+
|
215 |
+
## Training Details
|
216 |
+
|
217 |
+
### Training Dataset
|
218 |
+
|
219 |
+
#### Unnamed Dataset
|
220 |
+
|
221 |
+
|
222 |
+
* Size: 4,093 training samples
|
223 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
224 |
+
* Approximate statistics based on the first 1000 samples:
|
225 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
226 |
+
|:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
227 |
+
| type | string | string | string |
|
228 |
+
| details | <ul><li>min: 11 tokens</li><li>mean: 57.82 tokens</li><li>max: 197 tokens</li></ul> | <ul><li>min: 30 tokens</li><li>mean: 76.95 tokens</li><li>max: 149 tokens</li></ul> | <ul><li>min: 22 tokens</li><li>mean: 22.0 tokens</li><li>max: 22 tokens</li></ul> |
|
229 |
+
* Samples:
|
230 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
231 |
+
|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------|
|
232 |
+
| <code>Lately, I've been feeling really stressed out and I keep having this thought that my partner might be cheating on me. It started when I noticed changes in their behavior a couple of months ago. At times, I feel like my partner acts distant, but when I confront them about it, they assure me that everything is fine between us. I feel like maybe it's not a strong enough thought to be classified as a delusion, but it's really bothering me and affecting our relationship.</code> | <code>Thank you for sharing your concerns and feelings with me. It's important to know that it's normal for people to experience doubts and insecurities in relationships sometimes. I understand that feeling stressed and having negative thoughts is difficult to deal with. If you're open to it, perhaps we could explore these thoughts and feelings further to better understand their origin and find ways to cope with them. Additionally, involving your partner in a constructive conversation about your feelings and concerns might help to improve your relationship and provide reassurance.</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
|
233 |
+
| <code>I think counting would mean that somebody cares about me and I matter to them. In my life, for example, I think Ms. Jones is upset that I am not as involved or doing the things that she saw me do last year. So I guess I matter to her.</code> | <code>It's important to recognize that feeling valued and cared for by others can contribute to our overall well-being. It's great that you have someone like Ms. Jones in your life who notices your changes and seems to care about you. Can you tell me more about your relationship with Ms. Jones or the changes that have occurred in the past year?</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
|
234 |
+
| <code>Lately, I've been feeling quite down, and it's been affecting my daily life. I've also been having a hard time concentrating and making decisions, and often find myself lost in thought. On top of that, I've started hearing voices that aren't there, but it only happens occasionally. It all started when I moved to a new city a few months ago. I feel so isolated, and the stress of adjusting to a new environment might be causing these issues.</code> | <code>Thank you for sharing your feelings and experiences with me. It sounds like you are struggling with some challenging emotions and experiences since your move to a new city. Adjusting to a new environment can be a significant source of stress. It's essential to acknowledge what you're going through and understand that seeking help is healthy. Can you provide further details about the voices you've been hearing? How often do they occur, and how exactly do they affect your daily life? Also, can you share your previous mental health history, if any? Your additional information will help me get a better understanding of your situation and provide the best support possible.</code> | <code>If you are a counsellor, please answer the questions based on the description of the patient.</code> |
|
235 |
+
* Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
|
236 |
+
```json
|
237 |
+
{
|
238 |
+
"distance_metric": "TripletDistanceMetric.EUCLIDEAN",
|
239 |
+
"triplet_margin": 5
|
240 |
+
}
|
241 |
+
```
|
242 |
+
|
243 |
+
### Training Hyperparameters
|
244 |
+
#### Non-Default Hyperparameters
|
245 |
+
|
246 |
+
- `per_device_train_batch_size`: 16
|
247 |
+
- `per_device_eval_batch_size`: 16
|
248 |
+
- `num_train_epochs`: 10
|
249 |
+
- `multi_dataset_batch_sampler`: round_robin
|
250 |
+
|
251 |
+
#### All Hyperparameters
|
252 |
+
<details><summary>Click to expand</summary>
|
253 |
+
|
254 |
+
- `overwrite_output_dir`: False
|
255 |
+
- `do_predict`: False
|
256 |
+
- `eval_strategy`: no
|
257 |
+
- `prediction_loss_only`: True
|
258 |
+
- `per_device_train_batch_size`: 16
|
259 |
+
- `per_device_eval_batch_size`: 16
|
260 |
+
- `per_gpu_train_batch_size`: None
|
261 |
+
- `per_gpu_eval_batch_size`: None
|
262 |
+
- `gradient_accumulation_steps`: 1
|
263 |
+
- `eval_accumulation_steps`: None
|
264 |
+
- `torch_empty_cache_steps`: None
|
265 |
+
- `learning_rate`: 5e-05
|
266 |
+
- `weight_decay`: 0.0
|
267 |
+
- `adam_beta1`: 0.9
|
268 |
+
- `adam_beta2`: 0.999
|
269 |
+
- `adam_epsilon`: 1e-08
|
270 |
+
- `max_grad_norm`: 1
|
271 |
+
- `num_train_epochs`: 10
|
272 |
+
- `max_steps`: -1
|
273 |
+
- `lr_scheduler_type`: linear
|
274 |
+
- `lr_scheduler_kwargs`: {}
|
275 |
+
- `warmup_ratio`: 0.0
|
276 |
+
- `warmup_steps`: 0
|
277 |
+
- `log_level`: passive
|
278 |
+
- `log_level_replica`: warning
|
279 |
+
- `log_on_each_node`: True
|
280 |
+
- `logging_nan_inf_filter`: True
|
281 |
+
- `save_safetensors`: True
|
282 |
+
- `save_on_each_node`: False
|
283 |
+
- `save_only_model`: False
|
284 |
+
- `restore_callback_states_from_checkpoint`: False
|
285 |
+
- `no_cuda`: False
|
286 |
+
- `use_cpu`: False
|
287 |
+
- `use_mps_device`: False
|
288 |
+
- `seed`: 42
|
289 |
+
- `data_seed`: None
|
290 |
+
- `jit_mode_eval`: False
|
291 |
+
- `use_ipex`: False
|
292 |
+
- `bf16`: False
|
293 |
+
- `fp16`: False
|
294 |
+
- `fp16_opt_level`: O1
|
295 |
+
- `half_precision_backend`: auto
|
296 |
+
- `bf16_full_eval`: False
|
297 |
+
- `fp16_full_eval`: False
|
298 |
+
- `tf32`: None
|
299 |
+
- `local_rank`: 0
|
300 |
+
- `ddp_backend`: None
|
301 |
+
- `tpu_num_cores`: None
|
302 |
+
- `tpu_metrics_debug`: False
|
303 |
+
- `debug`: []
|
304 |
+
- `dataloader_drop_last`: False
|
305 |
+
- `dataloader_num_workers`: 0
|
306 |
+
- `dataloader_prefetch_factor`: None
|
307 |
+
- `past_index`: -1
|
308 |
+
- `disable_tqdm`: False
|
309 |
+
- `remove_unused_columns`: True
|
310 |
+
- `label_names`: None
|
311 |
+
- `load_best_model_at_end`: False
|
312 |
+
- `ignore_data_skip`: False
|
313 |
+
- `fsdp`: []
|
314 |
+
- `fsdp_min_num_params`: 0
|
315 |
+
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
|
316 |
+
- `fsdp_transformer_layer_cls_to_wrap`: None
|
317 |
+
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
|
318 |
+
- `deepspeed`: None
|
319 |
+
- `label_smoothing_factor`: 0.0
|
320 |
+
- `optim`: adamw_torch
|
321 |
+
- `optim_args`: None
|
322 |
+
- `adafactor`: False
|
323 |
+
- `group_by_length`: False
|
324 |
+
- `length_column_name`: length
|
325 |
+
- `ddp_find_unused_parameters`: None
|
326 |
+
- `ddp_bucket_cap_mb`: None
|
327 |
+
- `ddp_broadcast_buffers`: False
|
328 |
+
- `dataloader_pin_memory`: True
|
329 |
+
- `dataloader_persistent_workers`: False
|
330 |
+
- `skip_memory_metrics`: True
|
331 |
+
- `use_legacy_prediction_loop`: False
|
332 |
+
- `push_to_hub`: False
|
333 |
+
- `resume_from_checkpoint`: None
|
334 |
+
- `hub_model_id`: None
|
335 |
+
- `hub_strategy`: every_save
|
336 |
+
- `hub_private_repo`: False
|
337 |
+
- `hub_always_push`: False
|
338 |
+
- `gradient_checkpointing`: False
|
339 |
+
- `gradient_checkpointing_kwargs`: None
|
340 |
+
- `include_inputs_for_metrics`: False
|
341 |
+
- `eval_do_concat_batches`: True
|
342 |
+
- `fp16_backend`: auto
|
343 |
+
- `push_to_hub_model_id`: None
|
344 |
+
- `push_to_hub_organization`: None
|
345 |
+
- `mp_parameters`:
|
346 |
+
- `auto_find_batch_size`: False
|
347 |
+
- `full_determinism`: False
|
348 |
+
- `torchdynamo`: None
|
349 |
+
- `ray_scope`: last
|
350 |
+
- `ddp_timeout`: 1800
|
351 |
+
- `torch_compile`: False
|
352 |
+
- `torch_compile_backend`: None
|
353 |
+
- `torch_compile_mode`: None
|
354 |
+
- `dispatch_batches`: None
|
355 |
+
- `split_batches`: None
|
356 |
+
- `include_tokens_per_second`: False
|
357 |
+
- `include_num_input_tokens_seen`: False
|
358 |
+
- `neftune_noise_alpha`: None
|
359 |
+
- `optim_target_modules`: None
|
360 |
+
- `batch_eval_metrics`: False
|
361 |
+
- `eval_on_start`: False
|
362 |
+
- `eval_use_gather_object`: False
|
363 |
+
- `batch_sampler`: batch_sampler
|
364 |
+
- `multi_dataset_batch_sampler`: round_robin
|
365 |
+
|
366 |
+
</details>
|
367 |
+
|
368 |
+
### Training Logs
|
369 |
+
| Epoch | Step | Training Loss |
|
370 |
+
|:------:|:----:|:-------------:|
|
371 |
+
| 1.9531 | 500 | 0.2321 |
|
372 |
+
| 3.9062 | 1000 | 0.0 |
|
373 |
+
| 5.8594 | 1500 | 0.0001 |
|
374 |
+
| 7.8125 | 2000 | 0.0 |
|
375 |
+
| 9.7656 | 2500 | 0.0 |
|
376 |
+
|
377 |
+
|
378 |
+
### Framework Versions
|
379 |
+
- Python: 3.10.12
|
380 |
+
- Sentence Transformers: 3.0.1
|
381 |
+
- Transformers: 4.44.2
|
382 |
+
- PyTorch: 2.4.0+cu121
|
383 |
+
- Accelerate: 0.33.0
|
384 |
+
- Datasets: 2.21.0
|
385 |
+
- Tokenizers: 0.19.1
|
386 |
+
|
387 |
+
## Citation
|
388 |
+
|
389 |
+
### BibTeX
|
390 |
+
|
391 |
+
#### Sentence Transformers
|
392 |
+
```bibtex
|
393 |
+
@inproceedings{reimers-2019-sentence-bert,
|
394 |
+
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
395 |
+
author = "Reimers, Nils and Gurevych, Iryna",
|
396 |
+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
397 |
+
month = "11",
|
398 |
+
year = "2019",
|
399 |
+
publisher = "Association for Computational Linguistics",
|
400 |
+
url = "https://arxiv.org/abs/1908.10084",
|
401 |
+
}
|
402 |
+
```
|
403 |
+
|
404 |
+
#### TripletLoss
|
405 |
+
```bibtex
|
406 |
+
@misc{hermans2017defense,
|
407 |
+
title={In Defense of the Triplet Loss for Person Re-Identification},
|
408 |
+
author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
|
409 |
+
year={2017},
|
410 |
+
eprint={1703.07737},
|
411 |
+
archivePrefix={arXiv},
|
412 |
+
primaryClass={cs.CV}
|
413 |
+
}
|
414 |
+
```
|
415 |
+
|
416 |
+
<!--
|
417 |
+
## Glossary
|
418 |
+
|
419 |
+
*Clearly define terms in order to be accessible across audiences.*
|
420 |
+
-->
|
421 |
+
|
422 |
+
<!--
|
423 |
+
## Model Card Authors
|
424 |
+
|
425 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
426 |
+
-->
|
427 |
+
|
428 |
+
<!--
|
429 |
+
## Model Card Contact
|
430 |
+
|
431 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
432 |
+
-->
|
config.json
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilroberta-base",
|
3 |
+
"architectures": [
|
4 |
+
"RobertaModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "gelu",
|
11 |
+
"hidden_dropout_prob": 0.1,
|
12 |
+
"hidden_size": 768,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 514,
|
17 |
+
"model_type": "roberta",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 6,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"position_embedding_type": "absolute",
|
22 |
+
"torch_dtype": "float32",
|
23 |
+
"transformers_version": "4.44.2",
|
24 |
+
"type_vocab_size": 1,
|
25 |
+
"use_cache": true,
|
26 |
+
"vocab_size": 50265
|
27 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "3.0.1",
|
4 |
+
"transformers": "4.44.2",
|
5 |
+
"pytorch": "2.4.0+cu121"
|
6 |
+
},
|
7 |
+
"prompts": {},
|
8 |
+
"default_prompt_name": null,
|
9 |
+
"similarity_fn_name": null
|
10 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c91b0eef6b288b62d98c15ce75d0da0cb94736f830ebb0612ac611e22bfb96bb
|
3 |
+
size 328485128
|
modules.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
}
|
14 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 512,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"content": "<mask>",
|
7 |
+
"lstrip": true,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"pad_token": "<pad>",
|
13 |
+
"sep_token": "</s>",
|
14 |
+
"unk_token": "<unk>"
|
15 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"0": {
|
5 |
+
"content": "<s>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"1": {
|
13 |
+
"content": "<pad>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": true,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"2": {
|
21 |
+
"content": "</s>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": true,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
},
|
28 |
+
"3": {
|
29 |
+
"content": "<unk>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": true,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false,
|
34 |
+
"special": true
|
35 |
+
},
|
36 |
+
"50264": {
|
37 |
+
"content": "<mask>",
|
38 |
+
"lstrip": true,
|
39 |
+
"normalized": false,
|
40 |
+
"rstrip": false,
|
41 |
+
"single_word": false,
|
42 |
+
"special": true
|
43 |
+
}
|
44 |
+
},
|
45 |
+
"bos_token": "<s>",
|
46 |
+
"clean_up_tokenization_spaces": true,
|
47 |
+
"cls_token": "<s>",
|
48 |
+
"eos_token": "</s>",
|
49 |
+
"errors": "replace",
|
50 |
+
"mask_token": "<mask>",
|
51 |
+
"model_max_length": 512,
|
52 |
+
"pad_token": "<pad>",
|
53 |
+
"sep_token": "</s>",
|
54 |
+
"tokenizer_class": "RobertaTokenizer",
|
55 |
+
"trim_offsets": true,
|
56 |
+
"unk_token": "<unk>"
|
57 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|