96abhishekarora
commited on
Commit
•
3d73ad8
1
Parent(s):
5e651fe
Update README.md
Browse files
README.md
CHANGED
@@ -11,9 +11,13 @@ tags:
|
|
11 |
- news
|
12 |
---
|
13 |
|
14 |
-
This model was contrastively trained for entity coreference on a dataset constructed by mentions of the same entity. The model requires text with entities detected via NER and focuses specifically on Person [PER] tags.
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
-
We start with a base S-BERT MPNet bi-encoder model (18). This is constrastively trained on 179 million pairs taken from mentions of entities on Wikipedia, where positives are mentions of the same individual. Hard negatives are mined using individuals that appear on the same disambiguation pages. Embeddings from the tuned co-reference resolution model are then clustered using Hierarchical Agglomerative Clustering.
|
17 |
|
18 |
More information about its training (and use) can be found on the associated code [repo](https://github.com/dell-research-harvard/newswire/tree/main) and [paper](https://arxiv.org/pdf/2406.09490).
|
19 |
|
|
|
11 |
- news
|
12 |
---
|
13 |
|
14 |
+
This model was contrastively trained for entity coreference on a dataset constructed by mentions of the same entity. The model requires text with entities detected via NER and focuses specifically on Person [PER] tags. The mention of an entity of interest is then decorated a set of special tokens with the syntax - ([M ] Entity [\M ] .
|
15 |
+
|
16 |
+
We start with a base S-BERT MPNet bi-encoder [model](https://huggingface.co/sentence-transformers/all-mpnet-base-v2). This is contrastively trained on 179 million pairs taken from mentions of entities on Wikipedia, where positives are mentions of the same individual. Hard negatives are mined using individuals that appear on the same disambiguation pages.
|
17 |
+
This essentially trains a semantic similarity model.
|
18 |
+
|
19 |
+
At inference time, to collect mentions of the same entity, we reccomend a clustering contexts with deocrated entities in the embedding space - each cluster is then a set of mentions of the same entity, clusterd through the contextual information contained in the enclosing text.
|
20 |
|
|
|
21 |
|
22 |
More information about its training (and use) can be found on the associated code [repo](https://github.com/dell-research-harvard/newswire/tree/main) and [paper](https://arxiv.org/pdf/2406.09490).
|
23 |
|