dardem commited on
Commit
3fb5045
1 Parent(s): 278fe3c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -73,6 +73,25 @@ You can also check out our [web-demo](https://detoxifier.nlp.zhores.net/junction
73
  abstract = "We present a novel pipeline for the collection of parallel data for the detoxification task. We collect non-toxic paraphrases for over 10,000 English toxic sentences. We also show that this pipeline can be used to distill a large existing corpus of paraphrases to get toxic-neutral sentence pairs. We release two parallel corpora which can be used for the training of detoxification models. To the best of our knowledge, these are the first parallel datasets for this task.We describe our pipeline in detail to make it fast to set up for a new language or domain, thus contributing to faster and easier development of new parallel resources.We train several detoxification models on the collected data and compare them with several baselines and state-of-the-art unsupervised approaches. We conduct both automatic and manual evaluations. All models trained on parallel data outperform the state-of-the-art unsupervised models by a large margin. This suggests that our novel datasets can boost the performance of detoxification systems.",
74
  }
75
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
 
77
  ## Contacts
78
 
 
73
  abstract = "We present a novel pipeline for the collection of parallel data for the detoxification task. We collect non-toxic paraphrases for over 10,000 English toxic sentences. We also show that this pipeline can be used to distill a large existing corpus of paraphrases to get toxic-neutral sentence pairs. We release two parallel corpora which can be used for the training of detoxification models. To the best of our knowledge, these are the first parallel datasets for this task.We describe our pipeline in detail to make it fast to set up for a new language or domain, thus contributing to faster and easier development of new parallel resources.We train several detoxification models on the collected data and compare them with several baselines and state-of-the-art unsupervised approaches. We conduct both automatic and manual evaluations. All models trained on parallel data outperform the state-of-the-art unsupervised models by a large margin. This suggests that our novel datasets can boost the performance of detoxification systems.",
74
  }
75
  ```
76
+ and
77
+ ```
78
+ @inproceedings{dementieva2021crowdsourcing,
79
+ title = "Crowdsourcing of Parallel Corpora: the Case of Style Transfer for Detoxification",
80
+ author = {Dementieva, Daryna
81
+ and Ustyantsev, Sergey
82
+ and Dale, David
83
+ and Kozlova, Olga
84
+ and Semenov, Nikita
85
+ and Panchenko, Alexander
86
+ and Logacheva, Varvara},
87
+ booktitle = "Proceedings of the 2nd Crowd Science Workshop: Trust, Ethics, and Excellence in Crowdsourced Data Management at Scale co-located with 47th International Conference on Very Large Data Bases (VLDB 2021 (https://vldb.org/2021/))",
88
+ year = "2021",
89
+ address = "Copenhagen, Denmark",
90
+ publisher = "CEUR Workshop Proceedings",
91
+ pages = "35--49",
92
+ url={http://ceur-ws.org/Vol-2932/paper2.pdf}
93
+ }
94
+ ```
95
 
96
  ## Contacts
97