{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,10]],"date-time":"2024-09-10T06:38:34Z","timestamp":1725950314478},"publisher-location":"Stroudsburg, PA, USA","reference-count":0,"publisher":"Association for Computational Linguistics","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.18653\/v1\/2020.findings-emnlp.151","type":"proceedings-article","created":{"date-parts":[[2020,11,29]],"date-time":"2020-11-29T14:58:51Z","timestamp":1606661931000},"page":"1675-1681","source":"Crossref","is-referenced-by-count":20,"title":["Cost-effective Selection of Pretraining Data: A Case Study of Pretraining BERT on Social Media"],"prefix":"10.18653","author":[{"given":"Xiang","family":"Dai","sequence":"first","affiliation":[]},{"given":"Sarvnaz","family":"Karimi","sequence":"additional","affiliation":[]},{"given":"Ben","family":"Hachey","sequence":"additional","affiliation":[]},{"given":"Cecile","family":"Paris","sequence":"additional","affiliation":[]}],"member":"1643","event":{"name":"Findings of the Association for Computational Linguistics: EMNLP 2020","start":{"date-parts":[[2020,11]]},"location":"Online","end":{"date-parts":[[2020,11]]}},"container-title":["Findings of the Association for Computational Linguistics: EMNLP 2020"],"original-title":[],"deposited":{"date-parts":[[2020,11,29]],"date-time":"2020-11-29T15:01:02Z","timestamp":1606662062000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.aclweb.org\/anthology\/2020.findings-emnlp.151"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"references-count":0,"URL":"http:\/\/dx.doi.org\/10.18653\/v1\/2020.findings-emnlp.151","relation":{},"subject":[],"published":{"date-parts":[[2020]]}}}