Efficiently and Thoroughly Anonymizing a Transformer Language Model for Dutch Electronic Health Records:a Two-Step Method

Neural Network (NN) architectures are used more and more to model large amounts of data, such as text data available online. Transformer-based NN architectures have shown to be very useful for language modelling. Although many researchers study how such Language Models (LMs) work, not much attention has been paid to the privacy risks of training LMs on large amounts of data and publishing them online. This paper presents a new method for anonymizing a language model by presenting the way in which MedRoBERTa.nl, a Dutch language model for hospital notes, was anonymized. The two step method invo... Mehr ...

Verfasser: Verkijk, Stella
Vossen, Piek
Dokumenttyp: contributionToPeriodical
Erscheinungsdatum: 2022
Verlag/Hrsg.: European Language Resources Association (ELRA)
Schlagwörter: Anonymization / Language Model / Medical Text Data / /dk/atira/pure/sustainabledevelopmentgoals/partnerships / name=SDG 17 - Partnerships for the Goals
Sprache: Englisch
Permalink: https://search.fid-benelux.de/Record/base-27463936
Datenquelle: BASE; Originalkatalog
Powered By: BASE
Link(s) : https://research.vu.nl/en/publications/be39e5ba-cce8-4af1-ab99-6d588be24855