Over the last few years, deep learning has revolutionized the field of machine learning by dramatically improving the stateof-the-art in various domains. However, as the size of supervised artificial neural networks grows, typically so does the need for larger labeled datasets. Recently, crowdsourcing has established itself as an efficient and cost-effective solution for labeling large sets of data in a scalable manner, but it often requires aggregating labels from multiple noisy contributors with different levels of expertise. In this paper, we address the problem of learning deep neural networks from crowds. We begin by describing an EM algorithm for jointly learning the parameters of the network and the reliabilities of the annotators. Then, a novel general-purpose crowd layer is proposed, which allows us to train deep neural networks end-to-end, directly from the noisy labels of multiple annotators, using only backpropagation. We empirically show that the proposed approach is able to internally capture the reliability and biases of different annotators and achieve new state-of-the-art results for various crowdsourced datasets across different settings, namely classification, regression and sequence labeling.
|Title of host publication||Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18)|
|Publication status||Published - 2018|
|Event||The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI), 2018 - |
Duration: 2 Feb 2018 → 7 Feb 2018
Conference number: 32
|Conference||The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI), 2018|
|Period||02/02/2018 → 07/02/2018|