Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels
Citizen science has become a valuable and reliable method for interpreting and processing big datasets, and is vital in the era of ever-growing data volumes. However, there are inherent difficulties in the generating labels from citizen scientists, due to the inherent variability between the members...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Ubiquity Press
2024-12-01
|
Series: | Citizen Science: Theory and Practice |
Subjects: | |
Online Access: | https://account.theoryandpractice.citizenscienceassociation.org/index.php/up-j-cstp/article/view/731 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1841554925594607616 |
---|---|
author | Ramanakumar Sankar Kameswara Mantha Cooper Nesmith Lucy Fortson Shawn Brueshaber Candice Hansen-Koharcheck Glenn Orton |
author_facet | Ramanakumar Sankar Kameswara Mantha Cooper Nesmith Lucy Fortson Shawn Brueshaber Candice Hansen-Koharcheck Glenn Orton |
author_sort | Ramanakumar Sankar |
collection | DOAJ |
description | Citizen science has become a valuable and reliable method for interpreting and processing big datasets, and is vital in the era of ever-growing data volumes. However, there are inherent difficulties in the generating labels from citizen scientists, due to the inherent variability between the members of the crowd, leading to variability in the results. Sometimes, this is useful — such as with serendipitous discoveries, which corresponds to rare/unknown classes in the data — but it might also be due to ambiguity between classes. The primary issue is then to distinguish between the intrinsic variability in the dataset and the uncertainty in the citizen scientists’ responses, and leveraging that to extract scientifically useful relationships. In this paper, we explore using a neural network to interpret volunteer confusion across the dataset, to increase the purity of the downstream analysis. We focus on the use of learned features from the network to disentangle feature similarity across the classes, and the ability of the machines’ “attention” in identifying features that lead to confusion. We use data from Jovian Vortex Hunter, a citizen science project to study vortices in Jupiter’s atmosphere, and find that the latent space from the model helps effectively identify different sources of image-level features that lead to low volunteer consensus. Furthermore, the machine’s attention highlights features corresponding to specific classes. This provides meaningful image-level feature-class relationships, which is useful in our analysis for identifying vortex-specific features to better understand vortex evolution mechanisms. Finally, we discuss the applicability of this method to other citizen science projects. |
format | Article |
id | doaj-art-2fac5c5f839745b8949aa502c8bb1a9f |
institution | Kabale University |
issn | 2057-4991 |
language | English |
publishDate | 2024-12-01 |
publisher | Ubiquity Press |
record_format | Article |
series | Citizen Science: Theory and Practice |
spelling | doaj-art-2fac5c5f839745b8949aa502c8bb1a9f2025-01-08T07:54:40ZengUbiquity PressCitizen Science: Theory and Practice2057-49912024-12-0191414110.5334/cstp.731713Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer LabelsRamanakumar Sankar0https://orcid.org/0000-0002-6794-7587Kameswara Mantha1https://orcid.org/0000-0002-6016-300XCooper Nesmith2Lucy Fortson3https://orcid.org/0000-0002-1067-8558Shawn Brueshaber4https://orcid.org/0000-0002-3669-0539Candice Hansen-Koharcheck5Glenn Orton6https://orcid.org/0000-0001-7871-2823University of California, BerkeleyUniversity of Minnesota, Twin CitiesUniversity of Minnesota, Twin CitiesUniversity of Minnesota, Twin CitiesMichigan Technological UniversityPlanetary Science InstituteJet Propulsion Laboratory/California Institute of TechnologyCitizen science has become a valuable and reliable method for interpreting and processing big datasets, and is vital in the era of ever-growing data volumes. However, there are inherent difficulties in the generating labels from citizen scientists, due to the inherent variability between the members of the crowd, leading to variability in the results. Sometimes, this is useful — such as with serendipitous discoveries, which corresponds to rare/unknown classes in the data — but it might also be due to ambiguity between classes. The primary issue is then to distinguish between the intrinsic variability in the dataset and the uncertainty in the citizen scientists’ responses, and leveraging that to extract scientifically useful relationships. In this paper, we explore using a neural network to interpret volunteer confusion across the dataset, to increase the purity of the downstream analysis. We focus on the use of learned features from the network to disentangle feature similarity across the classes, and the ability of the machines’ “attention” in identifying features that lead to confusion. We use data from Jovian Vortex Hunter, a citizen science project to study vortices in Jupiter’s atmosphere, and find that the latent space from the model helps effectively identify different sources of image-level features that lead to low volunteer consensus. Furthermore, the machine’s attention highlights features corresponding to specific classes. This provides meaningful image-level feature-class relationships, which is useful in our analysis for identifying vortex-specific features to better understand vortex evolution mechanisms. Finally, we discuss the applicability of this method to other citizen science projects.https://account.theoryandpractice.citizenscienceassociation.org/index.php/up-j-cstp/article/view/731citizen sciencecrowdsourcingauto-encodersemi-supervised networkmachine attentionplanetary atmospheres |
spellingShingle | Ramanakumar Sankar Kameswara Mantha Cooper Nesmith Lucy Fortson Shawn Brueshaber Candice Hansen-Koharcheck Glenn Orton Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels Citizen Science: Theory and Practice citizen science crowdsourcing auto-encoder semi-supervised network machine attention planetary atmospheres |
title | Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels |
title_full | Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels |
title_fullStr | Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels |
title_full_unstemmed | Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels |
title_short | Understanding Confusion: A Case Study of Training a Machine Model to Predict and Interpret Consensus From Volunteer Labels |
title_sort | understanding confusion a case study of training a machine model to predict and interpret consensus from volunteer labels |
topic | citizen science crowdsourcing auto-encoder semi-supervised network machine attention planetary atmospheres |
url | https://account.theoryandpractice.citizenscienceassociation.org/index.php/up-j-cstp/article/view/731 |
work_keys_str_mv | AT ramanakumarsankar understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT kameswaramantha understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT coopernesmith understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT lucyfortson understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT shawnbrueshaber understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT candicehansenkoharcheck understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels AT glennorton understandingconfusionacasestudyoftrainingamachinemodeltopredictandinterpretconsensusfromvolunteerlabels |