Citizen science

The educational technology and digital learning wiki
Jump to navigation Jump to search

Draft

I will start working on this piece on January 2012 and try to do some quick literature review, in particular with respect to topics like "how do participants learn", "in what respect are citizens creative", "what is their motivation", "how do communities work".

- Daniel K. Schneider 16:43, 23 December 2011 (CET)

Introduction

Citizen science does not have a uniquely accepted definition. It could mean:

  • Participation of citizen for collection of data, for example observation of animals, pollution, or plant growth.
  • Participation of citizen for analyzing data, in various forms. For example, some provide computing power (a typical example is the captcha mechanism in this wiki for curbing spam). Another would be helping to recognize patterns (e.g. forms of galaxies)
  • Dissemination of scientific thought and result in schools in order to promote engagement with science or with the intent to help updating the curriculum.
  • Amateur science, i.e. citizen create scientific thoughts and other products.
  • Citizen assessment of science and scientific projects.

See also: e-science

Types of citizen science

The variety of citizen science programs is important with respect to many criteria, e.g.: aims, target population, locations (schools, museums, media, Internet groups), forms, subject areas, tasks, etc.

Cooper et al. (2007) distinguish between the citizen science model and the participatory action research model. The former uses citizens as data collectors and will receive recognition (including results of the study). Participatory action research “begins with the interests of participants, who work collaboratively with professional researchers through all steps of the scientific process to find solutions to problems of community relevance Finn (1994) outlined three key elements of participatory research: (1) it responds to the experiences and needs of the community, (2) it fosters collaboration between researchers and community in research activities, and (3) it promotes common knowledge and increases community awareness. Although citizen science can have research and education goals similar to many participatory action research projects (Finn 1994 and below), citizen science is distinct from participatory action research in that it occurs at larger scales and typically does not incorporate iterative or collaborative action.”. However, some citizen science project do actively try to engage citizens to participate as we shall try to show below.

Wiggins and Crowston (2011) identified “five mutually exclusive and exhaustive types of projects, which [they] labelled Action, Conservation, Investigation, Virtual and Education. Action projects employ volunteer-initiated participatory action research to encourage participant intervention in local concerns. Conservation projects address natural resource management goals, involving citizens in stewardship for outreach and increased scope. Investigation projects focus on scientific research goals in a physical setting, while Virtual projects have goals similar to Investigation projects, but are entirely ICT-mediated and differ in a number of other characteristics. Finally, Education projects make education and outreach primary goals”.

Wiggins and Crowston (2012) then created two typologies of citizen science projets, using data from 63 projects. A first typology is based on twelve participation tasks, i.e. Observation, Species identification, Classification or tagging, Data entry, Finding entities, Measurement, Specimen/sample collection, Sample analysis, Site selection &/or description, Geolocation, Photography, Data analysis, and Number of tasks. The other typology is based on ten project goals, i.e. Science, Management, Action, Education, Conservation, Monitoring, Restoration, Outreach, Stewardship, and Discovery.

Participation clusters:

  1. Involve observation and identification tasks, but never require analysis. (13)
  2. Involve observation, data entry, and analysis, but include no locational tasks such as site selection or geolocation (6)
  3. Engage participants in a variety of tasks, the only participation task not represented is sample analysis (17)
  4. Engage the public in every participation task considered (15)
  5. Involve reporting, using observations, species identification, and data entry, but with very few additional participation opportunities (12)

Goal clusters:

  1. Afforded nearly equal weights (midpoint of the scale or higher) to each of the goal areas
  2. Most strongly focused on science
  3. Science is the most important goal, but education, monitoring, and discovery are only slightly less important on average
  4. Science, conservation, monitoring and stewardship are most important, while discovery is less valued than in the preceding clusters
  5. Outlier

For other typologies, see Wiggins and Crowston (2012).

Topics in the study of citizen science projects

Motivation

Motivation and motivation in education are complex constructs. A typical general model includes several components. E.g. Herzberg's (1959) five component model of job satisfaction includes achievement, recognition, work itself, responsibility, and advancement.

Some recent research by Raddick et al. (2010) about motivations of participants in cyberscience projects analyzed free forum messages and structured interviews. The first mentioned categories include the subject (astronomy), contribution and vastness. Learning (3%) and Science (1%) were minor. Looking at all responses, the same motivations - subject (46%), contribution (22%), vastness (24%) - still dominate, but other motivations exceed 10%, e.g. beauty (16%), fun (11%) and learning (10%).

It might be interesting to also look at personality traits of participants. E.g. Furnham et al. (1999) in a study about personality and work motivation found out that extroverts on the Eysenck Personality Profile correlate with a preference for Herzberg's motivators, whereas "neuroticism" correlates with so-called hygiene factors, i.e. the environment. In other words, extraverts are rather motivated by intrinsic factors whereas others by extrinsic ones.

Virtual organisations for citizen science

An important variant of citizen science uses citizens as helpers for research. “Citizen science is a form of organisation design for collaborative scientific research involving scientists and volunteers, for which internet-based modes of participation enable massive virtual collaboration by thousands of members of the public.” (Wiggins and Crowson, 2010:148). The authors argue that virtual organisations for citizen science are a bit different from other virtual organizations: “The project level of group interaction is distinct from those of small work groups and organisations (Grudin, 1994), which has implications for organisation design efforts. Project teams and communities of practice can be distinguished by their goal orientation among other features (Wenger, 1999), but empirical observation of citizen science VOs to date indicates a hybrid ‘community of purpose’ might better describe many projects, with characteristics of both a project team and a community of practice or interest.” (p 159).

Wiggins and Crowson then suggest a conceptual model of citizen science VOs:

Source: Wiggins and Crowston, Preprint. Reproduced without permission

Learning in citizen science projects

Luther et al. (2009) related learning in citizen science projects to Wikipedia and open source (OSS) participation. “researchers have identified some surprising commonalities between Wikipedia and OSS worth mentioning here. The first is legitimate peripheral participation (LPP), a theory of social learning which holds that novices in a community of practice may transform into experts by observing expert practices and taking on increasingly complex work (Lave & Wenger, 1991). In the case of Wikipedia, a user may begin by fixing typos, gradually making more substantial edits until she is writing entire articles from scratch (Brynt et al. 2005). In OSS, bug reporters may write a few lines of code to fix a glitch and eventually learn to code new modules”.

The Caise report argues that “educational research shows that people have greater motivation to engage and learn if the subject matter is directly relevant to their lives and interests and/or if the learning process is interactive—one in which the learner can directly affect the learning process, content, and/or outcomes of the experience (Falk 2001)”.

According to Cooper et al (2007) “Citizen science also provides informal learning experiences that improve science literacy (Krasny and Bonney 2005, Evans et al. 2005). Several studies have found that citizen science projects at the Cornell Laboratory of Ornithology have improved participants’ knowledge about biology and natural history, engaged them directly in the process of inquisitive thinking, and increased their ability to frame questions scientifically (Bonney and Dhondt 1997, Trumbull et al. 2000, 2005, Lewenstein 2001, Bonney 2004, Krasny and Bonney 2005).”

Evaluation of informal science education

Informal science education refers mostly to larger or smaller top-down initiatives that aim to raise interest for STEM (Science, technology, engineering and mathematics) subjects. As an example, “The Informal Science Education (ISE) program at the National Science Foundation (NSF) invests in projects designed to increase interest in, engagement with, and understanding of science, technology, engineering, and mathematics (STEM) by individuals of all ages and backgrounds through self-directed learning experiences” (Ucko, 2008: 9)

Dierking (2008:19), a contributor to the US Framework for Evaluating Impacts of Informal Science Education Projectsin Friedman (2008), suggests that a project should be able to answer at least the following questions at the outset of initiating a project:

  1. What audience impacts will this project facilitate?
  2. What approach/type of project will best enable us to accomplish these goals and why do we feel that this is the best approach to take?
  3. How will we know whether the activities of the project accomplished these intended goals and objectives and with what evidence will we support the assertion that they did?
  4. How will we ensure that unanticipated outcomes are also documented?

This US "Informal education and outreach framework" (Uko 2008:11, Dierking 2008:21) identifies six impact categories with respect to both public audiences and professional audiences.

The Informal Education and Outreach Framework (Uko 2008:11, Dierking 2008:21)
Impact Category Public Audiences Professional Audiences Generic Definition
Awareness, knowledge or understanding (of) STEM concepts, processes, or careers Informal STEM education/outreach research or practice. Measurable demonstration of assessment of, change in, or exercise of awareness, knowledge, understanding of a particular scientific topic, concept, phenomena, theory, or careers central to the project.
Engagement or interest (in) STEM concepts, processes, or careers Advancing informal STEM education/outreach field Measurable demonstration of assessment of, change in, or exercise of engagement/interest in a particular scientific topic, concept, phenomena, theory, or careers central to the project.
Attitude (towards) STEM-related topic or capabilities Informal STEM education/outreach research or practice Measurable demonstration of assessment of, change in, or exercise of attitude toward a particular scientific topic, concept, phenomena, theory, or careers central to the project or one’s capabilities relative to these areas. Although similar to awareness/interest/engagement, attitudes refer to changes in relatively stable, more intractable constructs such as empathy for animals and their habitats, appreciation for the role of scientists in society or attitudes toward stem cell research.
Behavior (related to) STEM concepts, processes, or careers Informal STEM education/outreach research or practice Measurable demonstration of assessment of, change in, or exercise of behavior related to a STEM topic. These types of impacts are particularly relevant to projects that are environmental in nature or have some kind of a health science focus since action is a desired outcome.
Skills (based on) STEM concepts, processes, or careers Informal STEM education/outreach research or practice Measurable demonstration of the development and/or reinforcement of skills, either entirely new ones or the reinforcement, even practice, of developing skills. These tend to be procedural aspects of knowing, as opposed to the more declarative aspects of knowledge impacts. Although they can sometimes manifest as engagement, typically observed skills include a level of depth and skill such as engaging in scientific inquiry skills (observing, classifying, exploring, questioning, predicting, or experimenting), as well as developing/practicing very specific skills related to the use of scientific instruments and devices (e.g. using microscopes or telescopes successfully).
Other Project specific Project specific Project specific

From this table, Dierking (2008:23) then derives a simple worksheet for Developing Intended Impacts, Indicators & Evidence

worksheet for Developing Intended Impacts, Indicators & Evidence
ISE Category of Impact Potential indicators Evidence that impact was attained
Awareness, knowledge or understanding of STEM concepts, processes or careers
Engagement or interest in STEM concepts, processes, or careers
Attitude towards STEM-related topics or capabilities
Behavior resulting from experience
Skills based on experience
Other (describe)

Software

Guides

Robert D. Stevenson et al. (2003) at the University of Massachusetts have developed open source software tools that enable nonexperts to produce customised field guides. Field guides are used to identify species, a task that is common to many projects.

  • R.D. Stevenson, et al. Electronic field guides and user communities in the eco-informatics revolution. Conserv. Ecol., 7 (2003), p. 3.

Data collection

Pathfinder (Luther et al. 2009: abstract) “is an online environment that challenges this traditional division of labor by providing tools for citizen scientists to collaboratively discuss and analyze the data they collect.”. The design of pathfinder was inspired by three main phenomena: Complex online participation (e.g. Wikipedia or OSS), Argument mapping systems such as Compendium and social data analysis (SDA), e.g. the sense.us system. There are two main components to Pathfinder: tracks (data sets) and discussions. Both are linked. In addition, pathfinder allows to annotate a pattern or trend found in data set as finding. Finally, every data set (trac)k page also contains comments and a basic wiki. The latter can be used for general casual discussions. “The discussions component of Pathfinder allows users to engage in collaborative, structured analyses around tracks. Each discussion page is organized into three sections: the topic, responses to the topic, and an overview of these responses, displayed below the topic but before the responses themselves” (Luther et al., 2009: 242). Responses can be tagged as so-called "Milestones", i.e. Questions, Hypotheses, Evidence (Pro or Con), Background, Prediction, Conclusion, or To-Do.

Ellul et al (2011) developed a Flexible Database-Centric Platform for Citizen Science Data Capture.


Links

General

  • Citizen Science (Scientific American). Also includes a larger list of projects.

Organizations

Index pages

On-line environments

Evaluation

Bibliography

  • Ballard, H., Pilz, D., Jones, E.T., and Getz, C. (2005). Training Curriculum for Scientists and Managers: Broadening Participation in Biological Monitoring. Corvalis, OR: Institute for Culture and Ecology.
  • Baretto, C., Fastovsky, D. and Sheehan, P. (2003). A Model for Integrating the Public into Scientific Research. Journal of Geoscience Education. 50 (1). p. 71-75.
  • Bauer, M., Petkova, K., and Boyadjieva, P. (2000). Public Knowledge of and Attitudes to Science: Alternative Measures That May End the "Science War". Science Technology and Human Values. 25 (1). p. 30-51.
  • Bonney, R. 2004. Understanding the process of research. Chapter 12 in D. Chittenden, G. Farmelo, and B. Lewenstein, editors. Creating connections: museums and public understanding of current research. Altamira Press, California, USA.
  • Bonney, R. and LaBranche, M. (2004). Citizen Science: Involving the Public in Research. ASTC Dimensions. May/June 2004, p. 13.
  • Bonney, R., Cooper, C.B., Dickinson, J., Kelling, S., Phillips, T., Rosenberg, K.V. and Shirk, J. (2009). Citizen Science: A Developing Tool for Expanding Science Knowledge and Scientific Literacy. BioScience. 59 (11). P. 977-984.
  • Bonney, R., and A. A. Dhondt. 1997. Project FeederWatch. Chapter 3 in K. C. Cohen, editor. Internet links to science education: student scientist partnerships. Plenum Press, New York, New York, USA.
  • Bouwen, R., and T. Taillieu. 2004. Multi-party collaboration as social learning for interdependence: developing relational knowing for sustainable natural resource management. Journal of Community and Applied Social Psychology 14:137–153.
  • Bracey, G.L. 2009. “The developing field of citizen science: A review of the literature.” Submitted to Science Education. (not published yet as of 1/2011 ?)
  • Bradford, B. M., and G. D. Israel. 2004. Evaluating volunteer motivation for sea turtle conservation in Florida. University of Florida: Agriculture Education and Communication Department, Institute of Agriculture and Food Sciences, AEC 372. http://edis.ifas.ufl.edu.
  • Brossard, D., Lewenstein, B., and Bonney, R. (2005). Scientific Knowledge and Attitude Change: The Impact of a Citizen Science Project. International Journal of Science Education. 27 (9). p. 1099-1121.
  • Bryant, S.L., Forte, A., and Bruckman, A. Becoming Wikipedian: transformation of participation in a collaborative online encyclopedia. Proc. SIGGROUP 2005, ACM (2005), 1-10.
  • Cooper, Seth, Firas Khatib, Adrien Treuille, Janos Barbero, Jeehyung Lee, Michael Beenen, Andrew Leaver-Fay, David Baker, Zoran Popović and Foldit players. Predicting protein structures with a multiplayer online games. Nature 466, 756-760 (2010).
  • Ellul, C., Francis, L. and Haklay, M., 2011, A Flexible Database-Centric Platform for Citizen Science Data Capture, Computing for Citizen Science Workshop, in Proceedings of the 2011 Seventh IEEE International Conference on eScience (eScience 2011)
  • Evans, C., E. Abrams, R. Reitsma, K. Roux, L. Salmonsen, and P. P. Marra. 2005. The neighborhood nestwatch program: participant outcomes of a citizen-science ecological research project. Conservation Biology 19:589–594.
  • Falk, J., (Ed). (2001). Free-choice science education: How we learn science outside of school. New York: Teachers College Press.
  • Fernandez-Gimenez, M. E., Ballard, H. L., and Sturtevant, V. E. (2008). Adaptive management and social learning in collaborative and community-based monitoring: A study of five community-based forestry organizations in the western USA. Ecology and Society, 13(2). http://www.ecologyandsociety.org/vol13/iss2/art4/
  • Firehock, K. and West, J. (2001). A brief history of volunteer biological water monitoring using macroinvertebrates. Journal of the North American Benthological Society. 14 (2) p. 197-202.
  • Hand, Eric (2010). "Citizen science: People power". Nature 466, 685-687
  • Hartanto, H., M. C. B. Lorenzo, and A. L. Frio. 2002. Collective action and learning in developing a local monitoring system. International Forestry Review 4:184–195.
  • Herzberg, F., Mausner, B., and Snyderman, B. The Motivation to Work. (2nd rev. ed.) New York: Wiley, 1959.
  • Khatib, Firas; Seth Cooper, Michael D. Tyka, Kefan Xu, Ilya Makedon, Zoran Popović, David Baker, and Foldit Players. Algorithm discovery by protein folding game players. In Proceedings of the National Academy of Sciences (2011).
  • King, K., C.V. Lynch 1998. “The motivation of volunteers in the nature conservancy - Ohio chapter, a non-profit environmental organization.” Journal of Volunteer Administration 16, (2): 5.
  • Kolb, D. A. 1984. Experiential learning: experience as the source of learning and development. Prentice-Hall, Englewood Hills, New Jersey, USA.
  • Kountoupes, D. and Oberhauser, K. S. (2008). Citizen science and youth audiences: Educational outcomes of the Monarch Larva Monitoring Project. Journal of Community Engagement and Scholarship, 1(1):10–20.
  • Krasny, M., and R. Bonney. 2005. Environmental education through citizen science and paticipatory action research. Chapter 13 in E. A. Johnson and M. J. Mappin, editors. Environmental education or advocacy: perspectives of ecology and education in environmental education. Cambridge University Press, Cambridge, UK.
  • Lave, J. and Wenger, E. Situated Learning: Legitimate Peripheral Participation. Cambridge UP, 1991.
  • Lewenstein 2001. PIPE Evaluation Report. Year 2, 1999–2000. Department of Communications, Cornell University, Ithaca, New York, USA.
  • Luther, Kurt; Scott Counts, Kristin B. Stecher, Aaron Hoff, and Paul Johns. 2009. Pathfinder: an online collaboration environment for citizen scientists. In Proceedings of the 27th international conference on Human factors in computing systems (CHI '09). ACM, New York, NY, USA, 239-248. DOI=10.1145/1518701.1518741 http://doi.acm.org/10.1145/1518701.1518741
  • McCurley, S. & Lynch, R. (1997). Volunteer management. Downers Grove, Illinois: Heritage Arts.
  • National Research Council. (2009). Learning science in informal environments: People. places, and pursuits. Committee on Learning Science in Informal Environments, P. Bell, B. Lewenstein, A. W. Shouse, and M. A. Feder (Eds.) Board on Science Education, Center for Education, Division of Behavioral and Social Sciences and Education. Washington, DC: The National Academies Press.
  • Osborn, D., Pearse, J. and Roe, A. Monitoring Rocky Intertidal Shorelines: A Role for the Public in Resource Management. In California and the World Ocean: Revisiting and Revising California's Ocean Agenda. Magoon, O., Converse, H., Baird, B., Jines, B, and Miller-Henson, M., Eds. p. 624-636. Reston, VA: ASCE.
  • Raddick, M. Jordan; Georgia Bracey, Pamela L. Gay, Chris J. Lintott, Phil Murray, Kevin Schawinski, Alexander S. Szalay, and Jan Vandenberg (2010). Galaxy Zoo: Exploring the Motivations of Citizen Science Volunteers, Astronomy Education Review 9, 010103 (2010), DOI:10.3847/AER2009036 - PDF Preprint
  • Silvertown, J. (2009). A New Dawn for Citizen Science. Trends in Ecology & Evolution. 24 (9). p. 467-471
  • Spiro, M. (2004). What should the citizen know about science? Journal of the Royal Society of Medicine, 97 (1).
  • Trumbull, D. J., R. Bonney, D. Bascom, and A. Cabral. 2000. Thinking scientifically during participation in a citizen-science project. Science Education 84:265–27.
  • Trumbull, D. J., R. Bonney, and N. Grudens-Schuck. 2005. Developing materials to promote inquiry: lessons learned. Science Education 89(6):879–900.
  • Wiggins, A. and Crowston, K. (2010) ‘Developing a conceptual model of virtual organisations for citizen science’, Int. J. Organisational Design and Engineering, Vol. 1, Nos. 1/2, pp.148–162.
  • Wiggins, A., & Crowston K. (2011). From Conservation to Crowdsourcing: A Typology of Citizen Science. Proceedings of the Forty-fourth Hawai'i International Conference on System Science (HICSS-44).
  • Wiggins, A., & Crowston K. (2012). Goals and tasks: Two typologies of citizen science projects. Forty-fifth Hawai’i International Conference on System Science (HICSS-45)
  • Ye, Y. and Kishida, K. Toward an understanding of the motivation Open Source Software developers. Proc. ICSE 2003, IEEE (2003), 419-429.