Leveraging Peer Review in Visualization Education: A Proposal for a New Model
Abstract
In visualization education, both science and humanities, the literature is often divided into two parts: the design aspect and the analysis of the visualization. However, we find limited discussion on how to motivate and engage visualization students in the classroom. In the field of Writing Studies, researchers develop tools and frameworks for student peer review of writing. Based on the literature review from the field of Writing Studies, this paper proposes a new framework to implement visualization peer review in the classroom to engage today’s students. This framework can be customized for incremental and double-blind review to inspire students and reinforce critical thinking about visualization.
\ul
Keywords: Visualization education, peer review, student motivation, new model of visual peer review.
I Introduction
Rushmeier et. al [1], defined visual education as a work in progress. Meanwhile, the number of students enrolled in visualization courses, both face-to-face and online, has grown dramatically in recent years—a growth that makes it difficult to provide students the feedback they need to improve the quality of their work.
Visualization evaluation has been addressed by a number of researchers [2, 3, 4]. While many techniques and algorithms can be objectively measured, most recognize that there are also significant subjective measurements necessary to completely evaluate the effectiveness of a visualization [5]. Such forms of subjective measurement are not just important for the instructor’s assessment of students, but the students also need the ability to critically evaluate a visualization.
For practical purposes, visualization education can be split into two categories. The first is learning the proper construction of visualization—using the right visualization principals in creating your own visualization. This tends to be the primary focus of visualization courses. In an era with increasing proliferation of visualization in science and mass media, the second category is focused on evaluation of the quality and accuracy of visualizations in general. These skills tend to be taught through less formal methods, such as group or class discussion. Less formal methods, while critical, leave students’ skills underdeveloped.
To this end, we propose a framework for engaging students in peer review as part of their formal visualization coursework. This framework contains 2 parts. First, we provide a reconfigurable assessment rubric for students to learn how to properly peer review visualizations. Second, we provide a step-by-step approach to educate students in proper evaluation techniques by gradually reducing the constraints on their peer reviews over the course of a semester.
Ii Education and peer review
Assessment is the heart of formal higher education. Bransford et. al [6] defines assessment as a core component for effective learning. Student assessment has many forms in higher education, and one of these forms, used most commonly in the liberal and creative arts, is peer review. Peer review was established as an essential component for many professional practices, such as the scholarly publication process. The fundamental principle for peer review is that experts in any given domain appraise the professional performance, creativity, or quality of scientific work produced by others in their area of competence. Peer review is often defined as the evaluation of work by one or more people of similar competence to the producers of the work [7]. The principle component of student peer review is assessment of students by other students, both formative reviews to provide feedback and summative grading.
Iii Moxley Rubric
Moxley [8] reports that the use of peer review technologies is changing the ecology of assessment and student motivation. “Rather than happening after the fact, online assessment systems are becoming part of the teaching and learning process, part of the digital experience that students are so motivated to be part of.” Moxley utilizes this rubric to develop online peer review, where the instructor can choose between two versions of the common rubric: (1) the numeric rubric, which requires students to score rubric criteria (Focus, Evidence, Organization, Style, and Format) based on a five-point scale; and (2) the discussion rubric, which requires students to write textual comments regarding these criteria rather than scores. While the textual feedback provides more detail, research has shown that instructors have favored the numeric version of the rubric over the discussion version [8]. A reproduction of the rubric is seen in Figure 1.
Iv Background in Visualization Education
In 2006, the ACM Siggraph Education Committee started the discussion on what visualization education should be. Domik [9] recaptured this discussion by posting the question: “Do we need formal education in visualization?” She answered this question by stating that we do at least need informal visualization education. In the 2000s, humanities departments started to offer visualization courses to their students, and as a result, Pat Hanrahan proposed implementation of core topics to fit the humanities, including: data and image models; perception and cognition; interaction; space; color; shape and lines; texture; interior structure with volumetric techniques; and narrative with animation [10].
For this study, we embraced Hanrahan’s core topics in our classes, due in part to the different backgrounds our students bring to our classes. Understanding the breadth of these topic areas is not necessarily sufficient to be considered skilled in visualization. Gilbert [11] described 5 levels of competency that give insight to the depth of topical understanding in visualization. His method outlines step-by-step instruction on visualization design that includes: step 1, representation as depiction; step 2, converting early symbolic skills; step 3, the syntactic use of formal representation; step 4, semantic use of formal representations; and step 5, reflective, rhetorical use of presentation. However, we did not find any discussion or references on how to embed these steps into the analysis of visualization or peer review of visualization.
V Approach
In 2013 and 2014, the University of South Florida hired two faculty into the College of Arts and Sciences and College of Engineering, respectively, to build visualization into their curriculum. This effort was taking place in the context of a larger university-led effort towards improving visualization skills and access to visualization equipment. Both faculty focus on teaching visualization as any technique for creating images, diagrams, or animations to communicate a message using different platforms and programming languages.
In the School of Information under the College of Arts and Sciences, a course called “Visualization of Big Data” was offered. The class is part of new area of concentration in Data Science and was offered to students via an online setting only. In the College of Engineering the Department of Computer Science and Engineering offered the class called “Scientific Visualization.” The class was offered in traditional classroom format only as an elective. The key differences between these courses were format (i.e., online vs. classroom) and emphasis (i.e., design for the Information Science course and algorithmic for the Computer Science course). In both classes the instructors focused not just on design and algorithmic technique, but they also focused on the analysis of visualizations using models such as the “what, why, and how” model of Liu and Stasko [12] as a framework for discussion.
During these 2 courses, we employed several peer review techniques. We collected the students’ peer review comments with the hope to find patterns. Our premises for the investigation was based on the three principals of Monk et al. [13]:
-
Careless mapping from data to pictures may lead to erroneous interpretation.
-
A substantial amount of knowledge is necessary to generate images depicting complex information in a way that prevents erroneous interpretation.
-
Decision making is increasingly based on visual representations.

Vi Results
We did not find a single student who directly refereed to the Monk’s principles in their peer review, as they were covered in class and in their assignments. Many of the students selected one aspect of visualization without considering the other aspects that are involved in creating and evaluating visualization work. For example, many students simply ignored considerations such as the variables embedded, visual patterns, or time. To illustrate our findings, the following is a common example from a student peer review. The student in this case was asked to review their peer’s histogram created using R, based on Monk’s principles. The histogram had to be displayed in multiple colors as part of the assignment.
This is the first time in my life I had to provide comments and criticism on my peersâ work using open-source R. I certainly had my struggles with R to create visualization, but in this assignment, I have problems understanding to the complexity of visualization and visualization creation using R.
Furthermore, many of the comments we examined did not have any direct connection to the content of the visualization itself. Thus, we propose a new framework to better teach the evaluation of visualizations.
Vii Proposed Framework
In the field of Writing Studies, researchers have reported on the practice of using generic rubrics to make global claims about writing development. Beaufort [14], Nowacek [15], and Wardle and Roozen [16] have explored how the ability to write well is a skill that can be mastered in one context and then simply carried over to another context. Moxley [8] opposed this approach stating that arguments that make grand claims about student ability based on a handful of rubric scores need to be seriously challenged. Students’ scores on one rubric are not necessarily predictive of how they will do when facing alternative genres. To address the limited availability of visualization peer review tools and techniques in the classroom, we propose the following framework adhering to Moxley’s recommendation.
Vii-a Peer Review Rubric for visualization education
The basic structure of our rubric is to divide topics into 5 major assessment categories, with each category having sub-assessments. The sub-assessments contain 3 scoring categories affixed to a 5-point scale (though no scale is shown). The category on the left is reserved for poor performance on the assessment, middle for average performance, and right for excellent performance. Each sub-assessment contains a comments box for providing details about the scoring.
For our rubric, the 5 major assessment categories are: algorithm design, interaction design, visual design, design consideration, and visualization narrative. The categories cover the following subjects: The algorithm design category is concerned with algorithm selection and implementation. Interaction design is concerned with how the user interacts with the visualization. Visual design is concerned with the technical aspects of how data are placed in the visualization. Examples include, visual encoding channels, their expressiveness, and their effectiveness. Design consideration focuses on composition and aesthetics aspects of the visualization, such as embellishments. The final category visualization narrative is critical in projects where the story surrounding the visualization is as important as the visualization itself. This category provides a basic framework for assessing whether the visualization supports the story and vice versa.
Vii-B Customize Rubric
We also consider customization of the rubric, so it can be built based upon the content of an assignment or customized to fit the course content. Instructors should take our full template, extract the critical components, and add missing components with respect to their assignment and course content. For example, on a computer science assignment, the rubric provides the ability to remove or reduce the number of components dealing with aesthetics and narrative, while emphasizing those for algorithm. On the other hand, for projects that do not implement an algorithm or those for which the algorithm is not critical may skip this category.
Another aspect in which this rubric can also be customized is to reduce the constraints of the rubric itself. Assuming the critical thinking skills of students, particularly in the topic of visualization, are weak, the rubric itself can be a tool to help improve those skills. At the beginning of the course, the rubric can be provided that includes all scoring categories to build an analytic framework in the students’ minds. As their skills improve, the 5-point scoring scale can be removed, leaving only the comments section behind. Finally, as students begin to perfect their skills, the sub-assessments can be removed altogether, leaving only the major categories of evaluation. In this way, students will go from highly constrained and guided evaluation to free-form evaluation over the course of a semester.
Finally, we foresee the peer review process appearing much like conference or journal reviews. First, double-blind review helps to reduce the risk of collusion or malicious review. Second, multiple reviews per visualization help to ensure that consensus is built around the evaluation. In addition, each student having to produce multiple reviews helps to reinforce the critical thinking skills. Some of these functionalities are available in learning management systems, such as Canvas.
Viii Risks in Peer Review
As we implement this framework in our own classrooms, we are keeping an eye out for a couple of key issues. For example, peer review is a compliment to existing educational approaches, but the instructor, as the expert, still needs to provide feedback. Otherwise, some important subtleties may be missed. Second, the risk for collusion or malice are great, even with double-blind reviews. Students talk, and when they talk, they will undoubtedly discover they are evaluating each other’s visualizations. Therefore, instructors should take care to randomize peer reviewers, and grades should only be loosely based upon the results of the peer review. It is also important to remember that there is both art and science to design with no single optimal design. Therefore, a peer review may not necessarily be the best mechanism for providing grades. However, peer review naturally supports iterative design improvements [5].
Ix Conclusion
We have presented a new framework for evaluating student visualizations based upon peer review rubric. This rubric has a couple of important aspects. First, it provided students a framework for critical thinking and evaluation of visualizations. Second, it provides a mechanism to help with issues of larger class sizes, while still providing students feedback about their work.
Finally, we do not believe that the rubric as presented will be the final static version. We anticipate it will be a growing and evolving document as community members provide their input and the focus of the visualization community changes. As a result, Appendices A-E represents our proposal for student peer review rubrics based on specific subjects from algorithmic design, interaction design, visual design, design consideration, and visualization narrative. For the full proposed visualization rubric, see Appendix A-E. A LaTeXversion of the rubric can be cloned/forked at https://github.com/USFDataVisualization/VisPeerReview.
Appendix
|
□
|
□ | □ Average | □ | □
|
||||||||||||
Comments: | |||||||||||||||||
|
□ No | □ | □
|
□ | □
|
||||||||||||
Comments: | |||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||
Comments: | |||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||
Comments: | |||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||
Comments: |
|
|
|||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Comments: | ||||||||||||||||||
|
□
|
□ | □ Expected | □ | □
|
|||||||||||||
Comments: |
|
|
||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Comments: | |||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□ No | □ | □
|
□ | □ Yes | ||||||||||||||||||
Comments: | |||||||||||||||||||||||
|
□ No | □ | □
|
□ | □ Redundant | ||||||||||||||||||
Comments: |
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
|
□ |
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
|
□ |
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: | |||||||||||||||||||||||||
|
□
|
□ | □
|
□ | □
|
||||||||||||||||||||
Comments: |
|
□
|
□ |
|
□ | □
|
|||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Comments: | ||||||||||||||
|
□
|
□ |
|
□ | □
|
|||||||||
Comments: | ||||||||||||||
|
□
|
□ | □ Partially | □ | □ Completely | |||||||||
Comments: |
References
- H. Rushmeier, J. Dykes, J. Dill, and P. Yoon, “Revisiting the need for formal education in visualization,” IEEE Computer Graphics and Applications, vol. 27, no. 6, 2007.
- A. Shamim, V. Balakrishnan, and M. Tahir, “Evaluation of opinion visualization techniques,” Information Visualization, vol. 14, no. 4, 2014.
- M. J. Eppler, “A comparison between concept maps, mind maps, conceptual diagrams, and visual metaphors as complementary tools for knowledge construction and sharing,” Information visualization, vol. 5, no. 3, pp. 202–210, 2006.
- N. JD and A. Cañas, “The origins of the concept mapping tool and the continuing evolution of the tool,” Information Visualization, vol. 5, no. 3, 2006.
- F. Viégas and M. Wattenberg, “Design and redesign in data visualization,” March 2015. [Online]. Available: https://medium.com/@hint_fm/design-and-redesign-4ab77206cf9
- N. R. Council et al., How people learn: Brain, mind, experience, and school: Expanded edition. National Academies Press, 2000.
- R. Spier, “The history of the peer-review process,” TRENDS in Biotechnology, vol. 20, no. 8, pp. 357–358, 2002.
- J. Moxley, “Big data, learning analytics, and social assessment,” Journal of Writing Assessment, vol. 6, no. 1, 2013.
- G. Domik, “Do we need formal education in visualization?” IEEE Computer Graphics and Applications, vol. 20, no. 4, pp. 16–19, 2000.
- K.-L. Ma, “Teaching visualization,” ACM Siggraph Computer Graphics, vol. 39, no. 1, pp. 4–5, 2005.
- J. K. Gilbert, “Visualization: An emergent field of practice and enquiry in science education,” Visualization: Theory and practice in science education, pp. 3–24, 2008.
- Z. Liu and J. Stasko, “Theories in information visualization: What, why and how,” in Workshop on the Role of Theory in Information Visualization, 2010.
- A. Monk, M. Hassenzahl, M. Blythe, and D. Reed, “Funology: designing enjoyment,” in CHI’02 Extended Abstracts on Human Factors in Computing Systems. ACM, 2002, pp. 924–925.
- A. Beaufort, College writing and beyond: A new framework for university writing instruction. University Press of Colorado, 2008.
- R. S. Nowacek, Agents of integration: Understanding transfer as a rhetorical act. SIU Press, 2011.
- E. Wardle and K. Roozen, “Addressing the complexity of writing development: Toward an ecological model of assessment,” Assessing Writing, vol. 17, no. 2, pp. 106–119, 2012.
