|
Post by fannie on Apr 10, 2016 0:54:52 GMT
[written jointly by Fannie and Judith] SummaryThe authors of this paper present Kinetica, a data visualization system that incorporates physics-based affordances and multi-touch interaction techniques. They explore how these naturalistic interaction techniques, which fall under the post-WIMP (window, icon, menu, and pointer) metaphor, might improve exploratory data visualization. Kinetica:The authors built Kinetica, an Objective C based app deployed on the iPad. They encoded data into circles colliding in a sandbox, by tilting the device so gravity took hold and pulling points with a magnet such that the forces balanced each other, data readily sorted itself and separated, highlighting outliers. Images of the Kinetica can be seen below: User Studies:To evaluate the Kinetica system, the authors conducted users studies comparing Kinetica to Excel. The authors were interested in whether or not participants’ findings would differ based on the software they were using. They had the participants look at two different datasets using either the Kinetica or Excel. The first was additional training, where participants looked at cereal brands and answered questions about them. The second was choosing a car they liked from a dataset of 133 car models. Results:The authors found that Kinetica users were able to build a more holistic understanding of the data than Excel users, as they were able to make more descriptive, comparative, and relationship findings while Excel users mostly made point (e.g. a finding regarding a particular row of data) or statistical findings. Kinetica participants also reported that Kinetica was easy to use, helped to explore and play around with data, and could be used to make informative decisions. Implications:Using their work with Kinetica, the authors developed the following framework for physics-based affordances in multivariate data visualization: The authors identify potential limitations and benefits of the Kinetica system/physics-based affordances in data visualizations: Limitations:
- Scalability: iPad is limited in computational resources, screen real estate, rendering sluggish at > 500 points
- Effect attribution: can be hard to attribute constraints/forces to specific causes when there are too many of them
- Expert usability and overload: experts may want quantitative tools which Kinetica wasn’t great for
- Overconstraining: relationships may not be faithfully represented
Benefits
- Minimal training: easy to use, could be useful for teaching data literacy to young students or visual learners
- Awareness of distribution: users could gain a qualitative understanding of data without needing chart frequencies
- Understanding process: fluidity of the data and physics interactions could help users interpret changes and come up with new ideas for future
- Augmenting working memory: the system may have helped users consider/keep track of more dimensions by leaving traces of interactions, helping users improve their memory of the current state of data and past actions
Thought Questions:Here are a few thought questions to get you guys started: - [From existing slides] What kinds of data could you use this tool on? Do you see any limitations in the kinds of data?
- Using the authors’ framework, can you come up with new tools/applications that use physics-based affordances? Or alternatively, what other HCI applications could naturalistic interactions benefit?
- The authors incorporated physics-based affordances to their visualization by using collisions and a magnet tool. Can you think of other physics or natural interaction based affordances that can enhance data visualization?
- The authors argued for the use of multi-touch gestures for interacting with the data and used either one or two finger gestures for drawing histograms, boundaries, lenses, and selection. How do you think this compares to the traditional WIMP metaphor? Can you think of other multi-touch gestures that could have enhanced the experience?
|
|
|
Post by julian on Apr 10, 2016 20:27:33 GMT
I think what I like the most about this reading, although I think is not highlighted enough in the paper, is on the role of Kinetica towards attention guidance provided by adding the Physics component of the visualization. By adding Physics, many participants were observing how data points were clustering themselves into groups, by observing the trajectories of these points, participants became aware of a reason for data points to belong together and the groups themselves. Basically, this dynamic visualization helped reveal aspects of a visualization that normally may have gone unnoticed (I’m comparing against just showing a scatter plot with zero dynamics). However (according to the authors), this comes at the expense of being able to show only few(500) data points, which I think is not a problem, for example the user could always first sample the data set and then use Kinetica. The sample will have statistical properties similar to those of the entire data set with the added benefit of the physics manipulations. I think another aspect the paper overlooks is that of being able to show both the information the user cares about and information that does not(immediately) care about but could be of importance. For instance in Figure 1, the filter example shows both the data that is being filtered out and the one that does not pass. The last usually is overlooked by pretty much any filter in any software package I can think of, while I do not know immediately how to use that information I would think it is useful and it comes at almost no expense. One of the participants I believe mentioned this feature was very useful at identifying he/she had made a mistake with the filter which was not having the desired effect on the data. Last, this work would be awesome for doing fast data exploration, for instance, learning very quickly about an unfamiliar data set.
|
|
mkery
New Member
Posts: 22
|
Post by mkery on Apr 11, 2016 1:04:53 GMT
A participant in the paper mentions that Kinetica could be used to teach young children data literacy. I’m excited by that idea, since in grade school I recall many lessons on how to make a bar chart or a scatterplot (hours of hand drawing graphs for homework), but with this focus on the mechanics of graph-making, I gained little sense of learning how to work with data or how to think about statistics. Kinetica is thus interesting as an education tool because of it’s ease of quick exploration and potentially, has a sense of play. The tangible physics metaphor reminds me of Nesra’s work “Tangible Collaborative Learning with a Mixed-Reality Game: EarthShake” However, access would be a tremendous issue, classes+iPads. Are there ways to translate this software into a WIMP non-touch application without utterly disrupting the physics metaphor?
|
|
|
Post by sciutoalex on Apr 11, 2016 16:09:13 GMT
I'd like to pickup Mary Beth's question about translating the application to a non-touch device. I came to the conclusion through reading the paper and looking at the demos that the multi-touch interface was not the essential part of the application that resulted in improved performance. Instead, it was the physics simulation and assorted tools. I'm reminded of browser-based before-and-after photo viewers that effectively utilize dragging back and forth to give users a sense of a scene from two different moments. This interaction is a very basic physics simulation and it is effective. I agree that there are some interactions—expanding a circular filter comes to mind—that would be difficult to replicate with a mouse, but I think a click (define center) and drag (define radius) would works almost as well.
I'm trying to think of other metaphors one could use as the basis for interacting with data. I remember the sedimentation metaphor (http://www.visualsedimentation.org/examples/sedivn/sedivn.html) that was moderately successful at showing real-time data. I also think trees and plant growth would be useful to show growth. Flow (like wind or a river) could be a metaphor to show how particular events change a pre-existing flow of data. These seem to be more opinionated metaphors than Kinetica's.
|
|
|
Post by JoselynMcD on Apr 11, 2016 16:56:15 GMT
RE: What kinds of data could you use this tool on? Do you see any limitations in the kinds of data?
I thought this was a very interesting research particularly because of a rather stress-inducing study I found demonstrating that 1/3 of US citizens and Germans had significant trouble with graph literacy and interpretation (http://mdm.sagepub.com/content/31/3/444.short). To that end, I think a tool like Kinetica is relevant and could be potentially beneficial as an alternative way for novices to interpret data. I wasn't compelled by the use of physics, or rather I'm not convinced that it's necessary to get the same effect.
This study relates to some of my earlier research on abstract data visualizations of personal data in order to provide people with low graph literacy feedback that wasn't difficult to interpret or stress-inducing. We moved onto another study area for a myriad number of reasons, but in reflecting on the potential limitations of this kind of tool, I am reminded of a sticking-point we kept coming back to: could the metaphor potentially affect the interpreation of the data in a way that is not faithful to what is presented. I didn't perceive that Kinetica's experience was doing this, but if this work were to be expanded, I think it's worthy of further research.
|
|
|
Post by stdang on Apr 11, 2016 18:27:29 GMT
Replying to Question #2, I think that the notion of physics-derived operators along the lines of the framework created could be combined with spatial operators for interesting hybrid data analysis and visualization techniques. For instance, while visualizing tweets within Kinectica, you could create barriers to segregate spatial areas and use forces to cluster data within each spatial area. Within a spatial area you can apply either a similar wordcloud operator to generate word clouds based on which data points are present so as to quickly see how your clustering might make semantic sense. Alternatively you could apply different operators within each space such as a different topic model that is intended to model slightly different social phenomena. Thus you can create a spatial arrangement for political tweets that visualizes the top 3 mentioned topics in the political space, while capturing the most retweeted topics in the pop music/culture space. This form of visualization seems less useful for rapid analytics, but instead rapid dashboard prototyping/configuration.
|
|
|
Post by Cole on Apr 11, 2016 19:19:40 GMT
I agree that this could be great for teaching statistical concepts, since most statistics graphs are not as interactive and intuitive as Kinetica seems to be. I would also love an interface like this for some tasks, as there are many cases where a Google search isn't enough control over the data, but the Excel spreadsheet interface is too much work.
I would be very interested to see how an interface like Kinetica could be tied into data journalism. News stories that let readers interpret the data themselves have become more popular, especially on sites like the New York Times or Five Thirty Eight. Because many visualizations used require some expertise to correctly interpret, natural physics affordances could be a perfect interface to the data. I would love to see some news stories embed Kinetica online (since I also don't think the multi-touch gestures are critical, just nice).
|
|
aato
New Member
Posts: 16
|
Post by aato on Apr 11, 2016 20:39:45 GMT
Re: The authors incorporated physics-based affordances to their visualization by using collisions and a magnet tool. Can you think of other physics or natural interaction based affordances that can enhance data visualization? There was a really cool paper at CHI 2015 that falls into this same category of 'natural' interactions with data, "Exploring Interactions with Physically Dynamic Bar Charts." They push the metaphor of natural interaction much further, but in doing so seem to amplify both the benefits and the limitations of Kinetica. Users can tap different areas of the 10x10 3D bar graph to annotate data, hide or highlight sections, and sort (I think?) data if they wish. I think there's an idea here that comes from Kinetica that is really interesting - that you can engage different types of learners using different perceptual and attentional cues to help make information and data more salient? This kind of digitally-augmented physical representation of data is really cool and maybe over time further developers can use the digital augmentation to solve some of the scale and processing limitations. Paper Link: dl.acm.org/citation.cfm?id=2702604Video link: www.youtube.com/watch?v=TIKuE6GT_P4
|
|
|
Post by Anna on Apr 11, 2016 22:52:29 GMT
Alexandra's point about engaging different types of learners using different perceptual/attentional cues is well put. Also, in my last post (in response to the graphical perception paper), I basically said graphs mean nothing to me. But actually, I find Kinetica really exciting and cool-- I remember seeing Jeff give a demo of it pretty early on in my first semester in HCI, I believe. With regards to the affordances of Kinetica/the limits of the iPad- discussed above- it does seem that having some more direct, kinesthetic touch aspect is crucial to the enjoyment of Kinetica, even if it doesn't directly impact users' understanding. I would hypothesize that Kinetica users' increased understanding at least partially stems from the enjoyability of the system-- because they enjoyed using the system more, they were perhaps more attentive to and thinking more critically about the data.
All this said, I'm still a little skeptical-- how much do I even really gain from understanding data better? Obviously, context matters (aka personally, maybe I don't really care about understanding a lot of data when it comes to choosing a car, but I would want to have a deeper understanding of a data-based topic I'm curious about such as the Titanic example).
But as we accumulate more and more data, how relevant is data visualization (referring also to Julian's point that the system can't handle above 500 data points)? I feel like now that we have more data, we're like cool, let's keep visualizing it in novel and more interesting ways so we can understand it better. But is that what we should be doing? Like, instead of trying to make sense of big data, should we be distributing the understanding of small subsets of big data? I feel like these are to a certain extent ridiculous/impractical questions, but they're also kind of real to me.
|
|
|
Post by jseering on Apr 12, 2016 0:13:27 GMT
I think there are really two points here; some of us have been talking about this in the context of making visualizations that increase our understanding of the data, which is obviously important, but there's also the educational value of simply getting somebody engaged in the data regardless of whether a specific visualization actually increases understanding. Like Mary Beth, I'm excited about this as a way to teach data literacy. There are obviously barriers to adoption, and cost is likely prohibitive for most of the educational environments where it would make the most difference, but I think there's a larger point to be made about playing with your data. While the ~kinetic component of this is interesting, I think the most interesting concept is simply the freedom it gives users to explore data in a semi-freeform way. I agree with Alex in that this is quite separable from the kinetic approach, but I would go further. What types of interfaces or apps could be designed that help students learn to enjoy playing with data?
|
|
|
Post by francesx on Apr 12, 2016 1:03:24 GMT
I see an entire space of possibilities to use this tool in learning sciences, specifically for allowing students to monitor their own learning, progress and performance in an Intelligent Tutoring System or in a platform without teacher feedback such as in MOOCS. Despite ITSs being a great platform for learning, students seem disconnected from understanding their own performance or finding ways to pace their work or increase their performance. This is in particular useful to novices, who do not know yet much of the space or domain they are working in. While ITSs produce a wealth of data, and students are not in touch with any of it, a tool such as the one presented in this paper could be the solution (or a solution) to our problems.
|
|
toby
New Member
Posts: 21
|
Post by toby on Apr 12, 2016 1:18:03 GMT
This paper presents a new data visualization system that enables the user to make sense of data more easily by introducing physics-based affordances for the data, at the cost of limited scalability on data size (~500 data points). For the purpose of quick figuring out relations and trends in a dataset, this tool is quite useful. And as Julian said, one can just use Kinetica on a sample of data to avoid the limitation in scalability. But for the use of such system in education, like Mary Beth proposed, I think it would be very interesting and important to explore if the adoption of such system (and the underlying analogy between "data" and physical object) would affect students' ability in data analysis with larger dataset.
An analogy would be, does teaching addition to toddlers as "adding one apple to another apple, you get two apples" makes it harder for them to apply addition on non-apple objects? Or a more complex example, visualization helps us learn 2d and 3d geometry more easily. But does using visualization to learn 2d or 3d geometry makes it more difficult for us to learn high dimensional geometry?
|
|
|
Post by mmadaio on Apr 12, 2016 1:32:34 GMT
Franceska, was that a pun to "put students more in touch with their data"? I also think this is really interesting, as much for its educational and rhetorical power as for its implications for exploratory data analysis. I think the ed psych community has pretty thoroughly debunked the notion of "learning styles" (e.g. visual learner, auditory learner) as having any educational benefit, but I would be intrigued to see how self-professed "kinesthetic" learners either use Kinetica differently than students with other "learning styles" or how it impacts their learning differently, if at all. From a rhetorical perspective, I love that Alex mentioned that some of the physics-based metaphors were more "opinionated" than others. One thing I didn't see in this paper was any sense of how the choice of velocity or weight (or any of the physics attributes for that matter) encoded any semantic properties. For example, in the "flow", "growth", or "sedimentation" metaphors, the choice of object-attribute pairings could communicate very different concepts, with very different effects on the viewer, though that might be more true for a designed visualization than for an exploratory interaction paradigm, like Kinetica. For a class project last year, I made a flocking algorithm to visualize tweets about the UN Climate Rally protest hashtag, clustered by a rudimentary sentiment analysis. It was... a "beautiful disaster", according to Johanna Drucker, but it's one possible way to do this kind of physics-based rhetorical visualization. I did wish there was a more detailed discussion of the mental model hypotheses in the paper. They say they want to "maintain users mental model of the data landscape"... rather than breaking it, leading to uncertainty, but it's not entirely clear that that's what might happen, or that their solution prevents that.
|
|
|
Post by kjholste on Apr 12, 2016 1:52:46 GMT
(0) Joseph wrote: "but there's also the educational value of simply getting somebody engaged in the data regardless of whether a specific visualization actually increases understanding" This is an interesting concept - I wonder how we might characterize the educational value here. Is the idea that this would increase the chances that a student would want to play with data in the future (effectively increasing opportunities for future learning)? Or is there an assumption that, even if the student's understanding of the underlying domain is not increasing, the student is likely to come away with an increased understanding of domain-general properties of data? (1) Above, Anna wrote "...how much do I even really gain from understanding data better?". I share her skepticism about many existing visualization tools that are intended to augment human decision making in real world settings: I would be very interested in seeing empirical studies that attempt to examine whether and how data visualization tools -- especially those intended for use in real-time, or over short time intervals -- actually influence human decision making in complex/noisy environments (e.g. social decision making). ...I've probably phrased this a bit too broadly. But does anyone have recommendations? Anna also wrote: "But as we accumulate more and more data, how relevant is data visualization .... instead of trying to make sense of big data, should we be distributing the understanding of small subsets of big data? I feel like these are to a certain extent ridiculous/impractical questions, but they're also kind of real to me." ^ I think about this all the time too... To me, one of the main purposes of relatively low-level data visualizations (e.g. visualizations presenting data that have not been *heavily* pre-processed and/or pre-interpreted -- pushing more of the interpretation work to the user) is to enable inferences that humans are likely to make (given their rich and structured prior knowledge, which contributes to their inductive reasoning advantages over machines in many learning, reasoning, and decision-making contexts), but that automated pre-processing methods are likely to miss. Another reason (strongly linked to the last point), is to enable more flexible use of data (e.g. for different decision-making contexts) via a single visualization. Pre-processing may remove much of the work of interpretation, while also constraining the way the data can be used. And similarly, visualizing lower-level data may promote trust between the user and the system (even if the presentation of lower-level data does not ultimately result in different inferences or decisions than would have been reached via presentation of more pre-processed data). ...In cases where it may be desirable to visualize less pre-processed data, I'd agree with Julian that it would probably just make sense to sample. ...But I'm really interested in this idea of distributing understanding of small subsets of a larger dataset. Can anyone think of a contexts where this might be beneficial (above and beyond presenting one or more samples [not simultaneously] to an individual?) (2) A while back, I saw an interesting app in the NYT that asked users to draw their best guess of the curve (from a fairly broad space of functional forms) that best characterizes some real-world statistical relationship between two variables: www.nytimes.com/interactive/2015/05/28/upshot/you-draw-it-how-family-income-affects-childrens-college-chances.htmlI wonder whether there have been any empirical studies on the effects of this use of visualization tools as a instructional devices (i.e. having users reveal [or formulate] their preconceptions about some statistical relationship, and then receiving reasonably detailed feedback about how their preconceptions differ from the measured reality)?
|
|
judy
New Member
Posts: 22
|
Post by judy on Apr 12, 2016 1:54:08 GMT
What I find most exciting about Kinetica is the opportunity to play with the data. The physics metaphor gives the user more tangible ways to dig in and manipulate the data. In other words, the user takes an active role in data interpretation. I would call, for example, the magnet metaphor a "mechanic," following Anthropy & Clark's simple definition of 'an action that allows the player to change the state of the game.' Good mechanics, like this magnet, allow the player to make interesting and meaningful choices...
which is why, as many people have mentioned, the "sandbox" quality of Kinetica, makes it particularly suited as a learning tool.
|
|