|
Post by francesx on Mar 27, 2016 0:36:42 GMT
These questions were written by Michael Madaio and Franceska collaboratively. Looking forward to the discussion here and in class!!
- How do you see the results of this paper apply to HCI? What about your own research (research you have done previously)?
- What do you think of the experiment(s) where participants were asked to list only one difference? Would it have been better to list 5 differences (10)?
- How do you think the 5 min time (in the first experiment) could have affected for good/bad the results? Would have been better to have a longer time/no time, or show the participants the pairs and record the time until they gave a response? How would you deal with this topic differently?
- Do you think we might we see different effects for the similarity ratings for different tasks, or if the rater is primed to notice certain features over others? If so, can the findings about structural similarity really be generalizable?
- How do you think the results might change if the task were to list the differences between objects or experiences with some perceptual features instead of listing the differences between words and concepts?
- To what extent are the categories of commonalities and differences they discuss here “surface” features or “deep” features of the words shown?
|
|
|
Post by mrivera on Mar 27, 2016 1:42:06 GMT
1) I can’t justify writing a 3 paragraph response because I don't really get the point of this paper. Maybe I missed something, but what is the generalizable knowledge? I won't know until class when we hear a summary of it . 2) I don't understand why 1 stated difference was enough of an indication for their experiment. 3) Time adds pressure to the task. It probably would have been better to let them just list off all the difference they could about each of a smaller set of words :8. Weekly Snark: “Theories of similarity generally agree that the similar between a pair increases with its commonalities and decreases with its differences.” - www.reactiongifs.com/wp-content/uploads/2014/01/Tom-Hanks-orly.gif
|
|
mkery
New Member
Posts: 22
|
Post by mkery on Mar 27, 2016 2:51:26 GMT
I agree with mrivera that it’s a bit hard to pull generalizations that we can apply to HCI out of this paper without seeing more of how this study fits into the broader space of research on how people think about comparisons/relationships. The author’s findings, however, that people notice more differences between two objects they find similar (objects they can draw a relationship between) is pretty obvious. In HCI practice, this may manifest in trying to make UI components consistent, for example, so that the user can more easily focus on and recall many small differences than being disorientated by large changes.
I worry, however, about trying to draw generalization for HCI out of studies this reductive. This paper studies a narrow, specific perceptual phenomena. Trying to pull too much applicable theme from it, I worry, will dangerously lean towards making specious, non-scientific, and completely opinion-based claims. In HCI we would like to be informed by psychology and cognitive science. How do we do this without completely contorting, misrepresenting or misunderstanding less-applied science?
As for Question 5, I think using words as a test set was one of the more valid, measurable ways for the authors to test their hypothesis. A word carries some depth of meaning, but also an agreed upon dictionary definition. Of course there may be variability of one person’s personal or cultural associations with a word different from another person. However words have pretty standard core meaning. If we open this up to testing people’s comparisons of objects or experiences (I’m sure this must have been studied somewhere) we open up to a much broader range of person biases, associations, that just might be too strong confounds to make any claims on the data.
|
|
Qian
New Member
Posts: 20
|
Post by Qian on Mar 27, 2016 15:11:43 GMT
I echo Michael’s point that I don't understand why 1 stated difference was enough of an indication for their experiment. An alternative explanation of the experiment result can be high-difference pairs have less possible differences available, so the search for answers is easier and more efficient. Also I don’t think more “kinds of differences” stated can prove that people found more differences. Participants chose to list differences of a single dimension for low-difference pairs because of the need of efficiency, unlikely because they are unaware of other differences. (Kind of related to question 6) I think the experiment result might be different if comparing abstract words (i.e. freedom, design, science - -*) than concrete items.
As for implications for HCI systems design, I think it is still a meaningful point that people make sense of concepts by comparison/alignment. AI or mixed-initiative systems, for example, should first establish a common ground (i.e. structural alignment) with human decision-making system before claiming its differences.
|
|
|
Post by jseering on Mar 27, 2016 20:34:07 GMT
When encountering a new thing, our go-to mode for interpreting it is to say that "it's like [x] but a little different." It's hard to even imagine how we would process something that was completely different than anything we'd ever encountered before (if we could even conceptualize such a thing). I don't think that this paper was trying to imply that similar things are more different than different things, but rather that it's easier to pick out and verbalize simple distinctions when they're obvious. A red circle is different from an otherwise-identical blue circle because one is red and the other is blue. It takes more effort to think of discrete differences between a red circle and a walrus, even though there are obviously more.
I'm not sure I can think of broad implications for HCI, but I have personal experience with how this is useful in creativity outside HCI in a task somewhat similar to the "draw people who don't exist" task discussed in this paper. In a creative writing class I once took, half of the class was tasked with simply generating ideas for stories while the other half was tasked with generating ideas for stories that started with two people walking out of a building. Despite the latter category being hugely more restricted, those people had a far easier time generating ideas because of the restriction.
|
|
nhahn
New Member
Posts: 22
|
Post by nhahn on Mar 27, 2016 23:39:40 GMT
Yes, just echoing Joseph, the point of this paper was to say that things with similar things, because of the structural alignment process, it is much easier to pickout differences between the two items. This furthers the idea that individuals are doing this "structural alignment" process when doing comparisons (which was / is still a theory). Structural alignment, which they referred to in the introduction, has it's roots in analogical thinking. In analogy, individuals are performing this alignment process between two potential pairs of things, and then utilizing that alignment for generalization about a situation / scenario. Then, when teaching someone, you could leverage this contrast to point out maybe why a particular approach to solve a situation works in one case but not another.
I do think applying this, however, is pretty hard and in general it seems like taking these broad psychological concepts and generating interesting solutions is a problem. I am curious what the right way is to take these theories and apply them, other than just having a sudden moment of inspiration.
|
|
|
Post by Brandon on Mar 28, 2016 15:03:46 GMT
I'll take a stab at how this might be applied to HCI: If you're trying to create a new interface, there might be value in departing farther from existing interfaces than making incremental changes. If the differences are only incremental, people may be fixate on the differences since they more readily come to mind. Perhaps a more radical departure (or the addition of additional superficial differences) might weaken the mental comparison with existing interfaces and lead people to consider the new interface more on it's own terms. Obviously, I'm just conjecturing here and there are other aspects which will influence interface acceptance/adoption, but there might be something worth exploring here.
I think the experimental design is fine. I did wonder about the validity of the types of differences. I don't think it really impacts the main findings here, but I'm curious how consistent the labels (alignable vs. non-alignable, types) are. Are those really different categories or are those the categories *these* researchers have focused on. Citing previous work you did to justify the method isn't necessarily the most convincing thing (though that cited paper may well provide strong justification for the technique).
|
|
|
Post by judithodili on Mar 28, 2016 16:08:36 GMT
+1 with Qian and Joseph on "I think it is still a meaningful point that people make sense of concepts by comparison/alignment. AI or mixed-initiative systems, for example, should first establish a common ground (i.e. structural alignment) with human decision-making system before claiming its differences." The results of the paper are intuitive, same concept as people being able to read and comprehend paragraphs with typos scattered all over. There is ample research that proves that the human brain tries to associate new stimuli with something it has perceived in the past, and that association dramatically improves recall.
The results may have been different if they had to list the differences between a 3D object, although I suspect that the shape and symmetry will influence their perception of difference just like with words. It may be interesting to look at the features that people ignore e.g color vs the ones they subconsciously can't ignore.
|
|
|
Post by Felicia Ng on Mar 28, 2016 16:13:10 GMT
I think the experiment design was clever, and the authors did a commendable job of teasing out a potential confounding variable in Experiment 1 by conducting Experiment 2. 100%-perfect experiments are hard to design. I will stand by them and defend their rigor against critics.
In thinking about how this applies to HCI, it seems like a lot of people are fixated on UI evaluation, but what about implications for other areas of HCI? For example, if it's difficult for people to make comparisons between totally dissimilar items, then maybe one implication is that shopping/recommender systems should provide people with suggested lists of important features to consider instead of making people go through the cognitive struggle of finding nonalignable differences themselves. Along the same lines, maybe inspiration systems for creative tasks should provide people with lists of suggested criteria, constraints, or features to focus on when coming up with analogies to help them think outside the box. On a kind of random note: One thing that really struck a nerve in me was the last paragraph before the Conclusion about how people put more weight on alignable test scores when making decisions. THIS IS WHAT'S WRONG WITH THE EDUCATION SYSTEM RIGHT NOW!!! Standardized tests are used to evaluate students now, because they're an easy way to produce alignable information on student progress. But there are so many other aspects of learning that can't be as easily quantified or aligned- like grit, discipline, character building, emotional maturity, empathy development, etc. Just because something is easier to align doesn't mean it should be the primary measure for evaluating a person. This really ruffled my feathers. How can we overcome these cognitive biases from structural alignment to make this world a better place?
|
|
|
Post by sciutoalex on Mar 28, 2016 19:48:52 GMT
Reading this paper reminded me of a comment a previous design professor made to me in a motion design class. He was critiquing a particularly good animation, and after it was done, said “Theme and variation. Find three similar things and change one thing about them. Place them side by side. What more does a design need?” Gentler and Markman’s work explains why my professor’s comment often leads to good design solutions.
Gentler and Markman’s conclusion is that to understand difference, there needs to be enough underlying similarity to create common categories that can be differentiated. This applies as much to visual perception as it does to object recognition. In visual perception, humans can see weight, shape, color, pattern, texture, value/brightness. If a design varies all of these, then the viewer cannot understand a design’s order because there are no underlying connections or themes. But if some are held constant while others are varied, the viewer can understand both the whole and how the parts are unique.
When it comes to the design and engineering of technology, I think we forget the importance of theme and variation and instead opt for maximum variation in the name of giving the user maximum control. Gentler and Markman’s work urges us to choose what we vary carefully.
|
|
|
Post by Anna on Mar 28, 2016 22:02:13 GMT
Building off Felicia's post, I have a (tenuous) example of how this might apply to HCI research that is not about UIs: A lot of the work in Geoff's lab is around encountering, processing, and challenging how we view interpersonal differences (hm, I think this is an accurate statement?). So from this perspective, if we want to prime study participants to notice or mentally engage with differences among people more, this work suggests that we should present participants with humans that lean towards the similar side. And conversely, if we want to participants not to be as readily aware of interpersonal differences, then we should present them with humans that are more dissimilar (?).
This just doesn't seem right, though. Perhaps if users were asked to create a list of differences between such humans, they would generate more ideas for the more similar pair than for the more dissimilar pair, per the findings of this work. But does that doesn't tell us much about how they actually view those differences-- perhaps having a shorter, more salient list of differences still means a lot more (depending on what you are studying, etc.). For example, let's say I want to create an in-group identity among a group of users in an artificial online community for the purposes of a study. I would think that the more similar these users are on a number of different measures, the more readily they will be able to view others in the community as part of their in-group. Even if they can list more differences between themselves and other users in this situation, I would think the connection and extent to which they feel similar to the other participants would still be stronger than if the other participants had been more dissimilar.
(Btw, I don't think this work would refute such an idea. Because obviously people might be able to list more differences between a dog and a cat than between a dog and the Grand Canyon, but then if you ask them which is more similar, they'll still say the dog and the cat, not the dog and the Grand Canyon. But just trying to think of ways to extend...).
|
|
|
Post by fannie on Mar 28, 2016 22:15:42 GMT
I mainly tried to think about this in terms of my own work. Since my work deals with people/interactions, I thought about differences that show up as you compare people, or compare yourself to other people. When you’re more similar to someone, say your friends or maybe your undergraduate population, I could see that it’d be easier to point out differences in the same way that Joseph pointed out, except with “we’re both like this, but I’m [x] and she’s [y].” But at the same time, I wonder about the types of differences that come out and at what level of similarity. Like a kitten vs a magazine, compared to a kitten vs a bunny, compared to a kitten vs a kitten. Going back to people, a person vs another person would maybe fall into the same category because they’re both people, but if you broke it down into a male vs. female, according to this it should be harder to find differences between a male and female than between female and female--but perhaps those few differences that are found in the former case are what are easier to focus on (thinking about stereotypes).
Also I was typing this and just saw what Anna posted, so..what she said aha. But yeah I agree about what Anna is saying that people would still feel similar to those they are similar with. Not sure if it’s because in the comparison process they would also be able to think of more similarities, or if they would focus on the few drastic differences they could list for people who are different from them.
|
|
|
Post by Yang Zhang on Mar 28, 2016 23:33:33 GMT
I agree with Judith on the argument that similarity making it easier to establish a common ground. The AI example is also very appropriate. I would also argue that more valuable difference can be found when we compare between similar works. For example in HCI research, we always refer previous work and compare our work with them to point out the novelty and importance of our work. There won't be much to dig out if we compare between different research fields.
|
|
toby
New Member
Posts: 21
|
Post by toby on Mar 29, 2016 1:19:43 GMT
+1 for what Felicia said. The "easier to align" bias can be scary. It would be very useful to build systems that can better support not merely numeric or quantitative comparisons, but basically just help us identify the differences in a structured and organized way.
|
|
|
Post by xuwang on Mar 29, 2016 1:36:56 GMT
I agree with Joseph that the point of this paper is saying it’s easier to pick out distinctions between similar things. And I was wondering is it because similar words tend to appear together more often, so we’ve thought about their differences in previous experiences? For example, given motel and hotel, probably we’ve already compared their differences a lot of times in our daily life. I wonder if this effect comes from prior-knowledge or from alignment of differences in similar words. For example, when giving completely new words with meanings to participants (either high-similar ones, or low-similar ones), and give them the same amount of time to analyze the difference, will they still be better at generating differences for the high-similar ones? And I also think pointing out one difference is not very persuasive.
|
|