Capturing sentiment alongside awareness/usage #183
Replies: 14 comments 30 replies
-
Thanks for creating a separate thread, it was getting hard navigating the other one! I won't address everything point by point because I mostly agree with you about the tradeoffs of each solution. But I guess my overall position would be that based purely on my own subjective intuition and experience, I do not feel like Option 2 is clearly superior to Option, 1 given our requirements. Now I'm happy to be proven wrong, and it's true we have a chance to do just that with user testing. BUT – I do not think changing the UI should be our priority right now when the survey outline isn't even done yet. To be honest, when I heard about the original end-of-August deadline, it already seemed very tight WITHOUT any user testing or UI changes. So if the decision is left up to me, I would recommend against investing our resources into: implementing this new UI, testing it, discussing the test results, and implementing new improvements – at least at this time (we can of course revisit this topic for future surveys, when we have a little bit more time). That being said I'm well aware I'm not the only person involved in the project (and I'm very grateful that this is the case!), so I don't intend to force a decision that others are not happy with. What does everybody think? |
Beta Was this translation helpful? Give feedback.
-
Hi, I'm coming from the position of user-first principles, and to share with you an industry proven design approach. Two UIs for capturing sentiment alongside awareness of web platform features and tools. The right thing to do when variant testing is to test as early as pre-launch, and have a continuous testing post-launch when there's enough audience to build on sentiment and awareness, to make informed decisions. Yet if you must turn into which UI we should consider using right now? Let's base assumptions over a top-down model of user archetype(s). The context is in answering multiple choice questions. We have Option 1: 5-answer multiple choice questions and Option 2: Quick sentiment. Within these two UIs (Opt 2) has elements that merge the standard Pros of an already Familiar set of UI used in other surveys. To guide user behaviour and improve their experiences is a controversial topic so here are some key considerations why and how familiarity and comfort matters in UX. Familiarity plays a crucial role in the usability and overall UX of a product or service. Familiar designs are more likely to be easily understood and navigated by users. It is closely tied to cognitive processes such as attention, perception, and memory. When users encounter a design that's familiar to them, their attention is more likely captured, and they are more likely to perceive and remember the design's elements and interactions. There are 2 other factors that influence familiarity. Design elements and patterns, Navigation and information architecture. Familiar design elements and patterns include common layout structures, navigation systems, and common design elements. A clear and well-organised navigation system can make it easier to find information and understand the overall structure of an interface. Familiarity affects ease of learning, speed of completion, and error recovery / prevention. Where am I going with this? Keep the design simple and straightforward, avoid unnecessary complexity, and give clear and concise information. As Lea mentioned that if we implement Option 2 in time, we can even present user study participants with both (within-subjects design), and see how they do across both. And with a limited capacity, you could fall forward with to guide user behaviour with familiarity and comfort because that's what matters in UX. |
Beta Was this translation helpful? Give feedback.
-
I agree with you on all points, Lea. I'd just add that the first one increases scrolling by 66%, but the second one increases clicks by at least that much. I also used to feel strongly about providing a neutral option, but I'm now convinced by the evidence that it's often not a good idea, as you pointed out, this is not a Likert scale, so maybe not applicable here. In the end, I'd probably still prefer the second option, but only just so, and I'm not sure the impact will be materially different. I think focusing on the questions for the user testing is probably going to be more impactful. |
Beta Was this translation helpful? Give feedback.
-
I’m pasting here the heuristic eval by @michaelquiapos, as it think it's a shame if it's lost in an email thread. I’ve styled the items common across both options in
I’ll post some responses in a reply below. |
Beta Was this translation helpful? Give feedback.
-
I fleshed out the sentiment chips idea a bit more today, and made a higher fidelity prototype: https://lea.verou.me/files/stateof/mocks/sentiment/ Interaction Videos3-point Feature questionsDesktopfeature-desktop.mp4Mobilefeature-mobile.mp4Checkbox questions (mini-feature questions)Desktopminifeature-desktop.mp4Mobileminifeature-mobile.mp4Followup commentsoption-comments.mp4Changes from earlier prototypeThere are several changes wrt the UI & UX:
Non-user facing changes:
Caveats: Protoype does not currently work in Firefox or older browser versions. This is fine for the user study, though we'd need to fix it pre-launch. ArchitectureChanges to Data modelEach question that includes sentiment chips needs to store a separate Changes to question spec (Specifying sentiment labels)Sentiment label pairs (e.g. Interested/Not interested or Want to use again/Don’t want to use again) can be defined at the question level, or the option level. Both are needed: in features each option has a different pair, in checkbox questions the whole question has the same pair. If an option doesn't have a defined pair, it is inherited from the question. This also allows individual options to opt-out of sentiment chips (e.g. you don't want "None of the above" to have sentiment chips). You can see how this works in the sample questions. There are three predefined sentiment pairs:
These can be used by simply referencing their name. My prototype uses an array ( Potential compromises to reduce implementation effort:
Changes to MarkupWhile I tried to work with the existing markup & CSS as much as possible, it will require one small markup change: moving the option LogicThe logic is spread across:
The prototype is implemented with Vue, though Vue is only used for reactivity in the HTML: there are no Vue components, no plugins, even the app spec is empty, so it should be easy to follow even for someone who doesn't speak Vue. Basically all you'd need to understand is: Mini Vue Template Syntax Cheatsheet
A few notes:
|
Beta Was this translation helpful? Give feedback.
-
Hey everybody, the work done here is impressive! I can't add anything relevant on the UX side, but just wanted to give a bit more context on the data analysis side. If I sum up the point of sentiment analysis as "do people like/want this feature or not", we don't need that much data to answer this question reliably, just enough to achieve significance in a binomial hypothesis test. For instance if we get 500 respondent, with 270 having a positive opinion on a feature, the test is significant (p-value is roughly 4%). With 100 respondents, 60 positive answers is enough. My maths might be wrong so I'll double check that properly later on, to compute how many answers we need to get to a 5% significance level depending on the ratio of positive answers. But you get the idea: we need more data for controversial features, less for non-controversial features, sometimes even a very small number of respondants can be enough. This is something we could take into account to find a compromise between mental workload and precision of the responses. Perhaps some features are deemed "risky" by vendors and that's where we could accept an higher workload for respondents. We could even indicate that explicitly that the question is controversial and therefore their answer all the more appreciated. For other less controversial features, a more discrete feature could be sufficient to get significant data. Now there is the selection bias, more involved respondent are more likely to answer which affects the result in an unknown fashion. But I think we might want to keep that a separate issue and not care too much about that for now. We are currently working on detecting sub-populations among respondents via clustering, which could be a direction to better understand those biases. |
Beta Was this translation helpful? Give feedback.
-
I've deployed the different UIs under consideration here: https://survey-staging.devographics.com/en-US/survey/state-of-html/2023/
(We're not really considering the "Tell us more…" UI for this survey but I will probably use it for other surveys) |
Beta Was this translation helpful? Give feedback.
-
Hi again, Here are more details about the data analysis concerns regarding capturing sentiment. This is strictly from the data analyst point of view, many other dimensions should be taken into account, but I hope this can fuel the decisions being made. 0) Awareness vs sentimentTo capture awareness, it makes total sense to collect as much data as possible from the ecosystem, as the sheer number of respondents is a valuable information. To capture people sentiment, proportions are more interesting than headcount. And statistics can tell us exactly how many responses to collect to precisely compute a proportion. Sometimes it's much less than we expect. 1) Yes/no answersSay that we want to answer the following question: "Do people have a positive opinion of the We want to collect enough responses to answer this question confidently. The number of response we need actually depends on how controversial the question is. For instance, if you ask 10 people randomly and 9 have a positive opinion, you are done. If you ask 10 people randomly and have 6 positive opinion, it's not enough, you can't conclude anything. Here are the exact numbers, for a 95% confidence (industry standard for AB test):
If you have less data than that for each level of positiveness in the table, you can still conclude but with less "confidence". The math behind: this is a binomial hypothesis test, I've computed a rough value using a normal approximation, and computed an exact value using R from there. 2) EstimatesNow, say that you have 63 positive responses, on a total of 107 responses (so 69.8% of positive responses). You'll want to answer the following question: "Is 69.8% of positive sentiment a good estimation, given that I got a total of 107 responses?" The table above confirms that the sentiment is positive (= over 50% of the population is happy with the I didn't crunch the exact numbers but basically the value given above are a minimum. More data are still welcome so the actual value (69.8% here) can be considered reliable. What this imply for the UX and takeaways
|
Beta Was this translation helpful? Give feedback.
-
Hey Eric, You're right about all those things, but unfortunately that only works when your sampling is random and representative of the audience. We have convenience sampling, so anyone can participate, but the audience itself might not represent the population, so we can't calculate confidence intervals, just descriptive stats. |
Beta Was this translation helpful? Give feedback.
-
Applying Lea's sentiment chip idea to the 5-option format to improve readability! |
Beta Was this translation helpful? Give feedback.
-
Oh I can't remember if we discussed that, and it's not a huge detail, but does it make sense to have sentiment next to "don't know what it is"? What will people base their sentiment on? |
Beta Was this translation helpful? Give feedback.
-
Hi there! I just checked the survey. I am seeing this. |
Beta Was this translation helpful? Give feedback.
-
If I suddenly decide to just skip the question after initially selecting an option, I can no longer deselect my initially selected option. Can 'deselecting' be possible? This happens to questions whose options are in radio buttons. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Starting a new thread here since Idea: Quick Context (#163) was getting too long and had evolved quite a bit from its original proposal (which I don’t support anymore now that we have a clearer picture of the requirements).
Problem description
Option 1: 5-answer multiple choice questions
This is a template already used in some questions in other surveys, mainly around tooling.
Pros:
Cons:
Option 2: Sentiment Chips
Update: 👉🏼 Latest prototype 👈🏼
feature-desktop.mp4
feature-mobile.mp4
Older prototype
in-answer-lite.mp4
This interaction preserves the 3-answer format for awareness, with followups for sentiment expressed as green/red1 buttons next to each awareness answer. The followups are visible on hover/focus without any UI shifts. Followups of selected answers are always visible. Clicking on a followup also selects the answer associated with it, facilitating responses to both with a single click. To save space on mobile, followups could be coded as 👍🏼 👎🏼 buttons, with only the description of the selected sentiment visible.
We could also reword the 3 answers to match the shorter ones in the 5-answer template ("Know what it is, but haven't used it" → "Heard of it", "I've used it" → Used it).
Pros:
Cons:
Footnotes
Color scheme would need to be swapped in many Asian locales. ↩ ↩2
Beta Was this translation helpful? Give feedback.
All reactions