On Oct 4, 2009, at 5:40 AM, James Page wrote:
The issue I have with testing with just a few users is that it can
exclude a significant issue.
James,
I think that's the major flaw in your thinking. You're trying to use
usability testing primarily for issue detection and it's a very
inefficient tool for that.
Nielsen makes a claim that his useit site might look awful, but that
it is readable, which is is not the case for me. I am Dyslexic, and
I find Nielsen's useit website hard going, because he uses very wide
column widths.
I too am dyslexic, but the column widths aren't the big issue I have
with Jakob's site. The big issue issue I have is his content.
By only using a few people for user research in one location, are
you not excluding a significant number of your site's audience?
Yes.
Which is why using usability testing as a sole source for issue
detection will inevitably fail.
There's no way you could put together a cost-effective study (even
with super-duper remote testing applications) that would participants
at chance for every possible variance found in humans.
By trying to use usability testing in this way, you're creating a
final inspection mentality, which Demming and the world of statistical
quality control has taught us (since the 40s) is the most expensive
and least reliable way of ensuring high quality. Issues will be missed
and users will be less satisfied using this approach.
Instead, a better approach is to prevent the usability problems from
being built into the design in the first place. Jakob shouldn't need
to conduct usability tests to discover that longer column widths could
be a problem with people with reading disabilities. In fact, those of
us who've paid attention to the research on effective publishing
practices have known for a long time that shorter columns are better.
Larger sample sizes, even when the testing is dirt cheap, is too
expensive for finding problems like this. We need to shift away from
the mentality that usability testing is a quality control technique.
Because of this, we've found in our research that teams get the most
value from usability testing (along the other user research
techniques) when they use it to inform their design process. By
getting exposure to the users, the teams can make informed decisions
about their design. The more exposure, the better the outcomes of the
designs.
To research this, we studied teams building a variety of online
experiences. We looked for correlations between those teams' user
research practices and how effective the team was at producing great
designs. We looked at the range of techniques they employed, whether
they hired experienced researchers, how many studies they ran, how
frequently the studies were, and about 15 other related variables.
We found that many of the variables, including the nature of the
studies (lab versus field, for example) or number of study
participants did not correlate to better designs.
More importantly, we found that 2 key variables did correlate
substantially to better designs: the % of hours of exposure each team
member had to primary observation and the frequency of primary
observation.
This led us to start recommending that teams try to get every team
member exposed to as many hours of observing users throughout the
design process. The minimum we're recommending is 2 hours of
observation every 6 weeks. The best teams have their team members
observing users for several hours every week or so.
Based on our research, we can confidently predict that having each
team member watch two users for two hours every 3 weeks will result in
a substantially better design than hiring the world's most experienced
user researchers to conduct a 90-participant study that none of the
team members observe.
So, number of participants in the study is a red herring. The real
value is number of hours each team member is exposed to users.
That's my opinion, and it's worth what you paid for it.
Jared
p.s. Is Webnographer an unmoderated remote usability testing tool? It
occurred to me this morning that it would be great to combine
unmoderated remote usability testing with eye tracking. Then we could
throw out all the data in a single step, instead of having to ignore
it piecemeal. A huge step forward in efficiency, I would think.
Jared M. Spool
User Interface Engineering
510 Turnpike St., Suite 102, North Andover, MA 01845
e: [email protected] p: +1 978 327 5561
http://uie.com Blog: http://uie.com/brainsparks Twitter: @jmspool
________________________________________________________________
Welcome to the Interaction Design Association (IxDA)!
To post to this list ....... [email protected]
Unsubscribe ................ http://www.ixda.org/unsubscribe
List Guidelines ............ http://www.ixda.org/guidelines
List Help .................. http://www.ixda.org/help