|
Table tents for focus groups, 224
Tabulating customer comments, 401
Tabulating survey results, 340–349
bimodal distribution, 344, 345
counting results, 340–342
cross-tabulation, 345–349
histogram display, 342
mean calculation, 342–343
median calculation, 344
mode calculation, 343–344
proportion chart, 348–349
Talkative participants in focus groups, 234–235
Target audience. See audiences
Task analysis, 182–192
analyzing results, 186–192
benefits and pitfalls, 69
contextual inquiry vs., 182, 184
defined, 182
gathering information, 184–186
hierarchical task analysis (HTA), 189–192
for information architecture development, 47
for interaction design, 49
interview format, 184
overview, 69
preparation, 184
process of, 183–186
schedule for, 183
task decomposition, 186–189
time requirements, 76
usability testing combined with, 474–475
uses for, 182–183
Task-based usability testing interview, 291
Task decomposition, 186–189
describing actions, 187–188
difficulties of, 186–187
informal versions, 189
preparing for, 189
See also task analysis
Task flows, 49
Tasks
as audience attribute, 143
for competitive usability testing, 429
creating for usability testing, 270–275
defined, 12
estimating time for, 272
of focus group assistant moderators, 228–229, 239
of focus group moderators, 227–229
for friends and family usability test, 11–12
for scheduling research participants, 104
in script for usability testing, 281–284
of usability test moderators, 288–290
for Web sites vs. dedicated software, 275
Teams, development. See development teams
Teams, user. See user advisory boards
Techniques. See tools and techniques
Technological profiles
as audience attributes, 137–138
defined, 85
Technological survey questions, 308, 537–538
Teenagers, recruiting, 112
Telephone focus groups, 460–461
Telephone screener for recruiting, 97–102
introduction for the participants, 97–98
introduction for the recruiter, 97–98
logistics, 98
questions, 99–102
target audience, 97–98
terminating, 99, 102
Telephone surveys, 339–340
Telescoping, 393
Terminology
contextual inquiry and, 165, 175
focus groups and, 230
information architecture development and, 46
for presentations, 495
Testing
discussion guide for focus groups, 223
recruiting, 94
reports, 493
surveys, 327
Tier 1 competitors, 422–423
Tier 2 competitors, 423
Timbuktu Pro (Netopia), 465
Time requirements, 76, 84. See also duration
Timing
for card sorting, 192–193
for focus groups, 201, 203
latecomers to focus groups, 233–234
survey bias from, 334
for surveys, 304–305
tracking survey responses, 325–326
for usability testing, 260
for using specialists, 448–449
See also duration
Tone, communicating functionality of, 26–27
Tools and techniques
choosing, 60–71
contextual inquiry and, 171, 175
free Web survey tools, 325
for identity design, 52
for information architecture, 47–48
for interaction design, 49–60
for log analysis, 414–418
variations on techniques, 459–460
See also specific tools and techniques
Topics for focus groups, 207–208, 214, 236
Total Quality Management (TQM), 30–31, 468. See also iterative development
Tradeoffs
inevitability of, 33
iterative development and, 33
for research techniques, 69–71
Traffic
published information, 441–443
as success criterion, 22
Transcribing focus group data, 241–242
Trend extraction
from diary study data, 382–384
from focus group data, 243–246
from usability test data, 296
Trends
focus groups for explaining, 204–205, 207, 247–248
usage trends of audience, 142
Trust, user-centered processes and, 517
TRUSTe Web site, 408
Truth, belief vs., 355
Typhoon example, 4–8
|