Friday, June 8 - 8am PDT / 10am MDT / 11am EDT / 4pm GMT
- Feedback on the coordination and process documentation ("Working Group Coordination" and "Evaluation Process" sections below)
- Discuss reporting format (please take a look at/comment on/edit example template)
- Discuss prioritization scale for issues
Note taker rotation:
User Experience Inspections of uPortal, Moodle, Sakai
(Usability and accessibility Heuristic Evaluations and Cognitive Walkthroughs)
One of the first design activities we are working on in the Fluid project is to identify current user "pain points" by performing heuristic evaluation and cognitive walk-throughs of uPortal, Sakai and Moodle.
"Heuristic evaluation is a discount usability engineering method for quick, cheap, and easy evaluation of a user interface design" (Jacob Nielson, http://www.useit.com/papers/heuristic/). Design, usability, and accessibility experts will engage in systematic inspections of the the user interface with the goal of identifying usability and accessibility problems based on recognized principles ("heuristics"). Our particular technique will combine heuristic evaluations with cognitive walk-throughs of the user interface so that we also look at expected user flows through the system and identify potential work flow problems. Heuristic evaluation isn't, however, meant to be a replacement for watching real users' activity on a system, so we intend to use heuristics in conjunction with user testing.
These evaluations will help us identify areas of the user interface that are most in need of improvement. We can thus prioritize our work on the most important usability and accessibility problems. Based on the findings of these evaluations, we will focus on UI issues that can be solved by designing well-tested, user-centered UI components. These components will encompass common interactions that can be reused within and across community source applications. On the other hand, we don't expect that all problems can be solved by creating UI components. We'll also ensure that findings and identified solutions outside the component realm will be shared with the communities. Heuristic evaluations and walk-throughs will identify areas of focus; we will engage in solid user-centered design practices and user testing to create the right solution.
The working list of heuristics and cognitive walk-through questions for the Fluid project is being compiled at the User Experience Inspection wiki page. We welcome your input, and would very much appreciate additional volunteers to help with the evaluation process.
The checklist is organic and will continue to be refined as we learn from doing the hybrid inspections/evaluations.
Please feel free to add, delete and/or more your name around this list. This list of names came from volunteers at the May 11th meeting.
Moodle UX Inspection Subgroup
- Ron (coordinator)
uPortal UX Inspection Subgroup
- Paul (coordinator)
Sakai UX Inspection Subgroup
- Daphne (coordinator)
Accesibility UX Inspection Subgroup
- Mike (coordinator)
"To Do" list - User Experience Inspection Protocol
As a group:
- In Progress - Create list of usabilty accessibility heuristics - use draft list to get started. In the spirit of agile, we'll refactor the list as we learn from the experience of combining the usability and accesibiliy heuristics and the cognitive walkthrough methods
- Agree on evaluation reporting format
Within the "application" teams:
- Agree on user profiles
- Define scenarios for cognitive walkthroughs
- Define priority settings for reporting out (evaluators will go away do their evaluation and come back together to synthesize the results)
Working Group Coordination
There is one inspection team per project/application. Teams are expected to be self-organizing and to form their own plans on how to proceed, but to communicate actively with the other teams on their plans and decisions - primarily through the wiki. Much in our approach is experimental, and it will be valuable to record what works, and what does not. Here is an outline for consideration by the team members and coordinators:
- The coordinator arranges an initial team meeting, using the Breeze meeting room or other convenient venue.
- The team members identify their areas of experience, expertise, and interest in:
- Accessibility - cognitive, visual, etc
- Cognitive walkthroughs
- The team discusses the protocol (See Clayton's outline)
- What usability heuristics do the members find most suitable?
- What accessibility measures/tools are to be used?
- What user profiles are to be assumed?
- What cognitive walkthrough scenarios are to be attempted?
- What refinements are required in the protocol?
- The team assesses coverage. What areas are covered, and with how much (desirable) redundancy? What areas aren't covered?
- Team member partnerships are arranged where possible to address usability and accessibility synchronously. Team leads are assigned in areas of expertise.
- The team discusses the logistics of actual inspection activities:
- What are the problems with geographically distributed teams?
- Can the Breeze facility help to overcome the problems?
- The team discusses reporting:
- Does the proposed template meet the team's needs?
- Are refinements to the template required?
- What additional information will be reported?
- How can results be aggregated with those from other teams (consistency, style, references to heuristic principles, etc.)
- The team determines the test target and records a clear definition of it, sufficient to permit repeatability of the assessments. (See Defining Inspection Targets)
- The team creates a test plan covering:
- Activity assignments
- Selected heuristcs, and CW methods
- User profiles
- Deliverables (what is to be captured from the inspections)
- Reporting template
- The test plan is published in the wiki.
- The team commences the evaluation process.
Evaluation Process (details of #11 above)
Assumptions: Protocol has been created
- Break application into "chunks" for evaluation (highest priority areas first)
- Create usage scenarios for cognitive walkthroughs
- Individual evaluation by 3 - 5 evaluators
- Synthesize and prioritize findings
- Brainstorm design session (identify conceptual solutions to high priority issues). Are there good component candidates?
- Write and share out report
- Incorporate findings into community (some will drive component development - others can be used for general product development in the communities)
- Sakai - Integrate into requirements group. Do we need to create jira tickets? Are these really "design bugs" conceptually and thus have a different status than requirements?
- Moodle - how does this get fed back into the process?
- uPortal - how do we integrate into their requirements process? Deliver findings to the community?
- Look for pain across applications? Are there issues a component(s) can address well?
Selecting a Target Instance of a Product for Inspection
With complex and flexible products such as uPortal, Sakai, and Moodle, which are highly configurable, customizable, extendable, and responsive to their local institutional environment, defining a test-bed environment for inspection presents some challenges. Some thoughts and suggestions are expressed in: Defining Inspection Targets.
Heuristic evaluation & Cognitive walkthrough reference material:
Jacob Nielsen'sdescription and overview
Applying Heuristics to Perform a Rigorous Accessibility Inspection in a Commercial Context
Heuristic Evaluation Report Example
Usability Heuristic checklists
Accessibility Heuristic checklists