User Testing of E-learning Courses

Back to Topic: Instructional Design > User Testing of E-Learning Courses > Conducting User Testing

The following

12 Tips to Successful E-Learning
The following tips can be helpful conducting and/or observing the evaluation of an e-learning course.

Use technology that’s easy to access such as an easy synchronous technology,, i.e. WebEx, if it’s a virtual evaluation to Microsoft Office products and e-mail. Many synchronous technologies contain built in recording features. Other tools that can be used are Captivate which will create a recording of the learner’s actions and SnagIt is a static screen capture utility similar to Print-Screen for static pictures that can be pasted into a document. Make sure each person performing the evaluation has their own computer. Use virtual recording within the synchronous technology, screen captures, or complete paper or online evaluations. Virtual evaluations cannot observe body language but they can hear what the learner is saying. Examples of positive trigger words include wow, cool, and great. Examples of negative trigger words include hmm, just a minute, and I don’t understand, where do I go now. If an observer is not present, a background program such as WebTrends can report how many times a user clicks on a certain page within the course or runs a particular interface element. Time may be a factor in the evaluation process so it’s important to prioritize up-front on which activities/content the learner should spend their time. Learners have a more positive evaluation experience if they can see their suggestions implemented while they’re testing. Provide additional feedback tools as links within the course they’re testing. This feedback can be submitted directly to a database or via email to the designer.
 * Select effortless technology
 * Have a means of capturing/recording screens.
 * Focus on a single subject per session
 * Describe the learning Scenario
 * 1) Always communicate to the learner the goals of the test. For user testing, the goal is to assess the course usability and learner attitudes while completing the instruction. It’s important to a ssure them they are not being tested, the instruction.
 * 2) Don’t provide too much detail or commentary about the course to explain various parts because it should be self explanatory. If the learner runs into problems, that’s a sign there’s something wrong with the instruction.
 * 3) Ask the learner to talk out loud while performing the testing. Not only the steps they’re taking but also what they’re thinking about the course itself and various parts such as activities, images, etc.
 * Record the test
 * If the evaluation includes an observer, listen for “trigger words” and read body language
 * Watch for “mouse mania”. If the learner is being observed virtually or in person, the following actions from the learner are considered “mouse mania” and signal the learner might be confused or impatient with the e-learning course:
 * 1) Rapid, undirected movement of the mouse.
 * 2) Frequently scrolling up and down the screen.
 * 3) Rapidly clicking on buttons and other interface elements within the course
 * Cherry-pick key learning activities
 * Allow the client to “listen in”
 * 1) If it’s a virtual evaluation, the client can be in the same room as the observer listening to the conversation and watching the computer screen.
 * 2) If the evaluation is in-person, the learner may be uncomfortable with an extra person in the room but they can listen to audio or view screen captures of the completed evaluation.
 * On-The-Fly Updates
 * Feedback and Subsequent Engagement

Resources
Adapted from...

I'm trying to upload it but don't have the licensing, emailing Dr. H.

Also having problem linking the file to a different name..

For more information and a complete description of each tip, read, “” by Dr. Peter Honebein and Darryl Sink.