We conducted our testing using the same briefing and 9 tasks from GR3 Paper Prototyping. Throughout GR4 and GR5 the intended purpose of our application and it's features did not change, so we decided the same briefing and tasks would still be suitable and sufficient to test our finished implementation.
During pilot and user testing our users were given the briefing and then began completing the tasks one by one as they were dictated. There was no demo given since our testing involved tasks that highlighted every main feature of the application. Therefore, any sort of demo would have caused us to bias user performance on some of our tasks. We opted to have more thorough testing with all 9 tasks, rather than give a demo and have to omit some of the tasks from the testing process.
During user testing, in addition to the briefing users were given instructions to think out loud, to not worry about how quickly they completed the tasks, and to feel free to give up and ask for the solution to a task if they got frustrated.
In both situations our users were completely representative our our target audience, as they were all current MIT students.
Pilot testing was conducted on four 6.813 / 6.831 students on Demo Day. All four users had no trouble completing each of the 9 tasks but provided feedback that brought to light the following issues.
User testing was conducted on three students who have never taken either course and have never been educated in user interface design through any other sources.