HeaderSIS.jpg

IS480 Team wiki: 2012T1 6-bit Project Management UT4

From IS480
Jump to navigation Jump to search
6-bit logo.png
6-bit's Chapalang! is a social utility that connects people with friends and new friends
by offering a place for exchanging ideas and information on its public domain.
http://www.chapalang.com

Final Wikipage
Home Technical Overview Project Deliverables Project Management Learning Outcomes


Schedule

Planned Schedule

6-bit ScheduleDiagramOverview.png

































Meeting Minutes

Team Meeting Minutes

[|Meeting Minute 1] [|Meeting Minute 11] [|Meeting Minute 21]
[|Meeting Minute 2] [|Meeting Minute 12] [|Meeting Minute 22]
[|Meeting Minute 3] [|Meeting Minute 13] [|Meeting Minute 23]
[|Meeting Minute 4] [|Meeting Minute 14] [|Meeting Minute 24]
[|Meeting Minute 5] [|Meeting Minute 15] [|Meeting Minute 25]
[|Meeting Minute 6] [|Meeting Minute 16]
[|Meeting Minute 7] [|Meeting Minute 17]
[|Meeting Minute 8] [|Meeting Minute 18]
[|Meeting Minute 9] [|Meeting Minute 19]
[|Meeting Minute 10] [|Meeting Minute 20]

Supervisor Meeting Minutes

|Meeting Minute 1
|Meeting Minute 2
|Meeting Minute 3
|Meeting Minute 4
|Meeting Minute 5
|Meeting Minute 6
|Meeting Minute 7
|Meeting Minute 8
|Meeting Minute 9
|Meeting Minute 10
|Meeting Minute 11

Testing

Test Cases

|Test Cases

Test Plans

Test Plan 1 on 17 September 2012
Test Plan 2 on 28 September 2012
Test Plan 3 on 19 October 2012
Test Plan 4 on 4 November 2012

User Testing

User Testing 1 User Testing 2 User Testing 3 User Testing 4

User Testing 4

6bituser-testing4.png

Test Description

The objective of User Test 4 is on scalability, performance and analytics testing of the system. This is a 2-part test session, firstly on scalability and performance which does not require physical testers, and secondly on inter-rater reliability which requires rating judges.

The coverage of the scalability and performance test is focused on the bottleneck functions, which are the discussion forums and marketplace. The terms “performance” and “scalability” are commonly used interchangeably, but the two are distinct: performance measures the speed with which a single request can be executed, while scalability measures the ability of a request to maintain its performance under increasing load.

Additionally, Inter-Rater Reliability Test is performed on the Personalized Dashboard to determine the concordance of the personalized results and actual personality of user stereotypes.

Testers Background

There are a total of 46 testers who attended the User Test, of which 61% (28) are male and 39% (18) are female, representing various schools in SMU with SIS students being the majority. It is also observed that most testers are users of Chrome and Firefox web browser. There is no significant difference between testers’ background in User Test 1 and 2.
6bituser-testing2results.png

Test Groups

There are 2 test groups employed in this test.

Group A (Control Group) consists of 17 testers who were testers from User Test 1. The purpose is to test if their return experience with the system has any improvements.
Group B (Test Group) consists of 29 new testers who have not participated in any of our previous user tests. The purpose of their participation is for comparison, to find out if they have a different experience from the returning testers in Group A.

Test Procedures

Testers are invited to attend the User Test session and required to bring their own laptops. They are informed on the purpose of the test and given a brief description of the system objective of Chapalang!.

Subsequently, they are provided with an instruction sheet for a guided test experience. Testers will be required to perform a series of system tasks based on a test case that will go through all the system features and use-case. Thereafter, testers will fill in details in an open-ended textbox appended after each question, should they encounter any bugs or suggestions for improvements.

Test Instruction

Click Here to Download User Testing 2 Instruction

Test Results

6bituser-testing2resultsfigure.png
Based on the survey questions, the results are positive with most testers having a good level of comfort using our web application.

There are also 21 reported bugs, and 69 recommendations for improvements received. The following are the top 10 bugs reported.
6bituser-testing2bug.png

The top 10 most frequently mentioned or important recommendations will be published and appended below.
6bituser-testing2recommendation.png

Click Data Analysis(User Test 1 vs. User Test 2 – Forum Functions Only)

Additionally, click data of each test session has also been collected and analysed. They are also being compared to that with the results of User Test 1.
6bituser-testing2cda1.png
The above box-plot represents 3 sets of data comparing the number of clicks per task, for discussion forum functions only. UT1 represents the results from User Test 1, UT2A represents the results from Group A testers of User Test 2, while UT2B represents the results from Group B testers of User Test 2. For the objective of fair comparisons, the results from User Test 2 has been drilled down to consists of data

The median number of clicks it takes per tester to accomplish a forum-related task in User Test 2 ranges from 1 to 3 clicks with 2 clicks being the median, a decrement from the median of 3 clicks, as well as a smaller variance as compared to User Test 1. Additionally, it can also be observed that there is no significant difference in the results between Group A and Group B users.

Preliminary, we can observe an improvement in the user experience for Group A users between the 2 tests. The improvement can be broadly attributed to the improvements made as well as the high learnability of the system interface design. However, this observation is not conclusive and more data is required.
6bituser-testing2cda2.png
Additional statistics were computer and observed that the median time spent to accomplish a forum-related task for Group A tester is 4 seconds, and Group B tester is 5 seconds. Again, this is a significant decrement in the time spent from User Test 1, where testers spent a median of 10 seconds between each task. Based on the finding, we can reasonably derive that there is an improved user experience between User Test 1 and User Test 2, attributing to the improvements made and high learnability of the system. In addition, the improved user experience is shared between both Group A and Group B users, possibly suggesting that the improved system does not require much training or high learning curve.

Click Data Analysis (Group A vs. Group B – Marketplace Functions)

Prospectively, we will also study the user experience difference between Group A and Group B testers on marketplace functions, based on the click data which measures the number of clicks involved per task and time taken in seconds between each task.

In the box-plot diagram above, UT2A refers to User Test 2 Group A testers, while UT2B refers to User Test 2 Group B testers. Each box-plot is represented by data of a specific group of users, and the results computed based on the number of clicks of time pertaining to forum or marketplace functions.
6bituser-testing2cda3.png
Comparing marketplace functions, both Group A and Group B testers have made a median of 2 clicks to accomplish each task. While Group B testers have a wider variance of clicks of up to 4 clicks, it can be broadly attributed to outliers, user experiments or some learning curve involved in getting used to the functions or interface objects placements.
6bituser-testing2cda4.png
The result when comparing the time taken is consistent with the preliminary conclusion when comparing the number of clicks per task. The median time taken for Group A and Group B testers for forum and marketplace functions are within the range of 4 to 5 seconds. The difference between the median records is insignificant.

Overall, the result is consistent across forums and marketplace functions, between testers from both User Tests and test groups. It is also consistent with our earlier preliminary conclusion that the improvements made between the two User Tests have resulted in improved user experience, and there is a good level of learnability in the interface design.

While there are limitations in this test, where there are other externalities such as network performance, computing habits of testers and response time of each users, the macro results of the test provide a reasonable sampling on the objective of the test.

Milestones

6-bit milestones.png

Schedule Metric


6-bit ScheduleMetric.png

Bug Metric


6-bit BugMetric.png Bug Log: |Click Here

Risk & Mitigation


6-bit RiskDiagram.png