HeaderSIS.jpg

Difference between revisions of "IS480 Team wiki:2017T2 Zenith Midterm Wiki"

From IS480
Jump to navigation Jump to search
(creation of page)
 
(update project highlights)
Line 43: Line 43:
 
===Deployed site link===
 
===Deployed site link===
 
===Project Highlights===
 
===Project Highlights===
(Highlight changes since project acceptance)
+
<br>
 +
Our project schedule is divided into 13 iterations.
 +
* We are currently on our 9th iteration.
 +
* Up till (date), we have completed (percentage)% of our development progress.
 +
* 1 user testing was conducted before midterms. The results is shown (here).
  
Describe the project progress briefly here. Have the project continued as planned? If not, is the team confident to complete? This is a crossroad for the team to make a decision. Proceed with confident or file an incomplete.
+
Unexpected events:
 
+
* New team of clients
What unexpected events occurred?
+
* List of requirement changes (here).
*Team members too busy with other work
 
*List of requirement changes  
 
** CRUD items replaced with CU/Sync/Archive items
 
** Business analytics replaced with iPad client
 
*Took 8 weeks to learn Ruby on Rails
 
*etc.
 
Be brief.
 
  
 
==Project Management==
 
==Project Management==

Revision as of 14:11, 3 February 2018

Zenith banner.png

Home

Team

Project Overview

Project Management

Documentation

Main Wiki

Midterm Wiki Final Wiki


Zenith midterm header.PNG


Project Progress Summary

Midterm Slides

Deployed site link

Project Highlights


Our project schedule is divided into 13 iterations.

  • We are currently on our 9th iteration.
  • Up till (date), we have completed (percentage)% of our development progress.
  • 1 user testing was conducted before midterms. The results is shown (here).

Unexpected events:

  • New team of clients
  • List of requirement changes (here).

Project Management

Provide more details about the status, schedule and the scope of the project. Describe the complexity of the project.

Project Status:

Highlight changes to modules, the completion status (implemented, user testing done, client approved, deployed, etc), the confidence level (0-1 where 0 is no confident of getting it done, 1 is 100% confident in getting it done) and comments (who has been assigned to do it, new scope, removed scoped, etc). Please use a table format to summarize with links to function details.

Project Schedule (Plan Vs Actual):

Compare the project plan during acceptance with the actual work done at this point. Briefly describe a summary here. Everything went as plan, everything has changed and the team is working on a new project with new sponsors or the supervisor is missing. A good source for this section comes from the project weekly report.

Provide a comparison of the plan and actual schedule. Has the project scope expanded or reduced? You can use the table below or your own gantt charts.

Project Metrics:

Summary of analysis for the metrics collected. You may refer to another page for the details about the metrics and how it is collected.

Project Risks:

Update the proposal assumptions and risks. Describe what you learn from the risk update and mitigation steps taken.

Be sure to prioritize the risks.

Technical Complexity:

Describe and list the technical complexity of your project in order of highest complexity first. For example, deploying on iPhone using Objective-C, customizing Drupal with own database, quick search for shortest flight path, database structure, etc.

Quality of product

Provide more details about the quality of your work. For example, you designed a flexible configurable system using XML.config files, uses Strategy Design Pattern to allow plugging in different strategy, implement a regular expression parser to map a flexible formula editor, etc.

Intermediate Deliverables:

There should be some evidence of work in progress.

Not all parts of the deliverables are necessary but the evidence should be convincing of the progress. Try to include design deliverables that shows the quality of your project.

Deployment:

In an iterative approach, ready to use system should be available (deployed) for client and instructions to access the system described here (user name). If necessary, provide a deployment diagram link.

Testing:

Describe the testing done on your system. For example, the number of user testing, tester profile, test cases, survey results, issue tracker, bug reports, etc.

Reflection

In this section, describe what have the team learn? Be brief. Sometimes, the client writes a report to feedback on the system; this sponsor report can be included or linked from here.

Team Reflection:

Any training and lesson learn? What are the take-away so far? It would be very convincing if the knowledge is share at the wiki knowledge base and linked here.

Benjamin Gan Reflection:

You may include individual reflection if that make sense at this point. The team is uncooperative and did not follow my instructions.