Three horrible guys Data

From Geospatial Analytics for Urban Planning
Jump to navigation Jump to search
3hg.png

Back to Project Home

 

PROPOSAL

 

DATA TRANSFORMATION

 

POSTER

 

WEB MAPS

 

REPORT


Preliminary Data Observations

These are the datasets we are using for data transformation:

Dataset Rationale
Administrative Boundaries, Taiwan
  • A dataset containing SHP files of the administrative boundaries of taiwan (county, town, village)
  • Used as a reference to digitize IFC branch trade areas
Branch location of IFC, Taiwan
  • A dataset containing the geographical information of each individual branch.
  • Used as the main target of our project
  • Branches (FID) that we are looking into: 5, 162, 195, 7, 57, 25, 144, 129, 54, 81, 69, 132, 53, 122
Point of Interests (POI) , Taiwan
  • A dataset containing each individual Point-Of-Interests in Taiwan (e.g. ATMs, Amusement Parks, Banks)
  • Used as features for analysis with regards to each branch
Powerpoint Slides of trade areas for each branch (PPT slides)
  • A dataset containing each individual hand drawn trade area for each branch in Taiwan
  • Used as features for analysis with regards to each branch
Outlets Daily Sales Data
  • A dataset containing the daily sales information of each individual branch
  • Used to study the sales data along with the profile of each branch to generate yielding patterns (e.g. top and bottom performer)


Steps for data transformation:

  1. Align to Geometry and Coordinate Reference System for all files: EPSG:3828
  2. Extracting each POI from the different shp files
  3. Formation of polygons for each branch to align to ppt slides' trade area
  4. Creation of competitors' POI
  5. Aggregation of count of each POI to each trade area
  6. Investigation on areas not covered by trade area
  7. Pre-processing and aggregating of Sales Data

1. Geometry and Coordinate Reference System

Every qgis file created have to align to the same reference system used by Taiwan Maps for accuracy and unit measurement.

Information on 3828 Reference system: https://epsg.io/3828

3828.png

2. Extracting each POI from the different shp files

We are given the following POIs to extract:

  1. ATM
  2. Bank
  3. Bar or Pub
  4. Bookstore
  5. Bowling Centre
  6. Bus Station
  7. Business Facility
  8. Cinema
  9. Clothing Store
  10. Coffee Shop
  11. Commuter Rail Station
  12. Consumer Electronics Store
  13. Convenience Store
  14. Department Store
  15. Government Office
  16. Grocery Store
  17. Higher Education
  18. Hospital
  19. Hotel
  20. Medical Service
  21. Pharmacy
  22. Residential Area/ Building
  23. Restaurant
  24. School
  25. Shopping
  26. Sports Centre
  27. Sports Complex
  28. Train Station
  29. Night Life
  30. Industrial Zone
  31. Speciality Store
  32. Performing Arts


These POIs have to be extracted by finding their respective Facility Type in the shp files provided. For example, Facility Type of 9583 was filtered using the filter function in QGIS and exported as a layer to the geopackage.

FAC.png

3. Formation of polygons for each branch to align to ppt slides' trade area

As the trade areas are predefined from the ppt slides, we have to manually digitalise the trade area for each of the outlets. QGIS tools were used in the process: Split Features, Merge Features and our artistic skills. Moreover, we have added the store codes and area code given by the ppt slides so that we can easily identify these polygons.

Store.png


Below is the generated polygons for all of our trade area.

Do.png

4. Creation of competitors' POI

5 Competitors are identified and they are Dominos Pizza, Napolean Pizza, Mcdonalds, KFC and MosBurger. These data points are crucial in our analysis as they may have a correlation with the sales data. Hence, these competitors branches location have to be extracted from "Restaurant 5800".

After extraction, they are exported as layers in the geopackage to be used for aggregation.

Mac.png

5. Aggregation of count of each POI to each trade area (Script)

It is an exhausting task to aggregating each of the POIs into each trade area using the "Count Points in Polygon" tool.

The batch processing tool was tried but it is unable to append each newly created POI into an existing geopackage. Hence, a python script was written and it ran under QGIS's Python Console. It utilised "qgis:countpointsinpolygon" function, an inbuilt function by QGIS.

Learn more about the function here: https://docs.qgis.org/2.8/en/docs/user_manual/processing_algs/qgis/vector_analysis_tools/countpointsinpolygon.html

Py.png

After the aggregation of POIs into each trade area, below is the screenshot of the columns for a particular trade area.

Ta.png

Since all POIs are contained in their own geopackage, we need to find a way to merge these geopackages into one. Gladly, we have the OSGeo4W Shell. It allows us to run shell commands to merge these geopackages.

First, open the OSGeo4W Shell and cd to the directory that has the master geopackage (Taiwan_stores0x.gpkg) we want to merge into.

cd C:\Users\.......\Project\data\GeoPackage

And as all the layers in each geopackage is named as Taiwan_stores37, I want to rename them to their respective area. The test folder contains all the geopackages that I want to merge.

for %f in (./test/*.gpkg); do ogrinfo ./test/%f -sql "ALTER TABLE Taiwan_stores37 RENAME TO %~nf"

Lastly, the below command append and merges all the geopackages in the test folder into Taiwan_stores0x.gpkg.

for %f in (./test/*.gpkg); do ogr2ogr -f "format" -append ./Taiwan_stores0x.gpkg ./test/%f

6. Investigation on areas not covered by trade area

As we observe the generated trade areas, it seems that there are some areas that are not covered by any of the branches. Hence, we decided to anaylse these areas.

F1.png


In the above figure, the uncovered area is 大安森林公园 which is a park. Hence, it is explainable that such locations should not be covered in the delivery area.

F2.png


In the above figure, we do see buildings that are not in any trade area. Therefore, this uncovered area may have be omitted and should be fulfilled by one of the branches.


7. Extracting Sales Data

We obtained raw sales data which was generated from the company's Sales system. This was provided to us in a Report format which wasn't directly importable into QGIS. There were about 200 over outlet's data and it would have been impossible to do it manually one by one. Therefore we wrote 2 script that was used to extract key information from the Report and automated the output for the whole directory.
We found that the report although wasn't exact directly importable into QGIS. There was a pattern in each reports which allow us to easily parse and obtain key information which can be outputted into 2 different types of output: Daily Sales Data, Regional Sales Data for a Year.

Sales script.jpg

Both scripts were written in Python. We found that valid rows has date on its left for the Daily Sales Data. Therefore we made us of this pattern to export the daily sales data into a separate CSV. In another similar script, we identified that once a certain word is encountered in the first column, any rows after that consisted of Regional Sales data and therefore made us of that fact to wrote a second script for a different output.

For all files.png

This method is then called in both scripts to enable all the Reports in the directory to be processed and outputted as a separate CSV in another directory. This allowed us to easily distribute the Regional Sales Data to the other groups who were working on different regions.

Comments

Feel free to leave comments / suggestions!