Hochman and Manovich  analyze the sociocultural effects of specific places during specific
periods of time on user uploaded photos. Hu et al.
 quantifies various different properties based on
the users and the images on Instagram which helps
to gain insight about various meta-data and their
distribution. Camila et al.  research focused on
time of the day, week and its relationship to user
behavior, resulting in new clustering strategies.
Hochman et al.  analyzed the volume, spatial
patterns and aggregated visual features of photos
from Instagram to offer social, cultural and political
insights. Highfield et al.  discussed methodology
for research using Instagram data based on the
learning on twitter which will help us understand
the strength of these methods and their applications.
Christian et al. makes use of deep convolutional
neural networks to solve the ImageNet classification
fine-tuned the next to last layer and retrained the network to output 11 activation values which are then
converted into probability values by the softmax
layer which is the last layer of the network. Using a
small dataset of about 150 to 500 images for various
categories and by making use of random cropping,
scaling, brightness and horizontal flipping we got
an accuracy of about 88% on the test set which
we also found to be reasonable on other arbitrary
images from the instagram data collected.
B. Data Analysis
After the data preparation, we performed initial
analysis on the images and found some patterns
emerging from the data. Some interesting correlations and patterns are shown in Figures 2, 3, 4, 5.
III. DATA C OLLECTION & A NALYSIS
We have collected 600,000 images along with
their meta-data per city using the Instagram APIs. A
total of 2.4 million image dataset has been collected
across 4 cities. The collected data has the following
id, link, tags, filter, comments, likes, latitude,
longitude, locationname, locationid, createdtime,
imageurl, userid, username, realname
Using the created time attribute we find out the
stripped time from Unix timestamp and divide into
month and hour of the day for inferring seasons
and period of the day. To detect the image category
we use a Deep Neural Network which categorizes
images into 11 categories including city, selfie,
food, animal, flower, beach, nature, abstract, group,
fashion and quote. These initial set of categories
are carefully chosen after analyzing a sample set of
Instagram images and also several literature reviews.
A. Classifying images into categories
Fig. 2. Variation of filter usage during periods of the day. Clarendon
filter is being widely used.
Fig. 3. Variation of filter usage during seasons of the year. Clarendon
is more popular during Fall and Winter.
The ImageNet dataset consists of images belonging to 1000 different categories. The categories are
IV. S OLUTION & I MPLEMENTATION
varied but mainly consists of different types of
animals, flowers, objects, clothing etc. We created a A. Feature Construction
Based on the initial data analysis we found that
generic category class, for example a panda and a
dog would be classified as an animal. We therefore the following features have the most impact on filter