diff --git a/.gitattributes b/.gitattributes deleted file mode 100644 index f31e32f..0000000 --- a/.gitattributes +++ /dev/null @@ -1 +0,0 @@ -Edge_List.csv filter=lfs diff=lfs merge=lfs -text diff --git a/Amsterdam/data_storage/GEO_metadata.csv b/Amsterdam/data_storage/GEO_metadata.csv deleted file mode 100644 index 7f11cec..0000000 --- a/Amsterdam/data_storage/GEO_metadata.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6c0f0848d1eb578275c4beb41da8db3575aff21aed18d02e2aad1962b008bed2 -size 10128566 diff --git a/Amsterdam/data_storage/GEO_node_dist.csv b/Amsterdam/data_storage/GEO_node_dist.csv deleted file mode 100644 index f83ef24..0000000 --- a/Amsterdam/data_storage/GEO_node_dist.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6c990f43741653497a46944154d7cc300e2860eb2edac3c9384ae7bd6ec9200d -size 12649416 diff --git a/Amsterdam/data_storage/metadata_compare.csv b/Amsterdam/data_storage/metadata_compare.csv deleted file mode 100644 index d2d0644..0000000 --- a/Amsterdam/data_storage/metadata_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f9f7f7c457127a09cc15d88bef21ff8219495b54fa3d2e29b566ae8745897306 -size 680712 diff --git a/Amsterdam/data_storage/sentences.csv b/Amsterdam/data_storage/sentences.csv deleted file mode 100644 index 9de8d83..0000000 --- a/Amsterdam/data_storage/sentences.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6e8d9945a9ac50e6b66a67b3dc03319968c7b6a0fa56b2e5110226725c4fb05d -size 1086932 diff --git a/Amsterdam/data_storage/sentences_compare.csv b/Amsterdam/data_storage/sentences_compare.csv deleted file mode 100644 index df5209f..0000000 --- a/Amsterdam/data_storage/sentences_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1cd42f13ef8494a90176a237235ebd9d3040137f14bd71b18867be35b37fb101 -size 2618045 diff --git a/Amsterdam/data_storage/social_links.csv b/Amsterdam/data_storage/social_links.csv deleted file mode 100644 index d459663..0000000 --- a/Amsterdam/data_storage/social_links.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b342ee11811728f5a087ca3677402bef7894f225d2696e6ea9199009783203da -size 1910439 diff --git a/Amsterdam/data_storage/tags_embedding.csv b/Amsterdam/data_storage/tags_embedding.csv deleted file mode 100644 index a9568cc..0000000 --- a/Amsterdam/data_storage/tags_embedding.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bfea92b875225a8ab84eb95db9805bf5195cd003704764bffd0941609ce10970 -size 1503602 diff --git a/Amsterdam/save_image.py b/Amsterdam/save_image.py deleted file mode 100644 index 4768121..0000000 --- a/Amsterdam/save_image.py +++ /dev/null @@ -1,162 +0,0 @@ -import json -import flickrapi -import pandas as pd -import numpy as np -from argparse import Namespace -from collections import Counter -import pickle -import os -import urllib - -args = Namespace( - # Data and Path information - api_key = u'[api_key]', - api_secret = u'[api_secret]', - radius = 2, - save_dir = 'data_storage/', - tags = None, - len_grid = 1, - image_dir = 'data_storage/images/' - ) - -def get_latlon(id_x, id_y, num = args.len_grid): - lat = 52.365 - lon = 4.887777778 - return lat,lon - -def collect_ids(flickr, lat, lon, radius, x,y, tags = None): - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].astype(str).unique().tolist() - else: - Ids = [] - - walk = flickr.walk(has_geo = 1, lat = lat, lon = lon, radius = args.radius, tags=tags) - for photo in walk: - id_now = photo.get('id') - if id_now in Ids: - continue - Ids.append(id_now) - if len(Ids)%200 == 0: - print('{} photo ids collected'.format(len(Ids))) - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - if len(Ids) >= 5000: - return Ids - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - return Ids - -def update_df(Photos): - return Photos - - -def get_photos(flickr, Photos, Ids): - Photos = {str(k):v for k,v in Photos.items()} - processed = Photos.keys() - print(len(processed)) - for id_now in Ids: - if id_now in processed: - continue - else: - Photos[id_now] = {} - sizes = json.loads(flickr.photos.getSizes(photo_id = id_now, format='json')) - info = json.loads(flickr.photos.getInfo(photo_id = id_now, format='json')) - try: - url_c = sizes['sizes']['size'][8]['source'] - url_q = sizes['sizes']['size'][1]['source'] - url_n = sizes['sizes']['size'][4]['source'] - url_largest = sizes['sizes']['size'][-1]['source'] - can = sizes['sizes']['candownload'] - Photos[id_now]['candownload'] = can - Photos[id_now]['url_c'] = url_c - Photos[id_now]['url_q'] = url_q - Photos[id_now]['url_n'] = url_n - Photos[id_now]['url_largest'] = url_largest - Photos[id_now]['others'] = sizes - Photos[id_now]['info'] = info - Photos[id_now]['owner'] = info['photo']['owner']['nsid'] - Photos[id_now]['owner_loc'] = info['photo']['owner']['location'] - Photos[id_now]['title'] = info['photo']['title']['_content'] - Photos[id_now]['description'] = info['photo']['description']['_content'] - Photos[id_now]['comments'] = info['photo']['comments']['_content'] - Photos[id_now]['taken'] = info['photo']['dates']['taken'] - Photos[id_now]['views'] = info['photo']['views'] - Photos[id_now]['people'] = info['photo']['people']['haspeople'] - Photos[id_now]['tags'] = info['photo']['tags']['tag'] - Photos[id_now]['lat'] = info['photo']['location']['latitude'] - Photos[id_now]['lon'] = info['photo']['location']['longitude'] - Photos[id_now]['neighbourhood'] = info['photo']['location']['neighbourhood']['_content'] - Photos[id_now]['url'] = info['photo']['urls']['url'][0]['_content'] - - if can: - urllib.request.urlretrieve(url_q, args.image_dir+'150/{}.jpg'.format(id_now)) - urllib.request.urlretrieve(url_n, args.image_dir+'320/{}.jpg'.format(id_now)) - - if len(processed)%20 ==1: - print('{}/{} photos collected'.format(len(processed),len(Ids))) - with open(args.image_dir+'Photo_sizes_pre_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - except Exception as e: - print(e) - print(id_now) - continue - - with open(args.image_dir+'Photo_sizes_pre.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - return Photos - -def main(): - - flickr = flickrapi.FlickrAPI(args.api_key, args.api_secret) - - if 'completed.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'completed.p', 'rb') as fp: - completed = pickle.load(fp) - else: - completed = {} - - if 'Photo_sizes.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'Photo_sizes.p', 'rb') as fp: - Photos = pickle.load(fp) - else: - Photos = {} - - for x in range(args.len_grid): - for y in range(args.len_grid): - if (x,y) in completed.keys(): - continue - - lat,lon = get_latlon(x,y) - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - else: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - - - Photos = get_photos(flickr, Photos, Ids) - - completed[(x,y)] = {} - completed[(x,y)]['lat'] = lat - completed[(x,y)]['lon'] = lon - completed[(x,y)]['collected'] = len(Ids) - completed[(x,y)]['total'] = len(Photos) - - with open(args.save_dir+'completed.p', 'wb') as fp: - pickle.dump(completed,fp, protocol=pickle.HIGHEST_PROTOCOL) - - completed_df = pd.DataFrame(completed).T - completed_df.to_csv(args.save_dir+'completed.csv') - - -if __name__ == "__main__": - main() -"""## END""" \ No newline at end of file diff --git a/Suzhou/data_storage/GEO_metadata.csv b/Suzhou/data_storage/GEO_metadata.csv deleted file mode 100644 index 5304d3a..0000000 --- a/Suzhou/data_storage/GEO_metadata.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bcfa7a5aa4d3866cc26838c23afa0c524356e170d9889a4ac8fa17893ec524e3 -size 3279387 diff --git a/Suzhou/data_storage/GEO_node_dist.csv b/Suzhou/data_storage/GEO_node_dist.csv deleted file mode 100644 index 5708acc..0000000 --- a/Suzhou/data_storage/GEO_node_dist.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9530067faffb092406b5bf5c609266ffd5e2440b69e9d2754375e8e55052039f -size 1418528 diff --git a/Suzhou/data_storage/metadata_compare.csv b/Suzhou/data_storage/metadata_compare.csv deleted file mode 100644 index acc54cf..0000000 --- a/Suzhou/data_storage/metadata_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9c4b740acc52eb10938f0755de199b30b77f267cb515f74f9c69d593e0ad81b3 -size 573046 diff --git a/Suzhou/data_storage/sentences.csv b/Suzhou/data_storage/sentences.csv deleted file mode 100644 index 62f7d84..0000000 --- a/Suzhou/data_storage/sentences.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:25b63d93b8ca7ddf40aae5a1f4cc6be9882538f3586f10e9206c880d57d3a509 -size 83411 diff --git a/Suzhou/data_storage/sentences_compare.csv b/Suzhou/data_storage/sentences_compare.csv deleted file mode 100644 index 8b27a0f..0000000 --- a/Suzhou/data_storage/sentences_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:badf2939b3ae6d14667e790686b17e566e14e5a409131f7f04c7c8123e0fdfcb -size 241512 diff --git a/Suzhou/data_storage/social_links.csv b/Suzhou/data_storage/social_links.csv deleted file mode 100644 index d13c11d..0000000 --- a/Suzhou/data_storage/social_links.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:133dc949c63ce7e031727b0a2afb034d227111fc60363478571ec4a0f824b0dd -size 424187 diff --git a/Suzhou/data_storage/tags_embedding.csv b/Suzhou/data_storage/tags_embedding.csv deleted file mode 100644 index 7e83285..0000000 --- a/Suzhou/data_storage/tags_embedding.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:da5232c7a1da8eee79e89e325a72bd0b1cecbc823c861374491bbbf34eac2d55 -size 122852 diff --git a/Suzhou/save_image.py b/Suzhou/save_image.py deleted file mode 100644 index dee81fd..0000000 --- a/Suzhou/save_image.py +++ /dev/null @@ -1,161 +0,0 @@ -import json -import flickrapi -import pandas as pd -import numpy as np -from argparse import Namespace -from collections import Counter -import pickle -import os -import urllib - -args = Namespace( - # Data and Path information - api_key = u'[api_key]', - api_secret = u'[api_secret]', - radius = 5, - save_dir = 'data_storage/', - tags = None, - len_grid = 1, - image_dir = 'data_storage/images/' - ) - -def get_latlon(id_x, id_y, num = args.len_grid): - lat = 31.302300 - lon = 120.631300 - return lat,lon - -def collect_ids(flickr, lat, lon, radius, x,y, tags = None): - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].astype(str).unique().tolist() - else: - Ids = [] - - walk = flickr.walk(has_geo = 1, lat = lat, lon = lon, radius = args.radius, tags=tags) - for photo in walk: - id_now = photo.get('id') - if id_now in Ids: - continue - Ids.append(id_now) - if len(Ids)%200 == 0: - print('{} photo ids collected'.format(len(Ids))) - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - if len(Ids) >= 4200: - return Ids - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - return Ids - -def update_df(Photos): - return Photos - - -def get_photos(flickr, Photos, Ids): - Photos = {str(k):v for k,v in Photos.items()} - processed = Photos.keys() - print(len(processed)) - for id_now in Ids: - if id_now in processed: - continue - else: - Photos[id_now] = {} - sizes = json.loads(flickr.photos.getSizes(photo_id = id_now, format='json')) - info = json.loads(flickr.photos.getInfo(photo_id = id_now, format='json')) - try: - url_c = sizes['sizes']['size'][8]['source'] - url_q = sizes['sizes']['size'][1]['source'] - url_n = sizes['sizes']['size'][4]['source'] - url_largest = sizes['sizes']['size'][-1]['source'] - can = sizes['sizes']['candownload'] - Photos[id_now]['candownload'] = can - Photos[id_now]['url_c'] = url_c - Photos[id_now]['url_q'] = url_q - Photos[id_now]['url_n'] = url_n - Photos[id_now]['url_largest'] = url_largest - Photos[id_now]['others'] = sizes - Photos[id_now]['info'] = info - Photos[id_now]['owner'] = info['photo']['owner']['nsid'] - Photos[id_now]['owner_loc'] = info['photo']['owner']['location'] - Photos[id_now]['title'] = info['photo']['title']['_content'] - Photos[id_now]['description'] = info['photo']['description']['_content'] - Photos[id_now]['comments'] = info['photo']['comments']['_content'] - Photos[id_now]['taken'] = info['photo']['dates']['taken'] - Photos[id_now]['views'] = info['photo']['views'] - Photos[id_now]['people'] = info['photo']['people']['haspeople'] - Photos[id_now]['tags'] = info['photo']['tags']['tag'] - Photos[id_now]['lat'] = info['photo']['location']['latitude'] - Photos[id_now]['lon'] = info['photo']['location']['longitude'] - Photos[id_now]['neighbourhood'] = info['photo']['location']['neighbourhood']['_content'] - Photos[id_now]['url'] = info['photo']['urls']['url'][0]['_content'] - - if can: - urllib.request.urlretrieve(url_q, args.image_dir+'150/{}.jpg'.format(id_now)) - urllib.request.urlretrieve(url_n, args.image_dir+'320/{}.jpg'.format(id_now)) - - if len(processed)%20 ==1: - print('{}/{} photos collected'.format(len(processed),len(Ids))) - with open(args.image_dir+'Photo_sizes_pre_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - except Exception as e: - print(e) - print(id_now) - continue - - with open(args.image_dir+'Photo_sizes_pre.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - return Photos - -def main(): - - flickr = flickrapi.FlickrAPI(args.api_key, args.api_secret) - - if 'completed.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'completed.p', 'rb') as fp: - completed = pickle.load(fp) - else: - completed = {} - - if 'Photo_sizes.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'Photo_sizes.p', 'rb') as fp: - Photos = pickle.load(fp) - else: - Photos = {} - - for x in range(args.len_grid): - for y in range(args.len_grid): - if (x,y) in completed.keys(): - continue - - lat,lon = get_latlon(x,y) - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - else: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - - - Photos = get_photos(flickr, Photos, Ids) - - completed[(x,y)] = {} - completed[(x,y)]['lat'] = lat - completed[(x,y)]['lon'] = lon - completed[(x,y)]['collected'] = len(Ids) - completed[(x,y)]['total'] = len(Photos) - - with open(args.save_dir+'completed.p', 'wb') as fp: - pickle.dump(completed,fp, protocol=pickle.HIGHEST_PROTOCOL) - - completed_df = pd.DataFrame(completed).T - completed_df.to_csv(args.save_dir+'completed.csv') - -if __name__ == "__main__": - main() -"""## END""" \ No newline at end of file diff --git a/Tripoli/data_storage/Face_preds.csv b/Tripoli/data_storage/Face_preds.csv deleted file mode 100644 index 2226879..0000000 --- a/Tripoli/data_storage/Face_preds.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c551ded8f33e75071b4280a0acc5792f1dcd056ac08d61dc8c703316f1fe1514 -size 36674 diff --git a/Tripoli/data_storage/IMG_pred_150.csv b/Tripoli/data_storage/IMG_pred_150.csv deleted file mode 100644 index 0151609..0000000 --- a/Tripoli/data_storage/IMG_pred_150.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:84e9cbcb2db4d31fd565896e852a66a7841e5ec1c35e8b57aadecbbf39a35e0e -size 9420972 diff --git a/Tripoli/model_storage/stack_classifier.joblib b/Tripoli/model_storage/stack_classifier.joblib deleted file mode 100644 index 783bb85..0000000 Binary files a/Tripoli/model_storage/stack_classifier.joblib and /dev/null differ diff --git a/Tripoli/model_storage/vote_classifier.joblib b/Tripoli/model_storage/vote_classifier.joblib deleted file mode 100644 index fd23dee..0000000 Binary files a/Tripoli/model_storage/vote_classifier.joblib and /dev/null differ diff --git a/Venezia/collect_data.py b/Venezia/collect_data.py deleted file mode 100644 index 977f201..0000000 --- a/Venezia/collect_data.py +++ /dev/null @@ -1,141 +0,0 @@ -import flickrapi -import json -import pandas as pd -import numpy as np -from argparse import Namespace -from collections import Counter -import pickle -import os - -args = Namespace( - # Data and Path information - api_key = u'[api_key]', - api_secret = u'[api_secret]', - radius = 0.3, - save_dir = 'data_storage/', - tags = None, - len_grid = 20, - ) - -def get_latlon(id_x, id_y, num = args.len_grid): - lat_min = 45.420855 - lon_min = 12.291054 - - lat_max = 45.448286 - lon_max = 12.369234 - - lat_d = (lat_max - lat_min)/(num-1) - lon_d = (lon_max - lon_min)/(num-1) - - lat = lat_min + id_x * lat_d - lon = lon_min + id_y * lon_d - - return lat,lon - -def collect_ids(flickr, lat, lon, radius, x,y, tags = None): - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].tolist() - else: - Ids = [] - - walk = flickr.walk(has_geo = 1, lat = lat, lon = lon, radius = args.radius, tags=tags) - for photo in walk: - id_now = photo.get('id') - if id_now in Ids: - continue - Ids.append(id_now) - if len(Ids)%200 == 0: - print('{} photo ids collected'.format(len(Ids))) - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - return Ids - -def update_df(Photos): - return Photos - -def get_photos(flickr, Photos, Ids): - processed = Photos.keys() - for id_now in Ids: - if str(id_now) in processed: - continue - - Photos[id_now] = {} - info = json.loads(flickr.photos.getInfo(photo_id = id_now, format='json')) - Photos[id_now]['info'] = info - Photos[id_now]['owner'] = info['photo']['owner']['nsid'] - Photos[id_now]['owner_loc'] = info['photo']['owner']['location'] - Photos[id_now]['title'] = info['photo']['title']['_content'] - Photos[id_now]['description'] = info['photo']['description']['_content'] - Photos[id_now]['comments'] = info['photo']['comments']['_content'] - Photos[id_now]['taken'] = info['photo']['dates']['taken'] - Photos[id_now]['views'] = info['photo']['views'] - Photos[id_now]['people'] = info['photo']['people']['haspeople'] - Photos[id_now]['tags'] = info['photo']['tags']['tag'] - Photos[id_now]['lat'] = info['photo']['location']['latitude'] - Photos[id_now]['lon'] = info['photo']['location']['longitude'] - Photos[id_now]['neighbourhood'] = info['photo']['location']['neighbourhood']['_content'] - Photos[id_now]['url'] = info['photo']['urls']['url'][0]['_content'] - - if len(processed)%100 ==1: - print('{}/{} photos collected'.format(len(processed),len(Ids))) - with open(args.save_dir+'Photo_info_last.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop('info',axis=1) - photo_df.to_csv(args.save_dir+'photos_last.csv', sep='\t',encoding='utf-8-sig') - - with open(args.save_dir+'Photo_info_last_pre.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - - with open(args.save_dir+'Photo_info_last.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop('info',axis=1) - photo_df.to_csv(args.save_dir+'photos_last.csv', sep='\t',encoding='utf-8-sig') - return Photos - -def main(): - - flickr = flickrapi.FlickrAPI(args.api_key, args.api_secret) - - if 'completed.p' in [files for root, dirs, files in os.walk(args.save_dir)][0]: - with open(args.save_dir+'completed.p', 'rb') as fp: - completed = pickle.load(fp) - else: - completed = {} - - for x in range(args.len_grid): - for y in range(args.len_grid): - if (x,y) in completed.keys(): - continue - - lat,lon = get_latlon(x,y) - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].tolist() - else: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - - if 'Photo_info_last.p' in [files for root, dirs, files in os.walk(args.save_dir)][0]: - with open(args.save_dir+'Photo_info_last.p', 'rb') as fp: - Photos = pickle.load(fp) - else: - Photos = {} - - Photos = get_photos(flickr, Photos, Ids) - - completed[(x,y)] = {} - completed[(x,y)]['lat'] = lat - completed[(x,y)]['lon'] = lon - completed[(x,y)]['collected'] = len(Ids) - completed[(x,y)]['total'] = len(Photos) - - with open(args.save_dir+'completed.p', 'wb') as fp: - pickle.dump(completed,fp, protocol=pickle.HIGHEST_PROTOCOL) - - completed_df = pd.DataFrame(completed).T - completed_df.to_csv(args.save_dir+'completed.csv') - - -if __name__ == "__main__": - main() -"""## END""" \ No newline at end of file diff --git a/Venezia/data_storage/GEO_metadata.csv b/Venezia/data_storage/GEO_metadata.csv deleted file mode 100644 index 5c55587..0000000 --- a/Venezia/data_storage/GEO_metadata.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a856890f7021086bf141a636999f31e536762dae0d2bb425f321f476143a6d78 -size 5854463 diff --git a/Venezia/data_storage/GEO_node_dist.csv b/Venezia/data_storage/GEO_node_dist.csv deleted file mode 100644 index e57e85f..0000000 --- a/Venezia/data_storage/GEO_node_dist.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:78ceb17b16a1145a3d33e89b5ee9c76dd3279eea84c55956aad2c135f5f32764 -size 12553485 diff --git a/Venezia/data_storage/metadata_compare.csv b/Venezia/data_storage/metadata_compare.csv deleted file mode 100644 index a5756b9..0000000 --- a/Venezia/data_storage/metadata_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:0aee50838d329e50d8900982fea43ac660c26d66202438e814e246e6da7253aa -size 538286 diff --git a/Venezia/data_storage/sentences.csv b/Venezia/data_storage/sentences.csv deleted file mode 100644 index bf48153..0000000 --- a/Venezia/data_storage/sentences.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bc1945111884ed9e77ec18d2e968965ff9c3d25956bca5970c4dff0d27a5a58a -size 462315 diff --git a/Venezia/data_storage/sentences_compare.csv b/Venezia/data_storage/sentences_compare.csv deleted file mode 100644 index 2ec28b2..0000000 --- a/Venezia/data_storage/sentences_compare.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b79d39e7dc8566268f2e5b0e7cb1b030f34fc77b85bc68e18a49c6d0ed7df1a4 -size 1190367 diff --git a/Venezia/data_storage/social_links.csv b/Venezia/data_storage/social_links.csv deleted file mode 100644 index 2458dab..0000000 --- a/Venezia/data_storage/social_links.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:081c615d05cbbc3a464e0e6a14fcc7e71a354d8dde6e0bcd5a7c3c0c078cae77 -size 5620189 diff --git a/Venezia/data_storage/tags_embedding.csv b/Venezia/data_storage/tags_embedding.csv deleted file mode 100644 index fccad92..0000000 --- a/Venezia/data_storage/tags_embedding.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b2f491c05e65d24b9673e4e3e8e2ef69ce40000fd67655304ffa1681c3afb3d6 -size 775969 diff --git a/Venezia/save_image.py b/Venezia/save_image.py deleted file mode 100644 index e2f5c5e..0000000 --- a/Venezia/save_image.py +++ /dev/null @@ -1,162 +0,0 @@ -import json -import flickrapi -import pandas as pd -import numpy as np -from argparse import Namespace -from collections import Counter -import pickle -import os -import urllib - -args = Namespace( - # Data and Path information - api_key = u'[api_key]', - api_secret = u'[api_secret]', - radius = 5, - save_dir = 'data_storage/', - tags = None, - len_grid = 1, - image_dir = 'data_storage/images/' - ) - -def get_latlon(id_x, id_y, num = args.len_grid): - lat = 45.438759 - lon = 12.327145 - return lat,lon - -def collect_ids(flickr, lat, lon, radius, x,y, tags = None): - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].astype(str).unique().tolist() - else: - Ids = [] - - walk = flickr.walk(has_geo = 1, lat = lat, lon = lon, radius = args.radius, tags=tags) - for photo in walk: - id_now = photo.get('id') - if id_now in Ids: - continue - Ids.append(id_now) - if len(Ids)%200 == 0: - print('{} photo ids collected'.format(len(Ids))) - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - if len(Ids) >= 5000: - return Ids - pd.Series(Ids, name = 'ids').to_csv(args.save_dir + 'photo_ids_{}_{}.csv'.format(x,y), index=False) - return Ids - -def update_df(Photos): - return Photos - - -def get_photos(flickr, Photos, Ids): - Photos = {str(k):v for k,v in Photos.items()} - processed = Photos.keys() - print(len(processed)) - for id_now in Ids: - if id_now in processed: - continue - else: - Photos[id_now] = {} - sizes = json.loads(flickr.photos.getSizes(photo_id = id_now, format='json')) - info = json.loads(flickr.photos.getInfo(photo_id = id_now, format='json')) - try: - url_c = sizes['sizes']['size'][8]['source'] - url_q = sizes['sizes']['size'][1]['source'] - url_n = sizes['sizes']['size'][4]['source'] - url_largest = sizes['sizes']['size'][-1]['source'] - can = sizes['sizes']['candownload'] - Photos[id_now]['candownload'] = can - Photos[id_now]['url_c'] = url_c - Photos[id_now]['url_q'] = url_q - Photos[id_now]['url_n'] = url_n - Photos[id_now]['url_largest'] = url_largest - Photos[id_now]['others'] = sizes - Photos[id_now]['info'] = info - Photos[id_now]['owner'] = info['photo']['owner']['nsid'] - Photos[id_now]['owner_loc'] = info['photo']['owner']['location'] - Photos[id_now]['title'] = info['photo']['title']['_content'] - Photos[id_now]['description'] = info['photo']['description']['_content'] - Photos[id_now]['comments'] = info['photo']['comments']['_content'] - Photos[id_now]['taken'] = info['photo']['dates']['taken'] - Photos[id_now]['views'] = info['photo']['views'] - Photos[id_now]['people'] = info['photo']['people']['haspeople'] - Photos[id_now]['tags'] = info['photo']['tags']['tag'] - Photos[id_now]['lat'] = info['photo']['location']['latitude'] - Photos[id_now]['lon'] = info['photo']['location']['longitude'] - Photos[id_now]['neighbourhood'] = info['photo']['location']['neighbourhood']['_content'] - Photos[id_now]['url'] = info['photo']['urls']['url'][0]['_content'] - - if can: - urllib.request.urlretrieve(url_q, args.image_dir+'150/{}.jpg'.format(id_now)) - urllib.request.urlretrieve(url_n, args.image_dir+'320/{}.jpg'.format(id_now)) - - if len(processed)%20 ==1: - print('{}/{} photos collected'.format(len(processed),len(Ids))) - with open(args.image_dir+'Photo_sizes_pre_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - except Exception as e: - print(e) - print(id_now) - continue - - with open(args.image_dir+'Photo_sizes_pre.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - - with open(args.image_dir+'Photo_sizes.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop(['others','info'],axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes.csv', sep='\t',encoding='utf-8-sig') - return Photos - -def main(): - - flickr = flickrapi.FlickrAPI(args.api_key, args.api_secret) - - if 'completed.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'completed.p', 'rb') as fp: - completed = pickle.load(fp) - else: - completed = {} - - if 'Photo_sizes.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'Photo_sizes.p', 'rb') as fp: - Photos = pickle.load(fp) - else: - Photos = {} - - for x in range(args.len_grid): - for y in range(args.len_grid): - if (x,y) in completed.keys(): - continue - - lat,lon = get_latlon(x,y) - - if 'photo_ids_{}_{}.csv'.format(x,y) in [files for root, dirs, files in os.walk(args.save_dir)][0]: - #Ids = pd.read_csv(args.save_dir+'photo_ids_{}_{}.csv'.format(x,y),sep='\t')['ids'].tolist() - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - else: - Ids = collect_ids(flickr, lat,lon, args.radius, tags=args.tags, x=x,y=y) - - - Photos = get_photos(flickr, Photos, Ids) - - completed[(x,y)] = {} - completed[(x,y)]['lat'] = lat - completed[(x,y)]['lon'] = lon - completed[(x,y)]['collected'] = len(Ids) - completed[(x,y)]['total'] = len(Photos) - - with open(args.save_dir+'completed.p', 'wb') as fp: - pickle.dump(completed,fp, protocol=pickle.HIGHEST_PROTOCOL) - - completed_df = pd.DataFrame(completed).T - completed_df.to_csv(args.save_dir+'completed.csv') - -if __name__ == "__main__": - main() -"""## END""" \ No newline at end of file diff --git a/Venezia/save_image_all.py b/Venezia/save_image_all.py deleted file mode 100644 index d3fc7b2..0000000 --- a/Venezia/save_image_all.py +++ /dev/null @@ -1,81 +0,0 @@ -import json -import flickrapi -import pandas as pd -import numpy as np -from argparse import Namespace -from collections import Counter -import pickle -import os -import urllib - -args = Namespace( - # Data and Path information - api_key = u'[api_key]', - api_secret = u'[api_secret]', - radius = 0.3, - save_dir = 'data_storage/', - tags = None, - len_grid = 20, - image_dir = 'data_storage/images/grid/' - #tags = 'landscape,urban,heritage,culture,building,architecture,park,street' -) - -def get_photos(flickr, Photos, Ids): - processed = Photos.keys() - for id_now in Ids: - if id_now in processed: - continue - else: - Photos[id_now] = {} - sizes = json.loads(flickr.photos.getSizes(photo_id = id_now, format='json')) - try: - url_c = sizes['sizes']['size'][-2]['source'] - url_q = sizes['sizes']['size'][1]['source'] - can = sizes['sizes']['candownload'] - Photos[id_now]['candownload'] = can - Photos[id_now]['url_c'] = url_c - Photos[id_now]['url_q'] = url_q - Photos[id_now]['others'] = sizes - - if can: - urllib.request.urlretrieve(url_q, args.image_dir+'{}.jpg'.format(id_now)) - - if len(processed)%20 ==1: - print('{}/{} photos collected'.format(len(processed),len(Ids))) - with open(args.image_dir+'Photo_sizes_pre_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - with open(args.image_dir+'Photo_sizes_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop('others',axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes_sep.csv', sep='\t',encoding='utf-8-sig') - except: - print(id_now) - continue - - with open(args.image_dir+'Photo_sizes_pre_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - - with open(args.image_dir+'Photo_sizes_sep.p', 'wb') as fp: - pickle.dump(Photos,fp, protocol=pickle.HIGHEST_PROTOCOL) - photo_df = pd.DataFrame(Photos).T.drop('others',axis=1) - photo_df.to_csv(args.image_dir+'photos_sizes_sep.csv', sep='\t',encoding='utf-8-sig') - return Photos - -def main(): - - flickr = flickrapi.FlickrAPI(args.api_key, args.api_secret) - - df=pd.read_csv(args.save_dir+'grid/photos_sep.csv',sep='\t').rename(columns={'Unnamed: 0':'num'}) - Ids = df['ids'] - - if 'Photo_sizes_sep.p' in [files for root, dirs, files in os.walk(args.image_dir)][0]: - with open(args.image_dir+'Photo_sizes_sep.p', 'rb') as fp: - Photos = pickle.load(fp) - else: - Photos = {} - - Photos = get_photos(flickr, Photos, Ids) - -if __name__ == "__main__": - main() -"""## END""" \ No newline at end of file diff --git a/dataset/Amsterdam/Attribute_Labels.csv b/dataset/Amsterdam/Attribute_Labels.csv deleted file mode 100644 index 402f98e..0000000 --- a/dataset/Amsterdam/Attribute_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:79dbd3f76f5840b49597ad30baef16d114f2575b4c9bf8c56b18ef43cd5dc18a -size 776973 diff --git a/dataset/Amsterdam/Edge_List.csv b/dataset/Amsterdam/Edge_List.csv deleted file mode 100644 index 974854c..0000000 --- a/dataset/Amsterdam/Edge_List.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:01ad6f31b4775de00fc360f43990eb480939c3a1a13d91c8f5737914297c904f -size 345050826 diff --git a/dataset/Amsterdam/Textual_Features.csv b/dataset/Amsterdam/Textual_Features.csv deleted file mode 100644 index e0c9b3d..0000000 --- a/dataset/Amsterdam/Textual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6c1548b0b36164da6d2e96cc6cd320719df860c9994cd00eb3f7b56ffc38ed71 -size 26599660 diff --git a/dataset/Amsterdam/Value_Labels.csv b/dataset/Amsterdam/Value_Labels.csv deleted file mode 100644 index 537d8b9..0000000 --- a/dataset/Amsterdam/Value_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:028d1f4b022b5918bc775b68a6cedd83ecf44894b80d0ff272148841cd61c99a -size 1282248 diff --git a/dataset/Amsterdam/Visual_Features.csv b/dataset/Amsterdam/Visual_Features.csv deleted file mode 100644 index a3b2d5b..0000000 --- a/dataset/Amsterdam/Visual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1857147c46574147f7837ab48eb6ae2cc6fc89090ac5e4690a89e7b7296e4ded -size 73703739 diff --git a/dataset/Suzhou/Attribute_Labels.csv b/dataset/Suzhou/Attribute_Labels.csv deleted file mode 100644 index 4d035dd..0000000 --- a/dataset/Suzhou/Attribute_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e3245966aa7cec07e22d32ffaae62d6b1a40c4598ce9f1b4a3649088515e8133 -size 654094 diff --git a/dataset/Suzhou/Edge_List.csv b/dataset/Suzhou/Edge_List.csv deleted file mode 100644 index 0393333..0000000 --- a/dataset/Suzhou/Edge_List.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1c5df35b441193229079589715161b5df084f2a88431c444305f532079f8dcde -size 244516908 diff --git a/dataset/Suzhou/Textual_Features.csv b/dataset/Suzhou/Textual_Features.csv deleted file mode 100644 index d3f3ed8..0000000 --- a/dataset/Suzhou/Textual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:39ccaf018439df202b83c556c21b3f332d929e1adb487c3c0fb8d45b51296bba -size 8519945 diff --git a/dataset/Suzhou/Value_Labels.csv b/dataset/Suzhou/Value_Labels.csv deleted file mode 100644 index b0119b4..0000000 --- a/dataset/Suzhou/Value_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ca060292fa17b260176748b923b9c6747214b04e27343b47664123daf9afc7b1 -size 448781 diff --git a/dataset/Suzhou/Visual_Features.csv b/dataset/Suzhou/Visual_Features.csv deleted file mode 100644 index d0df029..0000000 --- a/dataset/Suzhou/Visual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:befcb748a8f0e0694a9644ef7a23d8d1bb82f42fcf828bd670542f649ffdbe0c -size 62044956 diff --git a/dataset/Venice-XL/Attribute_Labels.csv b/dataset/Venice-XL/Attribute_Labels.csv deleted file mode 100644 index 1b9434d..0000000 --- a/dataset/Venice-XL/Attribute_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9f12c4f1fc6de7efe7acec7f0df9b3a24acf243639562ca2e50fe7d0db51a649 -size 17767357 diff --git a/dataset/Venice-XL/Textual_Features.csv b/dataset/Venice-XL/Textual_Features.csv deleted file mode 100644 index 13c851e..0000000 --- a/dataset/Venice-XL/Textual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c4717d5ed7d9fb275e40dd40b8087cfd0c324a252a9761b224ddd27b7c939292 -size 469868603 diff --git a/dataset/Venice-XL/Value_Labels.csv b/dataset/Venice-XL/Value_Labels.csv deleted file mode 100644 index b4db081..0000000 --- a/dataset/Venice-XL/Value_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c9ee32cc98b2540f7f3355511fe5693980efe86034e288c73387464d47016171 -size 22952340 diff --git a/dataset/Venice-XL/Visual_Features.csv b/dataset/Venice-XL/Visual_Features.csv deleted file mode 100644 index 5b562fe..0000000 --- a/dataset/Venice-XL/Visual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8ca510c21b512cb1ecbbcad5e65e80a0749cb2c886c2f3c7e2cbe6746320f89c -size 1600771549 diff --git a/dataset/Venice/Attribute_Labels.csv b/dataset/Venice/Attribute_Labels.csv deleted file mode 100644 index caee275..0000000 --- a/dataset/Venice/Attribute_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:23d2a06d4188d4f49d0d7928d4ac6823e6a7a0cd47ff4402431ad5527ce46798 -size 640728 diff --git a/dataset/Venice/Edge_List.csv b/dataset/Venice/Edge_List.csv deleted file mode 100644 index e70cc0c..0000000 --- a/dataset/Venice/Edge_List.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3aed6312f8db718cf9fef219aef3482542abc05e39c8ef23f78d7ad1839ca7a4 -size 147681581 diff --git a/dataset/Venice/Textual_Features.csv b/dataset/Venice/Textual_Features.csv deleted file mode 100644 index 46e38b7..0000000 --- a/dataset/Venice/Textual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b7e4083873cf4ff047bd9ab965e8429b52b8a13451ab0b2739d1524add468be5 -size 16523427 diff --git a/dataset/Venice/Value_Labels.csv b/dataset/Venice/Value_Labels.csv deleted file mode 100644 index 2604b80..0000000 --- a/dataset/Venice/Value_Labels.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:258d51703ea3f09af9b44feec6f7f4fc2f3b4fc1a06a7e8aaed77e5557ea0858 -size 813972 diff --git a/dataset/Venice/Visual_Features.csv b/dataset/Venice/Visual_Features.csv deleted file mode 100644 index bf227a7..0000000 --- a/dataset/Venice/Visual_Features.csv +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:fa333f2da3fe16c0c11bcf17ad974e99051104273a2920deaf5c844943d3f60e -size 58381915