You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
193 lines
6.2 KiB
193 lines
6.2 KiB
3 years ago
|
# Copyright (c) 2022, ETH Zurich and UNC Chapel Hill.
|
||
|
# All rights reserved.
|
||
|
#
|
||
|
# Redistribution and use in source and binary forms, with or without
|
||
|
# modification, are permitted provided that the following conditions are met:
|
||
|
#
|
||
|
# * Redistributions of source code must retain the above copyright
|
||
|
# notice, this list of conditions and the following disclaimer.
|
||
|
#
|
||
|
# * Redistributions in binary form must reproduce the above copyright
|
||
|
# notice, this list of conditions and the following disclaimer in the
|
||
|
# documentation and/or other materials provided with the distribution.
|
||
|
#
|
||
|
# * Neither the name of ETH Zurich and UNC Chapel Hill nor the names of
|
||
|
# its contributors may be used to endorse or promote products derived
|
||
|
# from this software without specific prior written permission.
|
||
|
#
|
||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
|
||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||
|
# POSSIBILITY OF SUCH DAMAGE.
|
||
|
#
|
||
|
# Author: Johannes L. Schoenberger (jsch-at-demuc-dot-de)
|
||
|
|
||
|
import os
|
||
|
import time
|
||
|
import datetime
|
||
|
import urllib
|
||
|
import urllib2
|
||
|
import urlparse
|
||
|
import socket
|
||
|
import argparse
|
||
|
import multiprocessing
|
||
|
import xml.etree.ElementTree as ElementTree
|
||
|
|
||
|
PER_PAGE = 500
|
||
|
SORT = "date-posted-desc"
|
||
|
URL = "https://api.flickr.com/services/rest/?method=flickr.photos.search&" \
|
||
|
"api_key=%s&text=%s&sort=%s&per_page=%d&page=%d&min_upload_date=%s&" \
|
||
|
"max_upload_date=%s&format=rest&extras=url_o,url_l,url_c,url_z,url_n"
|
||
|
MAX_PAGE_REQUESTS = 5
|
||
|
MAX_PAGE_TIMEOUT = 20
|
||
|
MAX_IMAGE_REQUESTS = 3
|
||
|
TIME_SKIP = 24 * 60 * 60
|
||
|
MAX_DATE = time.time()
|
||
|
MIN_DATE = MAX_DATE - TIME_SKIP
|
||
|
|
||
|
|
||
|
def parse_args():
|
||
|
parser = argparse.ArgumentParser()
|
||
|
parser.add_argument("--search_text", required=True)
|
||
|
parser.add_argument("--api_key", required=True)
|
||
|
parser.add_argument("--image_path", required=True)
|
||
|
parser.add_argument("--num_procs", type=int, default=10)
|
||
|
parser.add_argument("--max_days_without_image", type=int, default=365)
|
||
|
args = parser.parse_args()
|
||
|
return args
|
||
|
|
||
|
|
||
|
def compose_url(page, api_key, text, min_date, max_date):
|
||
|
return URL % (api_key, text, SORT, PER_PAGE, page,
|
||
|
str(min_date), str(max_date))
|
||
|
|
||
|
|
||
|
def parse_page(page, api_key, text, min_date, max_date):
|
||
|
f = None
|
||
|
for _ in range(MAX_PAGE_REQUESTS):
|
||
|
try:
|
||
|
f = urllib2.urlopen(compose_url(page, api_key, text, min_date,
|
||
|
max_date), timeout=MAX_PAGE_TIMEOUT)
|
||
|
except socket.timeout:
|
||
|
continue
|
||
|
else:
|
||
|
break
|
||
|
|
||
|
if f is None:
|
||
|
return {'pages': '0',
|
||
|
'total': '0',
|
||
|
'page': '0',
|
||
|
'perpage': '0'}, tuple()
|
||
|
|
||
|
response = f.read()
|
||
|
root = ElementTree.fromstring(response)
|
||
|
|
||
|
if root.attrib["stat"] != "ok":
|
||
|
raise IOError
|
||
|
|
||
|
photos = []
|
||
|
for photo in root.iter("photo"):
|
||
|
photos.append(photo.attrib)
|
||
|
|
||
|
return root.find("photos").attrib, photos
|
||
|
|
||
|
|
||
|
class PhotoDownloader(object):
|
||
|
|
||
|
def __init__(self, image_path):
|
||
|
self.image_path = image_path
|
||
|
|
||
|
def __call__(self, photo):
|
||
|
# Find the URL corresponding to the highest image resolution. We will
|
||
|
# need this URL here to determine the image extension (typically .jpg,
|
||
|
# but could be .png, .gif, etc).
|
||
|
url = None
|
||
|
for url_suffix in ("o", "l", "k", "h", "b", "c", "z"):
|
||
|
url_attr = "url_%s" % url_suffix
|
||
|
if photo.get(url_attr) is not None:
|
||
|
url = photo.get(url_attr)
|
||
|
break
|
||
|
|
||
|
if url is not None:
|
||
|
# Note that the following statement may fail in Python 3. urlparse
|
||
|
# may need to be replaced with urllib.parse.
|
||
|
url_filename = urlparse.urlparse(url).path
|
||
|
image_ext = os.path.splitext(url_filename)[1]
|
||
|
|
||
|
image_name = "%s_%s%s" % (photo["id"], photo["secret"], image_ext)
|
||
|
path = os.path.join(self.image_path, image_name)
|
||
|
if not os.path.exists(path):
|
||
|
print(url)
|
||
|
for _ in range(MAX_IMAGE_REQUESTS):
|
||
|
try:
|
||
|
urllib.urlretrieve(url, path)
|
||
|
except urllib.ContentTooShortError:
|
||
|
continue
|
||
|
else:
|
||
|
break
|
||
|
|
||
|
|
||
|
def main():
|
||
|
args = parse_args()
|
||
|
|
||
|
downloader = PhotoDownloader(args.image_path)
|
||
|
pool = multiprocessing.Pool(processes=args.num_procs)
|
||
|
|
||
|
num_pages = float("inf")
|
||
|
page = 0
|
||
|
|
||
|
min_date = MIN_DATE
|
||
|
max_date = MAX_DATE
|
||
|
|
||
|
days_in_row = 0;
|
||
|
|
||
|
search_text = args.search_text.replace(" ", "-")
|
||
|
|
||
|
while num_pages > page:
|
||
|
page += 1
|
||
|
|
||
|
metadata, photos = parse_page(page, args.api_key, search_text,
|
||
|
min_date, max_date)
|
||
|
|
||
|
num_pages = int(metadata["pages"])
|
||
|
|
||
|
print(78 * "=")
|
||
|
print("Page:\t\t", page, "of", num_pages)
|
||
|
print("Min-Date:\t", datetime.datetime.fromtimestamp(min_date))
|
||
|
print("Max-Date:\t", datetime.datetime.fromtimestamp(max_date))
|
||
|
print("Num-Photos:\t", len(photos))
|
||
|
print(78 * "=")
|
||
|
|
||
|
try:
|
||
|
pool.map_async(downloader, photos).get(1e10)
|
||
|
except KeyboardInterrupt:
|
||
|
pool.wait()
|
||
|
break
|
||
|
|
||
|
if page >= num_pages:
|
||
|
max_date -= TIME_SKIP
|
||
|
min_date -= TIME_SKIP
|
||
|
page = 0
|
||
|
|
||
|
if num_pages == 0:
|
||
|
days_in_row = days_in_row + 1
|
||
|
num_pages = float("inf")
|
||
|
|
||
|
print(" No images in", days_in_row, "days in a row")
|
||
|
|
||
|
if days_in_row == args.max_days_without_image:
|
||
|
break
|
||
|
else:
|
||
|
days_in_row = 0
|
||
|
|
||
|
|
||
|
if __name__ == "__main__":
|
||
|
main()
|