subways/mapsme_subways.py
2017-10-04 15:58:58 +03:00

230 lines
8.6 KiB
Python
Executable file

#!/usr/bin/env python3
import argparse
import json
import logging
import os
import sys
import time
import urllib.parse
import urllib.request
from subway_structure import download_cities
def overpass_request(bboxes=None):
query = '[out:json][timeout:1000];('
if bboxes is None:
bboxes = [None]
for bbox in bboxes:
bbox_part = '' if not bbox else '({})'.format(','.join(bbox))
for t, k, v in (('rel', 'route', 'subway'),
('rel', 'route', 'light_rail'),
('rel', 'route_master', 'subway'),
('rel', 'route_master', 'light_rail'),
('rel', 'public_transport', 'stop_area'),
('rel', 'public_transport', 'stop_area_group'),
('node', 'railway', 'subway_entrance')):
query += '{}["{}"="{}"]{};'.format(t, k, v, bbox_part)
query += ');(._;>);out body center qt;'
logging.debug('Query: %s', query)
url = 'http://overpass-api.de/api/interpreter?data={}'.format(urllib.parse.quote(query))
response = urllib.request.urlopen(url, timeout=1000)
if response.getcode() != 200:
raise Exception('Failed to query Overpass API: HTTP {}'.format(response.getcode()))
return json.load(response)['elements']
def multi_overpass(bboxes):
if not bboxes:
return overpass_request(None)
SLICE_SIZE = 10
result = []
for i in range(0, len(bboxes) + SLICE_SIZE - 1, SLICE_SIZE):
if i > 0:
time.sleep(5)
result.append(overpass_request(bboxes[i:i+SLICE_SIZE]))
return result
def load_xml(f):
try:
from lxml import etree
except ImportError:
import xml.etree.ElementTree as etree
elements = []
nodes = {}
for event, element in etree.iterparse(f):
if element.tag in ('node', 'way', 'relation'):
el = {'type': element.tag, 'id': int(element.get('id'))}
if element.tag == 'node':
for n in ('lat', 'lon'):
el[n] = float(element.get(n))
nodes[el['id']] = (el['lat'], el['lon'])
tags = {}
nd = []
members = []
for sub in element:
if sub.tag == 'tag':
tags[sub.get('k')] = sub.get('v')
elif sub.tag == 'nd':
nd.append(int(sub.get('ref')))
elif sub.tag == 'member':
members.append({'type': sub.get('type'),
'ref': int(sub.get('ref')),
'role': sub.get('role', '')})
if tags:
el['tags'] = tags
if nd:
el['nodes'] = nd
if members:
el['members'] = members
elements.append(el)
element.clear()
logging.info('Read %s elements, now finding centers of ways and relations', len(elements))
# Now make centers, assuming relations go after ways
ways = {}
relations = {}
for el in elements:
if el['type'] == 'way' and 'nodes' in el:
center = [0, 0]
count = 0
for nd in el['nodes']:
if nd in nodes:
center[0] += nodes[nd][0]
center[1] += nodes[nd][1]
count += 1
if count > 0:
el['center'] = {'lat': center[0]/count, 'lon': center[1]/count}
ways[el['id']] = (el['center']['lat'], el['center']['lon'])
elif el['type'] == 'relation' and 'members' in el:
center = [0, 0]
count = 0
for m in el['members']:
if m['type'] == 'node' and m['ref'] in nodes:
center[0] += nodes[m['ref']][0]
center[1] += nodes[m['ref']][1]
count += 1
elif m['type'] == 'way' and m['ref'] in ways:
center[0] += ways[m['ref']][0]
center[1] += ways[m['ref']][1]
count += 1
if count > 0:
el['center'] = {'lat': center[0]/count, 'lon': center[1]/count}
relations[el['id']] = (el['center']['lat'], el['center']['lon'])
# Iterating again, now filling relations that contain only relations
for el in elements:
if el['type'] == 'relation' and 'members' in el:
center = [0, 0]
count = 0
for m in el['members']:
if m['type'] == 'node' and m['ref'] in nodes:
center[0] += nodes[m['ref']][0]
center[1] += nodes[m['ref']][1]
count += 1
elif m['type'] == 'way' and m['ref'] in ways:
center[0] += ways[m['ref']][0]
center[1] += ways[m['ref']][1]
count += 1
elif m['type'] == 'relation' and m['ref'] in relations:
center[0] += relations[m['ref']][0]
center[1] += relations[m['ref']][1]
count += 1
if count > 0:
el['center'] = {'lat': center[0]/count, 'lon': center[1]/count}
relations[el['id']] = (el['center']['lat'], el['center']['lon'])
return elements
def merge_mapsme_networks(networks):
result = {}
for k in ('stops', 'transfers', 'networks'):
result[k] = sum([n[k] for n in networks], [])
return result
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument(
'-i', '--source', help='File to write backup of OSM data, or to read data from')
parser.add_argument('-x', '--xml', help='OSM extract with routes, to read data from')
parser.add_argument(
'-b', '--bbox', action='store_true',
help='Use city boundaries to query Overpass API instead of querying the world')
parser.add_argument('-q', '--quiet', action='store_true', help='Show only warnings and errors')
parser.add_argument('-c', '--city', help='Validate only a single city')
parser.add_argument('-o', '--output', help='JSON file for MAPS.ME')
parser.add_argument('-n', '--networks', type=argparse.FileType('w'), help='File to write the networks statistics')
options = parser.parse_args()
if options.quiet:
log_level = logging.WARNING
else:
log_level = logging.INFO
logging.basicConfig(level=logging.INFO, datefmt='%H:%M:%S',
format='%(asctime)s %(levelname)-7s %(message)s')
# Downloading cities from Google Spreadsheets
cities = download_cities()
if options.city:
cities = [c for c in cities if c.name == options.city]
logging.info('Read %s metro networks', len(cities))
if not cities:
sys.exit(2)
# Reading cached json, loading XML or querying Overpass API
if options.source and os.path.exists(options.source):
logging.info('Reading %s', options.source)
with open(options.source, 'r') as f:
osm = json.load(f)
if 'elements' in osm:
osm = osm['elements']
elif options.xml:
logging.info('Reading %s', options.xml)
osm = load_xml(options.xml)
if options.source:
with open(options.source, 'w') as f:
json.dump(osm, f)
else:
if options.bbox:
bboxes = [c.bbox for c in cities]
else:
bboxes = None
logging.info('Downloading data from Overpass API')
osm = multi_overpass(bboxes)
if options.source:
with open(options.source, 'w') as f:
json.dump(osm, f)
logging.info('Downloaded %s elements, sorting by city', len(osm))
# Sorting elements by city and prepare a dict
for el in osm:
for c in cities:
if c.contains(el):
c.add(el)
logging.info('Building routes for each city')
good_cities = []
for c in cities:
c.extract_routes()
c.validate()
if c.errors == 0:
good_cities.append(c)
logging.info('%s good cities: %s', len(good_cities), ', '.join([c.name for c in good_cities]))
if options.networks:
from collections import Counter
for c in cities:
networks = Counter()
for r in c.routes.values():
networks[str(r.network)] += 1
print('{}: {}'.format(c.name, '; '.join(
['{} ({})'.format(k, v) for k, v in networks.items()])), file=options.networks)
# Finally, preparing a JSON file for MAPS.ME
if options.output:
networks = [c.for_mapsme() for c in cities]
with open(options.output, 'w') as f:
json.dump(merge_mapsme_networks(networks), f)