diff --git a/3party/just_gtfs b/3party/just_gtfs index d8510d0a4e..0b3559b70b 160000 --- a/3party/just_gtfs +++ b/3party/just_gtfs @@ -1 +1 @@ -Subproject commit d8510d0a4e9893a9ff6fc0521477773fdf991905 +Subproject commit 0b3559b70b183ddbf6fd7355bbfc62a383d431e4 diff --git a/README.md b/README.md index fd26e6605f..621a918818 100644 --- a/README.md +++ b/README.md @@ -57,6 +57,8 @@ repository: see `3party` directory. The team uses mostly XCode and Qt Creator, though these are not mandatory. We have an established [c++ coding style](https://github.com/omapsapp/omapsapp/blob/master/docs/CPP_STYLE.md) and [Objective-C coding style](https://github.com/omapsapp/omapsapp/blob/master/docs/OBJC_STYLE.md). +**You can turn on experimental public transport support.** For details please read [simple instruction.](https://github.com/omapsapp/omapsapp/blob/master/docs/EXPERIMENTAL_PUBLIC_TRANSPORT_SUPPORT.md) + See [CONTRIBUTING.md](https://github.com/omapsapp/omapsapp/blob/master/docs/CONTRIBUTING.md) for the repository initialization process, the description of all the directories of this repository and other development-related information. diff --git a/android/build.gradle b/android/build.gradle index 0022743d15..3d78e2b092 100644 --- a/android/build.gradle +++ b/android/build.gradle @@ -80,7 +80,9 @@ dependencies { implementation 'androidx.constraintlayout:constraintlayout:1.1.3' implementation 'androidx.work:work-runtime:2.2.0' implementation 'com.android.billingclient:billing:1.1' - implementation 'uk.co.samuelwall:material-tap-target-prompt:2.12.1' + implementation('uk.co.samuelwall:material-tap-target-prompt:3.1.0', { + exclude group: 'androidx.lifecycle', module: 'lifecycle-extensions' + }) implementation 'com.firebase:firebase-jobdispatcher:0.8.5' implementation 'com.google.android:flexbox:1.0.0' implementation 'com.trafi:anchor-bottom-sheet-behavior:0.13-alpha' diff --git a/android/gradle.properties b/android/gradle.properties index 79cb13a587..8e1f366421 100644 --- a/android/gradle.properties +++ b/android/gradle.properties @@ -2,8 +2,8 @@ propMinSdkVersion=21 propTargetSdkVersion=29 propCompileSdkVersion=29 propBuildToolsVersion=27.0.3 -propVersionCode=1051 -propVersionName=10.5.1 +propVersionCode=1052 +propVersionName=10.5.2 propDebugNdkFlags=V=1 NDK_DEBUG=1 DEBUG=1 propReleaseNdkFlags=V=1 NDK_DEBUG=0 PRODUCTION=1 propMultiDexVersion=2.0.1 diff --git a/android/src/com/mapswithme/maps/routing/NavigationController.java b/android/src/com/mapswithme/maps/routing/NavigationController.java index d35df1e4e0..12b01b9074 100644 --- a/android/src/com/mapswithme/maps/routing/NavigationController.java +++ b/android/src/com/mapswithme/maps/routing/NavigationController.java @@ -6,8 +6,8 @@ import android.content.Context; import android.content.Intent; import android.location.Location; import android.media.MediaPlayer; -import android.os.Build; import android.os.Bundle; +import android.text.SpannableStringBuilder; import android.text.TextUtils; import android.util.Pair; import android.view.View; @@ -90,14 +90,7 @@ public class NavigationController implements TrafficManager.TrafficCallback, Vie { mFrame = activity.findViewById(R.id.navigation_frame); mBottomFrame = mFrame.findViewById(R.id.nav_bottom_frame); - mBottomFrame.setOnClickListener(new View.OnClickListener() - { - @Override - public void onClick(View v) - { - switchTimeFormat(); - } - }); + mBottomFrame.setOnClickListener(v -> switchTimeFormat()); mNavMenu = createNavMenu(); mNavMenu.refresh(); @@ -114,7 +107,7 @@ public class NavigationController implements TrafficManager.TrafficCallback, Vie mStreetFrame = topFrame.findViewById(R.id.street_frame); mNextStreet = (TextView) mStreetFrame.findViewById(R.id.street); View shadow = topFrame.findViewById(R.id.shadow_top); - UiUtils.showIf(Build.VERSION.SDK_INT < Build.VERSION_CODES.LOLLIPOP, shadow); + UiUtils.show(shadow); UiUtils.extendViewWithStatusBar(mStreetFrame); UiUtils.extendViewMarginWithStatusBar(turnFrame); @@ -196,12 +189,17 @@ public class NavigationController implements TrafficManager.TrafficCallback, Vie private void updateVehicle(RoutingInfo info) { - mNextTurnDistance.setText(Utils.formatUnitsText(mFrame.getContext(), - R.dimen.text_size_nav_number, - R.dimen.text_size_nav_dimension, - info.distToTurn, - info.turnUnits)); - info.carDirection.setTurnDrawable(mNextTurnImage); + if (!TextUtils.isEmpty(info.distToTurn)) + { + SpannableStringBuilder nextTurnDistance = Utils.formatUnitsText(mFrame.getContext(), + R.dimen.text_size_nav_number, + R.dimen.text_size_nav_dimension, + info.distToTurn, + info.turnUnits); + mNextTurnDistance.setText(nextTurnDistance); + info.carDirection.setTurnDrawable(mNextTurnImage); + } + if (RoutingInfo.CarDirection.isRoundAbout(info.carDirection)) UiUtils.setTextAndShow(mCircleExit, String.valueOf(info.exitNum)); else diff --git a/android/src/com/mapswithme/maps/routing/RoutingInfo.java b/android/src/com/mapswithme/maps/routing/RoutingInfo.java index 789ae9c9d9..2189279066 100644 --- a/android/src/com/mapswithme/maps/routing/RoutingInfo.java +++ b/android/src/com/mapswithme/maps/routing/RoutingInfo.java @@ -1,13 +1,11 @@ package com.mapswithme.maps.routing; -import android.location.Location; import androidx.annotation.DrawableRes; import androidx.annotation.NonNull; import android.widget.ImageView; import com.mapswithme.maps.R; -import com.mapswithme.maps.bookmarks.data.DistanceAndAzimut; public class RoutingInfo { diff --git a/android/src/com/mapswithme/maps/tips/Tutorial.java b/android/src/com/mapswithme/maps/tips/Tutorial.java index b2578024ed..660a82abe7 100644 --- a/android/src/com/mapswithme/maps/tips/Tutorial.java +++ b/android/src/com/mapswithme/maps/tips/Tutorial.java @@ -156,7 +156,8 @@ public enum Tutorial .setBackgroundColour(ThemeUtils.getColor(activity, R.attr.tipsBgColor)) .setFocalColour(activity.getResources().getColor(android.R.color.transparent)) .setPromptBackground(new ImmersiveModeCompatPromptBackground(activity.getWindowManager())) - .setPromptStateChangeListener(listener); + .setPromptStateChangeListener(listener) + .setIgnoreStatusBar(true); builder.show(); } diff --git a/android/src/com/mapswithme/util/BatteryState.java b/android/src/com/mapswithme/util/BatteryState.java index 9dbbe64931..64be40c1d6 100644 --- a/android/src/com/mapswithme/util/BatteryState.java +++ b/android/src/com/mapswithme/util/BatteryState.java @@ -38,6 +38,8 @@ public final class BatteryState return new State(getLevel(batteryStatus), getChargingStatus(batteryStatus)); } + // Called from JNI. + @SuppressWarnings("unused") @IntRange(from=0, to=100) public static int getLevel(@NonNull Context context) { @@ -50,6 +52,8 @@ public final class BatteryState return batteryStatus.getIntExtra(BatteryManager.EXTRA_LEVEL, 0); } + // Called from JNI. + @SuppressWarnings("unused") @ChargingStatus public static int getChargingStatus(@NonNull Context context) { diff --git a/docs/EXPERIMENTAL_PUBLIC_TRANSPORT_SUPPORT.md b/docs/EXPERIMENTAL_PUBLIC_TRANSPORT_SUPPORT.md new file mode 100644 index 0000000000..1141100662 --- /dev/null +++ b/docs/EXPERIMENTAL_PUBLIC_TRANSPORT_SUPPORT.md @@ -0,0 +1,81 @@ +# Experimental support of public transport from GTFS + +## Introduction to the experimental public transport feature + +At the moment our app does not have full support for public transport. What we have now: + +- Scarce transit data collected from OSM which includes subway, light rail, monorail and train routes and stops. Let's call it the [OSM transit](https://github.com/mapsme/omim/blob/master/docs/SUBWAY_GENERATION.md) from now on. +- The "Subway layer" for visual representation of the OSM transit. +- "Subway navigation" - OSM transit routing. + + +:bus: But you can turn on the experimental feature of [GTFS](https://developers.google.com/transit/gtfs/reference) public transport and use it inside the MAPS.ME app. It includes all [transit types definded by GTFS specification:](https://developers.google.com/transit/gtfs/reference/extended-route-types) bus, train, ferry, aerial lifts of different kinds, trolleybus and much more. Let's call this version of transit data **GTFS transit** from now on. + +To mixin the experimental GTFS transit into the OSM transit data you should follow 2 simple steps: +- Run the pipeline for downloading and preparing GTFS data about public transport. +- Switch to the new version of the transit routing section in maps: build maps with the GTFS transit section with the help of special options for generator_tool. + +After completing these steps you will have maps with: + +:star: Proper public transport navigation. You'll have the opportunity to build public transport routes with transfers from one type of public transport to another (e.g. bus -> train, monorail -> subway, etc). The route is built according to the transit schedules, days off and exceptions in timetables. + +:star: "Subway layer" with old OSM transit data but in new visual representation. It differs from the current layer in handling parallel segments of different routes with the same color. In current version we draw separate polyline for each route. It looks a little bit messy, and the new version fixes this: we collapse single-colored parallel routes into one line. + +Most of the data collisions between OSM and GTFS sources are excluded, because we filter all subway data from GTFS on purpose. There still could be collisions in light rail, monorail and train data, but it is not a problem: on the "Subway layer" we show data only from OSM. In routing we choose optimal route from GTFS and OSM. + + +**Why is this feature experimental?** The main reason is that it is not properly tested. Also it lacks new user interfaces for filtering public transport by type, setting departure time and routing "follow mode". + + +## Instructions for turning on full public transport support + +1. Run the python script for collecting GTFS feeds from two GTFS aggregators: [Transitland](https://www.transit.land/) and [OpenMobilityData.](http://transitfeeds.com/feeds) You can find the script in omim/tools/python/transit/gtfs/. It is called download_gtfs.py. It crawls all the feeds from these aggregators, loads feed zips and extracts as subdirectories to the specified directory. + +Example: +``` +python3 download_gtfs.py --path=dir_for_storing_feeds --mode=fullrun --source=all --omd_api_key=YOUR_KEY_FOR_OPEN_MOBILITY_DATA_API +``` + +In this example all accessible feeds from Transitland and OpenMobilityData will be downloaded to the `dir_for_storing_feeds` directory. But if you don't want to create key for OpenMobilityData API you can run this tool for crawling Transitland only: + +``` +python3 download_gtfs.py --path=dir_for_storing_feeds --mode=fullrun --source=transitland +``` + +After completing this step you will have directory containing subdirectories with GTFS feeds. + +2. Build and run the C++ tool for filtering, combining and transforming GTFS feeds (downloaded on the previous step) to the intermediate json files. Later they will be passed to the generator_tool for creating transit section in new format. You can find the tool in `omim/transit/world_feed/gtfs_converter.` The tool is` gtfs_converter.` + +It can be run in two modes: for handling GTFS-only or for merging data from GTFS and OSM. + + +Example of running `gtfs_converter` for merging GTFS and OSM data: + +``` +./gtfs_converter --path_mapping=mapping.txt --path_mapping_edges=mapping_edges.txt --path_gtfs_feeds=dir_for_storing_feeds --path_subway_json=subway_latest.json --path_json=result_json_dir --path_resources=omim/data +``` + +Example of running `gtfs_converter` for saving only GTFS data (without OSM): + +``` +./gtfs_converter --path_mapping=mapping.txt --path_mapping_edges=mapping_edges.txt --path_gtfs_feeds=dir_for_storing_feeds --path_json=result_json_dir --path_resources=omim/data +``` + +:exclamation: GTFS data often changes: some routes are cancelled, some stops are built, timetables constantly change. **If you're planning to build maps periodically, more than once, then you must take care of the safety of files with transit entities ids which are generated by `gtfs_converter.`** The goal is to make transit entities consistent between re-runs of the tool so that routing between old regions and the new ones doesn't fail. It is better to backup files specified as `path_mapping` and `path_mapping_edges`. When you run `gtfs_converter` the first time these files are created. Subsequent launches update existing files. + + +After completing this step you will have directory with line-by-line jsons in the directory specified by `path_json` option of `gtfs_converter.` + + +3. In map_generator.ini uncomment `SUBWAY_URL` parameter and set `TRANSIT_URL` parameter to the directory created on the previous step. Example: +``` +TRANSIT_URL: file:///home/result_json_dir +``` + +Run generator tool [as usual](https://github.com/mapsme/omim/tree/master/tools/python/maps_generator) with this ini config. After it is done you'll have mwms with transit section in experimental GTFS format. + +:checkered_flag: Use the resulting mwms in your app. Enjoy the experimental public transport in MAPS.ME! + + +## If you have questions +Feel free to ask [mesozoic-drones](https://github.com/mesozoic-drones), you can also do it by mail: mesozoic.drones@gmail.com diff --git a/generator/final_processor_country.cpp b/generator/final_processor_country.cpp index 645077de21..aae240e282 100644 --- a/generator/final_processor_country.cpp +++ b/generator/final_processor_country.cpp @@ -254,20 +254,18 @@ void CountryFinalProcessor::ProcessBuildingParts() m4::Tree buildingPartsKDTree; ForEachFeatureRawFormat(path, [&](auto && fb, auto /* pos */) { - if (!(fb.IsArea() && fb.IsValid())) - return; - - if (fb.HasType(buildingPartClassifType)) + if (fb.IsArea() && fb.HasType(buildingPartClassifType)) buildingPartsKDTree.Add(fb); }); FeatureBuilderWriter writer(path, true /* mangleName */); ForEachFeatureRawFormat(path, [&](auto && fb, auto /* pos */) { - if (fb.IsArea() && fb.IsValid() && + if (fb.IsArea() && fb.HasType(buildingClassifType) && DoesBuildingConsistOfParts(fb, buildingPartsKDTree)) { fb.AddType(buildingWithPartsClassifType); + fb.GetParams().FinishAddingTypes(); } writer.Write(fb); diff --git a/generator/generator_integration_tests/features_tests.cpp b/generator/generator_integration_tests/features_tests.cpp index c6e1051239..89b8c54f08 100644 --- a/generator/generator_integration_tests/features_tests.cpp +++ b/generator/generator_integration_tests/features_tests.cpp @@ -519,47 +519,47 @@ private: feature::GenerateInfo m_genInfo; }; -//UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCoasts) -//{ -// FeatureIntegrationTests::BuildCoasts(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCoasts) +{ + FeatureIntegrationTests::BuildCoasts(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, BuildWorldMultithread) -//{ -// FeatureIntegrationTests::BuildWorld(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, BuildWorldMultithread) +{ + FeatureIntegrationTests::BuildWorld(); +} UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCountriesMultithread) { FeatureIntegrationTests::BuildCountries(); } -//UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCountriesWithComplex) -//{ -// FeatureIntegrationTests::BuildCountriesWithComplex(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCountriesWithComplex) +{ + FeatureIntegrationTests::BuildCountriesWithComplex(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, CheckMixedTagsAndNodes) -//{ -// FeatureIntegrationTests::CheckMixedTagsAndNodes(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, CheckMixedTagsAndNodes) +{ + FeatureIntegrationTests::CheckMixedTagsAndNodes(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, CheckGeneratedDataMultithread) -//{ -// FeatureIntegrationTests::CheckGeneratedData(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, CheckGeneratedDataMultithread) +{ + FeatureIntegrationTests::CheckGeneratedData(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, BuildWorldOneThread) -//{ -// FeatureIntegrationTests::BuildWorldOneThread(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, BuildWorldOneThread) +{ + FeatureIntegrationTests::BuildWorldOneThread(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCountriesOneThread) -//{ -// FeatureIntegrationTests::BuildCountriesOneThread(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, BuildCountriesOneThread) +{ + FeatureIntegrationTests::BuildCountriesOneThread(); +} -//UNIT_CLASS_TEST(FeatureIntegrationTests, CheckGeneratedDataOneThread) -//{ -// FeatureIntegrationTests::CheckGeneratedDataOneThread(); -//} +UNIT_CLASS_TEST(FeatureIntegrationTests, CheckGeneratedDataOneThread) +{ + FeatureIntegrationTests::CheckGeneratedDataOneThread(); +} diff --git a/generator/generator_tool/generator_tool.cpp b/generator/generator_tool/generator_tool.cpp index 75730a6795..b71eed20b2 100644 --- a/generator/generator_tool/generator_tool.cpp +++ b/generator/generator_tool/generator_tool.cpp @@ -151,7 +151,7 @@ DEFINE_bool(disable_cross_mwm_progress, false, DEFINE_string(srtm_path, "", "Path to srtm directory. If set, generates a section with altitude information " "about roads."); -DEFINE_string(worldroads_path, "", +DEFINE_string(world_roads_path, "", "Path to a file with roads that should end up on the world map. If set, generates a " "section with these roads in World.mwm. The roads may be used to identify which mwm " "files are touched by an arbitrary route."); @@ -491,10 +491,10 @@ MAIN_WITH_ERROR_HANDLING([](int argc, char ** argv) } } - if (country == WORLD_FILE_NAME && !FLAGS_worldroads_path.empty()) + if (country == WORLD_FILE_NAME && !FLAGS_world_roads_path.empty()) { LOG(LINFO, ("Generating routing section for World.")); - if (!routing::BuildWorldRoads(dataFile, FLAGS_worldroads_path)) + if (!routing::BuildWorldRoads(dataFile, FLAGS_world_roads_path)) { LOG(LCRITICAL, ("Generating routing section for World has failed.")); return EXIT_FAILURE; diff --git a/openlr/router.cpp b/openlr/router.cpp index 342d18d052..98ec70b4cb 100644 --- a/openlr/router.cpp +++ b/openlr/router.cpp @@ -648,20 +648,22 @@ bool Router::ReconstructPath(std::vector & edges, std::vector & edges, std::vectorm_u, false /* outgoing */, m_points[0].m_lfrcnp, [&](routing::Edge const & edge) { + auto toPairRev = [](auto && e) { return e.ToPairRev(); }; double const score = GetMatchingScore( edge.GetEndJunction().GetPoint(), edge.GetStartJunction().GetPoint(), - make_transform_iterator(EdgeItRev(e), std::mem_fn(&Edge::ToPairRev)), - make_transform_iterator(edges.rend(), std::mem_fn(&Edge::ToPairRev))); + make_transform_iterator(EdgeItRev(e), toPairRev), + make_transform_iterator(edges.rend(), toPairRev)); if (score > frontEdgeScore) { frontEdgeScore = score; @@ -688,10 +691,11 @@ bool Router::ReconstructPath(std::vector & edges, std::vectorm_v, true /* outgoing */, m_points[m_points.size() - 2].m_lfrcnp, [&](routing::Edge const & edge) { + auto toPair = [](auto && e) { return e.ToPair(); }; double const score = GetMatchingScore( edge.GetStartJunction().GetPoint(), edge.GetEndJunction().GetPoint(), - make_transform_iterator(e.base(), std::mem_fn(&Edge::ToPair)), - make_transform_iterator(edges.end(), std::mem_fn(&Edge::ToPair))); + make_transform_iterator(e.base(), toPair), + make_transform_iterator(edges.end(), toPair)); if (score > backEdgeScore) { backEdgeScore = score; @@ -727,7 +731,7 @@ void Router::FindSingleEdgeApproximation(std::vector const & edges, { double const kCoverageThreshold = 0.5; - CHECK(all_of(edges.begin(), edges.end(), std::mem_fn(&Edge::IsFake)), ()); + CHECK(all_of(edges.begin(), edges.end(), [](auto && e) { return e.IsFake(); }), ()); double expectedLength = 0; for (auto const & edge : edges) diff --git a/routing/routing_benchmarks/car_routing_tests.cpp b/routing/routing_benchmarks/car_routing_tests.cpp index 297ea28bf8..ec8cd4546b 100644 --- a/routing/routing_benchmarks/car_routing_tests.cpp +++ b/routing/routing_benchmarks/car_routing_tests.cpp @@ -60,15 +60,15 @@ UNIT_CLASS_TEST(CarTest, InCity) TestCarRouter(ms::LatLon(55.75785, 37.58267), ms::LatLon(55.76082, 37.58492), 30); } -//// Start and finish are located near a big road. -//UNIT_CLASS_TEST(CarTest, BigRoad) -//{ -// TestCarRouter(ms::LatLon(55.75826, 37.39476), ms::LatLon(55.7605, 37.39003), 30); -//} +// Start and finish are located near a big road. +UNIT_CLASS_TEST(CarTest, BigRoad) +{ + TestCarRouter(ms::LatLon(55.75826, 37.39476), ms::LatLon(55.7605, 37.39003), 30); +} -//// Start are located near an airport center. It's far from road network. -//UNIT_CLASS_TEST(CarTest, InAirport) -//{ -// TestCarRouter(ms::LatLon(55.97285, 37.41275), ms::LatLon(55.96396, 37.41922), 30); -//} +// Start are located near an airport center. It's far from road network. +UNIT_CLASS_TEST(CarTest, InAirport) +{ + TestCarRouter(ms::LatLon(55.97285, 37.41275), ms::LatLon(55.96396, 37.41922), 30); +} } // namespace diff --git a/routing/routing_tests/followed_polyline_test.cpp b/routing/routing_tests/followed_polyline_test.cpp index b9f9b889d1..77225ca0e0 100644 --- a/routing/routing_tests/followed_polyline_test.cpp +++ b/routing/routing_tests/followed_polyline_test.cpp @@ -10,8 +10,8 @@ using namespace routing; namespace { - static const m2::PolylineD kTestDirectedPolyline1({{0.0, 0.0}, {3.0, 0.0}, {5.0, 0.0}}); - static const m2::PolylineD kTestDirectedPolyline2({{6.0, 0.0}, {7.0, 0.0}}); + static const m2::PolylineD kTestDirectedPolyline1(std::vector{{0.0, 0.0}, {3.0, 0.0}, {5.0, 0.0}}); + static const m2::PolylineD kTestDirectedPolyline2(std::vector{{6.0, 0.0}, {7.0, 0.0}}); } // namespace UNIT_TEST(FollowedPolylineAppend) @@ -100,7 +100,7 @@ UNIT_TEST(FollowedPolylineDistanceCalculationTest) UNIT_TEST(FollowedPolylineDirectionTest) { - m2::PolylineD testPolyline({{0, 0}, {1.00003, 0}, {1.00003, 1}}); + m2::PolylineD testPolyline(std::vector{{0, 0}, {1.00003, 0}, {1.00003, 1}}); FollowedPolyline polyline(testPolyline.Begin(), testPolyline.End()); TEST_EQUAL(polyline.GetCurrentIter().m_ind, 0, ()); m2::PointD directionPoint; @@ -115,7 +115,7 @@ UNIT_TEST(FollowedPolylineDirectionTest) UNIT_TEST(FollowedPolylineGetDistanceFromBeginM) { - m2::PolylineD testPolyline({{0, 0}, {1, 0}, {2, 0}, {3, 0}, {5, 0}, {6, 0}}); + m2::PolylineD testPolyline(std::vector{{0, 0}, {1, 0}, {2, 0}, {3, 0}, {5, 0}, {6, 0}}); FollowedPolyline polyline(testPolyline.Begin(), testPolyline.End()); m2::PointD point(4, 0); polyline.UpdateProjection(mercator::RectByCenterXYAndSizeInMeters(point, 2)); diff --git a/search/geocoder.cpp b/search/geocoder.cpp index b54f3af383..34ef83db37 100644 --- a/search/geocoder.cpp +++ b/search/geocoder.cpp @@ -1223,7 +1223,7 @@ void Geocoder::GreedilyMatchStreetsWithSuburbs(BaseContext & ctx, auto & layers = ctx.m_layers; ASSERT(layers.empty(), ()); layers.emplace_back(); - SCOPE_GUARD(cleanupGuard, bind(&vector::pop_back, &layers)); + SCOPE_GUARD(cleanupGuard, [&]{ layers.pop_back(); }); auto & layer = layers.back(); InitLayer(Model::TYPE_SUBURB, suburb.m_tokenRange, layer); @@ -1262,7 +1262,7 @@ void Geocoder::CreateStreetsLayerAndMatchLowerLayers(BaseContext & ctx, auto & layers = ctx.m_layers; layers.emplace_back(); - SCOPE_GUARD(cleanupGuard, bind(&vector::pop_back, &layers)); + SCOPE_GUARD(cleanupGuard, [&]{ layers.pop_back(); }); auto & layer = layers.back(); InitLayer(Model::TYPE_STREET, prediction.m_tokenRange, layer); @@ -1353,7 +1353,7 @@ void Geocoder::MatchPOIsAndBuildings(BaseContext & ctx, size_t curToken, CBV con // Following code creates a fake layer with buildings and // intersects it with the streets layer. layers.emplace_back(); - SCOPE_GUARD(cleanupGuard, bind(&vector::pop_back, &layers)); + SCOPE_GUARD(cleanupGuard, [&]{ layers.pop_back(); }); auto & layer = layers.back(); InitLayer(Model::TYPE_BUILDING, m_postcodes.m_tokenRange, layer); @@ -1367,7 +1367,7 @@ void Geocoder::MatchPOIsAndBuildings(BaseContext & ctx, size_t curToken, CBV con } layers.emplace_back(); - SCOPE_GUARD(cleanupGuard, bind(&vector::pop_back, &layers)); + SCOPE_GUARD(cleanupGuard, [&]{ layers.pop_back(); }); // Clusters of features by search type. Each cluster is a sorted // list of ids. diff --git a/tools/python/maps_generator/__main__.py b/tools/python/maps_generator/__main__.py index 0cedd6d164..5734b08d21 100644 --- a/tools/python/maps_generator/__main__.py +++ b/tools/python/maps_generator/__main__.py @@ -278,6 +278,9 @@ def main(): if not settings.NEED_PLANET_UPDATE: skipped_stages.add(sd.StageUpdatePlanet) + if not settings.NEED_BUILD_WORLD_ROADS: + skipped_stages.add(sd.StageRoutingWorld) + # Make env and run maps generation. env = Env( countries=countries, diff --git a/tools/python/maps_generator/generator/env.py b/tools/python/maps_generator/generator/env.py index 52808d7024..cb63c545ed 100644 --- a/tools/python/maps_generator/generator/env.py +++ b/tools/python/maps_generator/generator/env.py @@ -231,10 +231,10 @@ class PathProvider: ) @property - def worldroads_path(self) -> AnyStr: + def world_roads_path(self) -> AnyStr: return ( - os.path.join(self.intermediate_data_path, "worldroads.txt") - if settings.WORLDROADS_URL + os.path.join(self.intermediate_data_path, "world_roads.txt") + if settings.NEED_BUILD_WORLD_ROADS else "" ) @@ -434,6 +434,8 @@ class Env: if item.endswith(".download"): os.remove(os.path.join(self.paths.status_path, item)) + self.world_roads_builder_tool = self.setup_world_roads_builder_tool() + self.main_status = status.Status() # self.countries_meta stores log files and statuses for each country. self.countries_meta = collections.defaultdict(dict) @@ -523,6 +525,14 @@ class Env: raise Exception(exceptions) + @staticmethod + def setup_world_roads_builder_tool() -> AnyStr: + logger.info(f"Check world_roads_builder_tool. Looking for it in {settings.BUILD_PATH} ...") + world_roads_builder_tool_path = find_executable(settings.BUILD_PATH, "world_roads_builder_tool") + logger.info(f"world_roads_builder_tool found - {world_roads_builder_tool_path}") + return world_roads_builder_tool_path + + @staticmethod def setup_osm_tools() -> Dict[AnyStr, AnyStr]: path = settings.OSM_TOOLS_PATH diff --git a/tools/python/maps_generator/generator/gen_tool.py b/tools/python/maps_generator/generator/gen_tool.py index 3f4de37c43..b11216e3b1 100644 --- a/tools/python/maps_generator/generator/gen_tool.py +++ b/tools/python/maps_generator/generator/gen_tool.py @@ -74,7 +74,7 @@ class GenTool: "srtm_path": str, "transit_path": str, "transit_path_experimental": str, - "worldroads_path": str, + "world_roads_path": str, "ugc_data": str, "uk_postcodes_dataset": str, "us_postcodes_dataset": str, diff --git a/tools/python/maps_generator/generator/settings.py b/tools/python/maps_generator/generator/settings.py index 0b7a38f5cb..a16dd0a0fd 100644 --- a/tools/python/maps_generator/generator/settings.py +++ b/tools/python/maps_generator/generator/settings.py @@ -115,7 +115,7 @@ PROMO_CATALOG_COUNTRIES_URL = "" POPULARITY_URL = "" SUBWAY_URL = "" TRANSIT_URL = "" -WORLDROADS_URL = "" +NEED_BUILD_WORLD_ROADS = True FOOD_URL = "" FOOD_TRANSLATIONS_URL = "" UK_POSTCODES_URL = "" @@ -277,7 +277,7 @@ def init(default_settings_path: AnyStr): global POPULARITY_URL global SUBWAY_URL global TRANSIT_URL - global WORLDROADS_URL + global NEED_BUILD_WORLD_ROADS global FOOD_URL global UK_POSTCODES_URL global US_POSTCODES_URL @@ -301,7 +301,7 @@ def init(default_settings_path: AnyStr): POPULARITY_URL = cfg.get_opt_path("External", "POPULARITY_URL", POPULARITY_URL) SUBWAY_URL = cfg.get_opt("External", "SUBWAY_URL", SUBWAY_URL) TRANSIT_URL = cfg.get_opt("External", "TRANSIT_URL", TRANSIT_URL) - WORLDROADS_URL = cfg.get_opt("External", "WORLDROADS_URL", WORLDROADS_URL) + NEED_BUILD_WORLD_ROADS = cfg.get_opt("External", "NEED_BUILD_WORLD_ROADS", NEED_BUILD_WORLD_ROADS) FOOD_URL = cfg.get_opt("External", "FOOD_URL", FOOD_URL) UK_POSTCODES_URL = cfg.get_opt("External", "UK_POSTCODES_URL", UK_POSTCODES_URL) diff --git a/tools/python/maps_generator/generator/stages_declaration.py b/tools/python/maps_generator/generator/stages_declaration.py index 1468abf2c2..d1aa76ae8b 100644 --- a/tools/python/maps_generator/generator/stages_declaration.py +++ b/tools/python/maps_generator/generator/stages_declaration.py @@ -112,7 +112,9 @@ class StagePreprocess(Stage): D(settings.FOOD_URL, PathProvider.food_paths, "p"), D(settings.FOOD_TRANSLATIONS_URL, PathProvider.food_translations_path, "p"), ) -@test_stage(Test(st.make_test_booking_data(max_days=7), lambda e, _: e.production, True)) +@test_stage( + Test(st.make_test_booking_data(max_days=7), lambda e, _: e.production, True) +) class StageFeatures(Stage): def apply(self, env: Env): extra = {} @@ -179,7 +181,13 @@ class StageMwm(Stage): @staticmethod def make_mwm(country: AnyStr, env: Env): world_stages = { - WORLD_NAME: [StageIndex, StageCitiesIdsWorld, StageRoutingWorld, StageMwmStatistics], + WORLD_NAME: [ + StageIndex, + StageCitiesIdsWorld, + StagePrepareRoutingWorld, + StageRoutingWorld, + StageMwmStatistics, + ], WORLD_COASTS_NAME: [StageIndex, StageMwmStatistics], } @@ -232,7 +240,13 @@ class StageCitiesIdsWorld(Stage): @country_stage -@depends_from_internal(D(settings.WORLDROADS_URL, PathProvider.worldroads_path),) +@helper_stage_for("StageRoutingWorld") +class StagePrepareRoutingWorld(Stage): + def apply(self, env: Env, country, **kwargs): + steps.step_prepare_routing_world(env, country, **kwargs) + + +@country_stage class StageRoutingWorld(Stage): def apply(self, env: Env, country, **kwargs): steps.step_routing_world(env, country, **kwargs) @@ -295,12 +309,11 @@ class StageRoutingTransit(Stage): class StageMwmDiffs(Stage): def apply(self, env: Env, country, logger, **kwargs): data_dir = diffs.DataDir( - mwm_name=env.build_name, new_version_dir=env.build_path, - old_version_root_dir=settings.DATA_ARCHIVE_DIR - ) - diffs.mwm_diff_calculation( - data_dir, logger, depth=settings.DIFF_VERSION_DEPTH + mwm_name=env.build_name, + new_version_dir=env.build_path, + old_version_root_dir=settings.DATA_ARCHIVE_DIR, ) + diffs.mwm_diff_calculation(data_dir, logger, depth=settings.DIFF_VERSION_DEPTH) @country_stage @@ -316,7 +329,8 @@ class StageMwmStatistics(Stage): settings.PROMO_CATALOG_COUNTRIES_URL, PathProvider.promo_catalog_countries_path, "p", - ) + ), + D(settings.PROMO_CATALOG_CITIES_URL, PathProvider.promo_catalog_cities_path, "p"), ) class StageCountriesTxt(Stage): def apply(self, env: Env): @@ -329,9 +343,8 @@ class StageCountriesTxt(Stage): env.paths.mwm_version, ) if env.production: - countries_json = json.loads(countries) inject_promo_ids( - countries_json, + countries, env.paths.promo_catalog_cities_path, env.paths.promo_catalog_countries_path, env.paths.mwm_path, @@ -339,8 +352,8 @@ class StageCountriesTxt(Stage): env.paths.mwm_path, ) - with open(env.paths.counties_txt_path, "w") as f: - json.dump(countries_json, f, ensure_ascii=True, indent=1) + with open(env.paths.counties_txt_path, "w") as f: + json.dump(countries, f, ensure_ascii=True, indent=1) @outer_stage diff --git a/tools/python/maps_generator/generator/steps.py b/tools/python/maps_generator/generator/steps.py index 94567b4484..46fcfcbd6a 100644 --- a/tools/python/maps_generator/generator/steps.py +++ b/tools/python/maps_generator/generator/steps.py @@ -16,6 +16,7 @@ from maps_generator.generator.env import WORLDS_NAMES from maps_generator.generator.env import WORLD_NAME from maps_generator.generator.env import get_all_countries_list from maps_generator.generator.exceptions import ValidationError +from maps_generator.generator.exceptions import wait_and_raise_if_fail from maps_generator.generator.gen_tool import run_gen_tool from maps_generator.generator.osmtools import osmconvert from maps_generator.generator.osmtools import osmupdate @@ -200,6 +201,22 @@ def step_cities_ids_world(env: Env, country: AnyStr, **kwargs): ) +def step_prepare_routing_world(env: Env, country: AnyStr, **kwargs): + world_roads_builder_tool_with_args = [env.world_roads_builder_tool, + f"--path_roads_file={env.paths.planet_o5m}", + f"--path_resources={env.paths.user_resource_path}", + f"--path_res_file={env.paths.world_roads_path}"] + logger.info(f"Starting {world_roads_builder_tool_with_args}") + sub_proc = subprocess.Popen( + world_roads_builder_tool_with_args, + stdout=env.get_subprocess_out(country), + stderr=env.get_subprocess_out(country), + env=os.environ + ) + + wait_and_raise_if_fail(sub_proc) + + def step_routing_world(env: Env, country: AnyStr, **kwargs): run_gen_tool_with_recovery_country( env, @@ -209,7 +226,7 @@ def step_routing_world(env: Env, country: AnyStr, **kwargs): data_path=env.paths.mwm_path, user_resource_path=env.paths.user_resource_path, output=country, - worldroads_path=env.paths.worldroads_path, + world_roads_path=env.paths.world_roads_path, **kwargs, ) diff --git a/tools/python/maps_generator/utils/file.py b/tools/python/maps_generator/utils/file.py index 1036d8b9c7..f97bcfae2a 100644 --- a/tools/python/maps_generator/utils/file.py +++ b/tools/python/maps_generator/utils/file.py @@ -54,7 +54,14 @@ def download_file(url: AnyStr, name: AnyStr, download_if_exists: bool = True): session.mount("file://", FileAdapter()) with open(tmp_name, "wb") as handle: response = session.get(url, stream=True) - file_length = int(response.headers["Content-Length"]) + file_length = None + try: + file_length = int(response.headers["Content-Length"]) + except KeyError: + logger.warning( + f"There is no attribute Content-Length in headers [{url}]: {response.headers}" + ) + current = 0 max_attempts = 32 attempts = max_attempts @@ -63,7 +70,7 @@ def download_file(url: AnyStr, name: AnyStr, download_if_exists: bool = True): current += len(data) handle.write(data) - if file_length == current: + if file_length is None or file_length == current: break logger.warning( diff --git a/tools/python/maps_generator/var/etc/map_generator.ini.default b/tools/python/maps_generator/var/etc/map_generator.ini.default index fd3c447e07..b9a2b7b377 100644 --- a/tools/python/maps_generator/var/etc/map_generator.ini.default +++ b/tools/python/maps_generator/var/etc/map_generator.ini.default @@ -57,15 +57,15 @@ DIFF_VERSION_DEPTH: 2 # The https://somesite.com/download/latest_coasts.geom url will be used to download latest_coasts.geom and # the https://somesite.com/download/latest_coasts.rawgeom url will be used to download latest_coasts.rawgeom. # PLANET_COASTS_URL: +# Set to 'true' to build special routing section in World.mwm for alerting about absent regions without which the +# route can't be built. +NEED_BUILD_WORLD_ROADS: true # The url to the subway file. SUBWAY_URL: http://osm-subway.maps.me/mapsme/latest.json # The url of the location with the transit files extracted from GTFS. # TRANSIT_URL: -# The url of the file with info about main roads between mwms for special routing section in World mwm. -# WORLDROADS_URL: - # Urls for production maps generation. # UGC_URL: # HOTELS_URL: diff --git a/tools/python/post_generation/__main__.py b/tools/python/post_generation/__main__.py index 0b97f9647a..dd0a0a686a 100644 --- a/tools/python/post_generation/__main__.py +++ b/tools/python/post_generation/__main__.py @@ -90,7 +90,7 @@ The post_generation commands are: help="Output countries.txt file (default is stdout)", ) args = parser.parse_args(sys.argv[2:]) - countries_json = hierarchy_to_countries_( + countries = hierarchy_to_countries_( args.old, args.osm, args.countries_synonyms, @@ -100,9 +100,9 @@ The post_generation commands are: ) if args.output: with open(args.output, "w") as f: - f.write(countries_json) + json.dump(countries, f, ensure_ascii=True, indent=1) else: - print(countries_json) + print(json.dumps(countries, ensure_ascii=True, indent=1)) @staticmethod def inject_promo_ids(): @@ -152,7 +152,7 @@ The post_generation commands are: ) with open(args.output, "w") as f: - json.dump(countries, f, indent=1) + json.dump(countries, f, ensure_ascii=True, indent=1) PostGeneration() diff --git a/tools/python/post_generation/hierarchy_to_countries.py b/tools/python/post_generation/hierarchy_to_countries.py index 777f4f0079..9853580080 100755 --- a/tools/python/post_generation/hierarchy_to_countries.py +++ b/tools/python/post_generation/hierarchy_to_countries.py @@ -185,4 +185,4 @@ def hierarchy_to_countries( stack[-1]["g"].append(g) collapse_single(stack[-1]) - return json.dumps(stack[-1], ensure_ascii=True, indent=1) + return stack[-1] diff --git a/tools/python/transit/gtfs/download_gtfs.py b/tools/python/transit/gtfs/download_gtfs.py index ac18c55ce8..ac895cd1bf 100644 --- a/tools/python/transit/gtfs/download_gtfs.py +++ b/tools/python/transit/gtfs/download_gtfs.py @@ -14,8 +14,7 @@ import zipfile import requests -MAX_RETRIES = 3 -AVG_SLEEP_TIMEOUT_S = 10 +MAX_RETRIES = 2 MAX_SLEEP_TIMEOUT_S = 30 URLS_FILE_TRANSITLAND = "feed_urls_transitland.txt" @@ -35,9 +34,6 @@ def get_feeds_links(data): for feed in data: if feed["feed_format"] != "gtfs" or feed["spec"] != "gtfs": - # Warning about strange format - not gtfs and not real-time gtfs: - if feed["feed_format"] != "gtfs-rt": - logger.warning(f"Skipped feed: feed_format {feed['feed_format']}, spec {feed['spec']}") continue if "url" in feed and feed["url"] is not None and feed["url"]: @@ -51,16 +47,9 @@ def parse_transitland_page(url): retries = MAX_RETRIES while retries > 0: - with requests.get(url) as response: - if response.status_code != 200: - logger.error(f"Failed loading feeds: {response.status_code}") - if response.status_code == 429: - logger.error("Too many requests.") - time.sleep(MAX_SLEEP_TIMEOUT_S) - else: - time.sleep(AVG_SLEEP_TIMEOUT_S) - retries -= 1 - continue + try: + response = requests.get(url) + response.raise_for_status() data = json.loads(response.text) if "feeds" in data: @@ -71,6 +60,17 @@ def parse_transitland_page(url): next_page = data["meta"]["next"] if "next" in data["meta"] else "" return gtfs_feeds_urls, next_page + except requests.exceptions.HTTPError as http_err: + logger.error(f"HTTP error {http_err} downloading zip from {url}") + if http_err == 429: + time.sleep(MAX_SLEEP_TIMEOUT_S) + except requests.exceptions.RequestException as ex: + logger.error( + f"Exception {ex} while parsing Transitland url {url} with code {response.status_code}" + ) + + retries -= 1 + return [], "" @@ -82,7 +82,10 @@ def extract_to_path(content, out_path): return True except zipfile.BadZipfile: logger.exception("BadZipfile exception.") - return False + except Exception as e: + logger.exception(f"Exception while unzipping feed: {e}") + + return False def load_gtfs_feed_zip(path, url): @@ -90,24 +93,21 @@ def load_gtfs_feed_zip(path, url): retries = MAX_RETRIES while retries > 0: try: - with requests.get(url, stream=True) as response: - if response.status_code != 200: - logger.error(f"HTTP code {response.status_code} loading gtfs {url}") - retries -= 1 - time.sleep(MAX_SLEEP_TIMEOUT_S) - continue + response = requests.get(url, stream=True) + response.raise_for_status() - if not extract_to_path(response.content, path): - retries -= 1 - logger.error(f"Could not extract zip: {url}") - continue + if not extract_to_path(response.content, path): + retries -= 1 + logger.error(f"Could not extract zip: {url}") + continue - return True + return True + except requests.exceptions.HTTPError as http_err: + logger.error(f"HTTP error {http_err} downloading zip from {url}") except requests.exceptions.RequestException as ex: - logger.error(f"Exception {ex} for url {url}") - retries -= 1 - time.sleep(AVG_SLEEP_TIMEOUT_S) + logger.error(f"Exception {ex} downloading zip from {url}") + retries -= 1 return False @@ -125,35 +125,52 @@ def parse_openmobilitydata_pages(omd_api_key): "location": "undefined", "descendants": 1, "limit": 100, - "type": "gtfs" + "type": "gtfs", } - with requests.get(url_page, params=params, headers=HEADERS_OMD) as response: - if response.status_code != 200: - logger.error(f"Http code {response.status_code} loading feed ids: {url_page}") - return [], "" + try: + with requests.get(url_page, params=params, headers=HEADERS_OMD) as response: + if response.status_code != 200: + logger.error( + f"Http code {response.status_code} loading feed ids: {url_page}" + ) + return [], "" - data = json.loads(response.text) + data = json.loads(response.text) - if page == 1: - pages_count = data["results"]["numPages"] - logger.info(f"Pages count {pages_count}") + if page == 1: + pages_count = data["results"]["numPages"] + logger.info( + f"There are {pages_count} Openmobilitydata pages with feed urls." + ) - for feed in data["results"]["feeds"]: - params = { - "key": omd_api_key, - "feed": feed["id"] - } + for feed in data["results"]["feeds"]: + params = {"key": omd_api_key, "feed": feed["id"]} + try: + with requests.get( + url_with_redirect, + params=params, + headers=HEADERS_OMD, + allow_redirects=True, + ) as response_redirect: + if response_redirect.history: + urls.append(response_redirect.url) + else: + logger.error( + f"Could not get link to zip with feed {feed['id']} from {url_with_redirect}" + ) + except requests.exceptions.RequestException as ex_redirect: + logger.error( + f"Exception {ex_redirect} while getting link to zip with " + f"feed {feed['id']} from {url_with_redirect}" + ) + except requests.exceptions.RequestException as ex: + logger.error( + f"Exception {ex} while getting {url_page} (page {page}) from Openmobilitydata." + ) - with requests.get(url_with_redirect, params=params, headers=HEADERS_OMD, allow_redirects=True) \ - as response_redirect: - if response_redirect.history: - urls.append(response_redirect.url) - else: - logger.error(f"Could not get link to zip with feed {feed['id']} from {url_with_redirect}") - - logger.info(f"page {page}") - page += 1 + logger.info(f"Crawled {page}/{pages_count} page of Openmobilitydata.") + page += 1 return urls @@ -184,10 +201,7 @@ def crawl_transitland_for_feed_urls(out_path): def get_filename(file_prefix, index): - index_str = str(index) - index_len = len(index_str) - zeroes_prefix = "" if MAX_INDEX_LEN < index_len else "0" * (MAX_INDEX_LEN - index_len) - return file_prefix + "_" + zeroes_prefix + index_str + return f"{file_prefix}_{index:0{MAX_INDEX_LEN}d}" def load_gtfs_zips_from_urls(path, urls_file, threads_count, file_prefix): @@ -200,11 +214,18 @@ def load_gtfs_zips_from_urls(path, urls_file, threads_count, file_prefix): err_count = 0 with concurrent.futures.ThreadPoolExecutor(max_workers=threads_count) as executor: - future_to_url = {executor.submit(load_gtfs_feed_zip, - os.path.join(path, get_filename(file_prefix, i)), url): - url for i, url in enumerate(urls)} + future_to_url = { + executor.submit( + load_gtfs_feed_zip, + os.path.join(path, get_filename(file_prefix, i)), + url, + ): url + for i, url in enumerate(urls) + } - for j, future in enumerate(concurrent.futures.as_completed(future_to_url), start=1): + for j, future in enumerate( + concurrent.futures.as_completed(future_to_url), start=1 + ): url = future_to_url[future] loaded = future.result() @@ -227,25 +248,45 @@ def main(): """Downloads urls of feeds from feed aggregators and saves to the file. Downloads feeds from these urls and saves to the directory.""" - parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter) + parser = argparse.ArgumentParser( + formatter_class=argparse.ArgumentDefaultsHelpFormatter + ) parser.add_argument("-p", "--path", required=True, help="working directory path") - parser.add_argument("-m", "--mode", required=True, - help="fullrun | load_feed_urls | load_feed_zips") + parser.add_argument( + "-m", "--mode", required=True, help="fullrun | load_feed_urls | load_feed_zips" + ) - parser.add_argument("-s", "--source", default="transitland", - help="source of feeds: transitland | openmobilitydata | all") + parser.add_argument( + "-s", + "--source", + default="transitland", + help="source of feeds: transitland | openmobilitydata | all", + ) - parser.add_argument("-t", "--threads", type=int, default=THREADS_COUNT, - help="threads count for loading zips") + parser.add_argument( + "-t", + "--threads", + type=int, + default=THREADS_COUNT, + help="threads count for loading zips", + ) - parser.add_argument("-k", "--omd_api_key", default="", - help="user key for working with openmobilitydata API") + parser.add_argument( + "-k", + "--omd_api_key", + default="", + help="user key for working with openmobilitydata API", + ) args = parser.parse_args() - logging.basicConfig(filename=os.path.join(args.path, "crawling.log"), filemode="w", level=logging.INFO) + logging.basicConfig( + filename=os.path.join(args.path, "crawling.log"), + filemode="w", + level=logging.INFO, + ) if args.mode in ["fullrun", "load_feed_urls"]: @@ -253,14 +294,18 @@ def main(): crawl_transitland_for_feed_urls(args.path) if args.source in ["all", "openmobilitydata"]: if not args.omd_api_key: - logger.error("No key provided for openmobilitydata. Set omd_api_key argument.") + logger.error( + "No key provided for openmobilitydata. Set omd_api_key argument." + ) return crawl_openmobilitydata_for_feed_urls(args.path, args.omd_api_key) if args.mode in ["fullrun", "load_feed_zips"]: if args.source in ["all", "transitland"]: - load_gtfs_zips_from_urls(args.path, URLS_FILE_TRANSITLAND, args.threads, "tl") + load_gtfs_zips_from_urls( + args.path, URLS_FILE_TRANSITLAND, args.threads, "tl" + ) if args.source in ["all", "openmobilitydata"]: load_gtfs_zips_from_urls(args.path, URLS_FILE_OMD, args.threads, "omd") diff --git a/transit/world_feed/feed_helpers.cpp b/transit/world_feed/feed_helpers.cpp index 82e34582ca..f88299d7ad 100644 --- a/transit/world_feed/feed_helpers.cpp +++ b/transit/world_feed/feed_helpers.cpp @@ -159,13 +159,19 @@ std::pair PrepareNearestPointOnTrack(m2::PointD const & point, return {proj->m_indexOnShape, proj->m_needsInsertion}; } -bool IsRelevantType(const gtfs::RouteType & routeType) +bool IsRelevantType(gtfs::RouteType const & routeType) { // All types and constants are described in GTFS: // https://developers.google.com/transit/gtfs/reference + auto const isSubway = [](gtfs::RouteType const & routeType) { + return routeType == gtfs::RouteType::Subway || + routeType == gtfs::RouteType::MetroService || + routeType == gtfs::RouteType::UndergroundService; + }; + // We skip all subways because we extract subway data from OSM, not from GTFS. - if (routeType == gtfs::RouteType::Subway) + if (isSubway(routeType)) return false; auto const val = static_cast(routeType); @@ -183,8 +189,6 @@ bool IsRelevantType(const gtfs::RouteType & routeType) gtfs::RouteType::CarTransportRailService, gtfs::RouteType::LorryTransportRailService, gtfs::RouteType::VehicleTransportRailService, - gtfs::RouteType::MetroService, - gtfs::RouteType::UndergroundService, gtfs::RouteType::PostBusService, gtfs::RouteType::SpecialNeedsBus, gtfs::RouteType::MobilityBusService,