# coding: utf-8 """ This module contains all the filtering functions. It exposes ``first_pass`` and ``second_pass`` functions which are a set of filters applied during the first pass and the second pass. """ from __future__ import absolute_import, print_function, unicode_literals import logging from flatisfy import tools from flatisfy.filters import duplicates from flatisfy.filters import metadata LOGGER = logging.getLogger(__name__) def refine_with_housing_criteria(flats_list, config): """ Filter a list of flats according to criteria. Housings posts websites tend to return broader results that what was actually asked for. Then, we should filter out the list to match the user criteria, and avoid exposing unwanted flats. :param flats_list: A list of flats dict to filter. :param config: A config dict. :return: A tuple of flats to keep and flats to delete. """ # For each flat, the associated `is_ok` value indicate whether it should be # kept or discarded. is_ok = [True for _ in flats_list] for i, flat in enumerate(flats_list): # Check postal code postal_code = flat["flatisfy"].get("postal_code", None) if ( postal_code and postal_code not in config["constraints"]["postal_codes"] ): LOGGER.info("Postal code for flat %s is out of range.", flat["id"]) is_ok[i] = is_ok[i] and False # Check time_to for place_name, time in flat["flatisfy"].get("time_to", {}).items(): time = time["time"] is_within_interval = tools.is_within_interval( time, *(config["constraints"]["time_to"][place_name]["time"]) ) if not is_within_interval: LOGGER.info("Flat %s is too far from place %s: %ds.", flat["id"], place_name, time) is_ok[i] = is_ok[i] and is_within_interval # Check other fields for field in ["area", "cost", "rooms", "bedrooms"]: interval = config["constraints"][field] is_within_interval = tools.is_within_interval( flat.get(field, None), *interval ) if not is_within_interval: LOGGER.info("%s for flat %s is out of range.", field.capitalize(), flat["id"]) is_ok[i] = is_ok[i] and is_within_interval return ( [ flat for i, flat in enumerate(flats_list) if is_ok[i] ], [ flat for i, flat in enumerate(flats_list) if not is_ok[i] ] ) def first_pass(flats_list, config): """ First filtering pass. Flatboob only fetches data from the listing of the available housing. Then, we should do a first pass to filter based on the already available data and only request more data for the remaining housings. :param flats_list: A list of flats dict to filter. :param config: A config dict. :return: A dict mapping flat status and list of flat objects. """ LOGGER.info("Running first filtering pass.") # Handle duplicates based on ids # Just remove them (no merge) as they should be the exact same object. flats_list, duplicates_by_id = duplicates.detect( flats_list, key="id", merge=False, should_intersect=False ) # Also merge duplicates based on urls (these may come from different # flatboob backends) # This is especially useful as some websites such as entreparticuliers # contains a lot of leboncoin housings posts. flats_list, duplicates_by_urls = duplicates.detect( flats_list, key="urls", merge=True, should_intersect=True ) # Guess the postal codes flats_list = metadata.guess_postal_code(flats_list, config) # Try to match with stations flats_list = metadata.guess_stations(flats_list, config) # Remove returned housing posts that do not match criteria flats_list, ignored_list = refine_with_housing_criteria(flats_list, config) return { "new": flats_list, "ignored": ignored_list, "duplicate": duplicates_by_id + duplicates_by_urls } def second_pass(flats_list, config): """ Second filtering pass. This pass is expected to have as most information as possible on the available housings. Plus it runs after first pass which already consolidated data. It should consolidate everything and try to extract as many data as possible from the fetched housings. :param flats_list: A list of flats dict to filter. :param config: A config dict. :return: A dict mapping flat status and list of flat objects. """ LOGGER.info("Running second filtering pass.") # Assumed to run after first pass, so there should be no obvious duplicates # left and we already tried to find postal code and nearby stations. # Confirm postal code flats_list = metadata.guess_postal_code(flats_list, config) # Better match with stations (confirm and check better) flats_list = metadata.guess_stations(flats_list, config) # Compute travel time to specified points flats_list = metadata.compute_travel_times(flats_list, config) # Remove returned housing posts that do not match criteria flats_list, ignored_list = refine_with_housing_criteria(flats_list, config) return { "new": flats_list, "ignored": ignored_list, "duplicate": [] } def third_pass(flats_list, config): """ Third filtering pass. This pass is expected to perform deep duplicate detection on available flats. :param flats_list: A list of flats dict to filter. :param config: A config dict. :return: A dict mapping flat status and list of flat objects. """ # Deduplicate the list using every available data flats_list, duplicate_flats = duplicates.deep_detect(flats_list) return { "new": flats_list, "ignored": [], "duplicate": duplicate_flats }