| """Places Dataset Loading Script for Hugging Face""" |
|
|
| import csv |
| import json |
| import os |
| import datasets |
| from typing import Dict, List, Any |
|
|
|
|
| _CITATION = """\ |
| @dataset{places_dataset_2025, |
| title={Places Dataset}, |
| author={patrick piemonte}, |
| year={2025}, |
| publisher={Hugging Face} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| This dataset contains information about close to 70,000 places with associated metadata including |
| locations, attribution tags, and contact details. The data includes geographic coordinates, |
| place descriptions, categorization through attribution tags, and social media presence information. |
| """ |
|
|
| _HOMEPAGE = "" |
| _LICENSE = "cc-by-4.0" |
|
|
| _URLS = { |
| "place": "place.csv", |
| "location": "location.csv", |
| "place_contact": "place_contact.csv", |
| "tag": "tag.csv", |
| "place_tag": "place_tag.csv", |
| } |
|
|
|
|
| class PlacesDataset(datasets.GeneratorBasedBuilder): |
| """Places dataset with multiple related tables.""" |
|
|
| VERSION = datasets.Version("1.0.0") |
|
|
| BUILDER_CONFIGS = [ |
| datasets.BuilderConfig(name="all", version=VERSION, description="Load all tables"), |
| datasets.BuilderConfig(name="place", version=VERSION, description="Main places table"), |
| datasets.BuilderConfig(name="location", version=VERSION, description="Geographic coordinates"), |
| datasets.BuilderConfig(name="place_contact", version=VERSION, description="Contact information"), |
| datasets.BuilderConfig(name="tag", version=VERSION, description="Categorization tags"), |
| datasets.BuilderConfig(name="place_tag", version=VERSION, description="Place-tag relationships"), |
| datasets.BuilderConfig(name="denormalized", version=VERSION, description="Denormalized view with places, locations, and primary tags"), |
| ] |
|
|
| DEFAULT_CONFIG_NAME = "all" |
|
|
| def _info(self): |
| if self.config.name == "place": |
| features = datasets.Features({ |
| "id": datasets.Value("string"), |
| "created_at": datasets.Value("string"), |
| "name": datasets.Value("string"), |
| "description": datasets.Value("string"), |
| "address": datasets.Value("string"), |
| "address_formatted": datasets.Value("string"), |
| "cross_street": datasets.Value("string"), |
| "locality": datasets.Value("string"), |
| "administrative_area": datasets.Value("string"), |
| "postal_code": datasets.Value("string"), |
| "country_code": datasets.Value("string"), |
| "verified": datasets.Value("bool"), |
| "flagged": datasets.Value("bool"), |
| "place_contact_id": datasets.Value("string"), |
| "location_id": datasets.Value("string"), |
| "author_id": datasets.Value("string"), |
| "owner_id": datasets.Value("string"), |
| "locale_id": datasets.Value("string"), |
| "primary_tag_id": datasets.Value("string"), |
| "country": datasets.Value("string"), |
| "sublocality": datasets.Value("string"), |
| "subadministrative_area": datasets.Value("string"), |
| "updated_at": datasets.Value("string"), |
| "radius_in_meters": datasets.Value("int32"), |
| "stamp_id": datasets.Value("string"), |
| "z_priority": datasets.Value("int32"), |
| "clustering_category": datasets.Value("string"), |
| "places_token_id": datasets.Value("string"), |
| "nano_id": datasets.Value("string"), |
| "slug": datasets.Value("string"), |
| }) |
| elif self.config.name == "location": |
| features = datasets.Features({ |
| "id": datasets.Value("string"), |
| "latitude": datasets.Value("float64"), |
| "longitude": datasets.Value("float64"), |
| "horizontal_accuracy": datasets.Value("float32"), |
| "altitude": datasets.Value("float32"), |
| "vertical_accuracy": datasets.Value("float32"), |
| "geom": datasets.Value("string"), |
| "geog": datasets.Value("string"), |
| }) |
| elif self.config.name == "place_contact": |
| features = datasets.Features({ |
| "id": datasets.Value("string"), |
| "instagram": datasets.Value("string"), |
| "x": datasets.Value("string"), |
| "website": datasets.Value("string"), |
| }) |
| elif self.config.name == "tag": |
| features = datasets.Features({ |
| "id": datasets.Value("string"), |
| "created_at": datasets.Value("string"), |
| "name": datasets.Value("string"), |
| "slug": datasets.Value("string"), |
| "private_tag": datasets.Value("bool"), |
| "tag_type": datasets.Value("string"), |
| "image_asset_id": datasets.Value("string"), |
| "theme_asset_id": datasets.Value("string"), |
| "search_tokens": datasets.Value("string"), |
| "keywords": datasets.Value("string"), |
| "radius_in_meters": datasets.Value("int32"), |
| "content_rating": datasets.Value("string"), |
| "stamp_id": datasets.Value("string"), |
| "hidden_tag": datasets.Value("bool"), |
| }) |
| elif self.config.name == "place_tag": |
| features = datasets.Features({ |
| "id": datasets.Value("string"), |
| "tag_id": datasets.Value("string"), |
| "place_id": datasets.Value("string"), |
| "created_at": datasets.Value("string"), |
| }) |
| elif self.config.name == "denormalized": |
| features = datasets.Features({ |
| |
| "place_id": datasets.Value("string"), |
| "name": datasets.Value("string"), |
| "description": datasets.Value("string"), |
| "address": datasets.Value("string"), |
| "address_formatted": datasets.Value("string"), |
| "locality": datasets.Value("string"), |
| "administrative_area": datasets.Value("string"), |
| "postal_code": datasets.Value("string"), |
| "country_code": datasets.Value("string"), |
| "country": datasets.Value("string"), |
| "verified": datasets.Value("bool"), |
| |
| "latitude": datasets.Value("float64"), |
| "longitude": datasets.Value("float64"), |
| "horizontal_accuracy": datasets.Value("float32"), |
| "altitude": datasets.Value("float32"), |
| |
| "primary_tag_name": datasets.Value("string"), |
| "primary_tag_slug": datasets.Value("string"), |
| "primary_tag_type": datasets.Value("string"), |
| |
| "website": datasets.Value("string"), |
| "instagram": datasets.Value("string"), |
| "twitter": datasets.Value("string"), |
| }) |
| else: |
| features = datasets.Features({ |
| "table_name": datasets.Value("string"), |
| "data": datasets.Value("string"), |
| }) |
|
|
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| if self.config.name == "denormalized": |
| downloaded_files = dl_manager.download_and_extract(_URLS) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "filepaths": downloaded_files, |
| }, |
| ), |
| ] |
| elif self.config.name != "all": |
| downloaded_file = dl_manager.download_and_extract(_URLS[self.config.name]) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "filepath": downloaded_file, |
| "table_name": self.config.name, |
| }, |
| ), |
| ] |
| else: |
| downloaded_files = dl_manager.download_and_extract(_URLS) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "filepaths": downloaded_files, |
| }, |
| ), |
| ] |
|
|
| def _generate_examples(self, filepath=None, table_name=None, filepaths=None): |
| if self.config.name == "denormalized": |
| |
| places = {} |
| locations = {} |
| tags = {} |
| contacts = {} |
| |
| |
| with open(filepaths["place"], encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| for row in reader: |
| places[row["id"]] = row |
| |
| |
| with open(filepaths["location"], encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| for row in reader: |
| locations[row["id"]] = row |
| |
| |
| with open(filepaths["tag"], encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| for row in reader: |
| tags[row["id"]] = row |
| |
| |
| with open(filepaths["place_contact"], encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| for row in reader: |
| contacts[row["id"]] = row |
| |
| |
| idx = 0 |
| for place_id, place in places.items(): |
| result = { |
| "place_id": place_id, |
| "name": place.get("name"), |
| "description": place.get("description"), |
| "address": place.get("address"), |
| "address_formatted": place.get("address_formatted"), |
| "locality": place.get("locality"), |
| "administrative_area": place.get("administrative_area"), |
| "postal_code": place.get("postal_code"), |
| "country_code": place.get("country_code"), |
| "country": place.get("country"), |
| "verified": place.get("verified") == "true" if place.get("verified") else False, |
| } |
| |
| |
| if place.get("location_id") and place["location_id"] in locations: |
| loc = locations[place["location_id"]] |
| |
| lat = None |
| lon = None |
| if loc.get("latitude"): |
| try: |
| lat = float(loc["latitude"]) |
| if not (-90 <= lat <= 90): |
| print(f"Warning: Invalid latitude {lat} for place {place_id}") |
| lat = None |
| except ValueError: |
| pass |
| if loc.get("longitude"): |
| try: |
| lon = float(loc["longitude"]) |
| if not (-180 <= lon <= 180): |
| print(f"Warning: Invalid longitude {lon} for place {place_id}") |
| lon = None |
| except ValueError: |
| pass |
| |
| result.update({ |
| "latitude": lat, |
| "longitude": lon, |
| "horizontal_accuracy": float(loc["horizontal_accuracy"]) if loc.get("horizontal_accuracy") else None, |
| "altitude": float(loc["altitude"]) if loc.get("altitude") else None, |
| }) |
| else: |
| result.update({ |
| "latitude": None, |
| "longitude": None, |
| "horizontal_accuracy": None, |
| "altitude": None, |
| }) |
| |
| |
| if place.get("primary_tag_id") and place["primary_tag_id"] in tags: |
| tag = tags[place["primary_tag_id"]] |
| result.update({ |
| "primary_tag_name": tag.get("name"), |
| "primary_tag_slug": tag.get("slug"), |
| "primary_tag_type": tag.get("tag_type"), |
| }) |
| else: |
| result.update({ |
| "primary_tag_name": None, |
| "primary_tag_slug": None, |
| "primary_tag_type": None, |
| }) |
| |
| |
| if place.get("place_contact_id") and place["place_contact_id"] in contacts: |
| contact = contacts[place["place_contact_id"]] |
| result.update({ |
| "website": contact.get("website"), |
| "instagram": contact.get("instagram"), |
| "twitter": contact.get("twitter"), |
| }) |
| else: |
| result.update({ |
| "website": None, |
| "instagram": None, |
| "twitter": None, |
| }) |
| |
| yield idx, result |
| idx += 1 |
| |
| elif self.config.name != "all": |
| with open(filepath, encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| for idx, row in enumerate(reader): |
| |
| for key, value in row.items(): |
| if value in ["true", "false"]: |
| row[key] = value == "true" |
| elif value == "": |
| row[key] = None |
| |
| elif key in ["radius_in_meters", "z_priority"]: |
| try: |
| row[key] = int(value) if value else None |
| except ValueError: |
| row[key] = None |
| elif key in ["latitude", "longitude", "horizontal_accuracy", "altitude", "vertical_accuracy"]: |
| try: |
| if value: |
| val = float(value) |
| |
| if key == "latitude" and not (-90 <= val <= 90): |
| print(f"Warning: Invalid latitude {val} in row {idx}") |
| row[key] = None |
| elif key == "longitude" and not (-180 <= val <= 180): |
| print(f"Warning: Invalid longitude {val} in row {idx}") |
| row[key] = None |
| else: |
| row[key] = val |
| else: |
| row[key] = None |
| except ValueError: |
| row[key] = None |
| |
| yield idx, row |
| else: |
| |
| idx = 0 |
| for table_name, filepath in filepaths.items(): |
| with open(filepath, encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| rows = list(reader) |
| yield idx, { |
| "table_name": table_name, |
| "data": json.dumps(rows) |
| } |
| idx += 1 |
|
|
|
|
| |
| def load_places_as_dict(data_dir: str) -> Dict[str, List[Dict[str, Any]]]: |
| """ |
| Load all CSV files from the directory into a dictionary of tables. |
| |
| Args: |
| data_dir: Directory containing the CSV files |
| |
| Returns: |
| Dictionary where keys are table names and values are lists of row dictionaries |
| """ |
| tables = {} |
| |
| for table_name, filename in _URLS.items(): |
| filepath = os.path.join(data_dir, filename) |
| if os.path.exists(filepath): |
| with open(filepath, encoding="utf-8") as f: |
| reader = csv.DictReader(f) |
| tables[table_name] = list(reader) |
| |
| return tables |
|
|