| | ''' |
| | Script on Generating Wikipedia Data that are dumped into https://dumps.wikimedia.org/ |
| | More info can be read on https://huggingface.co/datasets/wikipedia |
| | ------------------- |
| | Check here to see available indexed data: https://dumps.wikimedia.org/backup-index.html |
| | Also check here to see language meta from its code: https://meta.wikimedia.org/wiki/List_of_Wikipedias |
| | ''' |
| |
|
| | import os, gc |
| | import logging |
| | import argparse |
| |
|
| | import pandas as pd |
| | from datasets import load_dataset |
| |
|
| |
|
| | def set_logger(): |
| | |
| | logging.basicConfig( |
| | level=logging.INFO, |
| | format='%(asctime)s [%(levelname)s]: %(message)s', |
| | datefmt='%Y-%m-%d %H:%M:%S' |
| | ) |
| |
|
| | |
| | file_handler = logging.FileHandler('app.log') |
| |
|
| | |
| | file_handler.setLevel(logging.INFO) |
| |
|
| | |
| | file_formatter = logging.Formatter('%(asctime)s [%(levelname)s]: %(message)s', datefmt='%Y-%m-%d %H:%M:%S') |
| | file_handler.setFormatter(file_formatter) |
| |
|
| | logger = logging.getLogger("Wiki Dataset Generation") |
| | logger.addHandler(file_handler) |
| |
|
| | return logger |
| |
|
| |
|
| | |
| | if __name__ == "__main__": |
| | parser = argparse.ArgumentParser() |
| |
|
| | parser.add_argument("--lang-id", help="Lang ID from Wikipedia Data to extract") |
| |
|
| | parser.add_argument("--date-ver", help="Date of Wikipedia Data (YYYYMMDD) generation to extract") |
| |
|
| | parser.add_argument("--save-dir-path", help="""Relative dir path of saved Wikipedia CSV data |
| | to the `extract_raw_wiki_data.py` script dir""", |
| | default=os.path.dirname(os.path.abspath(__file__))) |
| |
|
| | args = parser.parse_args() |
| |
|
| |
|
| | dset_name = "wikipedia" |
| |
|
| | logger = set_logger() |
| | logger.info("Parsing arguments...") |
| |
|
| | lang_id = args.lang_id |
| | date_ver = args.date_ver |
| | save_dir = args.save_dir_path |
| |
|
| | logger.info("Loading the dataset from Wikipedia...") |
| | df = load_dataset(dset_name, language=lang_id, date=date_ver, beam_runner='DirectRunner', split="train").to_pandas() |
| | logger.info("Loading done!") |
| | logger.info(f"#Data collected: {df.shape[0]}") |
| | logger.info("Saving dataset raw form...") |
| | df.to_csv(f"{save_dir}/wiki_{lang_id}_{date_ver}_raw_dataset.csv", index=False) |
| |
|
| | del df |
| | gc.collect() |
| |
|