| from __future__ import absolute_import, division, print_function |
|
|
| import json |
| import os |
| import sys |
|
|
| import datasets |
| from pyarrow import csv |
|
|
| _DESCRIPTION = """Papers with aspects from paperswithcode.com dataset""" |
|
|
| _HOMEPAGE = "https://github.com/malteos/aspect-document-embeddings" |
|
|
| _CITATION = '''@InProceedings{Ostendorff2022, |
| title = {Specialized Document Embeddings for Aspect-based Similarity of Research Papers}, |
| booktitle = {Proceedings of the {ACM}/{IEEE} {Joint} {Conference} on {Digital} {Libraries} ({JCDL})}, |
| author = {Ostendorff, Malte and Blume, Till, Ruas, Terry and Gipp, Bela and Rehm, Georg}, |
| year = {2022}, |
| }''' |
|
|
| DATA_URL = "http://datasets.fiq.de/paperswithcode_aspects.tar.gz" |
|
|
| DOC_A_COL = "from_paper_id" |
| DOC_B_COL = "to_paper_id" |
| LABEL_COL = "label" |
|
|
| |
| LABEL_CLASSES = labels = ['y', 'n'] |
|
|
| ASPECTS = ['task', 'method', 'dataset'] |
|
|
|
|
| def get_train_split(aspect, k): |
| return datasets.Split(f'fold_{aspect}_{k}_train') |
|
|
|
|
| def get_test_split(aspect, k): |
| return datasets.Split(f'fold_{aspect}_{k}_test') |
|
|
|
|
| class PWCConfig(datasets.BuilderConfig): |
| def __init__(self, features, data_url, aspects, **kwargs): |
| super().__init__(version=datasets.Version("0.1.0"), **kwargs) |
| self.features = features |
| self.data_url = data_url |
| self.aspects = aspects |
|
|
|
|
| class PWCAspects(datasets.GeneratorBasedBuilder): |
| """Paper aspects dataset.""" |
|
|
| BUILDER_CONFIGS = [ |
| PWCConfig( |
| name="docs", |
| description="document text and meta data", |
| |
| |
| features={ |
| "paper_id": datasets.Value("string"), |
| "paper_url": datasets.Value("string"), |
| "title": datasets.Value("string"), |
| "abstract": datasets.Value("string"), |
| "arxiv_id": datasets.Value("string"), |
| "url_abs": datasets.Value("string"), |
| "url_pdf": datasets.Value("string"), |
| "aspect_tasks": datasets.Sequence(datasets.Value('string', id='task')), |
| "aspect_methods": datasets.Sequence(datasets.Value('string', id='method')), |
| "aspect_datasets": datasets.Sequence(datasets.Value('string', id='dataset')), |
| }, |
| data_url=DATA_URL, |
| aspects=ASPECTS, |
| ), |
| PWCConfig( |
| name="relations", |
| description=" relation data", |
| features={ |
| DOC_A_COL: datasets.Value("string"), |
| DOC_B_COL: datasets.Value("string"), |
| LABEL_COL: datasets.Value("string"), |
| }, |
| data_url=DATA_URL, |
| aspects=ASPECTS, |
| ), |
| ] |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION + self.config.description, |
| features=datasets.Features(self.config.features), |
| homepage=_HOMEPAGE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| arch_path = dl_manager.download_and_extract(self.config.data_url) |
|
|
| if "relations" in self.config.name: |
| train_file = "train.csv" |
| test_file = "test.csv" |
|
|
| generators = [] |
|
|
| |
| for aspect in self.config.aspects: |
| for k in ["sample"] + [1, 2, 3, 4]: |
| folds_path = os.path.join(arch_path, 'folds', aspect, str(k)) |
| generators += [ |
| datasets.SplitGenerator( |
| name=get_train_split(aspect, k), |
| gen_kwargs={'filepath': os.path.join(folds_path, train_file)} |
| ), |
| datasets.SplitGenerator( |
| name=get_test_split(aspect, k), |
| gen_kwargs={'filepath': os.path.join(folds_path, test_file)} |
| ) |
| ] |
| return generators |
|
|
| elif "docs" in self.config.name: |
| |
| docs_file = os.path.join(arch_path, "docs.jsonl") |
|
|
| return [ |
| datasets.SplitGenerator(name=datasets.Split('docs'), gen_kwargs={"filepath": docs_file}), |
| ] |
| else: |
| raise ValueError() |
|
|
| @staticmethod |
| def get_dict_value(d, key, default=None): |
| if key in d: |
| return d[key] |
| else: |
| return default |
|
|
| def _generate_examples(self, filepath): |
| """Generate docs + rel examples.""" |
|
|
| if "relations" in self.config.name: |
| df = csv.read_csv(filepath).to_pandas() |
|
|
| for idx, row in df.iterrows(): |
| yield idx, { |
| DOC_A_COL: str(row[DOC_A_COL]), |
| DOC_B_COL: str(row[DOC_B_COL]), |
| LABEL_COL: row['label'], |
| } |
|
|
| elif self.config.name == "docs": |
| with open(filepath, 'r') as f: |
| for i, line in enumerate(f): |
| doc = json.loads(line) |
| |
| features = {k: doc[k] if k in doc else None for k in self.config.features.keys()} |
|
|
| yield i, features |
|
|