|
""" |
|
The Harvard USPTO Patent Dataset (HUPD) is a large-scale, well-structured, and multi-purpose corpus |
|
of English-language patent applications filed to the United States Patent and Trademark Office (USPTO) |
|
between 2004 and 2018. With more than 4.5 million patent documents, HUPD is two to three times larger |
|
than comparable corpora. Unlike other NLP patent datasets, HUPD contains the inventor-submitted versions |
|
of patent applications, not the final versions of granted patents, allowing us to study patentability at |
|
the time of filing using NLP methods for the first time. |
|
""" |
|
|
|
from __future__ import absolute_import, division, print_function |
|
|
|
import os |
|
import datetime |
|
import pandas as pd |
|
import numpy as np |
|
from pathlib import Path |
|
try: |
|
import ujson as json |
|
except: |
|
import json |
|
|
|
import datasets |
|
|
|
|
|
_CITATION = """\ |
|
@InProceedings{suzgun2021:hupd, |
|
title = {The Harvard USPTO Patent Dataset}, |
|
authors={Mirac Suzgun and Suproteem Sarkar and Luke Melas-Kyriazi and Scott Kominers and Stuart Shieber}, |
|
year={2021} |
|
} |
|
""" |
|
|
|
_DESCRIPTION = """ |
|
The Harvard USPTO Patent Dataset (HUPD) is a large-scale, well-structured, and multi-purpose corpus |
|
of English-language patent applications filed to the United States Patent and Trademark Office (USPTO) |
|
between 2004 and 2018. With more than 4.5 million patent documents, HUPD is two to three times larger |
|
than comparable corpora. Unlike other NLP patent datasets, HUPD contains the inventor-submitted versions |
|
of patent applications, not the final versions of granted patents, allowing us to study patentability at |
|
the time of filing using NLP methods for the first time. |
|
""" |
|
|
|
RANDOM_STATE = 1729 |
|
|
|
_FEATURES = [ |
|
"patent_number", |
|
"decision", |
|
"title", |
|
"abstract", |
|
"claims", |
|
"background", |
|
"summary", |
|
"description", |
|
"cpc_label", |
|
"ipc_label", |
|
"filing_date", |
|
"patent_issue_date", |
|
"date_published", |
|
"examiner_id" |
|
] |
|
|
|
|
|
def str_to_date(s): |
|
"""A helper function to convert strings to dates""" |
|
return datetime.datetime.strptime(s, '%Y-%m-%d') |
|
|
|
|
|
class PatentsConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for Patents""" |
|
|
|
def __init__( |
|
self, |
|
metadata_url: str, |
|
data_url: str, |
|
data_dir: str, |
|
ipcr_label: str = None, |
|
cpc_label: str = None, |
|
train_filing_start_date: str = None, |
|
train_filing_end_date: str = None, |
|
val_filing_start_date: str = None, |
|
val_filing_end_date: str = None, |
|
query_string: str = None, |
|
val_set_balancer=False, |
|
uniform_split=False, |
|
**kwargs |
|
): |
|
""" |
|
If train_filing_end_date is None, then a random train-val split will be used. If it is |
|
specified, then the specified date range will be used for the split. If train_filing_end_date |
|
if specified and val_filing_start_date is not specifed, then val_filing_start_date defaults to |
|
train_filing_end_date. |
|
|
|
Args: |
|
metadata_url: `string`, url from which to download the metadata file |
|
data_url: `string`, url from which to download the json files |
|
data_dir: `string`, folder (in cache) in which downloaded json files are stored |
|
ipcr_label: International Patent Classification code |
|
cpc_label: Cooperative Patent Classification code |
|
train_filing_start_date: Start date for patents in train set (and val set if random split is used) |
|
train_filing_end_date: End date for patents in train set |
|
val_filing_start_date: Start date for patents in val set |
|
val_filing_end_date: End date for patents in val set (and train set if random split is used) |
|
**kwargs: keyword arguments forwarded to super |
|
""" |
|
super().__init__(**kwargs) |
|
self.metadata_url = metadata_url |
|
self.data_url = data_url |
|
self.data_dir = data_dir |
|
self.ipcr_label = ipcr_label |
|
self.cpc_label = cpc_label |
|
self.train_filing_start_date = train_filing_start_date |
|
self.train_filing_end_date = train_filing_end_date |
|
self.val_filing_start_date = val_filing_start_date |
|
self.val_filing_end_date = val_filing_end_date |
|
self.query_string = query_string |
|
self.val_set_balancer = val_set_balancer |
|
self.uniform_split = uniform_split |
|
|
|
|
|
class Patents(datasets.GeneratorBasedBuilder): |
|
_DESCRIPTION |
|
|
|
VERSION = datasets.Version("1.0.2") |
|
|
|
|
|
|
|
|
|
BUILDER_CONFIG_CLASS = PatentsConfig |
|
BUILDER_CONFIGS = [ |
|
PatentsConfig( |
|
name="sample", |
|
description="Patent data from January 2016, for debugging", |
|
metadata_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/hupd_metadata_jan16_2022-02-22.feather", |
|
data_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/data/sample-jan-2016.tar.gz", |
|
data_dir="data", |
|
), |
|
PatentsConfig( |
|
name="all", |
|
description="Patent data from all years (2004-2018)", |
|
metadata_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/hupd_metadata_2022-02-22.feather", |
|
data_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/data/all-years.tar.gz", |
|
data_dir="data", |
|
), |
|
] |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=datasets.Features( |
|
{k: datasets.Value("string") for k in _FEATURES} |
|
), |
|
|
|
|
|
|
|
supervised_keys=("claims", "decision"), |
|
homepage="https://github.com/suzgunmirac/hupd", |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager: datasets.DownloadManager): |
|
"""Returns SplitGenerators.""" |
|
print(f'Loading dataset with config: {self.config}') |
|
|
|
|
|
|
|
metadata_url = self.config.metadata_url |
|
metadata_file = dl_manager.download_and_extract(self.config.metadata_url) |
|
print(f'Using metadata file: {metadata_file}') |
|
|
|
|
|
|
|
|
|
download_dir = dl_manager.download_and_extract(self.config.data_url) |
|
json_dir = os.path.join(download_dir, self.config.data_dir) |
|
|
|
|
|
print(f'Reading metadata file: {metadata_file}') |
|
if metadata_url.endswith('.feather'): |
|
df = pd.read_feather(metadata_file) |
|
elif metadata_url.endswith('.csv'): |
|
df = pd.read_csv(metadata_file) |
|
elif metadata_url.endswith('.tsv'): |
|
df = pd.read_csv(metadata_file, delimiter='\t') |
|
elif metadata_url.endswith('.pickle'): |
|
df = pd.read_pickle(metadata_file) |
|
else: |
|
raise ValueError(f'Metadata file invalid: {metadata_url}') |
|
|
|
|
|
if self.config.ipcr_label: |
|
print(f'Filtering by IPCR label: {self.config.ipcr_label}') |
|
df = df[df['main_ipcr_label'].str.startswith(self.config.ipcr_label)] |
|
elif self.config.cpc_label: |
|
print(f'Filtering by CPC label: {self.config.cpc_label}') |
|
df = df[df['main_cpc_label'].str.startswith(self.config.cpc_label)] |
|
|
|
|
|
if self.config.query_string: |
|
df = df.query(self.config.query_string) |
|
|
|
|
|
if self.config.uniform_split: |
|
|
|
|
|
if self.config.train_filing_start_date: |
|
df = df[df['filing_date'] >= self.config.train_filing_start_date] |
|
if self.config.val_filing_end_date: |
|
df = df[df['filing_date'] <= self.config.val_filing_end_date] |
|
df = df.sample(frac=1.0, random_state=RANDOM_STATE) |
|
num_train_samples = int(len(df) * 0.85) |
|
train_df = df.iloc[0:num_train_samples] |
|
val_df = df.iloc[num_train_samples:-1] |
|
|
|
else: |
|
|
|
|
|
if not (self.config.train_filing_start_date and self.config.train_filing_end_date and |
|
self.config.val_filing_start_date and self.config.train_filing_end_date): |
|
raise ValueError("Please either use uniform_split or specify your exact \ |
|
training and validation split dates.") |
|
|
|
|
|
print(f'Filtering train dataset by filing start date: {self.config.train_filing_start_date}') |
|
print(f'Filtering train dataset by filing end date: {self.config.train_filing_end_date}') |
|
print(f'Filtering val dataset by filing start date: {self.config.val_filing_start_date}') |
|
print(f'Filtering val dataset by filing end date: {self.config.val_filing_end_date}') |
|
train_df = df[ |
|
(df['filing_date'] >= self.config.train_filing_start_date) & |
|
(df['filing_date'] < self.config.train_filing_end_date) |
|
] |
|
val_df = df[ |
|
(df['filing_date'] >= self.config.val_filing_start_date) & |
|
(df['filing_date'] < self.config.val_filing_end_date) |
|
] |
|
|
|
|
|
if self.config.val_set_balancer: |
|
rejected_df = val_df[val_df.status == 'REJECTED'] |
|
num_rejected = len(rejected_df) |
|
accepted_df = val_df[val_df.status == 'ACCEPTED'] |
|
num_accepted = len(accepted_df) |
|
if num_rejected < num_accepted: |
|
accepted_df = accepted_df.sample(frac=1.0, random_state=RANDOM_STATE) |
|
accepted_df = accepted_df[:num_rejected] |
|
else: |
|
rejected_df = rejected_df.sample(frac=1.0, random_state=RANDOM_STATE) |
|
rejected_df = rejected_df[:num_accepted] |
|
val_df = pd.concat([rejected_df, accepted_df]) |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs=dict( |
|
df=train_df, |
|
json_dir=json_dir, |
|
split='train', |
|
), |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
gen_kwargs=dict( |
|
df=val_df, |
|
json_dir=json_dir, |
|
split='val', |
|
), |
|
), |
|
] |
|
|
|
def _generate_examples(self, df, json_dir, split): |
|
""" Yields examples by loading JSON files containing patent applications. """ |
|
|
|
|
|
for id_, x in enumerate(df.itertuples()): |
|
|
|
|
|
application_year = str(x.filing_date.year) |
|
application_number = x.application_number |
|
filepath = os.path.join(json_dir, application_year, application_number + '.json') |
|
try: |
|
with open(filepath, 'r') as f: |
|
patent = json.load(f) |
|
except Exception as e: |
|
print('------------') |
|
print(f'ERROR WITH {filepath}\n') |
|
print(repr(e)) |
|
print() |
|
yield id_, {k: "error" for k in _FEATURES} |
|
|
|
|
|
decision = x.decision |
|
yield id_, { |
|
"patent_number": application_number, |
|
"decision": decision, |
|
"title": patent["title"], |
|
"abstract": patent["abstract"], |
|
"claims": patent["claims"], |
|
"description": patent["full_description"], |
|
"background": patent["background"], |
|
"summary": patent["summary"], |
|
"cpc_label": patent["main_cpc_label"], |
|
'filing_date': patent['filing_date'], |
|
'patent_issue_date': patent['patent_issue_date'], |
|
'date_published': patent['date_published'], |
|
'examiner_id': patent['examiner_id'], |
|
"ipc_label": patent["main_ipcr_label"], |
|
} |
|
|