PatentsView / PatentsView-Evaluation

Evaluation and benchmarking of PatentsView disambiguation algorithms

Home Page:https://patentsview.github.io/PatentsView-Evaluation/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Python package pages-build-deployment

📊 PatentsView-Evaluation: Benchmark Disambiguation Algorithms

pv_evaluation is a Python package built to help advance research on author/inventor name disambiguation systems such as PatentsView. It provides:

  1. A large set of benchmark datasets for U.S. patents inventor name disambiguation.
  2. Disambiguation summary statistics, evaluation methodology, and performance estimators through the ER-Evaluation Python package.

See the project website for full documentation. The Examples page provides real-world examples of the use of pv_evaluation submodules.

Submodules

pv_evaluation has the following submodules:

  • benchmark.data: Access to evaluation datasets and standardized comparison benchmarks. The following benchmark datasets are available:
    • Academic Life Sciences (ALS) inventors benchmark.
    • Israeli inventors benchmark.
    • Engineering and Sciences (ENS) inventors benchmark.
    • Lai's 2011 inventors benchmark.
    • PatentsView's 2021 inventors benchmark.
    • Binette et al.'s 2022 inventors benchmark.
  • benchmark.report: Visualization of key monitoring and performance metrics.
  • templates: Templated performance summary reports.

Installation

Install the released version of pv_evaluation using

pip install pv-evaluation

Rendering reports requires the installation of quarto from quarto.org.

Examples

Note: Working with the full patent data requires large amounts of memory (we suggest having 64GB RAM available).

See the examples page for complete reproducible examples. The examples below only provide a quick overview of pv_evaluation's functionality.

Metrics and Summary Statistics

Generate an html report summarizing properties of the current disambiguation algorithm (see this example):

from pv_evaluation.templates import render_inventor_disambiguation_report

render_inventor_disambiguation_report(
    ".", 
    disambiguation_files=["disambiguation_20211230.tsv", "disambiguation_20220630.tsv"],
    inventor_not_disambiguated_file="g_inventor_not_disambiguated.tsv"
)

Benchmark Datasets

Access PatentsView-Evaluation's large collection of benchmark datasets:

from pv_evaluation.benchmark import *

load_lai_2011_inventors_benchmark()
load_israeli_inventors_benchmark()
load_patentsview_inventors_benchmark()
load_als_inventors_benchmark()
load_ens_inventors_benchmark()
load_binette_2022_inventors_benchmark()
load_air_umass_assignees_benchmark()
load_nber_subset_assignees_benchmark()

Representative Performance Evaluation

See this example of how representative performance estimates are obtained from Binette's 2022 benchmark dataset.

Citation

Contributing

Contribute code and documentation

Look through the GitHub issues for bugs and feature requests. To contribute to this package:

  1. Fork this repository
  2. Make your changes and update CHANGELOG.md
  3. Submit a pull request
  4. For maintainers: if needed, update the "release" branch and create a release.

A conda environment is provided for development convenience. To create or update this environment, make sure you have conda installed and then run make env. You can then activate the development environment using conda activate pv-evaluation.

The makefile provides other development utilities such as make black to format Python files, make data to re-generate benchmark datasets from raw data located on AWS S3, and make docs to generate the documentation website.

Raw data

Raw public data is located on PatentsView's AWS S3 server at https://s3.amazonaws.com/data.patentsview.org/PatentsView-Evaluation/data-raw.zip. This zip file should be updated as needed to reflect datasets provided by this package and to ensure that original data sources are preserved without modification.

Testing

The minimal testing requirement for this package is a check that all code executes without error. We recommend placing execution checks in a runnable notebook and using the testbook package for execution within unit tests. User examples should also be provided to exemplify usage on real data.

Report bugs and submit feedback

Report bugs and submit feedback at https://github.com/PatentsView/PatentsView-Evaluation/issues.

About

Evaluation and benchmarking of PatentsView disambiguation algorithms

https://patentsview.github.io/PatentsView-Evaluation/

License:GNU General Public License v3.0


Languages

Language:Python 51.1%Language:HTML 40.6%Language:Jupyter Notebook 6.6%Language:Makefile 1.8%