Transformer-Based Patent Novelty Search by Training Claims to Their Own Description

  • Freunek M
  • Bodmer A
N/ACitations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

In this paper we present a method to concatenate patent claims to their own description. By applying this method, bidirectional encoder representations from transformers (BERT) train suitable descriptions for claims. Such a trained BERT could be able to identify novelty relevant descriptions for patents. In addition, we introduce a new scoring scheme: relevance score or novelty score to interprete the output of BERT. We test the method on patent applications by training BERT on the first claims of patents and corresponding descriptions. The output is processed according to the relevance score and the results compared with the cited X documents in the search reports. The test shows that BERT score some of the cited X documents as highly relevant.

Cite

CITATION STYLE

APA

Freunek, M., & Bodmer, A. (2021). Transformer-Based Patent Novelty Search by Training Claims to Their Own Description. Applied Economics and Finance, 8(5), 37. https://doi.org/10.11114/aef.v8i5.5182

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free