BEST PATH RESEARCH
  • HOME
  • ABOUT US
  • SERVICE
  • COMPANY
  • NEWS
  • BLOG
  • CONTACT
  • English
  • 日本語

  • English
  • 日本語

  • By  Kenji Saito
  • 0 comments
  • 2024-03-09

Publication of the paper “Large Language Models for Named Entity Extraction and Spelling Correction”

As the field of AI/ML makes revolutionary strides, Best Path Research has just published a paper on arxiv which adds another contribution to the field:

https://arxiv.org/abs/2403.00528

While a large part of the advancements is in the user experience, the perceived “intelligence” of Large Language models (LLMs) is often due to simply using more training data and training larger models. However, we like to emphasize that just making things “bigger” is not the only way to make a model better. To advance the field and make systems which are more usable and interpretable, we need a combination of better algorithms, data and modularity. This is where Best Path Research has deep knowledge and experience.

In our latest work, Best Path Research compared the performance of eight, open-source, “generative” LLMs with two existing state-of-the-art BERT language models on the task of extracting Named Entities (NEs), such as shop names, addresses and product names, from Japanese shop receipts.

Based on decades of experience with speech and handwriting recognition, machine learning, and natural language processing, our method combines known and proven language modelling algorithms with LLMs in an effective way. One such integrated approach is Question Answering using Language Models, that our CEO first developed and wrote about while he was a post-doc at Philips and Tokyo Institute of Technology in the early 2000s. At that time the method was used to answer factoid questions in the Text REtrieval Conference (TREC) and Question Answering Competitions.

In our proposed approach, text is first extracted from scanned images of paper shop receipts using Best Path Research’s proprietary and productized Optical Character Recognition (OCR) system (<1% character error rate). The text is then fed to a fine-tuned language model (either an LLM or BERT, in our work), so as to answer a question about a desired NE category that we want to extract from the text, such as the shop name, address or a product name in the original receipt.

We show that the best LLMs perform as well as, or better than, the existing state-of-the-art methods, achieving 100% precision and recall on some numerical NE categories, while also demonstrating additional benefits, such as being able to correct some of the original OCR recognition errors.

While our research demonstrates the effectiveness of our approach using LLMs on OCR’d text from Japanese shop receipts, the method is equally applicable to any language and any scenario that requires the extraction of Named Entities from text documents, e.g., counterparty names, addresses and payment conditions from legal contracts, revenue changes and market conditions from financial reports, and many other scenarios.

Get in touch with us to discuss where we can assist you in your digitization and document extraction requirements.

Tags:
news_

Search

Recent News

  • Publication of the paper “Large Language Models for Named Entity Extraction and Spelling Correction”
  • BEST PATH RESEARCH Exhibits at Intel Connection 2023
  • Navit K.K. publishes an Interview with Best Path Research in “Sohos-Style”
  • Publication of the paper “Automatic Detection and Rectification of Paper Receipts on Smartphones”
  • Publication of the paper “Extending TrOCR for Text Localization-Free OCR of Full-Page Scanned Receipt Images”

Recent Blog

  • Japanese Receipt OCR and Named-entity Extraction: Low-cost Inference with Multiple Models using AWS SageMaker Serverless and Triton Inference Server
  • Privacy-preserving Image Search using Natural Language on iOS

BEST PATH RESEARCH © 2023 All Rights Reserved