Publication of the paper “Large Language Models for Named Entity Extraction and Spelling Correction”
As the field of AI/ML makes revolutionary strides, Best Path Research has just published a paper on arxiv which adds another contribution to the field:
While a large part of the advancements is in the user experience, the perceived “intelligence” of Large Language models (LLMs) is often due to simply using more training data and training larger models. However, we like to emphasize that just making things “bigger” is not the only way to make a model better. To advance the field and make systems which are more usable and interpretable, we need a combination of better algorithms, data and modularity. This is where Best Path Research has deep knowledge and experience.
In our latest work, Best Path Research compared the performance of eight, open-source, “generative” LLMs with two existing state-of-the-art BERT language models on the task of extracting Named Entities (NEs), such as shop names, addresses and product names, from Japanese shop receipts.
Based on decades of experience with speech and handwriting recognition, machine learning, and natural language processing, our method combines known and proven language modelling algorithms with LLMs in an effective way. One such integrated approach is Question Answering using Language Models, that our CEO first developed and wrote about while he was a post-doc at Philips and Tokyo Institute of Technology in the early 2000s. At that time the method was used to answer factoid questions in the Text REtrieval Conference (TREC) and Question Answering Competitions.
In our proposed approach, text is first extracted from scanned images of paper shop receipts using Best Path Research’s proprietary and productized Optical Character Recognition (OCR) system (<1% character error rate). The text is then fed to a fine-tuned language model (either an LLM or BERT, in our work), so as to answer a question about a desired NE category that we want to extract from the text, such as the shop name, address or a product name in the original receipt.
We show that the best LLMs perform as well as, or better than, the existing state-of-the-art methods, achieving 100% precision and recall on some numerical NE categories, while also demonstrating additional benefits, such as being able to correct some of the original OCR recognition errors.
While our research demonstrates the effectiveness of our approach using LLMs on OCR’d text from Japanese shop receipts, the method is equally applicable to any language and any scenario that requires the extraction of Named Entities from text documents, e.g., counterparty names, addresses and payment conditions from legal contracts, revenue changes and market conditions from financial reports, and many other scenarios.
Get in touch with us to discuss where we can assist you in your digitization and document extraction requirements.