We develop custom solutions based on large language models (LLMs), actively participate in related research and regularly publish technical content about various aspects of LLMs.
From OCR to LLMs: The journey to reliable data extraction from complex retail documents
Axel Besinger and
Augusto Stoffel (PhD)
May 23rd, 2025
AI-powered data extraction works - until it doesn’t. When handling structured tables in invoices, orders, or financial documents, we expect OCR, LLMs, and Vision AI to extract data reliably. However, complex documents - e.g. nested tables, irregular structures, and edge cases - pose real challenges for document data extraction AI models. With our solution smartextract, we tackled a real-world customer challenge: automating order entry from complex order documents and tables for a German shoe retailer: OCR and text-based LLMs struggled, Vision LLMs were inconsistent. Only extensive customization could solve the appearing problems - including segmentation, few-shot prompting, fine-tuning, and even the possibility of training a custom computer vision model. In this talk, we will show why standard AI models struggle with complex tables and demonstrate in which cases segmentation helps. Further, we will show benchmarks of commercial vs. open-source models and discuss the trade-offs between OCR, LLMs, and computer vision models.
dida talks
Fabian Dechent
Decision process automation with Large Language Models
Fabian Dechent
May 31st, 2024
Large Language Models impress with their adeptness in context-aware text generation, logic, as well as reasoning. Typically, downstream models fine tuned on chat data possess the remarkable ability to be directed towards solving tasks described in natural language without explicit further weight adaptation. In relevant applications, interesting use cases often relate multiple external data sources with each other and are characterized by a complex multistep decision process. In this talk, we discuss how predefining decision steps and integrating external data filtering can break down multifaceted problems into manageable, self-contained language processing tasks, which can readily be solved by LLMs.
dida talks
Jona Welsch
Information extraction with BERT from free-form text
Jona Welsch
April 28th, 2023
Jona Welsch's talk centers on using Deep Learning methods like BERT to extract information from unstructured text. A project with idealo serves as a case study, showcasing how rule-based algorithms and Deep Learning can be combined to turn product descriptions into structured data. The talk also touches on creating weakly labeled training data to ease the labeling process.
dida talks
Mattes Mollenhauer
Automated answering of questions with neural networks: BERT
Mattes Mollenhauer
May 26th, 2021
In this webinar we will present a method based on the BERT model for automated answering of questions. The potential applications are manifold: the ideas of this approach can be used for example in chatbots, information extraction from texts and Q&A sections of websites. As a concrete example, we discuss the extraction of information from biomedical research using the open CORD-19 data set for COVID-19 research.
dida talks
Konrad Schultka (PhD)
Jona Welsch
Semantic search and understanding of natural text with neural networks: BERT
Konrad Schultka (PhD) and
Jona Welsch
May 26th, 2021
In this webinar you will get an introduction to the application of BERT for Semantic Search using a real case study: Every year millions of citizens interact with public authorities and are regularly overwhelmed by the technical language used there. We have successfully used BERT to deliver the right answer from government documents with the help of colloquial queries - without having to use technical terms in the queries.