On-the-job training for the AI software model
Piprek’s work began with an ordinary off-the-shelf AI software program. With a few tweaks, he and his team adapted it to the characteristics of the technical documents. “A good way to think of the package is like the brain of a toddler: huge potential, but still at a low level.”
Children are encouraged to develop this potential in school, but AI software needs to be trained. In Piprek’s project, the software applies deep learning techniques and is fed with countless texts from MTU’s PLM system. From a software perspective, it’s kind of like on-the-job training. “We know precisely what knowledge is available in the PLM system and how it is structured. That provides us with an excellent basis for training and evaluating the AI.” Such background knowledge is necessary because the AI relies on verified training data as it runs through the decision scenarios.
“In simple terms, what happens is that the AI module first builds itself a statistical model,” Piprek explains. “The next step is for us to give it rules for evaluating the content in a specific way.” Equipped with this combination of statistics and rules, the software gradually figures out a way to make sense of the documents. Over and over again, it retrieves familiar patterns and applies them to new scenarios requiring a decision. By matching the calculated result with the desired target result, the AI model learns. With each training run, it gets a little smarter, a little faster, a little more accurate. The more data available and the higher its quality, the better the training. The technical term for what emerges is a neural network.
In this network, the software searches for terms that it recognizes from a similar context or that it identifies as synonyms. At some point, it will even be able to make sense of the content of documents that have little structure—and it will no longer be confused by barely legible letters on a scan of yellowed paper.