Ramamurthy, RajkumarRajkumarRamamurthyPielka, MarenMarenPielkaStenzel, Marc RobinMarc RobinStenzelBauckhage, ChristianChristianBauckhageSifa, RafetRafetSifaKhameneh, Tim DilmaghaniTim DilmaghaniKhamenehWarning, UlrichUlrichWarningKliem, BerndBerndKliemLoitz, RĂ¼digerRĂ¼digerLoitz2022-11-252022-11-252021-08-16https://publica.fraunhofer.de/handle/publica/41290110.1145/3469096.3474928We consider Automated List Inspection (ALI), a content-based text recommendation system that assists auditors in matching relevant text passages from notes in financial statements to specific law regulations. ALI follows a ranking paradigm in which a fixed number of requirements per textual passage are shown to the user. Despite achieving impressive ranking performance, the user experience can still be improved by showing a dynamic number of recommendations. Besides, existing models rely on a feature-based language model that needs to be pre-trained on a large corpus of domain-specific datasets. Moreover, they cannot be trained in an end-to-end fashion by jointly optimizing with language model parameters. In this work, we alleviate these concerns by considering a multi-label classification approach that predicts dynamic requirement sequences. We base our model on pre-trained BERT that allows us to fine-tune the whole model in an end-to-end fashion, thereby avoiding the need for training a language representation model. We conclude by presenting a detailed evaluation of the proposed model on two German financial datasets.enNeural NetworksText ClassificationNatural Language Processing005006629ALiBERT: Improved automated list inspection (ALI) with BERTconference paper