In the past few years, object detection has attracted a lot of attention in the context of human–robot collaboration and Industry 5.0 due to enormous quality improvements in deep learning technologies. In many applications, object detection models have to be able to quickly adapt to a changing environment, i.e., to learn new objects. A crucial but challenging prerequisite for this is the automatic generation of new training data which currently still limits the broad application of object detection methods in industrial manufacturing. In this work, we discuss how to adapt state-of-the-art object detection methods for the task of automatic bounding box annotation in a use case where the background is homogeneous and the object’s label is provided by a human. We compare an adapted version of Faster R-CNN and the Scaled-YOLOv4-p5 architecture and show that both can be trained to distinguish unknown objects from a complex but homogeneous background using only a small amount of training data. In contrast to most other state-of-the-art methods for bounding box labeling, our proposed method neither requires human verification, a predefined set of classes, nor a very large manually annotated dataset. Our method outperforms the state-of-the-art, transformer-based object discovery method LOST on our simple fruits dataset by large margins.
In the last decade, industry’s demand for deep learning (DL) has increased due to its high performance in complex scenarios. Due to the DL method’s complexity, experts and non-experts rely on blackbox software packages such as Tensorflow and Pytorch. The frameworks are constantly improving, and new versions are released frequently. As a natural process in software development, the released versions contain improvements/changes in the methods and their implementation. Moreover, versions may be bug-polluted, leading to the model performance decreasing or stopping the model from working. The aforementioned changes in implementation can lead to variance in obtained results. This work investigates the effect of implementation changes in different major releases of these frameworks on the model performance. We perform our study using a variety of standard datasets. Our study shows that users should consider that changing the framework version can affect the model performance. Moreover, they should consider the possibility of a bug-polluted version before starting to debug source code that had an excellent performance before a version change. This also shows the importance of using virtual environments, such as Docker, when delivering a software product to clients.
In enterprises, data is usually distributed across multiple data sources and stored in heterogeneous formats. The harmonization and integration of data is a prerequisite to leverage it for AI initiatives. Recently, data catalogs pose a promising solution to...
T. Hütter, N. Augsten, C. Kirsch, M. Carey, and C. Li. Proceedings of the 2022 International Conference on Management of Data, page 1584–1597. New York, NY, USA, Association for Computing Machinery, (Jun 11, 2022)
R. Ramler, M. Moser, L. Fischer, M. Nissl, and R. Heinzl. Proceedings of the 1st International Workshop on Large Language Models for Code, page 1–7. New York, NY, USA, Association for Computing Machinery, (Sep 10, 2024)
M. Bechny, F. Sobieczky, J. Zeindl, and L. Ehrlinger. Proceedings of the 33rd International Conference on Scientific and Statistical Database Management, page 214–219. New York, NY, USA, Association for Computing Machinery, (Aug 11, 2021)
A. Hackl, J. Zeindl, and L. Ehrlinger. Proceedings of the 35th International Conference on Scientific and Statistical Database Management, page 1–2. New York, NY, USA, Association for Computing Machinery, (Aug 27, 2023)