Feature extraction layers like Local Binary Patterns (LBP) transform can be very useful for improving the accuracy of machine learning and deep learning models depending on the problem type. Direct implementations of such layers in Python may result in long running times, and training a computer vision model may be delayed significantly. For this purpose, TensorFlow framework enables developing accelerated custom operations based on the existing operations which already have support for accelerated hardware such as multicore CPU and GPU. In this study, LBP transform which is used for feature extraction in various applications, was implemented based on TensorFlow operations. The evaluations were done using both standard Python operations and TensorFlow library for performance comparisons. The experiments were realized using images in various dimensions and various batch sizes. Numerical results show that algorithm based on TensorFlow operations provides good acceleration rates over Python runs. The implementation of LBP can be used for the accelerated computing for various feature extraction purposes including machine learning as well as in deep learning applications.
Primary Language | English |
---|---|
Subjects | Engineering |
Journal Section | Research Article |
Authors | |
Publication Date | June 30, 2021 |
Published in Issue | Year 2021 |
Manas Journal of Engineering