@ Code for Paper: categorical boosting based tabnet
Recent deep learning models perform well in image and natural language processing. However, in tabular data, there is a problem that good performance is not achieved due to data-level problems. Recently, TabNet, a model that overcomes these shortcomings, has been widely used for tabular data learning. However, categorical variable data does not perform significantly in tabular data. To solve this problem, Catboost Encoding method is used to solve the problem. In the case of this model, the pre-processing of categorical variable data was well utilized to derive more performance than other models, and it showed better performance than other encoding techniques.
Results |
---|
Categorical boosting demonstrates superior performance compared to TabNet and exhibits excellent results across diverse datasets. |
Mask |
---|
In TabNet, if attention is distributed equally among features, in this approach, attention to different features varies across layers. |