I'm a bit confused with the use of Random Forest in Sklearn in case we have categorical variables. I've read this article stating that one hot encoding affects performance negatively when using Decision Tree-based methods.
https://roamanalytics.com/2016/10/28/are-categorical-variables-getting-lost-in-your-random-forests/
I'm facing this exact problem with a dataset. I have a high number of features with continuous values and three other categorical features that I know are highly correlated with the target variable. However when I use mutual_info_classif and RandomForest these categorical variables are considered not important. I have tried keeping them as categorical variables but as far as I know, sklearn implementation can't handle categorical variables. I've also one-hot encoded them but then I'm facing the problem explained in the link.
What are the options left?. Should I just change the library used for Random Forest?. Maybe change to another model that can be used more efffectively with one hot encoding like SVM?