-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
robust irmi, bug with bcancer data #70
Comments
I would not remove rlm, because this fallback is often used when lmrob does not work. The problem is also not with rml, it is with rml and lmrob and any method, since the algorithm decided that the variable to impute is numeric for any reason, but it is not.
So, either the print output is wrong, or the algorithm tooks the wrong regression method for a categorical variable. I can't go into details before Christmas, but can check it later in more detail. What I propose is to not exclude rlm. |
P.S. and it happens in iteration 3, the first 2 iterations was fine...
|
The problem is happening in rlm, probably in the init part where a sample of the data is fitted with LS regression.
Btw, we can support all the methods provided by rlm with lmrob and I would not think that rlm is more stable than lmrob. |
the init is actual happening in lqs which is used by rlm as init method. |
Or simply use the following as the function argument of This way, rlm is still inside as a fallback and That would actually be the fastest solution, wouldn't it? I have committed it this way. |
yes, but the parameter robMethod is quite confusing now as it is. We could improve it by removing rlm and then the parameter is actual stating only the method to be used for the robust estimation and not the function. Because robMethod = "lmrob" is actually doing a MM estimation. |
If I remeber correclty, the aim was always to use lmrob at first glance and as a fallback rlm, because rlm (at least) was more robust in terms of its implementation than lmrob. So there was (at least in the past) a lot of situation, where lmrob does not give a solution, but rlm did. It seems that once we even then changed rlm to default. To not risk more failures for other situations/data, I recommend just to update the documentation instead of kicking out rlm. We can write that we use - when a fallback to rlm is used - also MM regression but then in function rlm. I still think it is a good fallback (when setting |
yeah, ok. Let's do this. |
Ok. Thanks for your efforts! I will update the documentation to be more precise on this. |
This is an unpleasant bug, because it is very hard to debug. It does not happen all the time.
What will be: since so few different values in individual variables, it no longer interprets these variables as numeric.
If you add very little noise, everything fits with no errors:
The text was updated successfully, but these errors were encountered: