Here is a Cornell CS document that compares the performance of different classifiers. It does not reach speed, but it exceeds the predictive ability of almost all classification algorithms that are widely used today. The fastest of these will be algorithms that are not ensemble classifiers of learning. Any algorithm that builds several models and averages the results will essentially take longer. However, as can be seen in Table 2 on page 5, ensemble methods are the most effective classifiers. If you want to build the model as quickly as possible, then you should probably use only one decision tree or logistic regression. Otherwise, you need to spend some time learning the ensemble's learning techniques and figure out how to optimize the speed of this particular algorithm. I got good results by parallelizing my random forests using a technique similar to this .
Edit to more accurately solve your memory problems: Memory usage is less than what you choose than how you use this algorithm. Assuming you used the default random left call for your original model, you would build 500 decision trees, each of which has ~ 450 predictor variables and as many terminal nodes as you have data points in the sample. It takes a whole bunch of memory. What I'm trying to do is that you can customize any of these classification models to reduce the amount of memory and work more efficiently in R. As mentioned earlier, methods without ensemble (logistic regression, naive bays, CHAID / CART / etc) will use at least the default memory.
source share