-
I'm running apriori on a dataset with 25,000 rows and 24 columns (features). When I try to run this with min_support set to anything less than 0.6, it runs out of memory. I'm watching the swapfile grow up to about 200G and then Mac OS shuts it down. itemsets = apriori(df, min_support=min_support, use_colnames=True) Any suggestions about how to deal with this? Break the problem into smaller parts? Change parameters to allow more swap space, etc? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
I suggest trying |
Beta Was this translation helpful? Give feedback.
I suggest trying
fpgrowth
instead ofapriori
. They are both different approaches that lead to the same solution, butfpgrowth
is usually more memory efficient. It's a simple drop-in replacement, so you can just swap it in without making any other changes.