Alexandre,

You are mining for association rules with an absolute support count of 1001*0.01 = 10.1. From your 40 minutes processing time I guess that your survey data set is very dense and that results in an enormous number of frequent itemsets (potentially up to 2^k - k - 1 which is in your case for k=71 about 10^21) which causes you to run out of memory. You have the following options:

* increase minimum support (e.g., start with 0.1) and see how low you can go without using up all your memory (I don't know how to watch memory usage on Windows).

* restrict the maximal length of frequent itemsets by using parameter=list(maxlen = 3, support = ...)

If that does not help I will need your dataset and some code to reproduce and study the problem.

-Michael


Hi I'm a windows XP user. My notebook have 1gb ram, 160gb hd, processor amd turion64 1,6gh. For processing, it 
takes about 40 minutes. This is the code i used: dados=read.csv("C:/Documents and 
Settings/Administrador/Meus documentos/My Dropbox/Estat?stica/Association Rules/Top2009 alterado.csv", 
header=T, sep=";") library(arules) bin=as(dados, "transactions") rules <- apriori(bin, 
parameter = list(support = 0.01, confidence = 0.6)) Below is a sample of the file. I export data as CSV from 
excel. I have 71 columns (variabels) an 1001 lines (responses) 1. Churrascaria 2. Supermercado 3. Restaurante 
Self Service 4. Restaurante Chin?s 5. Buffet 6. Sorvete Galp?o Nelore Super Muffato N?o Sabe Jin Jin Planalto 
S?vio Vento Sul Super Muffato N?o Sabe N?o Sabe N?o Sabe Doce Ver?o Ga?cha Super Muffato N?o Sabe N?o Sabe N?o 
Sabe Kibon Tradi??o Ga?cha Super Muffato N?o Sabe N?o Sabe N?o Sabe Nestl? N?o Sabe Super Muffato N?o Sabe N?o 
Sabe Estilo S?vio Rancho Grill Viscardi Akira Akira N?
o Sabe N?o Sabe Thank you very for your helping!!! On 3 dez, 01:46, Steve Lianoglou 
<mailinglist.honey...@gmail.com> wrote:
> Hi,
>
> On Wed, Dec 2, 2009 at 6:57 PM, Alexandre - UEL <shima...@gmail.com> wrote:
>
> > Hello everybody!
> > I'm trying some datamining, but i'm having some problems with arule
> > package, in the end of processing R "had to be closed". I already
> > tryied to reinstall the 2.10 version, change the computer and
> > realocated more virtual memory.
>
> > Does anyone had this problem to?
>
> > I had a hiphoteses that i have to prepare the data, somehow i don't
> > know.
>
> > Thanks for helping!!!
>
> Can you provide more info here?
>
> 1. I'm assuming since you're talking about reallocating virtual
> memory, or whatever, you're on windows?
> 2. What's the exact error you're getting (what's it saying before "R
> 'had to be closed'"?
> 3. What's the size of your data? What type of data is it?
> 4. How much RAM do you have?
> 5. Are you on a 32 or 64 bit system?
> 6 What happens if you cut your data in half?
> 6. Can you provide a (very small) reproducible example of your data + code?
> ...
>
> -steve
>
> --
> Steve Lianoglou
> Graduate Student: Computational Systems Biology
>  | Memorial Sloan-Kettering Cancer Center
>  | Weill Medical College of Cornell University
> Contact Info:http://cbio.mskcc.org/~lianos/contact
>
> ______________________________________________
> r-h...@r-project.org mailing listhttps://stat.ethz.ch/mailman/listinfo/r-help
> PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html
> and provide commented, minimal, self-contained, reproducible code.





--
  Michael Hahsler
  email: mich...@hahsler.net
  web: http://michael.hahsler.net

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to