Re: [R] Arules - Association Rules

2009-12-04 Thread Alexandre Shimada
Michael, thanks a lot, it works!
Now I have to study the dataset to mining some interesting rules.
Just one more question, I saw ways to find rules that contains some itens.
But,  Is there a method to find rules that doesn't have a item?

Thanks again! You were very helpfull!!!


2009/12/4 Michael Hahsler hahs...@ai.wu.ac.at

 Alexandre,

 You are mining for association rules with an absolute support count of
 1001*0.01 = 10.1. From your 40 minutes processing time I guess that your
 survey data set is very dense and that results in an enormous number of
 frequent itemsets (potentially up to 2^k - k - 1 which is in your case for
 k=71 about 10^21) which causes you to run out of memory. You have the
 following options:

 * increase minimum support (e.g., start with 0.1) and see how low you can
 go without using up all your memory (I don't know how to watch memory usage
 on Windows).

 * restrict the maximal length of frequent itemsets by using
 parameter=list(maxlen = 3, support = ...)

 If that does not help I will need your dataset and some code to reproduce
 and study the problem.

 -Michael


  Hi I'm a windows XP user. My notebook have 1gb ram, 160gb hd, processor
 amd turion64 1,6gh. For processing, it takes about 40 minutes. This is the
 code i used: dados=read.csv(C:/Documents and Settings/Administrador/Meus
 documentos/My Dropbox/Estat?stica/Association Rules/Top2009 alterado.csv,
 header=T, sep=;) library(arules) bin=as(dados, transactions) rules -
 apriori(bin, parameter = list(support = 0.01, confidence = 0.6)) Below is a
 sample of the file. I export data as CSV from excel. I have 71 columns
 (variabels) an 1001 lines (responses) 1. Churrascaria 2. Supermercado 3.
 Restaurante Self Service 4. Restaurante Chin?s 5. Buffet 6. Sorvete Galp?o
 Nelore Super Muffato N?o Sabe Jin Jin Planalto S?vio Vento Sul Super Muffato
 N?o Sabe N?o Sabe N?o Sabe Doce Ver?o Ga?cha Super Muffato N?o Sabe N?o Sabe
 N?o Sabe Kibon Tradi??o Ga?cha Super Muffato N?o Sabe N?o Sabe N?o Sabe
 Nestl? N?o Sabe Super Muffato N?o Sabe N?o Sabe Estilo S?vio Rancho Grill
 Viscardi Akira Akira N?

 o Sabe N?o Sabe Thank you very for your helping!!! On 3 dez, 01:46, Steve
 Lianoglou mailinglist.honey...@gmail.com wrote:

  Hi,

 
  On Wed, Dec 2, 2009 at 6:57 PM, Alexandre - UEL shima...@gmail.com
 wrote:
 

   Hello everybody!
   I'm trying some datamining, but i'm having some problems with arule
   package, in the end of processing R had to be closed. I already
   tryied to reinstall the 2.10 version, change the computer and
   realocated more virtual memory.

 

   Does anyone had this problem to?

 

   I had a hiphoteses that i have to prepare the data, somehow i don't
   know.

 

   Thanks for helping!!!

 
  Can you provide more info here?
 
  1. I'm assuming since you're talking about reallocating virtual
  memory, or whatever, you're on windows?
  2. What's the exact error you're getting (what's it saying before R
  'had to be closed'?
  3. What's the size of your data? What type of data is it?
  4. How much RAM do you have?
  5. Are you on a 32 or 64 bit system?
  6 What happens if you cut your data in half?
  6. Can you provide a (very small) reproducible example of your data +
 code?
  ...
 
  -steve
 
  --
  Steve Lianoglou
  Graduate Student: Computational Systems Biology
   | Memorial Sloan-Kettering Cancer Center
   | Weill Medical College of Cornell University
  Contact 
  Info:http://cbio.mskcc.org/~lianos/contacthttp://cbio.mskcc.org/%7Elianos/contact
 
  __
  r-h...@r-project.org mailing listhttps://
 stat.ethz.ch/mailman/listinfo/r-help
  PLEASE do read the posting guidehttp://
 www.R-project.org/posting-guide.html

  and provide commented, minimal, self-contained, reproducible code.






 --
  Michael Hahsler
  email: mich...@hahsler.net
  web: http://michael.hahsler.net




-- 
É inútil e estúpido lastimar o mundo.
Sofrimento e prazer são frutos do que faz
Nissen Shounin

[[alternative HTML version deleted]]

__
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.


Re: [R] Arules - Association Rules

2009-12-03 Thread Alexandre Shimada
Hi
I'm a windows XP user.
My notebook have 1gb ram, 160gb hd, processor amd turion64 1,6gh. For
processing, it takes about 40 minutes.

This is the code i used:

dados=read.csv(C:/Documents and Settings/Administrador/Meus documentos/My
Dropbox/Estatística/Association Rules/Top2009 alterado.csv, header=T,
rep=;)
library(arules)
bin=as(dados, transactions)
rules - apriori(bin, parameter = list(support = 0.01, confidence = 0.6))

Above is a sample os fthe file. I export data as CSV from excel. I have 71
columns (variabels) an 1001 lines (responses)

 1. Churrascaria 2. Supermercado 3. Restaurante Self Service 4. Restaurante
Chinês 5. Buffet 6. Sorvete 7. Frango  Galpão Nelore Super Muffato Não Sabe Jin
Jin Planalto Sávio Big Frango  Vento Sul Super Muffato Não Sabe Não Sabe Não
Sabe Doce Verão Frangão  Gaúcha Super Muffato Não Sabe Não Sabe Não Sabe
Kibon Sadia  Tradição Gaúcha Super Muffato Não Sabe Não Sabe Não Sabe
Nestlé Big
Frango  Não Sabe Super Muffato Não Sabe Não Sabe Estilo Sávio Big
Frango  Rancho
Grill Viscardi Akira Akira Não Sabe Não Sabe Não Sabe


Thank you very for your helping!!!

2009/12/3 Steve Lianoglou mailinglist.honey...@gmail.com

 Hi,

 On Wed, Dec 2, 2009 at 6:57 PM, Alexandre - UEL shima...@gmail.com
 wrote:
  Hello everybody!
  I'm trying some datamining, but i'm having some problems with arule
  package, in the end of processing R had to be closed. I already
  tryied to reinstall the 2.10 version, change the computer and
  realocated more virtual memory.
 
  Does anyone had this problem to?
 
  I had a hiphoteses that i have to prepare the data, somehow i don't
  know.
 
  Thanks for helping!!!

 Can you provide more info here?

 1. I'm assuming since you're talking about reallocating virtual
 memory, or whatever, you're on windows?
 2. What's the exact error you're getting (what's it saying before R
 'had to be closed'?
 3. What's the size of your data? What type of data is it?
 4. How much RAM do you have?
 5. Are you on a 32 or 64 bit system?
 6 What happens if you cut your data in half?
 6. Can you provide a (very small) reproducible example of your data + code?
 ...

 -steve

 --
 Steve Lianoglou
 Graduate Student: Computational Systems Biology
  | Memorial Sloan-Kettering Cancer Center
  | Weill Medical College of Cornell University
 Contact Info: 
 http://cbio.mskcc.org/~lianos/contacthttp://cbio.mskcc.org/%7Elianos/contact




-- 
É inútil e estúpido lastimar o mundo.
Sofrimento e prazer são frutos do que faz
Nissen Shounin

[[alternative HTML version deleted]]

__
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.