Scanning various lists for R, I've noticed that a few people have raised the
question of getting the "rgl" package to run in R on the Mac operating
system. As far as I can tell (and I must admit to being a novice here), the
problem has to do with the inclusion of the /usr/X11R6/lib in R's
environm
Thanks for everyone's help, there seems to be many ways of solving the
problem that work well.
Peter
At 12:36 AM 7/1/2004, Gabor Grothendieck wrote:
Assuming all the entries are non-negative and non-NA this will do it:
DF[rowSums(DF) > 0,]
Peter Wilkinson videotron.ca> writes:
:
: here is a
his ...
h looks like your working with microarray data as well The actual
matrix that I am working with is 19000 x 340
Thanks for the help, that was perfect. I am still getting used to the R
language way of doing things.I will look into the apply function as you
have written it.
Pete
Assuming all the entries are non-negative and non-NA this will do it:
DF[rowSums(DF) > 0,]
Peter Wilkinson videotron.ca> writes:
:
: here is a snippet of data where I would like to drop all rows that have
: zeros across them, and keep the rest of the rows while maintaining the row
: n
You right its a matrix (I ran an is.matrix() on my object).
Thanks,
Peter
At 12:13 AM 7/1/2004, Patrick Connolly wrote:
On Wed, 30-Jun-2004 at 11:57PM -0400, Peter Wilkinson wrote:
|> here is a snippet of data where I would like to drop all rows that have
|> zeros across them, and keep the rest of
On Wed, 30-Jun-2004 at 11:57PM -0400, Peter Wilkinson wrote:
|> here is a snippet of data where I would like to drop all rows that have
|> zeros across them, and keep the rest of the rows while maintaining the row
|> names (1,2,3, ...10). The idea here is that a row of zeros is an indication
|>
here is a snippet of data where I would like to drop all rows that have
zeros across them, and keep the rest of the rows while maintaining the row
names (1,2,3, ...10). The idea here is that a row of zeros is an indication
that the row must be dropped. There will never be the case where there is
For those interested in experimenting with (compiling / developing)
a version of R for the Common Language Runtime (CLR)
environment (Microsoft .Net, Novell Ximan Mono & DotGNU)
-- a few links to some "free" compliers:
C
DotGNU (the official GNU project)
http://dotgnu.org/
http://www.southern-s
<[EMAIL PROTECTED]>:
Rejected for the following Virus
Virus Name: Worm.SomeFool.Gen-1
User Rejected Message (#5.1.1)
--- Below this line is a copy of the message headers.
>From [EMAIL PROTECTED] Wed Jun 30 22:26:39 2004
Received: from fairfieldi.com (68-233-97-39.ironoh.adelphia.net [68.233.97.
> From: [EMAIL PROTECTED]
>
> Hi,
> I´m new in R. I´m working with similarity coefficients for clustering
> items. I created one function (coef), to calculate the
> coefficients from
> two pairs of vectors and then, as an example, the function
> simple_matching,
> taking a data.frame(X) and using
An R-News or J. of Statistical Software note titled, "Getting to the
Source
of the Problem" detailing the basic strategies for these adventures in
the
new world of S4 methods and namespaces would be very useful.
url:www.econ.uiuc.edu/~rogerRoger Koenker
email [EMAIL PROTEC
That's because `assist' has a namespace:
> library(assist)
Loading required package: nlme
> predict.ssr
Error: Object "predict.ssr" not found
> methods("predict")
[1] predict.ar*predict.Arima*
[3] predict.arima0*predict.glm
[5] predict.gls
> Oh yes. The "load package" under the "packages menu" in the Windows
> version
> does that. To check I typed "library(assist)" after starting R. Same
> behavior, ssr is found, but others like predict.ssr, and plot.ssr, give a
> "not found" message.
Short answer: Try using "predict" instead of "p
Olivier Collignon wrote:
> I have been doing data analysis/modeling in R, connecting to SQL
databases
> with RODBC (winXP client with R1.9.0 and win2k SQL server 2000).
>
> I am now trying to leverage some of the OLAP features to keep the data
> intensive tasks on the DB server side and only ke
Oh yes. The "load package" under the "packages menu" in the Windows version
does that. To check I typed "library(assist)" after starting R. Same
behavior, ssr is found, but others like predict.ssr, and plot.ssr, give a
"not found" message.
Thanks for the suggestion.
Mike
___
Hi, I can' t load a variable tcltk declared with tclVar, why is this?, the
exmple above explain me ,Thanks Ruben
a<-tclVar(init="")
> f<-function(){
+ a<<-"pipo"
+ }
> f()
> a
[1] "pipo"
> tclvalue(a)
Error in structure(.External("dotTcl", ..., PACKAGE = "tcltk"), class =
"tclObj") :
[tcl]
On 06/30/04 16:32, Peter Gaffney wrote:
>Hi!
>
>I'm having some issues on both conceptual and
>technical levels for selecting the right combination
>of variables for this model I'm working on. The basic,
>all inclusive form looks like
>
>lm(mic ~ B * D * S * U * V * ICU)
When you do this, you are
I am a new user to R. I installed and loaded the smoothing spline package
called "assist." The function "ssr" seems to work, but "predict.ssr" and
some others don't seem to be available, I get a "can't find" message. But
they appear in the extensions folder, c:program
files\R\rw1091\library\ass
Hi!
I'm having some issues on both conceptual and
technical levels for selecting the right combination
of variables for this model I'm working on. The basic,
all inclusive form looks like
lm(mic ~ B * D * S * U * V * ICU)
Where mic, U, V, and ICU are numeric values and B D
and S are factors with
Thanks. Adding the -shared to the Makeconf file fixed the problem. I
installed R from R-1.9.1-0.fdr.2.rh90.i386.rpm , but for some reason the
flag to share libraries was never set in SHLIB_CXXLDFLAGS.
rgl_0.64-13.tar.gz now installed without problems.
Thanks again,
Enrique
On Wed, 2004-06-30 a
It is amazing the amount of time that has been spent on this issue. In
most cases, if you do some timing studies using 'scan', you will find that
you can read some quite large data structures in a reasonable time. If you
initial concern was having to wait 10 minutes to have your data read in,
> <[EMAIL PROTECTED]> writes:
>
>> I did not use R ten years ago, but "reasonable" RAM amounts have
>> multiplied by roughly a factor of 10 (from 128Mb to 1Gb), CPU speeds
>> have gone up by a factor of 30 (from 90Mhz to 3Ghz), and disk space
>> availabilty has gone up probably by a factor of 10. S
Without trying to understand your code in detail let me just
assume you are trying to create a matrix, ret, whose i,j-th
entry is some function, f, of row i of X and row j of X.
In that case this should do it:
apply(X,1,function(x)apply(X,1,function(y)f(x,y)))
Date: Wed, 30 Jun 20
hi chaps: I just found out that this R post list is actively harvested
by some spammers. I used this account exclusively for r-help posts,
and promptly received a nice email with links to nigeria, india, and
china. this is not a good thing. can we obfuscate the email addresses
of posters, s
Sorry, I missed the original message, so piggybacking off of a reply.
> On Tue, 29 Jun 2004, Ruben Solis wrote:
> > I've tried installing the MacOS X binaries for R available at:
> > http://www.bioconductor.org/CRAN/
> > I'm running MacOS X version 10.2.8.
Since this is coming out of our mirror
Thanks for the pointers.
I've read them with care but I don't seem capable of making it work.
For example, if I do:
data(USArrests)
USArrests2<-USArrests
USArrests2[1,1]<-NA
princomp(USArrests2, cor = TRUE, na.action = "na.omit")
I get the error message:
Error in cov.wt(z) : x must contain finite v
I believe from your description that this is interval-censored survival
and can be handled by survreg.
It would also be very easy to amend polr to do this.
On Wed, 30 Jun 2004, Thomas Ribarits wrote:
> does anyone have a quick answer to the question of how to carry out
> interval regression in R
<[EMAIL PROTECTED]> writes:
> I did not use R ten years ago, but "reasonable" RAM amounts have
> multiplied by roughly a factor of 10 (from 128Mb to 1Gb), CPU speeds have
> gone up by a factor of 30 (from 90Mhz to 3Ghz), and disk space availabilty
> has gone up probably by a factor of 10. So, unle
Hi,
I´m new in R. I´m working with similarity coefficients for clustering
items. I created one function (coef), to calculate the coefficients from
two pairs of vectors and then, as an example, the function
simple_matching,
taking a data.frame(X) and using coef in a for cicle.
It works, but I believ
Hi,
does anyone have a quick answer to the question of how to carry out
interval regression in R. I have found "ordered logit" and "ordered
probit" as well as multinomial logit etc. The thing is, though, that I
want to apply logit/probit to interval-coded data and I know the cell
limits which are
Hi, all;
I have a question on memory utilization of R.
Does R use only RAM memory or can I make it use virtual memory?
Currently, I am using Mac OS X 10.3.3 with 1.5 GB RAM.
However, I need more memory for some large size problem right now.
Thanks in advance,
Tae-Hoon Chung, Ph.D
Post-doctoral Rese
> I suspect that what has happened is that memory, CPU speed, and I/O
> speed have evolved at different rates, so what used to be acceptable
> code in read.table() (in both R and S-plus) is now showing its
> limitations and has reached the point where it can take half an hour to
> read in, on a
To be careful, there's lots more to I/O than the functions read.table() &
scan() -- I was only commenting on those, and no inference should be made
about other aspects of S-plus I/O based on those comments!
I suspect that what has happened is that memory, CPU speed, and I/O speed
have evolved a
grp <- rep(1:5, each=3)
resp <- rnorm(15)
mu <- tapply(resp, grp, mean)
s <- tapply(resp, grp, sd)
stopifnot( identical( names(mu), names(s) ) )
LCL <- mu - 2*s # lower confidence limit
UCL <- mu + 2*s
Here I choose 2 as we expect 95% of the data to fall under 4 sd.
# Type 1
plot(names(mu),
On Tue, 29 Jun 2004, Ruben Solis wrote:
> I've tried installing the MacOS X binaries for R available at:
>
> http://www.bioconductor.org/CRAN/
>
> I'm running MacOS X version 10.2.8.
>
> I get a message indicating the installation is successful, but when I
> double-click on the R icon that shows u
The problem is that you compiled the code with wide characters, and on
your 10.2.8 the support functions are not compiled into the R binary
(which is what I would expect).
As to why you are getting wide chars, I suggest you ask on the R-SIG-Mac
list (https://www.stat.math.ethz.ch/mailman/listinf
boxplot(Response_time ~ seconds, data = my.data.frame)
Coburn Watson comcast.net> writes:
:
: Hello R'ers,
:
: I have a large set of data which has many y samples for each unit x. The
data
: might look like:
:
: Seconds Response_time
: --
: 0
Hello All,
I have built a package depmix and build a source package file from it for
distribution. I have done all this on the following platform:
platform powerpc-apple-darwin6.8
arch powerpc
os darwin6.8
system powerpc, darwin6.8
status
major1
minor
Thank you! It's interesting about S-Plus, since they apparently try to support
work with much larger data sets by writing everything out to disk (thus getting
around the, eg, address space limitations, I guess), so it is a little surprising
that they did not tweak the I/O more...
Thanks a
Hello R'ers,
I have a large set of data which has many y samples for each unit x. The data
might look like:
Seconds Response_time
--
0 0.150
0 0.202
0 0.065
1 0.110
1
Tony Plate blackmesacapital.com> writes:
> I get the best read performance out of S-plus by using
> a homegrown binary file format with each column stored in a contiguous
> block of memory and meta data (i.e., column types and dimensions) stored at
> the start of the file. The S-plus read fun
As far as I know, read.table() in S-plus performs similarly to read.table()
in R with respect to speed. So, I wouldn't put high hopes in finding much
satisfaction there.
I do frequently read large tables in S-plus, and with a considerable amount
of work was able to speed things up significantl
From Legendre & Legendre Numerical Ecology I read: "The Australians
Lance & Williams (1967a) give several variants of the Manhattan metric
including their Canberra metric (Lance & Williams 1966c)"
Lance &Williams (1967a) Mixed-data classificatory programs
I.Agglomerative systems. Aust. Comput.J.
I thought outlier tests were mainly superseded two decades ago by the use
of robust methods -- they certainly were in analytical chemistry, for
example. All outlier tests are "bad" in the sense that outliers will
damage the results long before they are detected. See e.g.
@Article{AMC.89a,
auth
I have been learning about some outlier tests -- Dixon
and Grubb, specifically -- for small data sets. When
I try help.start() and search for outlier tests, the
only response I manage to find is the Bonferroni test
avaiable from the CAR package... are there any other
packages the offer outlier te
Dear All
Do you know some functions that can perform
the PRINCIPAL SURFACE estimation?
Please give me a hint.
Thanks for your help in advance.
Fred
[[alternative HTML version deleted]]
__
[EMAIL PROTECTED] mailing list
https://www.stat.math.e
[EMAIL PROTECTED] wrote on 30/06/2004 16:05:15:
>
>
> I could not figure this out from the documentation: is there a way to
send
> formatted non-graphical data to a fancy output device (eg, latex, pdf...)
> For example, if I want to include the summary of a linear model in a
> document, I migh
I could not figure this out from the documentation: is there a way to send
formatted non-graphical data to a fancy output device (eg, latex, pdf...)
For example, if I want to include the summary of a linear model in a
document, I might want to have it automatically texified.
Thanks!
Igor
Douglas Bates <[EMAIL PROTECTED]> writes:
> Often when nls doesn't converge there is a good reason for it.
>
> I'm on a very slow internet connection these days and will not be able
> to look at the data myself but I ask you to bear in mind that, when
> dealing with nonlinear models, there are mo
Hi!
I'm afraid we need some more bits of information. Do you have
administrative privileges?, which version of R are you trying to install?,
Have you read the RMacOSX FAQ?
On Tue, 29 Jun 2004, Ruben Solis wrote:
> I've tried installing the MacOS X binaries for R available at:
>
> http://www.
Often when nls doesn't converge there is a good reason for it.
I'm on a very slow internet connection these days and will not be able
to look at the data myself but I ask you to bear in mind that, when
dealing with nonlinear models, there are model/data set combinations for
which there are no pa
On Wed, 30 Jun 2004, Angel Lopez wrote:
> I need to perform a principal components analysis on a matrix with
> missing values.
> I've searched the R web site but didn't manage to find anything.
?princomp
has a description of an na.action argument, and
help.search("missing values")
co
I need to perform a principal components analysis on a matrix with
missing values.
I've searched the R web site but didn't manage to find anything.
Any pointers/guidelines are much appreciatted.
Angel
__
[EMAIL PROTECTED] mailing list
https://www.stat.m
I am curious. What were the dimensions of this data set? Did this
person know use read.table(), or scan(). Did they know about the
possibility of reading the data one part at a time?
The way that SAS processes the data row by row limits what can be done.
It is often possible with scant loss
Martyn Plummer <[EMAIL PROTECTED]> writes:
> On Wed, 2004-06-30 at 00:52, E GCP wrote:
> > Thanks for your replies. I do not HTML-ize my mail, but free email accounts
> > do that and there is not a switch to turn it off. I apologize in advance.
> >
> > I installed R from the redhat package prov
On Wed, 2004-06-30 at 00:52, E GCP wrote:
> Thanks for your replies. I do not HTML-ize my mail, but free email accounts
> do that and there is not a switch to turn it off. I apologize in advance.
>
> I installed R from the redhat package provided by Martyn Plummer. It
> installed fine and witho
__
[EMAIL PROTECTED] mailing list
https://www.stat.math.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
Hi!
Have you downloaded the package multilevel first?
Best wishes,
Karl
___
Bestellen Sie Y! DSL und erhalten Sie die AVM "FritzBox SL" für 0.
Sie sparen 119 und bekommen 2 Monate Grundgebührbefreiung.
help.search("time")
gives
proc.time(base) Running Time of R
system.time(base) CPU Time Used
both of which could be used.
On Wed, 30 Jun 2004, zze-PELAY Nicolas FTRD/DMR/BEL wrote:
> Is there any function to mesurate the duration of a procedure (like tic
> and toc in matlab) ?
>
?system.time
*** REPLY SEPARATOR ***
On 30.06.2004 at 10:59 zze-PELAY Nicolas FTRD/DMR/BEL wrote:
>Hello ,
>Is there any function to mesurate the duration of a procedure (like tic
>and toc in matlab) ?
>
>Tic
>Source("procedure.R")
>Toc
>
>(toc is the duration between the exe
Hello ,
Is there any function to mesurate the duration of a procedure (like tic
and toc in matlab) ?
Tic
Source("procedure.R")
Toc
(toc is the duration between the execution of tic and the execution of
toc)
Thank you
nicolas
[[alternative HTML version deleted]]
__
Prof Brian Ripley <[EMAIL PROTECTED]> writes:
> and don't call the result `glm' (or call your data, `data'). Also, this
> is not a question about GLMs with that family but about linear models, so
> why not use lm.
>
> > summary(glm)
> > anova(glm, test="Chisq")
*
And sh
On Wed, 30 Jun 2004, Varrin muriel wrote:
> HI, I am a studient, so don't be surprise if my question seems so simple
> for you...
> I have a dataframe with 6 qualitative variables divided in 33
> modalities, 2 qualitatives variables and 78 lines. I use a glm to know
> wich variables have interact
It's undesirable to need Visual Studio (I suspect later versions than 6
also work), and Duncan Murdoch wrote some standalone code to do this which
he may be willing to share (but he is away until next week).
Meanwhile I have put edited .exe's for rw1091 at
http://www.stats.ox.ac.uk/pub/RWin/Edite
HI, I am a studient, so don't be surprise if my question seems so simple for you...
I have a dataframe with 6 qualitative variables divided in 33 modalities, 2
qualitatives variables and 78 lines. I use a glm to know wich variables have
interactions... I would like to know if its normal that one
I'm running R under a 32-bit Win2003 Server with 24 Go RAM ...
I read the FAQ and found that I have to edit my boot.ini file with the
"/PAE" switch; Done.
I noticed also that I have to set the R image header with the flag
"/LARGEADDRESSAWARE", but to do this I need Microsoft Visual Studio 6.0
so
Hi,
I wonder, why SAS should be better in time for reading a data in the system.
I have an example, that shows that R is (sometimes?, always?) faster.
-
Data with 14432 observations and 120 variables.
Time for reading the data:
SAS 8e:
data testt;
set l1.lse01;run;
real
Douglas Bates stat.wisc.edu> writes:
: If you are routinely working with very large data sets it would be
: worthwhile learning to use a relational database (PostgreSQL, MySQL,
: even Access) to store the data and then access it from R with RODBC or
: one of the specialized database packages.
68 matches
Mail list logo