546
FAILURE_ID_HASH: {3e222f7c-7009-e3ad-3df7-4c44c21d59ab}
Followup: MachineOwner
-
Regards
Mike Francis
Peirianydd SCCM / SCCM Engineer
Cyfoeth Naturiol Cymru / Natural Resources Wales
Ffon / Mobile: 07876847084
E-Bost / Email: mike.fran...@cyfoethnaturiolcymru.gov
Hi have a large number of zipped files, each containing 3 xml files that I want
to read. I would like to read one of the xml files without having to
decompress each zip file first.
If I run gzfile(path2zipped file) I get
A connection with
description "X:/Mkt Science/Projects/ tv/202109.ext/t
My experience is that the combination of OneDrive and R leads to lack of
productivity.
--
W. Michael Conklin
EVP Marketing & Data Sciences
GfK
M +1 612 567 8287
-Original Message-
From: R-help On Behalf Of Kevin Thorpe
Sent: Thursday, October 7, 2021 10:50 AM
To: Jeff Newmiller
Cc: R
all posts moderated?
Now that the weekend is upon me, I'll have some time to come up to speed on
all this. Thanks for your patience.
Thanks again!
Mike
On Fri, Sep 4, 2020 at 8:05 AM Marc Schwartz wrote:
>
> > On Sep 3, 2020, at 9:26 PM, Michael Feher wrote:
> >
> > Gree
Patrick, Rui, Murdoch, and Rasmus,
Thank you for your responses. It may not be fair of me to send out that
request just before leaving on vacation for a week. I appreciate you getting
back to me.
All the best,
Mike
-Original Message-
From: Rasmus Liland
Sent: Saturday, August
eturning 1.
But which.min only does so if the values don't contain fractions.
And I get
> identical (data3ba, c(2.9,2.9))
[1] FALSE
Why is which.min not always returning 1 but which.max does?
Mike
__
R-help@r-project.org mailing list -- To
type="n"' should be passed.
As a workaround for now I do something similar. I check if the range
to be plotted is completely NA. If so I manipulate one of those
observations (which can only be 0, 1 or NA) to an "illegal" value of
0.5.
Mike
__
s no indication what date these questions and answers were posted. If
it's true that R is only tested on supported versions of Windows, that would
exclude Vista and Windows 7.
The download site doesn't mention operating system requirements. Does 4.0.2
work on Windows 7?
Thank you
d).
This post
https://stat.ethz.ch/pipermail/r-sig-finance/2020q3/015000.html
suggests to generate the plot object with chart_Series and then to
explicitly set x$Env$ylim[[4]] before plotting - without success.
Can I tell plot/plot.window to ignore such errors and simply generate
an empty region instea
New package: mortyr. A wrapper to the The Rick and Morty API
https://rickandmortyapi.com/ and the most rickdiculous package in the R-verse.
Use it to return information about your favourite characters, locations, and
episodes from the show.
The released version can be found on CRAN:
https://C
I am trying to use a shiny app to update records in an sqlite database. I keep
running into the following error:
unable to find an inherited method for function 'dbSendQuery' for signature
'"src_dbi", "character"'
The query I am trying to send is:
[1] "update kpquestions set mrisupercat = 'Dem
alling into the global library and not a
user specific library
--
W. Michael Conklin
EVP Marketing & Data Sciences
GfK
T +1 763 417 4545 | M +1 612 567 8287
-Original Message-----
From: Conklin, Mike (GfK)
Sent: Friday, June 23, 2017 4:38 PM
To: Don Cohen; Duncan Murdoch
Cc: Martin
esn't appear to be R3.4
related. I am now reaching out to some other sources who may have installed R
on similar systems.
From: Conklin, Mike (GfK)
Sent: Friday, June 23, 2017 8:50 AM
To: Don Cohen; Duncan Murdoch
Cc: Martin Maechler; r-help@r-p
h
Cc: Conklin, Mike (GfK); Martin Maechler; r-help@r-project.org
Subject: Re: [R] Missing dependencies in pkg installs
Duncan Murdoch writes:
> On 22/06/2017 5:02 PM, Conklin, Mike (GfK) wrote:
> > I am using debug on the .install_packages function...stepping through.
> > Once t
11:44 NEWS
drwxr-xr-x 2 root root 4096 Jun 22 22:26 R
drwxr-xr-x 3 root root 8192 Jun 22 22:26 src
[root@dcex1102shinypr ~]#
If I CTRL-Z out of R in the first session I confirm the same result - the
system shows the file as executable
--
W. Michael Conklin
EVP Marketing & Data Sciences
G
ckages() exit status 1
--
W. Michael Conklin
EVP Marketing & Data Sciences
GfK
T +1 763 417 4545 | M +1 612 567 8287
-Original Message-
From: Duncan Murdoch [mailto:murdoch.dun...@gmail.com]
Sent: Thursday, June 22, 2017 11:09 AM
To: Conklin, Mike (GfK); Martin Maechler; David W
also it seems to be no issue to execute configure from the command line
./configure
--
W. Michael Conklin
EVP Marketing & Data Sciences
GfK
T +1 763 417 4545 | M +1 612 567 8287
-Original Message-
From: R-help [mailto:r-help-boun...@r-project.org] On Behalf Of Conklin, Mike
expand it using
tar zxvf stringi_1.1.5.tar.gz
and use
ls -l stringi/configure
to see if the system thinks it is executable, and in R,
file_test("-x", "stringi/configure")
to see what R sees.
Duncan Murdoch
On 22/06/2017 3:38 AM, Martin Maechler wrote:
>>>>>> David Wi
I have a Ubuntu server with an R installation that has 384 packages installed.
We are trying to replicate the system on a Red Hat Enterprise server. I
downloaded the list of packages from the Ubuntu machine and read it into an R
session on the new machine. Then I ran install.packages(listofpack
Ah, that works! Thank you!
From: PIKAL Petr
Sent: Tuesday, May 2, 2017 7:56:33 AM
To: C W; r-help
Subject: RE: [R] How create columns for squared values from previous columns?
Hi
you can use data.frame
data.frame(dat, dat[,1:3]^2)
and you can set names afterwar
Thanks Rolf.
I was just a bit frustrated that R wouldn't generate dummy variable names on
the fly.
Also, another question, if I want to put column 5 at column 3,
dat[, 3:5] <- dat[, c(5,3,4)]
It does not work, why?
From: Rolf Turner
Sent: Friday, April 28, 201
I was using OS X native R editor. I would imagine that editor is as simple and
native as it gets. But, if it's truly native, why would Gmail think of my code
chunk so differently.
I'm just throwing it out there! I can always remove format in Gmail after
pasting as a precaution. :)
On Fri,
forts. I look forward to the results.
Best regards,
Michael
unaffiliatd
> Gesendet: Freitag, 21. Oktober 2016 um 09:39 Uhr
> Von: "Berend Hasselman"
> An: "Mike meyer" <1101...@gmx.net>
> Cc: ProfJCNash , "r-help@r-project.org"
>
> Betr
For that reason it is (in my view) a bad idea to force the user to set up his
problem in
R-model notation.
Michael
unaffiliated
> Gesendet: Donnerstag, 20. Oktober 2016 um 15:26 Uhr
> Von: ProfJCNash
> An: "S Ellison" , "Mike meyer" <1101...@gmx.net>
>
How do you reply to a specific post on this board instead of the thread?
I am too incompetent to find this out myself.
Thanks,
Michael
unaffiliated
__
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listin
And finally, to put to rest the notion that the number of residuals is in any
way significant for the
solution of the least squares problem I submit to you the function
f(x,y)=(x²+y²)²
of 2 variables but only one residual f_1(x,y)=x²+y² which nonetheless has a
unique minimum
at the point (0,0).
>From my reading of the above cited document I get the impression that the
>algorithm
(algorithm 3.16, p27) can easily be adapted to handle the case m 0 and so the system becomes ill conditioned.
Why can we not get around this as follows: as soon as mu is below some threshold
we solve instead the
Make that f(x,u)=||x||².
__
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-c
@SE: yes, not every system of equations with more variables than equations is
solvable,
we need an additional condition e.g. full rank of the coefficient matrix.
Uniqueness of the solution was not required.
@BH:
Yes this is the document, it is a nice presentation.
I did not read the first page b
@pd: you know that a System of equations with more variables than equations is
always solvable
and if a unique solution is desired one of mimimal norm can be used.
According to "Methods for nonlinear least squares problems" by Madsen, Nielsen
and Tingleff the LM-algorithm
solves Systems of the f
Greetings,
The description of nls.lm specifies that in minimizing a sum of squares of
residuals
the number of residuals must be no less than the dimension of the independent
variable
("par").
In fact the algorithm does not work otherwise (termination code 0).
But this condition is sensel
Hello,
I have both 32 and 64 bit verions of R installed. What happens if I open a
workspace saved from 64 bit R
in the 32 bit version or conversely?
I am fairly careless but never noticed any problems.
__
R-help@r-project.org mailing list -- To UN
show.
Does anyone have any ideas?
-Mike
[[alternative HTML version deleted]]
__
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-
fined"
-is there a more elegant/memory efficient way of doing this?!
Any help much appreciated!
best wishes
mike
PS Two sample images here:
http://www.hsm.org.uk/1.tif
http://www.hsm.org.uk/2.tif
#Scan directory and store filenames in string, then count total files
files <- as.charact
the pixel values look like they might have been scaled. For example,
below. Is writeTIFF doing something Im not aware of??
> m_image_tiff1[1,1,1]
[1] 0.9745098
> test_tiff[1,1,1]
[1] 0.06532387
Thanks very much!
mike
#Scan directory and store filenames in string, then count total files
e <- rownames(df)
US> #Turn data from wide to long
US> ds<-melt(df, id.vars = "case")
US> ggplot(ds, aes(x = variable, y = value, group = case)) +
US> geom_point () + geom_line()
US> Hope this helps,
US> Ulrik
US> On Sat, 14 May 2016 at 10:20 Mike Smith
. the first row in
the original dataset). geo_segment allows me to specify start-end but I need to
do this over multiple time periods
Any help much appreciated
thanks
mike
library(reshape2)
library(ggplot2)
library(ggthemes)
library(cowplot)
#Read raw data
df = read.table("
om_abline(intercept = 0, slope = 1, size=1) +
coord_fixed()
p<-plot_grid(gg1,gg2,labels=c("A", "B"), nrow = 2)
title <- ggdraw() + draw_label("A Title", fontface='bold')
plot_grid(title, p, ncol=1, rel_heights=c(0.1, 1), align=c("h")) # re
Is there a way to get ggplot scale_colour_distiller to display all values in
the legend?
Currently using this code.
thanks!
mike
library(ggplot2)
#Input data: insert the filename for raw data
data <-
read.csv("http://www.lecturematerials.co.uk/data/learning_bands.csv",header=T)
>>> On Apr 30, 2016, at 12:58 PM, Mike Smith wrote:
>>> Hi
>>> First post and a relative R newbie
>>> I am using the vioplot library to produce some violin plots.
DW> It's a package, not a library.
>>> I have an input CSV
names=c("Y6",
"Y5","Y4","Y3","Y2","Y1"), col = "lightblue")
Two queries:
1. Is there a more elegant way of automatically stripping the NAs, passing the
columns to the function along w
Hi all,
I am trying to figure out how to do this in R and I need your help.
My journey started from something like the following:
http://stackoverflow.com/questions/11365857/real-time-auto-updating-incremental-plot-in-r/1#1
n=1000
df=data.frame(time=1:n,y=runif(n))
window=100
for(i in 1:(n�
R silently converts the integer to a character for comparison in the subset
operation. But if we explicitly do the conversion we see that it does not work
with the default R settings.
> as.character(10)
[1] "1e+05"
> as.character(9)
[1] "9"
--
W. Michael Conklin
EVP Marketing & Da
te it.
Best regards,
Mike
--
W. Michael Conklin
Executive Vice President
Marketing & Data Sciences - North America
GfK | 8401 Golden Valley Road | Minneapolis | MN | 55427
mike.conk...@gfk.com
T +1 763 417 4545 | M +1 612 567 8287
www.gfk.com
___
. the long description
of the variable name?)
Thanks for any help,
Mike
--
W. Michael Conklin
Executive Vice President
Marketing & Data Sciences - North America
GfK | 8401 Golden Valley Road | Minneapolis | MN | 55427
mike.conk...@gfk.com
T +1 763 417 4545 | M +1 612 567 8287
www.gfk.co
Hello,
In the following code, any characters matching "/|@| \\|") will be changed to
a space.
> library(tm)
> toSpace <- content_transformer(function(x, pattern) gsub(pattern, " ", x))
> docs <- tm_map(docs, toSpace, "/|@| \\|")
What code would transform all non-letters to a space? (What goes
findAssocs() is not working, as is seen below. "Lucid" and "dreaming" occur
together quite often in the book.
The corpus is a single document, the text version of a book. Does this
function require at least two documents? If so, if I split the book in half
will I get the correlations regardi
Here's the code and results. The corpus is the text version of a single book.
(r vs. 3.2)
> docs <- tm_map(docs, stemDocument)
> dtm <- DocumentTermMatrix(docs)
> freq <- colSums(as.matrix(dtm))
> ord <- order(freq)
> freq[tail(ord)]
one experi will can lucid dream
287 312 363 452 1
e/master.zip
Installing package from
C:\Users\MIKE~1.CON\AppData\Local\Temp\RtmpOG7HXq/master.zip
Installing swirl
"C:/Program Files/R/R-31~1.1/bin/i386/R" --vanilla CMD INSTALL \
"C:\Users\mike.conklin\AppData\Local\Temp\RtmpOG7HXq\devtoolsd1c48a85b74\swirl-master"
\
--lib
On Sat, 7 Feb 2015, Mike Miller wrote:
res <- residuals( model )
resStd <- ( res - mean( res, na.rm=TRUE ) ) / sd( res, na.rm=TRUE )
Another issue is how to make the theoretical quantiles for the normal
distribution. There are a few methods:
https://www.statsdirect.co
1 ) )
resNorm[ is.nan( resNorm ) ] <- NA
Then you could plot it directly:
plot(resNorm, resStd)
When we use qqnorm() in R, it looks like R is using a Blom method with
c=1/2 instead of c=3/8. I believe Blom recommended 3/8 and programs that
offer Blom normal scores use c=3/8.
Best,
Mike
S and R could have been written so that
"<-" had to be followed by a space or an error would result. That might
have been a good idea, but it's too late now. So we just have to not make
mistakes. I'm working on it.
Mike
_
stat.com/pages/Tutor/R_inferno.pdf
http://biostat.mc.vanderbilt.edu/wiki/Main/LinuxWorkshopRProgramingTipsAndGotchas
http://tim-smith.us/arrgh/
I didn't happen to see my example on any list, but I didn't read them
thoroughly, so it's probably there.
Mike
__
h(snps)
4.0146146985 bytes
That saves 93.8% of the memory by dropping 0.28% of the markers and
encoding as integers instead of strings. I might end up doing this by
encoding the other characters as negative integers.
Mike
__
R-hel
Thanks, Jeff. You really know the packages. I search and I guess I
didn't use the right terms. That package seems to do exactly what I
wanted.
Mike
On Tue, 13 Jan 2015, Jeff Newmiller wrote:
On Tue, 13 Jan 2015, Mike Miller wrote:
I have many pairs of data frames each with abo
e done about it?
The inspiration for my question comes partly from the way GNU comm works.
If you have any ideas about this, I'd love to hear them.
Thanks in advance.
Mike
--
Michael B. Miller, Ph.D.
University of Minnesota
http://scholar.google.com/citatio
g,
though.
Henrik -- I think you are saying that your experience has shown that the
code you wrote for catching a corner case was not needed. Is that right?
Mike
On Tue, 13 Jan 2015, Henrik Bengtsson wrote:
I/we've been utilizing both read and write seek():s on *binary*
connections a
run under Windows that can seek (is it fseek in C?) reliably to a
position in a file. If that is the case, it's going to be hard to develop
good systems for managing bioinformatic data on Windows.
Thanks in advance.
Mike
--
Michael B. Miller, Ph.D.
University of Minnesot
r again, so that speed is excellent.
Mike
On Mon, 5 Jan 2015, Peter Alspach wrote:
Tena koe Mike
An alternative, which is slightly fast:
diffv <- diff(v)
starts <- c(1, which(diffv!=1)+1)
cbind(v[starts], c(diff(starts), length(v)-starts[length(starts)]+1))
Peter Alspach
-Ori
there are many consecutive variables
to be read, I can multiply the X in n=X by that number instead of doing
many different seek() calls. (The data are in a transposed format where I
read in every record for some variable as sequential elements.) I'm
probably not the first person to deal with
On Sun, 4 Jan 2015, Duncan Murdoch wrote:
On 04/01/2015 5:13 PM, Mike Miller wrote:
The help doc for readBin writeBin tells me this:
Handling R's missing and special (Inf, -Inf and NaN) values is discussed
in the ‘R Data Import/Export’ manual.
So I go here:
http://cran.r-project.or
but I guess that
doesn't work with the binary read system. I don't think I can scan the
readBin input because it isn't a file or stdin.
Mike
__
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/l
s me
wonder what else can handle it. uint16 seems the safer bet, and there is
no loss of precision. Of course, the downside is that the uint16 file is
twice as big as the uint8 file, and these files may be several hundred GB
in size.
Mike
On Sun, 4 Jan 2015, Uwe Ligges wrote:
Following
.1.1
Also, this is how the hexbin package is described:
"Description" Binning and plotting functions for hexagonal bins."
So I guess that suggestion wasn't helping me much, either.
Mike
On Sat, 3 Jan 2015, Jeff Newmiller wrote:
Your message is missing either a reproducible exam
answer for people who ask why I make
them divide by 1000 all the time. ;-)
Mike
__
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-proj
On Fri, 2 Jan 2015, Duncan Murdoch wrote:
On 01/01/2015 10:05 PM, Mike Miller wrote:
This is how big those errors are:
512*.Machine$double.eps
[1] 1.136868e-13
Under other conditions you also were seeing errors of twice that, or
1024*.Machine$double.eps. It might not be a coincidence
On Thu, 1 Jan 2015, Duncan Murdoch wrote:
On 01/01/2015 1:21 PM, Mike Miller wrote:
I understand that it's all about the problem of representing digital
numbers in binary, but I still find some of the results a little
surprising, like that list of numbers from the table() output.
UE
So they give the same answer, but converting to character takes about 25
times longer.
Mike
On Thu, 1 Jan 2015, ted.hard...@wlandres.net wrote:
I've been followeing this little tour round the murkier bistros
in the back-streets of R with interest! Then it occurred to me:
What is wrong
On Thu, 1 Jan 2015, Duncan Murdoch wrote:
On 01/01/2015 1:21 PM, Mike Miller wrote:
On Thu, 1 Jan 2015, Duncan Murdoch wrote:
On 31/12/2014 8:44 PM, David Winsemius wrote:
On Dec 31, 2014, at 3:24 PM, Mike Miller wrote:
This is probably a FAQ, and I don't really have a question abo
rmats,
so I had the data already in the format I described.
Mike
On Thu, 1 Jan 2015, Richard M. Heiberger wrote:
Interesting. Following someone on this list today the goal is input
the data correctly.
My inclination would be to read the file as text, pad each number to
the right, drop the de
On Thu, 1 Jan 2015, Duncan Murdoch wrote:
On 31/12/2014 8:44 PM, David Winsemius wrote:
On Dec 31, 2014, at 3:24 PM, Mike Miller wrote:
This is probably a FAQ, and I don't really have a question about it, but I just
ran across this in something I was working on:
as.integer(1000*
e faster
in my application:
as.integer( 1000*1.003 + .1 )
[1] 1003
FYI - I'm reading in a long vector of numbers from a text file with no
more than three digits to the right of the decimal. I'm converting them
to integers and saving them in bi
(for some notion of vector) in the sense that
all Lisp objects are lists, all APL objects are arrays and all tcl objects
are character strings." That's how I've been thinking about it, too, but
I'm not sure that *all* data objects are vectors in this sense. If that
were the ca
On Thu, 25 Dec 2014, David Winsemius wrote:
On Dec 25, 2014, at 1:04 AM, Mike Miller wrote:
I just wanted to put this out there. It's just some of my observations
about things that happen with R, or happened in this particular
investigation. There were definitely some lessons for
On Thu, 25 Dec 2014, Mike Miller wrote:
I was going to ask a question about it how to test that an object is a
vector, but then I found this:
"is.vector() does not test if an object is a vector. Instead it returns
TRUE only if the object is a vector with no attributes apart from names.
On Thu, 25 Dec 2014, Duncan Murdoch wrote:
On 25/12/2014 1:57 PM, Mike Miller wrote:
I do think I get what is going on with this, but why should I buy into
this conceptualization? Why is it better to say that a matrix *is* a
vector than to say that a matrix *contains* a vector? The latter
Thanks, but I was already in touch with Rob Kirkpatrick about it. We all
work together at U Minnesota, or did until Rob went to VCU.
Mike
On Thu, 25 Dec 2014, Uwe Ligges wrote:
This is a rather detailed analysis, thanks, but I think it should be
send to the maintainer of the "
On Thu, 25 Dec 2014, peter dalgaard wrote:
On 25 Dec 2014, at 08:15 , Mike Miller wrote:
"is.vector returns TRUE if x is a vector of the specified mode having
no attributes other than names. It returns FALSE otherwise."
So that means that a vector in R has no attributes other
On Thu, 25 Dec 2014, Mike Miller wrote:
On Thu, 25 Dec 2014, Jeff Newmiller wrote:
You have written a lot, Mike, as though we did not know it. You are
not the only one with math and multiple computing languages under your
belt.
I'm not assuming that you and Bert don't know thi
On Thu, 25 Dec 2014, Jeff Newmiller wrote:
You have written a lot, Mike, as though we did not know it. You are not
the only one with math and multiple computing languages under your belt.
I'm not assuming that you and Bert don't know things, but I do expect to
have a wider audien
probably would have caught it sooner and I
would have understood the problem.
This is how I'll recommend they fix the bug in the code (thanks to those
of you who helped with this):
temp.vec <- as.character( test.dat$FAMID[ test.dat$FTYPE != 6 ] )
test.dat$famsize[ test.dat$FTYPE !
tp://en.wikipedia.org/wiki/Vectorization_(mathematics)
But in math as in computing, we wouldn't say that a matrix *is* a vector.
If vec(A) = v, that does not mean that A = v. In R, it looks like
as.vector() can do what vec() does, and more.
Mike
--
Michael B. Miller, Ph.D.
Unive
On Wed, 24 Dec 2014, Jeff Newmiller wrote:
On December 24, 2014 6:49:47 PM PST, Mike Miller wrote:
On Wed, 24 Dec 2014, Mike Miller wrote:
Also, regarding the sacred text, "x A numeric." is a bit terse. The
same text later refers to length(x), so I suspect that "A numeric&
On Wed, 24 Dec 2014, Mike Miller wrote:
Also, regarding the sacred text, "x A numeric." is a bit terse. The
same text later refers to length(x), so I suspect that "A numeric" is
short for "A numeric vector", but that might not mean "a vector of
'
:
as.vector(table( as.character(vec) )[as.character(vec)])
If there are, say, 10,000 different elements in vec, each repeated an
average of 5-10 times, will this still work correctly? In other words,
the length of the table output array is unlimited, right?
Mike
htly trickier with an integer vector:
intvec <- c(4,4,5,6,6,6)
table( intvec )[intvec]
intvec
NA NA NA NA NA NA
as.vector(table( intvec )[as.character(intvec)])
[1] 2 2 1 3 3 3
So I think this will always work for vectors of either type:
as.vector(table( as.character(vec) )[as
ork, but is it OK?:
ave( rep(1, length(charvec)), as.factor(charvec), FUN=sum )
[1] 1 2 3 2 3 3
I suspect that ave() isn't the best choice, but what is the best way to do
this?
Thanks in advance.
Mike
__
R-help@r-project.org mailing
home/meb/source/R-3.1.1/src/library/tools'
make[2]: *** [R] Error 1
make[2]: Leaving directory `/home/meb/source/R-3.1.1/src/library'
make[1]: *** [R] Error 1
make[1]: Leaving directory `/home/meb/source/R-3.1.1/src'
make: *** [R] Error 1
$
The read.table.ffdf function in the ff package can read in delimited files
and store them to disk as individual columns. The ffbase package provides
additional data management and analytic functionality. I have used these
packages on 15 Gb files of 18 million rows and 250 columns.
On Tuesday
statistics to work with a few eigenvectors instead of with all of them.
Mike
On Fri, 2 May 2014, Yixuan Qiu wrote:
Exactly.
The syntax is intended to mimic eigs() in Matlab and Octave.
library(rARPACK)
eigs(X, 10) # If your X is of class "dsyMatrix"
eigs_sym(X, 10) # If X i
to the [V,d] = eigs(X,K) function in Octave/MATLAB.
Best,
Mike
--
Michael B. Miller, Ph.D.
Minnesota Center for Twin and Family Research
Department of Psychology
University of Minnesota
http://scholar.google.com/citations?user=EV_phq4J
__
ersion as expected. That's an
Octave problem, so I'll deal with them on that one. I might not have zlib
compiled in, or maybe they still have a bug in that function.
Thanks!
Mike
__
R-help@r-project.org mailing list
https://stat.e
ger(), 7845*5000, size=2,
signed=FALSE ), ncol=7845, byrow=TRUE ) )
user system elapsed
3.769 0.138 3.906
close(con)
The writeBin() 16-bit integer gzipped file, converted to numeric by
dividing by 1000 on the fly:
system.time( D <- matrix( readBin( con, integer(), 7845*5000, size=2
On Mon, 17 Mar 2014, Duncan Murdoch wrote:
On 14-03-17 6:22 PM, Mike Miller wrote:
Thanks! Another thing I've figured out: Use of "drop0trailing=T" in
format() fixes the .0 stuff that I didn't like:
write.table(format(data[1:10,], digits=5, trim=T, drop0trail
On Mon, 17 Mar 2014, Berend Hasselman wrote:
On 17-03-2014, at 21:03, Mike Miller wrote:
…...
data[,c(5:9,11,13,17:21)] <- signif(data[,c(5:9,11,13,17:21)], digits=5)
Then write.table(data) does what I'd want. It works better than format().
Example:
data2 <- data
data2[,c(5
On Sun, 16 Mar 2014, Duncan Murdoch wrote:
On 14-03-16 2:13 AM, Mike Miller wrote:
I always knew there was some numerical reason why I was getting very
long stretches of 9s or 0s in the write.table() output, but my concern
is really with how to prevent that from happening. So the question
On Sat, 15 Mar 2014, peter dalgaard wrote:
On 15 Mar 2014, at 20:54 , Mike Miller wrote:
$ cat data1.txt
0.005
0.00489
I don't know why it shows 17 digits and doesn't round to 15, but it is showing
that the numbers are different, for some reason.
Aiding my weakenin
some small amount so that machine precision isn't
constantly causing annoying output?
Thanks.
Mike
--
Michael B. Miller, Ph.D.
Minnesota Center for Twin and Family Research
Department of Psychology
University of Minnesota
http://scholar.google.com/citations?user=EV_phq4J
___
2503131e-16
I thing it is an undesirable feature that write.table() has to try to use
so many digits -- all the way out to eps -- because it causes these kinds
of inconsistencies. Is there no way to get it to put out numbers in the
obviously desirable way without having to write format statements?
Mike
On Sat, 15 Mar 2014, peter dalgaard wrote:
I don't think so. I think some of your numbers differ sufficiently from
numbers with only a few digits to the right of the decimal that
write.table needs to write them with increased precision. You didn't
read them like that, didn't you? You did some
1 - 100 of 858 matches
Mail list logo