Dear Edzer,
 
I do have a similar problem (extensive computation time, memory limits). I 
would like to run interpolation of 60,000 points for 500,000 new points using 
OK and nmax of 30 closest points (I am using a windowsxp machine with 2GB RAM).
 
But I have doubts that it can ever run in less than few hours in gstat. I mean, 
at each new location, the algorithm has to derive distances to all locations 
(60,000^2 matrix) and then find the 30 closest ones. There is no more 
intelligent way to pick up the closest points - or is there?
 
Thanx,
 
Tom
 

________________________________

From: [EMAIL PROTECTED] on behalf of Dave Depew
Sent: Fri 9/19/2008 5:00 PM
To: Edzer Pebesma
Cc: Chris Taylor; r-sig-geo@stat.math.ethz.ch
Subject: Re: [R-sig-Geo] cokriging question



Thanks for the quick responses.
I've often done global OK or UK that can take ~ 2-3 days to complete. I
always assumed that it was b/c the matrices were so large. Looking at
task manager indicated that Rgui.exe only consumed ~ 800 Mb of RAM
during the processes.
I'll try it by passing the maxdist to gstat() rather than predict.gstat().
The only other time I encounter the memory.c line is when I neglect to
remove duplicate observations.

I think I only should have at most 500 or so points within a 100-200m
maxdist...




Edzer Pebesma wrote:
> Good morning Dave (late afternoon here),
>
> Chris Taylor wrote:
>> Good morning Edzer and Dave,
>> Thanks for bringing up this point.  I had a similar issue recently
>> using krige().  Observations at 5800 locations, attempting to krige()
>> predictions at 112,000 locations resulted in the same "memory.c"
>> error message.  Reducing predicted locations to <<50,000 and reducing
>> max.dist seemed to help, but the predictions still took a very long
>> time (>2 hours).  (Running winxp with 4GB memory.)
>> Can you clarify your suspicion regarding the "lack of standardization
>> of coordinates"?
> In this message, a trend was modelled based on x and y coordinates, as
> follows:
>       x                y                DN4          indicator4
> Min.   :670462   Min.   :4215236   Min.   :18.00   Min.   :0.0000
> 1st Qu.:670683   1st Qu.:4215456   1st Qu.:24.00   1st Qu.:0.0000
> Median :670904   Median :4215677   Median :32.50   Median :0.0000
> Mean   :670904   Mean   :4215677   Mean   :43.26   Mean   :0.4795
> 3rd Qu.:671125   3rd Qu.:4215898   3rd Qu.:64.00   3rd Qu.:1.0000
> Max.   :671346   Max.   :4216119   Max.   :87.00   Max.   :1.0000
> >
> g<-gstat(id="indicator6",formula=indicator6~x+y+x*y+sqrt(x)+sqrt(y),location=~x+y,data=band6.data,...
>
>
> computing the x*y will give numbers many orders of magnitude larger
> than sqrt(x),or the intercept. The advice is usually to (somewhat)
> standardize coordinates before using them as a trend. But I doubt this
> helps you very much.
>
> I find it hard to consider > 2 hours as a very long time before I know
> all the details (e.g. how many points were there within your
> maxdist?), the reason why you want an instant answer, and preferably
> have heard a comparison with other software. If you then tell me what
> your budget is, I might come up with possible solutions (starting very
> cheap, e.g. look at demo(snow) in package gstat, or use an OS that can
> assign this 4Gb to a single process).
> --
> Edzer
>>
>> Chris
>>
>> Edzer Pebesma wrote:
>>> Dave,
>>>
>>> 12000 observations fit, in the c representation, in less than 1 Mb
>>> (64 bytes per observation).
>>>
>>> The issue is that you think that passing maxdist to predict.gstat
>>> has an effect. It doesn't; you need to pass it to function gstat().
>>>
>>> The same thing happened in this
>>> https://stat.ethz.ch/pipermail/r-sig-geo/2008-September/004182.html
>>> message, where nmax was passed to predict.gstat, and simulation took
>>> forever. The other issue in that question was, I suspect, lack of
>>> standardization of coordinates, used in a trend surface.
>>> --
>>> Edzer
>>>
>>> Dave Depew wrote:
>>>> Is there a limit to the # of observations or size of file that can
>>>> be co-kriged in gstat?
>>>> I have a ~12000 observation data set (2 variables), the variograms,
>>>> cross variogram and lmc are fit well, and co-kriging starts ok
>>>>
>>>> Linear Model of Coregionalization found. Good.
>>>> [using ordinary cokriging]
>>>>
>>>> then immediately outputs
>>>>
>>>> "memory.c", line 57: can't allocate memory in function m_get()
>>>> Error in predict.gstat(fit.ck, newdata = EcoSAV.grid, maxdist = 100) :
>>>>  m_get
>>>>
>>>> Iv tried different maxdist from 10 to 1000, with exactly the same
>>>> result.
>>>> I recently upgraded my RAM to 4Gb and flipped the windows XP /3GB
>>>> switch.
>>>>
>>>>
>>>
>>
>


--
David Depew
PhD Candidate
Department of Biology
University of Waterloo
200 University Ave W.
Waterloo, ON. Canada
N2L 3G1

(T) 1-519-888-4567 x33895
(F) 1-519-746-0614

http://www.science.uwaterloo.ca/~ddepew

_______________________________________________
R-sig-Geo mailing list
R-sig-Geo@stat.math.ethz.ch
https://stat.ethz.ch/mailman/listinfo/r-sig-geo



        [[alternative HTML version deleted]]

_______________________________________________
R-sig-Geo mailing list
R-sig-Geo@stat.math.ethz.ch
https://stat.ethz.ch/mailman/listinfo/r-sig-geo

Reply via email to