Duncan Murdoch wrote:
On 11/12/2008 6:04 PM, Terry Therneau wrote:
I'm making the move of the survival package from my own environment to,
and have stumbled into a vacuum. The R Extensions manual has really
nice
instructions about how to lay out the directories, order the files, and
run tests for DISTRIBUTION of a product, but I can't find anything on how
to set up a reasonable DEVELOPMENT environment.
In my local world, I had the .c and .s files in a common directory,
with
a Makefile that I had created, and the test suite in a subdirectory.
Debugging
and development was quite nice.
make
cd test
R
attach("..")
try something and perhaps it fails
q()
cd ..
Fix and repeat. The Makefile took some time to create, but paid for
itself a
hundred times over.
So, I've now rearranged everything into standard R order. Then I
did the
only thing I could find
R CMD INSTALL ~/myRlib survival where "survival" is said
directory. This turns out to be not useful at all.
The survival package is large, and I rather suspected that I would
goof something up, and I did, resulting in the following message
Error in parse(n = -1, file = file) : unexpected end of input at
14450: }
14451:
It is not exactly obvious which of the 132 files in my R/ directory is
the culprit here.
That's something I would like to fix too. There are (at least) two
possible ways: stop concatenating the files (which is not really needed
nowadays, most packages install saved images), or add some markup to the
concatenated file so that the parser can report on the original filename
and line number (like the #LINE directives output by the C preprocessor).
I would certainly appreciate a fix for this. When this has happened to
me, I usually end up just sourcing the individual files into an R
session until I find the bad file, and get a report with the right line
number. It always seems like a lot of work for something that should be
trivial.
In general:
1. The library is large, and recompiling/reparsing everything is very
far from
instantaneous. It is not the edit/load cycle I desire.
If you install from the directory, the compiling should only be done
once (unless you change a file, of course). (The alternative is
installing from the tarball, which is recommended for later stages of
testing before distribution, because it's possible something could go
wrong in building the tarball. But it won't include any object files,
so you'll recompile every time.)
You can also use option "--docs=none" to skip building the help system;
this will save a bit of time.
2. I take testing seriously: the test suite takes on the order of 15
minutes to
run on a fast machine. I most certainly don't want to run it in the
mid cycle.
I don't quite follow this. If you want to run all your tests, you would
use R CMD check. If you only want to run some of them, then you can
source things out of the tests directory while running interactively.
Someone must have tackled this. I'm hoping that there is some
documentation
that I have managed to overlook which discussess a good setup for this
middle
ground between concieving of a library and packaging it for delivery; the
"build, test, make sure it acually works" part of development.
In my experience there are two differ sorts of problems that I think
could benefit from some improvement. The first is that there should be a
standard way to have extra tests, that do not get run in the normal CRAN
testing cycle, or by developers when using a "quick" R CMD check, but
can be run with a standard mechanism. I do this by putting the tests in
a separate package, and I have seen reports of different mechanisms, but
I think they are all somewhat ad hoc. Currently, if you put too much
testing in your package then all the testing gets omitted on some CRAN
testing platforms. Just a common directory like extraTests/ would be a
good start.
The second problem is that a developer usually needs to run tests/ when
code in R/ has been changed, but probably not run tests/ when the
changes are only in man/, demos/, or inst/doc/. The checking that needs
to be done in those cases is reduced from the full R CMD check. A
common way to attack this is with a good Makefile. I wrote an article in
R News a few years ago about my attempts to do this, and my Makefiles
are available, but there is some customization necessary, and there is
lots of room for improvement. It does (mostly) work with make -j, so
days of testing on a single processor machine can be accomplished in a
few hours on multicore machines (for me, mileage may vary). I have not
addressed the idea of trying to specialize files in tests/ to specific
code files in R/. (I think others have tried to do this with a "unit
testing" approach.)
Paul Gilbert
I find the process I follow is to organize the files in the distribution
structure from the beginning. When adding new functions, I'll
generally use source() a few times to get the syntax right, and perhaps
run simple tests. (But remember, if you use a NAMESPACE, the functions
may not behave the same when they're sourced into the global
environment.) In the early stages, I'll do a lot of installs of the
packages.
If I was porting a big package and wanted to find syntax errors, to work
around the not-very-helpful error message you saw I'd do something like
for (f in list.files("pkg/R", full=TRUE)) source(f)
which will report the error more informatively.
Duncan Murdoch
______________________________________________
R-devel@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-devel
====================================================================================
La version française suit le texte anglais.
------------------------------------------------------------------------------------
This email may contain privileged and/or confidential in...{{dropped:26}}
______________________________________________
R-devel@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-devel