On 04/04/2018 06:21 PM, Phani Motamarri wrote:
I have just written a plan vanilla interface to call the Kelly Error Estimate
function passing a parallel distributed vector as input vector type. But
strangely enough when I run on MPI on multiple procs in debug mode, I get the
following error:
Hi,
I have just written a plan vanilla interface to call the Kelly Error
Estimate function passing a parallel distributed vector as input vector
type. But strangely enough when I run on MPI on multiple procs in debug
mode, I get the following error:
---
Sir
The Issue 6158 and 6159 have been opened for the proposed functionalities.
Tell you if you have any
thought.
By instruction, it's actually more for the testing. Maybe I will just
follow how the other functions are tested.
By the way, I am at UC Berkeley leading an open source group project
Weixiong,
Yes. It then allows non-uniform slicing along z-axis.
Great. Yes, please do submit a pull request for this function! (And let
us know if you need help figuring out how to create pull requests! I
know we have a how-to file somewhere, but I forgot where exactly it is.
Someone?)
Sir,
Yes. It then allows non-uniform slicing along z-axis.
Also, I am wondering if you guys are interested in making triangulation
based on a basic one for repeated structure as shown in the attachment. I
have to do this for my project but didn't find a function so I did
something like copy->s
On 04/04/2018 02:40 PM, Weixiong Zheng wrote:
I found out there is not a function for extruding triangulation by exact
position of cell vertices along z-axis. So I wrote a piece of code by
modifying
|
GridGenerator::extrude_triangulation.
|
The attachment is an example of what it does.
Did f
Hi,
I found out there is not a function for extruding triangulation by exact
position of cell vertices along z-axis. So I wrote a piece of code by
modifying
GridGenerator::extrude_triangulation.
The attachment is an example of what it does.
Did for my own needs. Just wonder if I can contribute
I wish I can only deal with small problem, but that is only a test problem
and
I guess we need to use this code to compute much larger 3-D cases. Now it
is too slow so I cannot test any larger cases than 400k dofs. I tried 400k
dofs,
still, 2 cores are slower than 1 core.
The weirdest thing is t