TDB backup is safer way to back it up. It is not safe to copy the files
of a live database because you may copy different files at different
points in time.
You can build from a backup, and copy files, then start TDB.
Or even capture changes; build from a backup, and replay patches. That
is what rdf-delta (not part of Jena) does.
TDB2 also offers a possibility that might work for you.
It has a compaction step that produces a new, smaller database (TDB2
databases grow faster then TDB1). Like TDB1, you can't safely copy a
live database around. But the old database after compaction is not live.
You can sync twice to produce a small copy.
Andy
On 05/01/18 06:35, Dimov, Stefan wrote:
Thanks Lorenz,
That could work if the DB is already deployed.
How about initial deployment on multiple nodes? Or vice versa – copying it up
from live nodes to some backup/archive storage?
S.
On 12/22/17, 11:46 PM, "Lorenz Buehmann" <[email protected]>
wrote:
Probably I misunderstand the workflow, but isn't it more efficient to
push the changeset to all nodes instead of copying the whole TDB? I.e.
synchronizing the TDB instances on the triple level instead of files.
Cheers,
Lorenz
On 23.12.2017 01:55, Dimov, Stefan wrote:
> Yes, this is high av., deployment and backup.
>
> We have cloud system with multiple nodes and TDB needs to be copied
to/from these nodes. Copied from – when we need to back it up and copied to – when
we deploy it. So, the problems come when we copy from/to the nodes – it takes
time, operations sometimes fail, downtime - the nodes don’t work while we copy the
TDB (it’s a limitation that comes from Jena/TDB) and also sometimes the hardware
OS is throwing errors just because there are big files on it (without even copying
them).
>
> Before you say “you need a better hardware”, I agree, but that’s what we
have right now, so I was just wondering if somebody had similar problems and how
they resolved it (apart from replacing the hardware with faster, bigger and high
av. one ( )
>
> Regards,
> Stefan
>
>
> On 12/22/17, 4:24 PM, "ajs6f" <[email protected]> wrote:
>
> Can you tell us about the workflow that is managing them? You
mention copying and maintaining TDB files-- is that for high availability? Are you
writing updates to a master and replicating the resulting store, or some other
scenario?
>
> ajs6f
>
> > On Dec 22, 2017, at 7:19 PM, Dimov, Stefan <[email protected]>
wrote:
> >
> > Our TDB now is about 32G and I see that some of its files are
almost 5G (single file size), but let’s assume it can/will grow 3, 4, 5 times …
> >
> > On 12/22/17, 2:16 PM, "Dick Murray" <[email protected]> wrote:
> >
> > How big? How many?
> >
> > On 22 Dec 2017 8:37 pm, "Dimov, Stefan" <[email protected]>
wrote:
> >
> >> Hi all,
> >>
> >> We have a project, which we’re trying to productize and we’re
facing
> >> certain operational issues with big size files. Especially with
copying and
> >> maintaining them on the productive cloud hardware (application
nodes).
> >>
> >> Did anybody have similar issues? How did you resolve them?
> >>
> >> I will appreciate if someone shares their
experience/problems/solutions.
> >>
> >> Regards,
> >> Stefan
> >>
> >
> >
>
>
>
>