That makes more sense. I've set it to 10 and inparallel to 2 and we'll see how it goes tonight.
On Mon, Nov 26, 2018 at 2:01 PM Cuttler, Brian R (HEALTH) <brian.cutt...@health.ny.gov> wrote: > > I believe maxdumps is max concurrent dumps across all clients. > > You might have 10 clients each with an inparallel of 2, giving 20 possible > concurrent dumps, but because of server limitations you might set maxdumps to > something between 2 and 20. > > -----Original Message----- > From: Chris Nighswonger <cnighswon...@foundations.edu> > Sent: Monday, November 26, 2018 1:57 PM > To: Cuttler, Brian R (HEALTH) <brian.cutt...@health.ny.gov> > Cc: amanda-users@amanda.org > Subject: Re: Another dumper question > > ATTENTION: This email came from an external source. Do not open attachments > or click on links from unknown senders or unexpected emails. > > > inparallel 10 > > maxdumps not listed, so I'm assuming the default of 1 is being observed. > > I'm not sure that the maxdumps parameter would affect dumping DLEs from > multiple clients in parallel, though. The manpage states, "The maximum number > of backups from a single host that Amanda will attempt to run in parallel." > That seems to indicate that this parameter controls parallel dumps of DLEs on > a single client. > > Kind regards, > Chris > On Mon, Nov 26, 2018 at 1:50 PM Cuttler, Brian R (HEALTH) > <brian.cutt...@health.ny.gov> wrote: > > > > Did you check your maxdumps and inparallel parameters? > > > > -----Original Message----- > > From: owner-amanda-us...@amanda.org <owner-amanda-us...@amanda.org> On > > Behalf Of Chris Nighswonger > > Sent: Monday, November 26, 2018 1:34 PM > > To: amanda-users@amanda.org > > Subject: Another dumper question > > > > ATTENTION: This email came from an external source. Do not open attachments > > or click on links from unknown senders or unexpected emails. > > > > > > So in one particular configuration I have the following lines: > > > > inparallel 10 > > dumporder "STSTSTSTST" > > > > I would assume that that amanda would spawn 10 dumpers in parallel and > > execute them giving priority to largest size and largest time alternating. > > I would assume that amanda would do some sort of sorting of the DLEs based > > on size and time, set them in descending order, and the run the first 10 > > based on the list thereby utilizing all 10 permitted dumpers in parallel. > > > > However, based on the amstatus excerpt below, it looks like amanda simply > > starts with the largest size and runs the DLEs one at a time, not making > > efficient use of parallel dumpers at all. This has the unhappy results at > > times of causing amdump to be running when the next backup is executed. > > > > I have changed the dumporder to STSTStstst for tonight's run to see if that > > makes any difference. But I don't have much hope it will. > > > > Any thoughts? > > > > Kind regards, > > Chris > > > > > > > > > > From Mon Nov 26 01:00:01 EST 2018 > > > > 1 4054117k waiting for dumping > > 1 6671k waiting for dumping > > 1 222k waiting for dumping > > 1 2568k waiting for dumping > > 1 6846k waiting for dumping > > 1 125447k waiting for dumping > > 1 91372k waiting for dumping > > 1 92k waiting for dumping > > 1 32k waiting for dumping > > 1 32k waiting for dumping > > 1 32k waiting for dumping > > 1 32k waiting for dumping > > 1 290840k waiting for dumping > > 1 76601k waiting for dumping > > 1 86k waiting for dumping > > 1 71414k waiting for dumping > > 0 44184811k waiting for dumping > > 1 281k waiting for dumping > > 1 6981k waiting for dumping > > 1 50k waiting for dumping > > 1 86968k waiting for dumping > > 1 81649k waiting for dumping > > 1 359952k waiting for dumping > > 0 198961004k dumping 159842848k ( 80.34%) (7:23:39) > > 1 73966k waiting for dumping > > 1 821398k waiting for dumping > > 1 674198k waiting for dumping > > 0 233106841k dump done (7:23:37), waiting for writing to tape > > 1 32k waiting for dumping > > 1 32k waiting for dumping > > 1 166876k waiting for dumping > > 1 32k waiting for dumping > > 1 170895k waiting for dumping > > 1 162817k waiting for dumping > > 0 failed: planner: [Request to client failed: Connection timed out] > > 1 32k waiting for dumping > > 1 32k waiting for dumping > > 0 53k waiting for dumping > > 0 77134628k waiting for dumping > > 1 2911k waiting for dumping > > 1 36k waiting for dumping > > 1 32k waiting for dumping > > 1 84935k waiting for dumping > > > > SUMMARY part real estimated > > size size > > partition : 43 > > estimated : 42 559069311k > > flush : 0 0k > > failed : 1 0k ( 0.00%) > > wait for dumping: 40 128740001k ( 23.03%) > > dumping to tape : 0 0k ( 0.00%) > > dumping : 1 159842848k 198961004k ( 80.34%) ( 28.59%) > > dumped : 1 233106841k 231368306k (100.75%) ( 41.70%) > > wait for writing: 1 233106841k 231368306k (100.75%) ( 41.70%) > > wait to flush : 0 0k 0k (100.00%) ( 0.00%) > > writing to tape : 0 0k 0k ( 0.00%) ( 0.00%) > > failed to tape : 0 0k 0k ( 0.00%) ( 0.00%) > > taped : 0 0k 0k ( 0.00%) ( 0.00%) > > 9 dumpers idle : 0 > > taper status: Idle > > taper qlen: 1 > > network free kps: 0 > > holding space : 436635431k ( 50.26%) > > chunker0 busy : 6:17:03 ( 98.28%) > > dumper0 busy : 6:17:03 ( 98.28%) > > 0 dumpers busy : 0:06:34 ( 1.72%) 0: 0:06:34 > > (100.00%) > > 1 dumper busy : 6:17:03 ( 98.28%) 0: 6:17:03 > > (100.00%)