Bruno,
Thanks for taking a look at the batch mysql issue. We started with inodb
and moved to MyISAM.  MyISAM seemed marginally faster.   I have included
the output of show table status.

Without the batch my troubled backup is no longer troubled it's working
at a decent rate based on the number of small files it has.   If this
was the same size with big files the transfer rates would be faster etc.

22-Jun 06:58 backup-server JobId 70468: Sending spooled attrs to the
Director. Despooling 2,897,722,241 bytes ...
22-Jun 07:33 backup-server JobId 70468: Bacula qwest-netbackup-dir 3.0.2
(18Jun09): 22-Jun-2009 07:33:15
  Build OS:               x86_64-unknown-linux-gnu redhat Enterprise
release
  JobId:                  70468
  Job:                    backup-client1.2009-06-22_01.03.05_56
  Backup Level:           Full
  Client:                 "backup-client1" 3.0.1 (30Apr09)
x86_64-unknown-linux-gnu,redhat,Enterprise release
  FileSet:                "LINUX_ALL_LOCAL_FS" 2009-05-17 11:11:12
  Pool:                   "Daily" (From Job resource)
  Catalog:                "MyCatalog" (From Client resource)
  Storage:                "Autochanger" (From Job resource)
  Scheduled time:         22-Jun-2009 01:03:05
  Start time:             22-Jun-2009 01:03:20
  End time:               22-Jun-2009 07:33:15
  Elapsed time:           6 hours 29 mins 55 secs
  Priority:               20
  FD Files Written:       9,448,749
  SD Files Written:       9,448,749
  FD Bytes Written:       242,193,953,548 (242.1 GB)
  SD Bytes Written:       243,431,447,949 (243.4 GB)
  Rate:                   10352.4 KB/s
  Software Compression:   None
  VSS:                    no
  Encryption:             no
  Accurate:               yes
  Volume name(s):         CA3833L4|CA3831L4
  Volume Session Id:      213
  Volume Session Time:    1245547751
  Last Volume Bytes:      871,293,198,336 (871.2 GB)
  Non-fatal FD errors:    1
  SD Errors:              0
  FD termination status:  OK
  SD termination status:  OK
  Termination:            Backup OK -- with warnings

                                 -Jason

On Mon, 2009-06-22 at 09:02 +0200, Bruno Friedmann wrote:
> hi Jason, hardware should be more than sufficient.
> 
> Could you post your my.cnf ?
> 
> and tell what type of db is used (mysisam / inodb )
> 
> ps :  I'm going to work, so it could take time before I answer ..
> 
> Jason A. Kates wrote:
> > In my /etc/my.cnf  I have this set: 
> > tmpdir = /tmp
> > 
> > /tmp is a memory file system to avoid I/O contention.
> > 
> > When looking you can see that temp tables are create in /tmp
> > thus /var/tmp should be out of the loop.
> > 
> > The db resides in /export/mysql which is part of a 6 disk 15K rpm
> > hardware raid 10 disk set.
> > 
> > df -Th
> > Filesystem    Type    Size  Used Avail Use% Mounted on
> > /dev/cciss/c0d0p1
> >               ext3    996M  489M  456M  52% /
> > /dev/mapper/Volume00-vartest
> >               ext3    2.0G  368M  1.5G  20% /var
> > /dev/mapper/Volume00-export
> >               ext3    186G   47G  130G  27% /export
> > /dev/mapper/Volume00-usr
> >               ext3    3.9G  1.8G  2.0G  47% /usr
> > none         tmpfs    8.0G  8.0K  8.0G   1% /tmp
> > /dev/cciss/c1d0p1
> >               ext3    943G  316G  579G  36% /var/spool/bacula
> > 
> > 
> > We are running Redhat ES 5 with Kernel 2.6.18-128.el5.
> > If you would like to test a mount option I would be glad to give it a
> > test.  
> > 
> >                                  -Jason
> > 
> > On Mon, 2009-06-22 at 07:38 +0200, Bruno Friedmann wrote:
> >> Hi Jason, can you tell us if the place mysql works typically /var/lib/mysql
> >> are on the same partition of /tmp or /var/tmp ?
> >>
> >> I've found (in another project) with mysql myisam table you can kill disk 
> >> io
> >> if the two are on same disk. It's worst when you insert lot of data from 
> >> the same.
> >>
> >> In a perfect world after tunning carefully the /etc/my.cnf ( huge isn't 
> >> suffisant in many big case )
> >> you should have one full io for mysql temp files, one full io for indexes 
> >> one full io for tables
> >> AND one full io for import data ( completely separate /tmp for example ).
> >>
> >> In many installation, /tmp and /var are on the same device if you add the 
> >> fact of lvm or raid 1 configuration
> >> you get some io trouble.
> >>
> >> I wouldn't explain all fine tuning of mysql here, there's a lot and each 
> >> off them should be review & tested.
> >>
> >> Next there's also ext3 tuning for those who use it as many distro are very 
> >> conservative.
> >> (Newer kernel >2.6.30 would reduce the impact)
> >> atime,diratime,data=ordered etc ... could change life :-)
> >>
> >>
> >> First time bacula came with batch-insert enabled, I've test it, and have 
> >> really poor performances.
> >> So I disable it. ( lvm on top of soft raid 1 ) my mistake to not mention 
> >> it on user/devel list.
> >>
> >> In some future here we decide to migrate all possible application to 
> >> postgresql.
> >>
> >> For debugging and making some sort of test case, it would be nice to have 
> >> a "joke" dump database
> >> containing lot of dir & files.
> >> so we can test large queries ....
> >>
> >> Perhaps someone with very large records could do it using the md5() on 
> >> Filename and directory ?
> >>
> >>
> >>
> >> Jason A. Kates wrote:
> >>> Based on following the thread I recompile bacula with the following
> >>> options:
> >>> ./configure --enable-smartalloc --with-mysql
> >>> --with-tcp-wrappers=/usr/lib/libwrap.so --enable-static-fd
> >>> --enable-static-dir --enable-static-sd --disable-batch-insert
> >>> And my 18+ hour backup ran in 9 hours 22 mins 27 secs last night.   It
> >>> seems to be back to the pre 3.0 time.
> >>>
> >>> As for my tables I have an Admin job that optimizes my bacula tables
> >>> within mysql.  
> >>> Job {
> >>>         Name = optimize_table_JobMedia
> >>>         JobDefs = "DefaultJob"
> >>>         Type = Admin
> >>>         Client = bacula-local-fd
> >>>         RunBeforeJob = "/etc/bacula/optimize_table_JobMedia.sh"
> >>>         Priority = 40
> >>>         Cancel Queued Duplicates=yes
> >>>         Maximum Concurrent Jobs = 1
> >>>         Run Before Job = "/etc/bacula/check-duplicate-queued-job.sh '%n'
> >>> '%i'"
> >>> }
> >>>
> >>> Please let me know what else I can do to provide useful data for the
> >>> improvement of bacula (Worlds best backup software!).
> >>>                              Thanks -Jason
> >>>
> >>> On Sun, 2009-06-21 at 08:11 +0200, Marc Cousin wrote:
> >>>> First thing to know is which of the queries takes time (there are the 
> >>>> multiple 
> >>>> inserts into the batch table, then insert missing filenames/path, and 
> >>>> the 
> >>>> final insert). Or is it slow with all the queries ? Which storage engine 
> >>>> ?
> >>>>
> >>>> There are two possibilities :
> >>>>
> >>>> - either you have a bad plan for one of the queries (but I don't see 
> >>>> how, 
> >>>> mysql having a very simple optimizer).
> >>>> - or you have a tuning problem with mysql (there I can't really help you 
> >>>> : not 
> >>>> enough memory ? fragmented indexes ? I don't really know how mysql works)
> >>>>
> >>>> Kern, I have not been following closely the 3.0 development : do the 
> >>>> xattr go 
> >>>> into the lstat field ?
> >>>>
> >>>> On Sunday 21 June 2009 01:52:34 Jason A. Kates wrote:
> >>>>> I have a backup that runs for 3 hours then spends another 15 hours in
> >>>>> status "Dir inserting Attributes".
> >>>>>
> >>>>> The backup has 9,442,578 files and 239,718,461,832 bytes.   It was much
> >>>>> faster with bacula before and I thought that that the time change was
> >>>>> due to  additional data with xattrsupport that didn't exist before 3.0.
> >>>>> The job a spool job with spooled attributes too.
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>> We are using mysql-server-5.0.45-7.el5.   I would be glad to collect and
> >>>>> share data.  Please let me know what you sould like to me set or do to
> >>>>> get the appropriate data.
> >>>>>
> >>>>>                              Thanks -Jason
> >>>>>
> >>>>> On Sat, 2009-06-20 at 17:48 +0200, Marc Cousin wrote:
> >>>>>> There may be something. But first we have to know why mysql behaves 
> >>>>>> that
> >>>>>> slowly for Tom Summers. The problem being that I don't have a big mysql
> >>>>>> database to test on ... So as long as we don't have the query plan or 
> >>>>>> an
> >>>>>> explanation ...
> >>>>>>
> >>>>>> On Saturday 20 June 2009 14:36:44 Kern Sibbald wrote:
> >>>>>>> Hello Marc,
> >>>>>>>
> >>>>>>> Thanks for looking at this and for your responses.  If I understand 
> >>>>>>> you
> >>>>>>> correctly, there is nothing to be done, which does not surprise me :-)
> >>>>>>>
> >>>>>>> Best regards,
> >>>>>>>
> >>>>>>> Kern
> >>>>>>>
> >>>>>>> On Saturday 20 June 2009 13:16:44 Marc Cousin wrote:
> >>>>>>>>> Subject: Re: [Bacula-users] Performance with MySQL queries since
> >>>>>>>>> 3.0.0 (Dir inserting attributes hang)
> >>>>>>>>> Date: Saturday 20 June 2009
> >>>>>>>>> From: Kern Sibbald <[email protected]>
> >>>>>>>>> To: Tom Sommer <[email protected]>
> >>>>>>>>>
> >>>>>>>>> On Saturday 20 June 2009 08:51:53 Tom Sommer wrote:
> >>>>>>>>>> Tom Sommer wrote:
> >>>>>>>>>>> Mike Holden wrote:
> >>>>>>>>>>>> Jari Fredriksson wrote:
> >>>>>>>>>>>>>> INSERT INTO Filename( Name )
> >>>>>>>>>>>>>> SELECT a.Name
> >>>>>>>>>>>>>> FROM (
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> SELECT DISTINCT Name
> >>>>>>>>>>>>>> FROM batch
> >>>>>>>>>>>>>> ) AS a
> >>>>>>>>>>>>>> WHERE NOT
> >>>>>>>>>>>>>> EXISTS (
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> SELECT Name
> >>>>>>>>>>>>>> FROM Filename AS f
> >>>>>>>>>>>>>> WHERE f.Name = a.Name
> >>>>>>>>>>>>>> )
> >>>>>>>>>>>>> That looks silly.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> I would write it shorter as
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> INSERT INTO Filename(Name)
> >>>>>>>>>>>>> SELECT DISTINCT Name
> >>>>>>>>>>>>> FROM batch AS a
> >>>>>>>>>>>>> WHERE NOT EXISTS
> >>>>>>>>>>>>> (
> >>>>>>>>>>>>>     SELECT Name
> >>>>>>>>>>>>>     FROM Filename AS f
> >>>>>>>>>>>>>     WHERE f.Name = a.Name
> >>>>>>>>>>>>> )
> >>>>>>>>>>>> You may also want to consider using a JOIN rather than a
> >>>>>>>>>>>> subquery with a NOT EXISTS, something like (untested and
> >>>>>>>>>>>> unvalidated!):
> >>>>>>>>>>>>
> >>>>>>>>>>>> INSERT INTO filename(name)
> >>>>>>>>>>>> SELECT DISTINCT name
> >>>>>>>>>>>> FROM batch AS a
> >>>>>>>>>>>> LEFT JOIN filename AS f USING (name)
> >>>>>>>>>>>> WHERE f.name IS NULL
> >>>>>>>>>>>>
> >>>>>>>>>>>> I know from experience as an Oracle DBA (my day job) that this
> >>>>>>>>>>>> can often produce far more efficient results.
> >>>>>>>>>>>>
> >>>>>>>>>>>> Of course, all options need testing for both speed of
> >>>>>>>>>>>> execution and resource usage, bearing in mind that data varies
> >>>>>>>>>>>> from one installation to another, and one size may not fit
> >>>>>>>>>>>> all!
> >>>>>>>>>>> Good suggestions, sounds like there might be an overall
> >>>>>>>>>>> performance problem with the current query for batch-inserts
> >>>>>>>>>>> with lots of data. I'm a bit unsure if I dare test these
> >>>>>>>>>>> queries on my current installation.
> >>>>>>>>>>>
> >>>>>>>>>>> I'll CC Kern on the thread, perhaps he has some insights.
> >>>>>>>>>> So this morning I had to kill the above query because it's been
> >>>>>>>>>> running for 24+ hours, preventing the new daily jobs from
> >>>>>>>>>> running. I think I'm going to try and disable batch-inserts, the
> >>>>>>>>>> current situation is simply not good enough, it's become a major
> >>>>>>>>>> headache to run backups suddenly. When I hit the end of this
> >>>>>>>>>> month and ALL servers have to run FULL backups, I'm gonna be in a
> >>>>>>>>>> world of trouble I think - I just don't understand what has
> >>>>>>>>>> changed, because it's all been running great up until now.
> >>>>>>>>>>
> >>>>>>>>>> // Tom
> >>>>>>>>> Hello,
> >>>>>>>>>
> >>>>>>>>> We certainly can use help optimizing SQL since we are not DBAs, and
> >>>>>>>>> we will look into the SQL optimization suggestions given above,
> >>>>>>>>> keeping in mind that there are often rather radical differences in
> >>>>>>>>> timing of particular SQL queries depending on the database engine
> >>>>>>>>> used.
> >>>>>>>>>
> >>>>>>>>> To the best of my knowledge nothing has changed in terms of the
> >>>>>>>>> Batch insert queries since when it was implemented, and it is
> >>>>>>>>> *very* unlikely (though I haven't checked the code) that something
> >>>>>>>>> changed from 2.4.4 to 3.0.x.
> >>>>>>>>>
> >>>>>>>>> More likely, your workload or MySQL has changed in some way -- e.g.
> >>>>>>>>> more Jobs, more machines backed up, Director machine with less
> >>>>>>>>> memory or other jobs that use memory, a new version of MySQL, ...
> >>>>>>>>>
> >>>>>>>>> - I would suggest that you ensure that your database has all the
> >>>>>>>>> recommended indexes (see the make_mysql_tables file), and that you
> >>>>>>>>> are running with the large memory /etc/my.cnf file.
> >>>>>>>>>
> >>>>>>>>> - Another thing to do is to compact your database.  One way to do
> >>>>>>>>> it is to write it to an ASCII file and then re-insert it.
> >>>>>>>>>
> >>>>>>>>> - If you are running certain programs that create and delete lots
> >>>>>>>>> of temporary files with different names, you Filename table may
> >>>>>>>>> need cleaning.
> >>>>>>>>>
> >>>>>>>>> - I would strongly recommend not starting *lots* of Full backups at
> >>>>>>>>> the same time or on the same day.  By lots, I mean more than say 10
> >>>>>>>>> or 20 (depends on the size of your system).  It is generally far
> >>>>>>>>> better to stage 1/4 of the backup every week for a full backup so
> >>>>>>>>> that the peak workload is spread out over the month.
> >>>>>>>>>
> >>>>>>>>> - If the bottleneck is in MySQL, you might consider moving it to
> >>>>>>>>> another machine that has more memory and faster disks.
> >>>>>>>>>
> >>>>>>>>> - If you really have a huge number of backups (say 50 or more) that
> >>>>>>>>> all run at the same time, it might be advisable to consider using
> >>>>>>>>> PostgreSQL, but in that case, you will probably need an onsite DBA
> >>>>>>>>> to properly tune and maintain it.
> >>>>>>>>>
> >>>>>>>>> - Regardless of what hardware you have, there are certain
> >>>>>>>>> limitations on how many simultaneous jobs you can run (this kicks
> >>>>>>>>> in on many systems around 50). Once a certain number is exceeded,
> >>>>>>>>> the total throughput can rather radically decrease so careful
> >>>>>>>>> monitoring is necessary.  Bweb can help a lot in these situations.
> >>>>>>>>>
> >>>>>>>>> - If you are really "in a world of trouble" and it is a performance
> >>>>>>>>> issue, there is not much we (the Bacula project) can do for you
> >>>>>>>>> other than the above tips.  However, ($$) Bacula Systems has tools
> >>>>>>>>> that can help more precisely identify bottlenecks and help balance
> >>>>>>>>> loads.
> >>>>>>>>>
> >>>>>>>>> Regards,
> >>>>>>>>>
> >>>>>>>>> Kern
> >>>>>>>>>
> >>>>>>>>> -------------------------------------------------------
> >>>>>>>> Sorry Kern, I wasn't monitoring this thread (mysql in the title  :) )
> >>>>>>>>
> >>>>>>>> I wrote these queries, so maybe I should explain (not why I'm right,
> >>>>>>>> I'm not sure of that, but why we've done it this way).
> >>>>>>>>
> >>>>>>>> A few forewords :
> >>>>>>>> - batch insert was created at the begining for postgresql. I don't
> >>>>>>>> have any experience with mysql, except that knowing at the moment we
> >>>>>>>> programmed batch insert it didn't really do anything but nested
> >>>>>>>> loops... - The main reason for batch inserts with postgresql was to
> >>>>>>>> use COPY instead of INSERT, which is much faster (we also had good
> >>>>>>>> results with stored procedures, but batch insert was even faster)
> >>>>>>>> - our current file table is nearly 200GB (1 billion rows), with a
> >>>>>>>> 250GB database, and we insert around 50 million records a day (rough
> >>>>>>>> estimate)
> >>>>>>>>
> >>>>>>>> * First question I've seen in the thread:
> >>>>>>>>
> >>>>>>>> Why
> >>>>>>>>
> >>>>>>>> INSERT INTO Filename( Name )
> >>>>>>>> SELECT a.Name
> >>>>>>>> FROM (
> >>>>>>>>      SELECT DISTINCT Name
> >>>>>>>>      FROM batch
> >>>>>>>> ) AS a
> >>>>>>>> WHERE NOT EXISTS (
> >>>>>>>>      SELECT Name
> >>>>>>>>      FROM Filename AS f
> >>>>>>>>      WHERE f.Name = a.Name
> >>>>>>>> )
> >>>>>>>>
> >>>>>>>> instead of the much simpler
> >>>>>>>>
> >>>>>>>> INSERT INTO Filename(Name)
> >>>>>>>> SELECT DISTINCT Name
> >>>>>>>> FROM batch AS a
> >>>>>>>> WHERE NOT EXISTS
> >>>>>>>> (
> >>>>>>>>      SELECT Name
> >>>>>>>>      FROM Filename AS f
> >>>>>>>>      WHERE f.Name = a.Name
> >>>>>>>> )
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> => Because the first one removes the duplicate rows from batch table
> >>>>>>>> before checking they exist in the filename table, while the second
> >>>>>>>> one checks after. Performance was better in benchmarks with the first
> >>>>>>>> one (with PostgreSQL, with our data).
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> * Second question :
> >>>>>>>> Why not :
> >>>>>>>>
> >>>>>>>> INSERT INTO filename(name)
> >>>>>>>> SELECT DISTINCT name
> >>>>>>>> FROM batch AS a
> >>>>>>>> LEFT JOIN filename AS f USING (name)
> >>>>>>>> WHERE f.name IS NULL
> >>>>>>>>
> >>>>>>>> In fact, if the engine choses the right execution path, the execution
> >>>>>>>> plan will be the roughly the same (nested loop) as
> >>>>>>>>
> >>>>>>>> INSERT INTO Filename(Name)
> >>>>>>>> SELECT DISTINCT Name
> >>>>>>>> FROM batch AS a
> >>>>>>>> WHERE NOT EXISTS
> >>>>>>>> (
> >>>>>>>> SELECT Name
> >>>>>>>> FROM Filename AS f
> >>>>>>>> WHERE f.Name = a.Name
> >>>>>>>> )
> >>>>>>>>
> >>>>>>>> Except if batch is close in size to filename (for instance on my
> >>>>>>>> bacula database, it would be a 50 million files backup). Anyway, If i
> >>>>>>>> recall correctly about mysql (and it hasn't changed recently), mysql
> >>>>>>>> will do a nested loop for both, as it doesn't know how to merge or
> >>>>>>>> hash join ...
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> But I agree that this second syntax theorically gives more lattitude
> >>>>>>>> to the engine. Except that recent oracle and postgresql versions will
> >>>>>>>> anyway be able to convert between the exists/not exists and the join
> >>>>>>>> (or IN) in this case, as they are semantically equivalent.
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> Now, I don't know how mysql would react to all these queries. An
> >>>>>>>> explain plan would be great to understand... but I would have thought
> >>>>>>>> that the NOT EXIST query would have been great for it's optimizer, as
> >>>>>>>> it tells mysql exactly what to do :)
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> A last note, for postgreSQL users on this mailing list :
> >>>>>>>> You can try to get better performance with batch inserts by lowering
> >>>>>>>> seq_page_cost and random_page_cost, so that the engine favours nested
> >>>>>>>> loop instead of hash join for the INSERT INTO File query : most of
> >>>>>>>> the time, most of filename and path will be in the database or
> >>>>>>>> filesystem cache, and the plan will be much faster (and save you big
> >>>>>>>> sorts)
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> Please tell me if you need more information.
> >>>>>>>>
> >>
> > 
> 
> 
> -- 
> 
>      Bruno Friedmann
> 
> 
> ------------------------------------------------------------------------------
> Are you an open source citizen? Join us for the Open Source Bridge conference!
> Portland, OR, June 17-19. Two days of sessions, one day of unconference: $250.
> Need another reason to go? 24-hour hacker lounge. Register today!
> http://ad.doubleclick.net/clk;215844324;13503038;v?http://opensourcebridge.org
> _______________________________________________
> Bacula-devel mailing list
> [email protected]
> https://lists.sourceforge.net/lists/listinfo/bacula-devel

-- 
----------------------------------------------------------------------------
Jason A. Kates ([email protected]) 
Fax:    208-975-1514
Phone:  212-400-1670 x2
============================================================================

[client]
port            = 3306
socket          = /var/lib/mysql/mysql.sock

[mysqld]

port            = 3306
socket          = /var/lib/mysql/mysql.sock
back_log = 50
max_connections = 100
max_connect_errors = 10
table_cache = 6144
max_allowed_packet = 16M
#binlog_cache_size = 1M
max_heap_table_size = 2048M
sort_buffer_size = 2048M
join_buffer_size = 2048M
thread_cache_size = 8
thread_concurrency = 8
query_cache_size = 64M
query_cache_limit = 2M
ft_min_word_len = 4
default_table_type = InnoDB
default-storage-engine = INNOBASE
thread_stack = 192K
transaction_isolation = REPEATABLE-READ
tmp_table_size = 3072M
#log-bin=mysql-bin
log_slow_queries=slow.log
long_query_time = 2
log_long_format
server-id = 1
key_buffer_size = 2048M
read_buffer_size = 2M
read_rnd_buffer_size = 16M
bulk_insert_buffer_size = 64M
myisam_sort_buffer_size = 128M
myisam_max_sort_file_size = 10G
myisam_max_extra_sort_file_size = 10G
myisam_repair_threads = 1
myisam_recover
skip-bdb
skip-innodb_adaptive_hash_index
innodb_additional_mem_pool_size = 16M
innodb_buffer_pool_size = 2G
#innodb_data_file_path = ibdata1:10M:autoextend
innodb_file_io_threads = 4
innodb_thread_concurrency = 16
innodb_flush_log_at_trx_commit = 1
innodb_log_buffer_size = 8M
innodb_log_file_size = 256M
innodb_log_files_in_group = 3
innodb_max_dirty_pages_pct = 90
innodb_lock_wait_timeout = 86400
innodb_file_per_table = 1
wait_timeout = 86400
set-variable = max_allowed_packet=32M
#datadir=/var/spool/bacula/mysql
datadir=/export/mysql
tmpdir = /tmp

[mysql.server]
user=mysql
#basedir=/var/lib
#basedir=/var/spool/bacula/mysql
basedir=/export/mysql
tmpdir = /tmp


[mysqldump]
quick
max_allowed_packet = 32M

[mysql]
no-auto-rehash


[isamchk]
key_buffer = 1024M
sort_buffer_size = 1024M
read_buffer = 8M
write_buffer = 8M

[myisamchk]
key_buffer = 2048M
sort_buffer_size = 2048M
read_buffer = 8M
write_buffer = 8M

[mysqlhotcopy]
interactive-timeout

[mysqld_safe]
#open-files-limit = 8192
open-files-limit = 65536
err-log=/var/log/mysqld.log
pid-file=/var/run/mysqld/mysqld.pid
Enter SQL query: show table status;
+--------------+--------+---------+------------+-------------+----------------+----------------+------------------------+---------------+---------------+----------------+---------------------+---------------------+---------------------+-------------------+----------+----------------------------------------+---------+
| Name         | Engine | Version | Row_format | Rows        | Avg_row_length | 
Data_length    | Max_data_length        | Index_length  | Data_free     | 
Auto_increment | Create_time         | Update_time         | Check_time         
 | Collation         | Checksum | Create_options                         | 
Comment |
+--------------+--------+---------+------------+-------------+----------------+----------------+------------------------+---------------+---------------+----------------+---------------------+---------------------+---------------------+-------------------+----------+----------------------------------------+---------+
| BaseFiles    | MyISAM |      10 | Fixed      |           0 |              0 | 
             0 |  7,036,874,417,766,399 |         1,024 |             0 |       
       1 | 2009-04-16 11:20:42 | 2009-04-16 11:20:42 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| CDImages     | MyISAM |      10 | Fixed      |           0 |              0 | 
             0 |  3,659,174,697,238,527 |         1,024 |             0 | NULL  
         | 2009-04-16 11:21:01 | 2009-04-16 11:21:01 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| Client       | MyISAM |      10 | Dynamic    |         128 |            100 | 
        12,852 |    281,474,976,710,655 |         8,192 |             0 |       
     199 | 2009-04-16 11:21:06 | 2009-06-22 08:11:20 | 2009-06-22 08:11:20 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Counters     | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 | NULL  
         | 2009-04-16 11:21:13 | 2009-04-16 11:21:13 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| Device       | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 |       
       1 | 2009-04-16 11:21:21 | 2009-04-16 11:21:21 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| File         | MyISAM |      10 | Dynamic    | 216,041,487 |            104 | 
23,991,885,416 |      1,099,511,627,775 | 9,863,255,040 | 1,504,156,820 |  
4,379,892,030 | 2009-06-21 18:21:05 | 2009-06-22 08:11:14 | 2009-06-21 19:38:47 
| latin1_swedish_ci | NULL     | max_rows=2115098112 avg_row_length=111 |       
  |
| FileSet      | MyISAM |      10 | Dynamic    |          29 |             57 | 
         1,664 |    281,474,976,710,655 |         2,048 |             0 |       
     109 | 2009-04-16 11:21:31 | 2009-06-21 17:06:17 | 2009-06-21 17:06:17 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Filename     | MyISAM |      10 | Dynamic    |   3,403,931 |             35 | 
   120,082,276 |    281,474,976,710,655 |   118,210,560 |             0 |      
7,527,514 | 2009-04-16 10:38:18 | 2009-06-22 08:11:14 | 2009-06-21 17:06:46 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Job          | MyISAM |      10 | Dynamic    |       6,438 |            154 | 
       996,696 |    281,474,976,710,655 |       116,736 |             0 |       
  70,469 | 2009-04-16 10:37:01 | 2009-06-22 08:11:20 | 2009-06-21 17:06:46 | 
latin1_swedish_ci | NULL     |                                        |         
|
| JobHisto     | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 | NULL  
         | 2009-04-16 11:21:47 | 2009-04-16 11:21:47 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| JobMedia     | MyISAM |      10 | Fixed      |  19,717,394 |             50 | 
   985,869,700 | 14,073,748,835,532,799 |   482,861,056 |             0 |     
59,485,362 | 2009-05-26 13:28:29 | 2009-06-22 08:10:54 | 2009-06-21 17:08:25 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Location     | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 |       
       1 | 2009-04-16 11:20:49 | 2009-04-16 11:20:49 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| LocationLog  | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 |       
       1 | 2009-04-16 11:22:03 | 2009-04-16 11:22:03 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| Log          | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 |       
       1 | 2009-04-16 16:01:52 | 2009-04-16 16:01:52 | 2009-05-26 19:35:44 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Media        | MyISAM |      10 | Dynamic    |         248 |            123 | 
        30,564 |    281,474,976,710,655 |        13,312 |             0 |       
     270 | 2009-04-16 11:22:13 | 2009-06-22 08:10:57 | 2009-06-21 17:08:25 | 
latin1_swedish_ci | NULL     |                                        |         
|
| MediaType    | MyISAM |      10 | Dynamic    |           1 |             20 | 
            20 |    281,474,976,710,655 |         2,048 |             0 |       
       2 | 2009-04-16 11:22:21 | 2009-04-16 11:22:21 | 2009-05-26 19:35:44 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Path         | MyISAM |      10 | Dynamic    |   2,004,911 |             89 | 
   178,885,860 |    281,474,976,710,655 |   142,074,880 |             0 |      
3,633,310 | 2009-04-16 09:59:26 | 2009-06-22 07:30:02 | 2009-06-21 17:08:53 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Pool         | MyISAM |      10 | Dynamic    |           3 |             42 | 
           128 |    281,474,976,710,655 |         3,072 |             0 |       
       4 | 2009-04-16 11:20:29 | 2009-06-22 02:22:35 | 2009-06-21 17:08:53 | 
latin1_swedish_ci | NULL     |                                        |         
|
| Status       | MyISAM |      10 | Dynamic    |           0 |              0 | 
             0 |    281,474,976,710,655 |         1,024 |             0 | NULL  
         | 2009-04-16 11:22:28 | 2009-04-16 11:22:28 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| Storage      | MyISAM |      10 | Dynamic    |           2 |             22 | 
            44 |    281,474,976,710,655 |         2,048 |             0 |       
       7 | 2009-04-16 11:22:36 | 2009-04-27 00:03:55 | 2009-05-26 19:35:44 | 
latin1_swedish_ci | NULL     |                                        |         
|
| UnsavedFiles | MyISAM |      10 | Fixed      |           0 |              0 | 
             0 |  4,785,074,604,081,151 |         1,024 |             0 |       
       1 | 2009-04-16 11:22:42 | 2009-04-16 11:22:42 | NULL                | 
latin1_swedish_ci | NULL     |                                        |         
|
| Version      | MyISAM |      10 | Fixed      |           1 |              7 | 
             7 |  1,970,324,836,974,591 |         1,024 |             0 | NULL  
         | 2009-04-16 11:20:34 | 2009-04-16 11:20:34 | 2009-05-26 19:35:44 | 
latin1_swedish_ci | NULL     |                                        |         
|
+--------------+--------+---------+------------+-------------+----------------+----------------+------------------------+---------------+---------------+----------------+---------------------+---------------------+---------------------+-------------------+----------+----------------------------------------+---------+

------------------------------------------------------------------------------
Are you an open source citizen? Join us for the Open Source Bridge conference!
Portland, OR, June 17-19. Two days of sessions, one day of unconference: $250.
Need another reason to go? 24-hour hacker lounge. Register today!
http://ad.doubleclick.net/clk;215844324;13503038;v?http://opensourcebridge.org
_______________________________________________
Bacula-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-devel

Reply via email to