Bug#946958: sa-compile failing on Graylisting.pm

2020-01-02 Thread Shannon Dealy

On Wed, 18 Dec 2019, Noah Meyerhans wrote:

[snip]

The problem is likely related to the fixes for CVE-2018-11805, which
involved malicious rulesets invoking arbitrary commands as the uid
running spamassassin/spamd.  In the case of sa-exim, the line triggering
the taint failure is performing an "eval" operation of configuration
data read directly from a .cf file, so changing spamasassin's behavior
is probably not ideal.

I've tested a backport of sa-exim 4.2.1-16 from stretch to jessie, and
have observed that the problem does not occur in this scenario.  So an
update of sa-exim in jessie might be the least disruptive path to a fix
here.  In the mean time, you might consider locally building it.


Hi Noah,

I tried backporting sa-exim 4.2.1-16 to jessie. While it installs and fixes 
the installation problem with sa-compile, for some reason it broke grey 
listing on my system. Basically, every email that gets grey listed will 
continually get temporary rejection until the other server gives up. It 
appears that no email once it is greylisted will ever get through.


I'm not sure if this is specific to my setup due to a quirk of my 
configuration settings and a minor incompatibility between the stretch 
and jessie versions of sa-exim or something more serious. Log files did not 
reveal anything obvious nor did a quick review of the differences between the 
sources of the two versions.


For now I have just disabled greylisting on my server, unfortunately I don't 
have time right now for a more in-depth analysis as I would have to re-learn 
how all of this works.


FWIW.

Shannon C. Dealy   |   DeaTech Research Inc.
de...@deatech.com  | Biotechnology Development Services
Telephone USA: +1 541-929-4089 |  USA and the Netherlands
Netherlands:   +31 85 208 5570 |  www.deatech.com



Bug#928121: Same problem on Lenovo T480

2019-11-04 Thread Shannon Dealy



I'm having what appears to be the same problem with a few differences of note:

 - Lenovo T480 (original report was for a T580)
 - Linux kernel 4.19.0-6-amd64
 - lxdm
 - xfce4
 - External monitor connected via HDMI

Shannon C. Dealy   |   DeaTech Research Inc.
de...@deatech.com  | Biotechnology Development Services
Telephone USA: +1 541-929-4089 |  USA and the Netherlands
Netherlands:   +31 85 208 5570 |  www.deatech.com



Bug#932101: release-notes: S3QL unable to resolve hostname due to S3 URL format change

2019-07-14 Thread Shannon Dealy
Package: release-notes
Severity: normal

Dear Maintainer,

After upgrading from Stretch to Buster, I was unable to use S3QL with any of
my Amazon S3 buckets. All attempts to access gave errors like:

 ERROR: Can't connect to backend: unable to resolve hostname

I was eventually able to track the problem down to a change in the format
of the Amazon S3 URL which occurred two years ago. The new format of the
URL is:

 s3:

Note the addition of a "region" section in the URL.

Unfortunately this change is not displayed by "apt-listchanges" during the
upgrade, nor is it documented in the release notes.

Please add this information to the release notes for Buster so that others 
don't waste hours trying to figure it out.


-- System Information:
Debian Release: 10.0
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'oldstable-updates'), (500, 'oldstable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 4.19.0-5-amd64 (SMP w/8 CPU cores)
Kernel taint flags: TAINT_OOT_MODULE, TAINT_UNSIGNED_MODULE
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8), 
LANGUAGE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: systemd (via /run/systemd/system)
LSM: AppArmor: enabled



Bug#930477: alltray disrupts GUI for application program it docks

2019-06-13 Thread Shannon Dealy
Package: alltray
Version: 0.71b-1+b2
Severity: important

Dear Maintainer,

When running the command:

   alltray zotero

the zotero program is opened, successfully docks and appears normal, however,
when you go to the system tray and display the zotero application window,
dragging and dropping items between collections (folders) is broken.  If 
zotero is launched outside of alltray, or launched instead using kdocker 
(the KDE equivalent of alltray), zotero has no problems.

NOTES:
  - When attempting to drag and drop in zotero, the folder name you are
dragging to becomes highlighted once you are aligned sufficiently
to allow dropping. When run under alltray, this highlighting never
occurs no matter where you position the item to be dropped, and
dropping the item doesn't work.

  - In zotero, this drag/drop mechanism is the only method which can be
used to copy or move items between the collection folders, so this
completely breaks copy/move functionality. I have not noticed any other
GUI issues when running zotero under alltray

  - Zotero version 5.0.66

  - For people seeking a workaround, when using kdocker (at least in debian
stretch), you need to use the "-n" option (which is not listed on the
kdocker man page):  kdocker -n Zotero zotero


-- System Information:
Debian Release: 9.9
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'stable-updates')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 4.9.0-8-amd64 (SMP w/8 CPU cores)
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8), 
LANGUAGE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: sysvinit (via /sbin/init)

Versions of packages alltray depends on:
ii  gconf-service3.2.6-4+b1
ii  libatk1.0-0  2.22.0-1
ii  libc62.24-11+deb9u4
ii  libcairo21.14.8-1
ii  libfontconfig1   2.11.0-6.7+b1
ii  libfreetype6 2.6.3-3.2
ii  libgconf-2-4 3.2.6-4+b1
ii  libgdk-pixbuf2.0-0   2.36.5-2+deb9u2
ii  libglib2.0-0 2.50.3-2
ii  libgtk2.0-0  2.24.31-2
ii  libpango-1.0-0   1.40.5-1
ii  libpangocairo-1.0-0  1.40.5-1
ii  libpangoft2-1.0-01.40.5-1
ii  libx11-6 2:1.6.4-3+deb9u1

alltray recommends no packages.

alltray suggests no packages.

-- debconf-show failed



Bug#882792: Source of problem with missing icons

2018-09-04 Thread Shannon Dealy



I found this bug report for the status notifier plugin:

   
https://bugs.launchpad.net/ubuntu/+source/xfce4-statusnotifier-plugin/+bug/1756627

Which had the following note:

   "This happens if you have the package "ayatana-indicator-application"
   installed. Note that although it appears to be stealing the indicators,
   they don't show up in indicator-plugin instead."

based on this I killed the process running the:

   "ayatana-indicator-application-service"

with the following results:

   - As soon as the above process was killed, the Remmina icon appeared in the
 system tray.

   - The VLC icon (though it was running) did not appear after killing the
 above process, however, exiting and restarting VLC caused its icon to
 appear in the system tray as well.

The difference in behavior of the two programs may be due to their using 
different frameworks (QT vs GTK).


So it appears (at least for my case) this problem is due to the
status notifier / ayatana-indicator-application problem in the above Ubuntu 
bug report.




Shannon C. Dealy   |   DeaTech Research Inc.
de...@deatech.com  | Biotechnology Development Services
Telephone USA: +1 541-929-4089 |  USA and the Netherlands
Netherlands:   +31 85 208 5570 |  www.deatech.com



Bug#882792: xfce4-panel: some icons not appearing in notification panel

2018-06-26 Thread Shannon Dealy
Package: xfce4-panel
Version: 4.12.1-2
Followup-For: Bug #882792

Dear Maintainer,

I am experiencing this problem as well. It started after I logged out and
rebooted the system for the first time in many weeks (I'm not sure exactly
how long). During that time many packages were installed / removed / upgraded,
though I am not sure if any of them are relevant to the problem.

Specifics:

  The following no longer display in the notification area, though they did
  before the reboot:

 vlc 3.0.2-0+deb
 remmina 1.2.30.1+dfsg-1~bpo9+1  

  The following all continue to work fine with the notification area:

 network-manager-gnome (nm-applet) 1.4.4-1
 xfce4-clipman 2:1.4.1-1
 hamster-applet 2.91.3+git2
 alltray 0.71b-1+b2

Worthy of note, vlc is QT based and remmina is GTK based, so it appears the
framework used may not matter.

I have tried rebooting, reinstalling recently deleted packages, deleting
recently installed packages, so far, nothing I can think of works. I didn't
bother with some of the things that had been tried above.

etckeeper is installed, and in reviewing the git log of package changes, I
didn't see any obvious issues, though perhaps someone with a better knowledge
of the notification area code might see something (a relevant library?), but
the section of the log that might apply (after stripping extraneous commits)
is about 800 lines long.


-- System Information:
Debian Release: 9.4
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'stable-updates')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 4.9.0-6-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8), 
LANGUAGE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: sysvinit (via /sbin/init)

Versions of packages xfce4-panel depends on:
ii  exo-utils0.10.7-1
ii  libatk1.0-0  2.22.0-1
ii  libc62.24-11+deb9u3
ii  libcairo21.14.8-1
ii  libdbus-1-3  1.10.26-0+deb9u1
ii  libdbus-glib-1-2 0.108-2
ii  libexo-1-0   0.10.7-1
ii  libfontconfig1   2.11.0-6.7+b1
ii  libfreetype6 2.6.3-3.2
ii  libgarcon-1-00.4.0-2
ii  libgdk-pixbuf2.0-0   2.36.5-2+deb9u2
ii  libglib2.0-0 2.50.3-2
ii  libgtk2.0-0  2.24.31-2
ii  libice6  2:1.0.9-2
ii  libpango-1.0-0   1.40.5-1
ii  libpangocairo-1.0-0  1.40.5-1
ii  libpangoft2-1.0-01.40.5-1
ii  libsm6   2:1.2.2-1+b3
ii  libwnck222.30.7-5.1
ii  libx11-6 2:1.6.4-3
ii  libxext6 2:1.3.3-1+b2
ii  libxfce4ui-1-0   4.12.1-2
ii  libxfce4util74.12.1-3
ii  libxfconf-0-24.12.1-1

xfce4-panel recommends no packages.

xfce4-panel suggests no packages.

-- debconf-show failed



Bug#896005: synergy: Changing mouse xinput coordinate transform matrix breaks synergy

2018-04-18 Thread Shannon Dealy
Package: synergy
Version: 1.8.8-stable+dfsg.1-1
Severity: important

Dear Maintainer,

Issuing the following command either before or after starting synergys:

   xinput --set-prop  "Coordinate Transformation Matrix" 3 0 0 0 3 0 
0 0 1

   Where  is the ID given by "xinput --list" for my trackpoint 
   mouse device.

Results in the following problem(s):

With the synergy client computer configured on the left side of my synergy
server, synergy does one of the two following behaviors when sliding the mouse
off the left side of the server screen:

  1 - The mouse immediately wraps around to the right side of my server display
  rather than showing up on the client.

  2 - The mouse is captured by the client display, but cannot be moved on the 
  client, however, when in this state, keyboard input from the server is
  correctly routed to the client machine.

Restoring the matrix to its default values on the server:

  xinput --set-prop  "Coordinate Transformation Matrix" 1 0 0 0 1 0 0 
0 1

fixes the problem.

I also tested this with version 1.4.18 of synergy installed on both computers
and experienced the same results.

I flagged this as important since I am left with a choice of usable mouse
performance (xinput scaling for proper acceleration on my trackpoint can
only be achieved by use of the transform matrix) or working synergy. 
Additionally, the transform matrix is modified by some systems in order to get
proper orientation of mouse output when using rotated screens. I tested
a 90 degree rotation using the following matrix values:

0 -1 1 1 0 0 0 0 1

and synergy again broke. This problem makes synergy completely unusable in
some working environments.

Client machine used for this is running Ubuntu 16.04.4 LTS, xenial, with the
Synergy 1.8.8 package installed from the Ubuntu "artful" archive.



-- System Information:
Debian Release: 9.4
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'stable-updates')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 4.9.0-6-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8), 
LANGUAGE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: sysvinit (via /sbin/init)

Versions of packages synergy depends on:
ii  libavahi-compat-libdnssd1  0.6.32-2
ii  libc6  2.24-11+deb9u3
ii  libcurl3-gnutls7.52.1-5+deb9u5
ii  libgcc11:6.3.0-18+deb9u1
ii  libqt4-network 4:4.8.7+dfsg-11
ii  libqtcore4 4:4.8.7+dfsg-11
ii  libqtgui4  4:4.8.7+dfsg-11
ii  libssl1.1  1.1.0f-3+deb9u2
ii  libstdc++6 6.3.0-18+deb9u1
ii  libx11-6   2:1.6.4-3
ii  libxext6   2:1.3.3-1+b2
ii  libxi6 2:1.7.9-1
ii  libxinerama1   2:1.1.3-1+b3
ii  libxrandr2 2:1.5.1-1
ii  libxtst6   2:1.2.3-1
ii  openssl1.1.0f-3+deb9u2

synergy recommends no packages.

synergy suggests no packages.

-- no debconf information



Bug#829300: - supplemental file

2016-07-02 Thread Shannon Dealy


The attached file can be used to demonstrate the bug.

 - Select "not empty" for the Y column filter and the chart will display
   correctly, sorted by the X values.

 - Change the filter selection so all rows display again. This will
   make the chart return to its original incorrect display

 - Now select cells B8 - C11 and delete them with the option to move
   cells up. The chart will now display incorrectly in a different
   manner, even though the data has not changed and remains in the
   same order, only empty cells have been removed.

 - Other changes to the number of rows of empty cells can result
   different versions of the wrong display.

Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
Telephone: +1 541-929-4089 |- Natural Building Instruction -
USA only: 800-467-5820 |www.deatech.com

Chart-sort-by-X-bug.ods
Description: bug example


Bug#829300: libreoffice-calc: Chart, x-y scatter with Sort by X values breaks with empty cells

2016-07-02 Thread Shannon Dealy
Package: libreoffice-calc
Version: 1:4.3.3-2+deb8u4
Severity: normal

Dear Maintainer,

I selected two columns of cells in a spreadsheet which had empty cells
between the rows of data in these columns, however, a filter was turned on
which displayed only the non-empty rows for these columns. I created an X-Y
scatter plot with points and lines in calc, with the "Sort by X values"
option and the chart was correctly created. Once the filter was turned off,
displaying the empty cells, the points remained in the correct places, however,
the order of the lines was rearranged so that the X sorting was no longer
being correctly observed. I had naturally expected that the chart would
not be affected by the filter being removed as only empty cells were added.

Some experimentation has shown that it is not the filter, but some strange
function of the number of empty cells between the rows that have data.
With the filter turned off and just inserting and deleting empty rows I get
different plots for the exact same data, given in the exact same order.
I created a spreadsheet with nothing but the data being ploted and a chart
and found that the order of the line plotting changed depending on the
number of rows of empty cells between the different rows of data values.
In other words, it does not just fail to sort the rows by the X value when
there are empty cells, but the order it uses for the plot changes based
on where and how many empty rows there are between the data values.


-- System Information:
Debian Release: 8.5
  APT prefers stable-updates
  APT policy: (990, 'stable-updates'), (990, 'stable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.16.0-4-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash
Init: sysvinit (via /sbin/init)

Versions of packages libreoffice-calc depends on:
ii  coinor-libcbc32.8.12-1
ii  coinor-libcoinmp1 1.7.6+dfsg1-1
ii  libboost-iostreams1.55.0  1.55.0+dfsg-3
ii  libc6 2.19-18+deb8u4
ii  libgcc1   1:4.9.2-10
ii  libicu52  52.1-8+deb8u3
ii  liblcms2-22.6-3+b3
ii  libmwaw-0.3-3 0.3.1-2
ii  libodfgen-0.1-1   0.1.1-2
ii  liborcus-0.8-00.7.0+dfsg-9
ii  libreoffice-base-core 1:4.3.3-2+deb8u4
ii  libreoffice-core  1:4.3.3-2+deb8u4
ii  librevenge-0.0-0  0.0.1-3
ii  libstdc++64.9.2-10
ii  libwps-0.3-3  0.3.0-2
ii  libxml2   2.9.1+dfsg1-5+deb8u2
ii  lp-solve  5.5.0.13-7+b1
ii  uno-libs3 4.3.3-2+deb8u4
ii  ure   4.3.3-2+deb8u4
ii  zlib1g1:1.2.8.dfsg-2+b1

libreoffice-calc recommends no packages.

Versions of packages libreoffice-calc suggests:
ii  ocl-icd-libopencl1  2.2.3-1+deb8u1

Versions of packages libreoffice-core depends on:
ii  fontconfig2.11.0-6.3
ii  fonts-opensymbol  2:102.6+LibO4.3.3-2+deb8u4
ii  libatk1.0-0   2.14.0-1
ii  libboost-date-time1.55.0  1.55.0+dfsg-3
ii  libc6 2.19-18+deb8u4
ii  libcairo2 1.14.0-2.1+deb8u1
ii  libclucene-contribs1  2.3.3.4-4
ii  libclucene-core1  2.3.3.4-4
ii  libcmis-0.4-4 0.4.1-7
ii  libcups2  1.7.5-11+deb8u1
ii  libcurl3-gnutls   7.38.0-4+deb8u3
ii  libdbus-1-3   1.8.20-0+deb8u1
ii  libdbus-glib-1-2  0.102-1
ii  libeot0   0.01-3
ii  libexpat1 2.1.0-6+deb8u3
ii  libexttextcat-2.0-0   3.4.4-1
ii  libfontconfig12.11.0-6.3
ii  libfreetype6  2.5.2-3+deb8u1
ii  libgcc1   1:4.9.2-10
ii  libgdk-pixbuf2.0-02.31.1-2+deb8u5
ii  libgl1-mesa-glx [libgl1]  10.3.2-1+deb8u1
ii  libglew1.10   1.10.0-3
ii  libglib2.0-0  2.42.1-1+b1
ii  libgltf-0.0-0 0.0.2-2
ii  libglu1-mesa [libglu1]9.0.0-2
ii  libgraphite2-31.3.6-1~deb8u1
ii  libgtk2.0-0   2.24.25-3+deb8u1
ii  libharfbuzz-icu0  0.9.35-2
ii  libharfbuzz0b 0.9.35-2
ii  libhunspell-1.3-0 1.3.3-3
ii  libhyphen02.8.8-1
ii  libice6   2:1.0.9-1+b1
ii  libicu52  52.1-8+deb8u3
ii  libjpeg62-turbo   1:1.3.1-12
ii  liblangtag1   0.5.1-3
ii  liblcms2-22.6-3+b3
ii  libldap-2.4-2 2.4.40+dfsg-1+deb8u2
ii  libmythes-1.2-0   2:1.2.4-1
ii  libneon27-gnutls  0.30.1-1
ii  libnspr4  2:4.10.7-1+deb8u1
ii  libnspr4-0d   2:4.10.7-1+deb8u1
ii  libnss3   2:3.17.2-1.1+deb8u2
ii  libnss3-1d2:3.17.2-1.1+deb8u2
ii  libodfgen-0.1-1   0.1.1-2
ii  libpango-1.0-01.36.8-3
ii  libpangocairo-1.0-0   1.36.8-3
ii  libpangoft2-1.0-0 1.36.8-3
ii  libpng12-01.2.50-2+deb8u2
ii  librdf0  

Bug#771969: Workaround

2014-12-13 Thread Shannon Dealy

On Tue, 9 Dec 2014, Nikolaus Rath wrote:


As a workaround, could you use the attached program instead of mount.s3ql?

It should work just like mount.s3ql, but it will retry on any kind of
DNS error.


I'm still planning to fix this properly (probably by not retrying on the
first resolution attempt), but that is going to take a while.


Thanks Nikolaus for this and all your efforts.

It should be noted for anyone else who tries to use this wrapper around 
the mount program, it requires python3-dugong version 3.4 (currently in 
experimental), it throws an exception with the 3.3 version in testing:


  AttributeError: 'module' object has no attribute 'HostnameNotResolvable'

I'm giving it try now which should be a good test of the code and local 
network as I'm moving about 400GB into an S3 file system.


Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-06 Thread Shannon Dealy

On Fri, 5 Dec 2014, Nikolaus Rath wrote:


Shannon Dealy  writes:

[snip]

Given the senario you were trying to fix with the change, perhaps a
better approach would be to fail if initial resolution fails, but if
the initial resolution succeeds, then the end point can reasonably be
assumed to exist and future failures should keep retrying, at least
for a substantial (possibly configurable) period of time.  This
provides the immediate feedback for manual or scripting related
interactions, but once the file system is mounted focuses on
maintaining/recovering the connection.


Yes, that would be the best solution. It's just ugly to code, either way
you to pass a "do_not_fail" parameter all the way from the main function
to the low level network routines, or you have to do the first lookup on
a higher level (which then needs to know details about all the storage
backends).


I would suggest that it should pass a timeout value rather than a 
"do_not_fail".  This would give the application level code the greatest 
flexibility allowing for both immediate and short term failure settings as 
well as an effective "never fail" by just passing a ridiculously large 
value.



Personally, I would love it if I could simply keep the file system
mounted at all times, even when there is no network link, so that when
there is a connection I can simply start using it, and when the
connection goes away (even for a day or two), everything blocks and
simply picks up where it left off when the connection returns.


Well, that should actually work already. When there is no network
connection, s3ql retries for a long amount. The problem only arises if
there is partial connectivity.


I tried it on the older version of the code I was using and it never 
recovered (not sure why).  Don't know on the current version.


Had another filesystem crash last night and while I can't say the exact 
series of events from the perspective of the file system, it was a 
complete network failure that crashed it (not just DNS).  This would imply 
that perhaps the test is too sensitive right at the boundary of a network 
failure (perhaps some packets get through, but most don't) and needs to 
retry over a longer period of time before deciding if the failure is 
network or DNS.


Upon further reflection:

I looked over your dugong code and have given some thought to what little 
I know of the local network topology, and my guess is that your test for 
live DNS will always decide that DNS is up at my location whenever the 
network connection fails (though it is just a guess).  The ISP appears to 
be in another country, and their local network in this building feeds 
about 500 rooms.  It is likely they are using a local DNS caching server 
(for the building, city or country, doesn't really matter which) which 
responds from its cache with anything it knows about, and forwards 
all other requests up to the ISP's main servers.  If that is the case, any 
time the network gets cut off between the local caching DNS server and the 
primary DNS servers, the test will fail because google.com will always 
resolve (since everyone uses it), but www.iana.org and C.root-servers.org 
will not since most internet users never have reason to do a direct lookup 
on the later two domains, and any recursive lookups of these domains as a 
result of a local query will always happen and be stored only at the 
primary server.


Based on this, I would suggest that a more robust test would be to declare 
a DNS failure if any of the three lookups fail (after the host lookup 
previously failed).  After all, a failure on any of these lookups would at 
least suggest a serious problem with the internet which is a reasonably 
likely cause of the initial host lookup failure.


FWIW.

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-05 Thread Shannon Dealy

On Fri, 5 Dec 2014, Nikolaus Rath wrote:


Shannon Dealy  writes:

On Thu, 4 Dec 2014, Nikolaus Rath wrote:


Shannon Dealy  writes:

[snip]

Unfortunately Linux does not provide an easy way to distinguish between
an unavailable DNS server, and an un-resolvable host name. To
distinguish these cases, S3QL/Dugong attempts to resolve a number of
"test" hostnames. If these resolve, but the S3 hostname does not,
S3QL/Dugong concludes that this hostname is not resolvable and
terminates. Otherwise it assumes that the DNS server is currently not
reachable and retries.

Attempting to resolve hostnames on your system frequently fails
(sometimes 3 times in a row), and sometimes it's apparently sufficiently
flaky that

1. server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com cannot
be resolved

2. Any of google.com, iana.org, or root-servers.net can be resolved

3. server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com cannot
be resolved

in this order, and without any waiting times.



It occurs to me that the above can be problematic as a test under some 
senarios.  When segmentation of the internet occurs, routing and DNS 
(from a given location) is often lost to some but not all major players. 
On a couple of occasions I have been unable to resolve google.com but 
could resolve debian.org, amazon.com and many other sites.  I understand 
your approach and reasoning, and at the moment can't think of a better 
solution to what you are trying to do here, however, even if my network 
link and DNS server are running flawlessly, events can occur on the 
internet where where my connection will fail the above test.


[snip]

The relevant code in S3QL 2.2 was last touched in version 2.2, so I
don't think that's it. However, relevant code in Dugong was last
modified in version 3.2. Prior to 3.2, *any* DNS problem was considered
temporary. In 3.2 and later, DNS problems are only considered temporary
of *no* hostname can be resolved.

The problem with the prior behavior was that it would permanently retry
even in situations like this:

mkfs.s3ql s3://mybucket.aws.s3.cmo

Having this command block for any significant amount of time in the hope
that the wrong hostname becomes resolvable was rather surprising, and
people complained that their scripts would just hang instead of properly
reporting an error.

It seems for you the situation is the other way around...


Which proves once again that you just can't win in life :-)

Given the senario you were trying to fix with the change, perhaps a better 
approach would be to fail if initial resolution fails, but if the 
initial resolution succeeds, then the end point can reasonably be assumed 
to exist and future failures should keep retrying, at least for a 
substantial (possibly configurable) period of time.  This provides the 
immediate feedback for manual or scripting related interactions, but once
the file system is mounted focuses on maintaining/recovering the 
connection.


Personally, I would love it if I could simply keep the file system mounted 
at all times, even when there is no network link, so that when there is a 
connection I can simply start using it, and when the connection goes away 
(even for a day or two), everything blocks and simply picks up where it 
left off when the connection returns.  I often leave my sshfs mount 
connected this way as it recovers when the link returns.


Of course my use-case may be different from others, I use it for 
maintaining laptop backups and access to data that won't fit on my 
laptop drive.  This means that network connections come and go multiple 
times each day as I move between different locations, and stay down 
whenever I don't use the laptop for a few days.


FWIW.

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-04 Thread Shannon Dealy

On Thu, 4 Dec 2014, Nikolaus Rath wrote:


Shannon Dealy  writes:

[snip]

Unfortunately Linux does not provide an easy way to distinguish between
an unavailable DNS server, and an un-resolvable host name. To
distinguish these cases, S3QL/Dugong attempts to resolve a number of
"test" hostnames. If these resolve, but the S3 hostname does not,
S3QL/Dugong concludes that this hostname is not resolvable and
terminates. Otherwise it assumes that the DNS server is currently not
reachable and retries.

Attempting to resolve hostnames on your system frequently fails
(sometimes 3 times in a row), and sometimes it's apparently sufficiently
flaky that

1. server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com cannot
be resolved

2. Any of google.com, iana.org, or root-servers.net can be resolved

3. server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com cannot
be resolved

in this order, and without any waiting times.


At this point, S3QL thus assumes that your bucket ceased to exist and
terminates.

To avoid this, you'll have to fix the DNS resolution issues on your
system. Maybe install a caching proxy nameserver like dnsmasq?


While I can try dnsmasq as a solution, I was not having the degree of 
problems I am seeing under the old version of s3ql, and this is not the 
only network I connect to where I am seeing these problems.  This could 
simply mean one or more of:


 - the network(s) have become more unstable (certainly possible)
 - that S3QL's test doesn't work quite the same way as before
 - there is a bug somewhere (S3QL or python libraries - I had to upgrade
   python to install the new S3QL) that makes it appear that DNS is still
   broken after it recovers
 - or something else is going on.

I see in the log I sent you that a few failures were reported 10 minutes 
before S3QL gave up, but it is not clear to me if S3QL was able to resolve 
DNS names between these.  Am I correct in assuming that there a timer as 
well as the three time in a row failure limit you mentioned?  If not, 
there should be.  Network failures often take a few minutes to 
resolve/recover so I would imagine the test should look something like:


   while failing and < 10 minutes (or possibly some configurable value)
  wait some short interval
  try again

It is certainly not unusual to see short network glitches (DNS, 
connection loss, routing problems) lasting a few seconds to (rarely) a 
minute here, the five to ten minute DNS or general outage that I would 
view as a minimum before S3QL gives up is very uncommon (I pretty much 
live on the network when I am awake, so I am pretty sure I would notice 
problems of this level).  Of course, I may be wrong, perhaps I should set 
up some kind of network monitor to see what kind of failure intervals 
there are.


FWIW.

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-04 Thread Shannon Dealy

On Wed, 3 Dec 2014, Nikolaus Rath wrote:


Hi,

Is it possible that your DNS server behaves as described in
https://bitbucket.org/nikratio/python-dugong/issue/16/?

If so, could you test if upgrading to python3-dugong 3.4 from
experimental fixes the problem?


I installed python-dugong from experimental and proceeded to transfer a 
lot of data to my S3 file system.  Stopped it a couple of times, unmounted 
once (successfully), remounted and continued transferring data.  After 
around 13-15 GB total had been uploaded, the file system crashed again (no 
hang):


   rsync: writefd_unbuffered failed to write 4 bytes to socket [sender]:
   Broken pipe (32)
   rsync: write failed on 
   Transport endpoint is not connected (107)
   rsync error: error in file IO (code 11) at receiver.c(322)
   [receiver=3.0.9]
   rsync: connection unexpectedly closed (51874 bytes received so far)
   [sender]
   rsync error: error in rsync protocol data stream (code 12) at io.c(605)
   [sender=3.0.9]

See attached mount log file for more details.

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com2014-12-04 22:21:27.735 31437:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-12-04 22:21:27.738 31437:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-12-04 22:21:36.322 31437:MainThread (name)s.get_metadata: Using cached 
metadata.
2014-12-04 22:21:36.325 31437:MainThread (name)s.main: Mounting filesystem...
2014-12-04 22:21:36.375 31445:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 31446
2014-12-05 00:18:35.346 31446:Thread-3 (name)s.wrapped: Encountered 
DNSUnavailable exception (Unable to resolve 
server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com, DNS server 
unavailable.), retrying call to ObjectW.close for the 3-th time...
2014-12-05 00:18:53.835 31446:Thread-10 (name)s.wrapped: Encountered 
DNSUnavailable exception (Unable to resolve 
server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com, DNS server 
unavailable.), retrying call to ObjectW.close for the 3-th time...
2014-12-05 00:18:58.710 31446:Thread-9 (name)s.wrapped: Encountered 
DNSUnavailable exception (Unable to resolve 
server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com, DNS server 
unavailable.), retrying call to ObjectW.close for the 3-th time...
2014-12-05 00:19:05.143 31446:Thread-4 (name)s.wrapped: Encountered 
DNSUnavailable exception (Unable to resolve 
server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com, DNS server 
unavailable.), retrying call to ObjectW.close for the 3-th time...
2014-12-05 00:19:13.463 31446:Thread-6 (name)s.wrapped: Encountered 
DNSUnavailable exception (Unable to resolve 
server-external-2014-11-21-deatech-com-s3ql.s3.amazonaws.com, DNS server 
unavailable.), retrying call to ObjectW.close for the 3-th time...
2014-12-05 00:28:34.696 31446:Thread-10 (name)s.excepthook: Uncaught top-level 
exception:
Traceback (most recent call last):
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 1445, in 
create_socket
return socket.create_connection(address)
  File "/usr/lib/python3.4/socket.py", line 491, in create_connection
for res in getaddrinfo(host, port, 0, SOCK_STREAM):
  File "/usr/lib/python3.4/socket.py", line 530, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -2] Name or service not known

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/usr/lib/s3ql/s3ql/mount.py", line 66, in run_with_except_hook
run_old(*args, **kw)
  File "/usr/lib/python3.4/threading.py", line 868, in run
self._target(*self._args, **self._kwargs)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 404, in _upload_loop
self._do_upload(*tmp)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 431, in _do_upload
% obj_id).get_obj_size()
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 258, in perform_write
return fn(fh)
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 477, in __exit__
self.close()
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 471, in close
self.fh.close()
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 636, in close
self.fh.close()
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 845, in close
headers=self.headers, body=self.fh)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 409, in _do_request
query_string=query_string, body=body)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 642, in _send_request
headers=headers, body=BodyFollowing(b

Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-04 Thread Shannon Dealy

On Thu, 4 Dec 2014, Nikolaus Rath wrote:


On 12/03/2014 11:44 PM, Shannon Dealy wrote:
the file system still hangs (or the next time it hangs), could you

follow the instructions at
https://bitbucket.org/nikratio/s3ql/wiki/Providing%20Debugging%20Info to
obtain a Python stacktrace, and attach it to this issue?


This time the file system didn't hang until I attempted to unmount it.
Attached are two log files with the mount and stack trace information.


This looks odd. Did you issue both "setfattr -n fuse_stacktrace" *and*
"kill -s USR1" commands?


Yes, I didn't realize that the setfattr command was being used to trigger 
a one time event, I thought it was more like turning on a debug mode with 
extra info.


If I am not mixing things up, I believe that initial setfattr was run 
right after mounting the file system, and the "kill -s USR1" was run when 
the file system was hung during the unmount, just before I did a kill -9

on mount.s3ql


The attached files indicate both, and moreover, the number of active
threads during the "kill -s USR1" command was much lower than during the
"setfattr" command. How much time did pass between these commands
(assuming that you issued both)?


I don't recall.


At least at the time that you issued the "kill" command, mount.s3ql was
busy doing the SSL handshake with the remote server, so while it looked
like it was hanging, it was probably waiting for the remote server.


It had been waiting a very long time, though I don't remember exactly, 
for this case.  I have limited the cache to 1GB so I don't have to wait 
too long if I need to kill a backup, pack up the computer and leave. 
Because of this, I know from past experience the maximum amount of time
it takes to flush the cache and unmount using this network.  I figure it 
is dead if the network has remained up, but network and cpu load for this 
process are minimal to zero and it has taken more than twice as long as it 
should for the unmount.  In this case it should be done in 15 to 20 
minutes max, so figure it had been at least twice that.  Sorry I don't 
have a better answer.  If you have a suggestion for a better way to tell 
if the filesystem is hung, it would be helpful.  On one occasion I let a 
hung system sit for hours to see if it would recover (it didn't).


Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#772052: (s3ql: umount.s3ql throws exception on s3ql file system mounted via sshfs file system)

2014-12-04 Thread Shannon Dealy

On Thu, 4 Dec 2014, Nikolaus Rath wrote:

[snip]
That's the reason for the exception on umount then. mount.s3ql 
currently isn't able to handle that. The existing data is ignored as you 
say, but mount.s3ql attempts to rmdir the cache directory after 
unmounting. That fails if there are still files in there.


Normally, the only way for this directory to exist is if the file 
system was not unmounted cleanly. In this case mount.s3ql will refuse to 
start, and you have to run fsck.s3ql instead. fsck.s3ql will then 
clean-up the cache directory, and mount.s3ql will start with an empty 
cache.


In your case...


It should be noted that the fsck run on this file system was not run
from my local machine, but from the remote server, so when I mount this
on my local machine, it says something to the effect that the local
cache is out of date, and then proceeds to download and unpack the
current file system information from the remote server.


.. you have neatly circumvented the clean-up of the cache directory. 
mount.s3ql now believes the file system is clean, but there is actually 
a cache directory with files in it.


This is certainly a bug in S3QL, but I have to think about what the 
correct behavior for mount.s3ql actually would be. Refusing to mount 
would be odd, because the file system is indeed clean. But silently 
deleting the data in there intuitively sounds like a dangerous choice as 
well.


While I don't know what is in the cache, it would seem to me that once the 
file system has been cleaned elsewhere that the cache data would cease to
be usable for any purpose (other than forensic) as it would be 
inconsistent with the state of the file system.  If so, I can't 
see that you have any real choice but to discard it.  Informing the user 
that you have data available but they can't use it for anything would 
serve no purpose.  You do provide warnings about multiple computers, 
including a warning when I run fsck on a computer which was not where the 
file system was last mounted.


FWIW.

At least this gives me a simple work-around for the bug :-)

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#772052: (s3ql: umount.s3ql throws exception on s3ql file system mounted via sshfs file system)

2014-12-04 Thread Shannon Dealy

On Thu, 4 Dec 2014, Nikolaus Rath wrote:


On 12/04/2014 10:07 AM, Shannon Dealy wrote:


Attached is the log file for a failed unmount.

While my previous attempts were simple umount.s3ql commands, this time a
couple of additional commands were run after rsync completed and before
the umount:

   s3qlctrl flushcache /media/server-external
   s3qlctrl upload-meta /media/server-external
   umount.s3ql /media/server-external


The logs looks as you also send a SIGUSR1 signal to mount.s3ql. Is that
correct?


No, it ran to completion on its own (though it took forever).  I forgot 
to mention that I did issue this command just before running umount.s3ql


   setfattr -n fuse_stacktrace /media/server-external

though I am not sure if that makes any difference with fuse as that 
command had previously been issued for that mount point when a different 
file system (the other one we have been debugging) was mounted there, and 
I have no idea if fuse retains this setting between unmounts and mounts

on a given mount point.


After the umount.s3ql command, what are the contents of
/root/.s3ql/local:=2F=2F=2Fmedia=2FTransferDrive=2FS3QL_server-external-cache?
Can you confirm that this directory did not exist when calling mount.s3ql?


No, in fact based on the timestamps, I can confirm that this directory did 
exist as it has over 4000 files in it with November 30th timestamps 
spanning roughly 1/2 hour.  I never gave this directory a thought as the 
mount indicates that the cache is out of date, so I assumed any old data
that was lying around would be discarded when it downloads the current 
file system info.


It should be noted that the fsck run on this file system was not run from 
my local machine, but from the remote server, so when I mount this on my 
local machine, it says something to the effect that the local cache is 
out of date, and then proceeds to download and unpack the current file 
system information from the remote server.  I realize this discards any 
chance of recovering data that might have been transferred before the 
crash, but past experience has shown me that in most cases it is 
significantly faster to run fsck at the other end and resend the data.
The problem is that fsck.s3ql is incredibly slow for my mounts using 
sshfs - well over an order of magnitude slower than my S3 mounts.


Not sure what the issue is with s3ql over sshfs, the file system 
performance while a bit on the slow side seems reasonable until I try to 
fsck or umount the file system, then it takes forever (though maybe I just 
notice it more as that is a more interactive use of the file system).


Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#772052: (s3ql: umount.s3ql throws exception on s3ql file system mounted via sshfs file system)

2014-12-04 Thread Shannon Dealy


Attached is the log file for a failed unmount.

While my previous attempts were simple umount.s3ql commands, this time a 
couple of additional commands were run after rsync completed and before 
the umount:


   s3qlctrl flushcache /media/server-external
   s3qlctrl upload-meta /media/server-external
   umount.s3ql /media/server-external

None displayed any errors at the command line level, but the end result 
was the same, next mount attempt indicated:


   File system damaged or not unmounted cleanly, run fsck!

Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com2014-12-04 13:14:27.483 13515:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-12-04 13:14:27.487 13515:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-12-04 13:14:41.599 13515:MainThread (name)s.get_metadata: Ignoring locally 
cached metadata (outdated).
2014-12-04 13:14:42.226 13515:MainThread (name)s.get_metadata: Downloading and 
decompressing metadata...
2014-12-04 13:23:26.960 13515:MainThread (name)s.get_metadata: Reading 
metadata...
2014-12-04 13:23:26.965 13515:MainThread (name)s.restore_metadata: ..objects..
2014-12-04 13:23:34.746 13515:MainThread (name)s.restore_metadata: ..blocks..
2014-12-04 13:24:10.699 13515:MainThread (name)s.restore_metadata: ..inodes..
2014-12-04 13:25:01.822 13515:MainThread (name)s.restore_metadata: 
..inode_blocks..
2014-12-04 13:26:16.473 13515:MainThread (name)s.restore_metadata: 
..symlink_targets..
2014-12-04 13:26:19.260 13515:MainThread (name)s.restore_metadata: ..names..
2014-12-04 13:26:46.256 13515:MainThread (name)s.restore_metadata: ..contents..
2014-12-04 13:29:10.849 13515:MainThread (name)s.restore_metadata: 
..ext_attributes..
2014-12-04 13:29:30.511 13515:MainThread (name)s.main: Setting cache size to 
42007 MB
2014-12-04 13:29:30.513 13515:MainThread (name)s.main: Mounting filesystem...
2014-12-04 13:29:30.535 13870:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 13874
2014-12-04 15:01:30.236 13874:Metadata-Upload-Thread (name)s.run: Dumping 
metadata...
2014-12-04 15:01:30.240 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..objects..
2014-12-04 15:01:33.701 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..blocks..
2014-12-04 15:01:40.118 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..inodes..
2014-12-04 15:02:22.569 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..inode_blocks..
2014-12-04 15:02:45.095 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..symlink_targets..
2014-12-04 15:02:45.892 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..names..
2014-12-04 15:02:48.911 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..contents..
2014-12-04 15:03:16.247 13874:Metadata-Upload-Thread (name)s.dump_metadata: 
..ext_attributes..
2014-12-04 15:03:17.194 13874:Metadata-Upload-Thread (name)s.run: Compressing 
and uploading metadata...
2014-12-04 15:06:19.607 13874:Metadata-Upload-Thread (name)s.run: Wrote 101 MiB 
of compressed metadata.
2014-12-04 15:06:19.609 13874:Metadata-Upload-Thread (name)s.run: Cycling 
metadata backups...
2014-12-04 15:06:19.609 13874:Metadata-Upload-Thread (name)s.cycle_metadata: 
Backing up old metadata...
2014-12-04 15:07:34.506 13874:Dummy-22 (name)s.stacktrace: 
# ThreadID: 140728286045952
pyapi.py:504, in stacktrace
for filename, lineno, name, line in traceback.extract_stack(frame):

# ThreadID: 140728294438656
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
queue.py:167, in get
self.not_empty.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140729826797312
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
block_cache.py:399, in _upload_loop
tmp = self.to_upload.get()
block_cache.py:97, in get
self.cv.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140728831309568
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
block_cache.py:677, in _removal_loop
tmp = self.to_remove.get(block=len(ids)==0)
queue.py:167, in get
self.not_empty.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140728822916864
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_

Bug#772052: s3ql: umount.s3ql throws exception on s3ql file system mounted via sshfs file system

2014-12-04 Thread Shannon Dealy
Package: s3ql
Version: 2.11.1+dfsg-2
Severity: important

Dear Maintainer,

I have an s3ql file system on a remote server and access it via a local mount
through sshfs.  It worked fine with version 2.8.1, but since upgrading to
to 2.11.1, umount.s3ql fails every time.  Further, umount.s3ql does not
report the failure, it appears to complete normally, so you only find out
about the failure when you try to remount it and get an error:

   Using 8 upload threads.
   Autodetected 4040 file descriptors available for cache entries
   Enter file system encryption passphrase: 
   Using cached metadata.
   File system damaged or not unmounted cleanly, run fsck!

The mount log file does show that an exception occured while unmounting the
file system, even though no error was reported on the command line (I would
guess that technically it is mount.s3ql that throws the exception)

I have tried this six or seven times after running fsck with the same result
each time.

-- System Information:
Debian Release: 7.7
  APT prefers stable-updates
  APT policy: (500, 'stable-updates'), (500, 'unstable'), (500, 'testing'), 
(500, 'stable'), (1, 'experimental')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.13-0.bpo.1-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages s3ql depends on:
ii  fuse   2.9.3-9
ii  libc6  2.18-4
ii  libjs-sphinxdoc1.1.3+dfsg-4
ii  libsqlite3-0   3.7.13-1+deb7u1
ii  psmisc 22.19-1+deb7u1
ii  python33.4.2-1
ii  python3-apsw   3.8.6-r1-1
ii  python3-crypto 2.6.1-5+b2
ii  python3-defusedxml 0.4.1-2
ii  python3-dugong 3.4+dfsg-1
ii  python3-llfuse 0.40-2+b2
ii  python3-pkg-resources  5.5.1-1
ii  python3-requests   2.4.3-4

s3ql recommends no packages.

s3ql suggests no packages.

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-04 Thread Shannon Dealy

On Wed, 3 Dec 2014, Nikolaus Rath wrote:


severity 771969 important
thanks

Thanks for the report! I'll set the severity to important for now,
because the package still works nicely for many other users. Not sure
why it's making so much trouble for you..


Because computers hate me :-)  I actually have yet another problem with 
an sshfs mounted s3ql file system but was trying to get the issues with 
my S3 mount resolved first.



but I notice that (like the
last bug you found) this seems related to the handling of unreliable
network connections. In this case there seems to be a temporary problem
with your DNS server which S3QL has trouble coping with.


From a high level perspective, the connection (generally) is pretty 

reliable, but I wouldn't notice momentary network issues.


The raw_streamXXX files are from the patch that you applied to the debug
the previous issue. I'd recommend to purge and reinstall both the s3ql
and python3-dugong packages to make sure that you're in vanilla state again.


I forgot about the dugong patch, so I reinstalled both packages, using 
the Debian-unstable version for s3ql



If the file system still hangs (or the next time it hangs), could you
follow the instructions at
https://bitbucket.org/nikratio/s3ql/wiki/Providing%20Debugging%20Info to
obtain a Python stacktrace, and attach it to this issue?


This time the file system didn't hang until I attempted to unmount it. 
Attached are two log files with the mount and stack trace information.



Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com2014-12-04 07:28:56.413 31144:MainThread (name)s.excepthook: Mountpoint does 
not exist.
2014-12-04 07:39:43.743 609:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-12-04 07:39:43.746 609:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-12-04 07:39:56.730 609:MainThread (name)s.get_metadata: Using cached 
metadata.
2014-12-04 07:39:56.734 609:MainThread (name)s.main: Mounting filesystem...
2014-12-04 07:39:56.821 626:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 630
2014-12-04 07:42:45.765 630:MainThread (name)s.main: FUSE main loop terminated.
2014-12-04 07:42:45.786 630:MainThread (name)s.unmount: Unmounting file 
system...
2014-12-04 07:42:46.272 630:MainThread (name)s.main: File system unchanged, not 
uploading metadata.
2014-12-04 07:42:46.309 630:MainThread (name)s.main: Cleaning up local 
metadata...
2014-12-04 07:42:51.966 630:MainThread (name)s.main: All done.
2014-12-04 07:42:58.936 3663:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-12-04 07:42:58.938 3663:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-12-04 07:43:05.572 3663:MainThread (name)s.get_metadata: Using cached 
metadata.
2014-12-04 07:43:05.582 3663:MainThread (name)s.main: Mounting filesystem...
2014-12-04 07:43:05.602 3669:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 3671
2014-12-04 07:43:12.769 3671:Dummy-22 (name)s.stacktrace: 
# ThreadID: 140042022409984
pyapi.py:504, in stacktrace
for filename, lineno, name, line in traceback.extract_stack(frame):

# ThreadID: 140042039195392
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
queue.py:167, in get
self.not_empty.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140043138082560
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
block_cache.py:677, in _removal_loop
tmp = self.to_remove.get(block=len(ids)==0)
queue.py:167, in get
self.not_empty.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140042576066304
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
block_cache.py:677, in _removal_loop
tmp = self.to_remove.get(block=len(ids)==0)
queue.py:167, in get
self.not_empty.wait()
threading.py:290, in wait
waiter.acquire()

# ThreadID: 140043154867968
threading.py:888, in _bootstrap
self._bootstrap_inner()
threading.py:920, in _bootstrap_inner
self.run()
mount.py:66, in run_with_except_hook
run_old(*args, **kw)
threading.py:868, in run
self._target(*self._args, **self._kwargs)
block_cache.py:677, in _removal_loop
tmp = self.to_remove.get(blo

Bug#771969: s3ql: mount.s3ql throws exception and hangs file system

2014-12-03 Thread Shannon Dealy
Package: s3ql
Version: 2.11.1+dfsg-2
Severity: grave
Justification: renders package unusable

Dear Maintainer,

I mounted an S3QL file system and ran rsync to transfer data to the file
system.  After a few minutes I noticed the rsync process was no longer 
scanning the file system or transfering data.  Ctrl-C would not terminate
the rsync (presumably locked in some I/O transaction).  mount.s3ql appears
hung and I could not unmount the file system.

The mount log file (see below) showed an exception had been thrown and
my .s3ql directory has about 600 raw_streamXXX files in it.

I killed everything, ran fsck on the file system, remounted and tried again
with the same result.

Contents of mount.log:

2014-12-03 23:17:03.472 23261:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-12-03 23:17:03.473 23261:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-12-03 23:17:12.025 23261:MainThread (name)s.get_metadata: Using cached 
metadata.
2014-12-03 23:17:12.029 23261:MainThread (name)s.main: Mounting filesystem...
2014-12-03 23:17:12.043 23269:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 23270
2014-12-03 23:20:46.372 23270:Thread-18 (name)s._parse_xml_response: Server did 
not provide Content-Type, assuming XML
2014-12-03 23:21:21.816 23270:Thread-18 (name)s.excepthook: Uncaught top-level 
exception:
Traceback (most recent call last):
  File "/usr/lib/s3ql/s3ql/mount.py", line 66, in run_with_except_hook
run_old(*args, **kw)
  File "/usr/lib/python3.4/threading.py", line 868, in run
self._target(*self._args, **self._kwargs)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 684, in _removal_loop
backend.delete_multi(['s3ql_data_%d' % i for i in ids])
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 273, in delete_multi
return self.backend.delete_multi(keys, force=force)
  File "/usr/lib/s3ql/s3ql/backends/s3.py", line 81, in delete_multi
self._delete_multi(tmp, force=force)
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/s3.py", line 116, in _delete_multi
resp = self._do_request('POST', '/', subres='delete', body=body, 
headers=headers)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 409, in _do_request
query_string=query_string, body=body)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 638, in _send_request
self.conn.send_request(method, path, body=body, headers=headers)
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 484, in 
send_request
self.timeout)
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 1373, in 
eval_coroutine
if not next(crt).poll(timeout=timeout):
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 511, in 
co_send_request
self.connect()
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 410, in connect
self._sock = socket.create_connection((self.hostname, self.port))
  File "/usr/lib/python3.4/socket.py", line 491, in create_connection
for res in getaddrinfo(host, port, 0, SOCK_STREAM):
  File "/usr/lib/python3.4/socket.py", line 530, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno -2] Name or service not known




-- System Information:
Debian Release: 7.7
  APT prefers stable-updates
  APT policy: (500, 'stable-updates'), (500, 'unstable'), (500, 'testing'), 
(500, 'stable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.13-0.bpo.1-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages s3ql depends on:
ii  fuse   2.9.3-9
ii  libc6  2.18-4
ii  libjs-sphinxdoc1.1.3+dfsg-4
ii  libsqlite3-0   3.7.13-1+deb7u1
ii  psmisc 22.19-1+deb7u1
ii  python33.4.2-1
ii  python3-apsw   3.8.6-r1-1
ii  python3-crypto 2.6.1-5+b2
ii  python3-defusedxml 0.4.1-2
ii  python3-dugong 3.3+dfsg-2
ii  python3-llfuse 0.40-2+b2
ii  python3-pkg-resources  5.5.1-1
ii  python3-requests   2.4.3-4

s3ql recommends no packages.

s3ql suggests no packages.

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Duplicate object check

2014-12-02 Thread Shannon Dealy

On Tue, 2 Dec 2014, Shannon Dealy wrote:

[snip]

I performed a capture as requested, ...


Sorry Nikolaus,

I have deleted that file from where I posted it.

I made a mistake, that file was captured with ssl enabled.  In fact, it 
appears based on a limited sample, that fsck.s3ql fails 100% of the time 
when ssl is enabled (35+ runs), and succeeds 100% of the time when ssl is 
disabled (6 runs so far).  I don't know if this is a generic work around, 
or if it just happens to work this way for my particular data set.


Given the above, there is no way I can provide you with a "no-ssl" version 
of the captured data (since it always succeeds).  Could you make a patch 
to fsck.s3ql which would provide you with the required data?  Or does the 
difference between ssl and no-ssl behavior give you some ideas?


Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Duplicate object check

2014-12-02 Thread Shannon Dealy

On Mon, 1 Dec 2014, Nikolaus Rath wrote:


severity -1 normal
retitle -1 fsck.s3ql sporadically crashes when checking objects
thanks

Retitling this bug and lowering severity for now. I don't think that the
mount.s3ql crash is related to the fsck.s3ql crashes, or that there is
any data corruption.


While I understand your assessment, I don't think I would characterize 
this bug as "normal".  Based on my understanding, the Debian guidelines 
would rate this bug at the very least as "important", and frankly I would 
still argue for "critical" given that it effectively "causes serious data 
loss" (one of the possible criteria for "critical").


My reasoning is that even assuming the file system data has not actually 
been lost or corrupted (and I believe you are likely correct in this), I 
ran fsck.s3ql roughly 30 times before getting a successful run.  There is 
nothing to indicate that someone else with similar circumstances would not 
have to run it 1000 or 100,000 more times in order to recover the file 
system.  For such a case, their data is unavailable to them and 
effectively 100% lost until they have a fixed version of fsck.s3ql


FWIW.

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Duplicate object check

2014-12-02 Thread Shannon Dealy

On Mon, 1 Dec 2014, Nikolaus Rath wrote:

[snip]

I still have the bucket that was copied using the aws command line
tools, and am in the process of copying that to a new bucket for testing
so we don't lose the corrupt version, but won't get to testing it
tonight.  I have not tried to use the original file system since
fsck.s3ql succeeded and am not entirely sure if it trust it without
knowing what was wrong with fsck.s3ql


At this point, I am pretty confident that this is a bug related to
object listing. Object listing is only used by fsck.s3ql, so I think
that using the file system normally (aka with mount.s3ql) should not
result in any problems.


My concern was not with the basic file system and mount.s3ql, but 
rather with whether or not fsck.s3ql had correctly reconstructed the file 
system after the failure given that there is (at least on occasion), a bug 
in it's handling of some of the file system data.


Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Duplicate object check

2014-12-01 Thread Shannon Dealy

On Mon, 1 Dec 2014, Nikolaus Rath wrote:

[snip]

I think at this point I can probably write you a patch to get the file
system functional again, but I'd very much like to find out what's
happening here.

Would you be able to run fsck with --backend-options no-ssl, and capture
the traffic using Wireshark?


Hi Nikolaus,

I performed several runs of fsck.s3ql while experimenting with wireshark 
(it has been years since I've used it or tcpdump) to get the settings 
right.  Each time, fsck.s3ql failed in the same manner.  Then when I did 
what was to be the final run/capture, it ran to completion without 
errors.


Given the behavior above, the first thing that leaps to mind is possibly a 
race condition.  I would usually expect more inconsistency (such as 
failing at different objects each time) if it was simply uninitialized 
data or corruption, though those may be possibilities too.


I still have the bucket that was copied using the aws command line tools, 
and am in the process of copying that to a new bucket for testing so we 
don't lose the corrupt version, but won't get to testing it tonight.  I 
have not tried to use the original file system since fsck.s3ql succeeded 
and am not entirely sure if it trust it without knowing what was wrong 
with fsck.s3ql


Something more for you to ponder.  FWIW.

Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Copied bucket

2014-11-30 Thread Shannon Dealy

On Sun, 30 Nov 2014, Nikolaus Rath wrote:


On 11/30/2014 12:50 AM, Shannon Dealy wrote:



I suspect this is because you copied the objects into a new bucket, but
did not include the associated metadata.

Which tool did you use to do the copy, and how did you call it?


I used the AWS command line tools:

   aws s3 sync s3://src-bucket s3://dest-bucket

It did fail part way through the transfer, so I restarted it with the
same command.


When you use the S3 Management Console
(https://console.aws.amazon.com/s3/home) to look at the "s3ql_metadata"
object in the original bucket and the copy, does it have the same
metadata?


While the s3ql_metadata file is there, in the new bucket it is missing
all of the keys except the Content-Type.  Not sure whether this means
the fsck without cached data trashed it, or if amazon's sync was
trashing the transfered data because I used it incorrectly or it is
broken somehow.


I think you have to blame the aws tool for this one. If you think
fsck.s3ql is at fault, that's easy enough to check: just run the sync
command again and see if the metadata is present before running fsck.s3ql.


I don't know if that's a bug in aws, or if you're using it incorrectly,
but "gsutil" is able to copy buckets including metadata (in case you
want to try that).


Poking around, it appears that the metadata was lost only on the 
s3ql_metadata files, so I deleted them and ran the "aws s3 sync"
command again.  The new files again were missing the metadata, so I just 
copied the 13 s3ql_metadata files using the aws console, and this time 
the metadata was preserved.


I didn't see anything special about these files with regard to permissions 
or other information, is there anything special about how these files
are created/stored on S3?  If not, then possibly simply having "metadata" 
in the filename may cause the aws s3 sync command to lose the file 
metadata.


In any case, once I replaced the s3ql_metadata files I was able to run 
fsck.s3ql on the copy of the file system bucket without the local data 
cache and it resulted in exactly the same failure mode/exception "PRIMARY 
KEY must be unique".


Will take a look at gsutil, thanks for the suggestion.

FWIW.

Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: Copied bucket

2014-11-30 Thread Shannon Dealy



I suspect this is because you copied the objects into a new bucket, but
did not include the associated metadata.

Which tool did you use to do the copy, and how did you call it?


I used the AWS command line tools:

   aws s3 sync s3://src-bucket s3://dest-bucket

It did fail part way through the transfer, so I restarted it with the same 
command.



When you use the S3 Management Console
(https://console.aws.amazon.com/s3/home) to look at the "s3ql_metadata"
object in the original bucket and the copy, does it have the same 
metadata?


While the s3ql_metadata file is there, in the new bucket it is missing all 
of the keys except the Content-Type.  Not sure whether this means the fsck 
without cached data trashed it, or if amazon's sync was trashing the 
transfered data because I used it incorrectly or it is broken somehow.



Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com


--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#771452: s3ql: fsck.s3ql on crashed file system results in uncaught exception

2014-11-29 Thread Shannon Dealy

On Sat, 29 Nov 2014, Nikolaus Rath wrote:


On 11/29/2014 09:49 AM, Shannon Dealy wrote:

Package: s3ql
Version: 2.11.1+dfsg-1
Severity: critical
Justification: causes serious data loss

Dear Maintainer,

While running rsync to backup data to an s3ql file system mounted from Amazon's
S3 services, the internet connection failed, resulting in the following

[snip]

Chances are good that this can be done, so don't give up hope yet.

Best,
-Nikolaus


Attached are both the mount and fsck log files, stripped of data from 
other (irrelevent) sessions and with slight mods to remove references to 
my S3 file system info (bucket/prefix).


I have left the original bucket and local cache unmodified except for 
whatever changes occured during my fsck attempt in case they can be of 
use in debugging this.


Regards,

Shannon C. Dealy   | DeaTech Research Inc.
de...@deatech.com  |- Custom Software Development -
USA Phone: +1 800-467-5820 |- Natural Building Instruction -
numbers  : +1 541-929-4089 |www.deatech.com2014-11-28 18:08:52.029 31159:MainThread (name)s.determine_threads: Using 8 
upload threads.
2014-11-28 18:08:52.032 31159:MainThread (name)s.main: Autodetected 4040 file 
descriptors available for cache entries
2014-11-28 18:09:00.435 31159:MainThread (name)s.get_metadata: Using cached 
metadata.
2014-11-28 18:09:00.439 31159:MainThread (name)s.main: Mounting filesystem...
2014-11-28 18:09:00.448 31163:MainThread (name)s.detach_process_context: 
Daemonizing, new PID is 31167
2014-11-29 02:14:44.787 31167:Thread-9 (name)s.wrapped: Encountered 
ConnectionTimedOut exception (send/recv timeout exceeded), retrying call to 
ObjectW.close for the 3-th time...
2014-11-29 10:49:36.636 31167:Thread-10 (name)s.excepthook: Uncaught top-level 
exception:
Traceback (most recent call last):
  File "/usr/lib/s3ql/s3ql/mount.py", line 66, in run_with_except_hook
run_old(*args, **kw)
  File "/usr/lib/python3.4/threading.py", line 868, in run
self._target(*self._args, **self._kwargs)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 404, in _upload_loop
self._do_upload(*tmp)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 431, in _do_upload
% obj_id).get_obj_size()
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 258, in perform_write
return fn(fh)
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 477, in __exit__
self.close()
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 471, in close
self.fh.close()
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 636, in close
self.fh.close()
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 845, in close
headers=self.headers, body=self.fh)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 409, in _do_request
query_string=query_string, body=body)
  File "/usr/lib/s3ql/s3ql/backends/s3c.py", line 642, in _send_request
headers=headers, body=BodyFollowing(body_len))
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 477, in 
send_request
self.timeout)
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 1361, in 
eval_coroutine
if not next(crt).poll(timeout=timeout):
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 504, in 
co_send_request
self.connect()
  File "/usr/lib/python3/dist-packages/dugong/__init__.py", line 408, in connect
self._sock = socket.create_connection((self.hostname, self.port))
  File "/usr/lib/python3.4/socket.py", line 509, in create_connection
raise err
  File "/usr/lib/python3.4/socket.py", line 500, in create_connection
sock.connect(sa)
OSError: [Errno 113] No route to host
2014-11-29 10:50:11.907 31167:Thread-9 (name)s.exchook: Unhandled top-level 
exception during shutdown (will not be re-raised)
2014-11-29 10:50:11.907 31167:Thread-9 (name)s.excepthook: Uncaught top-level 
exception:
Traceback (most recent call last):
  File "/usr/lib/s3ql/s3ql/mount.py", line 66, in run_with_except_hook
run_old(*args, **kw)
  File "/usr/lib/python3.4/threading.py", line 868, in run
self._target(*self._args, **self._kwargs)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 404, in _upload_loop
self._do_upload(*tmp)
  File "/usr/lib/s3ql/s3ql/block_cache.py", line 431, in _do_upload
% obj_id).get_obj_size()
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 46, in wrapped
return method(*a, **kw)
  File "/usr/lib/s3ql/s3ql/backends/common.py", line 258, in perform_write
return fn(fh)
  File "/usr/lib/s3ql/s3ql/backends/comprenc.py

Bug#771452: s3ql: fsck.s3ql on crashed file system results in uncaught exception

2014-11-29 Thread Shannon Dealy
Package: s3ql
Version: 2.11.1+dfsg-1
Severity: critical
Justification: causes serious data loss

Dear Maintainer,

While running rsync to backup data to an s3ql file system mounted from Amazon's
S3 services, the internet connection failed, resulting in the following 
error(s) from rsync:

   rsync: writefd_unbuffered failed to write 4 bytes to socket [sender]: Broken 
pipe (32)
   rsync: write failed on "": Software caused connection 
abort (103)
   rsync error: error in file IO (code 11) at receiver.c(322) [receiver=3.0.9]
   rsync: connection unexpectedly closed (17298 bytes received so far) [sender]
   rsync error: error in rsync protocol data stream (code 12) at io.c(605) 
[sender=3.0.9]

I attempted to unmount the file system with the following result (twice):

   # umount.s3ql /media/server-external
   File system appears to have crashed.

I then forced it to unmount as follows:

   # fusermount -u -z /media/server-external

Then attempted to fsck the file system (twice - both gave the same result):

   fsck.s3ql s3:///
   Enter file system encryption passphrase: 
   Starting fsck of s3:///
   Using cached metadata.
   Remote metadata is outdated.
   Checking DB integrity...
   Creating temporary extra indices...
   Checking lost+found...
   Checking cached objects...
   Committing block 14 of inode 442809 to backend
   Committing block 16 of inode 442809 to backend
   Committing block 17 of inode 442809 to backend
   Committing block 15 of inode 442809 to backend
   Committing block 19 of inode 442809 to backend
   Committing block 18 of inode 442809 to backend
   Checking names (refcounts)...
   Checking contents (names)...
   Checking contents (inodes)...
   Checking contents (parent inodes)...
   Checking objects (reference counts)...
   Checking objects (backend)...
   ..processed 10 objects so far..
   Dropping temporary indices...
   Uncaught top-level exception:
   Traceback (most recent call last):
 File "/usr/bin/fsck.s3ql", line 9, in 
   load_entry_point('s3ql==2.11.1', 'console_scripts', 'fsck.s3ql')()
 File "/usr/lib/s3ql/s3ql/fsck.py", line 1189, in main
   fsck.check()
 File "/usr/lib/s3ql/s3ql/fsck.py", line 85, in check
   self.check_objects_id()
 File "/usr/lib/s3ql/s3ql/fsck.py", line 848, in check_objects_id
   self.conn.execute('INSERT INTO obj_ids VALUES(?)', (obj_id,))
 File "/usr/lib/s3ql/s3ql/database.py", line 98, in execute
   self.conn.cursor().execute(*a, **kw)
 File "src/cursor.c", line 231, in resetcursor
   apsw.ConstraintError: ConstraintError: PRIMARY KEY must be unique

Next I copied the entire Amazon bucket to a new bucket and attempted an fsck
on the copy, minus the locally cached file system data:

   # fsck.s3ql s3:///
   Enter file system encryption passphrase: 
   Starting fsck of s3:///
   Uncaught top-level exception:
   Traceback (most recent call last):
 File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 381, in 
_convert_legacy_metadata
   meta_new['data'] = meta['data']
   KeyError: 'data'

   During handling of the above exception, another exception occurred:

   Traceback (most recent call last):
 File "/usr/bin/fsck.s3ql", line 9, in 
   load_entry_point('s3ql==2.11.1', 'console_scripts', 'fsck.s3ql')()
 File "/usr/lib/s3ql/s3ql/fsck.py", line , in main
   param = backend.lookup('s3ql_metadata')
 File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 72, in lookup
   meta_raw = self._convert_legacy_metadata(meta_raw)
 File "/usr/lib/s3ql/s3ql/backends/comprenc.py", line 383, in 
_convert_legacy_metadata
   raise CorruptedObjectError('meta key data is missing')
   s3ql.backends.common.CorruptedObjectError: meta key data is missing


NOTE: I'm not sure about the exact implication of "_convert_legacy_metadata"
  in the traceback above, but this was NOT a legacy file system, it was 
  just created using s3ql 2.11.1 as it is cheaper to rebuild it than
  to pull 700 GB in the old copy down from Amazon to do the "verify"
  specified as part of the upgrade procedure from older versions.

At the time of this failure I had uploaded between 200 and 300 GB of
deduplicated/compressed data to the new file system.

As things currently stand, unless I have overlooked or misunderstood something
(which I consider entirely possible), this network connection failure has
resulted in 100% data loss unless fsck can be fixed in a manner which will
allow it to complete correctly and recover the file system data.  As I
maintain other backups, no actual data has been lost (so far), but this
makes s3ql unsafe to use and further attempts to backup my data to S3
pointless.

Regards,

Shannon Dealy

-- Syste

Bug#747850: xfce4 incorrectly stores session configuration in .cache directory

2014-05-12 Thread Shannon Dealy
Package: xfce4
Version: 4.8.0.3
Severity: normal

Dear Maintainer,

xfce4 is currently storing session information under $HOME/.cache/sessions
when at least some of this information should be stored somewhere under
$HOME/.config/ instead.

Cache directories historically were intended for data which the program could
either regenerate or reload from external sources, so cache folders could
be deleted (or in my case not backed up) without data loss.

An example of this for "/var/cache" from the Filesystem Hierarchy Standard:

  "/var/cache is intended for cached data from applications. Such data is
   locally generated as a result of time-consuming I/O or calculation. The
   application must be able to regenerate or restore the data. Unlike
   /var/spool, the cached files can be deleted without data loss"

I do realize that FHS does not cover the $HOME/.cache directories however, 
their definition is consistent with the historical meaning and usage of 
"cache" directories. XDG which does define the standard for these directories
gives a much more vague definition of:

   "defines the base directory relative to which user specific non-essential
data files should be stored."

however, I would still interpret "non-essential" files as those which would
meet the FHS definition above.

As a saved session cannot be recovered by the program if $HOME/.cache is lost,
the user must manually rebuild the lost session layout (which in my case takes
about 10 minutes). This appears to violate the purpose of the $HOME/.cache
directory.


-- System Information:
Debian Release: 7.5
  APT prefers stable-updates
  APT policy: (500, 'stable-updates'), (500, 'stable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.9-0.bpo.1-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages xfce4 depends on:
ii  gtk2-engines-xfce  2.8.1-3
ii  orage  4.8.3-2
ii  thunar 1.2.3-4+b1
ii  xfce4-appfinder4.8.0-3
ii  xfce4-mixer4.8.0-3+b1
ii  xfce4-panel4.8.6-4
ii  xfce4-session  4.8.3-3
ii  xfce4-settings 4.8.3-2
ii  xfce4-utils4.8.3-2
ii  xfconf 4.8.1-1
ii  xfdesktop4 4.8.3-2
ii  xfwm4  4.8.3-2

Versions of packages xfce4 recommends:
ii  desktop-base  7.0.3
ii  tango-icon-theme  0.8.90-5
ii  thunar-volman 0.6.1-1
ii  xfce4-notifyd 0.2.2-2
ii  xorg  1:7.7+3~deb7u1

Versions of packages xfce4 suggests:
ii  xfce4-goodies  4.8.2
ii  xfprint4   4.6.1-3

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#739038: quitcount: Can't exit program

2014-02-15 Thread Shannon Dealy
Package: quitcount
Version: 2.0-1
Severity: normal

Dear Maintainer,

When quitcount starts, it shows up on my xfce4 panel.  When I right click on the
icon, a menu is displayed, when I click "quit", the program doesn't quit. It
appears the only way out is to use "kill" from the command line.

I tried running quitcount from an xterm and then clicking on "quit" this 
resulted in the following error being displayed in the xterm:

  (quitcount:5459): Gtk-CRITICAL **: gtk_main_quit: assertion `main_loops != 
NULL' failed




-- System Information:
Debian Release: 7.4
  APT prefers stable-updates
  APT policy: (500, 'stable-updates'), (500, 'stable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.9-0.bpo.1-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages quitcount depends on:
ii  libatk1.0-0 2.4.0-2
ii  libc6   2.13-38+deb7u1
ii  libcairo-gobject2   1.12.2-3
ii  libcairo2   1.12.2-3
ii  libfontconfig1  2.9.0-7.1
ii  libfreetype62.4.9-1.1
ii  libgdk-pixbuf2.0-0  2.26.1-1
ii  libglib2.0-02.33.12+really2.32.4-5
ii  libgtk-3-0  3.4.2-7
ii  libpango1.0-0   1.30.0-1

quitcount recommends no packages.

quitcount suggests no packages.

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#739037: quitcount: Auto-starts for all users

2014-02-15 Thread Shannon Dealy
Package: quitcount
Version: 2.0-1
Severity: normal

Dear Maintainer,

quitcount automatically starts up whenever any user logs in.  This ignores
the fact that many system users may not want quitcount to run, or more
importantly that no system user wants to run it.

quitcount was automatically installed on my system because I have the med-tools
task installed.  The next time I logged in, I was greeted with the quitcount
configuration screen.  This should not happen to any user who has not explicitly
chosen to have quitcount run for their account.



-- System Information:
Debian Release: 7.4
  APT prefers stable-updates
  APT policy: (500, 'stable-updates'), (500, 'stable')
Architecture: amd64 (x86_64)
Foreign Architectures: i386

Kernel: Linux 3.9-0.bpo.1-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages quitcount depends on:
ii  libatk1.0-0 2.4.0-2
ii  libc6   2.13-38+deb7u1
ii  libcairo-gobject2   1.12.2-3
ii  libcairo2   1.12.2-3
ii  libfontconfig1  2.9.0-7.1
ii  libfreetype62.4.9-1.1
ii  libgdk-pixbuf2.0-0  2.26.1-1
ii  libglib2.0-02.33.12+really2.32.4-5
ii  libgtk-3-0  3.4.2-7
ii  libpango1.0-0   1.30.0-1

quitcount recommends no packages.

quitcount suggests no packages.

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2013-08-15 Thread Shannon Dealy


Hi Moritz,

I have not seen this problem or any of the other iwl instabilities in a 
long time, however, I have been running with these option lines disabling 
11n functionality in order to achieve that stability:


options iwlagn 11n_disable50=1
options iwlwifi 11n_disable=1

since I am no longer running kernels which use iwlagn (currently running 
3.2.xx), I can't really speak to the original issue anymore.  Of course 
iwlwifi has had 11n related stability issues as well, however, I have 
never seen the:


  "MAC is in deep sleep"

bug (as far as I can recall) while running a newer kernel with iwlwifi 
instead of iwlagn.  Not sure how much code (if any) is common between 
iwlwifi and the older iwlagn module.


I suppose I should try enabling 11n again to see what happens (I had 
forgotten it was disabled).


The upgrade to wheezy is most notable for much shorter times required to 
make wireless connections, though I am not sure if it is the driver or 
network-manager responsible for that improvement.


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com


On Thu, 15 Aug 2013, Moritz Muehlenhoff wrote:


reassign 628444 src:linux
thanks

On Wed, Mar 14, 2012 at 08:37:17PM -0700, Shannon Dealy wrote:




Like others, the problems seemed to start around 2.6.39.


Thought I should note here, my system showed this problem with 2.6.36
through 2.6.39.  It seems to have stopped showing the problem (possibly
due to a memory upgrade many months ago), but it still has chronic
instability of the connection (drops out for varying intervals), and
often, network-manager doesn't seem aware of the failure - reloading the
driver isn't a reliable solution (sometimes appears to work, often does
not)


Does this bug still occur with more recent kernels/firmware, e.g. Wheezy?

Cheers,
   Moritz




--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#703851: scim: SCIM causes applications (gnucash, epiphany, ...) to segfault

2013-03-24 Thread Shannon Dealy
Package: scim
Version: 1.4.13-5
Severity: normal

Dear Maintainer,

SCIM causes segmentation faults in applications in two distinct senarios,
both related to having scim-tables-XX installed.  It appears that simply
having scim-modules-table installed alone is not sufficient to cause
either issue (though it may be where the problem lies).

1 - Having any scim-tables-XX installed will cause a segmentation fault
when exiting gnucash, epiphany browser and presumably other programs.
Most people wouldn't notice this unless they happened to launch the
program from the command line since (so far) I haven't seen any
side effects from this senario.

2 - Having any scim-tables-XX installed, run SCIM setup, select Global setup
under IMEngine, then disable any table(s) that are not in the "Other" list
(e.g. the "Japanese" group or "Nippon" individual table). At this point
attempting to launch gnucash or epiphany will segfault during startup of
these programs.

Things to note:

- The scim daemon does not need to be running when launching the program
  for these problems to occur.

- I have tried this with a variety of other tables installed instead of
  Japanese, the result is always the same.

- Purging and reinstalling scim and all its pieces had no effect

- Discarding my old SCIM configuration and configuring from scratch had
  no effect.

Work-around:

- Only install tables that are needed and don't disable any of them.
  This still leaves a potential problem for segfault on exit.


NOTE: I'm a bit wary of saying that the segfault at exit issue doesn't actually
cause any problems with the applications that use it, only that I haven't
seen any.  My guess is that it would depend on what order the application
code does its cleanup and shutdown.  If state saving / file updates are done
before wrapping up with SCIM they are probably fine, if SCIM is wrapped up
first, there is a chance of data loss.  So I would expect that it will
depend on the particular application as to whether or not there are
problems with using the obvious workaround above.


-- Package-specific info:
Related packages:
ii  libscim8c2a:i3 1.4.13-5 i386 library for SCIM platform
ii  scim   1.4.13-5 i386 smart common input method platfor
ii  scim-gtk-immod 1.4.13-5 i386 GTK+ input method module with SCI
ii  scim-modules-s 1.4.13-5 i386 socket modules for SCIM platform
ii  scim-modules-t 0.5.9-2  i386 generic tables IM engine module f
ii  scim-tables-ja 0.5.9-2  all  Japanese input method data tables

Related environment variables:
$XMODIFIERS=@im=SCIM
$GTK_IM_MODULE=scim
$QT_IM_MODULE=xim

Installed SCIM components:
/usr/lib/scim-1.0:
1.4.0
scim-helper-launcher
scim-helper-manager
scim-launcher
scim-panel-gtk

/usr/lib/scim-1.0/1.4.0:
Config
Filter
FrontEnd
Helper
IMEngine
SetupUI

/usr/lib/scim-1.0/1.4.0/Config:
simple.so

/usr/lib/scim-1.0/1.4.0/Filter:
sctc.so

/usr/lib/scim-1.0/1.4.0/FrontEnd:
x11.so

/usr/lib/scim-1.0/1.4.0/Helper:
setup.so

/usr/lib/scim-1.0/1.4.0/IMEngine:
rawcode.so
table.so

/usr/lib/scim-1.0/1.4.0/SetupUI:
aaa-frontend-setup.so
aaa-imengine-setup.so
panel-gtk-setup.so
table-imengine-setup.so

-- System Information:
Debian Release: 7.0
  APT prefers testing
  APT policy: (990, 'testing'), (500, 'testing-updates')
Architecture: i386 (x86_64)

Kernel: Linux 3.2.0-4-amd64 (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages scim depends on:
ii  libatk1.0-0 2.4.0-2
ii  libc6   2.13-38
ii  libcairo-gobject2   1.12.2-3
ii  libcairo2   1.12.2-3
ii  libgcc1 1:4.7.2-5
ii  libgdk-pixbuf2.0-0  2.26.1-1
ii  libglib2.0-02.33.12+really2.32.4-5
ii  libgtk-3-0  3.4.2-6
ii  libltdl72.4.2-1.1
ii  libpango1.0-0   1.30.0-1
ii  libscim8c2a 1.4.13-5
ii  libstdc++6  4.7.2-5
ii  libx11-62:1.5.0-1

Versions of packages scim recommends:
ii  im-switch  1.23
pn  scim-bridge-agent  
ii  scim-gtk-immodule  1.4.13-5

Versions of packages scim suggests:
pn  scim-anthy  
pn  scim-canna  
pn  scim-chewing
pn  scim-hangul 
pn  scim-m17n   
pn  scim-pinyin 
pn  scim-prime  
pn  scim-skk
pn  scim-tables-additional  
ii  scim-tables-ja  0.5.9-2
pn  scim-tables-ko  
pn  scim-tables-zh  
pn  scim-thai   
pn  scim-uim

-- debconf-show failed


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2012-03-17 Thread Shannon Dealy

On Fri, 16 Mar 2012, Bjørn Mork wrote:


Shannon Dealy  writes:


I created a file "/etc/modprobe.d/iwlagn.conf" and placed the
following line in it:

  options iwlagn 11n_disable=1 11n_disable50=1


Note that the 11n_disable50 options was removed in 3.0 and the iwlagn
module was renamed to iwlwifi in 3.2.


Unfortunate, since further testing shows that it is the 11n_disable50=1 
that matters.  Disabling either of the above options makes my link stable, 
and since 11n_disable=1 would effectively cause "11n_disable50" to happen 
as well, the implication is that the problem is in the code which is/was 
controlled by the 11n_disable50.


Would be nice if anyone can confirm if this fixes the deep sleep problem 
as well.



Which makes this workaround pretty much irrelevant to any current Debian
kernel as noone(?) has seen the bug in 2.6.32.


While it may not be relevant to you, it is completely relevant in that it:

  - provides developers with a narrowed range of places to look for the problem
in all versions of the kernel/driver code, including those were the
option was removed (they can easily check what part of the code it
previously disabled)

  - using just the 11n_disable option should still solve the stability
problems I am seeing and possibly the deep sleep as well (need
confirmation on that) for people running 3.x (though granted it is not
a great workaround given the performance hit)

  - I assume you are speaking with regard to 2.6.32 being the default
stable kernel so it won't affect people running stock Debian stable?
I don't recall seeing confirmation from anyone that the problem went
away by downgrading to 2.6.32.

Are you still using the  2.6.39-1 kernel you originally opened this bug 
against?

[snip]

Yes, though it has been rebuilt for debug tracing of the iwlagn driver. 
Unfortunately I can't afford the down time right now that a kernel 
upgrade beyond 2.6.39 would entail (other software would have to be 
upgraded and that might require that I write some code as well to 
deal with the kernel changes).  I also can't downgrade to 2.6.32 
(assuming that it would make any difference) as it doesn't support some of 
my hardware.


FWIW

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2012-03-15 Thread Shannon Dealy


Found these two pages discussing what appears to be the same problem:

   https://bugs.launchpad.net/ubuntu/+source/linux/+bug/575492
   http://ubuntuforums.org/showthread.php?t=1470820

in at least one case, it is definitely the same problem ("deep sleep" 
message shows up in the posted log).  Note: while there were Lenovo 
computers, there were a number of other brands also having problems. 
Wireless card in the discussion was the Intel 5100 AGN.


I created a file "/etc/modprobe.d/iwlagn.conf" and placed the following 
line in it:


  options iwlagn 11n_disable=1 11n_disable50=1

and reloaded my driver.  Wireless has been rock solid for 14 hours, no 
dropped packets, no long pauses, no mysterious disconnects.  Normally I 
would at least see short pauses, a few disconnects and 100's of dropped 
packets in this time frame.


Since I haven't been seeing the "deep sleep" problem lately, someone else 
will have to disable "N" on their system to test if this prevents that 
issue.


This doesn't really constitute a fix and it isn't a great work-around 
either given the major lost of network performance, but if it works for 
others, it would significantly narrow the range of possible causes for 
the problem.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2012-03-14 Thread Shannon Dealy




Like others, the problems seemed to start around 2.6.39.


Thought I should note here, my system showed this problem with 2.6.36 
through 2.6.39.  It seems to have stopped showing the problem (possibly 
due to a memory upgrade many months ago), but it still has chronic 
instability of the connection (drops out for varying intervals), and 
often, network-manager doesn't seem aware of the failure - reloading the 
driver isn't a reliable solution (sometimes appears to work, often does 
not)


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-03-07 Thread Shannon Dealy


Thought I should note that the addition of "pcie_aspm=off" to my command 
line has not cured my link problems, but it has changed the behavior of 
the problems.  It is now more likely to recover on its own without me 
reloading the driver and when I do reload the driver, it is far more 
likely to "fix" the problem for a while on the first or second try (in the 
past I often reloaded the driver six or more times to get it working again)


It also appears to be far more likely for the link to fail right when I 
begin using it.  This is a strange behavior where if I haven't been doing 
anything on the web and then attempt to use the web browser, the link 
fails at that point in time.  I leave a one second ping to my server 
running most of the time and usually have my email program open which 
maintains an imap connection to the server.  When I access the web browser 
and don't get a response from the web, a quick check shows the ping has 
stopped receiving responses, but that the email program has not detected a 
timeout on its link (which I believe takes 15 seconds of down time).  The 
implication is that web access from a relatively inactive state 
(just the ping and idle imap connection) actually kills the link for some 
reason.


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-03-07 Thread Shannon Dealy


Thought I should note that the addition of "pcie_aspm=off" to my command 
line has not cured my link problems, but it has changed the behavior of 
the problems.  It is now more likely to recover on its own without me 
reloading the driver and when I do reload the driver, it is far more 
likely to "fix" the problem for a while on the first or second try (in the 
past I often reloaded the driver six or more times to get it working again)


It also appears to be far more likely for the link to fail right when I 
begin using it.  This is a strange behavior where if I haven't been doing 
anything on the web and then attempt to use the web browser, the link 
fails at that point in time.  I leave a one second ping to my server 
running most of the time and usually have my email program open which 
maintains an imap connection to the server.  When I access the web browser 
and don't get a response from the web, a quick check shows the ping has 
stopped receiving responses, but that the email program has not detected a 
timeout on its link (which I believe takes 15 seconds of down time).  The 
implication is that web access from a relatively inactive state 
(just the ping and idle imap connection) actually kills the link for some 
reason.


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-29 Thread Shannon Dealy

On Thu, 1 Mar 2012, Ben Hutchings wrote:


On Wed, 2012-02-29 at 21:09 -0800, Shannon Dealy wrote:
[...]

Actually, a RAM upgrade can most definitely have an impact.  Replacing any
hardware on the bus changes bus loads, propagation delays and overall
timing of the bus.

[...]

Unfortunately, I'm not current on the hardware/buses involved here, so I
don't know how these potential issues might apply to this case.


I don't think RAM and PCI cards have ever been on the same bus.  And
with PCIe there is a separate link to each card.


Unfortunately, that doesn't actually isolate them, though it does usually 
reduce some of the sources of problems.  There are always interface chips 
tying them together (otherwise DMA transfers would not be possible) and 
the behavior of these chips will change depending on the load and 
temperature, so activity on the bus on one side of the chip can 
significantly affect the timing behavior of the chip on the other side
(I track down bugs like this for a living, just not on PC's in a very long 
time).  I have seen bugs in software show up due to a change in 
manufacturing processes used to produce an interface chip in the system 
and hardware glitches on an "isolated" bus that were due to transmission 
line effects that rippled right through the chip that was supposed to be 
isolating the two buses.  Of course there is always the old standby of 
noise passed between chips via the power supply traces.  There are many 
more examples.


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-29 Thread Shannon Dealy

On Wed, 29 Feb 2012, Juha Jäykkä wrote:


that the memory slots and mini-pcie slots are behind the same cover in
the X201 as well? Replugging it seems like a good idea...


I wish they were! They are in X4? and X30? but not X200s, where I had to
remove the keyboard and handrest to get to the card. Most annoying.


This is why I haven't tried it on my system.

Nevertheless, the RAM is quit directly on the opposite side of the 
motherboard

compared to the miniPCI slots (there are two: one is empty, I guess it is
meant for 3g), so I would not rule out the miniPCI card moving when addin RAM.
Especially since I now have almost 3 days of uptime without problems! This is
the record since my problems started except for disabling 802.11n and
powersave completely, which gave me about a week.

I will report back when I loose wifi again or in two weeks if I do not loose
it by then (in which case I will assume it was indeed loose in its socket).


You are forgetting that you also added:

   pcie_aspm=off

to your kernel boot command line.  Since I did the same thing and have 
also suddenly enjoyed quite stable WIFI for the first time, I would 
venture to guess that your problem was not the seating of the card but the 
"pcie_aspm" option setting.


If you want to confirm, you might want to remove that boot option and see 
if the problem returns.  If so, this would be the first really useful 
confirmation of a specific source for the "deep sleep" problem and 
provide a workaround for it that others can use, though someone who is 
familiar with this aspect of the Linux kernel will have to decide what 
this means in terms of hardware/software issues, what is causing the 
fault, and how to fix it.



Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-29 Thread Shannon Dealy

On Wed, 29 Feb 2012, Bjørn Mork wrote:


Juha Jäykkä  writes:


One thing just occurred to me: I added 2 GB of memory (total 4 GB) at about
the time these problems started. I cannot say for sure the problems did not
start earlier, but it is quite possible they started after! (Please do not

[snip]

This brings up an interesting point, while I have been having assorted 
other problems, my "deep sleep" issues may have stopped around the time 
that I upgraded to 8GB of RAM.



FWIW I am using the exact same card in an X301 with 8 GB memory and have
never seen any such problems. Don't think a RAM upgrade can have any
impact, except maybe by causing the card to move in its slot. I assume
that the memory slots and mini-pcie slots are behind the same cover in
the X201 as well? Replugging it seems like a good idea...

[snip]

Actually, a RAM upgrade can most definitely have an impact.  Replacing any 
hardware on the bus changes bus loads, propagation delays and overall 
timing of the bus.  These changes may be minor, but if the system is 
running near the limits of the specifications and/or any component on the 
bus violates the specs, then all sorts of flakey behavior can ensue.  Even 
if everyone plays by the rules, some aspects of bus timings are 
programmable on some hardware, so a software/firmware bug can also be a 
source of what would appear to be a "hardware" problems.


Unfortunately, I'm not current on the hardware/buses involved here, so I 
don't know how these potential issues might apply to this case.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-26 Thread Shannon Dealy

On Sun, 26 Feb 2012, Juha Jäykkä wrote:

[snip]

Shannon: FWIF, I have now rebooted with pci_aspm=off and we'll see what
happens.

Ben: I also replugged the wifi card.


I have also rebooted with pci_aspm=off, and while it has only been a few 
hours, I have not seen any of the usual irratic behavior that I am used to 
seeing from the wireless card.  Usually this is an ongoing source of 
irritation to me, though some times it works for a while without problems, 
so we will see if it lasts.


I would have replugged the wifi card a long time ago, but if I recall 
correctly, it is a little more complicated on this laptop than most I have 
owned, I don't have the time for it, and am a little wary of tinkering 
with my main computer when everything I am doing is time critical right 
now (maybe after I get out of school in a few months).



Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-26 Thread Shannon Dealy

On Fri, 24 Feb 2012, Juha Jäykkä wrote:

[snip]

I never had any issues with hibernate or suspend and certainly neither was
ever needed to trigger the bug. For example yesterday, I hit the bug after
about 3 hours without ever putting the laptop off my lap meanwhile.


It wasn't required to trigger the problem on my system either, but it 
seemed to happen a lot more often in the first few minutes after coming 
out of hibernation.


[snip]


What is VERY strange here is the fact that I did not have this problem before
going from 2.6.39 to 3.0.0, but not I have it even with 2.6.39! How can that
be?!? What else is involved with talking to the wifi card except the kernel
and the firmware? Do I need to downgrade iwconfig et al, too?

[snip]

On my system I could go for weeks without incident, and then have it 
happen three times in a couple hours.  It also appeared to be more 
prevalent with some versions of the kernel than others (though this may be 
an illusion due to the irratic nature of the problem).  How long did you 
run it on 2.6.39 previously?  Perhaps it would have happened with 2.6.39 
previously had you used it for a longer period?



Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-26 Thread Shannon Dealy

On Sun, 26 Feb 2012, Ben Hutchings wrote:

[snip]

No, wasn't even aware of its existance.  System shows it is currently at
"default" unfortunately, the debug kernel I am currently running has the
pcie_aspm regression which prevents it from being changed without
rebooting which is rather time consuming with my setup.  There doesn't
seem to be an obvious way to tell what the "default" is.

[...]

I'm not recommending that you set it.  I was concerned that forcing ASPM
on could possibly trigger this problem.


After looking up the setting, I had assumed that was your concern. 
However, it seems to me that if it is currently on, then forcing it to 
"off" might possibly help with the issues with this card.  Power 
management has caused a lot of issues on Linux over the years. 
Unfortunately, "default" doesn't tell me which state it is actually in.

Will give it a try next time I reboot.

FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-25 Thread Shannon Dealy

On Sun, 26 Feb 2012, Ben Hutchings wrote:


On Thu, 2012-02-23 at 17:59 -0800, Shannon Dealy wrote:
[...]

some point, outside software MUST be providing bogus information to the
driver.  I say this because after the deep sleep bug occcurs and the
hardware has been power cycled (through hibernate), the device driver and
firmware have been reloaded and right from the start they show the "deep
sleep" state again.


For a complete power-cycle, you may need to remove both the power cord
and the battery.  I don't think the Intel wireless cards have any
support for wake-on-WAN, but in general devices may still be partly
powered as long as any power source is connected to the system.


True enough, and I don't remember if I pulled the battery back when I was 
going after this problem (it has been almost a year since I did all my 
tests).  There is also a switch on the side that kills all RF devices 
(WLAN and Bluetooth) but I can't verify that it completely kills power to 
the devices.



The log messages you sent are indicative of a total failure of
communication with the card.  My suspicion would be that the hardware
has developed a fault, but it could be as simple as the card being
slightly loose in its slot.


Hardware failure might make sense except that it has always behaved this 
way, others have similar symptoms, and most importantly, regardless of 
what behaviors are occurring (there are a number of problems this card 
exhibits), while hibernate won't clear the "deep sleep" problem, rebooting 
the system (without power cycling) clears the problems EVERY time. 
Whatever the trigger (intermittent hardware fault or driver bug), it 
appears the kernel is getting into a state from which it is incapable of 
recovering.



Having said that, are you setting the pcie_aspm kernel parameter?


No, wasn't even aware of its existance.  System shows it is currently at 
"default" unfortunately, the debug kernel I am currently running has the 
pcie_aspm regression which prevents it from being changed without 
rebooting which is rather time consuming with my setup.  There doesn't 
seem to be an obvious way to tell what the "default" is.


I should note that I have not seen the "deep sleep" problem in many 
months, though it is hard to say when or if it has stopped since it often 
took weeks to show up and my usage patterns have changed considerably.  In 
particular I noticed that hibernate for my system seemed to be a 
significant contributor to triggering the bug since the problem often 
showed up in the first few minutes after bringing the computer out of 
hibernation.  This is not a requirement to trigger the problem, just 
significantly increased the odds.


NOTE: I stopped using hibernate in large part because of the "deep sleep" 
bug.  Unfortunately, the other problems of this card continue to plague me 
(chronic random communication failures being the big one).


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Info received (iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2012-02-23 Thread Shannon Dealy

On Wed, 22 Feb 2012, Juha Jäykkä wrote:


Replying to myself, really, but two new observations:


There were no problems in the 2.6-series. The bug occurs at least in the
Debian kernel versions 3.2.0-1-amd64, 3.0.0-2-amd64, and 3.0.0-1-amd64.


As much as I thought and hoped this was true, it is not. I just had the
problem with 2.6.39, so...

[snip]

As my original report noted, I have had the problem in everything from 
2.6.36 to 2.6.39 (I also think 2.6.32 but am not certain and didn't have 
the time to check).  I don't recall what I said about firmware, but have 
tried at least 5 different versions.  I have almost completely stopped 
using hibernate (suspend only) on my system and have not seen the problem 
in a long time and for my system at least, hibernate seems to be related 
to the number of incidents of the deep sleep bug.


While the firmware may play a role in the problem, at its core, there
are issues that must be occurring outside the firmware or even the iwlagn 
driver, namely a kernel bug or bug in a supporting driver - there is 
simply no way around this.  When a machine has been through a hibernation 
cycle and completely powered off with the driver unloaded before shutdown, 
it simply cannot come back up with the "deep sleep" problem still in place 
unless there is a bug in the kernel or some other driver involved.  At 
some point, outside software MUST be providing bogus information to the 
driver.  I say this because after the deep sleep bug occcurs and the 
hardware has been power cycled (through hibernate), the device driver and 
firmware have been reloaded and right from the start they show the "deep 
sleep" state again.  Everything related to the device has been completely 
reinitialized so they cannot have any "knowledge" of past history or 
status, so some piece of outside software is holding on to outdated 
information and not updating.  I demonstrated this repeatedly when I first 
started fighting with the problem.


I also have other stability issues with the driver, the most irritating 
of them is randomly "pausing" for periods of a few seconds up to several 
minutes where no data gets through, as well as assorted other symptoms 
consistent with race conditions between the driver and the kernel or 
other drivers.  Again, reloading the driver and/or firmware does not fix 
these problems, or only does so sporadically.


Based on this behavior and the fact that Juha appears to have similar 
hardware (though not the same model), and my previously noting that 
many of the people complaining on the internet seemed to be using Lenovo 
hardware, my recommendation to anyone who has time to investigate would be 
to look at the Linux driver(s) for the flavor of PCI interface bus these 
cards plug-in to and the particular chip sets used to implement this 
bus on the known Lenovo machines having the problem (x201i, x200, ...).
My guess is they are using the same hardware and therefore, the same 
interface driver.  I don't know where else the bogus device status 
information might come from or be stored, but I haven't been keeping up 
with the Linux kernel for quite a while.


Note: my contact at Intel has not responded to my request for status, so I 
don't know what if anything is happening at their end.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: Debian bug #628444

2012-02-15 Thread Shannon Dealy

On Sat, 11 Feb 2012, Juha Jäykkä wrote:


Hi Shannon!

What is the status of the fix from Intel? I have the same problem since
upgrading to 3.x series and it is VERY annoying - only way I can fix it is
rebooting the kernel, so it really seems a kernel bug. Hibernating the system
and doing a full power-off (unplug AC, remove battery, wait 60 seconds – this
should do it) the laptop does NOT fix it, so I concur: the kernel must retain
some bogus information somewhere since the hw has definitely lost its status
info.


Hi Juha,

Sorry, I haven't heard from him in a couple months.  I think both of us 
are pretty backed up with other things to deal with (as an embedded 
systems developer, I would have just tracked it down myself if I could 
spare the time).  One thing that seems to have made a difference for me is 
that I use suspend rather than hibernate these days.  This seems to give 
me much better system stability, particularly with the iwlagn driver, 
though it shows a number of quirks other than this bug, at least the other 
ones (so far) can be cured by simply waiting and/or reloading the device 
driver.  I'll send him a message and see what is happening.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#628444: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2011-11-25 Thread Shannon Dealy

On Wed, 23 Nov 2011, Jonathan Nieder wrote:


Shannon Dealy wrote:


A developer at Intel contacted me regarding this bug the other day (he was
following up on a similar bug report from another source) and I am working
with him on the problem (currently doing a debug build of the module to
collect data on what is happening).


That's good to hear.  Did it bear any fruit?  (Any test results or
public mailing list messages we can look at?)


Nothing so far, both he and I are very busy so it is a slow process, 
waiting for each of us to work the next step into our schedules. 
Currently I am waiting for his response to a set of data I captured from 
the driver surrounding one failure.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: linux-image-2.6.39-1-686-pae: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2011-09-17 Thread Shannon Dealy


A developer at Intel contacted me regarding this bug the other day (he 
was following up on a similar bug report from another source) and I 
am working with him on the problem (currently doing a debug build of the 
module to collect data on what is happening).


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: linux-image-2.6.39-1-686-pae: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2011-09-17 Thread Shannon Dealy

On Sat, 17 Sep 2011, maximilian attems wrote:


I wasn't aware that the 3.0 kernels were out until you sent this
message. Unfortunately, since this bug takes time to manifest, it
will require that I run with 3.0 as my default kernel, this means I
have to get vmware to run with it (since I use it heavily), and that
is usually a problem for bleeding edge kernels.  I will look into
it, but it may be a week or so before I can even try running a 3.0
kernel.


bleeding edge would be against linux-next, 3.1-rcX is in experimental.
that would be next target if problem persists.


Bleeding edge is anything which hasn't been in use long enough that there 
is still significant potential for major bugs.   Having had very long 
experience with the Linux kernel (and having fixed the bugs myself on 
several occasions), the 3.x kernels qualify.



Anyway according to your description you are loading binary blobs
into your OS, so you just voided our support and this bug report
can be closed away unless you can reproduce without.


1- Actually, VMWare provides the source code for their kernel interface
   code which uses loadable modules, but usually someone has to work up a
   source code patch for the latest versions of the kernel (< 6-9 months
   old) and then users have to patch and rebuild the kernel interface.  I
   have worked up patches myself on a few occasions, but I don't really
   have the time.

2- I wasn't running VMWare at the time I reported these problems, I use
   it only for customer software development and there weren't any
   customers back then.  Now I am rather busy.

So no, this bug report cannot be closed on that basis and as I stated 
before, even a cursory scan of the internet shows this is a common 
problem for people running similar hardware, so even if I were running 
"binary blobs", the bug should remain open.


NOTE: If one were to assume that VMWare were in any way relevant to this 
problem, then the bug should move from machine to machine as I migrate my 
development environment.  Instead, the bug showed up when I moved to a new 
machine with different hardware and device drivers, but the same kernel.

I upgraded to a variety of different kernels in the hope that one of them
would work and have settled on the current one I am running because it 
seems to have the problem less often.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#641845: upowerd: upowerd wastes power by writing to disk too often

2011-09-16 Thread Shannon Dealy
Package: upower
Version: 0.9.5-5
Severity: normal
File: upowerd


I was surprized to find that the reason my laptop disk keeps spinning up
is the "upowerd" power management daemon.  It is writing to its history files
at least a couple times a minute.  It seems to me that any program involved
in power management should be more careful in its use of power.  Ideally,
it would only push its data to disk when the disk is already spun up, possibly
with a maximum time limit of 20 minutes or when some maximum amount of buffered
data is reached, whichever comes first.  Alternatively, it should allow for
at least several minutes of data to buffer before flushing to disk.  Even
better would be if whichever of the above is implemented were configurable
(buffer size and time limit) in a run-time configuration file or as a command
line option.


-- System Information:
Debian Release: 6.0.2
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'stable-updates')
Architecture: i386 (i686)

Kernel: Linux 2.6.36-trunk-686-bigmem (SMP w/4 CPU cores)
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages upower depends on:
ii  dbus   1.2.24-4+squeeze1 simple interprocess messaging syst
ii  libc6  2.11.2-10 Embedded GNU C Library: Shared lib
ii  libdbus-1-31.2.24-4+squeeze1 simple interprocess messaging syst
ii  libdbus-glib-1-2   0.88-2.1  simple interprocess messaging syst
ii  libglib2.0-0   2.24.2-1  The GLib library of C routines
ii  libgudev-1.0-0 164-3 GObject-based wrapper library for 
ii  libimobiledevice1  1.0.2-1   Library for communicating with the
ii  libplist1  1.3-2 Library for handling Apple binary 
ii  libpolkit-gobject-1-0  0.96-4PolicyKit Authorization API
ii  libupower-glib10.9.5-5   abstraction for power management -
ii  libusb-1.0-0   2:1.0.8-2 userspace USB programming library
ii  udev   164-3 /dev/ and hotplug management daemo

Versions of packages upower recommends:
ii  pm-utils  1.3.0-3utilities and scripts for power ma
ii  policykit-1   0.96-4 framework for managing administrat

upower suggests no packages.

-- no debconf information



-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: linux-image-2.6.39-1-686-pae: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2011-09-15 Thread Shannon Dealy

On Tue, 13 Sep 2011, maximilian attems wrote:


tags 628444 moreinfo
stop


This bug actually applies to all Debian kernels I have tried including
2.6.36, 2.6.37, 2.6.38, and 2.6,39


Did you try since newer 3.0 images, how are they performing?

thank you


I wasn't aware that the 3.0 kernels were out until you sent this message. 
Unfortunately, since this bug takes time to manifest, it will require that 
I run with 3.0 as my default kernel, this means I have to get vmware to 
run with it (since I use it heavily), and that is usually a problem for 
bleeding edge kernels.  I will look into it, but it may be a week or so 
before I can even try running a 3.0 kernel.


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#637009: xdrawchem: Wrong path to "help" from molecule info window

2011-08-07 Thread Shannon Dealy
Package: xdrawchem
Version: 1.9.9-4.1
Severity: normal


If you click the "help" button from the "molecule info" window, it attempts
to display the file:

   /usr/share/xdrawchem/doc/molinfo.html

which results in a blank display.  The above path is incorrect and should
instead be:

   /usr/share/doc/xdrawchem/html/molinfo.html

which is where the file is currently located.

-- System Information:
Debian Release: 6.0.2
  APT prefers stable
  APT policy: (990, 'stable'), (500, 'stable-updates')
Architecture: i386 (i686)

Kernel: Linux 2.6.36-trunk-686-bigmem (SMP w/4 CPU cores)
Locale: LANG=en_US.utf8, LC_CTYPE=en_US.utf8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages xdrawchem depends on:
ii  libc6  2.11.2-10 Embedded GNU C Library: Shared lib
ii  libgcc11:4.4.5-8 GCC support library
ii  libopenbabel3  2.2.3-1+b1Chemical toolbox library
ii  libqt3-mt  3:3.3.8b-7+b1 Qt GUI Library (Threaded runtime v
ii  libstdc++6 4.4.5-8   The GNU Standard C++ Library v3

xdrawchem recommends no packages.

xdrawchem suggests no packages.

-- no debconf information



-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#607673: GTK_IM_MODULE setting

2011-06-28 Thread Shannon Dealy


Was changing window managers and discovered that the problem with gnucash 
and the scim daemon on my system is specific to the setting of 
GTK_IM_MODULE.  If it is set to "xim" gnucash editing is broken (arrows, 
delete/backspace), if it is set to "scim" then it appears to work fine.


My understanding is that scim is supposed to be able to work with xim (and 
it apparently does for my other apps), though I don't recall why I set it 
up this way years ago as it does not appear to be necessary for my current 
usage.  In any case, switching so:


   GTK_IM_MODULE="scim"

appears to solve my problem, though the bug is still there, this appears 
to make it a non-issue for me.  Perhaps this may give some clues to make 
it easier to track down the problem.


FWIW.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: Acknowledgement (linux-image-2.6.39-1-686-pae: iwlagn - "MAC is in deep sleep", cannot restore wifi operation)

2011-05-28 Thread Shannon Dealy


Oops, the previously closed bug I mentioned should have been:

   #589353
   http://bugs.debian.org/cgi-bin/bugreport.cgi?bug=589353

NOT #589383

Sorry for the confusion.

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#628444: linux-image-2.6.39-1-686-pae: iwlagn - "MAC is in deep sleep", cannot restore wifi operation

2011-05-28 Thread Shannon Dealy
Package: linux-2.6
Version: 2.6.39-1
Severity: important


After some arbitrary period of use, wifi fails and errors show up in syslog
indicating "MAC is in deep sleep" after which nothing can be done to restore
wifi functionality other than rebooting the system.  In most (possibly all)
cases, the failure occurs after the laptop has been suspended and resumed
at least once, though the actual occurance may not be immediately after
resuming, it often occurs within a few minutes after resume.

Unloading and reloading the iwlagn driver and the drivers it relies on have
no effect.

Speculation:
   there is a mechanical switch on the side of the laptop which
   cuts power to both bluetooth and wifi on this machine.  Unloading iwlagn
   turning off the switch waiting ten seconds, then turning it on and reloading 
   iwlagn also has no effect.  Assuming this switch actually is a hardware
   power off of the devices, this should reset the device and a reload of
   the driver should restore functionality.  Since this does not occur, I
   see only two possibilities, either the "power switch" does not completely
   power off the device, or the kernel is retaining some bogus information
   (possibly placed there by iwlagn) about the device in spite of the driver
   being unloaded and reloaded.  Even if it doesn't prevent the problem,
   simply curing this issue so the device can be restarted without rebooting
   would be a major improvement.

I have tried a number of different versions of the firmware as well as
different kernels, all combinations have the problem, though some seem to
be more stable than others.  In some combinations it make take one to two
weeks of usage with perhaps a dozen suspend resume cycles per day, in others
it seems to happen every day.

I have noted in my search of the net that Lenovo laptops seem to be most
(possibly all, I wasn't paying close attention) of the cases where this
bug occurs.

My system is a Lenovo x201i, the wifi adapter was listed as an:

"Intel Centrino Advanced-N 6200 (2x2 AGN)"


This bug actually applies to all Debian kernels I have tried including 2.6.36,
2.6.37, 2.6.38, and 2.6,39 (various versions of each and I think it included
earlier ones, but I don't remember which ones and no longer have them
installed).

NOTE: this appears to be the same as closed bug 589383 which was closed
upstream and tagged "unreproducible".  Even the most cursory search of 
the net shows there are lots of people having no problems reproducing it.
It strikes me as inappropriate to close a bug as unreproducible when it
is quite clear that many people are seeing it.


May 28 15:19:12 nashapur kernel: [16275.165421] iwlagn :02:00.0: Error 
sending REPLY_ADD_STA: time out after 500ms.
May 28 15:19:12 nashapur kernel: [16275.165428] HW problem - can not stop rx 
aggregation for tid 0
May 28 15:19:12 nashapur ifd[1923]: executing: 
'/usr/share/laptop-net/link-change wlan0 unmanaged up,running,connected 
up,running,disconnected'
May 28 15:19:12 nashapur kernel: [16275.664639] iwlagn :02:00.0: Error 
sending REPLY_QOS_PARAM: time out after 500ms.
May 28 15:19:12 nashapur kernel: [16275.664644] iwlagn :02:00.0: Failed to 
update QoS
May 28 15:19:12 nashapur kernel: [16275.748419] iwlagn :02:00.0: Queue 2 
stuck for 2000 ms.
May 28 15:19:12 nashapur kernel: [16275.748429] iwlagn :02:00.0: On demand 
firmware reload
May 28 15:19:13 nashapur kernel: [16276.163702] iwlagn :02:00.0: Error 
sending REPLY_RXON: time out after 500ms.
May 28 15:19:13 nashapur kernel: [16276.163709] iwlagn :02:00.0: Error 
clearing ASSOC_MSK on BSS (-110)
May 28 15:19:13 nashapur kernel: [16276.180377] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.196983] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.213584] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.230200] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.246801] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.263411] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.280011] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.296612] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.313214] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.329830] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.346433] iwlagn :02:00.0: MAC is in 
deep sleep!.  CSR_GP_CNTRL = 0x
May 28 15:19:13 nashapur kernel: [16276.363037]

Bug#607673: gnucash: Left arrow, right arrow and backspace keys don't work in register

2011-01-27 Thread Shannon Dealy

On Mon, 24 Jan 2011, Micha Lenk wrote:


Hi Shannon,

You wrote:

So the problem would appear to be SCIM related, but (not knowing the

[snip]

Gnucash 2.4.0 is now available in experimental. This new upstream
version can be considered rather stable, but was not uploaded to
experimental due to the freeze for Debian Squeeze. Would you mind to
install the package from experimental and try-out whether your problem
persists?


Hi Micha,

No change.  Installed gnucash from experimental and still can't edit the 
register (left/right arrows and backspace key) as long as SCIM is running 
(even though it is not active).  Terminate the SCIM daemon and immediately 
register editing works correctly, restart the daemon and editing is broken 
again.  This worked fine in Lenny, just not in squeeze or the current 
experimental.  So far I have not seen this problem with any other program, 
so it still appears that gnucash is the problem


Regards,

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#607673: gnucash: Left arrow, right arrow and backspace keys don't work in register

2010-12-21 Thread Shannon Dealy


Hi Micha,

Your idea that it was SCIM related was correct, however, it wasn't the 
patch that is the problem.


I tried rebuilding and installing gnucash as you suggested without the 
"10_fix_broken_SCIM_input_bug_587298" patch and it had no effect.
However, I do run with SCIM installed (for use with other programs), and 
terminating it allowed both versions of gnucash to start working 
correctly (with and without the above patch).  Restart the background 
daemon and gnucash immediately stops working correctly.


So the problem would appear to be SCIM related, but (not knowing the 
internals of gnucash), I do find it puzzling that the problem only 
affects transaction editing (register and scheduled transactions), but not 
any other dialogs that I have tried.  I would have expected that all 
keyboard input to the program would behave the same.


Regards,

Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com



--
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#607673: gnucash: Left arrow, right arrow and backspace keys don't work in register

2010-12-20 Thread Shannon Dealy
Package: gnucash
Version: 2.2.9-10
Severity: important


I just upgraded my system to squeeze and in gnucash I am no longer able to use
the left-arrow, right-arrow, or backspace keys to edit transactions or register
entries.  Other keys that I don't normally use for editing may be broken as
well.  Up-arrow and down-arrow keys work as expected for scrolling through the 
register.

These keys all work in every dialog box of gnucash that I have tried and all
other applications that I am using, just the main register and the scheduled
transaction editor windows where checks and other transactions are entered
or modified has the problem of them being non-functional.  This is a major
problem for data entry as the only way I have found to correct an error is to
delete the transaction and re-enter it from scratch.

As I found it hard to believe that a problem like this would not have been
noticed already by others, I have tried:

   starting the program after deleting my .gnucash directory

   purging/reinstalling gnucash (which caused about 20 other packages to
   also be removed/reinstalled)

   rebooting the system

assuming that it might be caused by some form of local corruption of my
system, none of these had any effect.


-- System Information:
Debian Release: squeeze/sid
  APT prefers testing
  APT policy: (990, 'testing'), (500, 'unstable')
Architecture: i386 (i686)

Kernel: Linux 2.6.36-trunk-686-bigmem (SMP w/4 CPU cores)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/dash

Versions of packages gnucash depends on:
ii  gconf2  2.28.1-6 GNOME configuration database syste
ii  gnucash-common  2.2.9-10 A personal finance and money track
ii  guile-1.6-libs  1.6.8-10 Main Guile libraries
ii  guile-1.6-slib  1.6.8-10 Guile SLIB support
ii  libaqbanking29  4.2.4-2  library for online banking applica
ii  libart-2.0-22.3.21-1 Library of functions for 2D graphi
ii  libatk1.0-0 1.30.0-1 The ATK accessibility toolkit
ii  libbonobo2-02.24.3-1 Bonobo CORBA interfaces library
ii  libbonoboui2-0  2.24.3-1 The Bonobo UI library
ii  libc6   2.11.2-7 Embedded GNU C Library: Shared lib
ii  libcairo2   1.8.10-6 The Cairo 2D vector graphics libra
ii  libcrypt-ssleay-perl0.57-2   Support for https protocol in LWP
ii  libdate-manip-perl  6.11-1   module for manipulating dates
ii  libenchant1c2a  1.6.0-1  a wrapper library for various spel
ii  libfinance-quote-perl   1.17-1   Perl module for retrieving stock q
ii  libfontconfig1  2.8.0-2.1generic font configuration library
ii  libfreetype62.4.2-2.1FreeType 2 font engine, shared lib
ii  libgconf2-4 2.28.1-6 GNOME configuration database syste
ii  libglade2-0 1:2.6.4-1library to load .glade files at ru
ii  libglib2.0-02.24.2-1 The GLib library of C routines
ii  libgnome2-0 2.30.0-1 The GNOME library - runtime files
ii  libgnomecanvas2-0   2.30.1-1 A powerful object-oriented display
ii  libgnomeui-02.24.3-1 The GNOME libraries (User Interfac
ii  libgnomevfs2-0  1:2.24.3-1   GNOME Virtual File System (runtime
ii  libgoffice-0.8-80.8.12-1 Document centric objects library -
ii  libgtk2.0-0 2.20.1-2 The GTK+ graphical user interface 
ii  libgtkhtml3.14-19   3.30.3-1 HTML rendering/editing library - r
ii  libguile-ltdl-1 1.6.8-10 Guile's patched version of libtool
ii  libgwenhywfar47 3.11.3-1 OS abstraction layer
ii  libice6 2:1.0.6-2X11 Inter-Client Exchange library
ii  libktoblzcheck1c2a  1.28-1   library for verification of accoun
ii  libltdl72.2.6b-2 A system independent dlopen wrappe
ii  libofx4 1:0.9.0-3library to support Open Financial 
ii  liborbit2   1:2.14.18-0.1libraries for ORBit2 - a CORBA ORB
ii  libpango1.0-0   1.28.3-1 Layout and rendering of internatio
ii  libpopt01.16-1   lib for parsing cmdline parameters
ii  libqthreads-12  1.6.8-10 QuickThreads library for Guile
ii  libsm6  2:1.1.1-1X11 Session Management library
ii  libx11-62:1.3.3-4X11 client-side library
ii  libxml2 2.7.8.dfsg-1 GNOME XML library
ii  slib3b1-3.1  Portable Scheme library
ii  zlib1g  1:1.2.3.4.dfsg-3 compression library - runtime

Versions of packages gnucash recommends:
ii  gnucash-docs  2.2.0-3Documentation for gnucash, a perso

gnucash suggests no packages.

-- no debconf information

Bug#603770: flashplugin-nonfree: Out of date checksum again

2010-11-16 Thread Shannon Dealy
Package: flashplugin-nonfree
Version: 1:2.8.2
Severity: grave
Justification: renders package unusable


Adobe has again updated their flash player version 10 (now dated Nov. 11, 2010)

Attempting to install gives the following error:

   ERROR: sha512sum rejected install_flash_player_10_linux.tar.gz
   More information might be available at:
 http://wiki.debian.org/FlashPlayer

For this package to be usable for any period of time, it needs to not use 
a fixed built-in checksum (which I assume is the source of the error).
Otherwise, it could be broken again by Adobe doing a new release the day
after this is fixed, or the day of the next Debian release for that matter.

Possibly it could give a warning on checksum error and allow the user to
decide whether or not to install rather than failing outright.  I realize
there are potential problems with this approach too, but the current approach
is badly broken and this might make it less of a problem.  FWIW.

-- Package-specific info:
Debian version: squeeze/sid
Architecture: i386
Package version: 1:2.8.2
MD5 checksums:
md5sum: /var/cache/flashplugin-nonfree/*: No such file or directory
md5sum: /usr/lib/flashplugin-nonfree/libflashplayer.so: No such file or 
directory
Alternatives:
update-alternatives: error: no alternatives for flash-mozilla.so.
ls: cannot access /usr/lib/mozilla/plugins/flash-mozilla.so: No such 
file or directory
/usr/lib/mozilla/plugins/flash-mozilla.so: ERROR: cannot open 
`/usr/lib/mozilla/plugins/flash-mozilla.so' (No such file or directory)

-- System Information:
Debian Release: squeeze/sid
  APT prefers testing
  APT policy: (990, 'testing'), (500, 'unstable'), (500, 'stable')
Architecture: i386 (i686)

Kernel: Linux 2.6.32-5-686 (SMP w/1 CPU core)
Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8)
Shell: /bin/sh linked to /bin/bash

Versions of packages flashplugin-nonfree depends on:
ii  debconf [debconf-2.0] 1.5.36 Debian configuration management sy
ii  gnupg 1.4.10-4   GNU privacy guard - a free PGP rep
ii  libatk1.0-0   1.30.0-1   The ATK accessibility toolkit
ii  libcairo2 1.8.10-6   The Cairo 2D vector graphics libra
ii  libcurl3-gnutls   7.21.0-1   Multi-protocol file transfer libra
ii  libfontconfig12.8.0-2.1  generic font configuration library
ii  libfreetype6  2.4.2-1FreeType 2 font engine, shared lib
ii  libgcc1   1:4.4.5-6  GCC support library
ii  libglib2.0-0  2.24.2-1   The GLib library of C routines
ii  libgtk2.0-0   2.20.1-2   The GTK+ graphical user interface 
ii  libnspr4-0d   4.8.6-1NetScape Portable Runtime Library
ii  libnss3-1d3.12.8-1   Network Security Service libraries
ii  libpango1.0-0 1.28.3-1   Layout and rendering of internatio
ii  libstdc++64.4.5-6The GNU Standard C++ Library v3
ii  libx11-6  2:1.3.3-3  X11 client-side library
ii  libxext6  2:1.1.2-1  X11 miscellaneous extension librar
ii  libxt61:1.0.7-1  X11 toolkit intrinsics library
ii  wget  1.12-2.1   retrieves files from the web

flashplugin-nonfree recommends no packages.

Versions of packages flashplugin-nonfree suggests:
pn  flashplugin-nonfree-extrasoun  (no description available)
ii  iceweasel 3.5.15-1   Web browser based on Firefox
pn  konqueror-nsplugins(no description available)
ii  msttcorefonts 2.7transitional dummy package
ii  ttf-dejavu2.31-1 Metapackage to pull in ttf-dejavu-
ii  ttf-mscorefonts-installer [ms 3.3Installer for Microsoft TrueType c
pn  ttf-xfree86-nonfree(no description available)
ii  x-ttcidfont-conf  32 TrueType and CID fonts configurati

-- debconf information:
  flashplugin-nonfree/httpget: false
  flashplugin-nonfree/not_exist:
  flashplugin-nonfree/local:



-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Bug#592434: xserver-xorg-core: Broken GLX causes flash to crash when switching to full screen

2010-08-11 Thread Shannon Dealy

On Wed, 11 Aug 2010, Michel Dänzer wrote:


On Mon, 2010-08-09 at 19:39 -0700, Shannon Dealy wrote:



(II) LoadModule: "glx"
(II) Loading /usr/lib/xorg/modules/extensions/libglx.so
(II) Module glx: vendor="FireGL - ATI Technologies Inc."
  compiled for 7.5.0, module version = 1.0.0
(II) Loading extension GLX


Looks like your /usr/lib/xorg/modules/extensions/libglx.so is from an
fglrx driver installation, which probably can't work with other drivers.
Try fixing that.


Not sure how that ended up installed on the system, but the dist-upgrade 
to testing was not clean/easy and dependency chaos reigned for a while. 
While removing the fglrx packages now allows glx to run without crashing 
flash in full screen mode (thanks), this still leaves the issue of why is 
X not aborting on what should in my view be a fatal error at startup?


Shannon C. Dealy  |   DeaTech Research Inc.
de...@deatech.com |  - Custom Software Development -
Phone: (800) 467-5820 |  - Natural Building Instruction -
   or: (541) 929-4089 |  www.deatech.com

Bug#592434: xserver-xorg-core: Broken GLX causes flash to crash when switching to full screen

2010-08-09 Thread Shannon Dealy
Package: xserver-xorg-core
Version: 2:1.7.7-3
Severity: important


When using flash under firefox or chromium, switching it to full screen mode
causes flash to segfault.

I just upgraded this computer to testing, however, both the adobe flash plugin
and chromium were already running the latest versions so they were not upgraded.
Both were working fine before the upgrade.

I found the following error in the Xorg log file:

   (EE) GLX error: Can not get required symbols.

Searching the net I found reference to this error with a different X driver
having a possible fix by setting in the monitor section:

   option  "DPMS"  "true"

this had no effect.

Finally in the Module section of xorg.conf I changed:

   Load   "glx"
to
   Disable "glx"

which resolved the immediate problem, except of course that glx is still broken.

As I see it, there are a few issues here which should be resolved:

  1 - The cause of the "(EE) GLX error: Can not get required symbols." error
  needs to be fixed.

  2 - X should abort when an error like this occurs rather than continuing
  to run with what is apparently a corrupt environment.

  3 - If X is going to continue to run after an error like this, it needs
  to cleanly remove the offending module.  This clearly did not happen,
  otherwise, there would be no difference between running flash with
  glx failing to load correctly and glx disabled.

As a copy of the full log file is included below from the successful run, I am
including here the diffs between the log files from successful and failed
sessions:

16c16
< (==) Log file: "/var/log/Xorg.0.log", Time: Mon Aug  9 19:29:07 2010
---
> (==) Log file: "/var/log/Xorg.0.log", Time: Mon Aug  9 19:13:23 2010
57c57
< (++) using VT number 9
---
> (++) using VT number 8
62d61
< (WW) "glx" will not be loaded unless you've specified it to be loaded 
elsewhere.
65c64
< (II) "glx" will be loaded even though the default is to disable it.
---
> (II) "glx" will be loaded. This was enabled by default and also specified in 
> the config file.
85a85,89
> (II) LoadModule: "glx"
> (II) Loading /usr/lib/xorg/modules/extensions/libglx.so
> (II) Module glx: vendor="FireGL - ATI Technologies Inc."
>   compiled for 7.5.0, module version = 1.0.0
> (II) Loading extension GLX
335a340
> (EE) GLX error: Can not get required symbols.




-- Package-specific info:
/var/lib/x11/X.roster does not exist.

/var/lib/x11/X.md5sum does not exist.

X server symlink status:
lrwxrwxrwx 1 root root 13 Oct 16  2006 /etc/X11/X -> /usr/bin/Xorg
-rwxr-xr-x 1 root root 1725304 Jul 15 09:15 /usr/bin/Xorg

/var/lib/x11/xorg.conf.roster does not exist.

VGA-compatible devices on PCI bus:
00:02.0 VGA compatible controller: Intel Corporation Mobile 915GM/GMS/910GML 
Express Graphics Controller (rev 03)

/var/lib/x11/xorg.conf.md5sum does not exist.

Xorg X server configuration file status:
-rw-r--r-- 1 root root 2586 Aug  9 18:09 /etc/X11/xorg.conf

Contents of /etc/X11/xorg.conf:
# xorg.conf.dpkg-new (Xorg X Window System server configuration file)
#
# This file was generated by dexconf, the Debian X Configuration tool, using
# values from the debconf database.
#
# Edit this file with caution, and see the xorg.conf.dpkg-new manual page.
# (Type "man xorg.conf.dpkg-new" at the shell prompt.)
#
# This file is automatically updated on xserver-xorg package upgrades *only*
# if it has not been modified since the last upgrade of the xserver-xorg
# package.
#
# If you have edited this file but would like it to be automatically updated
# again, run the following commands as root:
#
#   cp /etc/X11/xorg.conf.dpkg-new /etc/X11/xorg.conf.dpkg-new.custom
#   md5sum /etc/X11/xorg.conf.dpkg-new 
>/var/lib/xfree86/xorg.conf.dpkg-new.md5sum
#   dpkg-reconfigure xserver-xorg

Section "Files"
FontPath"/usr/share/fonts/X11/misc"
FontPath"/usr/share/fonts/X11/cyrillic"
FontPath"/usr/share/fonts/X11/100dpi/:unscaled"
FontPath"/usr/share/fonts/X11/75dpi/:unscaled"
FontPath"/usr/share/fonts/X11/Type1"
FontPath"/usr/share/fonts/X11/100dpi"
FontPath"/usr/share/fonts/X11/75dpi"
# path to defoma fonts
FontPath"/var/lib/defoma/x-ttcidfont-conf.d/dirs/TrueType"
EndSection

Section "Module"
Load"i2c"
Load"bitmap"
Load"dbe"
Load"ddc"
Load"dri"
Disable "glx"
#   Load"glx"
Load"extmod"
Load"freetype"
Load"int10"
Load"record"
Load"v4l"
Load"vbe"
EndSection

Section "InputDevice"
Identifier  "Generic Keyboard"
Driver  "kbd"
Option  "CoreKeyboard"
Option  "XkbRules"  "xorg"
Option  "XkbModel"  "pc104"
Option  "XkbLayout" "us"
EndSection

Section "InputDevice"
Identifier  "Configured Mouse"
 

Bug#537968: gnome-mount: Doesn't auto mount partitions greater than 1 TB in size

2009-07-21 Thread Shannon Dealy

Package: gnome-mount
Version: 0.7-2
Severity: normal


I have a 1.5 terabyte USB hard drive, if I partition it with less than 
one terabyte of data in a partion, it is automatically mounted when I 
plug it into the computer, if I cross the one terabyte boundary with the 
partition size, it won't automatically mount (though I can mount it 
manually without problems).  I tried this with both an ext3 file system
and an encrypted setup (my original config) which prompts for the 
pass phrase only if the partition is less than one terabyte, otherwise 
there is no prompt and nothing is mounted.  I haven't found anything 
that looks like an error message relating to this, but I am a little 
fuzzy on the exact chain of events between HAL and gnome-mount, so I may 
not be looking in the right place.


cfdisk partitioning:

   This works:
  size entered: 1099510 MB, actual size displayed: 1099506.08 MB

   This fails (as does any larger value):
  size entered: 1099511 MB, actual size displayed: 1099514.31 MB

the exact power of 2 definition of "one terabyte" falls between the 
above two values, so I assume it is in some way significant.


syslog output after connecting the drive:

Jul 21 18:19:00 nashapur kernel: [19786.993063] usb 1-2: new high speed USB 
device using ehci_hcd and address 13
Jul 21 18:19:00 nashapur kernel: [19787.142662] usb 1-2: New USB device found, 
idVendor=0bc2, idProduct=3001
Jul 21 18:19:00 nashapur kernel: [19787.142672] usb 1-2: New USB device 
strings: Mfr=1, Product=2, SerialNumber=3
Jul 21 18:19:00 nashapur kernel: [19787.142680] usb 1-2: Product: FreeAgent
Jul 21 18:19:00 nashapur kernel: [19787.142685] usb 1-2: Manufacturer: Seagate
Jul 21 18:19:00 nashapur kernel: [19787.142690] usb 1-2: SerialNumber: 2GEV3WR1
Jul 21 18:19:00 nashapur kernel: [19787.142939] usb 1-2: configuration #1 
chosen from 1 choice
Jul 21 18:19:00 nashapur kernel: [19787.180655] scsi10 : SCSI emulation for USB 
Mass Storage devices
Jul 21 18:19:00 nashapur NetworkManager:  [1248225540.458861] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1').
Jul 21 18:19:00 nashapur kernel: [19787.248678] usb-storage: device found at 13
Jul 21 18:19:00 nashapur kernel: [19787.248685] usb-storage: waiting for device 
to settle before scanning
Jul 21 18:19:00 nashapur chipcardd[4492]: devicemanager.c: 3373: Changes in 
hardware list
Jul 21 18:19:00 nashapur NetworkManager:  [1248225540.664732] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1_usbraw').
Jul 21 18:19:00 nashapur NetworkManager:  [1248225540.948645] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1_if0').
Jul 21 18:19:00 nashapur NetworkManager:  [1248225540.969204] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1_if0_scsi_host').
Jul 21 18:19:05 nashapur kernel: [19792.253791] usb-storage: device scan 
complete
Jul 21 18:19:05 nashapur kernel: [19792.255346] scsi 10:0:0:0: Direct-Access
 Seagate  FreeAgent102C PQ: 0 ANSI: 4
Jul 21 18:19:05 nashapur NetworkManager:  [1248225545.755447] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1_if0_scsi_host_0').
Jul 21 18:19:05 nashapur NetworkManager:  [1248225545.763170] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/usb_device_bc2_3001_2GEV3WR1_if0_scsi_host_0_scsi_device_lun0').
Jul 21 18:19:16 nashapur kernel: [19802.994361] sd 10:0:0:0: [sdb] 2930277168 
512-byte hardware sectors: (1.50 TB/1.36 TiB)
Jul 21 18:19:16 nashapur kernel: [19802.995706] sd 10:0:0:0: [sdb] Write 
Protect is off
Jul 21 18:19:16 nashapur kernel: [19802.995719] sd 10:0:0:0: [sdb] Mode Sense: 
1c 00 00 00
Jul 21 18:19:16 nashapur kernel: [19802.995728] sd 10:0:0:0: [sdb] Assuming 
drive cache: write through
Jul 21 18:19:16 nashapur kernel: [19802.997822] sd 10:0:0:0: [sdb] 2930277168 
512-byte hardware sectors: (1.50 TB/1.36 TiB)
Jul 21 18:19:16 nashapur kernel: [19802.999168] sd 10:0:0:0: [sdb] Write 
Protect is off
Jul 21 18:19:16 nashapur kernel: [19802.999177] sd 10:0:0:0: [sdb] Mode Sense: 
1c 00 00 00
Jul 21 18:19:16 nashapur kernel: [19802.999183] sd 10:0:0:0: [sdb] Assuming 
drive cache: write through
Jul 21 18:19:16 nashapur kernel: [19802.999195]  sdb: sdb1
Jul 21 18:19:16 nashapur kernel: [19803.008167] sd 10:0:0:0: [sdb] Attached 
SCSI disk
Jul 21 18:19:16 nashapur NetworkManager:  [1248225556.995099] 
nm_hal_device_added(): New device added (hal udi is 
'/org/freedesktop/Hal/devices/storage_serial_Seagate_FreeAgent_2GEV3WR1_0_0').
Jul 21 18:19:21 nashapur kernel: [19808.353084] ===>rt_ioctl_giwscan. 1(1) BSS 
returned, data->length = 83
Jul 21 18:19:26 nashapur kernel: [19813.340151] ===>rt_ioctl_giwscan. 1(1) BSS 
returned, data->length = 83


-- System Informatio

Bug#355637: mailman: Stale lock files break administrative web interface

2006-03-06 Thread Shannon Dealy
Package: mailman
Version: 2.1.5-7
Severity: normal


Under some circumstances (presumably mailman software or system crashes),
list specific stale lock files are left in the directory /var/lib/mailman/locks
this can permanently prevent administrative login for that specific list until
the lock file(s) are removed.  It also seems to result in irratic behavior for
some other aspects of the list software (heavy cpu loading, refusal of qrunner
process to cleanly shutdown requiring a hard kill).  There appears to be no
mechanism to cleanup these stale lock files (there were a couple in the
directory dating back several years), and restarting mailman or even rebooting
the system does not clean things up.  At the very least restarting mailman
should cleanup these stale lock files, in particular what I assume is the
master lock: listname.lock and probably the actual source of my problems.
A better solution would probably include actually checking the lock files
periodically to make sure they are still valid.

-- System Information:
Debian Release: 3.1
Architecture: i386 (i686)
Kernel: Linux 2.6.11.10-bs5deatech
Locale: LANG=C, LC_CTYPE=C (charmap=ANSI_X3.4-1968)

Versions of packages mailman depends on:
ii  apache [httpd]1.3.33-6sarge1 versatile, high-performance HTTP s
ii  apache2-mpm-prefork [http 2.0.54-5   traditional model for Apache2
ii  cron  3.0pl1-86  management of regular background p
ii  debconf   1.4.30.13  Debian configuration management sy
ii  exim [mail-transport-agen 3.36-16An MTA (Mail Transport Agent)
ii  libc6 2.3.2.ds1-22   GNU C Library: Shared libraries an
ii  logrotate 3.7-5  Log rotation utility
ii  pwgen 2.03-1 Automatic Password generation
ii  python2.3.5-2An interactive high-level object-o
ii  ucf   1.17   Update Configuration File: preserv

-- debconf information:
* mailman/queue_files_present:
  mailman/default_server_language: en
* mailman/gate_news: false
* mailman/site_languages: en
* mailman/used_languages: en
* mailman/create_site_list:


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]



Bug#304538: Still broken: 248622 - exim_tidydb failed to open DB file

2005-04-13 Thread Shannon Dealy
Package: exim
Version: 3.36-14
Severity: normal


The bug previously reported in #248622 (caused by incompatible database 
formats), and closed by this version of exim (3.36-14) still exists in 
this version.  Among the changes noted in the message closing the previous 
bug report was that the postinst script was calling db3_upgrade to fix 
this problem, however, if it is doing so, it doesn't do it directly and I 
was unable to locate any indirect calls to db3_upgrade either.


-- System Information:
Debian Release: 3.1
  APT prefers testing
  APT policy: (500, 'testing')
Architecture: i386 (i586)
Kernel: Linux 2.4.27-2-586tsc
Locale: LANG=C, LC_CTYPE=C (charmap=ANSI_X3.4-1968)

Versions of packages exim depends on:
ii  cron3.0pl1-86management of regular background p
ii  libc6   2.3.2.ds1-20 GNU C Library: Shared libraries an
ii  libdb3  3.2.9-22 Berkeley v3 Database Libraries [ru
ii  libident0.22-3   simple RFC1413 client library - ru
ii  libldap22.1.30-3 OpenLDAP libraries
ii  libpam0g0.76-22  Pluggable Authentication Modules l
ii  libpcre34.5-1.1  Perl 5 Compatible Regular Expressi

-- no debconf information


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]