What do you mean by it cannot be reproduced? How often does it occur?
What is the OS? What is the browser?

What is the last version (and date) that worked without this problem?

massimo

On Sep 27, 7:16 am, Phyo Arkar <phyo.arkarl...@gmail.com> wrote:
> It cannot be reproduced ..
>
> On 9/27/11, Phyo Arkar <phyo.arkarl...@gmail.com> wrote:
>
>
>
>
>
>
>
> > its not the same warning notices , client recieves 0 byte and timed
> > out at brower side looking at firebug.
>
> > On 9/27/11, Massimo Di Pierro <massimo.dipie...@gmail.com> wrote:
> >> Other than the error, do you see wrong behavior from the browser?
> >> I do not not believe that there is anything wrong as we ship the
> >> "official" rocket. We are just printing an exception that before was
> >> being ignored.
>
> >> Massimo
>
> >> On Sep 26, 2:47 pm, Phyo Arkar <phyo.arkarl...@gmail.com> wrote:
> >>> Hello Web2py.
>
> >>>  Rocket server was very  well stable for me until today , after 2
> >>> weeks of update and while testing usual Json output which was never
> >>> changed and working fine fore 8 months.
>
> >>> This happens!
>
> >>> ERROR:Rocket.Errors.Thread-6:Unhandled Error when serving connection:
> >>> Traceback (most recent call last):
>
> >>>   File "/root/web2py/gluon/rocket.py", line 1064, in run
> >>>     self.run_app(conn)
>
> >>>   File "/root/web2py/gluon/rocket.py", line 1552, in run_app
> >>>     self.write(data, sections)
>
> >>>   File "/root/web2py/gluon/rocket.py", line 1472, in write
> >>>     self.send_headers(data, sections)
>
> >>>   File "/root/web2py/gluon/rocket.py", line 1456, in send_headers
> >>>     self.conn.sendall(b(header_data))
>
> >>>   File "/usr/lib/python2.7/socket.py", line 224, in meth
> >>>     return getattr(self._sock,name)(*args)
>
> >>> error: [Errno 32] Broken pipe
>
> >>> and it is on and off randomly.
>
> >>> Why?
>
> >>> heres my Json Controller :
>
> >>> def rows():
>
> >>>         if REQUEST.vars.db_name:
> >>>                 casesdb = DAL( 'mysql://root@localhost/' +
> >>> REQUEST.vars.db_name )
> >>>                 casesdb.define_table( 'email_data', migrate = False,
> >>> *email_halfschema )
> >>>                 casesdb.define_table( 'loosefile_data', migrate = False,
> >>> *file_halfschema )
> >>>                 casesdb.define_table( 'attach_data', migrate = False,
> >>> *file_halfschema )
> >>>         elif SESSION.db_name:
> >>>                 casesdb = DAL( 'mysql://root@localhost/' +
> >>> SESSION.db_name
> >>> )
> >>>                 casesdb.define_table( 'email_data', migrate = False,
> >>> *email_halfschema )
> >>>                 casesdb.define_table( 'loosefile_data', migrate = False,
> >>> *file_halfschema )
> >>>                 casesdb.define_table( 'attach_data', migrate = False,
> >>> *file_halfschema )
>
> >>>         fields = ['id', 'filePath', 'fileName', 'cus', 'sentto',\
> >>>                          'emailFrom', 'subject', 'cc', 'bcc',
> >>> 'extracted',
> >>> 'hdrs',\
>
> >>> 'DateTime','OriginalDateTime','TimeZone','reply_to',"master",'duplicated','
> >>> MD5Hash','msgID','fqdn']
> >>>         rows = []
> >>>         page = int( request.vars.page )
>
> >>>         max = request.vars.totalrows
> >>>         if max:
> >>>                 pagesize = int( max )
> >>>         else:
> >>>                 pagesize = int( request.vars.rows )
> >>>         print pagesize
> >>>         limitby = ( page * pagesize - pagesize, page * pagesize )
> >>> #       limitby = ( 1, 25 )
> >>>         print "str( page * pagesize - pagesize )" + str( page * pagesize
> >>> -
> >>> pagesize ) + " limitby " + str( page * pagesize )
> >>> #       orderby = request.vars.sidx
> >>>         orderby = casesdb.email_data[request.vars.sidx]
> >>>         print orderby
> >>>         if request.vars.sord == 'desc':
> >>>                 orderby = ~orderby
>
> >>>         query = casesdb.email_data.id > 0
>
> >>>         for r in casesdb( query ).select( orderby = orderby, limitby =
> >>> limitby ):
> >>>                 vals = []
> >>>                 for f in fields:
> >>>                         if f == 'extracted':
> >>>                                 vals.append( _sentisize(r[f]) )
> >>>                         else :
> >>>                                 vals.append( r[f] )
> >>>                 rows.append( dict( id = r.id, cell = vals ) )
>
> >>>         total = casesdb( query ).count()
> >>>         print total
> >>>         if total % pagesize >0:
> >>>                 pages = int( total / pagesize ) +1
> >>>         else:
> >>>                 pages = int( total / pagesize )
> >>>         #if total % pagesize == 0: pages -= 1
> >>>         data = dict( total = pages, page = page, rows = rows )
> >>>         return data

Reply via email to