NP: as everything it's not the silver bullet but with the redis incarnation I'm sure you can achieve less than 3 second (if you tune heartbeat even less than 1 second) from when the task gets queued to when it gets processed.
On Tuesday, May 3, 2016 at 12:32:13 PM UTC+2, Mirek Zvolský wrote: > > Hi, Niphlod. > > After I have read something about scheduler, > I am definitively sorry for my previous notes > and I choose web2py scheduler of course. > > It will be my first use of it (with much older ~3 years web2py app I have > used cron only), > so it will take some time to learn with scheduler. But it is sure worth to > redesign it so. > > Thanks you are patient with me. > Mirek > > > > > Dne pondělí 2. května 2016 20:35:05 UTC+2 Mirek Zvolský napsal(a): >> >> You are right. >> At this time it works for me via ajax well and I will look carefully for >> problems. >> If so, I will move to scheduler. >> >> I see this is exactly what Massimo(?) writes at the bottom of Ajax >> chapter of the book. >> >> PS: about times: >> At notebook with mobile connection it takes 20-40s. So it could be danger. >> At cloud server with SSD it takes 2-10s. But this will be my case. And I >> feel better when the user can have typical response in 3s instead in 8s. >> >> >> >> >> >> Dne neděle 1. května 2016 22:10:31 UTC+2 Niphlod napsal(a): >>> >>> the statement "I don't need to use the scheduler, because I want to >>> start it as soon as possible" is flaky at best. If your "fetching" varies >>> from 2 to 20 seconds and COULD extend further to 60 seconds, waiting a few >>> seconds for the scheduler to start the process is .... uhm... debatable. >>> Of course relying to ajax if your "feching" can be killed in the process >>> is the only other way. >>> >>> On Sunday, May 1, 2016 at 8:09:23 PM UTC+2, Mirek Zvolský wrote: >>>> >>>> Thanks for info and tips, 6 years later. >>>> >>>> What I try to do >>>> is a form with single input, where user gives a query string >>>> and then data about (usually ~300) books will be retrieved via z39 and >>>> marc protocol/format, parsed and saved into local database. >>>> >>>> Of course this will take a time (2? 5? 20? seconds) and I decided >>>> not to show the result immediately, >>>> but show the same form with possibility to enter the next query + there >>>> is a list of pending queries (and their status - via ajax testing every 5 >>>> seconds) >>>> >>>> So my idea was to provide a return from the controller fast and before >>>> the return to start a new thread to retrieve/parse/save/commit data. >>>> >>>> From this discussion I understand that open new thread isn't best idea. >>>> I think it could be still possible, because if my new thread could be >>>> killed 60s later from the web server together with the original thread - >>>> such possibility is not fatal problem for me here. >>>> >>>> However when (as I read here) this would be a little wild technology, >>>> and because other technologies mentioned here: >>>> https://en.wikipedia.org/wiki/Comet_(programming) -paragraph >>>> Aternatives, are too difficult for me, >>>> and because I don't want use a scheduler, because I need to start as >>>> soon as possible, >>>> >>>> I will solve it so, >>>> that I will make 2 http accesses from my page: one with submit (will >>>> validate/save the query to database) and one with ajax/javascript >>>> (onSubmit >>>> from the old page or better: onPageLoaded from the next page where I give >>>> the query in .html DOM as some hidden value), which will start the z39 >>>> protocol/retrieve/parse/save data. >>>> This will be much better, because web2py in the ajax call will prepare >>>> the db variable with proper db model for me (which otherwise I must handle >>>> myselves in the separate thread). >>>> Callback from this ajax call should/could be some dummy javascript >>>> function, because it is not sure, and not important, if the page still >>>> exists when the server job will finish. >>>> >>>> So, if somebody is interesting and will read this very old thread, >>>> maybe this can give him some idea for time consumming actions. >>>> And maybe somebody will add other important hints or comments (thanks >>>> in advance). >>>> >>>> >>>> >>>> >>>> >>>> >>>> Dne středa 26. května 2010 0:33:02 UTC+2 Giuseppe Luca Scrofani >>>> napsal(a): >>>>> >>>>> Hi all, as promised I'm here to prove you are patient and nice :) >>>>> I' have to make this little app where there is a function that read >>>>> the html content of several pages of another website (like a spider) >>>>> and if a specified keyword is found the app refresh a page where there >>>>> is the growing list of "match". >>>>> Now, the spider part is already coded, is called search(), it uses >>>>> twill to log in the target site, read the html of a list of pages, >>>>> perform some searching procedures and keep adding the result to a >>>>> list. I integrated this in a default.py controller and make a call in >>>>> def index(): >>>>> This make the index.html page loading for a long time, because now it >>>>> have to finish to scan all pages before return all results. >>>>> What I want to achieve is to automatically refresh index every 2 >>>>> second to keep in touch with what is going on, seeing the list of >>>>> match growing in "realtime". Even better, if I can use some sort of >>>>> ajax magic to not refresh the entire page... but this is not vital, a >>>>> simple page refresh would be sufficient. >>>>> Question is: I have to use threading to solve this problem? >>>>> Alternative solutions? >>>>> I have to made the list of match a global to read it from another >>>>> function? It would be simpler if I made it write a text file, adding a >>>>> line for every match and reading it from the index controller? If I >>>>> have to use thread it will run on GAE? >>>>> >>>>> Sorry for the long text and for my bad english :) >>>>> >>>>> gls >>>>> >>>>> -- Resources: - http://web2py.com - http://web2py.com/book (Documentation) - http://github.com/web2py/web2py (Source code) - https://code.google.com/p/web2py/issues/list (Report Issues) --- You received this message because you are subscribed to the Google Groups "web2py-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to web2py+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.