Thanks Daniel, It works fine now !
On Mon, Jun 30, 2014 at 10:20 PM, Daniel Graña <[email protected]> wrote: > Hi Umar, > > the package is available now in ubuntu repositories. > > thanks > Daniel. > > El sábado, 28 de junio de 2014 08:29:57 UTC-3, Umar Shah escribió: > >> Anyone tried to install the latest ubuntu package? >> I am not able to see version 0.24 after i changed the apt line and ran >> "apt-get update " >> >> >> On Fri, Jun 27, 2014 at 3:38 AM, Daniel Graña <[email protected]> wrote: >> >>> Hello all, >>> >>> It has been 4 months of development, 30 authors, more than 80 issues >>> closed, 225 commits, 177 files changed, 6740 insertions and 4134 deletions. >>> New and old faces has been seen in the past months reporting and fixing >>> issues, discussing and helping get new features in shape. >>> Pretty amazing work, thanks to everyone that contributed in one or other >>> way to make Scrapy 0.24 possible! >>> >>> I'd like to take this opportunity to ask for help with the scrapy.org >>> website. Its design is old (hasn't changed much since 2008!) and we would >>> like to give it a proper makeover, with a fresher, modern look, maybe >>> including a snippet of simple, self-contained code that shows the power of >>> Scrapy. Anyone out there that would like to become famous for designing the >>> new scrapy.org website? :) >>> >>> Check out the Release Notes >>> <http://doc.scrapy.org/en/latest/news.html#id1>, from where I would >>> like to highlight the now simpler top-level imports and selector's >>> shortcuts:: >>> >>> import scrapy >>> >>> class MySpider(scrapy.Spider): >>> # ... >>> def parse(self, response): >>> for href in response.xpath('//a/@href').extract(): >>> yield scrapy.Request(url) >>> >>> >>> At last but not less important, the credits: >>> >>> A.J. Welch (1): >>> Generalize the file pipeline log messages so they are not specific to >>> downloading images. >>> >>> Alex Cepoi (2): >>> improvements to scrapy check/contracts >>> fix contracts tests >>> >>> Alexander Chekunkov (5): >>> test for RFPDupeFilter.request_fingerprint overriding >>> added note about RFPDupeFilter.request_fingerprint overriding to the >>> settings documentation >>> added short RFPDupeFilter.request_fingerprint interface description >>> DOWNLOADER setting >>> DOWNLOADER setting >>> >>> Alexey Bezhan (6): >>> Clarify MapCompose documentation >>> Fix some typos, whitespace and small errors in docs >>> Add a note about reporting security issues >>> Bind telnet console and webservice to 127.0.0.1 by default >>> Fix PEP8 warnings in project template files >>> Fix PEP8 warnings in spider templates >>> >>> Ana Sabina Uban (1): >>> Fixed SgmlLinkExtractor constructor to properly handle both string >>> and list parameters (attrs, tags, deny_extensions) >>> >>> Benoit Blanchon (3): >>> BaseSgmlLinkExtractor: Fixed unknown_endtag() so that it only set >>> current_link=None when the end tag match the opening tag >>> BaseSgmlLinkExtractor: Added unit test of a link with an inner tag >>> BaseSgmlLinkExtractor: Fixed the missing space when the link has an >>> inner tag >>> >>> Breno Colom (1): >>> Update scrapy command line doc with additional scrapy parse options >>> >>> Cameron Lane (2): >>> [#744] Ensure domain is not None before building regex >>> [#744] Test for allowed domains including NoneTypes >>> >>> Capi Etheriel (4): >>> fixes dynamic itemclass example usage of type() >>> Running lucasdemarchi/codespell to fix typos in docs >>> Running lucasdemarchi/codespell to fix typos in SEPs >>> Running lucasdemarchi/codespell to fix typos in code >>> >>> Carlos Rivera (1): >>> grammatical issue >>> >>> Cash Costello (1): >>> Added missing word in practices.rst >>> >>> Claudio Salazar (4): >>> Fixed XXE flaw in sitemap reader >>> Fixed XML selector against XXE attacks >>> Added test against XXE attacks for Sitemap >>> Added resolve_entities to kwargs in SafeXMLParser >>> >>> Daniel Graña (45): >>> Merge 0.22.0 release notes >>> bump version to 0.23 >>> fix 0.22.0 release date >>> Update Ubuntu installation instructions >>> fix apt-get line >>> replace warning about updating package lists by a note on package >>> upgrade >>> show ubuntu setup instructions as literal code >>> replace unencodeable codepoints with html entities. fixes #562 and >>> #285 >>> Fix wrong checks on subclassing of deprecated classes. closes #581 >>> test inspect.stack failure >>> localhost666 can resolve under certain circumstances >>> Add 0.22.1 release notes >>> fix a reference to unexistent engine.slots. closes #593 >>> Add 0.22.2 release notes >>> try to restore pypy tests >>> Run testsuite with py.test >>> cleanup toplevel namespace >>> Add basic top-level shortcuts >>> remove .re() shortcut >>> update docs >>> update spider templates >>> Remove "sel" shortcut from scrapy shell$ >>> document shortcuts in TextResponse class >>> Ammend example nesting selectors >>> Restore and deprecate "sel" shortcut >>> limit Twisted support to pre-14.0.0 while #718 is fixed >>> fix tests after changes introduced by scrapy/w3lib#21 >>> force installation of w3lib and queuelib for trunk env >>> Avoid IPython warning. thanks @bryant1410. closes #623 >>> sort spiders in "scrapy list" cmd. closes #736 >>> Add a LevelDB cache backend >>> add leveldb cache backend docs >>> indent parsed-literal as part of ordered list >>> Upload sdist and wheel packages to pypi using travis-ci deploys >>> Add bumpversion config >>> Revert "limit Twisted support to pre-14.0.0 while #718 is fixed" >>> hold a reference to backwards compatible _contextFactory >>> Restore compatibility with Settings.overrides while still deprecating >>> it >>> recognize jl extension as jsonlines exporter and update docs >>> promote LxmlLinkExtractor as default in docs >>> address latest comments >>> No need to keep extracted links as instance attribute. fixes #763 >>> Add 0.24.0 release notes >>> Bump version: 0.23.0 → 0.24.0 >>> set 0.24.0 release date >>> >>> Denys Butenko (5): >>> Resolved issue #546. Output format parsing from filename extension. >>> Added back `-t` option. If `--output-format` not defined parse from >>> extension `--output` >>> Fix default value. >>> Add import os for crawl. >>> Added more verbose error message for unrecognized output format. PEP8. >>> >>> Edwin O Marshall (32): >>> Converted sep-001 to rst format >>> converted sep 002 to rst >>> - decided that removing files would cause conflicts on merge >>> - readded file to prevent future merge conflicts >>> converted sep 3 for #629 >>> sep 4 for #629 >>> sep 11 for #629 >>> - sep 15 for #629 >>> sep 6 for #629 >>> - sep 10 for #629 >>> - didn't like the way blockquotes rendered >>> - trying to separate quote context >>> - changing indentation so contexts are recognized >>> - given that it'sa block quote, quotation marks seem redundant >>> - removing trac file again to see if merges play well together >>> - removed trac file >>> - removed trac file >>> - removed trac file >>> - removed track file >>> removed trac file >>> removed trac file >>> - removed trac file >>> converted sep 7 for #629 >>> sep 12 for #629 >>> - converted sep 18 >>> converted sep 16 >>> converted sep 13 >>> converted sep 5 >>> - convertd sep 8 >>> converted sep 9 >>> converted sep017 >>> sep 14 for #629 >>> >>> Irhine (2): >>> add encoding utf-8 to the first line >>> support i18n by using utf-8 coding template files >>> >>> Julia Medina (34): >>> New doc: clickdata in Formrequest.from_response >>> New tests: clickdata's nr in Formrequest.from_response >>> FormRequest doc improvements >>> More appropriate assert in FormRequest test >>> Tests for loading download handlers >>> Fix minor typo in DownloaderHandlers comment >>> Doc for disabling download handler >>> Minor fixes in LoadTestCase in test_downloader_handlers >>> Trial functionality for running tests with pytest >>> Add py33 environment to allowed failures in travis-ci >>> Support doctest and __init__.py test discover in pytest >>> Ignore files with import errors on pytest test discover >>> Change function name so it does not mess up with pytest autodiscover >>> Fix httpcache doctest that assumed dictionary order >>> Ensure spiders module reload between spider manager tests >>> New tox env: docs >>> Ignore known broken links in docs linkcheck >>> Fix broken links in documentation >>> sep#19 proposed changes >>> New SettingsAttribute class >>> Settings priorities dictionary >>> New set and setdict method using SettingsAttribute in Settings >>> Deprecate CrawlerSettings, as its functionality is replicable by >>> Settings class >>> Settings and SettingsAtribute tests >>> Fix and extend the documentation of the new Settings api >>> Settings topic updated >>> Fix settings repr on the logs of the shell and tutorial docs topics >>> setmodule helper method on Settings class >>> Update get_crawler method in utils/test.py with new Settings interface >>> get_project_settings now returns a Settings instance >>> Change command's default_settings population in cmdline.py >>> Change how settings are overriden in ScrapyCommands >>> Fix settings usage in runspider and crawl commands >>> Fix settings usage across tests >>> >>> Mikhail Korobov (18): >>> fix typos in news.rst and remove (not released yet) header >>> Handle cases when inspect.stack() fails >>> testing PIL dependency is removed because there is a new mitmproxy >>> version >>> TST Improved twisted installation in tox.ini for Python 3.3 >>> reduce code duplication in test_spidermiddleware_httperror >>> scrapy.utils.test.docrawl function >>> Fix for #612 + integration-style tests for HttpErrorMiddleware >>> TST fix file descriptor leak and a bad variable name in get_testlog >>> make scrapy.version_info a tuple of integers >>> remove unused import >>> use "import scrapy" in templates >>> DOC use top-level shortcuts in docs >>> suggest scrapy.Selector in deprecation warnings >>> TST fix tests that became broken after adding top-level imports and >>> switching to py.test. >>> fix scrapy.version_info when SCRAPY_VERSION_FROM_GIT is set >>> response.selector, response.xpath(), response.css() and response.re() >>> DOC selectors.rst cleanup >>> add utf8 encoding header to spider templates >>> >>> Nikita Nikishin (1): >>> Fixed #441. >>> >>> Nikolaos-Digenis Karagiannis (5): >>> downloaderMW doc typo (spiderMW doc copy remnant) >>> SpiderMW doc typo: SWP request, response >>> ItemLoader doc: missing args in replace_value() >>> document spider.closed() shortcut >>> Document signal "request_scheduled" >>> >>> Pablo Hoffman (11): >>> make 'basic' the default template spider in genspider, and added info >>> with next steps to startproject. closes #488 >>> add SEP-021 (Add-ons) - work in progress >>> remove references to deprecated scrapy-developers list >>> rename attribute to match conventions used for XXX_DEBUG settings (in >>> autothrottle and cookies mw) >>> remove no longer used setting: MAIL_DEBUG >>> remove unused setting: DOWNLOADER_DEBUG >>> signals doc: make argument order more consistent with code (although >>> it doesn't matter in practice) >>> add Julia to SEP-019 authors >>> crate release notes for 0.24 and #699 to it >>> minor change to request_scheduled signal doc >>> doc: use |version| substitution in ubuntu packages >>> >>> Paul Brown (1): >>> fixed typo >>> >>> Paul Tremberth (18): >>> Disable smart strings in lxml XPath evaluations >>> Make lxml smart strings functionality customizable >>> Add testcase to check is default Selector doesnt return smart strings >>> Use assertTrue/False >>> RegexLinkExtractor: encode URL unicode value when creating Links >>> Offsite: add 2 stats counters >>> Always enable offsite stats + refactor test to initialize crawler >>> Fix tests for Travis-CI build >>> CrawSpider: support process_links as generator >>> Fix HtmlParserLinkExtractor and tests after #485 merge >>> Docs: 4-space indent for final spider example >>> DupeFilter: add setting for verbose logging + stats counter for >>> filtered requests >>> Remove _log_level attribute as per comments >>> Support case-insensitive domains in url_is_from_any_domain() >>> Add tests for start requests, filtered and non-filtered >>> Check pending start_requests before calling _spider_idle() in engine >>> (fixes #706) >>> Add LxmlLinkExtractor class similar to SgmlLinkExtractor (#528) >>> Add doc on LxmlLinkExtractor class >>> >>> Rafal Jagoda (1): >>> add response arg to item_dropped signal handlers #710 >>> >>> Rendaw (1): >>> Elaborated request priority value. >>> >>> Rolando Espinoza (8): >>> Ignore None's values when using the ItemLoader. >>> Unused re import and PEP8 minor edits. >>> Expose current crawler in the scrapy shell. >>> PEP8 minor edits. >>> Updated shell docs with the crawler reference and fixed the actual >>> shell output. >>> Updated the tutorial crawl output with latest output. >>> DOC Fixed HTTPCACHE_STORAGE typo in the default value which is now >>> Filesystem instead Dbm. >>> DOC Use pipelines module name instead of pipieline following default >>> project files. >>> >>> Rolando Espinoza La fuente (1): >>> Alow to disable a downloader handler just like any other component. >>> >>> Ruben Vereecken (2): >>> Added content-type check as per issue #193 >>> Redefined test for #193 >>> >>> deed02392 (1): >>> Update httperror.py >>> >>> ncp1113 (1): >>> for loops have to have a : at the end of the line >>> >>> nyov (2): >>> better call to parent class >>> update a link reference >>> >>> stray-leone (1): >>> modify the version of scrapy ubuntu package >>> >>> tpeng (3): >>> add message when raise IngoreReques; fix item_scraped document >>> set the exit code to non-zero when contracts fails >>> print spider name even it has no contract tests when -v is specified >>> >>> tracicot (1): >>> Correct typos >>> >>> -- >>> You received this message because you are subscribed to the Google >>> Groups "scrapy-users" group. >>> To unsubscribe from this group and stop receiving emails from it, send >>> an email to [email protected]. >>> To post to this group, send email to [email protected]. >>> Visit this group at http://groups.google.com/group/scrapy-users. >>> For more options, visit https://groups.google.com/d/optout. >>> >> >> -- > You received this message because you are subscribed to the Google Groups > "scrapy-users" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To post to this group, send email to [email protected]. > Visit this group at http://groups.google.com/group/scrapy-users. > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "scrapy-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/scrapy-users. For more options, visit https://groups.google.com/d/optout.
