That does not sound expected or desired.  Could you point me to which
Chromium builders are responsible for so much data?

I suspect this is an artifact of new-run-webkit-tests or how the
Chromium builders are set up.

On Tue, Oct 19, 2010 at 9:12 AM, William Siegrist <[email protected]> wrote:
> Right now, /results/ is served from the new storage and is receiving test 
> results data since a day or two ago. For anything older, you will get 
> redirected to /old-results/ which is on the old storage. This probably breaks 
> your code if you are trying to load /results/ and walk backwards in 
> revisions. We should probably look at adding some sort of map to the 
> /json/builders/ data instead.
>
> On a side note, Chromium test results account for 75% of the 700GB of result 
> data, SnowLeopard is 11%, then everyone else. I assume Chromium generating so 
> much more data than everyone else is expected and desired?
>
> -Bill
>
>
>
>
> On Oct 18, 2010, at 5:04 PM, Eric Seidel wrote:
>
>> The most frequent consumer of the historical data is webkit-patch,
>> which uses it to map from revisions to builds:
>> http://trac.webkit.org/browser/trunk/WebKitTools/Scripts/webkitpy/common/net/buildbot.py#L109
>>
>> It's used when we're walking back through revisions trying to find
>> when the build broke, or when the user passes us a revision and
>> expects us to know build information about such.
>>
>> It's possible we could move off that map with some re-design.
>>
>>
>> One thing which would *hugely* speed up webkit-patch failure-reason
>> (and sherriff-bot, and other commands which use the
>> build_to_revision_map) is if we could make the results/ pages
>> paginated.  :)
>>
>>
>> I would be nice to keep all the build data for forever.  Even if after
>> some date in the past its on a slower server.
>>
>> -eric
>>
>>
>> On Sat, Oct 16, 2010 at 12:38 AM, William Siegrist <[email protected]> 
>> wrote:
>>> On Oct 14, 2010, at 10:13 AM, William Siegrist wrote:
>>>
>>>> On Oct 14, 2010, at 9:27 AM, William Siegrist wrote:
>>>>
>>>>> I am in the process of moving buildbot onto faster storage which should 
>>>>> help with performance. However, during the move, performance will be even 
>>>>> worse due to the extra i/o. There will be a downtime period in the next 
>>>>> few days to do the final switchover, but I won't know when that will be 
>>>>> until the preliminary copying is done. I am trying not to kill the master 
>>>>> completely, but there have been some slave disconnects due to the load 
>>>>> already this morning. I'll let everyone know when the downtime will be 
>>>>> once I know.
>>>>>
>>>>
>>>>
>>>> The copying of data will take days at the rate we're going, and the server 
>>>> is exhibiting some strange memory paging in the process. I am going to 
>>>> reboot the server and try copying with the buildbot master down. The 
>>>> master will be down for about 15m, if I can't get the copy done in that 
>>>> time I will schedule a longer downtime at a better time. Sorry for the 
>>>> churn.
>>>>
>>>
>>>
>>> Most of build.webkit.org is now running on the newer/faster storage. 
>>> However, the results data[1] is hundreds of gigabytes, going back 6 months, 
>>> and the new storage is not big enough. Does anyone have any opinion on how 
>>> much data to keep in results? Does anyone ever look back more than a month 
>>> or two? For now, the results will still come up a slowly, but hopefully the 
>>> rest of buildbot is a little more responsive. We're still planning to move 
>>> all of webkit.org to better hardware soon, but we hit some delays in that 
>>> process.
>>>
>>> [1] http://build.webkit.org/results/
>>>
>>> Thanks
>>> -Bill
>>>
>>> _______________________________________________
>>> webkit-dev mailing list
>>> [email protected]
>>> http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev
>>>
>
>
_______________________________________________
webkit-dev mailing list
[email protected]
http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev

Reply via email to