On 2017-06-22 04:12, Qu Wenruo wrote:
>
> And in that case even device of data stripe 2 is missing, btrfs don't really
> need to use parity to rebuild it, as btrfs knows there is no extent in that
> stripe, and data csum matches for data stripe 1.
You are assuming that there is no data in disk2
At 06/22/2017 10:53 AM, Marc MERLIN wrote:
Ok, first it finished (almost 24H)
(...)
ERROR: root 3862 EXTENT_DATA[18170706 135168] interrupt
ERROR: root 3862 EXTENT_DATA[18170706 1048576] interrupt
ERROR: root 3864 EXTENT_DATA[109336 4096] interrupt
ERROR: errors found in fs roots
found 5544779
At 06/22/2017 10:53 AM, Marc MERLIN wrote:
Ok, first it finished (almost 24H)
(...)
ERROR: root 3862 EXTENT_DATA[18170706 135168] interrupt
ERROR: root 3862 EXTENT_DATA[18170706 1048576] interrupt
ERROR: root 3864 EXTENT_DATA[109336 4096] interrupt
ERROR: errors found in fs roots
found 5544779
At 06/22/2017 10:43 AM, Chris Murphy wrote:
On Wed, Jun 21, 2017 at 8:12 PM, Qu Wenruo wrote:
Well, in fact, thanks to data csum and btrfs metadata CoW, there is quite a
high chance that we won't cause any data damage.
But we have examples where data does not COW, we see a partial stripe
Ok, first it finished (almost 24H)
(...)
ERROR: root 3862 EXTENT_DATA[18170706 135168] interrupt
ERROR: root 3862 EXTENT_DATA[18170706 1048576] interrupt
ERROR: root 3864 EXTENT_DATA[109336 4096] interrupt
ERROR: errors found in fs roots
found 5544779108352 bytes used, error(s) found
total csum by
On Wed, Jun 21, 2017 at 8:12 PM, Qu Wenruo wrote:
>
> Well, in fact, thanks to data csum and btrfs metadata CoW, there is quite a
> high chance that we won't cause any data damage.
But we have examples where data does not COW, we see a partial stripe
overwrite. And if that is interrupted it's cl
On Thu, Jun 22, 2017 at 10:01:21AM +0800, Qu Wenruo wrote:
> Commit 4751832da990 ("btrfs: fiemap: Cache and merge fiemap extent before
> submit it to user") introduced a warning to catch unemitted cached
> fiemap extent.
>
> However such warning doesn't take the following case into consideration:
At 06/21/2017 11:13 PM, Marc MERLIN wrote:
On Tue, Jun 20, 2017 at 08:43:52PM -0700, Marc MERLIN wrote:
On Tue, Jun 20, 2017 at 09:31:42PM -0600, Chris Murphy wrote:
On Tue, Jun 20, 2017 at 5:12 PM, Marc MERLIN wrote:
I'm now going to remount this with nospace_cache to see if your guess ab
At 06/22/2017 02:24 AM, Chris Murphy wrote:
On Wed, Jun 21, 2017 at 2:45 AM, Qu Wenruo wrote:
Unlike pure stripe method, one fully functional RAID5/6 should be written in
full stripe behavior, which is made up by N data stripes and correct P/Q.
Given one example to show how write sequence a
At 06/22/2017 01:03 AM, Goffredo Baroncelli wrote:
Hi Qu,
On 2017-06-21 10:45, Qu Wenruo wrote:
At 06/21/2017 06:57 AM, waxhead wrote:
I am trying to piece together the actual status of the RAID5/6 bit of BTRFS.
The wiki refer to kernel 3.19 which was released in February 2015 so I assume
th
I'm getting command hangs and service start failures during scrubs.
top says CPU idle is 58-64%.
Running 'perf top' takes more than 1 minute for results to appear.
Connecting to a web management service (cockpit) takes longer, maybe 2
minutes or sometimes the login times out. And just doing an ls
Commit 4751832da990 ("btrfs: fiemap: Cache and merge fiemap extent before
submit it to user") introduced a warning to catch unemitted cached
fiemap extent.
However such warning doesn't take the following case into consideration:
0 4K 8K
|< fiemap ran
On Wed, Jun 21, 2017 at 05:22:15PM -0600, Chris Murphy wrote:
> I don't know what it means. Maybe Qu has some idea. He might want a
> btrfs-image of this file system to see if it's a bug. There are still
> some bugs found with lowmem mode, so these could be bogus messages.
> But the file system cl
At 06/21/2017 08:10 PM, Adam Borowski wrote:
On Wed, Jun 21, 2017 at 05:28:50PM +0800, Qu Wenruo wrote:
Would you please try this patch based on v4.12-rc5 and try to reproduce
the kernel warning?
It would be better to eliminate the noisy by ensure there is no other fiemap
caller on btrfs.
H
On Wed, Jun 21, 2017 at 9:13 AM, Marc MERLIN wrote:
> On Tue, Jun 20, 2017 at 08:43:52PM -0700, Marc MERLIN wrote:
>> On Tue, Jun 20, 2017 at 09:31:42PM -0600, Chris Murphy wrote:
>> > On Tue, Jun 20, 2017 at 5:12 PM, Marc MERLIN wrote:
>> >
>> > > I'm now going to remount this with nospace_cache
On Wed, Jun 21, 2017 at 2:12 PM, Goffredo Baroncelli wrote:
>
> Generally speaking, when you write "two failure" this means two failure at
> the same time. But the write hole happens even if these two failures are not
> at the same time:
>
> Event #1: power failure between the data stripe writ
On 6/21/17 5:15 PM, Chris Mason wrote:
>
>
> On 06/21/2017 05:08 PM, Jeff Mahoney wrote:
>> On 6/21/17 4:31 PM, Chris Mason wrote:
>>> On 06/21/2017 04:14 PM, Jeff Mahoney wrote:
On 6/14/17 11:44 AM, je...@suse.com wrote:
> From: Jeff Mahoney
>
> In a heavy write scenario, we ca
On 06/21/2017 05:08 PM, Jeff Mahoney wrote:
On 6/21/17 4:31 PM, Chris Mason wrote:
On 06/21/2017 04:14 PM, Jeff Mahoney wrote:
On 6/14/17 11:44 AM, je...@suse.com wrote:
From: Jeff Mahoney
In a heavy write scenario, we can end up with a large number of pinned
bytes. This can translate int
On 6/21/17 4:31 PM, Chris Mason wrote:
> On 06/21/2017 04:14 PM, Jeff Mahoney wrote:
>> On 6/14/17 11:44 AM, je...@suse.com wrote:
>>> From: Jeff Mahoney
>>>
>>> In a heavy write scenario, we can end up with a large number of pinned
>>> bytes. This can translate into (very) premature ENOSPC becau
On 06/21/2017 04:14 PM, Jeff Mahoney wrote:
On 6/14/17 11:44 AM, je...@suse.com wrote:
From: Jeff Mahoney
In a heavy write scenario, we can end up with a large number of pinned
bytes. This can translate into (very) premature ENOSPC because pinned
bytes must be accounted for when allowing a re
On 6/14/17 11:44 AM, je...@suse.com wrote:
> From: Jeff Mahoney
>
> In a heavy write scenario, we can end up with a large number of pinned
> bytes. This can translate into (very) premature ENOSPC because pinned
> bytes must be accounted for when allowing a reservation but aren't
> accounted for
On 2017-06-21 20:24, Chris Murphy wrote:
> On Wed, Jun 21, 2017 at 2:45 AM, Qu Wenruo wrote:
>
>> Unlike pure stripe method, one fully functional RAID5/6 should be written in
>> full stripe behavior, which is made up by N data stripes and correct P/Q.
>>
>> Given one example to show how write seq
On Wed, May 17, 2017 at 10:56:22AM +0800, Qu Wenruo wrote:
> The remaining qgroup fixes patches, based on the Chris' for-linus-4.12
> branch with commit 9bcaaea7418d09691f1ffab5c49aacafe3eef9d0 as base.
>
> Can be fetched from github:
> https://github.com/adam900710/linux/tree/qgroup_fixes_non_sta
On Wed, Jun 21, 2017 at 12:51 AM, Marat Khalili wrote:
> On 21/06/17 06:48, Chris Murphy wrote:
>>
>> Another possibility is to ensure a new write is written to a new*not*
>> full stripe, i.e. dynamic stripe size. So if the modification is a 50K
>> file on a 4 disk raid5; instead of writing 3 64K
On Wed, Jun 21, 2017 at 2:45 AM, Qu Wenruo wrote:
> Unlike pure stripe method, one fully functional RAID5/6 should be written in
> full stripe behavior, which is made up by N data stripes and correct P/Q.
>
> Given one example to show how write sequence affects the usability of
> RAID5/6.
>
> Exi
On Tue, Jun 06, 2017 at 04:45:32PM -0700, Omar Sandoval wrote:
> From: Omar Sandoval
>
> Catch any future/remaining leaks or underflows of total_bytes_pinned.
>
> Signed-off-by: Omar Sandoval
This patch received some objections. As it's a debugging aid, I'd rather
merge a patch that other agre
On Tue, Jun 06, 2017 at 04:45:27PM -0700, Omar Sandoval wrote:
> From: Omar Sandoval
>
> Signed-off-by: Omar Sandoval
Reviewed-by: David Sterba
Added some changelog.
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
On 2017-06-21 13:20, Andrei Borzenkov wrote:
21.06.2017 16:41, Austin S. Hemmelgarn пишет:
On 2017-06-21 08:43, Christoph Anton Mitterer wrote:
On Wed, 2017-06-21 at 16:45 +0800, Qu Wenruo wrote:
Btrfs is always using device ID to build up its device mapping.
And for any multi-device implement
21.06.2017 16:41, Austin S. Hemmelgarn пишет:
> On 2017-06-21 08:43, Christoph Anton Mitterer wrote:
>> On Wed, 2017-06-21 at 16:45 +0800, Qu Wenruo wrote:
>>> Btrfs is always using device ID to build up its device mapping.
>>> And for any multi-device implementation (LVM,mdadam) it's never a
>>> g
21.06.2017 09:51, Marat Khalili пишет:
> On 21/06/17 06:48, Chris Murphy wrote:
>> Another possibility is to ensure a new write is written to a new*not*
>> full stripe, i.e. dynamic stripe size. So if the modification is a 50K
>> file on a 4 disk raid5; instead of writing 3 64K data strips + 1 64K
Hi Qu,
On 2017-06-21 10:45, Qu Wenruo wrote:
> At 06/21/2017 06:57 AM, waxhead wrote:
>> I am trying to piece together the actual status of the RAID5/6 bit of BTRFS.
>> The wiki refer to kernel 3.19 which was released in February 2015 so I assume
>> that the information there is a tad outdated (th
On Wed, Jun 21, 2017 at 11:52:36AM -0400, Chris Mason wrote:
> On 06/21/2017 11:16 AM, Dave Jones wrote:
> > WARNING: CPU: 2 PID: 7153 at fs/btrfs/ordered-data.c:753
> > btrfs_wait_ordered_roots+0x1a3/0x220
> > CPU: 2 PID: 7153 Comm: kworker/u8:7 Not tainted 4.12.0-rc6-think+ #4
> > Workqueue
WARNING: CPU: 2 PID: 7153 at fs/btrfs/ordered-data.c:753
btrfs_wait_ordered_roots+0x1a3/0x220
CPU: 2 PID: 7153 Comm: kworker/u8:7 Not tainted 4.12.0-rc6-think+ #4
Workqueue: events_unbound btrfs_async_reclaim_metadata_space
task: 8804f08d5380 task.stack: c9000895c000
RIP: 0010:btrfs_wait_
The XATTR_ITEM is a type of a directory item so we use the common
validator helper. We have to adjust the limits because of potential
data_len (ie. the xattr value), which is otherwise 0 for other directory
items.
Signed-off-by: David Sterba
---
fs/btrfs/dir-item.c | 12 +++-
1 file chan
On 06/21/2017 11:16 AM, Dave Jones wrote:
WARNING: CPU: 2 PID: 7153 at fs/btrfs/ordered-data.c:753
btrfs_wait_ordered_roots+0x1a3/0x220
CPU: 2 PID: 7153 Comm: kworker/u8:7 Not tainted 4.12.0-rc6-think+ #4
Workqueue: events_unbound btrfs_async_reclaim_metadata_space
task: 8804f08d5380 task.st
I've done with heuristic method,
so i post some performance test output:
(I store test data in /run/user/$UID/, and script just ran programm 2 times)
###
# Performance test will measure initialization time
# And remove it from run time of tests
# This may be inaccurate in some cases
# But this
On Tue, Jun 20, 2017 at 08:43:52PM -0700, Marc MERLIN wrote:
> On Tue, Jun 20, 2017 at 09:31:42PM -0600, Chris Murphy wrote:
> > On Tue, Jun 20, 2017 at 5:12 PM, Marc MERLIN wrote:
> >
> > > I'm now going to remount this with nospace_cache to see if your guess
> > > about
> > > space_cache was c
On 06/21/2017 07:17 AM, David Sterba wrote:
> On Mon, Jun 19, 2017 at 01:55:37PM +0300, Dan Carpenter wrote:
>> This function is supposed to return blk_status_t error codes now but
>> there was a stray -ENOMEM left behind.
>>
>> Fixes: 4e4cbee93d56 ("block: switch bios to blk_status_t")
>> Signed-o
On Tue, Jun 20, 2017 at 08:15:26AM -0400, je...@suse.com wrote:
> From: Jeff Mahoney
>
> On an uncontended system, we can end up hitting soft lockups while
> doing replace_path. At the core, and frequently called is
> btrfs_qgroup_trace_leaf_items, so it makes sense to add a cond_resched
> there
On 2017-06-21 08:43, Christoph Anton Mitterer wrote:
On Wed, 2017-06-21 at 16:45 +0800, Qu Wenruo wrote:
Btrfs is always using device ID to build up its device mapping.
And for any multi-device implementation (LVM,mdadam) it's never a
good
idea to use device path.
Isn't it rather the other way
On Mon, Jun 19, 2017 at 01:55:37PM +0300, Dan Carpenter wrote:
> This function is supposed to return blk_status_t error codes now but
> there was a stray -ENOMEM left behind.
>
> Fixes: 4e4cbee93d56 ("block: switch bios to blk_status_t")
> Signed-off-by: Dan Carpenter
Acked-by: David Sterba
Th
On Wed, 2017-06-21 at 16:45 +0800, Qu Wenruo wrote:
> Btrfs is always using device ID to build up its device mapping.
> And for any multi-device implementation (LVM,mdadam) it's never a
> good
> idea to use device path.
Isn't it rather the other way round? Using the ID is bad? Don't you
remember
Marc MERLIN posted on Tue, 20 Jun 2017 16:12:03 -0700 as excerpted:
> On Tue, Jun 20, 2017 at 08:44:29AM -0700, Marc MERLIN wrote:
>> On Tue, Jun 20, 2017 at 03:36:01PM +, Hugo Mills wrote:
>>
"space cache will be invalidated " => doesn't that mean that my
cache was already cleared
Hi Adam,
Would you please try this patch based on v4.12-rc5 and try to reproduce
the kernel warning?
It would be better to eliminate the noisy by ensure there is no other fiemap
caller on btrfs.
Thanks,
Qu
Signed-off-by: Qu Wenruo
---
fs/btrfs/extent_io.c | 23 +--
1 file
At 06/21/2017 06:57 AM, waxhead wrote:
I am trying to piece together the actual status of the RAID5/6 bit of
BTRFS.
The wiki refer to kernel 3.19 which was released in February 2015 so I
assume that the information there is a tad outdated (the last update on
the wiki page was July 2016)
http
At 06/18/2017 09:42 PM, Adam Borowski wrote:
On Sun, Jun 18, 2017 at 07:23:00PM +0800, Qu Wenruo wrote:
[ 39.726215] BTRFS warning (device sda1): unhandled fiemap cache
detected: offset=phys$35798867968 len1072 flags=0x2008
[ 151.882586] BTRFS warning (device sda1): unhandled fiemap cac
When changing the size of disks/filesystem we should always be
rounding down to a multiple of sectorsize
Signed-off-by: Nikolay Borisov
---
Changes since v1:
- Worked on incorporated feedback by Eryu
- Changed test number to 146 to avoid clashes
tests/btrfs/146 | 147 ++
> [ ... ] This will make some filesystems mostly RAID1, negating
> all space savings of RAID5, won't it? [ ... ]
RAID5/RAID6/... don't merely save space, more precisely they
trade lower resilience and a more anisotropic and smaller
performance envelope to gain lower redundancy (= save space).
--
T
48 matches
Mail list logo