[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-07-01 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Description: 
*Problem:* 

In our production environment,  we put file in hdfs with zstd compressor, 
recently, we find that a specific file may leads to zstandard compressor 
failures. 

And we can reproduce the issue with specific file(attached file: badcase.data)

!image-2020-06-30-11-51-18-026.png|width=1031,height=230!

 

*Analysis*: 

ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
size)  for both inBufferSize and outBufferSize 

!image-2020-06-30-11-35-46-859.png|width=1027,height=387!

but zstd indeed provides two separately recommending inputBufferSize and 
outputBufferSize  

!image-2020-06-30-11-39-17-861.png!

 

*Workaround*

One workaround,  using recommended in/out buffer size provided by zstd lib can 
avoid the problem, but we don't know why. 

zstd recommended input buffer size:  1301072 (128 * 1024)

zstd recommended ouput buffer size: 131591 

!image-2020-06-30-11-42-44-585.png|width=1023,height=196!

 

 

 

 

 

 

 

  was:
*Problem:* 

In our production environment,  we put file in hdfs with zstd compressor, 
recently, we find that a specific file may leads to zstandard compressor 
failures. 

And we can reproduce the issue with specific file(attached file: badcase.data)

!image-2020-06-30-11-51-18-026.png|width=699,height=156!

 

*Analysis*: 

ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
size)  for both inBufferSize and outBufferSize 

!image-2020-06-30-11-35-46-859.png|width=475,height=179!

but zstd indeed provides two separately recommending inputBufferSize and 
outputBufferSize  

!image-2020-06-30-11-39-17-861.png!

 

*Workaround*

One workaround,  using recommended in/out buffer size provided by zstd lib can 
avoid the problem, but we don't know why. 

zstd recommended input buffer size:  1301072 (128 * 1024)

zstd recommended ouput buffer size: 131591 

!image-2020-06-30-11-42-44-585.png!

 

 

 

 

 

 

 


> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: HDFS-15445.patch, badcase.data, 
> image-2020-06-30-11-35-46-859.png, image-2020-06-30-11-39-17-861.png, 
> image-2020-06-30-11-42-44-585.png, image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=1031,height=230!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=1027,height=387!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png|width=1023,height=196!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-30 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Attachment: (was: 15445.patch)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: HDFS-15445.patch, badcase.data, 
> image-2020-06-30-11-35-46-859.png, image-2020-06-30-11-39-17-861.png, 
> image-2020-06-30-11-42-44-585.png, image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-30 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Attachment: HDFS-15445.patch
Status: Patch Available  (was: Open)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: 15445.patch, HDFS-15445.patch, badcase.data, 
> image-2020-06-30-11-35-46-859.png, image-2020-06-30-11-39-17-861.png, 
> image-2020-06-30-11-42-44-585.png, image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-30 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Status: Open  (was: Patch Available)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: 15445.patch, badcase.data, 
> image-2020-06-30-11-35-46-859.png, image-2020-06-30-11-39-17-861.png, 
> image-2020-06-30-11-42-44-585.png, image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-29 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Attachment: 15445.patch
Status: Patch Available  (was: Open)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: 15445.patch, badcase.data, 
> image-2020-06-30-11-35-46-859.png, image-2020-06-30-11-39-17-861.png, 
> image-2020-06-30-11-42-44-585.png, image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-29 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Status: Open  (was: Patch Available)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: badcase.data, image-2020-06-30-11-35-46-859.png, 
> image-2020-06-30-11-39-17-861.png, image-2020-06-30-11-42-44-585.png, 
> image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-29 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Status: Patch Available  (was: Open)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: badcase.data, image-2020-06-30-11-35-46-859.png, 
> image-2020-06-30-11-39-17-861.png, image-2020-06-30-11-42-44-585.png, 
> image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-29 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Description: 
*Problem:* 

In our production environment,  we put file in hdfs with zstd compressor, 
recently, we find that a specific file may leads to zstandard compressor 
failures. 

And we can reproduce the issue with specific file(attached file: badcase.data)

!image-2020-06-30-11-51-18-026.png|width=699,height=156!

 

*Analysis*: 

ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
size)  for both inBufferSize and outBufferSize 

!image-2020-06-30-11-35-46-859.png|width=475,height=179!

but zstd indeed provides two separately recommending inputBufferSize and 
outputBufferSize  

!image-2020-06-30-11-39-17-861.png!

 

*Workaround*

One workaround,  using recommended in/out buffer size provided by zstd lib can 
avoid the problem, but we don't know why. 

zstd recommended input buffer size:  1301072 (128 * 1024)

zstd recommended ouput buffer size: 131591 

!image-2020-06-30-11-42-44-585.png!

 

 

 

 

 

 

 

  was:
*Problem:* 

In our production environment,  we put file in hdfs with zstd compressor, 
recently, we find that a specific file may leads to zstandard compressor 
failures. 

And we can reproduce the issue with specific file(attached file: badcase.data)

!image-2020-06-30-11-51-18-026.png|width=699,height=156!

 

*Analysis*: 

ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
size)  for both inBufferSize and outBufferSize 

!image-2020-06-30-11-35-46-859.png|width=475,height=179!

but zstd indeed provides two separately recommending inputBufferSize and 
outputBufferSize  

!image-2020-06-30-11-39-17-861.png!

 

*Workaround*

One workaround,  using recommended in/out buffer size provided by zstd lib can 
avoid the problem, but we don't know why. 

input buffer size:  1301072 (128 * 1024)

ouput buffer size: 131591 

!image-2020-06-30-11-42-44-585.png!

 

 

 

 

 

 

 


> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: badcase.data, image-2020-06-30-11-35-46-859.png, 
> image-2020-06-30-11-39-17-861.png, image-2020-06-30-11-42-44-585.png, 
> image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> zstd recommended input buffer size:  1301072 (128 * 1024)
> zstd recommended ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-15445) ZStandardCodec compression mail fail(generic error) when encounter specific file

2020-06-29 Thread Igloo (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igloo updated HDFS-15445:
-
Summary: ZStandardCodec compression mail fail(generic error) when encounter 
specific file  (was: ZStandardCodec compression mail fail when encounter 
specific file)

> ZStandardCodec compression mail fail(generic error) when encounter specific 
> file
> 
>
> Key: HDFS-15445
> URL: https://issues.apache.org/jira/browse/HDFS-15445
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.6.5
> Environment: zstd 1.3.3
> hadoop 2.6.5 
>  
> --- 
> a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> +++ 
> b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/compress/zstd/TestZStandardCompressorDecompressor.java
> @@ -62,10 +62,8 @@
>  @BeforeClass
>  public static void beforeClass() throws Exception {
>  CONFIGURATION.setInt(IO_FILE_BUFFER_SIZE_KEY, 1024 * 64);
> - uncompressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt").toURI());
> - compressedFile = new File(TestZStandardCompressorDecompressor.class
> - .getResource("/zstd/test_file.txt.zst").toURI());
> + uncompressedFile = new File("/tmp/badcase.data");
> + compressedFile = new File("/tmp/badcase.data.zst");
>Reporter: Igloo
>Priority: Blocker
> Attachments: badcase.data, image-2020-06-30-11-35-46-859.png, 
> image-2020-06-30-11-39-17-861.png, image-2020-06-30-11-42-44-585.png, 
> image-2020-06-30-11-51-18-026.png
>
>
> *Problem:* 
> In our production environment,  we put file in hdfs with zstd compressor, 
> recently, we find that a specific file may leads to zstandard compressor 
> failures. 
> And we can reproduce the issue with specific file(attached file: badcase.data)
> !image-2020-06-30-11-51-18-026.png|width=699,height=156!
>  
> *Analysis*: 
> ZStandarCompressor use buffersize( From zstd recommended compress out buffer 
> size)  for both inBufferSize and outBufferSize 
> !image-2020-06-30-11-35-46-859.png|width=475,height=179!
> but zstd indeed provides two separately recommending inputBufferSize and 
> outputBufferSize  
> !image-2020-06-30-11-39-17-861.png!
>  
> *Workaround*
> One workaround,  using recommended in/out buffer size provided by zstd lib 
> can avoid the problem, but we don't know why. 
> input buffer size:  1301072 (128 * 1024)
> ouput buffer size: 131591 
> !image-2020-06-30-11-42-44-585.png!
>  
>  
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org