Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/16927
I'm worrying about this now: have my attempts to fix the messages gone
horribly wrong. Admittedly, it was sitting in a Budapest airport with a
post-ApacheCon hangover, but @afs was g
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r102977480
--- Diff: docs/streaming-programming-guide.md ---
@@ -630,35 +630,106 @@ which creates a DStream from text
data received over a TCP socket
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
@uncleGen: reviewed this, tweaked the docs slightly but otherwise, there's
nothing left to do that I can see
---
If your project is set up for it, you can reply to this email and have
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/16927
I could add the 2.6.5 binaries if you want, though the 2.6.4 ones should be
compatible.
I think I just lifted the 2.6.x artifacts out of an HDP build; its only the
2.7.x ones where I
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
(apologies for not replying; rebuilding a deceased laptop)
My main concern is to have the ability to make spark releases which include
the object store client libraries and a set of
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
Still waiting reviews for this. Anyone? Ideally before my forthcoming Spark
Summit talk...
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/16815
LGTM, though checkpointing to S3 has its own separate issues related to
rename performance and listing inconsistency. While this fix lets people
request different filesystems for the data
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
@nchammas sorry, should be clearer: "you must never use an aws-sdk version
other than the one hadoop-aws was built with, else things will break". if you
pull in hadoop-aws, that h
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r98449488
--- Diff: pom.xml ---
@@ -2586,6 +2591,100 @@
3.4.6
2.6.0
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r98448893
--- Diff:
cloud/src/test/scala/org/apache/spark/cloud/AzureInstantiationSuite.scala ---
@@ -0,0 +1,29 @@
+/*
--- End diff --
They
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
@nchammas the AWS SDK you get will be in sync with hadoop-aws; you have to
keep them in sync.
what is more brittle is the transients: httpclient, joda time, jackson,
etc, which is
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/9168
looking at the HDFS patch, it's in branch-2.9. We could backport to
branch-2.8, though it's too late to get into the 2.8.0 RC
---
If your project is set up for it, you can reply to
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
@uncleGen I've updated it. Note that
[HADOOP-13946](https://issues.apache.org/jira/browse/HADOOP-13946) tracks the
changes in the Hadoop docs, which writes down what HDFS actually does,
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r97367042
--- Diff:
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
---
@@ -196,29 +191,29 @@ class FileInputDStream[K, V, F
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r97363419
--- Diff:
streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala ---
@@ -235,18 +236,97 @@ class InputStreamsSuite extends
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
let me do a quick review & update
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this fea
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
latest patch: has updated the dependency settings. As noted, works for
Hadoop versions from 2.7 to 3.0.2-alpha & the HADOOP-13345 branch, at least if
you build the last two wi
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
Here's why this matters, and why a simple "isn't this just a matter of
dropping in the JARs" isn't the solution:
*getting getting the right jars together wit
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
this patch is ready for review. Anyone?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r95079472
--- Diff: docs/streaming-programming-guide.md ---
@@ -630,35 +630,106 @@ which creates a DStream from text
data received over a TCP socket
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
Sean, I think I've managed to delete the lines where you were asking about
globs
> Am I right that the net change here is not an optimization but an
expansion of the beh
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r94410195
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,17 +644,90 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r94407382
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,17 +644,90 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r94407115
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,17 +644,90 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
Test failure is pretty unlikely to be related. Looks more like a timing or
timeout problem.
```
org.apache.spark.rdd.AsyncRDDActionsSuite.async failure handling
Failing for
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
done
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
stylecheck; unexpected, as I thought I'd run them in the `mvn install` of
the module.
```
[error]
/home/jenkins/workspace/SparkPullRequestBuilder/core/src/test/scala/org/apache/
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/13579#discussion_r91548419
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala ---
@@ -289,6 +289,30 @@ object HistoryServer extends Logging
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/13579#discussion_r91546615
--- Diff:
core/src/test/scala/org/apache/spark/deploy/history/HistoryServerSuite.scala ---
@@ -349,6 +349,17 @@ class HistoryServerSuite extends
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/13579#discussion_r91546575
--- Diff:
core/src/test/scala/org/apache/spark/deploy/history/HistoryServerSuite.scala ---
@@ -349,6 +349,17 @@ class HistoryServerSuite extends
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
yeah, I've just got so many other distractions. Let me do it again while
tests run in different windows
---
If your project is set up for it, you can reply to this email and have your
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
Test failure due to new artifacts
```
+++ b/dev/pr-deps/spark-deps-hadoop-2.7
@@ -16,8 +16,6 @@ arpack_combined_all-0.1.jar
avro-1.7.7.jar
avro-ipc-1.7.7.jar
avro
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
the latest patch moves to the suggested name `spark-hadoop-cloud`; the
external test repo is in sync. Those test are all working happily against s3
ireland, Azure and rackspace swift, on
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/16089
I ask about committers as I'm staring at the V1 and V2 committer APIs right
now related to S3 destinations; not directly related to this though.
---
If your project is set up for it, yo
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/16089#discussion_r90502669
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonGenerator.scala
---
@@ -194,4 +194,8 @@ private[sql] class
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/16089
AFAIK, the big thing the FileOutputFormat really adds is not the
compression, but the output committer and the stuff to go with that (working
directories, paths, etc etc). If you aren't
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/16089#discussion_r90499765
--- Diff:
common/unsafe/src/main/java/org/apache/spark/unsafe/types/UTF8String.java ---
@@ -147,6 +147,17 @@ public void writeTo(ByteBuffer buffer
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/16089#discussion_r90497882
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/json/JacksonGenerator.scala
---
@@ -194,4 +194,8 @@ private[sql] class
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
The latest patch
1. keeps the cloud package separate from hadoop-2.7. This is important
avoid outstanding problems related to org.json licensed artifacts in the aws
SDK JARs. The
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14038#discussion_r89839965
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategySuite.scala
---
@@ -441,6 +441,44 @@ class
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15594#discussion_r89787175
--- Diff:
sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIService.scala
---
@@ -57,7 +59,24 @@ private[hive
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15991#discussion_r89786503
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/HistoryPage.scala ---
@@ -33,21 +33,40 @@ private[history] class HistoryPage(parent
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15648
LGTM, as the javadocs say *If true check only for Active NNs status, else
check first NN's status*. But I don't know enough about HDFS HA to be
It'll check the first
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89402090
--- Diff: pom.xml ---
@@ -2558,6 +2660,26 @@
+
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89352877
--- Diff: docs/cloud-integration.md ---
@@ -0,0 +1,953 @@
+---
+layout: global
+displayTitle: Integration with Cloud Infrastructures
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89346124
--- Diff:
cloud/src/test/scala/org/apache/spark/cloud/AzureInstantiationSuite.scala ---
@@ -0,0 +1,29 @@
+/*
--- End diff --
In the
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89340962
--- Diff: docs/cloud-integration.md ---
@@ -0,0 +1,953 @@
+---
+layout: global
+displayTitle: Integration with Cloud Infrastructures
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89340299
--- Diff: docs/cloud-integration.md ---
@@ -0,0 +1,953 @@
+---
+layout: global
+displayTitle: Integration with Cloud Infrastructures
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89340198
--- Diff: docs/storage-openstack-swift.md ---
@@ -19,41 +19,32 @@ Although not mandatory, it is recommended to configure
the proxy server of Swift
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89315595
--- Diff: pom.xml ---
@@ -2558,6 +2660,26 @@
+
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15984
Like you note, tests will be good here. Don't forget the corner cases:
unknown app, duplicate POSTs, known app but unknown attempt. I'm also curious
about what the policy would be i
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15984#discussion_r89312793
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala ---
@@ -121,6 +123,12 @@ class HistoryServer(
def initialize
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15984#discussion_r89312600
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -535,6 +535,26 @@ private[history] class
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15991#discussion_r89311812
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/HistoryPage.scala ---
@@ -33,21 +33,40 @@ private[history] class HistoryPage(parent
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89202156
--- Diff: docs/cloud-integration.md ---
@@ -0,0 +1,953 @@
+---
+layout: global
+displayTitle: Integration with Cloud Infrastructures
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r89176373
--- Diff: pom.xml ---
@@ -2558,6 +2660,26 @@
+
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
This is the patch stripped down to the packaging and some tests to load the
direct and indirect dependencies, so verifying that the classpath is valid
within the module itself. It also
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14038
@maropu if you create a PR for your work I'll comment on it
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your pr
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15869
The plugin point is more generic than ATS integration; it lets you stick
anything in to come up in the driver. Weakness: it's actually yarn specific; I
could imagine uses in standalon
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
2.6 vs 2.7 vs later releases âa moving target, with AWS versions and
other issues to worry about.
[HADOOP-13687](https://issues.apache.org/jira/browse/HADOOP-13687) is going to
add a
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
I had something tangible: the integration tests. It's clear those aren't
wanted. Now I'm proposing something more minimal, yet still tangible for anyone
trying to build spa
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
sean: there's two things: tests and packaging.
1. The packaging has to go in as probably the only way to get whatever
spark is built with to be consistent. That includes excl
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
Has anyone had a chance to review this? Is there more clarification needed,
or some specific aspect of the patch which needs changing?
Without this it is near-impossible to have a
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14646
thanks
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15594
I'm not spark committer so can't review it well enough to get in; I was
just watching it out of concern for the word "kerberos". How about you ask on
the spark de
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14646
Has anyone had a chance to review this. It's nicely self-contained, makes
it easier to use Spark as regression testing for ASF prerelease binaries of any
dependent project.
---
If
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15556#discussion_r84267482
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/ReplayListenerBus.scala ---
@@ -43,38 +43,56 @@ private[spark] class ReplayListenerBus
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15490
Oh, I see, new UI had meant I'd left the comment partially incomplete.
Sorry. Just the one: printing out the actual log dir location. That makes it
much easier to identify a configur
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15490#discussion_r83530966
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -262,6 +263,17 @@ private[history] class
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15490
@ajbozarth I'm not a spark committer, I'm not capaclbe of getting stuff in.
I did dd one comment to some of the code, otherwise nothing I have issues with.
LGTM
---
If your proj
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
that's it warning that the manifest has changed. Which it has: there's now
hadoop-azure, hadoop-openstack and hadoop-aws JARs on the CP, along with
dependencies (amazon-aws SDK,
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
@srowen have you got any comments on the last patch?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
I see: you want the HS to set it? Yeah, that would work. I'll change this
patch accordingly
---
If your project is set up for it, you can reply to this email and have your
reply appe
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/13579
The main problem here is in a cluster where auth is turned on globally, the
HS gets really confused: it's enabled but doesn't have any secrets. This patch
sets things up so that
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15374
Interesting to compare this with Hadoop's annotation, where I have mixed
opinions.
A key advantage Apache spark has is that Scala language lets you really
scope out thin
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/12004#discussion_r82502588
--- Diff: cloud/src/main/scala/org/apache/spark/cloud/s3/S3AConstants.scala
---
@@ -0,0 +1,75 @@
+/*
+ * Licensed to the Apache Software
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
# Packaging:
1. this addresses the problem that it's not always immediately obvious to
people what they have to do to get, say s3a working. Do you know precisely
which versi
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/15377#discussion_r82441129
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -2474,25 +2474,36 @@ private[spark] class CallerContext(
val context
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15137
I see this in master; but the JIRA associated with the PR is still opened &
unversioned. Which version did it make it into?
---
If your project is set up for it, you can reply to this e
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/12004
Test failure is because the list of 2.7.3 diffs has been expanded to
declare dependencies on the new hadoop JARs
---
If your project is set up for it, you can reply to this email and have
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14644
...so if you ask for 1 GPU you may only get 0?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14644
1. Any plans to add documentation?
2. What happens if you ask for (any, more) GPUs than there are?
2. If it fails, that could be a good test: ask for a very large number and
expect
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15115
thanks for clarifying; sometimes I feel that my patches get under reviewed
âwhich holds for Hadoop too, where some have been outstanding for so long
they're approach school age
-
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/15115
This seems a duplicate of the #14827 patch I filed 3 weeks earlier.
Is there some aspect of the PR submission process that I'm missing out on?
I would like to get my patches in
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r79209668
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,13 +644,44 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r78932943
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,13 +644,44 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14731#discussion_r78932779
--- Diff: docs/streaming-programming-guide.md ---
@@ -644,13 +644,44 @@ methods for creating DStreams from files as input
sources
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14827
I don't know what the default Hadoop version should be, that's the kind of
thing to discuss on mailing lists
personally, I'd rush to make 2.6 the bare minimum version
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14601
Could an automated test be done here. propagation can be tested with a a
function run on the
executor (such as a map) which fails if the required properties are missing
1. (set
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14601#discussion_r78521683
--- Diff: core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala
---
@@ -102,11 +102,20 @@ class SparkHadoopUtil extends Logging
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14601#discussion_r78520985
--- Diff: core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala
---
@@ -102,11 +102,20 @@ class SparkHadoopUtil extends Logging
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14827
Sean, the reason for a 2.7 profile is more significant with SPARK-7481 and
cloud support, as it can explicitly pull in hadoop-azure (2.7+ only) and
hadoop-aws (2.6+ only).
---
If your
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/9571#discussion_r77193230
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -664,6 +707,116 @@ private[history] class
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14731
The latest patch pulls out the shortcutting of the globStatus call if
there's no wildcard chars in the path; closer to the original patch
---
If your project is set up for it, you can
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/9571#discussion_r76972524
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -664,6 +707,116 @@ private[history] class
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/9571#discussion_r76952864
--- Diff:
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---
@@ -664,6 +707,116 @@ private[history] class
Github user steveloughran commented on the issue:
https://github.com/apache/spark/pull/14659
This context is just something passed over IPC to provide a general string
for the audit logs, the main actual access of it is in the HDFS audit log
```
HdfsAuditLogger
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14659#discussion_r76952473
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -2418,6 +2418,18 @@ private[spark] object Utils extends Logging
Github user steveloughran commented on a diff in the pull request:
https://github.com/apache/spark/pull/14659#discussion_r76951863
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -2418,6 +2418,18 @@ private[spark] object Utils extends Logging
401 - 500 of 1133 matches
Mail list logo