[SPARK-13928] Move org.apache.spark.Logging into org.apache.spark.internal.Logging
## What changes were proposed in this pull request? Logging was made private in Spark 2.0. If we move it, then users would be able to create a Logging trait themselves to avoid changing their own code. ## How was this patch tested? existing tests. Author: Wenchen Fan <wenc...@databricks.com> Closes #11764 from cloud-fan/logger. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8ef3399a Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8ef3399a Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8ef3399a Branch: refs/heads/master Commit: 8ef3399aff04bf8b7ab294c0f55bcf195995842b Parents: ea9ca6f Author: Wenchen Fan <wenc...@databricks.com> Authored: Thu Mar 17 19:23:38 2016 +0800 Committer: Wenchen Fan <wenc...@databricks.com> Committed: Thu Mar 17 19:23:38 2016 +0800 ---------------------------------------------------------------------- .../scala/org/apache/spark/Accumulator.scala | 1 + .../scala/org/apache/spark/ContextCleaner.scala | 1 + .../spark/ExecutorAllocationManager.scala | 1 + .../org/apache/spark/HeartbeatReceiver.scala | 1 + .../scala/org/apache/spark/HttpServer.scala | 1 + .../main/scala/org/apache/spark/Logging.scala | 167 ------------------- .../org/apache/spark/MapOutputTracker.scala | 1 + .../scala/org/apache/spark/SSLOptions.scala | 2 + .../org/apache/spark/SecurityManager.scala | 1 + .../main/scala/org/apache/spark/SparkConf.scala | 4 +- .../scala/org/apache/spark/SparkContext.scala | 1 + .../main/scala/org/apache/spark/SparkEnv.scala | 1 + .../org/apache/spark/SparkHadoopWriter.scala | 1 + .../org/apache/spark/TaskContextImpl.scala | 1 + .../scala/org/apache/spark/TaskEndReason.scala | 1 + .../spark/api/python/PythonGatewayServer.scala | 2 +- .../spark/api/python/PythonHadoopUtil.scala | 3 +- .../org/apache/spark/api/python/PythonRDD.scala | 1 + .../spark/api/python/PythonWorkerFactory.scala | 1 + .../org/apache/spark/api/python/SerDeUtil.scala | 3 +- .../scala/org/apache/spark/api/r/RBackend.scala | 3 +- .../apache/spark/api/r/RBackendHandler.scala | 2 +- .../scala/org/apache/spark/api/r/RRDD.scala | 1 + .../org/apache/spark/broadcast/Broadcast.scala | 2 +- .../spark/broadcast/BroadcastManager.scala | 3 +- .../spark/broadcast/TorrentBroadcast.scala | 1 + .../scala/org/apache/spark/deploy/Client.scala | 5 +- .../spark/deploy/ExternalShuffleService.scala | 3 +- .../spark/deploy/FaultToleranceTest.scala | 3 +- .../apache/spark/deploy/LocalSparkCluster.scala | 3 +- .../org/apache/spark/deploy/RPackageUtils.scala | 2 +- .../apache/spark/deploy/SparkCuratorUtil.scala | 3 +- .../apache/spark/deploy/SparkHadoopUtil.scala | 3 +- .../apache/spark/deploy/client/AppClient.scala | 5 +- .../spark/deploy/history/ApplicationCache.scala | 2 +- .../deploy/history/FsHistoryProvider.scala | 3 +- .../spark/deploy/history/HistoryServer.scala | 3 +- .../deploy/history/HistoryServerArguments.scala | 3 +- .../master/FileSystemPersistenceEngine.scala | 2 +- .../org/apache/spark/deploy/master/Master.scala | 3 +- .../deploy/master/RecoveryModeFactory.scala | 3 +- .../master/ZooKeeperLeaderElectionAgent.scala | 3 +- .../master/ZooKeeperPersistenceEngine.scala | 3 +- .../spark/deploy/master/ui/MasterWebUI.scala | 2 +- .../deploy/mesos/MesosClusterDispatcher.scala | 3 +- .../mesos/MesosExternalShuffleService.scala | 3 +- .../deploy/rest/RestSubmissionClient.scala | 3 +- .../deploy/rest/RestSubmissionServer.scala | 3 +- .../spark/deploy/worker/CommandUtils.scala | 2 +- .../spark/deploy/worker/DriverRunner.scala | 3 +- .../spark/deploy/worker/ExecutorRunner.scala | 3 +- .../org/apache/spark/deploy/worker/Worker.scala | 3 +- .../spark/deploy/worker/WorkerWatcher.scala | 2 +- .../apache/spark/deploy/worker/ui/LogPage.scala | 2 +- .../spark/deploy/worker/ui/WorkerWebUI.scala | 2 +- .../executor/CoarseGrainedExecutorBackend.scala | 1 + .../org/apache/spark/executor/Executor.scala | 1 + .../spark/executor/MesosExecutorBackend.scala | 3 +- .../org/apache/spark/executor/TaskMetrics.scala | 1 + .../input/FixedLengthBinaryInputFormat.scala | 2 +- .../org/apache/spark/internal/Logging.scala | 167 +++++++++++++++++++ .../spark/mapred/SparkHadoopMapRedUtil.scala | 3 +- .../spark/memory/ExecutionMemoryPool.scala | 2 +- .../org/apache/spark/memory/MemoryManager.scala | 3 +- .../apache/spark/memory/StorageMemoryPool.scala | 2 +- .../apache/spark/metrics/MetricsConfig.scala | 3 +- .../apache/spark/metrics/MetricsSystem.scala | 3 +- .../spark/network/BlockTransferService.scala | 4 +- .../network/netty/NettyBlockRpcServer.scala | 2 +- .../org/apache/spark/rdd/AsyncRDDActions.scala | 3 +- .../apache/spark/rdd/DoubleRDDFunctions.scala | 3 +- .../scala/org/apache/spark/rdd/HadoopRDD.scala | 1 + .../scala/org/apache/spark/rdd/JdbcRDD.scala | 3 +- .../spark/rdd/LocalRDDCheckpointData.scala | 3 +- .../org/apache/spark/rdd/NewHadoopRDD.scala | 1 + .../apache/spark/rdd/OrderedRDDFunctions.scala | 6 +- .../org/apache/spark/rdd/PairRDDFunctions.scala | 1 + .../main/scala/org/apache/spark/rdd/RDD.scala | 1 + .../apache/spark/rdd/RDDOperationScope.scala | 3 +- .../spark/rdd/ReliableCheckpointRDD.scala | 1 + .../spark/rdd/ReliableRDDCheckpointData.scala | 1 + .../spark/rdd/SequenceFileRDDFunctions.scala | 2 +- .../org/apache/spark/rpc/RpcEndpointRef.scala | 3 +- .../org/apache/spark/rpc/netty/Dispatcher.scala | 3 +- .../org/apache/spark/rpc/netty/Inbox.scala | 3 +- .../spark/rpc/netty/NettyRpcCallContext.scala | 2 +- .../apache/spark/rpc/netty/NettyRpcEnv.scala | 3 +- .../org/apache/spark/rpc/netty/Outbox.scala | 3 +- .../apache/spark/scheduler/DAGScheduler.scala | 2 +- .../spark/scheduler/EventLoggingListener.scala | 3 +- .../spark/scheduler/InputFormatInfo.scala | 2 +- .../org/apache/spark/scheduler/JobWaiter.scala | 2 +- .../scheduler/OutputCommitCoordinator.scala | 1 + .../scala/org/apache/spark/scheduler/Pool.scala | 2 +- .../spark/scheduler/ReplayListenerBus.scala | 2 +- .../spark/scheduler/SchedulableBuilder.scala | 3 +- .../apache/spark/scheduler/ShuffleMapTask.scala | 1 + .../apache/spark/scheduler/SparkListener.scala | 3 +- .../org/apache/spark/scheduler/Stage.scala | 1 + .../spark/scheduler/TaskResultGetter.scala | 1 + .../spark/scheduler/TaskSchedulerImpl.scala | 1 + .../apache/spark/scheduler/TaskSetManager.scala | 2 +- .../cluster/CoarseGrainedSchedulerBackend.scala | 3 +- .../cluster/SparkDeploySchedulerBackend.scala | 3 +- .../mesos/MesosClusterPersistenceEngine.scala | 3 +- .../mesos/MesosSchedulerBackendUtil.scala | 3 +- .../cluster/mesos/MesosSchedulerUtils.scala | 3 +- .../cluster/mesos/MesosTaskLaunchData.scala | 2 +- .../spark/scheduler/local/LocalBackend.scala | 3 +- .../spark/serializer/KryoSerializer.scala | 1 + .../serializer/SerializationDebugger.scala | 2 +- .../spark/shuffle/BlockStoreShuffleReader.scala | 1 + .../shuffle/FileShuffleBlockResolver.scala | 3 +- .../shuffle/IndexShuffleBlockResolver.scala | 3 +- .../spark/shuffle/hash/HashShuffleManager.scala | 1 + .../spark/shuffle/hash/HashShuffleWriter.scala | 1 + .../spark/shuffle/sort/SortShuffleManager.scala | 1 + .../spark/shuffle/sort/SortShuffleWriter.scala | 1 + .../api/v1/EventLogDownloadResource.scala | 3 +- .../apache/spark/storage/BlockInfoManager.scala | 3 +- .../org/apache/spark/storage/BlockManager.scala | 1 + .../spark/storage/BlockManagerMaster.scala | 3 +- .../storage/BlockManagerMasterEndpoint.scala | 3 +- .../storage/BlockManagerSlaveEndpoint.scala | 3 +- .../apache/spark/storage/DiskBlockManager.scala | 3 +- .../spark/storage/DiskBlockObjectWriter.scala | 2 +- .../org/apache/spark/storage/DiskStore.scala | 3 +- .../storage/ShuffleBlockFetcherIterator.scala | 3 +- .../spark/storage/memory/MemoryStore.scala | 3 +- .../apache/spark/ui/ConsoleProgressBar.scala | 1 + .../scala/org/apache/spark/ui/JettyUtils.scala | 3 +- .../scala/org/apache/spark/ui/SparkUI.scala | 3 +- .../scala/org/apache/spark/ui/UIUtils.scala | 2 +- .../main/scala/org/apache/spark/ui/WebUI.scala | 3 +- .../spark/ui/jobs/JobProgressListener.scala | 1 + .../spark/ui/scope/RDDOperationGraph.scala | 2 +- .../org/apache/spark/util/ClosureCleaner.scala | 3 +- .../scala/org/apache/spark/util/EventLoop.scala | 2 +- .../org/apache/spark/util/ListenerBus.scala | 2 +- .../apache/spark/util/ShutdownHookManager.scala | 2 +- .../org/apache/spark/util/SizeEstimator.scala | 2 +- .../util/SparkUncaughtExceptionHandler.scala | 2 +- .../apache/spark/util/TimeStampedHashMap.scala | 2 +- .../scala/org/apache/spark/util/Utils.scala | 1 + .../util/collection/ExternalAppendOnlyMap.scala | 3 +- .../spark/util/collection/ExternalSorter.scala | 1 + .../spark/util/collection/Spillable.scala | 3 +- .../spark/util/logging/FileAppender.scala | 3 +- .../spark/util/logging/RollingPolicy.scala | 2 +- .../util/random/StratifiedSamplingUtils.scala | 2 +- .../org/apache/spark/ContextCleanerSuite.scala | 1 + .../SparkContextSchedulerCreationSuite.scala | 1 + .../scala/org/apache/spark/SparkFunSuite.scala | 2 + .../scala/org/apache/spark/ThreadingSuite.scala | 2 + .../apache/spark/deploy/SparkSubmitSuite.scala | 1 + .../spark/deploy/client/AppClientSuite.scala | 1 + .../deploy/history/ApplicationCacheSuite.scala | 3 +- .../deploy/history/FsHistoryProviderSuite.scala | 3 +- .../input/WholeTextFileRecordReaderSuite.scala | 3 +- .../org/apache/spark/rdd/SortingSuite.scala | 3 +- .../scheduler/EventLoggingListenerSuite.scala | 1 + .../scheduler/TaskSchedulerImplSuite.scala | 1 + .../spark/scheduler/TaskSetManagerSuite.scala | 1 + .../apache/spark/util/FileAppenderSuite.scala | 3 +- .../org/apache/spark/util/UtilsSuite.scala | 3 +- .../spark/util/collection/SorterSuite.scala | 3 +- .../spark/examples/graphx/Analytics.scala | 1 + .../examples/streaming/CustomReceiver.scala | 3 +- .../examples/streaming/StreamingExamples.scala | 2 +- .../kafka/DirectKafkaInputDStream.scala | 3 +- .../streaming/kafka/KafkaInputDStream.scala | 2 +- .../apache/spark/streaming/kafka/KafkaRDD.scala | 3 +- .../spark/streaming/kafka/KafkaTestUtils.scala | 3 +- .../streaming/kafka/ReliableKafkaReceiver.scala | 3 +- .../kafka/DirectKafkaStreamSuite.scala | 3 +- .../streaming/KinesisWordCountASL.scala | 3 +- .../kinesis/KinesisBackedBlockRDD.scala | 1 + .../streaming/kinesis/KinesisCheckpointer.scala | 2 +- .../streaming/kinesis/KinesisReceiver.scala | 2 +- .../kinesis/KinesisRecordProcessor.scala | 2 +- .../streaming/kinesis/KinesisTestUtils.scala | 2 +- .../org/apache/spark/graphx/GraphLoader.scala | 3 +- .../scala/org/apache/spark/graphx/Pregel.scala | 2 +- .../graphx/impl/VertexPartitionBaseOps.scala | 2 +- .../org/apache/spark/graphx/lib/PageRank.scala | 2 +- .../spark/graphx/util/GraphGenerators.scala | 1 + .../scala/org/apache/spark/ml/Pipeline.scala | 3 +- .../scala/org/apache/spark/ml/Transformer.scala | 2 +- .../spark/ml/classification/GBTClassifier.scala | 2 +- .../ml/classification/LogisticRegression.scala | 3 +- .../org/apache/spark/ml/clustering/LDA.scala | 2 +- .../spark/ml/feature/QuantileDiscretizer.scala | 2 +- .../IterativelyReweightedLeastSquares.scala | 2 +- .../spark/ml/optim/WeightedLeastSquares.scala | 2 +- .../apache/spark/ml/recommendation/ALS.scala | 3 +- .../ml/regression/AFTSurvivalRegression.scala | 3 +- .../spark/ml/regression/GBTRegressor.scala | 2 +- .../GeneralizedLinearRegression.scala | 3 +- .../ml/regression/IsotonicRegression.scala | 2 +- .../spark/ml/regression/LinearRegression.scala | 3 +- .../ml/tree/impl/GradientBoostedTrees.scala | 2 +- .../apache/spark/ml/tree/impl/NodeIdCache.scala | 2 +- .../spark/ml/tree/impl/RandomForest.scala | 2 +- .../apache/spark/ml/tuning/CrossValidator.scala | 3 +- .../spark/ml/tuning/TrainValidationSplit.scala | 2 +- .../org/apache/spark/ml/util/ReadWrite.scala | 3 +- .../spark/mllib/classification/NaiveBayes.scala | 3 +- .../mllib/clustering/BisectingKMeans.scala | 2 +- .../mllib/clustering/BisectingKMeansModel.scala | 2 +- .../apache/spark/mllib/clustering/KMeans.scala | 2 +- .../org/apache/spark/mllib/clustering/LDA.scala | 2 +- .../spark/mllib/clustering/LocalKMeans.scala | 2 +- .../clustering/PowerIterationClustering.scala | 3 +- .../mllib/clustering/StreamingKMeans.scala | 2 +- .../BinaryClassificationMetrics.scala | 2 +- .../spark/mllib/evaluation/RankingMetrics.scala | 2 +- .../mllib/evaluation/RegressionMetrics.scala | 4 +- .../spark/mllib/feature/StandardScaler.scala | 2 +- .../apache/spark/mllib/feature/Word2Vec.scala | 3 +- .../spark/mllib/fpm/AssociationRules.scala | 2 +- .../org/apache/spark/mllib/fpm/FPGrowth.scala | 3 +- .../spark/mllib/fpm/LocalPrefixSpan.scala | 2 +- .../org/apache/spark/mllib/fpm/PrefixSpan.scala | 2 +- .../spark/mllib/impl/PeriodicCheckpointer.scala | 3 +- .../org/apache/spark/mllib/linalg/BLAS.scala | 2 +- .../mllib/linalg/distributed/BlockMatrix.scala | 3 +- .../mllib/linalg/distributed/RowMatrix.scala | 2 +- .../mllib/optimization/GradientDescent.scala | 2 +- .../apache/spark/mllib/optimization/LBFGS.scala | 2 +- .../apache/spark/mllib/recommendation/ALS.scala | 2 +- .../MatrixFactorizationModel.scala | 3 +- .../regression/GeneralizedLinearAlgorithm.scala | 3 +- .../regression/StreamingLinearAlgorithm.scala | 2 +- .../stat/correlation/PearsonCorrelation.scala | 2 +- .../stat/correlation/SpearmanCorrelation.scala | 2 +- .../spark/mllib/stat/test/ChiSqTest.scala | 3 +- .../mllib/stat/test/KolmogorovSmirnovTest.scala | 2 +- .../spark/mllib/stat/test/StreamingTest.scala | 2 +- .../mllib/stat/test/StreamingTestMethod.scala | 2 +- .../apache/spark/mllib/tree/DecisionTree.scala | 2 +- .../spark/mllib/tree/GradientBoostedTrees.scala | 2 +- .../apache/spark/mllib/tree/RandomForest.scala | 2 +- .../mllib/tree/impl/DecisionTreeMetadata.scala | 2 +- .../mllib/tree/model/DecisionTreeModel.scala | 3 +- .../apache/spark/mllib/tree/model/Node.scala | 2 +- .../mllib/tree/model/treeEnsembleModels.scala | 3 +- .../spark/mllib/util/DataValidators.scala | 2 +- .../spark/ml/feature/VectorIndexerSuite.scala | 3 +- .../spark/ml/recommendation/ALSSuite.scala | 4 +- .../spark/mllib/linalg/VectorsSuite.scala | 3 +- .../spark/mllib/stat/CorrelationSuite.scala | 3 +- .../mllib/tree/GradientBoostedTreesSuite.scala | 3 +- project/MimaExcludes.scala | 8 + .../main/scala/org/apache/spark/repl/Main.scala | 4 +- .../org/apache/spark/repl/SparkExprTyper.scala | 2 +- .../org/apache/spark/repl/SparkILoop.scala | 6 +- .../org/apache/spark/repl/SparkIMain.scala | 3 +- .../spark/repl/SparkJLineCompletion.scala | 6 +- .../main/scala/org/apache/spark/repl/Main.scala | 1 + .../apache/spark/repl/ExecutorClassLoader.scala | 3 +- .../spark/repl/ExecutorClassLoaderSuite.scala | 1 + .../catalyst/expressions/BoundAttribute.scala | 2 +- .../expressions/codegen/CodeGenerator.scala | 2 +- .../expressions/codegen/GenerateOrdering.scala | 2 +- .../spark/sql/catalyst/parser/ParseDriver.scala | 2 +- .../sql/catalyst/planning/QueryPlanner.scala | 2 +- .../spark/sql/catalyst/planning/patterns.scala | 2 +- .../catalyst/plans/logical/LogicalPlan.scala | 2 +- .../apache/spark/sql/catalyst/rules/Rule.scala | 2 +- .../spark/sql/catalyst/rules/RuleExecutor.scala | 2 +- .../spark/sql/catalyst/trees/package.scala | 2 +- .../scala/org/apache/spark/sql/Column.scala | 2 +- .../org/apache/spark/sql/DataFrameReader.scala | 3 +- .../scala/org/apache/spark/sql/SQLContext.scala | 3 +- .../org/apache/spark/sql/UDFRegistration.scala | 2 +- .../spark/sql/execution/CacheManager.scala | 2 +- .../apache/spark/sql/execution/SparkPlan.scala | 3 +- .../aggregate/AggregationIterator.scala | 2 +- .../aggregate/TungstenAggregationIterator.scala | 3 +- .../aggregate/TypedAggregateExpression.scala | 2 +- .../spark/sql/execution/aggregate/udaf.scala | 2 +- .../columnar/GenerateColumnAccessor.scala | 2 +- .../compression/CompressibleColumnBuilder.scala | 2 +- .../spark/sql/execution/command/commands.scala | 2 +- .../spark/sql/execution/command/ddl.scala | 2 +- .../sql/execution/datasources/DataSource.scala | 2 +- .../datasources/DataSourceStrategy.scala | 3 +- .../datasources/FileSourceStrategy.scala | 2 +- .../execution/datasources/SqlNewHadoopRDD.scala | 1 + .../execution/datasources/WriterContainer.scala | 1 + .../execution/datasources/csv/CSVOptions.scala | 2 +- .../execution/datasources/csv/CSVParser.scala | 2 +- .../execution/datasources/csv/CSVRelation.scala | 2 +- .../datasources/jdbc/DriverRegistry.scala | 2 +- .../execution/datasources/jdbc/JDBCRDD.scala | 3 +- .../execution/datasources/jdbc/JdbcUtils.scala | 2 +- .../datasources/json/JSONRelation.scala | 2 +- .../parquet/CatalystReadSupport.scala | 2 +- .../parquet/CatalystRowConverter.scala | 2 +- .../parquet/CatalystWriteSupport.scala | 2 +- .../datasources/parquet/ParquetRelation.scala | 3 +- .../spark/sql/execution/debug/package.scala | 3 +- .../exchange/ExchangeCoordinator.scala | 3 +- .../spark/sql/execution/python/PythonUDF.scala | 2 +- .../sql/execution/stat/FrequentItems.scala | 2 +- .../sql/execution/stat/StatFunctions.scala | 2 +- .../execution/streaming/FileStreamSource.scala | 2 +- .../execution/streaming/StreamExecution.scala | 2 +- .../spark/sql/execution/streaming/memory.scala | 3 +- .../sql/execution/ui/AllExecutionsPage.scala | 2 +- .../spark/sql/execution/ui/ExecutionPage.scala | 2 +- .../spark/sql/execution/ui/SQLListener.scala | 3 +- .../apache/spark/sql/execution/ui/SQLTab.scala | 2 +- .../org/apache/spark/sql/internal/SQLConf.scala | 2 +- .../apache/spark/sql/sources/interfaces.scala | 3 +- .../spark/sql/util/QueryExecutionListener.scala | 2 +- .../apache/spark/sql/DataFrameStatSuite.scala | 2 +- .../execution/columnar/ColumnTypeSuite.scala | 3 +- .../hive/thriftserver/HiveThriftServer2.scala | 3 +- .../SparkExecuteStatementOperation.scala | 2 +- .../hive/thriftserver/SparkSQLCLIDriver.scala | 2 +- .../sql/hive/thriftserver/SparkSQLDriver.scala | 2 +- .../sql/hive/thriftserver/SparkSQLEnv.scala | 3 +- .../server/SparkSQLOperationManager.scala | 2 +- .../hive/thriftserver/ui/ThriftServerPage.scala | 2 +- .../ui/ThriftServerSessionPage.scala | 2 +- .../hive/thriftserver/ui/ThriftServerTab.scala | 3 +- .../spark/sql/hive/thriftserver/CliSuite.scala | 3 +- .../thriftserver/HiveThriftServer2Suites.scala | 3 +- .../org/apache/spark/sql/hive/HiveCatalog.scala | 2 +- .../org/apache/spark/sql/hive/HiveContext.scala | 3 +- .../spark/sql/hive/HiveMetastoreCatalog.scala | 2 +- .../org/apache/spark/sql/hive/HiveQl.scala | 2 +- .../org/apache/spark/sql/hive/HiveShim.scala | 2 +- .../org/apache/spark/sql/hive/SQLBuilder.scala | 2 +- .../org/apache/spark/sql/hive/TableReader.scala | 2 +- .../spark/sql/hive/client/HiveClientImpl.scala | 3 +- .../apache/spark/sql/hive/client/HiveShim.scala | 2 +- .../sql/hive/client/IsolatedClientLoader.scala | 3 +- .../hive/execution/ScriptTransformation.scala | 3 +- .../org/apache/spark/sql/hive/hiveUDFs.scala | 2 +- .../spark/sql/hive/hiveWriterContainers.scala | 1 + .../spark/sql/hive/orc/OrcFileOperator.scala | 2 +- .../apache/spark/sql/hive/orc/OrcFilters.scala | 2 +- .../apache/spark/sql/hive/orc/OrcRelation.scala | 2 +- .../spark/sql/hive/HiveSparkSubmitSuite.scala | 1 + .../spark/sql/hive/client/FiltersSuite.scala | 3 +- .../spark/sql/hive/client/VersionsSuite.scala | 3 +- .../org/apache/spark/streaming/Checkpoint.scala | 3 +- .../apache/spark/streaming/DStreamGraph.scala | 2 +- .../spark/streaming/StreamingContext.scala | 1 + .../spark/streaming/dstream/DStream.scala | 3 +- .../dstream/DStreamCheckpointData.scala | 2 +- .../streaming/dstream/RawInputDStream.scala | 2 +- .../streaming/dstream/SocketInputDStream.scala | 2 +- .../streaming/receiver/BlockGenerator.scala | 3 +- .../spark/streaming/receiver/RateLimiter.scala | 3 +- .../receiver/ReceivedBlockHandler.scala | 3 +- .../streaming/receiver/ReceiverSupervisor.scala | 3 +- .../receiver/ReceiverSupervisorImpl.scala | 3 +- .../streaming/scheduler/InputInfoTracker.scala | 2 +- .../streaming/scheduler/JobGenerator.scala | 3 +- .../streaming/scheduler/JobScheduler.scala | 2 +- .../scheduler/ReceivedBlockTracker.scala | 3 +- .../streaming/scheduler/ReceiverTracker.scala | 1 + .../scheduler/rate/PIDRateEstimator.scala | 2 +- .../spark/streaming/ui/StreamingPage.scala | 2 +- .../spark/streaming/ui/StreamingTab.scala | 3 +- .../streaming/util/BatchedWriteAheadLog.scala | 3 +- .../streaming/util/FileBasedWriteAheadLog.scala | 3 +- .../util/FileBasedWriteAheadLogReader.scala | 2 +- .../util/RateLimitedOutputStream.scala | 2 +- .../spark/streaming/util/RawTextSender.scala | 3 +- .../spark/streaming/util/RecurringTimer.scala | 2 +- .../streaming/util/WriteAheadLogUtils.scala | 3 +- .../apache/spark/streaming/FailureSuite.scala | 1 + .../spark/streaming/InputStreamsSuite.scala | 2 +- .../spark/streaming/MasterFailureTest.scala | 2 +- .../streaming/ReceivedBlockHandlerSuite.scala | 1 + .../streaming/ReceivedBlockTrackerSuite.scala | 3 +- .../spark/streaming/StreamingContextSuite.scala | 1 + .../streaming/StreamingListenerSuite.scala | 2 +- .../apache/spark/streaming/TestSuiteBase.scala | 3 +- .../deploy/yarn/AMDelegationTokenRenewer.scala | 3 +- .../spark/deploy/yarn/ApplicationMaster.scala | 1 + .../org/apache/spark/deploy/yarn/Client.scala | 3 +- .../yarn/ClientDistributedCacheManager.scala | 2 +- .../yarn/ExecutorDelegationTokenUpdater.scala | 3 +- .../spark/deploy/yarn/ExecutorRunnable.scala | 3 +- .../spark/deploy/yarn/YarnAllocator.scala | 3 +- .../apache/spark/deploy/yarn/YarnRMClient.scala | 3 +- .../cluster/SchedulerExtensionService.scala | 3 +- .../cluster/YarnClientSchedulerBackend.scala | 3 +- .../cluster/YarnSchedulerBackend.scala | 3 +- .../deploy/yarn/BaseYarnClusterSuite.scala | 1 + .../spark/deploy/yarn/YarnClusterSuite.scala | 1 + .../yarn/YarnShuffleIntegrationSuite.scala | 1 + .../deploy/yarn/YarnSparkHadoopUtilSuite.scala | 3 +- .../ExtensionServiceIntegrationSuite.scala | 3 +- 399 files changed, 742 insertions(+), 510 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/Accumulator.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/Accumulator.scala b/core/src/main/scala/org/apache/spark/Accumulator.scala index 0e4bcc3..9857529 100644 --- a/core/src/main/scala/org/apache/spark/Accumulator.scala +++ b/core/src/main/scala/org/apache/spark/Accumulator.scala @@ -23,6 +23,7 @@ import javax.annotation.concurrent.GuardedBy import scala.collection.mutable import scala.ref.WeakReference +import org.apache.spark.internal.Logging import org.apache.spark.storage.{BlockId, BlockStatus} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/ContextCleaner.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/ContextCleaner.scala b/core/src/main/scala/org/apache/spark/ContextCleaner.scala index 17014e4..8fc657c 100644 --- a/core/src/main/scala/org/apache/spark/ContextCleaner.scala +++ b/core/src/main/scala/org/apache/spark/ContextCleaner.scala @@ -23,6 +23,7 @@ import java.util.concurrent.{ConcurrentLinkedQueue, ScheduledExecutorService, Ti import scala.collection.JavaConverters._ import org.apache.spark.broadcast.Broadcast +import org.apache.spark.internal.Logging import org.apache.spark.rdd.{RDD, ReliableRDDCheckpointData} import org.apache.spark.util.{ThreadUtils, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala index 9b8279f..0926d05 100644 --- a/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala +++ b/core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala @@ -24,6 +24,7 @@ import scala.util.control.ControlThrowable import com.codahale.metrics.{Gauge, MetricRegistry} +import org.apache.spark.internal.Logging import org.apache.spark.metrics.source.Source import org.apache.spark.scheduler._ import org.apache.spark.util.{Clock, SystemClock, ThreadUtils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala b/core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala index 7f474ed..e8748dd 100644 --- a/core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala +++ b/core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala @@ -22,6 +22,7 @@ import java.util.concurrent.{ScheduledFuture, TimeUnit} import scala.collection.mutable import scala.concurrent.Future +import org.apache.spark.internal.Logging import org.apache.spark.rpc.{RpcCallContext, RpcEnv, ThreadSafeRpcEndpoint} import org.apache.spark.scheduler._ import org.apache.spark.storage.BlockManagerId http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/HttpServer.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/HttpServer.scala b/core/src/main/scala/org/apache/spark/HttpServer.scala index 3c80842..9fad1f6 100644 --- a/core/src/main/scala/org/apache/spark/HttpServer.scala +++ b/core/src/main/scala/org/apache/spark/HttpServer.scala @@ -28,6 +28,7 @@ import org.eclipse.jetty.servlet.{DefaultServlet, ServletContextHandler, Servlet import org.eclipse.jetty.util.security.{Constraint, Password} import org.eclipse.jetty.util.thread.QueuedThreadPool +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/Logging.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/Logging.scala b/core/src/main/scala/org/apache/spark/Logging.scala deleted file mode 100644 index efab61e..0000000 --- a/core/src/main/scala/org/apache/spark/Logging.scala +++ /dev/null @@ -1,167 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the "License"); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.spark - -import org.apache.log4j.{Level, LogManager, PropertyConfigurator} -import org.slf4j.{Logger, LoggerFactory} -import org.slf4j.impl.StaticLoggerBinder - -import org.apache.spark.util.Utils - -/** - * Utility trait for classes that want to log data. Creates a SLF4J logger for the class and allows - * logging messages at different levels using methods that only evaluate parameters lazily if the - * log level is enabled. - */ -private[spark] trait Logging { - - // Make the log field transient so that objects with Logging can - // be serialized and used on another machine - @transient private var log_ : Logger = null - - // Method to get the logger name for this object - protected def logName = { - // Ignore trailing $'s in the class names for Scala objects - this.getClass.getName.stripSuffix("$") - } - - // Method to get or create the logger for this object - protected def log: Logger = { - if (log_ == null) { - initializeLogIfNecessary(false) - log_ = LoggerFactory.getLogger(logName) - } - log_ - } - - // Log methods that take only a String - protected def logInfo(msg: => String) { - if (log.isInfoEnabled) log.info(msg) - } - - protected def logDebug(msg: => String) { - if (log.isDebugEnabled) log.debug(msg) - } - - protected def logTrace(msg: => String) { - if (log.isTraceEnabled) log.trace(msg) - } - - protected def logWarning(msg: => String) { - if (log.isWarnEnabled) log.warn(msg) - } - - protected def logError(msg: => String) { - if (log.isErrorEnabled) log.error(msg) - } - - // Log methods that take Throwables (Exceptions/Errors) too - protected def logInfo(msg: => String, throwable: Throwable) { - if (log.isInfoEnabled) log.info(msg, throwable) - } - - protected def logDebug(msg: => String, throwable: Throwable) { - if (log.isDebugEnabled) log.debug(msg, throwable) - } - - protected def logTrace(msg: => String, throwable: Throwable) { - if (log.isTraceEnabled) log.trace(msg, throwable) - } - - protected def logWarning(msg: => String, throwable: Throwable) { - if (log.isWarnEnabled) log.warn(msg, throwable) - } - - protected def logError(msg: => String, throwable: Throwable) { - if (log.isErrorEnabled) log.error(msg, throwable) - } - - protected def isTraceEnabled(): Boolean = { - log.isTraceEnabled - } - - protected def initializeLogIfNecessary(isInterpreter: Boolean): Unit = { - if (!Logging.initialized) { - Logging.initLock.synchronized { - if (!Logging.initialized) { - initializeLogging(isInterpreter) - } - } - } - } - - private def initializeLogging(isInterpreter: Boolean): Unit = { - // Don't use a logger in here, as this is itself occurring during initialization of a logger - // If Log4j 1.2 is being used, but is not initialized, load a default properties file - val binderClass = StaticLoggerBinder.getSingleton.getLoggerFactoryClassStr - // This distinguishes the log4j 1.2 binding, currently - // org.slf4j.impl.Log4jLoggerFactory, from the log4j 2.0 binding, currently - // org.apache.logging.slf4j.Log4jLoggerFactory - val usingLog4j12 = "org.slf4j.impl.Log4jLoggerFactory".equals(binderClass) - if (usingLog4j12) { - val log4j12Initialized = LogManager.getRootLogger.getAllAppenders.hasMoreElements - // scalastyle:off println - if (!log4j12Initialized) { - val defaultLogProps = "org/apache/spark/log4j-defaults.properties" - Option(Utils.getSparkClassLoader.getResource(defaultLogProps)) match { - case Some(url) => - PropertyConfigurator.configure(url) - System.err.println(s"Using Spark's default log4j profile: $defaultLogProps") - case None => - System.err.println(s"Spark was unable to load $defaultLogProps") - } - } - - if (isInterpreter) { - // Use the repl's main class to define the default log level when running the shell, - // overriding the root logger's config if they're different. - val rootLogger = LogManager.getRootLogger() - val replLogger = LogManager.getLogger(logName) - val replLevel = Option(replLogger.getLevel()).getOrElse(Level.WARN) - if (replLevel != rootLogger.getEffectiveLevel()) { - System.err.printf("Setting default log level to \"%s\".\n", replLevel) - System.err.println("To adjust logging level use sc.setLogLevel(newLevel).") - rootLogger.setLevel(replLevel) - } - } - // scalastyle:on println - } - Logging.initialized = true - - // Force a call into slf4j to initialize it. Avoids this happening from multiple threads - // and triggering this: http://mailman.qos.ch/pipermail/slf4j-dev/2010-April/002956.html - log - } -} - -private object Logging { - @volatile private var initialized = false - val initLock = new Object() - try { - // We use reflection here to handle the case where users remove the - // slf4j-to-jul bridge order to route their logs to JUL. - val bridgeClass = Utils.classForName("org.slf4j.bridge.SLF4JBridgeHandler") - bridgeClass.getMethod("removeHandlersForRootLogger").invoke(null) - val installed = bridgeClass.getMethod("isInstalled").invoke(null).asInstanceOf[Boolean] - if (!installed) { - bridgeClass.getMethod("install").invoke(null) - } - } catch { - case e: ClassNotFoundException => // can't log anything yet so just fail silently - } -} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/MapOutputTracker.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/MapOutputTracker.scala b/core/src/main/scala/org/apache/spark/MapOutputTracker.scala index 9cb6159..3a5caa3 100644 --- a/core/src/main/scala/org/apache/spark/MapOutputTracker.scala +++ b/core/src/main/scala/org/apache/spark/MapOutputTracker.scala @@ -25,6 +25,7 @@ import scala.collection.JavaConverters._ import scala.collection.mutable.{ArrayBuffer, HashMap, HashSet, Map} import scala.reflect.ClassTag +import org.apache.spark.internal.Logging import org.apache.spark.rpc.{RpcCallContext, RpcEndpoint, RpcEndpointRef, RpcEnv} import org.apache.spark.scheduler.MapStatus import org.apache.spark.shuffle.MetadataFetchFailedException http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SSLOptions.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SSLOptions.scala b/core/src/main/scala/org/apache/spark/SSLOptions.scala index d755f07..30db6cc 100644 --- a/core/src/main/scala/org/apache/spark/SSLOptions.scala +++ b/core/src/main/scala/org/apache/spark/SSLOptions.scala @@ -23,6 +23,8 @@ import javax.net.ssl.SSLContext import org.eclipse.jetty.util.ssl.SslContextFactory +import org.apache.spark.internal.Logging + /** * SSLOptions class is a common container for SSL configuration options. It offers methods to * generate specific objects to configure SSL for different communication protocols. http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SecurityManager.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SecurityManager.scala b/core/src/main/scala/org/apache/spark/SecurityManager.scala index 6132fa3..e8f6822 100644 --- a/core/src/main/scala/org/apache/spark/SecurityManager.scala +++ b/core/src/main/scala/org/apache/spark/SecurityManager.scala @@ -28,6 +28,7 @@ import com.google.common.io.Files import org.apache.hadoop.io.Text import org.apache.spark.deploy.SparkHadoopUtil +import org.apache.spark.internal.Logging import org.apache.spark.network.sasl.SecretKeyHolder import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SparkConf.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SparkConf.scala b/core/src/main/scala/org/apache/spark/SparkConf.scala index 22e9c5e..5da2e98 100644 --- a/core/src/main/scala/org/apache/spark/SparkConf.scala +++ b/core/src/main/scala/org/apache/spark/SparkConf.scala @@ -17,15 +17,15 @@ package org.apache.spark -import java.util.concurrent.{ConcurrentHashMap, TimeUnit} +import java.util.concurrent.ConcurrentHashMap import scala.collection.JavaConverters._ import scala.collection.mutable.LinkedHashSet import org.apache.avro.{Schema, SchemaNormalization} +import org.apache.spark.internal.Logging import org.apache.spark.internal.config.{ConfigEntry, OptionalConfigEntry} -import org.apache.spark.network.util.JavaUtils import org.apache.spark.serializer.KryoSerializer import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SparkContext.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala b/core/src/main/scala/org/apache/spark/SparkContext.scala index 5c7ae57..d2cf3bf 100644 --- a/core/src/main/scala/org/apache/spark/SparkContext.scala +++ b/core/src/main/scala/org/apache/spark/SparkContext.scala @@ -50,6 +50,7 @@ import org.apache.spark.broadcast.Broadcast import org.apache.spark.deploy.{LocalSparkCluster, SparkHadoopUtil} import org.apache.spark.input.{FixedLengthBinaryInputFormat, PortableDataStream, StreamInputFormat, WholeTextFileInputFormat} +import org.apache.spark.internal.Logging import org.apache.spark.io.CompressionCodec import org.apache.spark.partial.{ApproximateEvaluator, PartialResult} import org.apache.spark.rdd._ http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SparkEnv.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SparkEnv.scala b/core/src/main/scala/org/apache/spark/SparkEnv.scala index 668a913..459fab8 100644 --- a/core/src/main/scala/org/apache/spark/SparkEnv.scala +++ b/core/src/main/scala/org/apache/spark/SparkEnv.scala @@ -28,6 +28,7 @@ import com.google.common.collect.MapMaker import org.apache.spark.annotation.DeveloperApi import org.apache.spark.api.python.PythonWorkerFactory import org.apache.spark.broadcast.BroadcastManager +import org.apache.spark.internal.Logging import org.apache.spark.memory.{MemoryManager, StaticMemoryManager, UnifiedMemoryManager} import org.apache.spark.metrics.MetricsSystem import org.apache.spark.network.BlockTransferService http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala b/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala index 5864786..17daac1 100644 --- a/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala +++ b/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala @@ -27,6 +27,7 @@ import org.apache.hadoop.fs.Path import org.apache.hadoop.mapred._ import org.apache.hadoop.mapreduce.TaskType +import org.apache.spark.internal.Logging import org.apache.spark.mapred.SparkHadoopMapRedUtil import org.apache.spark.rdd.HadoopRDD import org.apache.spark.util.SerializableJobConf http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala index 7e96040..c9354b3 100644 --- a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala +++ b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala @@ -20,6 +20,7 @@ package org.apache.spark import scala.collection.mutable.ArrayBuffer import org.apache.spark.executor.TaskMetrics +import org.apache.spark.internal.Logging import org.apache.spark.memory.TaskMemoryManager import org.apache.spark.metrics.MetricsSystem import org.apache.spark.metrics.source.Source http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/TaskEndReason.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/TaskEndReason.scala b/core/src/main/scala/org/apache/spark/TaskEndReason.scala index 509fb2e..83af226 100644 --- a/core/src/main/scala/org/apache/spark/TaskEndReason.scala +++ b/core/src/main/scala/org/apache/spark/TaskEndReason.scala @@ -23,6 +23,7 @@ import scala.util.Try import org.apache.spark.annotation.DeveloperApi import org.apache.spark.executor.TaskMetrics +import org.apache.spark.internal.Logging import org.apache.spark.scheduler.AccumulableInfo import org.apache.spark.storage.BlockManagerId import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/python/PythonGatewayServer.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonGatewayServer.scala b/core/src/main/scala/org/apache/spark/api/python/PythonGatewayServer.scala index 164e950..6c40722 100644 --- a/core/src/main/scala/org/apache/spark/api/python/PythonGatewayServer.scala +++ b/core/src/main/scala/org/apache/spark/api/python/PythonGatewayServer.scala @@ -22,7 +22,7 @@ import java.net.Socket import py4j.GatewayServer -import org.apache.spark.Logging +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/python/PythonHadoopUtil.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonHadoopUtil.scala b/core/src/main/scala/org/apache/spark/api/python/PythonHadoopUtil.scala index d2beef2..6f67306 100644 --- a/core/src/main/scala/org/apache/spark/api/python/PythonHadoopUtil.scala +++ b/core/src/main/scala/org/apache/spark/api/python/PythonHadoopUtil.scala @@ -23,8 +23,9 @@ import scala.util.{Failure, Success, Try} import org.apache.hadoop.conf.Configuration import org.apache.hadoop.io._ -import org.apache.spark.{Logging, SparkException} +import org.apache.spark.SparkException import org.apache.spark.broadcast.Broadcast +import org.apache.spark.internal.Logging import org.apache.spark.rdd.RDD import org.apache.spark.util.{SerializableConfiguration, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala b/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala index 8f30677..f423b2e 100644 --- a/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala +++ b/core/src/main/scala/org/apache/spark/api/python/PythonRDD.scala @@ -36,6 +36,7 @@ import org.apache.spark._ import org.apache.spark.api.java.{JavaPairRDD, JavaRDD, JavaSparkContext} import org.apache.spark.broadcast.Broadcast import org.apache.spark.input.PortableDataStream +import org.apache.spark.internal.Logging import org.apache.spark.rdd.RDD import org.apache.spark.util.{SerializableConfiguration, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala b/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala index 433764b..3df87f6 100644 --- a/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala +++ b/core/src/main/scala/org/apache/spark/api/python/PythonWorkerFactory.scala @@ -26,6 +26,7 @@ import scala.collection.mutable import scala.collection.JavaConverters._ import org.apache.spark._ +import org.apache.spark.internal.Logging import org.apache.spark.util.{RedirectThread, Utils} private[spark] class PythonWorkerFactory(pythonExec: String, envVars: Map[String, String]) http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala b/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala index 55db938..1c632eb 100644 --- a/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala +++ b/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala @@ -28,8 +28,9 @@ import scala.util.Try import net.razorvine.pickle.{Pickler, Unpickler} -import org.apache.spark.{Logging, SparkException} +import org.apache.spark.SparkException import org.apache.spark.api.java.JavaRDD +import org.apache.spark.internal.Logging import org.apache.spark.rdd.RDD /** Utilities for serialization / deserialization between Python and Java, using Pickle. */ http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/r/RBackend.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/r/RBackend.scala b/core/src/main/scala/org/apache/spark/api/r/RBackend.scala index 8b3be0d..9d29a84 100644 --- a/core/src/main/scala/org/apache/spark/api/r/RBackend.scala +++ b/core/src/main/scala/org/apache/spark/api/r/RBackend.scala @@ -29,7 +29,8 @@ import io.netty.channel.socket.nio.NioServerSocketChannel import io.netty.handler.codec.LengthFieldBasedFrameDecoder import io.netty.handler.codec.bytes.{ByteArrayDecoder, ByteArrayEncoder} -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf +import org.apache.spark.internal.Logging /** * Netty-based backend server that is used to communicate between R and Java. http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/r/RBackendHandler.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/r/RBackendHandler.scala b/core/src/main/scala/org/apache/spark/api/r/RBackendHandler.scala index 9bddd72..c416e83 100644 --- a/core/src/main/scala/org/apache/spark/api/r/RBackendHandler.scala +++ b/core/src/main/scala/org/apache/spark/api/r/RBackendHandler.scala @@ -25,8 +25,8 @@ import scala.language.existentials import io.netty.channel.{ChannelHandlerContext, SimpleChannelInboundHandler} import io.netty.channel.ChannelHandler.Sharable -import org.apache.spark.Logging import org.apache.spark.api.r.SerDe._ +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/api/r/RRDD.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/api/r/RRDD.scala b/core/src/main/scala/org/apache/spark/api/r/RRDD.scala index 401f362..588a57e 100644 --- a/core/src/main/scala/org/apache/spark/api/r/RRDD.scala +++ b/core/src/main/scala/org/apache/spark/api/r/RRDD.scala @@ -29,6 +29,7 @@ import scala.util.Try import org.apache.spark._ import org.apache.spark.api.java.{JavaPairRDD, JavaRDD, JavaSparkContext} import org.apache.spark.broadcast.Broadcast +import org.apache.spark.internal.Logging import org.apache.spark.rdd.RDD import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/broadcast/Broadcast.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/broadcast/Broadcast.scala b/core/src/main/scala/org/apache/spark/broadcast/Broadcast.scala index 0d68872..24d953e 100644 --- a/core/src/main/scala/org/apache/spark/broadcast/Broadcast.scala +++ b/core/src/main/scala/org/apache/spark/broadcast/Broadcast.scala @@ -21,8 +21,8 @@ import java.io.Serializable import scala.reflect.ClassTag -import org.apache.spark.Logging import org.apache.spark.SparkException +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala b/core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala index be416c4..e88988f 100644 --- a/core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala +++ b/core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala @@ -21,7 +21,8 @@ import java.util.concurrent.atomic.AtomicLong import scala.reflect.ClassTag -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} +import org.apache.spark.internal.Logging private[spark] class BroadcastManager( val isDriver: Boolean, http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala b/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala index 8091aa8..2634d88 100644 --- a/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala +++ b/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala @@ -25,6 +25,7 @@ import scala.reflect.ClassTag import scala.util.Random import org.apache.spark._ +import org.apache.spark.internal.Logging import org.apache.spark.io.CompressionCodec import org.apache.spark.serializer.Serializer import org.apache.spark.storage.{BlockId, BroadcastBlockId, StorageLevel} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/Client.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/Client.scala b/core/src/main/scala/org/apache/spark/deploy/Client.scala index dcef03e..640f25f 100644 --- a/core/src/main/scala/org/apache/spark/deploy/Client.scala +++ b/core/src/main/scala/org/apache/spark/deploy/Client.scala @@ -22,11 +22,12 @@ import scala.concurrent.ExecutionContext import scala.reflect.ClassTag import scala.util.{Failure, Success} -import org.apache.log4j.{Level, Logger} +import org.apache.log4j.Logger -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.DeployMessages._ import org.apache.spark.deploy.master.{DriverState, Master} +import org.apache.spark.internal.Logging import org.apache.spark.rpc.{RpcAddress, RpcEndpointRef, RpcEnv, ThreadSafeRpcEndpoint} import org.apache.spark.util.{SparkExitCode, ThreadUtils, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/ExternalShuffleService.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/ExternalShuffleService.scala b/core/src/main/scala/org/apache/spark/deploy/ExternalShuffleService.scala index c514a1a..adc0de1 100644 --- a/core/src/main/scala/org/apache/spark/deploy/ExternalShuffleService.scala +++ b/core/src/main/scala/org/apache/spark/deploy/ExternalShuffleService.scala @@ -21,7 +21,8 @@ import java.util.concurrent.CountDownLatch import scala.collection.JavaConverters._ -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} +import org.apache.spark.internal.Logging import org.apache.spark.network.TransportContext import org.apache.spark.network.netty.SparkTransportConf import org.apache.spark.network.sasl.SaslServerBootstrap http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala b/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala index 305994a..abb98f9 100644 --- a/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala +++ b/core/src/main/scala/org/apache/spark/deploy/FaultToleranceTest.scala @@ -32,8 +32,9 @@ import scala.sys.process._ import org.json4s._ import org.json4s.jackson.JsonMethods -import org.apache.spark.{Logging, SparkConf, SparkContext} +import org.apache.spark.{SparkConf, SparkContext} import org.apache.spark.deploy.master.RecoveryState +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/LocalSparkCluster.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/LocalSparkCluster.scala b/core/src/main/scala/org/apache/spark/deploy/LocalSparkCluster.scala index 2dfb813..84aa894 100644 --- a/core/src/main/scala/org/apache/spark/deploy/LocalSparkCluster.scala +++ b/core/src/main/scala/org/apache/spark/deploy/LocalSparkCluster.scala @@ -19,9 +19,10 @@ package org.apache.spark.deploy import scala.collection.mutable.ArrayBuffer -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf import org.apache.spark.deploy.master.Master import org.apache.spark.deploy.worker.Worker +import org.apache.spark.internal.Logging import org.apache.spark.rpc.RpcEnv import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/RPackageUtils.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/RPackageUtils.scala b/core/src/main/scala/org/apache/spark/deploy/RPackageUtils.scala index 81718e0..3d2cabc 100644 --- a/core/src/main/scala/org/apache/spark/deploy/RPackageUtils.scala +++ b/core/src/main/scala/org/apache/spark/deploy/RPackageUtils.scala @@ -26,8 +26,8 @@ import scala.collection.JavaConverters._ import com.google.common.io.{ByteStreams, Files} -import org.apache.spark.Logging import org.apache.spark.api.r.RUtils +import org.apache.spark.internal.Logging import org.apache.spark.util.{RedirectThread, Utils} private[deploy] object RPackageUtils extends Logging { http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/SparkCuratorUtil.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/SparkCuratorUtil.scala b/core/src/main/scala/org/apache/spark/deploy/SparkCuratorUtil.scala index 8d5e716..8247110 100644 --- a/core/src/main/scala/org/apache/spark/deploy/SparkCuratorUtil.scala +++ b/core/src/main/scala/org/apache/spark/deploy/SparkCuratorUtil.scala @@ -23,7 +23,8 @@ import org.apache.curator.framework.{CuratorFramework, CuratorFrameworkFactory} import org.apache.curator.retry.ExponentialBackoffRetry import org.apache.zookeeper.KeeperException -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf +import org.apache.spark.internal.Logging private[spark] object SparkCuratorUtil extends Logging { http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala b/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala index 270ca84..06b7b38 100644 --- a/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala +++ b/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala @@ -35,8 +35,9 @@ import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifie import org.apache.hadoop.mapred.JobConf import org.apache.hadoop.security.{Credentials, UserGroupInformation} -import org.apache.spark.{Logging, SparkConf, SparkException} +import org.apache.spark.{SparkConf, SparkException} import org.apache.spark.annotation.DeveloperApi +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/client/AppClient.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/client/AppClient.scala b/core/src/main/scala/org/apache/spark/deploy/client/AppClient.scala index b9dec62..43b17e5 100644 --- a/core/src/main/scala/org/apache/spark/deploy/client/AppClient.scala +++ b/core/src/main/scala/org/apache/spark/deploy/client/AppClient.scala @@ -23,12 +23,13 @@ import java.util.concurrent.atomic.{AtomicBoolean, AtomicReference} import scala.util.control.NonFatal -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf import org.apache.spark.deploy.{ApplicationDescription, ExecutorState} import org.apache.spark.deploy.DeployMessages._ import org.apache.spark.deploy.master.Master +import org.apache.spark.internal.Logging import org.apache.spark.rpc._ -import org.apache.spark.util.{RpcUtils, ThreadUtils, Utils} +import org.apache.spark.util.{RpcUtils, ThreadUtils} /** * Interface allowing applications to speak with a Spark deploy cluster. Takes a master URL, http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala b/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala index 000f7e8..a370526 100644 --- a/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala +++ b/core/src/main/scala/org/apache/spark/deploy/history/ApplicationCache.scala @@ -28,7 +28,7 @@ import com.codahale.metrics.{Counter, MetricRegistry, Timer} import com.google.common.cache.{CacheBuilder, CacheLoader, LoadingCache, RemovalListener, RemovalNotification} import org.eclipse.jetty.servlet.FilterHolder -import org.apache.spark.Logging +import org.apache.spark.internal.Logging import org.apache.spark.metrics.source.Source import org.apache.spark.ui.SparkUI import org.apache.spark.util.Clock http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala b/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala index f885798..d5afb33 100644 --- a/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala +++ b/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala @@ -31,8 +31,9 @@ import org.apache.hadoop.hdfs.DistributedFileSystem import org.apache.hadoop.hdfs.protocol.HdfsConstants import org.apache.hadoop.security.AccessControlException -import org.apache.spark.{Logging, SecurityManager, SparkConf, SparkException} +import org.apache.spark.{SecurityManager, SparkConf, SparkException} import org.apache.spark.deploy.SparkHadoopUtil +import org.apache.spark.internal.Logging import org.apache.spark.scheduler._ import org.apache.spark.ui.SparkUI import org.apache.spark.util.{Clock, SystemClock, ThreadUtils, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala b/core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala index 076bdc5..d821474 100644 --- a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala +++ b/core/src/main/scala/org/apache/spark/deploy/history/HistoryServer.scala @@ -25,8 +25,9 @@ import scala.util.control.NonFatal import org.eclipse.jetty.servlet.{ServletContextHandler, ServletHolder} -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.SparkHadoopUtil +import org.apache.spark.internal.Logging import org.apache.spark.status.api.v1.{ApiRootResource, ApplicationInfo, ApplicationsListResource, UIRoot} import org.apache.spark.ui.{SparkUI, UIUtils, WebUI} import org.apache.spark.ui.JettyUtils._ http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala b/core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala index fc3790f..2eddb5f 100644 --- a/core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala +++ b/core/src/main/scala/org/apache/spark/deploy/history/HistoryServerArguments.scala @@ -19,7 +19,8 @@ package org.apache.spark.deploy.history import scala.annotation.tailrec -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala b/core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala index 1aa8cd5..f2b5ea7 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/FileSystemPersistenceEngine.scala @@ -21,7 +21,7 @@ import java.io._ import scala.reflect.ClassTag -import org.apache.spark.Logging +import org.apache.spark.internal.Logging import org.apache.spark.serializer.{DeserializationStream, SerializationStream, Serializer} import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala index c97ad4d..01901bb 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala @@ -31,7 +31,7 @@ import scala.util.Random import org.apache.hadoop.fs.Path -import org.apache.spark.{Logging, SecurityManager, SparkConf, SparkException} +import org.apache.spark.{SecurityManager, SparkConf, SparkException} import org.apache.spark.deploy.{ApplicationDescription, DriverDescription, ExecutorState, SparkHadoopUtil} import org.apache.spark.deploy.DeployMessages._ @@ -40,6 +40,7 @@ import org.apache.spark.deploy.master.DriverState.DriverState import org.apache.spark.deploy.master.MasterMessages._ import org.apache.spark.deploy.master.ui.MasterWebUI import org.apache.spark.deploy.rest.StandaloneRestServer +import org.apache.spark.internal.Logging import org.apache.spark.metrics.MetricsSystem import org.apache.spark.rpc._ import org.apache.spark.scheduler.{EventLoggingListener, ReplayListenerBus} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala b/core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala index c4c3283..ffdd635 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/RecoveryModeFactory.scala @@ -17,8 +17,9 @@ package org.apache.spark.deploy.master -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf import org.apache.spark.annotation.DeveloperApi +import org.apache.spark.internal.Logging import org.apache.spark.serializer.Serializer /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala b/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala index 336cb24..1e8dabf 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperLeaderElectionAgent.scala @@ -20,8 +20,9 @@ package org.apache.spark.deploy.master import org.apache.curator.framework.CuratorFramework import org.apache.curator.framework.recipes.leader.{LeaderLatch, LeaderLatchListener} -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf import org.apache.spark.deploy.SparkCuratorUtil +import org.apache.spark.internal.Logging private[master] class ZooKeeperLeaderElectionAgent(val masterInstance: LeaderElectable, conf: SparkConf) extends LeaderLatchListener with LeaderElectionAgent with Logging { http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala b/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala index b0cedef..79f7721 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/ZooKeeperPersistenceEngine.scala @@ -25,8 +25,9 @@ import scala.reflect.ClassTag import org.apache.curator.framework.CuratorFramework import org.apache.zookeeper.CreateMode -import org.apache.spark.{Logging, SparkConf} +import org.apache.spark.SparkConf import org.apache.spark.deploy.SparkCuratorUtil +import org.apache.spark.internal.Logging import org.apache.spark.serializer.Serializer http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/master/ui/MasterWebUI.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/master/ui/MasterWebUI.scala b/core/src/main/scala/org/apache/spark/deploy/master/ui/MasterWebUI.scala index d754392..ae16ce9 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/ui/MasterWebUI.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/ui/MasterWebUI.scala @@ -17,8 +17,8 @@ package org.apache.spark.deploy.master.ui -import org.apache.spark.Logging import org.apache.spark.deploy.master.Master +import org.apache.spark.internal.Logging import org.apache.spark.status.api.v1.{ApiRootResource, ApplicationInfo, ApplicationsListResource, UIRoot} import org.apache.spark.ui.{SparkUI, WebUI} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosClusterDispatcher.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosClusterDispatcher.scala b/core/src/main/scala/org/apache/spark/deploy/mesos/MesosClusterDispatcher.scala index 7091513..a057977 100644 --- a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosClusterDispatcher.scala +++ b/core/src/main/scala/org/apache/spark/deploy/mesos/MesosClusterDispatcher.scala @@ -19,9 +19,10 @@ package org.apache.spark.deploy.mesos import java.util.concurrent.CountDownLatch -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.mesos.ui.MesosClusterUI import org.apache.spark.deploy.rest.mesos.MesosRestServer +import org.apache.spark.internal.Logging import org.apache.spark.scheduler.cluster.mesos._ import org.apache.spark.util.{ShutdownHookManager, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosExternalShuffleService.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosExternalShuffleService.scala b/core/src/main/scala/org/apache/spark/deploy/mesos/MesosExternalShuffleService.scala index c0f9129..6b297c4 100644 --- a/core/src/main/scala/org/apache/spark/deploy/mesos/MesosExternalShuffleService.scala +++ b/core/src/main/scala/org/apache/spark/deploy/mesos/MesosExternalShuffleService.scala @@ -22,8 +22,9 @@ import java.util.concurrent.{ConcurrentHashMap, TimeUnit} import scala.collection.JavaConverters._ -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.ExternalShuffleService +import org.apache.spark.internal.Logging import org.apache.spark.network.client.{RpcResponseCallback, TransportClient} import org.apache.spark.network.shuffle.ExternalShuffleBlockHandler import org.apache.spark.network.shuffle.protocol.BlockTransferMessage http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala b/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala index d3e092a..c5a5876 100644 --- a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala +++ b/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionClient.scala @@ -30,7 +30,8 @@ import scala.io.Source import com.fasterxml.jackson.core.JsonProcessingException -import org.apache.spark.{Logging, SPARK_VERSION => sparkVersion, SparkConf} +import org.apache.spark.{SPARK_VERSION => sparkVersion, SparkConf} +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionServer.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionServer.scala b/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionServer.scala index 8e0862d..14244ea 100644 --- a/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionServer.scala +++ b/core/src/main/scala/org/apache/spark/deploy/rest/RestSubmissionServer.scala @@ -29,7 +29,8 @@ import org.eclipse.jetty.util.thread.QueuedThreadPool import org.json4s._ import org.json4s.jackson.JsonMethods._ -import org.apache.spark.{Logging, SPARK_VERSION => sparkVersion, SparkConf} +import org.apache.spark.{SPARK_VERSION => sparkVersion, SparkConf} +import org.apache.spark.internal.Logging import org.apache.spark.util.Utils /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala b/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala index ce02ee2..a4efafc 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/CommandUtils.scala @@ -22,9 +22,9 @@ import java.io.{File, FileOutputStream, InputStream, IOException} import scala.collection.JavaConverters._ import scala.collection.Map -import org.apache.spark.Logging import org.apache.spark.SecurityManager import org.apache.spark.deploy.Command +import org.apache.spark.internal.Logging import org.apache.spark.launcher.WorkerCommandBuilder import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/DriverRunner.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/DriverRunner.scala b/core/src/main/scala/org/apache/spark/deploy/worker/DriverRunner.scala index 7f4fe26..9c6bc5c 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/DriverRunner.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/DriverRunner.scala @@ -25,11 +25,12 @@ import scala.collection.JavaConverters._ import com.google.common.io.Files import org.apache.hadoop.fs.Path -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.{DriverDescription, SparkHadoopUtil} import org.apache.spark.deploy.DeployMessages.DriverStateChanged import org.apache.spark.deploy.master.DriverState import org.apache.spark.deploy.master.DriverState.DriverState +import org.apache.spark.internal.Logging import org.apache.spark.rpc.RpcEndpointRef import org.apache.spark.util.{Clock, SystemClock, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala index 208a1bb..f9c92c3 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala @@ -24,9 +24,10 @@ import scala.collection.JavaConverters._ import com.google.common.io.Files -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.{ApplicationDescription, ExecutorState} import org.apache.spark.deploy.DeployMessages.ExecutorStateChanged +import org.apache.spark.internal.Logging import org.apache.spark.rpc.RpcEndpointRef import org.apache.spark.util.{ShutdownHookManager, Utils} import org.apache.spark.util.logging.FileAppender http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala index c18c8c7..1b7637a 100755 --- a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala @@ -29,12 +29,13 @@ import scala.concurrent.ExecutionContext import scala.util.{Failure, Random, Success} import scala.util.control.NonFatal -import org.apache.spark.{Logging, SecurityManager, SparkConf} +import org.apache.spark.{SecurityManager, SparkConf} import org.apache.spark.deploy.{Command, ExecutorDescription, ExecutorState} import org.apache.spark.deploy.DeployMessages._ import org.apache.spark.deploy.ExternalShuffleService import org.apache.spark.deploy.master.{DriverState, Master} import org.apache.spark.deploy.worker.ui.WorkerWebUI +import org.apache.spark.internal.Logging import org.apache.spark.metrics.MetricsSystem import org.apache.spark.rpc._ import org.apache.spark.util.{ThreadUtils, Utils} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/WorkerWatcher.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/WorkerWatcher.scala b/core/src/main/scala/org/apache/spark/deploy/worker/WorkerWatcher.scala index ab56fde..af29de3 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/WorkerWatcher.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/WorkerWatcher.scala @@ -17,7 +17,7 @@ package org.apache.spark.deploy.worker -import org.apache.spark.Logging +import org.apache.spark.internal.Logging import org.apache.spark.rpc._ /** http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/ui/LogPage.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/ui/LogPage.scala b/core/src/main/scala/org/apache/spark/deploy/worker/ui/LogPage.scala index 09ae64a..6500cab 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/ui/LogPage.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/ui/LogPage.scala @@ -22,7 +22,7 @@ import javax.servlet.http.HttpServletRequest import scala.xml.Node -import org.apache.spark.Logging +import org.apache.spark.internal.Logging import org.apache.spark.ui.{UIUtils, WebUIPage} import org.apache.spark.util.Utils import org.apache.spark.util.logging.RollingFileAppender http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/deploy/worker/ui/WorkerWebUI.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/ui/WorkerWebUI.scala b/core/src/main/scala/org/apache/spark/deploy/worker/ui/WorkerWebUI.scala index b45b682..db696b0 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/ui/WorkerWebUI.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/ui/WorkerWebUI.scala @@ -20,8 +20,8 @@ package org.apache.spark.deploy.worker.ui import java.io.File import javax.servlet.http.HttpServletRequest -import org.apache.spark.Logging import org.apache.spark.deploy.worker.Worker +import org.apache.spark.internal.Logging import org.apache.spark.ui.{SparkUI, WebUI} import org.apache.spark.ui.JettyUtils._ import org.apache.spark.util.RpcUtils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala b/core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala index 3b5cb18..320a200 100644 --- a/core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala +++ b/core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala @@ -28,6 +28,7 @@ import org.apache.spark._ import org.apache.spark.TaskState.TaskState import org.apache.spark.deploy.SparkHadoopUtil import org.apache.spark.deploy.worker.WorkerWatcher +import org.apache.spark.internal.Logging import org.apache.spark.rpc._ import org.apache.spark.scheduler.TaskDescription import org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages._ http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/executor/Executor.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/executor/Executor.scala b/core/src/main/scala/org/apache/spark/executor/Executor.scala index 48372d7..6327d55 100644 --- a/core/src/main/scala/org/apache/spark/executor/Executor.scala +++ b/core/src/main/scala/org/apache/spark/executor/Executor.scala @@ -29,6 +29,7 @@ import scala.util.control.NonFatal import org.apache.spark._ import org.apache.spark.deploy.SparkHadoopUtil +import org.apache.spark.internal.Logging import org.apache.spark.memory.TaskMemoryManager import org.apache.spark.rpc.RpcTimeout import org.apache.spark.scheduler.{AccumulableInfo, DirectTaskResult, IndirectTaskResult, Task} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala b/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala index cfd9bcd..680cfb7 100644 --- a/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala +++ b/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala @@ -25,9 +25,10 @@ import org.apache.mesos.{Executor => MesosExecutor, ExecutorDriver, MesosExecuto import org.apache.mesos.Protos.{TaskStatus => MesosTaskStatus, _} import org.apache.mesos.protobuf.ByteString -import org.apache.spark.{Logging, SparkConf, SparkEnv, TaskState} +import org.apache.spark.{SparkConf, SparkEnv, TaskState} import org.apache.spark.TaskState.TaskState import org.apache.spark.deploy.SparkHadoopUtil +import org.apache.spark.internal.Logging import org.apache.spark.scheduler.cluster.mesos.MesosTaskLaunchData import org.apache.spark.util.Utils http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala b/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala index 9da9cb5..02219a8 100644 --- a/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala +++ b/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala @@ -22,6 +22,7 @@ import scala.collection.mutable.ArrayBuffer import org.apache.spark._ import org.apache.spark.annotation.DeveloperApi +import org.apache.spark.internal.Logging import org.apache.spark.scheduler.AccumulableInfo import org.apache.spark.storage.{BlockId, BlockStatus} http://git-wip-us.apache.org/repos/asf/spark/blob/8ef3399a/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala b/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala index bc98273..978afaf 100644 --- a/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala +++ b/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala @@ -22,7 +22,7 @@ import org.apache.hadoop.io.{BytesWritable, LongWritable} import org.apache.hadoop.mapreduce.{InputSplit, JobContext, RecordReader, TaskAttemptContext} import org.apache.hadoop.mapreduce.lib.input.FileInputFormat -import org.apache.spark.Logging +import org.apache.spark.internal.Logging /** * Custom Input Format for reading and splitting flat binary files that contain records, --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org