http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-cblock/tools/src/test/org/apache/hadoop/cblock/TestCBlockCLI.java ---------------------------------------------------------------------- diff --git a/hadoop-cblock/tools/src/test/org/apache/hadoop/cblock/TestCBlockCLI.java b/hadoop-cblock/tools/src/test/org/apache/hadoop/cblock/TestCBlockCLI.java deleted file mode 100644 index a3f53aa..0000000 --- a/hadoop-cblock/tools/src/test/org/apache/hadoop/cblock/TestCBlockCLI.java +++ /dev/null @@ -1,242 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.hadoop.cblock; - -import org.apache.hadoop.cblock.cli.CBlockCli; -import org.apache.hadoop.cblock.meta.VolumeDescriptor; -import org.apache.hadoop.cblock.util.MockStorageClient; -import org.apache.hadoop.conf.OzoneConfiguration; -import org.apache.hadoop.hdds.scm.client.ScmClient; -import org.apache.hadoop.test.GenericTestUtils; -import org.junit.After; -import org.junit.AfterClass; -import org.junit.BeforeClass; -import org.junit.Test; - -import java.io.ByteArrayOutputStream; -import java.io.File; -import java.io.IOException; -import java.io.PrintStream; -import java.util.List; - -import static org.apache.hadoop.cblock.CBlockConfigKeys - .DFS_CBLOCK_JSCSIRPC_ADDRESS_KEY; -import static org.apache.hadoop.cblock.CBlockConfigKeys - .DFS_CBLOCK_SERVICERPC_ADDRESS_KEY; -import static org.apache.hadoop.cblock.CBlockConfigKeys - .DFS_CBLOCK_SERVICE_LEVELDB_PATH_KEY; -import static org.junit.Assert.assertEquals; -import static org.junit.Assert.assertTrue; - -/** - * A testing class for cblock command line tool. - */ -public class TestCBlockCLI { - private static final long GB = 1 * 1024 * 1024 * 1024L; - private static final int KB = 1024; - private static CBlockCli cmd; - private static OzoneConfiguration conf; - private static CBlockManager cBlockManager; - private static ByteArrayOutputStream outContent; - private static PrintStream testPrintOut; - - @BeforeClass - public static void setup() throws IOException { - outContent = new ByteArrayOutputStream(); - ScmClient storageClient = new MockStorageClient(); - conf = new OzoneConfiguration(); - String path = GenericTestUtils - .getTempPath(TestCBlockCLI.class.getSimpleName()); - File filePath = new File(path); - if (!filePath.exists() && !filePath.mkdirs()) { - throw new IOException("Unable to create test DB dir"); - } - conf.set(DFS_CBLOCK_SERVICERPC_ADDRESS_KEY, "127.0.0.1:0"); - conf.set(DFS_CBLOCK_JSCSIRPC_ADDRESS_KEY, "127.0.0.1:0"); - conf.set(DFS_CBLOCK_SERVICE_LEVELDB_PATH_KEY, path.concat( - "/testCblockCli.dat")); - cBlockManager = new CBlockManager(conf, storageClient); - cBlockManager.start(); - testPrintOut = new PrintStream(outContent); - cmd = new CBlockCli(conf, testPrintOut); - } - - @AfterClass - public static void clean() { - if (cBlockManager != null) { - cBlockManager.stop(); - cBlockManager.join(); - cBlockManager.clean(); - } - } - - @After - public void reset() { - outContent.reset(); - } - - /** - * Test the help command. - * @throws Exception - */ - @Test - public void testCliHelp() throws Exception { - PrintStream initialStdOut = System.out; - System.setOut(testPrintOut); - String[] args = {"-h"}; - cmd.run(args); - String helpPrints = - "usage: cblock\n" + - " -c,--createVolume <user> <volume> <volumeSize in [GB/TB]> " + - "<blockSize> create a fresh new volume\n" + - " -d,--deleteVolume <user> <volume> " + - " delete a volume\n" + - " -h,--help " + - " help\n" + - " -i,--infoVolume <user> <volume> " + - " info a volume\n" + - " -l,--listVolume <user> " + - " list all volumes\n" + - " -s,--serverAddr <serverAddress>:<serverPort> " + - " specify server address:port\n"; - assertEquals(helpPrints, outContent.toString()); - outContent.reset(); - System.setOut(initialStdOut); - } - - /** - * Test volume listing command. - * @throws Exception - */ - @Test - public void testCliList() throws Exception { - String userName0 = "userTestCliList0"; - String userName1 = "userTestCliList1"; - String userTestNotExist = "userTestNotExist"; - String volumeName0 = "volumeTest0"; - String volumeName1 = "volumeTest1"; - String volumeSize0 = "30GB"; - String volumeSize1 = "40GB"; - String blockSize = Integer.toString(4); - String[] argsCreate0 = - {"-c", userName0, volumeName0, volumeSize0, blockSize}; - cmd.run(argsCreate0); - String[] argsCreate1 = - {"-c", userName0, volumeName1, volumeSize1, blockSize}; - cmd.run(argsCreate1); - String[] argsCreate2 = - {"-c", userName1, volumeName0, volumeSize0, blockSize}; - cmd.run(argsCreate2); - String[] argsList0 = {"-l"}; - cmd.run(argsList0); - String[] outExpected1 = { - "userTestCliList1:volumeTest0\t32212254720\t4096\n", - "userTestCliList0:volumeTest0\t32212254720\t4096\n", - "userTestCliList0:volumeTest1\t42949672960\t4096\n"}; - int length = 0; - for (String str : outExpected1) { - assertTrue(outContent.toString().contains(str)); - length += str.length(); - } - assertEquals(length, outContent.toString().length()); - outContent.reset(); - - String[] argsList1 = {"-l", userName1}; - cmd.run(argsList1); - String outExpected2 = "userTestCliList1:volumeTest0\t32212254720\t4096\n"; - assertEquals(outExpected2, outContent.toString()); - outContent.reset(); - - String[] argsList2 = {"-l", userTestNotExist}; - cmd.run(argsList2); - String outExpected3 = "\n"; - assertEquals(outExpected3, outContent.toString()); - } - - /** - * Test create volume command. - * @throws Exception - */ - @Test - public void testCliCreate() throws Exception { - String userName = "userTestCliCreate"; - String volumeName = "volumeTest"; - String volumeSize = "30GB"; - String blockSize = "4"; - String[] argsCreate = {"-c", userName, volumeName, volumeSize, blockSize}; - cmd.run(argsCreate); - List<VolumeDescriptor> allVolumes = cBlockManager.getAllVolumes(userName); - assertEquals(1, allVolumes.size()); - VolumeDescriptor volume = allVolumes.get(0); - assertEquals(userName, volume.getUserName()); - assertEquals(volumeName, volume.getVolumeName()); - long volumeSizeB = volume.getVolumeSize(); - assertEquals(30, (int)(volumeSizeB/ GB)); - assertEquals(4, volume.getBlockSize()/ KB); - } - - /** - * Test delete volume command. - * @throws Exception - */ - @Test - public void testCliDelete() throws Exception { - String userName = "userTestCliDelete"; - String volumeName = "volumeTest"; - String volumeSize = "30GB"; - String blockSize = "4"; - String[] argsCreate = {"-c", userName, volumeName, volumeSize, blockSize}; - cmd.run(argsCreate); - List<VolumeDescriptor> allVolumes = cBlockManager.getAllVolumes(userName); - assertEquals(1, allVolumes.size()); - VolumeDescriptor volume = allVolumes.get(0); - assertEquals(userName, volume.getUserName()); - assertEquals(volumeName, volume.getVolumeName()); - long volumeSizeB = volume.getVolumeSize(); - assertEquals(30, (int)(volumeSizeB/ GB)); - assertEquals(4, volume.getBlockSize()/ KB); - - String[] argsDelete = {"-d", userName, volumeName}; - cmd.run(argsDelete); - allVolumes = cBlockManager.getAllVolumes(userName); - assertEquals(0, allVolumes.size()); - } - - /** - * Test info volume command. - * @throws Exception - */ - @Test - public void testCliInfoVolume() throws Exception { - String userName0 = "userTestCliInfo"; - String volumeName0 = "volumeTest0"; - String volumeSize = "8000GB"; - String blockSize = "4"; - String[] argsCreate0 = { - "-c", userName0, volumeName0, volumeSize, blockSize}; - cmd.run(argsCreate0); - String[] argsInfo = {"-i", userName0, volumeName0}; - cmd.run(argsInfo); - // TODO : the usage field is not implemented yet, always 0 now. - String outExpected = " userName:userTestCliInfo " + - "volumeName:volumeTest0 " + - "volumeSize:8589934592000 " + - "blockSize:4096 (sizeInBlocks:2097152000) usageInBlocks:0\n"; - assertEquals(outExpected, outContent.toString()); - } -}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh b/hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh index 960bc63..bee1430 100755 --- a/hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh +++ b/hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh @@ -600,8 +600,6 @@ function hadoop_bootstrap HDDS_LIB_JARS_DIR=${HDDS_LIB_JARS_DIR:-"share/hadoop/hdds/lib"} OZONE_DIR=${OZONE_DIR:-"share/hadoop/ozone"} OZONE_LIB_JARS_DIR=${OZONE_LIB_JARS_DIR:-"share/hadoop/ozone/lib"} - CBLOCK_DIR=${CBLOCK_DIR:-"share/hadoop/cblock"} - CBLOCK_LIB_JARS_DIR=${CBLOCK_LIB_JARS_DIR:-"share/hadoop/cblock/lib"} HADOOP_TOOLS_HOME=${HADOOP_TOOLS_HOME:-${HADOOP_HOME}} HADOOP_TOOLS_DIR=${HADOOP_TOOLS_DIR:-"share/hadoop/tools"} http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-common-project/hadoop-common/src/main/conf/hadoop-env.sh ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/conf/hadoop-env.sh b/hadoop-common-project/hadoop-common/src/main/conf/hadoop-env.sh index 257b2fd..6573a81 100644 --- a/hadoop-common-project/hadoop-common/src/main/conf/hadoop-env.sh +++ b/hadoop-common-project/hadoop-common/src/main/conf/hadoop-env.sh @@ -403,25 +403,6 @@ esac # # export HDFS_DFSROUTER_OPTS="" - -### -# HDFS CBlock Server specific parameters -### -# Specify the JVM options to be used when starting the HDFS CBlock Server. -# These options will be appended to the options specified as HADOOP_OPTS -# and therefore may override any similar flags set in HADOOP_OPTS -# -# export HDFS_CBLOCKSERVER_OPTS="" - -### -# HDFS JSCSI specific parameters -### -# Specify the JVM options to be used when starting the HDFS JSCSI. -# These options will be appended to the options specified as HADOOP_OPTS -# and therefore may override any similar flags set in HADOOP_OPTS -# -# export HDFS_JSCSI_OPTS="" - ### # HDFS Key Space Manager specific parameters ### http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-dist/pom.xml ---------------------------------------------------------------------- diff --git a/hadoop-dist/pom.xml b/hadoop-dist/pom.xml index 3003550..48e6673 100644 --- a/hadoop-dist/pom.xml +++ b/hadoop-dist/pom.xml @@ -239,10 +239,6 @@ </dependency> <dependency> <groupId>org.apache.hadoop</groupId> - <artifactId>hadoop-cblock-server</artifactId> - </dependency> - <dependency> - <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-hdds-container-service</artifactId> </dependency> <dependency> @@ -257,10 +253,6 @@ <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-ozone-tools</artifactId> </dependency> - <dependency> - <groupId>org.apache.hadoop</groupId> - <artifactId>hadoop-cblock-tools</artifactId> - </dependency> </dependencies> <build> <plugins> http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-dist/src/main/compose/cblock/.env ---------------------------------------------------------------------- diff --git a/hadoop-dist/src/main/compose/cblock/.env b/hadoop-dist/src/main/compose/cblock/.env deleted file mode 100644 index af20d3e..0000000 --- a/hadoop-dist/src/main/compose/cblock/.env +++ /dev/null @@ -1,17 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -VERSION=${project.version} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-dist/src/main/compose/cblock/README.md ---------------------------------------------------------------------- diff --git a/hadoop-dist/src/main/compose/cblock/README.md b/hadoop-dist/src/main/compose/cblock/README.md deleted file mode 100644 index f951c87..0000000 --- a/hadoop-dist/src/main/compose/cblock/README.md +++ /dev/null @@ -1,42 +0,0 @@ -<!-- - Licensed to the Apache Software Foundation (ASF) under one or more - contributor license agreements. See the NOTICE file distributed with - this work for additional information regarding copyright ownership. - The ASF licenses this file to You under the Apache License, Version 2.0 - (the "License"); you may not use this file except in compliance with - the License. You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. ---> - - -# CBlock dozone configuration - -This directory contains example cluster definition for CBlock/jscsi servers. - -## How to use - -1. First of all Start the servers with `docker-compose up -d` - -2. Wait until the servers are up and running (check http://localhost:9876 and wait until you have a healthy node) - -3. Create a volume: `docker-compose exec cblock hdfs cblock -c bilbo volume2 1GB 4` - -4. Mount the iscsi volume (from host machine): - -``` -sudo iscsiadm -m node -o new -T bilbo:volume2 -p 127.0.0.1 -sudo iscsiadm -m node -T bilbo:volume2 --login -``` - -5. Check the device name from `dmesg` or `lsblk` (eg /dev/sdd). Errors in dmesg could be ignored: jscsi doesn't implement all the jscsi commands. - -6. Format the device (`mkfs.ext4 /dev/sdd`). (Yes, format the while device, not just a partition). - -7. Mount it (`mount /dev/sdd /mnt/target`). http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-dist/src/main/compose/cblock/docker-compose.yaml ---------------------------------------------------------------------- diff --git a/hadoop-dist/src/main/compose/cblock/docker-compose.yaml b/hadoop-dist/src/main/compose/cblock/docker-compose.yaml deleted file mode 100644 index fa4d267..0000000 --- a/hadoop-dist/src/main/compose/cblock/docker-compose.yaml +++ /dev/null @@ -1,66 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -version: "3" -services: - namenode: - image: apache/hadoop-runner - hostname: namenode - volumes: - - ../..//hadoop-${VERSION}:/opt/hadoop - ports: - - 9870:9870 - environment: - ENSURE_NAMENODE_DIR: /data/namenode - env_file: - - ./docker-config - command: ["/opt/hadoop/bin/hdfs","namenode"] - datanode: - image: apache/hadoop-runner - volumes: - - ../..//hadoop-${VERSION}:/opt/hadoop - ports: - - 9864 - command: ["/opt/hadoop/bin/ozone","datanode"] - env_file: - - ./docker-config - jscsi: - image: apache/hadoop-runner - ports: - - 3260:3260 - volumes: - - ../..//hadoop-${VERSION}:/opt/hadoop - env_file: - - ./docker-config - command: ["/opt/hadoop/bin/ozone","jscsi"] - cblock: - image: apache/hadoop-runner - volumes: - - ../..//hadoop-${VERSION}:/opt/hadoop - env_file: - - ./docker-config - command: ["/opt/hadoop/bin/ozone","cblockserver"] - scm: - image: apache/hadoop-runner - volumes: - - ../..//hadoop-${VERSION}:/opt/hadoop - ports: - - 9876:9876 - env_file: - - ./docker-config - command: ["/opt/hadoop/bin/ozone","scm"] - environment: - ENSURE_SCM_INITIALIZED: /data/metadata/scm/current/VERSION http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-dist/src/main/compose/cblock/docker-config ---------------------------------------------------------------------- diff --git a/hadoop-dist/src/main/compose/cblock/docker-config b/hadoop-dist/src/main/compose/cblock/docker-config deleted file mode 100644 index f69bef0..0000000 --- a/hadoop-dist/src/main/compose/cblock/docker-config +++ /dev/null @@ -1,39 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -CORE-SITE.XML_fs.defaultFS=hdfs://namenode:9000 -OZONE-SITE.XML_ozone.ksm.address=ksm -OZONE-SITE.XML_ozone.scm.names=scm -OZONE-SITE.XML_ozone.enabled=True -OZONE-SITE.XML_ozone.scm.datanode.id=/data/datanode.id -OZONE-SITE.XML_ozone.scm.block.client.address=scm -OZONE-SITE.XML_ozone.metadata.dirs=/data/metadata -OZONE-SITE.XML_ozone.handler.type=distributed -OZONE-SITE.XML_ozone.scm.client.address=scm - -OZONE-SITE.XML_dfs.cblock.jscsi.cblock.server.address=cblock -OZONE-SITE.XML_dfs.cblock.scm.ipaddress=scm -OZONE-SITE.XML_dfs.cblock.service.leveldb.path=/tmp -OZONE-SITE.XML_hdds.datanode.plugins=org.apache.hadoop.ozone.web.OzoneHddsDatanodeService -HDFS-SITE.XML_dfs.datanode.plugins=org.apache.hadoop.ozone.HddsDatanodeService -HDFS-SITE.XML_dfs.namenode.rpc-address=namenode:9000 -HDFS-SITE.XML_dfs.namenode.name.dir=/data/namenode -HDFS-SITE.XML_rpc.metrics.quantile.enable=true -HDFS-SITE.XML_rpc.metrics.percentiles.intervals=60,300 -LOG4J.PROPERTIES_log4j.rootLogger=info,stdout -LOG4J.PROPERTIES_log4j.appender.stdout=org.apache.log4j.ConsoleAppender -LOG4J.PROPERTIES_log4j.appender.stdout.layout=org.apache.log4j.PatternLayout -LOG4J.PROPERTIES_log4j.appender.stdout.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss} %-5p %c{1}:%L - %m%n http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-ozone/common/src/main/bin/ozone ---------------------------------------------------------------------- diff --git a/hadoop-ozone/common/src/main/bin/ozone b/hadoop-ozone/common/src/main/bin/ozone index 2f2c98c..7419743 100755 --- a/hadoop-ozone/common/src/main/bin/ozone +++ b/hadoop-ozone/common/src/main/bin/ozone @@ -33,8 +33,6 @@ function hadoop_usage hadoop_add_option "--workers" "turn on worker mode" - hadoop_add_subcommand "cblock" admin "cblock CLI" - hadoop_add_subcommand "cblockserver" daemon "run cblock server" hadoop_add_subcommand "classpath" client "prints the class path needed to get the hadoop jar and the required libraries" hadoop_add_subcommand "datanode" daemon "run a DFS datanode" hadoop_add_subcommand "envvars" client "display computed Hadoop environment variables" @@ -43,7 +41,6 @@ function hadoop_usage hadoop_add_subcommand "getozoneconf" client "get ozone config values from configuration" hadoop_add_subcommand "jmxget" admin "get JMX exported values from NameNode or DataNode." - hadoop_add_subcommand "jscsi" daemon "run cblock jscsi server" hadoop_add_subcommand "ksm" daemon "Ozone keyspace manager" hadoop_add_subcommand "o3" client "command line interface for ozone" hadoop_add_subcommand "noz" client "ozone debug tool, convert ozone metadata into relational data" @@ -65,13 +62,6 @@ function ozonecmd_case shift case ${subcmd} in - cblock) - HADOOP_CLASSNAME=org.apache.hadoop.cblock.cli.CBlockCli - ;; - cblockserver) - HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true" - HADOOP_CLASSNAME=org.apache.hadoop.cblock.CBlockManager - ;; classpath) hadoop_do_classpath_subcommand HADOOP_CLASSNAME "$@" ;; @@ -106,10 +96,6 @@ function ozonecmd_case getozoneconf) HADOOP_CLASSNAME=org.apache.hadoop.ozone.freon.OzoneGetConf; ;; - jscsi) - HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true" - HADOOP_CLASSNAME=org.apache.hadoop.cblock.jscsiHelper.SCSITargetDaemon - ;; ksm) HADOOP_SUBCMD_SUPPORTDAEMONIZATION="true" HADOOP_CLASSNAME=org.apache.hadoop.ozone.ksm.KeySpaceManager http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-ozone/common/src/main/shellprofile.d/hadoop-ozone.sh ---------------------------------------------------------------------- diff --git a/hadoop-ozone/common/src/main/shellprofile.d/hadoop-ozone.sh b/hadoop-ozone/common/src/main/shellprofile.d/hadoop-ozone.sh index c016165..2cd2bb3 100644 --- a/hadoop-ozone/common/src/main/shellprofile.d/hadoop-ozone.sh +++ b/hadoop-ozone/common/src/main/shellprofile.d/hadoop-ozone.sh @@ -19,7 +19,7 @@ if [[ "${HADOOP_SHELL_EXECNAME}" = ozone ]]; then fi -## @description Profile for hdds/cblock/ozone components. +## @description Profile for hdds/ozone components. ## @audience private ## @stability evolving function _ozone_hadoop_classpath @@ -40,7 +40,5 @@ function _ozone_hadoop_classpath hadoop_add_classpath "${HADOOP_HDFS_HOME}/${HDDS_DIR}"'/*' hadoop_add_classpath "${HADOOP_HDFS_HOME}/${OZONE_LIB_JARS_DIR}"'/*' hadoop_add_classpath "${HADOOP_HDFS_HOME}/${OZONE_DIR}"'/*' - hadoop_add_classpath "${HADOOP_HDFS_HOME}/${CBLOCK_LIB_JARS_DIR}"'/*' - hadoop_add_classpath "${HADOOP_HDFS_HOME}/${CBLOCK_DIR}"'/*' } http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/hadoop-project/pom.xml ---------------------------------------------------------------------- diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml index d762c44..4595c28 100644 --- a/hadoop-project/pom.xml +++ b/hadoop-project/pom.xml @@ -587,11 +587,6 @@ </dependency> <dependency> <groupId>org.apache.hadoop</groupId> - <artifactId>hadoop-cblock-server</artifactId> - <version>${project.version}</version> - </dependency> - <dependency> - <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-ozone-tools</artifactId> <version>${project.version}</version> </dependency> @@ -604,14 +599,6 @@ <dependency> <groupId>org.apache.hadoop</groupId> - <artifactId>hadoop-cblock-tools</artifactId> - <version>${project.version}</version> - </dependency> - - - - <dependency> - <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-hdds-server-framework</artifactId> <version>${project.version}</version> </dependency> http://git-wip-us.apache.org/repos/asf/hadoop/blob/ea85801c/pom.xml ---------------------------------------------------------------------- diff --git a/pom.xml b/pom.xml index 7d244fd..fc0a5d4 100644 --- a/pom.xml +++ b/pom.xml @@ -720,7 +720,6 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs </activation> <modules> <module>hadoop-ozone</module> - <module>hadoop-cblock</module> <module>hadoop-hdds</module> <module>hadoop-ozone/acceptance-test</module> </modules> --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
