AMBARI-7102. Slider View: Drop yarn-site and global config sections from Configs tab of app. (akovalenko)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/a41ef2a1 Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/a41ef2a1 Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/a41ef2a1 Branch: refs/heads/branch-alerts-dev Commit: a41ef2a1c01b978751cca6734a0fd6f4649e092e Parents: ad1d6a3 Author: Aleksandr Kovalenko <akovale...@hortonworks.com> Authored: Mon Sep 1 16:27:20 2014 +0300 Committer: Aleksandr Kovalenko <akovale...@hortonworks.com> Committed: Mon Sep 1 16:27:36 2014 +0300 ---------------------------------------------------------------------- .../resources/ui/app/assets/data/apps/apps.json | 675 +++++++++++++++---- .../main/resources/ui/app/models/slider_app.js | 51 +- .../ui/app/templates/slider_app/configs.hbs | 84 +-- 3 files changed, 600 insertions(+), 210 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/a41ef2a1/contrib/views/slider/src/main/resources/ui/app/assets/data/apps/apps.json ---------------------------------------------------------------------- diff --git a/contrib/views/slider/src/main/resources/ui/app/assets/data/apps/apps.json b/contrib/views/slider/src/main/resources/ui/app/assets/data/apps/apps.json index 46c5f45..60cc831 100644 --- a/contrib/views/slider/src/main/resources/ui/app/assets/data/apps/apps.json +++ b/contrib/views/slider/src/main/resources/ui/app/assets/data/apps/apps.json @@ -1,87 +1,421 @@ { - "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps?fields=*", + "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/?fields=*&_=1409569396360", "items" : [ { - "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/1401266754647_1", - "appVersion" : "0.96.0.2.1.1", - "components" : null, - "configs" : null, - "description" : "apache hbase is the hadoop database, a distributed, scalable, big data\n ...", - "diagnostics" : "org.apache.slider.core.exceptions.TriggerClusterTeardownException: Unstable Application Instance : - failed with role HBASE_MASTER failing 6 times (2 in startup); threshold is 5 - last failure: Failure container_1401266754647_0001_01_000008 on host c6401.ambari.apache.org, see http://c6401.ambari.apache.org:19888/jobhistory/logs/c6401.ambari.apache.org:45454/container_1401266754647_0001_01_000008/ctx/yarn", - "endTime" : 1401267215643, - "id" : "1401266754647_1", - "instance_name" : "SLIDER_1", - "jmx" : null, - "name" : "hoya30", - "startTime" : 1401266799783, - "state" : "FROZEN", - "type" : "hbase", - "urls" : null, - "user" : "yarn", - "version" : "1.0.0", - "view_name" : "SLIDER", - "yarnId" : "application_1401266754647_0001" - }, - { - "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/1401266754647_2", - "appVersion" : "0.96.0.2.1.1", - "components" : null, - "configs" : null, - "description" : "apache hbase is the hadoop database, a distributed, scalable, big data\n ...", - "diagnostics" : "Application application_1401266754647_0002 failed 2 times due to AM Container for appattempt_1401266754647_0002_000002 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException: \norg.apache.hadoop.util.Shell$ExitCodeException: \n\tat org.apache.hadoop.util.Shell.runCommand(Shell.java:505)\n\tat org.apache.hadoop.util.Shell.run(Shell.java:418)\n\tat org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:650)\n\tat org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)\n\tat org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:300)\n\tat org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:81)\n\tat java.util.concurrent.FutureTask.run(FutureTask.java:262)\n\tat java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoo lExecutor.java:1145)\n\tat java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)\n\tat java.lang.Thread.run(Thread.java:744)\n\n\nContainer exited with a non-zero exit code 1\n.Failing this attempt.. Failing the application.", - "endTime" : 1401298946291, - "id" : "1401266754647_2", + "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/1409333994422_5", + "appVersion" : "0.98.4-hadoop2", + "description" : "apache hbase is the hadoop database, a distributed, scalable, big data store....", + "diagnostics" : "", + "endTime" : 0, + "id" : "1409333994422_5", "instance_name" : "SLIDER_1", - "jmx" : null, - "name" : "hoya32", - "startTime" : 1401298942807, - "state" : "FAILED", + "metrics" : null, + "name" : "h4", + "startTime" : 1409348496653, + "state" : "RUNNING", "type" : "hbase", - "urls" : null, "user" : "yarn", "version" : "1.0.0", "view_name" : "SLIDER", - "yarnId" : "application_1401266754647_0002" + "yarnId" : "application_1409333994422_0005", + "components" : { + "HBASE_MASTER" : { + "componentName" : "HBASE_MASTER", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000002" : { + "released" : "false", + "startTime" : "1409348511279", + "createTime" : "1409348510529", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\"]", + "host" : "Ñ6402.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000002", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0005_01_000002___HBASE_MASTER --zk-quorum Ñ6401.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h4 ; ", + "state" : "3", + "role" : "HBASE_MASTER", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "1" + } + }, + "completedContainers" : { } + }, + "HBASE_REGIONSERVER" : { + "componentName" : "HBASE_REGIONSERVER", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000003" : { + "released" : "false", + "startTime" : "1409348510717", + "createTime" : "1409348510581", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\"]", + "host" : "Ñ6403.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000003", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0005_01_000003___HBASE_REGIONSERVER --zk-quorum Ñ6401.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h4 ; ", + "state" : "3", + "role" : "HBASE_REGIONSERVER", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "2" + } + }, + "completedContainers" : { } + }, + "HBASE_REST" : { + "componentName" : "HBASE_REST", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000004" : { + "released" : "false", + "startTime" : "1409348511120", + "createTime" : "1409348510581", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\"]", + "host" : "Ñ6401.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000004", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0005_01_000004___HBASE_REST --zk-quorum Ñ6401.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h4 ; ", + "state" : "3", + "role" : "HBASE_REST", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "3" + } + }, + "completedContainers" : { } + }, + "HBASE_THRIFT" : { + "componentName" : "HBASE_THRIFT", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000005" : { + "released" : "false", + "startTime" : "1409348511479", + "createTime" : "1409348510555", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\"]", + "host" : "Ñ6404.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000005", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0005_01_000005___HBASE_THRIFT --zk-quorum Ñ6401.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h4 ; ", + "state" : "3", + "role" : "HBASE_THRIFT", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "4" + } + }, + "completedContainers" : { } + }, + "HBASE_THRIFT2" : { + "componentName" : "HBASE_THRIFT2", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000006" : { + "released" : "false", + "startTime" : "1409348511727", + "createTime" : "1409348511700", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\"]", + "host" : "Ñ6402.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000006", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0005_01_000006___HBASE_THRIFT2 --zk-quorum Ñ6401.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h4 ; ", + "state" : "3", + "role" : "HBASE_THRIFT2", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "5" + } + }, + "completedContainers" : { } + }, + "slider-appmaster" : { + "componentName" : "slider-appmaster", + "instanceCount" : 1, + "activeContainers" : { + "container_1409333994422_0005_01_000001" : { + "released" : "false", + "startTime" : "0", + "createTime" : "0", + "host" : "Ñ6403.ambari.apache.org", + "name" : "container_1409333994422_0005_01_000001", + "command" : "", + "state" : "3", + "role" : "slider-appmaster", + "exitCode" : "0", + "diagnostics" : "", + "roleId" : "0" + } + }, + "completedContainers" : { } + } + }, + "configs" : { + "componentinstancedata" : { + "container_1409333994422_0005_01_000002.app.jmx" : "Ñ6402.ambari.apache.org:52349/jmx", + "container_1409333994422_0005_01_000002.app.monitor" : "Ñ6402.ambari.apache.org:52349/master-status" + }, + "core-site" : { + "fs.defaultFS" : "hdfs://Ñ6401.ambari.apache.org:8020", + "fs.trash.interval" : "360", + "hadoop.proxyuser.falcon.groups" : "users", + "hadoop.proxyuser.falcon.hosts" : "*", + "hadoop.proxyuser.hcat.groups" : "users", + "hadoop.proxyuser.hcat.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.proxyuser.hive.groups" : "users", + "hadoop.proxyuser.hive.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.proxyuser.oozie.groups" : "users", + "hadoop.proxyuser.oozie.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.security.auth_to_local" : "\n RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/\n RULE:[2:$1@$0](jhs@.*)s/.*/mapred/\n RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/\n RULE:[2:$1@$0](hm@.*)s/.*/hbase/\n RULE:[2:$1@$0](rs@.*)s/.*/hbase/\n DEFAULT", + "hadoop.security.authentication" : "simple", + "hadoop.security.authorization" : "false", + "io.compression.codecs" : "org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec", + "io.file.buffer.size" : "131072", + "io.serializations" : "org.apache.hadoop.io.serializer.WritableSerialization", + "ipc.client.connect.max.retries" : "50", + "ipc.client.connection.maxidletime" : "30000", + "ipc.client.idlethreshold" : "8000", + "mapreduce.jobtracker.webinterface.trusted" : "false" + }, + "global" : { + "app_container_id" : "container_1409333994422_0005_01_000002", + "app_input_conf_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002/propagatedconf", + "app_install_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002/app/install", + "app_log_dir" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000002", + "app_pid_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002/app/run", + "app_root" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002/app/install/hbase-0.98.4-hadoop2", + "app_user" : "yarn", + "ganglia_server_host" : "Ñ6401.ambari.apache.org", + "ganglia_server_id" : "Application1", + "ganglia_server_port" : "8667", + "hbase_instance_name" : "instancename", + "hbase_rest_port" : "49728", + "hbase_root_password" : "secret", + "hbase_thrift2_port" : "59165", + "hbase_thrift_port" : "53679", + "monitor_protocol" : "http", + "security_enabled" : "false", + "user_group" : "hadoop" + }, + "hbase-env" : { + "content" : "# Set environment variables here.\n\n# The java implementation to use. Java 1.6 required.\nexport JAVA_HOME={{java64_home}}\n\n# HBase Configuration directory\nexport HBASE_CONF_DIR=${HBASE_CONF_DIR:-{{conf_dir}}}\n\n# Extra Java CLASSPATH elements. Optional.\nexport HBASE_CLASSPATH=${HBASE_CLASSPATH}\n\n# The maximum amount of heap to use, in MB. Default is 1000.\n# export HBASE_HEAPSIZE=1000\n\n# Extra Java runtime options.\n# Below are what we set by default. May only work with SUN JVM.\n# For more on why as well as other possible settings,\n# see http://wiki.apache.org/hadoop/PerformanceTuning\nexport HBASE_OPTS=\"-XX:+UseConcMarkSweepGC -XX:ErrorFile={{log_dir}}/hs_err_pid%p.log\"\nexport SERVER_GC_OPTS=\"-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:{{log_dir}}/gc.log-`date +'%Y%m%d%H%M'`\"\n# Uncomment below to enable java garbage collection logging.\n# export HBASE_OPTS=\"$HBASE_OPTS -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStam ps -Xloggc:$HBASE_HOME/logs/gc-hbase.log\"\n\n# Uncomment and adjust to enable JMX exporting\n# See jmxremote.password and jmxremote.access in $JRE_HOME/lib/management to configure remote password access.\n# More details at: http://java.sun.com/javase/6/docs/technotes/guides/management/agent.html\n#\n# export HBASE_JMX_BASE=\"-Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.authenticate=false\"\nexport HBASE_MASTER_OPTS=\"-Xmx{{master_heapsize}}\"\nexport HBASE_REGIONSERVER_OPTS=\"-Xmn{{regionserver_xmn_size}} -XX:CMSInitiatingOccupancyFraction=70 -Xms{{regionserver_heapsize}} -Xmx{{regionserver_heapsize}}\"\n# export HBASE_THRIFT_OPTS=\"$HBASE_JMX_BASE -Dcom.sun.management.jmxremote.port=10103\"\n# export HBASE_ZOOKEEPER_OPTS=\"$HBASE_JMX_BASE -Dcom.sun.management.jmxremote.port=10104\"\n\n# File naming hosts on which HRegionServers will run. $HBASE_HOME/conf/regionservers by default.\nexport HBASE_REGIONSERVERS=${HBASE_CONF_DIR}/regionservers\n\n# Extra ssh options. Empty by default.\n# export HBASE_SSH_OPTS=\"-o ConnectTimeout=1 -o SendEnv=HBASE_CONF_DIR\"\n\n# Where log files are stored. $HBASE_HOME/logs by default.\nexport HBASE_LOG_DIR={{log_dir}}\n\n# A string representing this instance of hbase. $USER by default.\n# export HBASE_IDENT_STRING=$USER\n\n# The scheduling priority for daemon processes. See 'man nice'.\n# export HBASE_NICENESS=10\n\n# The directory where pid files are stored. /tmp by default.\nexport HBASE_PID_DIR={{pid_dir}}\n\n# Seconds to sleep between slave commands. Unset by default. This\n# can be useful in large clusters, where, e.g., slave rsyncs can\n# otherwise arrive faster than the master can service them.\n# export HBASE_SLAVE_SLEEP=0.1\n\n# Tell HBase whether it should manage it's own instance of Zookeeper or not.\nexport HBASE_MANAGES_ZK=false\n\n{% if security_enabled %}\nexport HBASE_OPTS=\"$HBASE_OPTS -Djava.security.auth.login.config={{client_jaas_config_file}}\"\nexport HBASE_MASTER_OPTS=\"$HBASE_M ASTER_OPTS -Djava.security.auth.login.config={{master_jaas_config_file}}\"\nexport HBASE_REGIONSERVER_OPTS=\"$HBASE_REGIONSERVER_OPTS -Djava.security.auth.login.config={{regionserver_jaas_config_file}}\"\n{% endif %}", + "hbase_master_heapsize" : "1024m", + "hbase_regionserver_heapsize" : "1024m", + "hbase_regionserver_xmn_max" : "512", + "hbase_regionserver_xmn_ratio" : "0.2" + }, + "hbase-log4j" : { + "content" : "# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\n# Define some default values that can be overridden by system properties\nhbase.root.logger=INFO,console\nhbase.security.logger=INFO,console\nhbase.log.dir=.\nhbase.lo g.file=hbase.log\n\n# Define the root logger to the system property \"hbase.root.logger\".\nlog4j.rootLogger=${hbase.root.logger}\n\n# Logging Threshold\nlog4j.threshold=ALL\n\n#\n# Daily Rolling File Appender\n#\nlog4j.appender.DRFA=org.apache.log4j.DailyRollingFileAppender\nlog4j.appender.DRFA.File=${hbase.log.dir}/${hbase.log.file}\n\n# Rollver at midnight\nlog4j.appender.DRFA.DatePattern=.yyyy-MM-dd\n\n# 30-day backup\n#log4j.appender.DRFA.MaxBackupIndex=30\nlog4j.appender.DRFA.layout=org.apache.log4j.PatternLayout\n\n# Pattern format: Date LogLevel LoggerName LogMessage\nlog4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n# Rolling File Appender properties\nhbase.log.maxfilesize=256MB\nhbase.log.maxbackupindex=20\n\n# Rolling File Appender\nlog4j.appender.RFA=org.apache.log4j.RollingFileAppender\nlog4j.appender.RFA.File=${hbase.log.dir}/${hbase.log.file}\n\nlog4j.appender.RFA.MaxFileSize=${hbase.log.maxfilesize}\nlog4j.appender.RFA.MaxBackupIndex=$ {hbase.log.maxbackupindex}\n\nlog4j.appender.RFA.layout=org.apache.log4j.PatternLayout\nlog4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n#\n# Security audit appender\n#\nhbase.security.log.file=SecurityAuth.audit\nhbase.security.log.maxfilesize=256MB\nhbase.security.log.maxbackupindex=20\nlog4j.appender.RFAS=org.apache.log4j.RollingFileAppender\nlog4j.appender.RFAS.File=${hbase.log.dir}/${hbase.security.log.file}\nlog4j.appender.RFAS.MaxFileSize=${hbase.security.log.maxfilesize}\nlog4j.appender.RFAS.MaxBackupIndex=${hbase.security.log.maxbackupindex}\nlog4j.appender.RFAS.layout=org.apache.log4j.PatternLayout\nlog4j.appender.RFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n\nlog4j.category.SecurityLogger=${hbase.security.logger}\nlog4j.additivity.SecurityLogger=false\n#log4j.logger.SecurityLogger.org.apache.hadoop.hbase.security.access.AccessController=TRACE\n\n#\n# Null Appender\n#\nlog4j.appender.NullAppender=org.apache.log4j.varia.NullAppender\n \n#\n# console\n# Add \"console\" to rootlogger above if you want to use this\n#\nlog4j.appender.console=org.apache.log4j.ConsoleAppender\nlog4j.appender.console.target=System.err\nlog4j.appender.console.layout=org.apache.log4j.PatternLayout\nlog4j.appender.console.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n# Custom Logging levels\n\nlog4j.logger.org.apache.zookeeper=INFO\n#log4j.logger.org.apache.hadoop.fs.FSNamesystem=DEBUG\nlog4j.logger.org.apache.hadoop.hbase=DEBUG\n# Make these two classes INFO-level. Make them DEBUG to see more zk debug.\nlog4j.logger.org.apache.hadoop.hbase.zookeeper.ZKUtil=INFO\nlog4j.logger.org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher=INFO\n#log4j.logger.org.apache.hadoop.dfs=DEBUG\n# Set this class to log INFO only otherwise its OTT\n# Enable this to get detailed connection error/retry logging.\n# log4j.logger.org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation=TRACE\n\n\n# Uncomment this line to enable trac ing on _every_ RPC call (this can be a lot of output)\n#log4j.logger.org.apache.hadoop.ipc.HBaseServer.trace=DEBUG\n\n# Uncomment the below if you want to remove logging of client region caching'\n# and scan of .META. messages\n# log4j.logger.org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation=INFO\n# log4j.logger.org.apache.hadoop.hbase.client.MetaScanner=INFO" + }, + "hbase-policy" : { + "security.admin.protocol.acl" : "*", + "security.client.protocol.acl" : "*", + "security.masterregion.protocol.acl" : "*" + }, + "hbase-site" : { + "dfs.domain.socket.path" : "/var/lib/hadoop-hdfs/dn_socket", + "hbase.client.keyvalue.maxsize" : "10485760", + "hbase.client.scanner.caching" : "100", + "hbase.cluster.distributed" : "true", + "hbase.defaults.for.version.skip" : "true", + "hbase.hregion.majorcompaction" : "86400000", + "hbase.hregion.max.filesize" : "10737418240", + "hbase.hregion.memstore.block.multiplier" : "2", + "hbase.hregion.memstore.flush.size" : "134217728", + "hbase.hregion.memstore.mslab.enabled" : "true", + "hbase.hstore.blockingStoreFiles" : "10", + "hbase.hstore.compactionThreshold" : "3", + "hbase.hstore.flush.retries.number" : "120", + "hbase.local.dir" : "${hbase.tmp.dir}/local", + "hbase.master.info.bindAddress" : "0.0.0.0", + "hbase.master.info.port" : "52349", + "hbase.master.kerberos.principal" : "hbase/_h...@example.com", + "hbase.master.keytab.file" : "/etc/security/keytabs/hbase.service.keytab", + "hbase.master.port" : "60000", + "hbase.regionserver.global.memstore.lowerLimit" : "0.38", + "hbase.regionserver.global.memstore.upperLimit" : "0.4", + "hbase.regionserver.handler.count" : "60", + "hbase.regionserver.info.port" : "0", + "hbase.regionserver.kerberos.principal" : "hbase/_h...@example.com", + "hbase.regionserver.keytab.file" : "/etc/security/keytabs/hbase.service.keytab", + "hbase.regionserver.port" : "0", + "hbase.rootdir" : "hdfs://Ñ6401.ambari.apache.org:8020/user/yarn/.slider/cluster/h4/database/data", + "hbase.security.authentication" : "simple", + "hbase.security.authorization" : "false", + "hbase.stagingdir" : "hdfs://Ñ6401.ambari.apache.org:8020/user/yarn/.slider/cluster/h4/database/staging", + "hbase.superuser" : "yarn", + "hbase.tmp.dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002/work/app/tmp", + "hbase.zookeeper.property.clientPort" : "2181", + "hbase.zookeeper.quorum" : "Ñ6401.ambari.apache.org", + "hbase.zookeeper.useMulti" : "true", + "hfile.block.cache.size" : "0.40", + "zookeeper.session.timeout" : "30000", + "zookeeper.znode.parent" : "/services/slider/users/yarn/h4" + }, + "hdfs-site" : { + "dfs.block.access.token.enable" : "true", + "dfs.blockreport.initialDelay" : "120", + "dfs.blocksize" : "134217728", + "dfs.client.read.shortcircuit" : "true", + "dfs.client.read.shortcircuit.streams.cache.size" : "4096", + "dfs.cluster.administrators" : " hdfs", + "dfs.datanode.address" : "0.0.0.0:50010", + "dfs.datanode.balance.bandwidthPerSec" : "6250000", + "dfs.datanode.data.dir" : "/hadoop/hdfs/data", + "dfs.datanode.data.dir.perm" : "750", + "dfs.datanode.du.reserved" : "1073741824", + "dfs.datanode.failed.volumes.tolerated" : "0", + "dfs.datanode.http.address" : "0.0.0.0:50075", + "dfs.datanode.ipc.address" : "0.0.0.0:8010", + "dfs.datanode.kerberos.principal" : "dn/_h...@example.com", + "dfs.datanode.keytab.file" : "/etc/security/keytabs/dn.service.keytab", + "dfs.datanode.max.transfer.threads" : "1024", + "dfs.domain.socket.path" : "/var/lib/hadoop-hdfs/dn_socket", + "dfs.heartbeat.interval" : "3", + "dfs.hosts.exclude" : "/etc/hadoop/conf/dfs.exclude", + "dfs.https.port" : "50470", + "dfs.journalnode.edits.dir" : "/grid/0/hdfs/journal", + "dfs.journalnode.http-address" : "0.0.0.0:8480", + "dfs.namenode.accesstime.precision" : "0", + "dfs.namenode.avoid.read.stale.datanode" : "true", + "dfs.namenode.avoid.write.stale.datanode" : "true", + "dfs.namenode.checkpoint.dir" : "/hadoop/hdfs/namesecondary", + "dfs.namenode.checkpoint.edits.dir" : "/hadoop/hdfs/namesecondary", + "dfs.namenode.checkpoint.period" : "21600", + "dfs.namenode.checkpoint.txns" : "1000000", + "dfs.namenode.handler.count" : "40", + "dfs.namenode.http-address" : "Ñ6401.ambari.apache.org:50070", + "dfs.namenode.https-address" : "Ñ6401.ambari.apache.org:50470", + "dfs.namenode.kerberos.https.principal" : "HTTP/_h...@example.com", + "dfs.namenode.kerberos.principal" : "nn/_h...@example.com", + "dfs.namenode.keytab.file" : "/etc/security/keytabs/nn.service.keytab", + "dfs.namenode.name.dir" : "/hadoop/hdfs/namenode", + "dfs.namenode.name.dir.restore" : "true", + "dfs.namenode.safemode.threshold-pct" : "1.0f", + "dfs.namenode.secondary.http-address" : "Ñ6402.ambari.apache.org:50090", + "dfs.namenode.stale.datanode.interval" : "30000", + "dfs.namenode.write.stale.datanode.ratio" : "1.0f", + "dfs.permissions.enabled" : "true", + "dfs.permissions.superusergroup" : "hdfs", + "dfs.replication" : "3", + "dfs.replication.max" : "50", + "dfs.secondary.namenode.kerberos.https.principal" : "HTTP/_h...@example.com", + "dfs.secondary.namenode.kerberos.principal" : "nn/_h...@example.com", + "dfs.secondary.namenode.keytab.file" : "/etc/security/keytabs/nn.service.keytab", + "dfs.support.append" : "true", + "dfs.web.authentication.kerberos.keytab" : "/etc/security/keytabs/spnego.service.keytab", + "dfs.web.authentication.kerberos.principal" : "HTTP/_h...@example.com", + "dfs.webhdfs.enabled" : "true", + "fs.permissions.umask-mode" : "022" + }, + "logfolders" : { + "Ñ6401.ambari.apache.org-container_1409333994422_0005_01_000004-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000004", + "Ñ6401.ambari.apache.org-container_1409333994422_0005_01_000004-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000004", + "Ñ6402.ambari.apache.org-container_1409333994422_0005_01_000002-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000002", + "Ñ6402.ambari.apache.org-container_1409333994422_0005_01_000002-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000002", + "Ñ6402.ambari.apache.org-container_1409333994422_0005_01_000006-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000006", + "Ñ6402.ambari.apache.org-container_1409333994422_0005_01_000006-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000006", + "Ñ6403.ambari.apache.org-container_1409333994422_0005_01_000003-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000003", + "Ñ6403.ambari.apache.org-container_1409333994422_0005_01_000003-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000003", + "Ñ6404.ambari.apache.org-container_1409333994422_0005_01_000005-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0005/container_1409333994422_0005_01_000005", + "Ñ6404.ambari.apache.org-container_1409333994422_0005_01_000005-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0005/container_1409333994422_0005_01_000005" + }, + "yarn-site" : { + "yarn.acl.enable" : "false", + "yarn.application.classpath" : "/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*", + "yarn.log-aggregation-enable" : "true", + "yarn.log-aggregation.retain-seconds" : "2592000", + "yarn.log.server.url" : "http://Ñ6402.ambari.apache.org:19888/jobhistory/logs", + "yarn.nodemanager.address" : "0.0.0.0:45454", + "yarn.nodemanager.admin-env" : "MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX", + "yarn.nodemanager.aux-services" : "mapreduce_shuffle", + "yarn.nodemanager.aux-services.mapreduce_shuffle.class" : "org.apache.hadoop.mapred.ShuffleHandler", + "yarn.nodemanager.container-executor.class" : "org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor", + "yarn.nodemanager.container-monitor.interval-ms" : "3000", + "yarn.nodemanager.delete.debug-delay-sec" : "0", + "yarn.nodemanager.disk-health-checker.min-healthy-disks" : "0.25", + "yarn.nodemanager.health-checker.interval-ms" : "135000", + "yarn.nodemanager.health-checker.script.timeout-ms" : "60000", + "yarn.nodemanager.linux-container-executor.group" : "hadoop", + "yarn.nodemanager.local-dirs" : "/hadoop/yarn/local", + "yarn.nodemanager.log-aggregation.compression-type" : "gz", + "yarn.nodemanager.log-dirs" : "/hadoop/yarn/log", + "yarn.nodemanager.log.retain-second" : "604800", + "yarn.nodemanager.remote-app-log-dir" : "/app-logs", + "yarn.nodemanager.remote-app-log-dir-suffix" : "logs", + "yarn.nodemanager.resource.memory-mb" : "2046", + "yarn.nodemanager.vmem-check-enabled" : "false", + "yarn.nodemanager.vmem-pmem-ratio" : "2.1", + "yarn.resourcemanager.address" : "Ñ6402.ambari.apache.org:8050", + "yarn.resourcemanager.admin.address" : "Ñ6402.ambari.apache.org:8141", + "yarn.resourcemanager.am.max-attempts" : "2", + "yarn.resourcemanager.hostname" : "Ñ6402.ambari.apache.org", + "yarn.resourcemanager.nodes.exclude-path" : "/etc/hadoop/conf/yarn.exclude", + "yarn.resourcemanager.resource-tracker.address" : "Ñ6402.ambari.apache.org:8025", + "yarn.resourcemanager.scheduler.address" : "Ñ6402.ambari.apache.org:8030", + "yarn.resourcemanager.scheduler.class" : "org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler", + "yarn.resourcemanager.webapp.address" : "Ñ6402.ambari.apache.org:8088", + "yarn.scheduler.maximum-allocation-mb" : "2046", + "yarn.scheduler.minimum-allocation-mb" : "682", + "yarn.timeline-service.address" : "Ñ6402.ambari.apache.org:10200", + "yarn.timeline-service.enabled" : "true", + "yarn.timeline-service.generic-application-history.store-class" : "org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore", + "yarn.timeline-service.leveldb-timeline-store.path" : "/hadoop/yarn/timeline", + "yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms" : "300000", + "yarn.timeline-service.store-class" : "org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore", + "yarn.timeline-service.ttl-enable" : "true", + "yarn.timeline-service.ttl-ms" : "2678400000", + "yarn.timeline-service.webapp.address" : "Ñ6402.ambari.apache.org:8188", + "yarn.timeline-service.webapp.https.address" : "Ñ6402.ambari.apache.org:8190" + } + }, + "jmx" : { + "ClusterId" : "d34a2f54-f73a-4166-87f4-501b465ac25b", + "ClusterRequests" : "2241", + "DeadRegionServers" : "0", + "IsActiveMaster" : "true", + "MasterActiveTime" : "1409348614409", + "MasterStartTime" : "1409348612516", + "MetricAverageLoad" : "2.0", + "RegionServers" : "1", + "ServerName" : "Ñ6402.ambari.apache.org,60000,1409348607747", + "ZookeeperQuorum" : "Ñ6401.ambari.apache.org:2181" + }, + "urls" : { + "JMX" : "http://Ñ6402.ambari.apache.org:52349/jmx", + "UI" : "http://Ñ6402.ambari.apache.org:52349/master-status", + "app.ganglia" : "http://Ñ6401.ambari.apache.org/ganglia?c=Application1", + "app.metrics" : "http://Ñ6401.ambari.apache.org/cgi-bin/rrd.py?c=Application1", + "org.apache.slider.hbase.rest" : "http://Ñ6401.ambari.apache.org:49728", + "org.apache.slider.hbase.thrift" : "http://Ñ6404.ambari.apache.org:53679", + "org.apache.slider.hbase.thrift2" : "http://Ñ6402.ambari.apache.org:59165" + } }, { - "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/1401266754647_3", - "appVersion" : "0.96.0.2.1.1", - "description" : "apache hbase is the hadoop database, a distributed, scalable, big data\n ...", + "href" : "http://c6401:8080/api/v1/views/SLIDER/versions/1.0.0/instances/SLIDER_1/apps/1409333994422_9", + "appVersion" : "0.98.4-hadoop2", + "description" : "apache hbase is the hadoop database, a distributed, scalable, big data store....", "diagnostics" : "", "endTime" : 0, - "id" : "1401266754647_3", + "id" : "1409333994422_9", "instance_name" : "SLIDER_1", - "name" : "hoya30", - "startTime" : 1401299043372, + "metrics" : null, + "name" : "h10", + "startTime" : 1409359943615, "state" : "RUNNING", "type" : "hbase", "user" : "yarn", "version" : "1.0.0", "view_name" : "SLIDER", - "yarnId" : "application_1401266754647_0003", - "metrics" : { - "metric1" : "", - "metric2" : "", - "metric3" : "", - "metric4" : "", - "metric5" : "", - "metric6" : "", - "metric7" : "", - "metric8" : "" - }, + "yarnId" : "application_1409333994422_0009", "components" : { "HBASE_MASTER" : { "componentName" : "HBASE_MASTER", - "instanceCount" : 2, + "instanceCount" : 1, "activeContainers" : { - "container_1401266754647_0003_01_000002" : { + "container_1409333994422_0009_01_000002" : { "released" : "false", - "startTime" : "1401299052926", - "createTime" : "1401299052732", - "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", MALLOC_ARENA_MAX=\"4\"]", - "host" : "c6401.ambari.apache.org", - "name" : "container_1401266754647_0003_01_000002", - "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1401266754647_0003_01_000002___HBASE_MASTER --host c6401.ambari.apache.org --port 36278 ; ", + "startTime" : "1409359961882", + "createTime" : "1409359961649", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\", MALLOC_ARENA_MAX=\"4\"]", + "host" : "Ñ6403.ambari.apache.org", + "name" : "container_1409333994422_0009_01_000002", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0009_01_000002___HBASE_MASTER --zk-quorum Ñ6403.ambari.apache.org:2181,Ñ6401.ambari.apache.org:2181,Ñ6402.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h10 ; ", "state" : "3", "role" : "HBASE_MASTER", "exitCode" : "0", @@ -95,14 +429,14 @@ "componentName" : "HBASE_REGIONSERVER", "instanceCount" : 1, "activeContainers" : { - "container_1401266754647_0003_01_000003" : { + "container_1409333994422_0009_01_000003" : { "released" : "false", - "startTime" : "1401299053317", - "createTime" : "1401299053307", - "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", MALLOC_ARENA_MAX=\"4\"]", - "host" : "c6401.ambari.apache.org", - "name" : "container_1401266754647_0003_01_000003", - "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1401266754647_0003_01_000003___HBASE_REGIONSERVER --host c6401.ambari.apache.org --port 36278 ; ", + "startTime" : "1409359962422", + "createTime" : "1409359961566", + "environment" : "[AGENT_WORK_ROOT=\"$PWD\", HADOOP_USER_NAME=\"yarn\", AGENT_LOG_ROOT=\"$LOG_DIRS\", PYTHONPATH=\"./infra/agent/slider-agent/\", SLIDER_PASSPHRASE=\"DEV\", MALLOC_ARENA_MAX=\"4\"]", + "host" : "Ñ6404.ambari.apache.org", + "name" : "container_1409333994422_0009_01_000003", + "command" : "python ./infra/agent/slider-agent/agent/main.py --label container_1409333994422_0009_01_000003___HBASE_REGIONSERVER --zk-quorum Ñ6403.ambari.apache.org:2181,Ñ6401.ambari.apache.org:2181,Ñ6402.ambari.apache.org:2181 --zk-reg-path /registry/org-apache-slider/h10 ; ", "state" : "3", "role" : "HBASE_REGIONSERVER", "exitCode" : "0", @@ -116,12 +450,12 @@ "componentName" : "slider-appmaster", "instanceCount" : 1, "activeContainers" : { - "container_1401266754647_0003_01_000001" : { + "container_1409333994422_0009_01_000001" : { "released" : "false", "startTime" : "0", "createTime" : "0", - "host" : "c6401.ambari.apache.org", - "name" : "container_1401266754647_0003_01_000001", + "host" : "Ñ6404.ambari.apache.org", + "name" : "container_1409333994422_0009_01_000001", "command" : "", "state" : "3", "role" : "slider-appmaster", @@ -134,26 +468,69 @@ } }, "configs" : { + "componentinstancedata" : { + "container_1409333994422_0009_01_000002.app.jmx" : "Ñ6403.ambari.apache.org:33940/jmx", + "container_1409333994422_0009_01_000002.app.monitor" : "Ñ6403.ambari.apache.org:33940/master-status" + }, "core-site" : { - "fs.defaultFS" : "hdfs://c6401.ambari.apache.org:8020" + "fs.defaultFS" : "hdfs://Ñ6401.ambari.apache.org:8020", + "fs.trash.interval" : "360", + "hadoop.proxyuser.falcon.groups" : "users", + "hadoop.proxyuser.falcon.hosts" : "*", + "hadoop.proxyuser.hcat.groups" : "users", + "hadoop.proxyuser.hcat.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.proxyuser.hive.groups" : "users", + "hadoop.proxyuser.hive.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.proxyuser.oozie.groups" : "users", + "hadoop.proxyuser.oozie.hosts" : "Ñ6402.ambari.apache.org", + "hadoop.security.auth_to_local" : "\n RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/\n RULE:[2:$1@$0](jhs@.*)s/.*/mapred/\n RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/\n RULE:[2:$1@$0](hm@.*)s/.*/hbase/\n RULE:[2:$1@$0](rs@.*)s/.*/hbase/\n DEFAULT", + "hadoop.security.authentication" : "simple", + "hadoop.security.authorization" : "false", + "io.compression.codecs" : "org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec", + "io.file.buffer.size" : "131072", + "io.serializations" : "org.apache.hadoop.io.serializer.WritableSerialization", + "ipc.client.connect.max.retries" : "50", + "ipc.client.connection.maxidletime" : "30000", + "ipc.client.idlethreshold" : "8000", + "mapreduce.jobtracker.webinterface.trusted" : "false" }, "global" : { - "app_install_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1401266754647_0003/container_1401266754647_0003_01_000002/app/install", - "app_log_dir" : "/hadoop/yarn/log/application_1401266754647_0003/container_1401266754647_0003_01_000002/app/log", - "app_pid_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1401266754647_0003/container_1401266754647_0003_01_000002/app/run", - "app_root" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1401266754647_0003/container_1401266754647_0003_01_000002/app/install/hbase-0.96.1-hadoop2", + "app_container_id" : "container_1409333994422_0009_01_000002", + "app_input_conf_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002/propagatedconf", + "app_install_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002/app/install", + "app_log_dir" : "/hadoop/yarn/log/application_1409333994422_0009/container_1409333994422_0009_01_000002", + "app_pid_dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002/app/run", + "app_root" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002/app/install/hbase-0.98.4-hadoop2", "app_user" : "yarn", - "ganglia_server_host" : "c6401.ambari.apache.org", + "ganglia_server_host" : "Ñ6402.ambari.apache.org", "ganglia_server_id" : "Application1", - "ganglia_server_port" : "8663", - "hbase_master_heapsize" : "1024m", - "hbase_regionserver_heapsize" : "1024m", - "hbase_regionserver_xmn_size" : "512", - "hbase_regionserver_xmn_ratio" : "0.2", + "ganglia_server_port" : "8667", + "hbase_instance_name" : "instancename", + "hbase_rest_port" : "0", + "hbase_root_password" : "secret", + "hbase_thrift2_port" : "0", + "hbase_thrift_port" : "0", + "monitor_protocol" : "http", "security_enabled" : "false", "user_group" : "hadoop" }, + "hbase-env" : { + "content" : "# Set environment variables here.\n\n# The java implementation to use. Java 1.6 required.\nexport JAVA_HOME={{java64_home}}\n\n# HBase Configuration directory\nexport HBASE_CONF_DIR=${HBASE_CONF_DIR:-{{conf_dir}}}\n\n# Extra Java CLASSPATH elements. Optional.\nexport HBASE_CLASSPATH=${HBASE_CLASSPATH}\n\n# The maximum amount of heap to use, in MB. Default is 1000.\n# export HBASE_HEAPSIZE=1000\n\n# Extra Java runtime options.\n# Below are what we set by default. May only work with SUN JVM.\n# For more on why as well as other possible settings,\n# see http://wiki.apache.org/hadoop/PerformanceTuning\nexport HBASE_OPTS=\"-XX:+UseConcMarkSweepGC -XX:ErrorFile={{log_dir}}/hs_err_pid%p.log\"\nexport SERVER_GC_OPTS=\"-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:{{log_dir}}/gc.log-`date +'%Y%m%d%H%M'`\"\n# Uncomment below to enable java garbage collection logging.\n# export HBASE_OPTS=\"$HBASE_OPTS -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStam ps -Xloggc:$HBASE_HOME/logs/gc-hbase.log\"\n\n# Uncomment and adjust to enable JMX exporting\n# See jmxremote.password and jmxremote.access in $JRE_HOME/lib/management to configure remote password access.\n# More details at: http://java.sun.com/javase/6/docs/technotes/guides/management/agent.html\n#\n# export HBASE_JMX_BASE=\"-Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.authenticate=false\"\nexport HBASE_MASTER_OPTS=\"-Xmx{{master_heapsize}}\"\nexport HBASE_REGIONSERVER_OPTS=\"-Xmn{{regionserver_xmn_size}} -XX:CMSInitiatingOccupancyFraction=70 -Xms{{regionserver_heapsize}} -Xmx{{regionserver_heapsize}}\"\n# export HBASE_THRIFT_OPTS=\"$HBASE_JMX_BASE -Dcom.sun.management.jmxremote.port=10103\"\n# export HBASE_ZOOKEEPER_OPTS=\"$HBASE_JMX_BASE -Dcom.sun.management.jmxremote.port=10104\"\n\n# File naming hosts on which HRegionServers will run. $HBASE_HOME/conf/regionservers by default.\nexport HBASE_REGIONSERVERS=${HBASE_CONF_DIR}/regionservers\n\n# Extra ssh options. Empty by default.\n# export HBASE_SSH_OPTS=\"-o ConnectTimeout=1 -o SendEnv=HBASE_CONF_DIR\"\n\n# Where log files are stored. $HBASE_HOME/logs by default.\nexport HBASE_LOG_DIR={{log_dir}}\n\n# A string representing this instance of hbase. $USER by default.\n# export HBASE_IDENT_STRING=$USER\n\n# The scheduling priority for daemon processes. See 'man nice'.\n# export HBASE_NICENESS=10\n\n# The directory where pid files are stored. /tmp by default.\nexport HBASE_PID_DIR={{pid_dir}}\n\n# Seconds to sleep between slave commands. Unset by default. This\n# can be useful in large clusters, where, e.g., slave rsyncs can\n# otherwise arrive faster than the master can service them.\n# export HBASE_SLAVE_SLEEP=0.1\n\n# Tell HBase whether it should manage it's own instance of Zookeeper or not.\nexport HBASE_MANAGES_ZK=false\n\n{% if security_enabled %}\nexport HBASE_OPTS=\"$HBASE_OPTS -Djava.security.auth.login.config={{client_jaas_config_file}}\"\nexport HBASE_MASTER_OPTS=\"$HBASE_M ASTER_OPTS -Djava.security.auth.login.config={{master_jaas_config_file}}\"\nexport HBASE_REGIONSERVER_OPTS=\"$HBASE_REGIONSERVER_OPTS -Djava.security.auth.login.config={{regionserver_jaas_config_file}}\"\n{% endif %}", + "hbase_master_heapsize" : "1024m", + "hbase_regionserver_heapsize" : "1024m", + "hbase_regionserver_xmn_max" : "512", + "hbase_regionserver_xmn_ratio" : "0.2" + }, + "hbase-log4j" : { + "content" : "# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\n# Define some default values that can be overridden by system properties\nhbase.root.logger=INFO,console\nhbase.security.logger=INFO,console\nhbase.log.dir=.\nhbase.lo g.file=hbase.log\n\n# Define the root logger to the system property \"hbase.root.logger\".\nlog4j.rootLogger=${hbase.root.logger}\n\n# Logging Threshold\nlog4j.threshold=ALL\n\n#\n# Daily Rolling File Appender\n#\nlog4j.appender.DRFA=org.apache.log4j.DailyRollingFileAppender\nlog4j.appender.DRFA.File=${hbase.log.dir}/${hbase.log.file}\n\n# Rollver at midnight\nlog4j.appender.DRFA.DatePattern=.yyyy-MM-dd\n\n# 30-day backup\n#log4j.appender.DRFA.MaxBackupIndex=30\nlog4j.appender.DRFA.layout=org.apache.log4j.PatternLayout\n\n# Pattern format: Date LogLevel LoggerName LogMessage\nlog4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n# Rolling File Appender properties\nhbase.log.maxfilesize=256MB\nhbase.log.maxbackupindex=20\n\n# Rolling File Appender\nlog4j.appender.RFA=org.apache.log4j.RollingFileAppender\nlog4j.appender.RFA.File=${hbase.log.dir}/${hbase.log.file}\n\nlog4j.appender.RFA.MaxFileSize=${hbase.log.maxfilesize}\nlog4j.appender.RFA.MaxBackupIndex=$ {hbase.log.maxbackupindex}\n\nlog4j.appender.RFA.layout=org.apache.log4j.PatternLayout\nlog4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n#\n# Security audit appender\n#\nhbase.security.log.file=SecurityAuth.audit\nhbase.security.log.maxfilesize=256MB\nhbase.security.log.maxbackupindex=20\nlog4j.appender.RFAS=org.apache.log4j.RollingFileAppender\nlog4j.appender.RFAS.File=${hbase.log.dir}/${hbase.security.log.file}\nlog4j.appender.RFAS.MaxFileSize=${hbase.security.log.maxfilesize}\nlog4j.appender.RFAS.MaxBackupIndex=${hbase.security.log.maxbackupindex}\nlog4j.appender.RFAS.layout=org.apache.log4j.PatternLayout\nlog4j.appender.RFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n\nlog4j.category.SecurityLogger=${hbase.security.logger}\nlog4j.additivity.SecurityLogger=false\n#log4j.logger.SecurityLogger.org.apache.hadoop.hbase.security.access.AccessController=TRACE\n\n#\n# Null Appender\n#\nlog4j.appender.NullAppender=org.apache.log4j.varia.NullAppender\n \n#\n# console\n# Add \"console\" to rootlogger above if you want to use this\n#\nlog4j.appender.console=org.apache.log4j.ConsoleAppender\nlog4j.appender.console.target=System.err\nlog4j.appender.console.layout=org.apache.log4j.PatternLayout\nlog4j.appender.console.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m%n\n\n# Custom Logging levels\n\nlog4j.logger.org.apache.zookeeper=INFO\n#log4j.logger.org.apache.hadoop.fs.FSNamesystem=DEBUG\nlog4j.logger.org.apache.hadoop.hbase=DEBUG\n# Make these two classes INFO-level. Make them DEBUG to see more zk debug.\nlog4j.logger.org.apache.hadoop.hbase.zookeeper.ZKUtil=INFO\nlog4j.logger.org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher=INFO\n#log4j.logger.org.apache.hadoop.dfs=DEBUG\n# Set this class to log INFO only otherwise its OTT\n# Enable this to get detailed connection error/retry logging.\n# log4j.logger.org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation=TRACE\n\n\n# Uncomment this line to enable trac ing on _every_ RPC call (this can be a lot of output)\n#log4j.logger.org.apache.hadoop.ipc.HBaseServer.trace=DEBUG\n\n# Uncomment the below if you want to remove logging of client region caching'\n# and scan of .META. messages\n# log4j.logger.org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation=INFO\n# log4j.logger.org.apache.hadoop.hbase.client.MetaScanner=INFO" + }, + "hbase-policy" : { + "security.admin.protocol.acl" : "*", + "security.client.protocol.acl" : "*", + "security.masterregion.protocol.acl" : "*" + }, "hbase-site" : { + "dfs.domain.socket.path" : "/var/lib/hadoop-hdfs/dn_socket", "hbase.client.keyvalue.maxsize" : "10485760", "hbase.client.scanner.caching" : "100", "hbase.cluster.distributed" : "true", @@ -167,36 +544,99 @@ "hbase.hstore.compactionThreshold" : "3", "hbase.hstore.flush.retries.number" : "120", "hbase.local.dir" : "${hbase.tmp.dir}/local", - "hbase.master.info.port" : "39505", + "hbase.master.info.bindAddress" : "0.0.0.0", + "hbase.master.info.port" : "33940", + "hbase.master.kerberos.principal" : "hbase/_h...@example.com", + "hbase.master.keytab.file" : "/etc/security/keytabs/hbase.service.keytab", + "hbase.master.port" : "60000", "hbase.regionserver.global.memstore.lowerLimit" : "0.38", "hbase.regionserver.global.memstore.upperLimit" : "0.4", "hbase.regionserver.handler.count" : "60", "hbase.regionserver.info.port" : "0", + "hbase.regionserver.kerberos.principal" : "hbase/_h...@example.com", + "hbase.regionserver.keytab.file" : "/etc/security/keytabs/hbase.service.keytab", "hbase.regionserver.port" : "0", - "hbase.rootdir" : "hdfs://c6401.ambari.apache.org:8020/apps/hbase/data", + "hbase.rootdir" : "hdfs://Ñ6401.ambari.apache.org:8020/user/yarn/.slider/cluster/h10/database/data", "hbase.security.authentication" : "simple", "hbase.security.authorization" : "false", - "hbase.stagingdir" : "hdfs://c6401.ambari.apache.org:8020/apps/hbase/staging", + "hbase.stagingdir" : "hdfs://Ñ6401.ambari.apache.org:8020/user/yarn/.slider/cluster/h10/database/staging", "hbase.superuser" : "yarn", - "hbase.tmp.dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1401266754647_0003/container_1401266754647_0003_01_000002/work/app/tmp", + "hbase.tmp.dir" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002/work/app/tmp", "hbase.zookeeper.property.clientPort" : "2181", - "hbase.zookeeper.quorum" : "c6401.ambari.apache.org", + "hbase.zookeeper.quorum" : "Ñ6403.ambari.apache.org,Ñ6401.ambari.apache.org,Ñ6402.ambari.apache.org", "hbase.zookeeper.useMulti" : "true", "hfile.block.cache.size" : "0.40", "zookeeper.session.timeout" : "30000", - "zookeeper.znode.parent" : "/hbase-unsecure" + "zookeeper.znode.parent" : "/services/slider/users/yarn/h10" }, "hdfs-site" : { - "dfs.namenode.http-address" : "c6401.ambari.apache.org:50070", - "dfs.namenode.https-address" : "c6401.ambari.apache.org:50470" + "dfs.block.access.token.enable" : "true", + "dfs.blockreport.initialDelay" : "120", + "dfs.blocksize" : "134217728", + "dfs.client.read.shortcircuit" : "true", + "dfs.client.read.shortcircuit.streams.cache.size" : "4096", + "dfs.cluster.administrators" : " hdfs", + "dfs.datanode.address" : "0.0.0.0:50010", + "dfs.datanode.balance.bandwidthPerSec" : "6250000", + "dfs.datanode.data.dir" : "/hadoop/hdfs/data", + "dfs.datanode.data.dir.perm" : "750", + "dfs.datanode.du.reserved" : "1073741824", + "dfs.datanode.failed.volumes.tolerated" : "0", + "dfs.datanode.http.address" : "0.0.0.0:50075", + "dfs.datanode.ipc.address" : "0.0.0.0:8010", + "dfs.datanode.kerberos.principal" : "dn/_h...@example.com", + "dfs.datanode.keytab.file" : "/etc/security/keytabs/dn.service.keytab", + "dfs.datanode.max.transfer.threads" : "1024", + "dfs.domain.socket.path" : "/var/lib/hadoop-hdfs/dn_socket", + "dfs.heartbeat.interval" : "3", + "dfs.hosts.exclude" : "/etc/hadoop/conf/dfs.exclude", + "dfs.https.port" : "50470", + "dfs.journalnode.edits.dir" : "/grid/0/hdfs/journal", + "dfs.journalnode.http-address" : "0.0.0.0:8480", + "dfs.namenode.accesstime.precision" : "0", + "dfs.namenode.avoid.read.stale.datanode" : "true", + "dfs.namenode.avoid.write.stale.datanode" : "true", + "dfs.namenode.checkpoint.dir" : "/hadoop/hdfs/namesecondary", + "dfs.namenode.checkpoint.edits.dir" : "/hadoop/hdfs/namesecondary", + "dfs.namenode.checkpoint.period" : "21600", + "dfs.namenode.checkpoint.txns" : "1000000", + "dfs.namenode.handler.count" : "40", + "dfs.namenode.http-address" : "Ñ6401.ambari.apache.org:50070", + "dfs.namenode.https-address" : "Ñ6401.ambari.apache.org:50470", + "dfs.namenode.kerberos.https.principal" : "HTTP/_h...@example.com", + "dfs.namenode.kerberos.principal" : "nn/_h...@example.com", + "dfs.namenode.keytab.file" : "/etc/security/keytabs/nn.service.keytab", + "dfs.namenode.name.dir" : "/hadoop/hdfs/namenode", + "dfs.namenode.name.dir.restore" : "true", + "dfs.namenode.safemode.threshold-pct" : "1.0f", + "dfs.namenode.secondary.http-address" : "Ñ6402.ambari.apache.org:50090", + "dfs.namenode.stale.datanode.interval" : "30000", + "dfs.namenode.write.stale.datanode.ratio" : "1.0f", + "dfs.permissions.enabled" : "true", + "dfs.permissions.superusergroup" : "hdfs", + "dfs.replication" : "3", + "dfs.replication.max" : "50", + "dfs.secondary.namenode.kerberos.https.principal" : "HTTP/_h...@example.com", + "dfs.secondary.namenode.kerberos.principal" : "nn/_h...@example.com", + "dfs.secondary.namenode.keytab.file" : "/etc/security/keytabs/nn.service.keytab", + "dfs.support.append" : "true", + "dfs.web.authentication.kerberos.keytab" : "/etc/security/keytabs/spnego.service.keytab", + "dfs.web.authentication.kerberos.principal" : "HTTP/_h...@example.com", + "dfs.webhdfs.enabled" : "true", + "fs.permissions.umask-mode" : "022" + }, + "logfolders" : { + "Ñ6403.ambari.apache.org-container_1409333994422_0009_01_000002-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0009/container_1409333994422_0009_01_000002", + "Ñ6403.ambari.apache.org-container_1409333994422_0009_01_000002-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000002", + "Ñ6404.ambari.apache.org-container_1409333994422_0009_01_000003-AGENT_LOG_ROOT" : "/hadoop/yarn/log/application_1409333994422_0009/container_1409333994422_0009_01_000003", + "Ñ6404.ambari.apache.org-container_1409333994422_0009_01_000003-AGENT_WORK_ROOT" : "/hadoop/yarn/local/usercache/yarn/appcache/application_1409333994422_0009/container_1409333994422_0009_01_000003" }, "yarn-site" : { - "yarn.acl.enable" : "true", - "yarn.admin.acl" : "*", + "yarn.acl.enable" : "false", "yarn.application.classpath" : "/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*", "yarn.log-aggregation-enable" : "true", "yarn.log-aggregation.retain-seconds" : "2592000", - "yarn.log.server.url" : "http://c6401.ambari.apache.org:19888/jobhistory/logs", + "yarn.log.server.url" : "http://Ñ6402.ambari.apache.org:19888/jobhistory/logs", "yarn.nodemanager.address" : "0.0.0.0:45454", "yarn.nodemanager.admin-env" : "MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX", "yarn.nodemanager.aux-services" : "mapreduce_shuffle", @@ -214,48 +654,49 @@ "yarn.nodemanager.log.retain-second" : "604800", "yarn.nodemanager.remote-app-log-dir" : "/app-logs", "yarn.nodemanager.remote-app-log-dir-suffix" : "logs", - "yarn.nodemanager.resource.memory-mb" : "2048", + "yarn.nodemanager.resource.memory-mb" : "2046", "yarn.nodemanager.vmem-check-enabled" : "false", "yarn.nodemanager.vmem-pmem-ratio" : "2.1", - "yarn.resourcemanager.address" : "c6401.ambari.apache.org:8050", - "yarn.resourcemanager.admin.address" : "c6401.ambari.apache.org:8141", + "yarn.resourcemanager.address" : "Ñ6402.ambari.apache.org:8050", + "yarn.resourcemanager.admin.address" : "Ñ6402.ambari.apache.org:8141", "yarn.resourcemanager.am.max-attempts" : "2", - "yarn.resourcemanager.hostname" : "c6401.ambari.apache.org", + "yarn.resourcemanager.hostname" : "Ñ6402.ambari.apache.org", "yarn.resourcemanager.nodes.exclude-path" : "/etc/hadoop/conf/yarn.exclude", - "yarn.resourcemanager.resource-tracker.address" : "c6401.ambari.apache.org:8025", - "yarn.resourcemanager.scheduler.address" : "c6401.ambari.apache.org:8030", + "yarn.resourcemanager.resource-tracker.address" : "Ñ6402.ambari.apache.org:8025", + "yarn.resourcemanager.scheduler.address" : "Ñ6402.ambari.apache.org:8030", "yarn.resourcemanager.scheduler.class" : "org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler", - "yarn.resourcemanager.webapp.address" : "c6401.ambari.apache.org:8088", - "yarn.scheduler.maximum-allocation-mb" : "2048", + "yarn.resourcemanager.webapp.address" : "Ñ6402.ambari.apache.org:8088", + "yarn.scheduler.maximum-allocation-mb" : "2046", "yarn.scheduler.minimum-allocation-mb" : "682", - "yarn.timeline-service.address" : "c6401.ambari.apache.org:10200", + "yarn.timeline-service.address" : "Ñ6402.ambari.apache.org:10200", "yarn.timeline-service.enabled" : "true", "yarn.timeline-service.generic-application-history.store-class" : "org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore", - "yarn.timeline-service.leveldb-timeline-store.path" : "/var/log/hadoop-yarn/timeline", + "yarn.timeline-service.leveldb-timeline-store.path" : "/hadoop/yarn/timeline", "yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms" : "300000", - "yarn.timeline-service.store-class" : "org.apache.hadoop.yarn.server.applicationhistoryservice.timeline.LeveldbTimelineStore", + "yarn.timeline-service.store-class" : "org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore", "yarn.timeline-service.ttl-enable" : "true", "yarn.timeline-service.ttl-ms" : "2678400000", - "yarn.timeline-service.webapp.address" : "c6401.ambari.apache.org:8188", - "yarn.timeline-service.webapp.https.address" : "c6401.ambari.apache.org:8190" + "yarn.timeline-service.webapp.address" : "Ñ6402.ambari.apache.org:8188", + "yarn.timeline-service.webapp.https.address" : "Ñ6402.ambari.apache.org:8190" } }, "jmx" : { - "ClusterId" : "67ee848a-8576-45d3-9b18-cdda4eb6c025", - "DeadRegionServers" : "1", + "ClusterId" : "125f1f5b-713e-421a-9179-aadfc24a2a05", + "ClusterRequests" : "2127", + "DeadRegionServers" : "0", "IsActiveMaster" : "true", - "MasterActiveTime" : "1401332251511", - "MasterStartTime" : "1401332248911", + "MasterActiveTime" : "1409360031151", + "MasterStartTime" : "1409360028629", + "MetricAverageLoad" : "2.0", "RegionServers" : "1", - "ServerName" : "c6401.ambari.apache.org,60000,1401332245896", - "ZookeeperQuorum" : "c6401.ambari.apache.org:2181", - "cluster_requests" : "14496", - "metricAverageLoad" : "2.0" + "ServerName" : "Ñ6403.ambari.apache.org,60000,1409360025510", + "ZookeeperQuorum" : "Ñ6401.ambari.apache.org:2181,Ñ6403.ambari.apache.org:2181,Ñ6402.ambari.apache.org:2181" }, "urls" : { - "JMX" : "http://c6401.ambari.apache.org:39505/jmx", - "Metrics" : "http://c6401.ambari.apache.org/cgi-bin/rrd.py?c=Application1", - "UI" : "http://c6401.ambari.apache.org:39505/master-status" + "JMX" : "http://Ñ6403.ambari.apache.org:33940/jmx", + "UI" : "http://Ñ6403.ambari.apache.org:33940/master-status", + "app.ganglia" : "http://Ñ6402.ambari.apache.org/ganglia?c=Application1", + "app.metrics" : "http://Ñ6402.ambari.apache.org/cgi-bin/rrd.py?c=Application1" } } ] http://git-wip-us.apache.org/repos/asf/ambari/blob/a41ef2a1/contrib/views/slider/src/main/resources/ui/app/models/slider_app.js ---------------------------------------------------------------------- diff --git a/contrib/views/slider/src/main/resources/ui/app/models/slider_app.js b/contrib/views/slider/src/main/resources/ui/app/models/slider_app.js index 0b45607..69f2479 100644 --- a/contrib/views/slider/src/main/resources/ui/app/models/slider_app.js +++ b/contrib/views/slider/src/main/resources/ui/app/models/slider_app.js @@ -99,38 +99,25 @@ App.SliderApp = DS.Model.extend({ jmx: DS.attr('object'), - /** - * Global configs - * @type {{key: string, value: *}[]} - */ - globals: function() { - var c = this.get('configs.global'); - return this.mapObject(c); - }.property('configs.@each'), - - /** - * HBase-Site configs - * @type {{key: string, value: *}[]} - */ - hbaseSite: function() { - var c = this.get('configs.hbase-site'); - return this.mapObject(c); - }.property('configs.@each'), - - /** - * Configs which are not in global or hbase-site - * @type {{key: string, value: *}[]} - */ - otherConfigs: function() { - var c = this.get('configs'), - ret = [], - self = this; - if (Ember.typeOf(c) !== 'object') return []; - Ember.keys(c).forEach(function(key) { - if (['hbase-site', 'global'].contains(key)) return; - ret = ret.concat(self.mapObject(c[key])); - }); - return ret; + // Config categories, that should be hidden on app page + hiddenCategories: ['yarn-site', 'global'], + + /** + * Configs grouped to categories by site-objects + * @type {Array} + */ + configsByCategories: function () { + var configs = this.get('configs'), + hiddenCategories = this.get('hiddenCategories'), + groupedConfigs = []; + Ember.keys(configs).forEach(function (site) { + groupedConfigs.push({ + name: site, + configs: this.mapObject(configs[site]), + isVisible: !hiddenCategories.contains(site) + }); + }, this); + return groupedConfigs; }.property('configs.@each'), /** http://git-wip-us.apache.org/repos/asf/ambari/blob/a41ef2a1/contrib/views/slider/src/main/resources/ui/app/templates/slider_app/configs.hbs ---------------------------------------------------------------------- diff --git a/contrib/views/slider/src/main/resources/ui/app/templates/slider_app/configs.hbs b/contrib/views/slider/src/main/resources/ui/app/templates/slider_app/configs.hbs index cf4a9f1..c3597d7 100644 --- a/contrib/views/slider/src/main/resources/ui/app/templates/slider_app/configs.hbs +++ b/contrib/views/slider/src/main/resources/ui/app/templates/slider_app/configs.hbs @@ -16,65 +16,27 @@ * limitations under the License. }} -{{#if model.otherConfigs}} - {{#bs-panel heading="General" collapsible=true}} - <table class="table table-striped table-bordered table-condensed"> - <thead> - <tr> - <th class="col-sm-4">{{t common.name}}</th> - <th>{{t common.value}}</th> - </tr> - </thead> - <tbody> - {{#each model.otherConfigs}} - <tr> - <td>{{key}}</td> - <td>{{formatWordBreak value}}</td> - </tr> - {{/each}} - </tbody> - </table> - {{/bs-panel}} -{{/if}} - -{{#if model.globals}} - {{#bs-panel heading="Global" collapsible=true}} - <table class="table table-striped table-bordered table-condensed"> - <thead> - <tr> - <th class="col-sm-4">{{t common.name}}</th> - <th>{{t common.value}}</th> - </tr> - </thead> - <tbody> - {{#each model.globals}} - <tr> - <td>{{key}}</td> - <td>{{formatWordBreak value}}</td> - </tr> - {{/each}} - </tbody> - </table> - {{/bs-panel}} -{{/if}} - -{{#if model.hbaseSite}} - {{#bs-panel heading="HBase-Site" collapsible=true}} - <table class="table table-striped table-bordered table-condensed"> - <thead> - <tr> - <th class="col-sm-4">{{t common.name}}</th> - <th>{{t common.value}}</th> - </tr> - </thead> - <tbody> - {{#each model.hbaseSite}} - <tr> - <td>{{key}}</td> - <td>{{formatWordBreak value}}</td> - </tr> - {{/each}} - </tbody> - </table> - {{/bs-panel}} +{{#if model.configsByCategories.length}} + {{#each configCategory in configsByCategories}} + {{#if configCategory.isVisible}} + {{#bs-panel heading=configCategory.name collapsible=true}} + <table class="table table-striped table-bordered table-condensed"> + <thead> + <tr> + <th class="col-sm-4">{{t common.name}}</th> + <th>{{t common.value}}</th> + </tr> + </thead> + <tbody> + {{#each configCategory.configs}} + <tr> + <td>{{key}}</td> + <td>{{formatWordBreak value}}</td> + </tr> + {{/each}} + </tbody> + </table> + {{/bs-panel}} + {{/if}} + {{/each}} {{/if}} \ No newline at end of file