[ https://issues.apache.org/jira/browse/AMBARI-24825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16663740#comment-16663740 ]
ASF GitHub Bot commented on AMBARI-24825: ----------------------------------------- oleewere closed pull request #13: AMBARI-24825. Log Feeder: Fix HDFS/S3 outputs URL: https://github.com/apache/ambari-logsearch/pull/13 This is a PR merged from a forked repository. As GitHub hides the original diff on merge, it is displayed below for the sake of provenance: As this is a foreign pull request (from a fork), the diff is supplied below (as it won't show otherwise due to GitHub magic): diff --git a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputFile.java b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputFile.java index 7dae1b8091..7495444cee 100644 --- a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputFile.java +++ b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputFile.java @@ -98,7 +98,7 @@ public void write(Map<String, Object> jsonObj, InputFileMarker inputMarker) thro String outStr = null; CSVPrinter csvPrinter = null; try { - if (codec.equals("csv")) { + if ("csv".equals(codec)) { csvPrinter = new CSVPrinter(outWriter, CSVFormat.RFC4180); //TODO: } else { diff --git a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputHDFSFile.java b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputHDFSFile.java index ed93aa41c2..93a264366c 100644 --- a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputHDFSFile.java +++ b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputHDFSFile.java @@ -39,6 +39,7 @@ import java.util.Date; import java.util.HashMap; import java.util.Iterator; +import java.util.Map; import java.util.concurrent.ConcurrentLinkedQueue; /** @@ -114,6 +115,12 @@ public synchronized void write(String block, InputFileMarker inputMarker) throws } } + @Override + public void write(Map<String, Object> jsonObj, InputFileMarker inputMarker) throws Exception { + String block = LogFeederUtil.getGson().toJson(jsonObj); + write(block, inputMarker); + } + @Override public String getShortDescription() { diff --git a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputS3File.java b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputS3File.java index 7d7e6afbe4..a2f6b08e67 100644 --- a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputS3File.java +++ b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/OutputS3File.java @@ -192,12 +192,22 @@ private synchronized void writeGlobalConfig(S3OutputConfiguration s3OutputConfig */ @Override public void write(String block, InputFileMarker inputMarker) { + createLogSpoolerIfRequired(inputMarker); + logSpooler.add(block); + } + + @Override + public void write(Map<String, Object> jsonObj, InputFileMarker inputMarker) throws Exception { + String block = LogFeederUtil.getGson().toJson(jsonObj); + write(block, inputMarker); + } + + private void createLogSpoolerIfRequired(InputFileMarker inputMarker) { if (logSpooler == null) { if (inputMarker.getInput().getClass().isAssignableFrom(InputFile.class)) { InputFile input = (InputFile) inputMarker.getInput(); logSpooler = createSpooler(input.getFilePath()); s3Uploader = createUploader(input.getInputDescriptor().getType()); - logSpooler.add(block); } else { logger.error("Cannot write from non local file..."); } @@ -261,4 +271,9 @@ public void close() { public void handleRollover(File rolloverFile) { s3Uploader.addFileForUpload(rolloverFile.getAbsolutePath()); } + + @Override + public String getShortDescription() { + return "output:destination=s3,bucket=" + s3OutputConfiguration.getS3BucketName(); + } } diff --git a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/spool/LogSpooler.java b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/spool/LogSpooler.java index 7d7d111250..82a3f1b0fb 100644 --- a/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/spool/LogSpooler.java +++ b/ambari-logsearch-logfeeder/src/main/java/org/apache/ambari/logfeeder/output/spool/LogSpooler.java @@ -45,8 +45,9 @@ public class LogSpooler { private static final Logger logger = LogManager.getLogger(LogSpooler.class); - public static final long TIME_BASED_ROLLOVER_DISABLED_THRESHOLD = 0; - static final String fileDateFormat = "yyyy-MM-dd-HH-mm-ss"; + + private static final String fileDateFormat = "yyyy-MM-dd-HH-mm-ss"; + private static final long TIME_BASED_ROLLOVER_DISABLED_THRESHOLD = 0; private String spoolDirectory; private String sourceFileNamePrefix; diff --git a/ambari-logsearch-logfeeder/src/main/scripts/logfeeder.sh b/ambari-logsearch-logfeeder/src/main/scripts/logfeeder.sh index e1257687c9..eaf00e130d 100755 --- a/ambari-logsearch-logfeeder/src/main/scripts/logfeeder.sh +++ b/ambari-logsearch-logfeeder/src/main/scripts/logfeeder.sh @@ -168,7 +168,7 @@ function start() { LOGFEEDER_DEBUG_PORT=${LOGFEEDER_DEBUG_PORT:-"5006"} if [ "$LOGFEEDER_DEBUG" = "true" ]; then - LOGFEEDER_JAVA_OPTS="$LOGFEEDER_JAVA_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,address=$LOGFEEDER_DEBUG_PORT,server=y,suspend=$LOGFEEDER_DEBUG_SUSPEND " + LOGFEEDER_JAVA_OPTS="$LOGFEEDER_JAVA_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,address=*:$LOGFEEDER_DEBUG_PORT,server=y,suspend=$LOGFEEDER_DEBUG_SUSPEND " fi if [ "$LOGFEEDER_SSL" = "true" ]; then diff --git a/ambari-logsearch-server/src/main/scripts/logsearch.sh b/ambari-logsearch-server/src/main/scripts/logsearch.sh index b941e25833..1b64832995 100755 --- a/ambari-logsearch-server/src/main/scripts/logsearch.sh +++ b/ambari-logsearch-server/src/main/scripts/logsearch.sh @@ -149,7 +149,7 @@ function start() { LOGSEARCH_DEBUG_PORT=${LOGSEARCH_DEBUG_PORT:-"5005"} if [ "$LOGSEARCH_DEBUG" = "true" ]; then - LOGSEARCH_JAVA_OPTS="$LOGSEARCH_JAVA_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,address=$LOGSEARCH_DEBUG_PORT,server=y,suspend=$LOGSEARCH_DEBUG_SUSPEND " + LOGSEARCH_JAVA_OPTS="$LOGSEARCH_JAVA_OPTS -Xdebug -Xrunjdwp:transport=dt_socket,address=*:$LOGSEARCH_DEBUG_PORT,server=y,suspend=$LOGSEARCH_DEBUG_SUSPEND " fi if [ "$LOGSEARCH_SSL" = "true" ]; then diff --git a/docker/cloud-docker-compose.yml b/docker/cloud-docker-compose.yml new file mode 100644 index 0000000000..3a9ec05165 --- /dev/null +++ b/docker/cloud-docker-compose.yml @@ -0,0 +1,129 @@ +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License +version: '3.3' +services: + zookeeper: + image: zookeeper:${ZOOKEEPER_VERSION:-3.4.10} + restart: always + hostname: zookeeper + networks: + - logsearch-network + ports: + - 2181:2181 + environment: + ZOO_MY_ID: 1 + ZOO_SERVERS: server.1=zookeeper:2888:3888 + solr: + image: solr:${SOLR_VERSION:-7.5.0} + restart: always + hostname: solr + ports: + - "8983:8983" + networks: + - logsearch-network + env_file: + - Profile + entrypoint: + - docker-entrypoint.sh + - solr + - start + - "-f" + - "-c" + - "-z" + - ${ZOOKEEPER_CONNECTION_STRING} + logsearch: + image: ambari-logsearch:v1.0 + restart: always + hostname: logsearch.apache.org + labels: + logfeeder.log.type: "logsearch_server" + networks: + - logsearch-network + env_file: + - Profile + ports: + - 61888:61888 + - 4444:4444 + - 5005:5005 + environment: + COMPONENT: logsearch + COMPONENT_LOG: logsearch + ZK_CONNECT_STRING: ${ZOOKEEPER_CONNECTION_STRING} + DISPLAY: $DISPLAY_MAC + volumes: + - $AMBARI_LOCATION:/root/ambari + - $AMBARI_LOCATION/ambari-logsearch/docker/test-logs:/root/test-logs + - $AMBARI_LOCATION/ambari-logsearch/docker/test-config:/root/test-config + logfeeder: + image: ambari-logsearch:v1.0 + restart: always + hostname: logfeeder.apache.org + privileged: true + labels: + logfeeder.log.type: "logfeeder" + networks: + - logsearch-network + env_file: + - Profile + ports: + - 5006:5006 + environment: + COMPONENT: logfeeder + COMPONENT_LOG: logfeeder + ZK_CONNECT_STRING: ${ZOOKEEPER_CONNECTION_STRING} + volumes: + - $AMBARI_LOCATION:/root/ambari + - $AMBARI_LOCATION/ambari-logsearch/docker/test-logs:/root/test-logs + - $AMBARI_LOCATION/ambari-logsearch/docker/test-config:/root/test-config + - /var/run/docker.sock:/var/run/docker.sock + - /usr/local/bin/docker:/usr/local/bin/docker + - /var/lib/docker:/var/lib/docker + fakes3: + image: localstack/localstack + hostname: fakes3 + ports: + - "4569:4569" + environment: + - SERVICES=s3:4569 + - DEBUG=s3 + networks: + logsearch-network: + aliases: + - testbucket.fakes3 + env_file: + - Profile + namenode: + image: flokkr/hadoop-hdfs-namenode:${HADOOP_VERSION:-3.0.0} + hostname: namenode + ports: + - 9870:9870 + - 9000:9000 + env_file: + - Profile + environment: + ENSURE_NAMENODE_DIR: "/tmp/hadoop-hdfs/dfs/name" + networks: + - logsearch-network + datanode: + image: flokkr/hadoop-hdfs-datanode:${HADOOP_VERSION:-3.0.0} + links: + - namenode + env_file: + - Profile + networks: + - logsearch-network +networks: + logsearch-network: + driver: bridge diff --git a/docker/test-config/logfeeder/shipper-conf/output.config.json b/docker/test-config/logfeeder/shipper-conf/output.config.json index a85b4a4fe9..62950d1e39 100644 --- a/docker/test-config/logfeeder/shipper-conf/output.config.json +++ b/docker/test-config/logfeeder/shipper-conf/output.config.json @@ -31,6 +31,44 @@ ] } } + }, + { + "comment": "S3 file output", + "is_enabled": "true", + "destination": "s3_file", + "type": "s3", + "s3_access_key" : "accessKey", + "s3_secret_key" : "secretKey", + "s3_bucket" : "docker-logsearch", + "s3_endpoint" : "http://fakes3:4569", + "s3_log_dir" : "/tmp", + "skip_logtime": "true", + "conditions": { + "fields": { + "rowtype": [ + "s3" + ] + } + } + }, + { + "comment": "HDFS file output", + "is_enabled": "true", + "destination": "hdfs", + "type": "hdfs", + "file_name_prefix":"service-logs-", + "hdfs_out_dir": "/logfeeder/$HOST/service", + "hdfs_host": "namenode", + "hdfs_port": "9000", + "rollover_sec":"10", + "skip_logtime": "true", + "conditions": { + "fields": { + "rowtype": [ + "hdfs" + ] + } + } } ] } ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Log Feeder: Fix HDFS/S3 outputs > ------------------------------- > > Key: AMBARI-24825 > URL: https://issues.apache.org/jira/browse/AMBARI-24825 > Project: Ambari > Issue Type: Bug > Components: ambari-logsearch > Affects Versions: 2.7.0 > Reporter: Olivér Szabó > Assignee: Olivér Szabó > Priority: Major > Labels: pull-request-available > Fix For: 2.8.0 > > > Currently hdfs / s3 outputs are not working properly, they just open an empty > file and nothing happens -- This message was sent by Atlassian JIRA (v7.6.3#76005)