[
https://issues.apache.org/jira/browse/STORM-1199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15062976#comment-15062976
]
ASF GitHub Bot commented on STORM-1199:
---------------------------------------
Github user roshannaik commented on a diff in the pull request:
https://github.com/apache/storm/pull/936#discussion_r47970930
--- Diff:
external/storm-hdfs/src/main/java/org/apache/storm/hdfs/common/HdfsUtils.java
---
@@ -0,0 +1,101 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ * <p/>
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * <p/>
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.storm.hdfs.common;
+
+import org.apache.hadoop.fs.FSDataOutputStream;
+import org.apache.hadoop.fs.FileAlreadyExistsException;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.LocatedFileStatus;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.fs.RemoteIterator;
+import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException;
+import org.apache.hadoop.ipc.RemoteException;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+
+public class HdfsUtils {
+ /** list files sorted by modification time that have not been modified
since 'olderThan'. if
+ * 'olderThan' is <= 0 then the filtering is disabled */
+ public static Collection<Path> listFilesByModificationTime(FileSystem
fs, Path directory, long olderThan)
+ throws IOException {
+ ArrayList<LocatedFileStatus> fstats = new ArrayList<>();
+
+ RemoteIterator<LocatedFileStatus> itr = fs.listFiles(directory, false);
+ while( itr.hasNext() ) {
+ LocatedFileStatus fileStatus = itr.next();
+ if(olderThan>0) {
+ if( fileStatus.getModificationTime()<olderThan )
+ fstats.add(fileStatus);
+ }
+ else {
+ fstats.add(fileStatus);
+ }
+ }
+ Collections.sort(fstats, new CmpFilesByModificationTime() );
+
+ ArrayList<Path> result = new ArrayList<>(fstats.size());
+ for (LocatedFileStatus fstat : fstats) {
+ result.add(fstat.getPath());
+ }
+ return result;
+ }
+
+ /**
+ * Returns true if succeeded. False if file already exists. throws if
there was unexpected problem
+ */
+ public static FSDataOutputStream tryCreateFile(FileSystem fs, Path file)
throws IOException {
--- End diff --
right. will update
> Create HDFS Spout
> -----------------
>
> Key: STORM-1199
> URL: https://issues.apache.org/jira/browse/STORM-1199
> Project: Apache Storm
> Issue Type: New Feature
> Reporter: Roshan Naik
> Assignee: Roshan Naik
> Attachments: HDFSSpoutforStorm v2.pdf, HDFSSpoutforStorm.pdf,
> hdfs-spout.1.patch
>
>
> Create an HDFS spout so that Storm can suck in data from files in a HDFS
> directory
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)