I am trying vectorize text data for a Naive Bayes classifier that will be trained in Hadoop and then the corresponding model will be deployed in a Java app. My basic approach is to tokenize a string of text data using Lucene and then encode each token using a StaticWordValueEncoder here are the relevant code snippets
private static FeatureVectorEncoder memoEncoder = new StaticWordValueEncoder("memo"); Vector v = new RandomAccessSparseVector(FEATURES); StringReader reader = new StringReader(text); StandardTokenizer source = new StandardTokenizer(Version.LUCENE_46, reader); ShingleFilter sf = new ShingleFilter(source); sf.setOutputUnigrams(true); CharTermAttribute charTermAttribute = sf.addAttribute(CharTermAttribute.class); sf.reset(); while (sf.incrementToken()) { memoEncoder.addToVector(charTermAttribute.toString(), 1,v); } In the Mahout in Action book I got the impression that the term "memo" will seed the random number generator and I wanted to confirm that means I will have consistency if I deploy this vectorizer in both my Hadoop environment as well as my Java app. In particular, I am fixing the vector size to be of length FEATURES and I am using "memo" as the name of my encoder. Will those two things guarantee consistency of my text vectorization?