marc magrans de abril wrote:
Hi!

...I have found a good enough solution, although it only works if the
number of patterns (clusters) is not very big:
def classify(f):
    THERESHOLD=0.1

    patterns={}
    for l in enumerate(f):
        found = False
        for p,c in patterns.items():
            if dist(l,p) < THERESHOLD:
                found=True
                patterns[p] = c +1

        if not found:
            patterns[l] = 1

    return patterns

This algorithm is O(n*np*m^2). Where n is the number of logs, np the
number of patterns, and m is the log length (i.e. m^2 is the distance
cost). So it's way better O(n^2*m^2) and I can run it for some hours
to get back the results.

I wonder if there is a single threaded/process clustering algorithm
than runs in O(n)?

Your original code used the first entry in the remaining logs for each
pattern, but your new code stores the patterns in a dict, which is
unordered, so you might get different results.

But that doesn't matter, because your new code increments the count when
it has found a match, and then continues looking, so it might match and
increment more than once.

Finally, your original code treated it as a match if distance <=
threshold but your new code treats it as a match if distance <
threshold.

patterns = []
for x in logs:
    for index, (pat, count) in enumerate(patterns):
        if dist(pat, x) <= THRESHOLD:
            patterns[index] = pat, count + 1
            break
    else:
        # Didn't break out of the loop, therefore no match.
        patterns.append((x, 1))
--
http://mail.python.org/mailman/listinfo/python-list

Reply via email to