Sorry for answering delay.. Yes, this is not pure "CEP", but quite close
for it or many similar "functionalities".
My case is not so easy, because i dont' want to compare against original
time schedule of route.
I want to compare how close (ITS) system has estimated arrival time to
bus stop.
That means i have to read more (LOG C) logs (and do little calculation)
to determine the estimated arrival time.
And then it is checked how much a difference (error) between bus real
and the system's estimated arrival time..
In practic the situation can be little more complex..
---
Esa Heikkinen
29.4.2016, 19:38, Michael Segel kirjoitti:
If you’re getting the logs, then it really isn’t CEP unless you
consider the event to be the log from the bus.
This doesn’t sound like there is a time constraint.
Your bus schedule is fairly fixed and changes infrequently.
Your bus stops are relatively fixed points. (Within a couple of meters)
So then you’re taking bus A who is scheduled to drive route 123 and
you want to compare their nearest location to the bus stop at time T
and see how close it is to the scheduled route.
Or am I missing something?
-Mike
On Apr 29, 2016, at 3:54 AM, Esa Heikkinen
<esa.heikki...@student.tut.fi <mailto:esa.heikki...@student.tut.fi>>
wrote:
Hi
I try to explain my case ..
Situation is not so simple in my logs and solution. There also many
types of logs and there are from many sources.
They are as csv-format and header line includes names of the columns.
This is simplified description of input logs.
LOG A's: bus coordinate logs (every bus has own log):
- timestamp
- bus number
- coordinates
LOG B: bus login/logout (to/from line) message log:
- timestamp
- bus number
- line number
LOG C: log from central computers:
- timestamp
- bus number
- bus stop number
- estimated arrival time to bus stop
LOG A are updated every 30 seconds (i have also another system by 1
seconds interval). LOG B are updated when bus starts from terminal
bus stop and arrives to final bus stop in a line. LOG C is updated
when central computer sends new arrival time estimation to bus stop.
I also need metadata for logs (and analyzer). For example coordinates
for bus stop areas.
Main purpose of analyzing is to check an accuracy (error) of the
estimated arrival time to bus stops.
Because there are many buses and lines, it is too time-comsuming to
check all of them. So i check only specific lines with specific bus
stops. There are many buses (logged to lines) coming to one bus stop
and i am interested about only certain bus.
To do that, i have to read log partly not in time order (upstream) by
sequence:
1. From LOG C is searched bus number
2. From LOG A is searched when the bus has leaved from terminal bus stop
3. From LOG B is searched when bus has sent a login to the line
4. From LOG A is searched when the bus has entered to bus stop
5. From LOG C is searched a last estimated arrival time to the bus
stop and calculates error between real and estimated value
In my understanding (almost) all log file analyzers reads all data
(lines) in time order from log files. My need is only for specific
part of log (lines). To achieve that, my solution is to read logs in
an arbitrary order (with given time window).
I know this solution is not suitable for all cases (for example for
very fast analyzing and very big data). This solution is suitable for
very complex (targeted) analyzing. It can be too slow and
memory-consuming, but well done pre-processing of log data can help a
lot.
---
Esa Heikkinen
28.4.2016, 14:44, Michael Segel kirjoitti:
I don’t.
I believe that there have been a couple of hack-a-thons like one
done in Chicago a few years back using public transportation data.
The first question is what sort of data do you get from the city?
I mean it could be as simple as time_stamp, bus_id, route and GPS
(x,y). Or they could provide more information. Like last stop,
distance to next stop, avg current velocity…
Then there is the frequency of the updates. Every second? Every 3
seconds? 5 or 6 seconds…
This will determine how much work you have to do.
Maybe they provide the routes of the busses via a different API call
since its relatively static.
This will drive your solution more than the underlying technology.
Oh and whileI focused on bus, there are also rail and other modes of
public transportation like light rail, trains, etc …
HTH
-Mike
On Apr 28, 2016, at 4:10 AM, Esa Heikkinen
<esa.heikki...@student.tut.fi
<mailto:esa.heikki...@student.tut.fi>> wrote:
Do you know any good examples how to use Spark streaming in
tracking public transportation systems ?
Or Storm or some other tool example ?
Regards
Esa Heikkinen
28.4.2016, 3:16, Michael Segel kirjoitti:
Uhm…
I think you need to clarify a couple of things…
First there is this thing called analog signal processing…. Is
that continuous enough for you?
But more to the point, Spark Streaming does micro batching so if
you’re processing a continuous stream of tick data, you will have
more than 50K of tics per second while there are markets open and
trading. Even at 50K a second, that would mean 1 every .02 ms or
50 ticks a ms.
And you don’t want to wait until you have a batch to start
processing, but you want to process when the data hits the queue
and pull it from the queue as quickly as possible.
Spark streaming will be able to pull batches in as little as
500ms. So if you pull a batch at t0 and immediately have a tick in
your queue, you won’t process that data until t0+500ms. And said
batch would contain 25,000 entries.
Depending on what you are doing… that 500ms delay can be enough to
be fatal to your trading process.
If you don’t like stock data, there are other examples mainly when
pulling data from real time embedded systems.
If you go back and read what I said, if your data flow is >> (much
slower) than 500ms, and / or the time to process is >> 500ms (
much longer ) you could use spark streaming. If not… and there
are applications which require that type of speed… then you
shouldn’t use spark streaming.
If you do have that constraint, then you can look at systems like
storm/flink/samza / whatever where you have a continuous queue and
listener and no micro batch delays.
Then for each bolt (storm) you can have a spark context for
processing the data. (Depending on what sort of processing you
want to do.)
To put this in perspective… if you’re using spark streaming / akka
/ storm /etc to handle real time requests from the web, 500ms
added delay can be a long time.
Choose the right tool.
For the OP’s problem. Sure Tracking public transportation could be
done using spark streaming. It could also be done using half a
dozen other tools because the rate of data generation is much
slower than 500ms.
HTH
On Apr 27, 2016, at 4:34 PM, Mich Talebzadeh
<mich.talebza...@gmail.com <mailto:mich.talebza...@gmail.com>> wrote:
couple of things.
There is no such thing as Continuous Data Streaming as there is
no such thing as Continuous Availability.
There is such thing as Discrete Data Streaming and High
Availability but they reduce the finite unavailability to
minimum. In terms of business needs a 5 SIGMA is good enough and
acceptable. Even the candles set to a predefined time interval
say 2, 4, 15 seconds overlap. No FX savvy trader makes a sell or
buy decision on the basis of 2 seconds candlestick
The calculation itself in measurements is subject to finite error
as defined by their Confidence Level (CL) using Standard
Deviation function.
OK so far I have never noticed a tool that requires that details
of granularity. Those stuff from Flink etc is in practical term
is of little value and does not make commercial sense.
Now with regard to your needs, Spark micro batching is perfectly
adequate.
HTH
Dr Mich Talebzadeh
LinkedIn
/https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/
http://talebzadehmich.wordpress.com
On 27 April 2016 at 22:10, Esa Heikkinen
<esa.heikki...@student.tut.fi> wrote:
Hi
Thanks for the answer.
I have developed a log file analyzer for RTPIS (Real Time
Passenger Information System) system, where buses drive lines
and the system try to estimate the arrival times to the bus
stops. There are many different log files (and events) and
analyzing situation can be very complex. Also spatial data
can be included to the log data.
The analyzer also has a query (or analyzing) language, which
describes a expected behavior. This can be a requirement of
system. Analyzer can be think to be also a test oracle.
I have published a paper (SPLST'15 conference) about my
analyzer and its language. The paper is maybe too technical,
but it is found:
http://ceur-ws.org/Vol-1525/paper-19.pdf
I do not know yet where it belongs. I think it can be some
"CEP with delays". Or do you know better ?
My analyzer can also do little bit more complex and
time-consuming analyzings? There are no a need for real time.
And it is possible to do "CEP with delays" reasonably some
existing analyzer (for example Spark) ?
Regards
PhD student at Tampere University of Technology, Finland,
www.tut.fi
Esa Heikkinen
27.4.2016, 15:51, Michael Segel kirjoitti:
Spark and CEP? It depends…
Ok, I know that’s not the answer you want to hear, but its a
bit more complicated…
If you consider Spark Streaming, you have some issues.
Spark Streaming isn’t a Real Time solution because it is a
micro batch solution. The smallest Window is 500ms. This
means that if your compute time is >> 500ms and/or your
event flow is >> 500ms this could work.
(e.g. 'real time' image processing on a system that is
capturing 60FPS because the processing time is >> 500ms. )
So Spark Streaming wouldn’t be the best solution….
However, you can combine spark with other technologies like
Storm, Akka, etc .. where you have continuous streaming.
So you could instantiate a spark context per worker in storm…
I think if there are no class collisions between Akka and
Spark, you could use Akka, which may have a better potential
for communication between workers.
So here you can handle CEP events.
HTH
On Apr 27, 2016, at 7:03 AM, Mich Talebzadeh
<mich.talebza...@gmail.com> wrote:
please see my other reply
Dr Mich Talebzadeh
LinkedIn
/https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/
http://talebzadehmich.wordpress.com
On 27 April 2016 at 10:40, Esa Heikkinen
<esa.heikki...@student.tut.fi> wrote:
Hi
I have followed with interest the discussion about CEP
and Spark. It is quite close to my research, which is a
complex analyzing for log files and "history" data
(not actually for real time streams).
I have few questions:
1) Is CEP only for (real time) stream data and not for
"history" data?
2) Is it possible to search "backward" (upstream) by
CEP with given time window? If a start time of the time
window is earlier than the current stream time.
3) Do you know any good tools or softwares for "CEP's"
using for log data ?
4) Do you know any good (scientific) papers i should
read about CEP ?
Regards
PhD student at Tampere University of Technology,
Finland, www.tut.fi
Esa Heikkinen
---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org
The opinions expressed here are mine, while they may reflect
a cognitive thought, that is purely accidental.
Use at your own risk.
Michael Segel
michael_segel (AT) hotmail.com <http://hotmail.com/>
The opinions expressed here are mine, while they may reflect a
cognitive thought, that is purely accidental.
Use at your own risk.
Michael Segel
michael_segel (AT) hotmail.com <http://hotmail.com/>