Hi Vladimir,Not really, we do not want to store historical data in cache or may 
be cache it for few hours and then evict it.But if recent data is missing in 
cache then yes we want to cache it.So it would require some custom caching 
logic to decide which data to cache.So seems like storing historical data in 
persistent storage seems reasonable.The only thing is I have to use ignite 
cache as side cache rather than a write through cache to HDFS, because i dont 
think it would be nice idea to store individual key value pair in cache.May be 
can think about HBase bt not directly to HDFS.
Regards,Vij
Sent from Yahoo Mail on Android 
 
  On Thu, 14 Apr, 2016 at 7:26 pm, Vladimir Ozerov<[email protected]> wrote: 
  Hi Vij,
Storing hot recent data in cache, and historical data in persistent store 
sounds like a perfectly reasonable idea. 
If you decide to store historical data in HDFS, then you should be able to 
construct HDFS path from the key because store interface accepts keys to 
store/load data. If this is possible, then I do not see any obvious problems 
with this approach.
On the other hand, do you want this historical data to be cached on access?
Vladimir.
On Thu, Apr 14, 2016 at 3:17 PM, vijayendra bhati <[email protected]> 
wrote:

Thanks Vladimir !!
The drawback with using HDFS as persistent store behind Ignite cache is how we 
will take care of appending single key value pair to HDFS file.Ideally we 
should use some NoSQL store or RDBMS as persistent back up behind Ignite cache 
and then run some scheduled batch to transfer the data to HDFS as it happens in 
normal Lambda Architecture.
Now question comes why we want to use Ignite Cache ? Answer is it gives SQL 
interface that means we can query on any attribute on the fly.Other we could 
have used any other NoSQL. But NoSQL data model is entirely based upon query 
pattern so to bring the flexibility at the time of query we think Ignite cache 
would be better.
For our use case we want to put the latest 2 week data in Ignite cache to meet 
the latency requirements and then for any back date get the data from backend 
persistent storage, for which we are thinking about HDFS.Thats why we were 
thinking if we can make Ignite cache write through cache with HDFS as backed up 
persistent storage it would serve the purpose.
Please let me know whats your view on this. 
Many thanks,Vij 

    On Wednesday, April 13, 2016 8:58 PM, Vladimir Ozerov 
<[email protected]> wrote:
 

 Vij,
No, it doesn't. IGFS serves the very different purpose - it is 
Hadoop-compatible file system. It means that, for example, you can load data to 
IGFS and then query it using Hive. But native Ignite SQL is not applicable here.
Vladimir.
On Wed, Apr 13, 2016 at 3:55 PM, vijayendra bhati <[email protected]> 
wrote:

Thanks Vladimir,
I have not gone through complete documentation but if you could let me know 
does IGFS provide SQL support like Ignite cache does ?
Regards,Vij 

    On Wednesday, April 13, 2016 5:54 PM, Vladimir Ozerov 
<[email protected]> wrote:
 

 Hi Vijayendra,
IGFS is designed to be a distributed file system which could cache data from 
Hadoop file systems. It cannot be used as cache store by design.Ignite doesn't 
have store implementation for HDFS, so you should implement your own if needed. 
Particularly, you should implement org.apache.ignite.cache.store.CacheStore 
interface.
Vladimir.
On Wed, Apr 13, 2016 at 2:38 PM, vijayendra bhati <[email protected]> 
wrote:

Hi,
Can some body please provide me any pointers regarding how I can use Ignite 
Data Grid/ In Memory caching with write through/write behind mode and writing 
to HDFS ?
I know Ignite provides IGFS but its different from what I am looking for.
The other way could be I can use IGFS as my In Memory store but is it the right 
approach ?
Regards,Vijayendra Bhati



   



   

  

Reply via email to