The link to hbase 1.2.3 is also dead. So I downloaded the last stable
version, which is 1.2.4. I hope it is fully compatible.

2016-12-22 10:27 GMT+01:00 Bruno LEBON <b.le...@redfakir.fr>:

> Ok thanks for your answer, i'll keep an eye for the announcement. At this
> address, right? http://actionml.com/docs/pio_versions
>
> Also I am following the tuto here : http://actionml.com/docs/
> small_ha_cluster
> and there is a dead link at:
> Download Services On All Hosts - 1 Download -> the link for Spark's
> archive is dead.
> It is:
> http://www.us.apache.org/dist/spark/spark-1.6.2/spark-1.6.3-
> bin-hadoop2.6.tgz
> I think it should be:
> http://www.us.apache.org/dist/spark/spark-1.6.3/spark-1.6.3-
> bin-hadoop2.6.tgz
>
> Thanks for your time and help, much appreciated!
> Bruno
>
> 2016-12-18 18:58 GMT+01:00 Pat Ferrel <p...@occamsmachete.com>:
>
>> There was a bug in this feature in the Apache PIO version that has been
>> fixed in the SNAPSHOT. We will do a source tag to fix it before the next
>> release . The page you reference is being changed now to advise an Apache
>> PIO install as the root source of the project going forward. Keep and eye
>> out here for the announcement.
>>
>>
>> On Dec 7, 2016, at 9:01 AM, Bruno LEBON <b.le...@redfakir.fr> wrote:
>>
>> Hi,
>>
>> I am following the instructions here: http://actionml.com/docs/pio_v
>> ersions
>> to add the possibility to delete events that are too old to template
>> other than Universal recommender. I want to add it to ecommerce
>> recommendation.
>>
>> However it doesnt seem to work.
>>
>> I have the PredicitonIO 097-aml and the template ecommerce-recommendation
>> from apache-incubator (0.4.0)
>>
>> I have the following 3  errors:
>> [ERROR] [Console$] [error] /home/aml/incubator-prediction
>> io-template-ecom-recommender/src/main/scala/DataSource.scala:21: not
>> found: type SelfCleaningDataSource
>> [ERROR] [Console$] [error] with SelfCleaningDataSource {
>> [ERROR] [Console$] [error]      ^
>> [ERROR] [Console$] [error] /home/aml/incubator-prediction
>> io-template-ecom-recommender/src/main/scala/DataSource.scala:26: value
>> eventWindow is not a member of org.template.ecommercerecommen
>> dation.DataSourceParams
>> [ERROR] [Console$] [error]   override def eventWindow = dsp.eventWindow
>> [ERROR] [Console$] [error]                                  ^
>> [ERROR] [Console$] [error] /home/aml/incubator-prediction
>> io-template-ecom-recommender/src/main/scala/DataSource.scala:31: not
>> found: value cleanPersistedPEvents
>> [ERROR] [Console$] [error]     cleanPersistedPEvents(sc)
>> [ERROR] [Console$] [error]     ^
>> [ERROR] [Console$] [error] three errors found
>>
>>
>>
>> The first is related to the fact that DataSourceParams has no object
>> declared, so I added it in the declaration of the object DataSourceParams
>> but which type should I give to it? Is it an Object? I dont know Scala, so
>> I am not sure how to modify the code to make it work. In engine.json it
>> refers to :
>>   "datasource": {
>>     "params" : {
>>       "name": "some-name",
>>       "appName": "autocleangites22",
>>       "eventNames": ["view"],
>>       "eventWindow": {
>>         "duration": "3650 days",
>>         "removeDuplicates": false,
>>         "compressProperties": false
>>       }
>>     }
>> so the code accesses to eventWindow.duration.
>>
>>
>> The second error says that the interface (that would be called so in Java
>> at least) SelfCleaningDataSource is not found in PredictionIO. When I have
>> a look at the code I find the interface, under some annotation
>> @DeveloperApi:
>> @DeveloperApi
>> trait SelfCleaningDataSource {
>>
>> and here too (for the method generating the third error):
>> @DeveloperApi
>>   def cleanPersistedPEvents(sc: SparkContext): Unit ={
>>
>> which makes me wonder whether there is a special way to compile the
>> source code so that the code under this annotation is to be added?
>>
>> Thanks in advance for your help,
>> Bruno
>>
>>
>>
>

Reply via email to