yes i have already started the bundle, but if i search from the web interface or via a command line like: curl -X POST -d "name=roma*&limit=10&offset=0" http://localhost:8080/entityhub/site/<SITE-NAME>/find
i have the error i pasted. Any suggestions? maybe i miss some configuration step? 2012/3/28 Michel Benevento <[email protected]> > Have you started your installed bundle in the admin console? Click the > little triangle next to it so it becomes a square and the status message > updates. > > Michel > > > On 28 mrt. 2012, at 17:02, seralf wrote: > > > Hi i'm trying to use the KeywordLinking as Rupert suggested me earlier. > > I've done the solr indexes as in the tutorial and they seems to be ok (i > > looked inside them with Luke), i've copied them in ROOT/sling/dataset, > and > > then installed the generated bundle via the console. > > > > Now i have a strange error: seems like stanbol is not actually load my > > indexes, or for some reason it has not activated the yard > > > > java.lang.IllegalStateException: Unable to initialize the Cache with Yard > >> <SITE-NAME> Index! This is usually caused by Errors while reading the > Cache > >> Configuration from the Yard. > >> at > >> > org.apache.stanbol.entityhub.core.site.CacheImpl.getCacheYard(CacheImpl.java:214) > >> at > >> > org.apache.stanbol.entityhub.core.site.CacheImpl.findRepresentation(CacheImpl.java:331) > >> ... > >> Caused by: org.apache.stanbol.entityhub.servicesapi.yard.YardException: > >> The SolrIndex '<SITE-NAME>' for SolrYard '<SITE-NAME> Index' is > currently > >> not active! > >> ... > >> > > > > does anyone has suggestion on this? > > > > i have two other related questions: > > 1) how can i start stanbol with specific config activated? > > 2) is there any way to manage the deploy/activation via some kind of rest > > interface? (for example curl? it could be helpful for doing some > > automatization... ) > > > > thanks in advance, > > Alfredo > > > > > > > > > > 2012/3/22 seralf <[email protected]> > > > >> Thanks very much Rupert, you help me a lot in clarify my ideas :-) > >> > >> i think i'll try to follow your suggestion, and try to use my thesaurus > >> with the workflow option 2) > >> i already use solr either, so it's probably the best choice for my > needs, > >> indeed > >> > >> on the other hand i'm still interested on give a try on opennlp italian > >> model construction, but i can to my experiments externally, as i correct > >> understand. > >> > >> thanks very much, i'll try to make some progress > >> Alfredo > >> > >> > >> > >> 2012/3/22 Rupert Westenthaler <[email protected]> > >> > >>> Hi Alfredo > >>> > >>> On 22.03.2012, at 12:24, seralf wrote: > >>> > >>>> Hi i'm new to stambol, i'm reading the documentation and examples, and > >>> i'd > >>>> like to start some testing with it on italian language, if it's > >>> possible. > >>>> > >>>> Could someone give me some hint regarding the steps to try to costruct > >>> my > >>>> model (Italian) and configure it inside the platform? I suppose it's > >>>> possible and it should be not very far to the steps taken for > construct > >>>> -let's say- the Spanish integration. > >>>> What i need to do? I know it could sound a very generic question, but > >>> it's > >>>> not so clear from the documentation, so i need help. > >>>> For my test i would like to be able to use a text corpora from the > >>> database > >>>> of a client, and a skos thesaurus from the same domain. > >>>> > >>>> thanks in advance for every help (suggestions, code examples, ideas, > >>> etc) > >>>> > >>> > >>> In principle there are two different workflows how to extract Entities > >>> form Text > >>> > >>> (1) NamedEntityExtraction (NER) [3] => NamedEntityLinking [4] > >>> (2) KeywordLinking [5] > >>> > >>> > >>> (1) requires a OpenNLP [1] NER model for the language of your > documents. > >>> However currently there are no models for the italian language > distributed > >>> by OpenNLP. This would require you to build your own models. For more > >>> information on how to do that please see the documentation of OpenNLP > [1]. > >>> As soon as you have such models you need only copy them into the > >>> {stanbol-workingdir}/sling/datafiles folder. If they follow the naming > >>> scheme used by OpenNLP ("{lang}-ner-{type}.bin" e.g. > "it-ner.location.bin" > >>> for the model that detects locations for italian) Stanbol will pick > them up > >>> automatically. > >>> > >>> (2) directly matches words of the text with labels of entities within > the > >>> controlled vocabulary. This process can be improved by Natural Langauge > >>> Processing (e.g. Part-of-Speech tagging) but this is not a requirement. > >>> Typically this works fine for datasets that contain named entities > such as > >>> concepts of an thesaurus; contacts of an company, projects, products … > It > >>> does not work well with datasets that contains entities with labels > that > >>> are also used as common words in the given language as this will > result in > >>> a lot of false positives. > >>> > >>> Based on the information you provided on you use case I suggest that > (2) > >>> should work just fine for you. This user scenario [2] should provide > you > >>> will all the needed information on how to configure Stanbol for your > use > >>> case. > >>> > >>> I hope this helps. If you have any further questions feel free to ask > >>> > >>> best > >>> Rupert Westenthaler > >>> > >>> [1] http://opennlp.apache.org/ > >>> [2] > http://incubator.apache.org/stanbol/docs/trunk/customvocabulary.html > >>> > >>> [3] > >>> > http://incubator.apache.org/stanbol/docs/trunk/enhancer/engines/namedentityextractionengine.html > >>> [4] > >>> > http://incubator.apache.org/stanbol/docs/trunk/enhancer/engines/namedentitytaggingengine.html > >>> [5] > >>> > http://incubator.apache.org/stanbol/docs/trunk/enhancer/engines/keywordlinkingengine.html > >>> > >>>> cheers, > >>>> Alfredo Serafini > >>> > >>> > >> > >
