4. Write an external program that fetches the file, fetches the metadata, combines them, and send them to Solr.
wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) > On Jul 9, 2017, at 3:03 PM, Giovanni De Stefano <giova...@servisoft.be> wrote: > > Hello all, > > I have to index (and search) data organised as followed: many files on the > filesystem and each file has extra metadata stored on a DB (the DB table has > a reference to the file path). > > I think I should have 1 Solr document per file with fields coming from both > the DB (through DIH) and from Tika. > > How do you suggest to proceed? > > 1. index into different cores and search across cores (I would rather not do > that but I would be able to reuse “standard” importers) > 2. extend the DIH (which one?) > 3. implement a custom import handler > > How would you do it? > > Developing in Java is not a problem, I would just need some ideas on where to > start (I have been away from Solr for many years…). > > Thanks! > G.