of the same document.
The framework is smart enough to not hand a document to a connector if
it hasn't changed (according to how the connector computes the
connector-specific output version string).
Karl
On Tue, Sep 10, 2019 at 11:00 AM Julien Massiera
<mailto:julien.massi...@francelabs.
Hi,
I would like to have an explanation about the behavior of a job when
several outputs are configured. My main question is : for each output,
how is the docs ingestion managed ? More precisely, are the ingest
processes synchronized or not ? (in other words, is the ingestion of the
next
Hi all,
I was doing some tests with the Web connector, and after several tries
with different configurations of my job to crawl a session based
website, I noticed that one configuration was not working. So I debugged
the job and noticed that the connector was using a wrong session cookie.
In
s still around and the metadata can be shared?)
Thanks in advance,
Karl
--
Julien MASSIERA
Directeur développement produit
France Labs – Les experts du Search
Retrouvez-nous à l’Enterprise Search & Discovery Summit à Washington DC
www.francelabs.com
Hi Karl,
No problem, it is what I would have proposed anyway !
Julien
Le 23/12/2017 à 16:27, Karl Wright a écrit :
Do you mind if I include this in the SharePoint connector documentation?
Thanks,
Karl
On Sat, Dec 23, 2017 at 10:13 AM, Julien Massiera
<julien.massi...@francelabs.
rks but
I would like to avoid an admin user to crawl my site.
Thanks for your help !
Julien Massiera
---
L'absence de virus dans ce courrier électronique a été vérifiée par le logiciel
antivirus Avast.
https://www.avast.com/antivirus
Hi MCF community,
I recently switched from MCF 2.5 to MCF 2.8.1 and I am a little bit
confused with the agent logs.
First of all I noticed that MCF is now using log4j2 and the
documentation is not up to date on this point (the old logging.ini
format is mentioned) :
On Tue, Sep 12, 2017 at 5:19 AM, Julien Massiera
<julien.massi...@francelabs.com
<mailto:julien.massi...@francelabs.com>> wrote:
Hi everybody,
I usually download MCF through the http://archive.apache.org
website and I noticed that the 2.8.1 version is dated
and
I can safely download MCF from this website or is it better to take it
from the MCF website ?
Regards
--
Julien MASSIERA
Expert en technologies de recherche
France Labs – Les experts du Search
Vainqueur du challenge Internal Search de EY à Viva Technologies 2016
www.francelabs.com
Tel : +33
right a écrit :
> Hi Julien,
>
> How are you starting the job? If you use "Start minimal", deletion would not
> take place. If your job is a continuous one, this is also the case.
>
> Thanks,
> Karl
>
> On Wed, Apr 26, 2017 at 9:52 AM, <julien.massi...@francelabs.com> wrote:
> Hi the MCF community,
>
> I am using MCF 2.6 with the JDBC connector to crawl an Oracle Database and
> index the data into a Solr server, and it works very well. However, when I
> perform a delta re-crawl, the new IDs are correctly retrieved from the
> Database but those who have been deleted are not "detected" by the connector
> and thus, are still present in my Solr index.
> I would like to know if normally it should work and that I maybe have missed
> something in the configuration of the job, or if this is not implemented ?
> The only way I found to solve this issue is to reset the seeding of the job,
> but it is very time and resource consuming.
>
> Best regards,
> Julien Massiera
Links:
--
[1] http://doc.id
d not
> take place. If your job is a continuous one, this is also the case.
>
> Thanks,
> Karl
>
> On Wed, Apr 26, 2017 at 9:52 AM, <julien.massi...@francelabs.com> wrote:
> Hi the MCF community,
>
> I am using MCF 2.6 with the JDBC connector to crawl an Oracle Database and
> index the data into a Solr server, and it works very well. However, when I
> perform a delta re-crawl, the new IDs are correctly retrieved from the
> Database but those who have been deleted are not "detected" by the connector
> and thus, are still present in my Solr index.
> I would like to know if normally it should work and that I maybe have missed
> something in the configuration of the job, or if this is not implemented ?
> The only way I found to solve this issue is to reset the seeding of the job,
> but it is very time and resource consuming.
>
> Best regards,
> Julien Massiera
> Hi Julien,
>
> How are you starting the job? If you use "Start minimal", deletion would not
> take place. If your job is a continuous one, this is also the case.
>
> Thanks,
> Karl
>
> On Wed, Apr 26, 2017 at 9:52 AM, <julien.massi...@francelabs.com> wrote:
> Hi the MCF community,
>
> I am using MCF 2.6 with the JDBC connector to crawl an Oracle Database and
> index the data into a Solr server, and it works very well. However, when I
> perform a delta re-crawl, the new IDs are correctly retrieved from the
> Database but those who have been deleted are not "detected" by the connector
> and thus, are still present in my Solr index.
> I would like to know if normally it should work and that I maybe have missed
> something in the configuration of the job, or if this is not implemented ?
> The only way I found to solve this issue is to reset the seeding of the job,
> but it is very time and resource consuming.
>
> Best regards,
> Julien Massiera
he seeding of the
job, but it is very time and resource consuming.
Best regards,
Julien Massiera
eally want to avoid the job to stop if a
lock file is encountered and not filtered.
Thanks
--
Julien MASSIERA
Expert en technologies de recherche
France Labs – Les experts du Search
Vainqueur du challenge Internal Search de EY à Viva Technologies 2016
www.francelabs.com
Tel : +33 (0) 663778847
nnections" while the "GET"
or "PUT" are available (ref method "executeDeleteCommand" in the class
"org/apache/manifoldcf/crawler/system/ManifoldCF").
Is there a specific reason for this ? If not, is it part of your plan to
implement it any time soon ?
Than
re.
Am I missing something ?
Julien
On 19/05/2016 21:14, Karl Wright wrote:
This sounds like it would work.
Karl
Sent from my Windows Phone
From: Julien Massiera
Sent: 5/19/2016 12:44 PM
To:user@manifoldcf.apache.org
Subject: Multiple output documents from one input document in
transformation co
the emails and send them for Solr ingestion through the activities object.
Is my approach correct ? or do I need to consider another solution ?
Thanks for your help.
Julien Massiera
17 matches
Mail list logo