Hi, On UI shows 20,002 documents (on a firts phase show 10,001,and after sometime of process raise to 20,002) . It looks like a hard limit, there is more files on sharepoint with the used criteria
Jorge Alonso Garcia El jue., 19 dic. 2019 a las 16:05, Karl Wright (<daddy...@gmail.com>) escribió: > Hi Jorge, > > When you run the job, do you see more than 20,000 documents as part of it? > > Do you see *exactly* 20,000 documents as part of it? > > Unless you are seeing a hard number like that in the UI for that job on > the job status page, I doubt very much that the problem is a numerical > limitation in the number of documents. I would suspect that the inclusion > criteria, e.g. the mime type or maximum length, is excluding documents. > > Karl > > > On Thu, Dec 19, 2019 at 8:51 AM Jorge Alonso Garcia <jalon...@gmail.com> > wrote: > >> Hi Karl, >> We had installed the shaterpoint plugin, and access properly http:/server/ >> _vti_bin/MCPermissions.asmx >> >> [image: image.png] >> >> Sharepoint has more than 20,000 documents, but when execute the jon only >> extract these 20,000. How Can I check where is the issue? >> >> Regards >> >> >> Jorge Alonso Garcia >> >> >> >> El jue., 19 dic. 2019 a las 12:52, Karl Wright (<daddy...@gmail.com>) >> escribió: >> >>> By "stop at 20,000" do you mean that it finds more than 20,000 but stops >>> crawling at that time? Or what exactly do you mean here? >>> >>> FWIW, the behavior you describe sounds like you may not have installed >>> the SharePoint plugin and may have selected a version of SharePoint that is >>> inappropriate. All SharePoint versions after 2008 limit the number of >>> documents returned using the standard web services methods. The plugin >>> allows us to bypass that hard limit. >>> >>> Karl >>> >>> >>> On Thu, Dec 19, 2019 at 6:37 AM Jorge Alonso Garcia <jalon...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> We have an isuse with sharepoint connector. >>>> There is a job that crawl a sharepoint 2016, but it is not recovering >>>> all files, it stop at 20.000 documents without any error. >>>> Is there any parameter that should be change to avoid this limitation? >>>> >>>> Regards >>>> Jorge Alonso Garcia >>>> >>>>