RE: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

2008-01-09 Thread Mick Badran
Kristen - my thoughts too.

Wilhelmina - check the crawl logs to see if the files are being hit in user's 
My Docs.

Cheers,

Mick.


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Hodges, Kristen [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 3:21 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

My understanding is that MySite documents are indexed however they won’t show 
up in Search results unless you have permission to that MySite (ie you are an 
admin or the owner of the site) or it’s a Shared document.  A quick test in my 
MOSS x64 environment confirmed that shared objects from a MySite are listed in 
search results.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of VonBuellen, 
Wilhelmina
Sent: Wednesday, 9 January 2008 3:13 PM
To: listserver@ozMOSS.com
Subject: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

Hello all,

Would someone be able to confirm, deny or clarify an internally held belief 
that My Site documents cannot be indexed by MOSS 2007 because of a permissions 
issue? They are currently not being indexed in our environment and no one here 
can find a solution.

thanks,

[cid:image001.gif@01C852D2.DD5A46B0]

WILHELMINA VONBUELLEN
SPOKES ARCHITECT (CONTRACTOR)
INFORMATION SERVICES
Tel +61(0) 2 62036076  Ext 156076  Mob +61(0) 406103293
www.afp.gov.au




**
WARNING

This email message and any attached files may contain information
that is confidential and subject of legal privilege intended only for
use by the individual or entity to whom they are addressed.   If you
are not the intended recipient or the person responsible for
delivering the message to the intended recipient be advised that you
have received this message in error and that any use, copying,
circulation, forwarding, printing or publication of this message or
attached files is strictly forbidden, as is the disclosure of the
information contained therein. If you have received this message in
error, please notify the sender immediately and delete it from your
inbox.

AFP Web site: http://www.afp.gov.au
**

--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net
--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net



--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.

Powered by mailenable.com - List managed by www.readify.net

<>

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Mick Badran
Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Yeah that’s correct.  Recommendations we were given from Microsoft was 15 
minute intervals – and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don’t do 15 minutes because we didn’t have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is running. That’s on a 64-bit box with 4 
CPUs and 4Gb memory.

You could prove the performance baseline if you create a dummy copy and run  5 
minute incrementals on that?  See what PerfMon shows.  Then you’ll know for 
sure.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:32 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Correct me if I’m wrong, but my understanding of how an incremental crawl works 
is to crawl the site in full an do a hash compare on each item to look for 
changes. So the concern is the load on the servers hosting the sites and the 
time it would take to do the crawl (depending on the amount and size of the 
documents).

(but I agree that it doesn’t make a whole load of sense)

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 3:56 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

That doesn’t seem to make a whole lot of sense…  If the search server is on a 
separate server there are no performance implications???  You’re only talking 
about an incremental crawl right?

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:23 PM
To: listserver@ozMOSS.com
Subject: [OzMOSS] programmatically add items to be crawled

Hi All,

I have a question regarding search.
I’m trying to see if it’s at all possible to programmatically add items into 
the search index. I have a situation where the client wants documents to be 
searchable almost immediately after they have added them.
I’ve suggested that the only solution is to schedule an incremental crawl every 
5min or so, but they are not happy with that solution. Mainly due to 
performance concerns (they are running a separate search server).

Any help / comments are appreciated.
Thanks
Roger

IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages to clients of Oconics Pty Ltd may contain information that is 
confidential and legally privileged. Please do not read, copy, forward or store 
this message unless you are an intended recipient of it. If you have received 
this message in error, please forward it back to the sender and delete it 
completely from your computer system.
--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net
--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net

IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages to clients of Oconics Pty Ltd may contain information that is 
confidential and legally privileged. Please do not read, copy, 

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Roger Noble
Hi Mick,
Thanks for clearing that up, I didn't realise that there was a difference 
between v2 and v3 incremental crawls (we are using v3).

Roger

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 2:34 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled
Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Yeah that's correct.  Recommendations we were given from Microsoft was 15 
minute intervals - and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don't do 15 minutes because we didn't have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is running. That's on a 64-bit box with 4 
CPUs and 4Gb memory.

You could prove the performance baseline if you create a dummy copy and run  5 
minute incrementals on that?  See what PerfMon shows.  Then you'll know for 
sure.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:32 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Correct me if I'm wrong, but my understanding of how an incremental crawl works 
is to crawl the site in full an do a hash compare on each item to look for 
changes. So the concern is the load on the servers hosting the sites and the 
time it would take to do the crawl (depending on the amount and size of the 
documents).

(but I agree that it doesn't make a whole load of sense)

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 3:56 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

That doesn't seem to make a whole lot of sense...  If the search server is on a 
separate server there are no performance implications???  You're only talking 
about an incremental crawl right?

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:23 PM
To: listserver@ozMOSS.com
Subject: [OzMOSS] programmatically add items to be crawled

Hi All,

I have a question regarding search.
I'm trying to see if it's at all possible to programmatically add items into 
the search index. I have a situation where the client wants documents to be 
searchable almost immediately after they have added them.
I've suggested that the only solution is to schedule an incremental crawl every 
5min or so, but they are not happy with that solution. Mainly due to 
performance concerns (they are running a separate search server).

Any help / comments are appreciated.
Thanks
Roger

IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages to clients of Oconics Pty Ltd may contain information that is 
confidential and legally privileged. Please do not read, copy, forward or store 
this message unless you are an intended recipient of it. If you have received 
this message in error, please forward it back to the sender and delete it 
completely from your computer system.
--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net
--- OzMOSS.com 
- to unsubscribe from this list, send a mess

RE: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

2008-01-09 Thread Aaron Saikovski
That's my understanding as well. The indexer will index everything but
the query server provides context trimming on the search on what can
actually be searched by the given user.

Please correct me if i am wrong. J

 

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf
Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 3:21 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

 

My understanding is that MySite documents are indexed however they won't
show up in Search results unless you have permission to that MySite (ie
you are an admin or the owner of the site) or it's a Shared document.  A
quick test in my MOSS x64 environment confirmed that shared objects from
a MySite are listed in search results.

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf
Of VonBuellen, Wilhelmina
Sent: Wednesday, 9 January 2008 3:13 PM
To: listserver@ozMOSS.com
Subject: [OzMOSS] Indexing My Site documents [SEC=UNCLASSIFIED]

 

Hello all,

 

Would someone be able to confirm, deny or clarify an internally held
belief that My Site documents cannot be indexed by MOSS 2007 because of
a permissions issue? They are currently not being indexed in our
environment and no one here can find a solution.

 

thanks,

 

 

WILHELMINA VONBUELLEN
SPOKES ARCHITECT (CONTRACTOR)
INFORMATION SERVICES

Tel +61(0) 2 62036076  Ext 156076  Mob +61(0) 406103293

www.afp.gov.au  

 

 


**
WARNING

This email message and any attached files may contain information
that is confidential and subject of legal privilege intended only for
use by the individual or entity to whom they are addressed.   If you
are not the intended recipient or the person responsible for
delivering the message to the intended recipient be advised that you
have received this message in error and that any use, copying,
circulation, forwarding, printing or publication of this message or
attached files is strictly forbidden, as is the disclosure of the
information contained therein. If you have received this message in
error, please notify the sender immediately and delete it from your
inbox.

AFP Web site: http://www.afp.gov.au
**

---
OzMOSS.com - to unsubscribe from this list, send a message back to the
list with 'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net 

---
OzMOSS.com - to unsubscribe from this list, send a message back to the
list with 'unsubscribe' as the subject.
Powered by mailenable.com - List managed by www.readify.net 




--- OzMOSS.com 
- to unsubscribe from this list, send a message back to the list with 
'unsubscribe' as the subject.

Powered by mailenable.com - List managed by www.readify.net

<>

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Sezai KOMUR
>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

also

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled
Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Yeah that's correct.  Recommendations we were given from Microsoft was 15 
minute intervals - and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don't do 15 minutes because we didn't have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is running. That's on a 64-bit box with 4 
CPUs and 4Gb memory.

You could prove the performance baseline if you create a dummy copy and run  5 
minute incrementals on that?  See what PerfMon shows.  Then you'll know for 
sure.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:32 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Correct me if I'm wrong, but my understanding of how an incremental crawl works 
is to crawl the site in full an do a hash compare on each item to look for 
changes. So the concern is the load on the servers hosting the sites and the 
time it would take to do the crawl (depending on the amount and size of the 
documents).

(but I agree that it doesn't make a whole load of sense)

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 3:56 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

That doesn't seem to make a whole lot of sense...  If the search server is on a 
separate server there are no performance implications???  You're only talking 
about an incremental crawl right?

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:23 PM
To: listserver@ozMOSS.com
Subject: [OzMOSS] programmatically add items to be crawled

Hi All,

I have a question regarding search.
I'm trying to see if it's at all possible to programmatically add items into 
the search index. I have a situation where the client wants documents to be 
searchable almost immediately after they have added them.
I've suggested that the only solution is to schedule an incremental crawl every 
5min or so, but they are not happy with that solution. Mainly due to 
performance concerns (they are running a separate search server).

Any help / comments are apprecia

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Ishai Sagi
Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

also

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled
Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Yeah that's correct.  Recommendations we were given from Microsoft was 15 
minute intervals - and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don't do 15 minutes because we didn't have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is running. That's on a 64-bit box with 4 
CPUs and 4Gb memory.

You could prove the performance baseline if you create a dummy copy and run  5 
minute incrementals on that?  See what PerfMon shows.  Then you'll know for 
sure.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:32 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Correct me if I'm wrong, but my understanding of how an incremental crawl works 
is to crawl the site in full an do a hash compare on each item to look for 
changes. So the concern is the load on the servers hosting the sites and the 
time it would take to do the crawl (depending on the amount and size of the 
documents).

(but I agree that it doesn't make a whole load of sense)

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 3:56 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

That doesn't seem to make a whole lot of sense...  If th

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Roger Noble
I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

also

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled
Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Yeah that's correct.  Recommendations we were given from Microsoft was 15 
minute intervals - and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don't do 15 minutes because we didn't have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is running. That's on a 64-bit box with 4 
CPUs and 4Gb memory.

You could prove the performance baseline if you create a dummy copy and run  5 
minute incrementals on that?  See what PerfMon shows.  Then you'll know for 
sure.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Wednesday, 9 January 2008 4:32 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Correct me if I'm wrong, but

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Ishai Sagi
Roger,

There is a reason MS are doing crawls on timers. Adding documents as they are 
added to the database will cause major performance headaches and potential 
locks. I recommend that your customer accepts that...



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 11:49 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled



I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body



Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.







From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled



>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))



Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.



But what happens with a BDC incremental crawl ? Can I set them to 5mins too?



I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx



"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "



also



"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"



So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source

Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled



Hi Roger - (sorry for coming in late)



Are your sites v2 or v3?



If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.



I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)



A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.



15 min intervals is what we used to do for V2 sites.



Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]

Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor

blogs:http://blogs.breezetraining.com.au/mickb



   _

From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Ok, that sounds like a good place to start.

Thanks for your help!



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body



Yeah that's correct.  Recommendations we were given from Microsoft was 15 
minute intervals - and we have a large ContentDb + same server for search and 
hosting.  I have to admit though we don't do 15 minutes because we didn't have 
any need.  That said, MOM alerts have been configured if CPU usage exceeds 80% 
- it never happens even when search is run

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Aaron Saikovski
I would second that

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 12:16 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

Roger,

There is a reason MS are doing crawls on timers. Adding documents as they are 
added to the database will cause major performance headaches and potential 
locks. I recommend that your customer accepts that...

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 11:49 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

 

Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.

 

 

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

 

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

 

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

 

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

 

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

 

also

 

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string" 

 

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source

Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

Hi Roger - (sorry for coming in late)

 

Are your sites v2 or v3?

 

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

 

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

 

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

 

15 min intervals is what we used to do for V2 sites.

 

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]

Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor

blogs:http://blogs.breezetraining.com.au/mickb 
 

 



From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Ok, that sounds like a good place to start.

Thanks for your help!

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Hodges, Kristen
Sent: Wednesday, 9 January 2008 4:11 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

 

Yeah that's correct.  Recomm

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Roger Noble
I've made this recommendation to the client, just waiting to hear back.

Thanks everyone for your help on this.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Aaron Saikovski
Sent: Thursday, 10 January 2008 1:21 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

I would second that

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 12:16 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Roger,
There is a reason MS are doing crawls on timers. Adding documents as they are 
added to the database will cause major performance headaches and potential 
locks. I recommend that your customer accepts that...

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 11:49 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

also

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb


From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Roger Noble [EMAIL 
PROTECTED]
Sent: Wednesday, 9 January 2008 4:43 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled
Ok, that sounds like a good place to start.
Thanks for your help!

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] 

RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Aaron Saikovski
Be interested to hear the outcome.

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 1:53 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

I've made this recommendation to the client, just waiting to hear back.

 

Thanks everyone for your help on this.

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Aaron Saikovski
Sent: Thursday, 10 January 2008 1:21 PM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

 

I would second that

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 12:16 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

Roger,

There is a reason MS are doing crawls on timers. Adding documents as they are 
added to the database will cause major performance headaches and potential 
locks. I recommend that your customer accepts that...

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 11:49 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

 

Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.

 

 

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

 

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

 

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

 

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

 

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

 

also

 

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string" 

 

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source

Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

 

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

 

Hi Roger - (sorry for coming in late)

 

Are your sites v2 or v3?

 

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

 

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

 

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

 

15 min intervals is what we used to do for V2 sites.

 

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]

Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor

blogs:http://blogs.breezetraining.com.au/mickb 


RE: [OzMOSS] programmatically add items to be crawled

2008-01-09 Thread Sezai KOMUR
I would "third" that J

A general problem I find is people always try to introduce unrealistic 
requirements in MOSS 2007 projects, you then end up pushing the product in 
directions it's not meant to go.

Too much complicated and time consuming custom development and once you get to 
the end of the journey you realise that it was a bad idea, but the problem is 
that's what the client has asked for.
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Aaron Saikovski
Sent: Thursday, 10 January 2008 11:51 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

I would second that

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 12:16 PM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Roger,
There is a reason MS are doing crawls on timers. Adding documents as they are 
added to the database will cause major performance headaches and potential 
locks. I recommend that your customer accepts that...

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Roger Noble
Sent: Thursday, 10 January 2008 11:49 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

I had thought of doing this as we are currently using a workflow with the 
documents. My concern was that if a document was added once a crawl was already 
in progress then the crawl could potentially miss the new document. My solution 
to this would to then queue the crawls potentially as part of a timer. So that 
once the crawl had completed it would be crawled again - but this seems a bit 
ridiculous to me.

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ishai Sagi
Sent: Thursday, 10 January 2008 11:14 AM
To: listserver@ozMOSS.com
Subject: [SPAM] - RE: [OzMOSS] programmatically add items to be crawled - Found 
word(s) list error in the Text body

Not sure if this is relevant, but I once had a similar requirement and solved 
it by writing an event handler (you can also do it with workflow) to force a 
incremental crawl of the specific index that held the information. That way, 
after every change to that specific list, a crawl was triggered. Offcourse the 
event handler could have been attached to more lists, or if you develop it as a 
workflow and deploy it as a policy then you have an easy to deploy solution - 
but beware of the performance implications! My solution was ok because it was 
on a rarely changed list - and the index it was connecting to was relatively 
small.



From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Sezai KOMUR
Sent: Thursday, 10 January 2008 11:19 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

>> the index service will only look at the change log of the sites/site coll in 
>> question (SPSite.GetChanges(...))

Thanks for the info Mick, I've always wondered how an incremental sharepoint 
site crawl works in V3.

But what happens with a BDC incremental crawl ? Can I set them to 5mins too?

I was just about to set incremental crawls to 5 mins on a BDC content source as 
a test, but then realised it obviously can't use SPSite.GetChanges to get the 
delta, so I googled bdc incremental crawl and found the following - 
http://geekswithblogs.net/Tariq/archive/2007/05/23/112680.aspx

"You would need some column on your table/view/or Stored-Proc to indicate the 
last modified time of that entity instance "

also

"this property needs to be called __BDCLastModifiedTimestamp and should be of 
type string"

So if your lob application doesn't store the last modified time for each 
individual row, incremental crawls won't work for a BDC content source
Sezai Kömür
Senior Developer  - BEng, BSc - Microsoft Certified Technology Specialist  -  
http://www.moss2007.com.au/

From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Mick Badran
Sent: Thursday, 10 January 2008 1:04 AM
To: listserver@ozMOSS.com
Subject: RE: [OzMOSS] programmatically add items to be crawled

Hi Roger - (sorry for coming in late)

Are your sites v2 or v3?

If they are v3 the index service will only look at the change log of the 
sites/site coll in question (SPSite.GetChanges(...)) so only the changes are 
indexed for that last schedule interval.

I generally use every 5 mins for the crawl interval (incremental) on large 
sites with no worries (even if the V3 sites are on other boxes/farms)

A good place to get a handle on all this is the "Enterprise Search Resource 
Kit" - bunch of goodies on there from the prod team.

15 min intervals is what we used to do for V2 sites.

Mick Badran (MVP - BizTalk) | mb: +61 404842833 | im:[EMAIL PROTECTED]
Breeze Training | Training + Integration Specialist | Microsoft Readiness 
Instructor
blogs:http://blogs.breezetraining.com.au/mickb