On Tuesday, 6 August 2013 at 04:10:57 UTC, Andre Artus wrote:
On Monday, 5 August 2013 at 13:59:24 UTC, jicman wrote:

Greetings!

I have this code,

foreach (...)
{

 if (std.string.tolower(fext[0]) == "doc" ||
   std.string.tolower(fext[0]) == "docx" ||
   std.string.tolower(fext[0]) == "xls" ||
   std.string.tolower(fext[0]) == "xlsx" ||
   std.string.tolower(fext[0]) == "ppt" ||
   std.string.tolower(fext[0]) == "pptx")
  continue;
}

foreach (...)
{
 if (std.string.tolower(fext[0]) == "doc")
   continue;
 if (std.string.tolower(fext[0]) == "docx")
   continue;
 if (std.string.tolower(fext[0]) == "xls")
   continue;
 if (std.string.tolower(fext[0]) == "xlsx")
   continue;
 if (std.string.tolower(fext[0]) == "ppt")
   continue;
 if (std.string.tolower(fext[0]) == "pptx")
  continue;
 ...
 ...
}

thanks.

josé

What exactly are you trying to do with this? I get the impression that there is an attempt at "local optimization" when broader approach could lead to better results.

For instance. Using the OS's facilities to filter (six requests, one each for "*.doc", "*.docx") could actually end up being a lot faster.

If you could give more detail about what you are trying to achieve then it could be possible to get better results.

The files are in a network drive and doing a list foreach *.doc, *.docx, etc. will be more expensive than getting the list of all the files at once and then processing them accordingly.

Reply via email to