Re: Templates are slow.
On Friday, 9 September 2016 at 18:17:02 UTC, deadalnix wrote: You need to compare the string to unique them, so it doesn't change anything. It changes the frequency of comparisons.
Re: Templates are slow.
On Friday, 9 September 2016 at 12:09:32 UTC, Steven Schveighoffer wrote: On 9/8/16 6:57 PM, Stefan Koch wrote: Hi Guys, I have some more data. In the binderoo example the main time is spent in the backend. generating code and writing objects files. If we ever get Rainer's patch to collapse repetitive templates, we may help this problem. https://github.com/dlang/dmd/pull/5855 The front-end spends most of it's time comparing strings of unique type-names :) I thought the front end was changed to use the string pointer for symbol names as the match so string comparisons aren't done? Hm... maybe to intern the string? That kind of makes sense. I just had a thought. If you hash the string, and then compare the length of the string and first and last character along with the hash, what are the chances of it being a false positive? -Steve You need to compare the string to unique them, so it doesn't change anything.
Re: Templates are slow.
On Thu, 2016-09-08 at 14:23 +0200, Andrei Alexandrescu via Digitalmars- d wrote: > On 9/8/16 7:02 AM, Stefan Koch wrote: > > > > > > I will write an article about why templates are slow. > > > > The gist will be however : "Templates being slow is an inherent > > property > > of templates." (We are only talking about templates as defined by > > (C++ > > and D)). > > That would be a great article. Are there any situations that we canĀ > special-case away? Raising the roof. -- Andrei One that could get published in CVu or Overload. https://accu.org/index.php/journal -- Russel. = Dr Russel Winder t: +44 20 7585 2200 voip: sip:russel.win...@ekiga.net 41 Buckmaster Roadm: +44 7770 465 077 xmpp: rus...@winder.org.uk London SW11 1EN, UK w: www.russel.org.uk skype: russel_winder signature.asc Description: This is a digitally signed message part
Re: Templates are slow.
On Friday, 9 September 2016 at 12:09:32 UTC, Steven Schveighoffer wrote: On 9/8/16 6:57 PM, Stefan Koch wrote: Hi Guys, I have some more data. In the binderoo example the main time is spent in the backend. generating code and writing objects files. If we ever get Rainer's patch to collapse repetitive templates, we may help this problem. https://github.com/dlang/dmd/pull/5855 The front-end spends most of it's time comparing strings of unique type-names :) I thought the front end was changed to use the string pointer for symbol names as the match so string comparisons aren't done? In this case the string is freshly and not available as reference to an already lexed string . Hm... maybe to intern the string? That kind of makes sense. Yes that would be the way to go. I just had a thought. If you hash the string, and then compare the length of the string and first and last character along with the hash, what are the chances of it being a false positive? This depends entirely on the distribution of strings. It's probably quite high :)
Re: Templates are slow.
On Friday, 9 September 2016 at 12:09:32 UTC, Steven Schveighoffer wrote: I just had a thought. If you hash the string, and then compare the length of the string and first and last character along with the hash, what are the chances of it being a false positive? Any chance of a false positive is nontrivial when hashes are being compared so often. Never, never ever in production code are you to assume that hashes and other non-definitive data can be used to guarantee uniqueness.
Re: Templates are slow.
On 9/8/16 6:57 PM, Stefan Koch wrote: Hi Guys, I have some more data. In the binderoo example the main time is spent in the backend. generating code and writing objects files. If we ever get Rainer's patch to collapse repetitive templates, we may help this problem. https://github.com/dlang/dmd/pull/5855 The front-end spends most of it's time comparing strings of unique type-names :) I thought the front end was changed to use the string pointer for symbol names as the match so string comparisons aren't done? Hm... maybe to intern the string? That kind of makes sense. I just had a thought. If you hash the string, and then compare the length of the string and first and last character along with the hash, what are the chances of it being a false positive? -Steve
Re: Templates are slow.
On Friday, 9 September 2016 at 01:38:40 UTC, deadalnix wrote: On Thursday, 8 September 2016 at 20:10:01 UTC, Stefan Koch wrote: generating separate object files for each template instanciation is and then only re-generating on change will only be effective if they do not change much. From one build to the next. You'd have tens of thousands of file and a big io problem. I already thought about that. The Idea is to stuff the object-code of all templates in one-file with a bit a meta-data. make a do a hash lookup at instanciation. And write the cached code in of the instanciation is found. I agree, file I/O would kill any speed win many times over!
Re: Templates are slow.
On Thursday, 8 September 2016 at 20:10:01 UTC, Stefan Koch wrote: generating separate object files for each template instanciation is and then only re-generating on change will only be effective if they do not change much. From one build to the next. You'd have tens of thousands of file and a big io problem.
Re: Templates are slow.
On Thursday, 8 September 2016 at 22:57:07 UTC, Stefan Koch wrote: The front-end spends most of it's time comparing strings of unique type-names :) (Waits for Walter to say, "Use a pool Luke!")
Re: Templates are slow.
Hi Guys, I have some more data. In the binderoo example the main time is spent in the backend. generating code and writing objects files. The front-end spends most of it's time comparing strings of unique type-names :) One particular outlier in the backend code is the function ecom which eliminates common subexpression. We would potentially save some time by not emitting those in the first-place.
Re: Templates are slow.
On Thursday, 8 September 2016 at 19:17:42 UTC, Lewis wrote: Am I crazy in wondering about caching template instantiations? I understand that an incremental build would kind of accomplish this goal, but that comes with its own set of problems. Not as good as what you propose, but: LDC 1.1.0 can do _codegen_ caching which I guess is some intermediate form of incremental building. My testcase is a unittest piece from Weka.io that instantiates, oh, I don't remember exactly, 100.000+ templates. It takes about 65 seconds to compile. With codegen caching, the re(!)compile time on cache-hit is reduced to 39s. Note that the front-end still instantiates all those templates, but LDC's codegen at -O0 is not as fast as DMD's (calculating the hash also takes time and could be optimized further). In summary, for LDC -O3 builds, you can expect a large speed boost by just adding `-ir2obj-cache=` to the commandline (LDC >= 1.1.0-alpha1). -Johan
Re: Templates are slow.
On Thursday, 8 September 2016 at 19:49:38 UTC, Ethan Watson wrote: On Thursday, 8 September 2016 at 19:17:42 UTC, Lewis wrote: I can't help but wonder if there were some way to automatically cache templates instantiations between runs of dmd? I'm running with Visual D, which has a "COMPILE ALL THE THINGS" mentality as the default. As part of the rapid iteration part of Binderoo, I plan on doing incremental linking. Of course, if all template instantiations go in to one object file, that really ruins it. Each template instantiation going in to a separate object file will actually make life significantly easier, as each compile will have less output. The only time those template instantiations need to recompile is if the invoking module changes; the template's dependencies change; or the module the template lives in changes. My opinion is that splitting up object files will do more to reduce compile time for me than anything else, the pipeline we had for Quantum Break was to compile and link in separate steps so it's not much effort at all for me to keep that idea running in Binderoo and make it incrementally link. But I don't know the DMD code and I'm not a compiler writer, so I cannot say that authoritatively. It sounds very reasonable to me at least. generating separate object files for each template instanciation is and then only re-generating on change will only be effective if they do not change much. From one build to the next. For binderoos purpose this could be rather effective. As long as no one adds fields at the beginning of the structs :) Without incremental linking however your compile-times will shoot through the roof. And will probably damage the moon as well.
Re: Templates are slow.
On Thursday, 8 September 2016 at 19:17:42 UTC, Lewis wrote: I can't help but wonder if there were some way to automatically cache templates instantiations between runs of dmd? I'm running with Visual D, which has a "COMPILE ALL THE THINGS" mentality as the default. As part of the rapid iteration part of Binderoo, I plan on doing incremental linking. Of course, if all template instantiations go in to one object file, that really ruins it. Each template instantiation going in to a separate object file will actually make life significantly easier, as each compile will have less output. The only time those template instantiations need to recompile is if the invoking module changes; the template's dependencies change; or the module the template lives in changes. My opinion is that splitting up object files will do more to reduce compile time for me than anything else, the pipeline we had for Quantum Break was to compile and link in separate steps so it's not much effort at all for me to keep that idea running in Binderoo and make it incrementally link. But I don't know the DMD code and I'm not a compiler writer, so I cannot say that authoritatively. It sounds very reasonable to me at least.
Re: Templates are slow.
It's true that templates are inherently slow, and there isn't a ton we can do about that. However, almost every time I compile the project (hundreds of times per day), the overwhelming majority of the time, the same templates are being re-instantiated in exactly the same way. I can't help but wonder if there were some way to automatically cache templates instantiations between runs of dmd? The heavytemplates.d workaround I've used kind of accomplishes this as a total hack job. However... - It adds complexity to the build process - It adds a small overhead of linking in an extra .lib (although this is dwarfed by the win from no longer rebuilding expensive templates every build) - It means that when heavytemplates.d changes, my rebuild is significantly longer than before since I'm running dmd twice - It means extra work to implement that we don't want every developer to do themselves Am I crazy in wondering about caching template instantiations? I understand that an incremental build would kind of accomplish this goal, but that comes with its own set of problems. I can't help but think that there's some way to make dmd smarter about not redoing the exact same work build after build, when the templates and their instantiations only change very rarely.
Re: Templates are slow.
I recently went through the process of optimizing the build time on one of my projects. I started at ~3.08s, and got it down to ~1.6s. The project is around 7000 non-comment-non-whitespace LOC. I timed the build in a pretty non-rigourous fashion (I just timed the python script that kicks off a command-line call to dmd and waits for it to finish), however I only cared about making changes that resulted in large improvements to the build time, so this was good enough for my purposes. I too found that template instantiation was responsible for a lot of the extra build time. I found running dmd -v very helpful in tracking down excessive template instantiations or other places where the compiler was doing a lot of work that could be avoided. The steps I took were as follows: - Start (3.08s) - I was using custom assert function that grabbed __LINE__ and __FILE__ as template arguments, meaning each of the ~130 assert calls required a separate instantiation. I switched to passing those in as run-time arguments (2.85s) - I had a similar wrapper around some logging functions in std.experimental.logger. I made a small change to std.experimental.logger to allow a call path with no template instantiations, and similarly fixed my own wrapper. I had ~70 logging calls (2.7s) - Recompiled DMD with VS2015 (2.5s) - Overclocked my CPU :D (2.3s) - Created a file called heavytemplates.d that built to a .lib in a separate build step. The first templates I pulled out were a couple std.regex calls and instantiations (1.9s) - Changed some tuples into structs (negligible improvement) - Pulled several templates into heavytemplates.d that instantiate recursively over the Gamestate (a very large struct) (1.75s) - Pulled out template instantiations used by msgpack-d, which also instantiate recursively over the Gamestate for save/load of the game (1.6s) Of all of these, I was most surprised by the gain I got from pulling out std.regex calls into a separate build (0.4ms). Whether or not I used compile-time regexes didn't seem to affect build time substantially, just that I used anything at all. Also, whether I had one regex call or five didn't seem to matter, likely because std.regex instantiates using the string type as a parameter, and I just used plain old 'string' for all regex uses. There's still work I could do, but at some point I start to get diminishing returns, and have to actually work on features instead of just optimizing my build :D
Re: Templates are slow.
On Thu, Sep 08, 2016 at 04:37:36PM +, Stefan Koch via Digitalmars-d wrote: [...] > Also we need to special case ranges in general. > And try harder to inline calls to range functions. > Maybe even early in the frontend. [...] Yeah, dmd's inliner is really pessimistic. It gives up too easily, and as a result often misses a whole series of further opportunities that could have opened up, had it decided to inline. Having range-specific optimizations is a good idea, I think. Since it's one of D's big selling points, it needs to be as high-performance as we can possibly make it. I think there is a lot we can do in this area that we can't do as general optimizations; range-based code has certain recurring structures that should be highly-exploitable in terms of optimization opportunities. IME I've found that GDC is much better at optimizing range-based code because of its aggressive inliner (and also better loop optimization algorithms); but there's probably still room for range-specific optimizations that a general optimizer like the gcc backend wouldn't have. T -- Democracy: The triumph of popularity over principle. -- C.Bond
Re: Templates are slow.
On Thursday, 8 September 2016 at 15:45:53 UTC, Jonathan M Davis wrote: It's critical that we do what we can to make templates fast. And if we can't make them fast enough, we'll definitely have to come up with techniques/guidelines for reducing their usage when they're not really needed. - Jonathan M Davis I agree. We need to make templates faster. But it will be like squeezing water out of stones. A few more oblivious optimizations I have tried did not have the desired effect at all. @Andrei Also we need to special case ranges in general. And try harder to inline calls to range functions. Maybe even early in the frontend. secondly we need to inline range-chains into each other whenever possible. If we do this the right way early on we can reduce the symbolName-length as well. All we need for this is pattern-matching on a type-resolved call-graph. Which is something I am working on as part of my ctfe work.
Re: Templates are slow.
On Thursday, September 08, 2016 07:43:10 Ethan Watson via Digitalmars-d wrote: > On Thursday, 8 September 2016 at 05:02:38 UTC, Stefan Koch wrote: > > I have just hit a barrier trying to optimize the compile-time > > in binderoo. > > I did a double take when Stefan told me the representative sample > code I gave him to run with Binderoo instantiated ~20,000 > templates and resulted in ~10,000,000 hash map look ups inside > the compiler. > > I can certainly write it to be more optimised, but one of the > goals I have is to make the codebase human readable so that it's > not just Manu and myself that can understand the code. As a > result, I figure this could be representative of how an ordinary > user would write templated code. IIRC, Don posted at one point about how the std.algorithm unit tests ended up with over a million template instantiations. All of the eponymous templates that we use for template constraints really add up, and having heavily range-based code is going to rack up the number of template instantiations as well. It's critical that we do what we can to make templates fast. And if we can't make them fast enough, we'll definitely have to come up with techniques/guidelines for reducing their usage when they're not really needed. Improvements to CTFE have really helped though. std.metastrings was dog slow in comparison to using CTFE, and at this point, we don't need std.metastrings anymore, and so it's gone. That's definitely not going to work in all cases though - just in cases where something is being done with templates that could be done with a function that could run at runtime now that CTFE can do most things that can be done at runtime. And we've probably gotten most everything we can out of that transition already - at least in Phobos. - Jonathan M Davis
Re: Templates are slow.
On Thursday, 8 September 2016 at 05:02:38 UTC, Stefan Koch wrote: If you are instantiating a template inside another template think very hard about the reason, often you can "inline" the template body of the inner template and get an instant speed win right there. (Don't do this preemptively, ONLY when you know that this template is a problem!) What about functions in std.algorithm, like reduce, that have templates with multiple functions within them and the headline function will call private impl functions within them?
Re: Templates are slow.
On Thursday, 8 September 2016 at 12:23:35 UTC, Andrei Alexandrescu wrote: Are there any situations that we can special-case away? Raising the roof. -- Andrei The rangefying functions in std.array come to mind. That will give a huge boost to everyone. (everyone who uses arrays anyway :))
Re: Templates are slow.
On 9/8/16 7:02 AM, Stefan Koch wrote: I will write an article about why templates are slow. The gist will be however : "Templates being slow is an inherent property of templates." (We are only talking about templates as defined by (C++ and D)). That would be a great article. Are there any situations that we can special-case away? Raising the roof. -- Andrei
Re: Templates are slow.
On Thursday, 8 September 2016 at 05:02:38 UTC, Stefan Koch wrote: I have just hit a barrier trying to optimize the compile-time in binderoo. I did a double take when Stefan told me the representative sample code I gave him to run with Binderoo instantiated ~20,000 templates and resulted in ~10,000,000 hash map look ups inside the compiler. I can certainly write it to be more optimised, but one of the goals I have is to make the codebase human readable so that it's not just Manu and myself that can understand the code. As a result, I figure this could be representative of how an ordinary user would write templated code.
Re: Templates are slow.
On Thursday, 8 September 2016 at 06:34:58 UTC, Sebastien Alaiwan wrote: On Thursday, 8 September 2016 at 05:02:38 UTC, Stefan Koch wrote: (Don't do this preemptively, ONLY when you know that this template is a problem!) How would you measure such things? Is there such a thing like a "compilation time profiler" ? (Running oprofile on a dmd with debug info comes first to mind ; however, this would only give me statistics on dmd's source code, not mine.) I use a special profilng-build of dmd. oprofile on dmd can give you a good first impression of where you run into problems and then you can wirte special profiling code for this. If you do not want to write such code send me a message and I will look into it for you :)
Re: Templates are slow.
On Thursday, 8 September 2016 at 05:02:38 UTC, Stefan Koch wrote: (Don't do this preemptively, ONLY when you know that this template is a problem!) How would you measure such things? Is there such a thing like a "compilation time profiler" ? (Running oprofile on a dmd with debug info comes first to mind ; however, this would only give me statistics on dmd's source code, not mine.)
Templates are slow.
Hi Guys, I have just hit a barrier trying to optimize the compile-time in binderoo. Roughly 90% of the compile-time is spent instantiating templates. The 10% for CTFE are small in comparison. I will write an article about why templates are slow. The gist will be however : "Templates being slow is an inherent property of templates." (We are only talking about templates as defined by (C++ and D)). That said: Templates are great! But you have to use them sanely. If you are instantiating a template inside another template think very hard about the reason, often you can "inline" the template body of the inner template and get an instant speed win right there. (Don't do this preemptively, ONLY when you know that this template is a problem!) Phobos is has many templates inside templates. In constraints for example. I have no idea how to cut down on template-instanciations in phobos while still maintaining the same user-friendliness. Of course myself and other will continue fighting on the compiler-front. To give you that fastest implementation possible! Cheers, Stefan
D vs C++ metaprogramming: why are c++ templates inherently slow
This post is in response to Nick's suggestion in this post: http://forum.dlang.org/thread/k524ke$gvt$1...@digitalmars.com#post-20121104000747.1f10:40unknown to repost the question here. (please note I have read the link Nick provided, but as noted in my last post to the c++ list, that link, although very convincing, did not answer my question about why ADL and partial specialization are one of the causes of slow compile times.) So here's the original post to the c++ newsgroup: In this post: http://article.gmane.org/gmane.comp.lib.boost.devel/189925 Eric indicates that Walter Bright believes, in c++: that instantiating a template is inherently expensive, and certain features of the C++ language (ADL, partial specialization, etc.) force that to be the case. Walter, if Eric is remembering correctly, could you provide a little more explanation on why this is so or provide a link to some document supporting this conclusion? TIA. -regards, Larry
Re: D vs C++ metaprogramming: why are c++ templates inherently slow
On Sunday, 4 November 2012 at 16:44:30 UTC, evansl wrote: that instantiating a template is inherently expensive Just a guess, but perhaps pattern-matching templates is NP-complete?
Re: D vs C++ metaprogramming: why are c++ templates inherently slow
On 11/04/2012 08:47 PM, Mehrdad wrote: On Sunday, 4 November 2012 at 16:44:30 UTC, evansl wrote: that instantiating a template is inherently expensive Just a guess, but perhaps pattern-matching templates is NP-complete? That does not say anything about evaluation speed.