Thanks, nice result, I'll try to run the performance benchmarks that are coming with libstdc++ to see if they spot anything.

That's tests in testsuite/performance folder in case you want to have a try yourself.

François


On 18/01/2024 10:26, Huanghui Nie wrote:

Yes, I have. I did a benchmark today.

The conclusion is: the time consumption can be reduced by 0.4% ~ 1.2% when unordered_set erase(begin()), and 1.2% ~ 2.4% when erase(begin(), end()).


My test environment:

CPU: Intel(R) Xeon(R) CPU E5-2680 v4 @ 2.40GHz, 2393.365 MHz, 56 CPUs

MEM: 256G

OS: CentOS-8.2

g++: gcc version 8.3.1 20191121 (Red Hat 8.3.1-5) (GCC)

Compile flags: -O3 -std=c++17


Test conclusion data (time taken to delete every 100 million elements):

erase(begin()):

|size of unordered_set |100 |1,000 |10,000|100,000 |1,000,000|10,000,000|

|base time consuming (ms)|3827.736|3807.725|3830.168|3807.373|3798.713 |3854.168|

|test time consuming (ms)|3783.406|3789.460|3791.146|3778.033|3783.494 |3808.137|

|Time-consuming reduction|1.16% |0.48% |1.02% |0.77% |0.40%|1.19% |

erase(begin(),end()):

|size of unordered_set |100 |1,000 |10,000|100,000 |1,000,000|10,000,000|

|base time consuming (ms)|2779.229|2768.550|2795.778|2767.385|2761.521 |2804.099|

|test time consuming (ms)|2712.759|2726.578|2752.224|2732.140|2718.953 |2739.727|

|Time-consuming reduction|2.39% |1.52% |1.56% |1.27% |1.54%|2.30% |


Please see the attachment for test code and detailed test result.


2024年1月18日(木) 4:04 François Dumont <frs.dum...@gmail.com>:

    Hi

    Looks like a great finding to me, this is indeed a useless check,
    thanks!

    Have you any figures on the performance enhancement ? It might
    help to get proper approval as gcc is currently in dev stage 4
    that is to say only bug fixes normally.

    François

    On 17/01/2024 09:11, Huanghui Nie wrote:

    Hi.

    When I implemented a hash table with reference to the C++ STL, I
    found that when the hash table in the C++ STL deletes elements,
    if the first element deleted is the begin element, the before
    begin node is repeatedly assigned. This creates unnecessary
    performance overhead.


    First, let’s see the code implementation:

    In _M_remove_bucket_begin, _M_before_begin._M_nxt is assigned
    when &_M_before_begin == _M_buckets[__bkt]. That also means
    _M_buckets[__bkt]->_M_nxt is assigned under some conditions.

    _M_remove_bucket_begin is called by _M_erase and _M_extract_node:

     1. Case _M_erase a range: _M_remove_bucket_begin is called in a
        for loop when __is_bucket_begin is true. And if
        __is_bucket_begin is true and &_M_before_begin ==
        _M_buckets[__bkt], __prev_n must be &_M_before_begin.
        __prev_n->_M_nxt is always assigned in _M_erase. That means
        _M_before_begin._M_nxt is always assigned, if
        _M_remove_bucket_begin is called and &_M_before_begin ==
        _M_buckets[__bkt]. So there’s no need to assign
        _M_before_begin._M_nxt in _M_remove_bucket_begin.
     2. Other cases: _M_remove_bucket_begin is called when __prev_n
        == _M_buckets[__bkt]. And __prev_n->_M_nxt is always assigned
        in _M_erase and _M_before_begin. That means
        _M_buckets[__bkt]->_M_nxt is always assigned. So there's no
        need to assign _M_buckets[__bkt]->_M_nxt in
        _M_remove_bucket_begin.

    In summary, there’s no need to check &_M_before_begin ==
    _M_buckets[__bkt] and assign _M_before_begin._M_nxt in
    _M_remove_bucket_begin.


    Then let’s see the responsibility of each method:

    The hash table in the C++ STL is composed of hash buckets and a
    node list. The update of the node list is responsible for
    _M_erase and _M_extract_node method. _M_remove_bucket_begin
    method only needs to update the hash buckets. The update of
    _M_before_begin belongs to the update of the node list. So
    _M_remove_bucket_begin doesn’t need to update _M_before_begin.


    Existing tests listed below cover this change:

    23_containers/unordered_set/allocator/copy.cc

    23_containers/unordered_set/allocator/copy_assign.cc

    23_containers/unordered_set/allocator/move.cc

    23_containers/unordered_set/allocator/move_assign.cc

    23_containers/unordered_set/allocator/swap.cc

    23_containers/unordered_set/erase/1.cc

    23_containers/unordered_set/erase/24061-set.cc

    23_containers/unordered_set/modifiers/extract.cc

    23_containers/unordered_set/operations/count.cc

    23_containers/unordered_set/requirements/exception/basic.cc

    23_containers/unordered_map/allocator/copy.cc

    23_containers/unordered_map/allocator/copy_assign.cc

    23_containers/unordered_map/allocator/move.cc

    23_containers/unordered_map/allocator/move_assign.cc

    23_containers/unordered_map/allocator/swap.cc

    23_containers/unordered_map/erase/1.cc

    23_containers/unordered_map/erase/24061-map.cc

    23_containers/unordered_map/modifiers/extract.cc

    23_containers/unordered_map/modifiers/move_assign.cc

    23_containers/unordered_map/operations/count.cc

    23_containers/unordered_map/requirements/exception/basic.cc


    Regression tested on x86_64-pc-linux-gnu. Is it OK to commit?


    ---

    ChangeLog:


    libstdc++: hashtable: No need to update before begin node in
    _M_remove_bucket_begin


    2024-01-16Huanghui Nie<nnn...@gmail.com>


    gcc/

    * libstdc++-v3/include/bits/hashtable.h


    ---


    diff --git a/libstdc++-v3/include/bits/hashtable.h
    b/libstdc++-v3/include/bits/hashtable.h

    index b48610036fa..6056639e663 100644

    --- a/libstdc++-v3/include/bits/hashtable.h

    +++ b/libstdc++-v3/include/bits/hashtable.h

    @@ -872,13 +872,10 @@ _GLIBCXX_BEGIN_NAMESPACE_VERSION

          if (!__next_n || __next_bkt != __bkt)

            {

              // Bucket is now empty

    -         // First update next bucket if any

    +         // Update next bucket if any

              if (__next_n)

                _M_buckets[__next_bkt] = _M_buckets[__bkt];

    -         // Second update before begin node if necessary

    -         if (&_M_before_begin == _M_buckets[__bkt])

    -           _M_before_begin._M_nxt = __next_n;

              _M_buckets[__bkt] = nullptr;

            }

        }

Reply via email to