Are you going through any middleware or proxy, like promxy?

rate(foo[1m]) should definitely give no answer at all, when the timeseries 
data is sampled at 1 minute intervals.

Here is a working query_range for rate[1m] where the scrape interval is 15s:

# curl -Ssg 
'http://localhost:9090/api/v1/query_range?query=rate(ifHCInOctets{instance="gw1",ifName="ether1"}[60s])&start=1649264340&end=1649264640&step=60'
 
| python3 -m json.tool
{
    "status": "success",
    "data": {
        "resultType": "matrix",
        "result": [
            {
                "metric": {
                    "ifIndex": "16",
                    "ifName": "ether1",
                    "instance": "gw1",
                    "job": "snmp",
                    "module": "mikrotik_secret",
                    "netbox_type": "device"
                },
                "values": [
                    [
                        1649264340,
                        "578.6444444444444"
                    ],
                    [
                        1649264400,
                        "651.4222222222221"
                    ],
                    [
                        1649264460,
                        "135.17777777777778"
                    ],
                    [
                        1649264520,
                        "1699.4888888888888"
                    ],
                    [
                        1649264580,
                        "441.5777777777777"
                    ],
                    [
                        1649264640,
                        "39768.08888888888"
                    ]
                ]
            }
        ]
    }
}

But if I make exactly the same query but with rate[15s] then there are no 
answers:

# curl -Ssg 
'http://localhost:9090/api/v1/query_range?query=rate(ifHCInOctets{instance="gw1",ifName="ether1"}[15s])&start=1649264340&end=1649264640&step=60'
 
| python3 -m json.tool
{
    "status": "success",
    "data": {
        "resultType": "matrix",
        "result": []
    }
}

I think the real reason for your problem is hidden; you're obfuscating the 
query and metric names, and I suspect it's hidden behind that.  Sorry, I 
can't help you any further given what I can see, but hopefully you have an 
idea where you can look further.

On Wednesday, 6 April 2022 at 18:45:10 UTC+1 [email protected] wrote:

> Hey Brian,
>
> In the original post I put the output of the raw time series as gathered 
> the way you suggest. I'll copy it again below:
>
> {
>     "data": {
>         "result": [
>             {
>                 "metric": {/* redacted */},
>                 "values": [
>                     [
>                         1649239253.4,
>                         "225201"
>                     ],
>                     [
>                         1649239313.4,
>                         "225226"
>                     ],
>                     [
>                         1649239373.4,
>                         "225249"
>                     ],
>                     [
>                         1649239433.4,
>                         "225262"
>                     ],
>                     [
>                         1649239493.4,
>                         "225278"
>                     ],
>                     [
>                         1649239553.4,
>                         "225310"
>                     ],
>                     [
>                         1649239613.4,
>                         "225329"
>                     ],
>                     [
>                         1649239673.4,
>                         "225363"
>                     ],
>                     [
>                         1649239733.4,
>                         "225402"
>                     ],
>                     [
>                         1649239793.4,
>                         "225437"
>                     ],
>                     [
>                         1649239853.4,
>                         "225466"
>                     ],
>                     [
>                         1649239913.4,
>                         "225492"
>                     ],
>                     [
>                         1649239973.4,
>                         "225529"
>                     ],
>                     [
>                         1649240033.4,
>                         "225555"
>                     ],
>                     [
>                         1649240093.4,
>                         "225595"
>                     ]
>                 ]
>             }
>         ],
>         "resultType": "matrix"
>     },
>     "status": "success"
> }
>
> The query was of the form `counter[15m]` at a given time. I don't see 
> duplicate scrape data in there.
>
> The version of prometheus is 2.26.0, revision 
> 3cafc58827d1ebd1a67749f88be4218f0bab3d8d, go version go1.16.2.
>
> On Wednesday, April 6, 2022 at 6:13:10 PM UTC+1 Brian Candler wrote:
>
>> What version of prometheus are you running?
>>
>> With prometheus, rate(counter[1m]) should give you no results at all when 
>> you are scraping at 1 minute intervals - unless something has changed very 
>> recently (I'm running 2.33.4).  So this is a big red flag.
>>
>> Now, for driving the query API, you should be able to do it like this:
>>
>> # curl -Ssg '
>> http://localhost:9090/api/v1/query?query=ifHCInOctets{instance="gw1",ifName="ether1"}[60s]'
>>  
>> | python3 -m json.tool
>>
>> {
>>     "status": "success",
>>     "data": {
>>         "resultType": "matrix",
>>         "result": [
>>             {
>>                 "metric": {
>>                     "__name__": "ifHCInOctets",
>>                     "ifIndex": "16",
>>                     "ifName": "ether1",
>>                     "instance": "gw1",
>>                     "job": "snmp",
>>                     "module": "mikrotik_secret",
>>                     "netbox_type": "device"
>>                 },
>>                 "values": [
>>                     [
>>                         1649264595.241,
>>                         "117857843410 <(785)%20784-3410>"
>>                     ],
>>                     [
>>                         1649264610.241,
>>                         "117858063821 <(785)%20806-3821>"
>>                     ],
>>                     [
>>                         1649264625.241,
>>                         "117858075769 <(785)%20807-5769>"
>>                     ]
>>                 ]
>>             }
>>         ]
>>     }
>> }
>>
>> There I gave a range vector of 60 seconds, and I got 3 data points 
>> because I'm scraping at 15 second intervals, so only 3 points fell within 
>> the time window of (current time) and (current time - 60s)
>>
>> Sending a query_range will sample the data at intervals.  Only an actual 
>> range vector query (as shown above) will show you *all* the data points in 
>> the time series, wherever they lie.
>>
>> I think you should do this.  My guess - and it's only a guess at the 
>> moment - is that there are multiple points being received for the same 
>> timeseries, and this is giving your spike.  This could be due to 
>> overlapping scrape jobs for the same timeseries, or relabelling removing 
>> some distinguishing label, or some HA setup which is scraping the same 
>> timeseries multiple times but not adding external labels to distinguish 
>> them.
>>
>> I do have some evidence for my guess.  If you are storing the same data 
>> points twice, this will give you the rate of zero most of the time, when 
>> doing rate[1m], because there are two adjacent identical points most of the 
>> time (whereas if there were only a single data point, you'd get no rate at 
>> all).  And you'll get a counter spike if two data points get transposed.
>>
>> On Wednesday, 6 April 2022 at 14:37:57 UTC+1 [email protected] wrote:
>>
>>> Here's the query inspector output from Grafana for rate(counter[2m]). It 
>>> makes the answer to question 1 in my original post more clear. You're 
>>> right, the graph for 1m is just plain wrong. We do still see the reset, 
>>> though.
>>>
>>> {
>>>   "request": {
>>>     "url": 
>>> "api/datasources/proxy/1/api/v1/query_range?query=rate(counter[2m])&start=1649239200&end=1649240100&step=60",
>>>
>>>     "method": "GET",
>>>     "hideFromInspector": false
>>>   },
>>>   "response": {
>>>     "status": "success",
>>>     "data": {
>>>       "resultType": "matrix",
>>>       "result": [
>>>         {
>>>           "metric": {/* redacted */},
>>>           "values": [
>>>             [
>>>               1649239200,
>>>               "0.2871886897537781"
>>>             ],
>>>             [
>>>               1649239260,
>>>               "0.3084619260318357"
>>>             ],
>>>             [
>>>               1649239320,
>>>               "0.26591545347572043"
>>>             ],
>>>             [
>>>               1649239380,
>>>               "0.2446422171976628"
>>>             ],
>>>             [
>>>               1649239440,
>>>               "0.13827603580737463"
>>>             ],
>>>             [
>>>               1649239500,
>>>               "0.1701858902244611"
>>>             ],
>>>             [
>>>               1649239560,
>>>               "0.3403717804489222"
>>>             ],
>>>             [
>>>               1649239620,
>>>               "0.20209574464154753"
>>>             ],
>>>             [
>>>               1649239680,
>>>               "0.3616450167269798"
>>>             ],
>>>             [
>>>               1649239740,
>>>               "2397.9404664989347"
>>>             ],
>>>             [
>>>               1649239800,
>>>               "2397.88728340824"
>>>             ],
>>>             [
>>>               1649239860,
>>>               "0.3084619260318357"
>>>             ],
>>>             [
>>>               1649239920,
>>>               "0.27655207161474926"
>>>             ],
>>>             [
>>>               1649239980,
>>>               "0.39355487114406623"
>>>             ],
>>>             [
>>>               1649240040,
>>>               "0.27655207161474926"
>>>             ],
>>>             [
>>>               1649240100,
>>>               "0.43610134370018155"
>>>             ]
>>>           ]
>>>         }
>>>       ]
>>>     }
>>>   }
>>> }
>>>
>>> On Wednesday, April 6, 2022 at 2:34:59 PM UTC+1 Sam Rose wrote:
>>>
>>>> We do see a graph with rate(counter[1m]). It even looks pretty close to 
>>>> what we see with rate(counter[2m]). We definitely scrape every 60 seconds, 
>>>> double checked our config to make sure.
>>>>
>>>> The exact query was `counter[15m]`. Counter is 
>>>> `django_http_responses_total_by_status_total` in reality, with a long list 
>>>> of labels attached to ensure I'm selecting a single time series.
>>>>
>>>> I didn't realise Grafana did that, thank you for the advice.
>>>>
>>>> I feel like we're drifting away from the original problem a little bit. 
>>>> Can I get you any additional data to make the original problem easier to 
>>>> debug?
>>>>
>>>> On Wednesday, April 6, 2022 at 2:31:27 PM UTC+1 Brian Candler wrote:
>>>>
>>>>> If you are scraping at 1m intervals, then you definitely need 
>>>>> rate(counter[2m]).  That's because rate() needs at least two data points 
>>>>> to 
>>>>> fall within the range window.  I would be surprised if you see any graph 
>>>>> at 
>>>>> all with rate(counter[1m]).
>>>>>
>>>>> > This is the raw data, as obtained through a request to /api/v1/query
>>>>>
>>>>> What is the *exact* query you gave? Hopefully it is a range vector 
>>>>> query, like counter[15m].  A range vector expression sent to the simple 
>>>>> query endpoint gives you the raw data points with their raw timestamps 
>>>>> from 
>>>>> the database.
>>>>>
>>>>> > and then we configure the minimum value of it to 1m per-graph
>>>>>
>>>>> Just in case you haven't realised: to set a minimum value of 1m, you 
>>>>> must set the data source scrape interval (in Grafana) to 15s - since 
>>>>> Grafana clamps the minimum value to 4 x Grafana-configured data source 
>>>>> scrape interval.
>>>>>
>>>>> Therefore if you are actually scraping at 1m intervals, and you want 
>>>>> the minimum of $__rate_interval to be 2m, then you must set the Grafana 
>>>>> data source interval to 30s.  This is weird, but it is what it is.
>>>>> https://github.com/grafana/grafana/issues/32169
>>>>>
>>>>> On Wednesday, 6 April 2022 at 14:07:13 UTC+1 [email protected] wrote:
>>>>>
>>>>>> We do make use of that variable, and then we configure the minimum 
>>>>>> value of it to 1m per-graph. I didn't realise you could configure this 
>>>>>> per-datasource, thanks for pointing that out!
>>>>>>
>>>>>> We did used to scrape at 15s intervals but we're using AWS's managed 
>>>>>> prometheus workspaces, and each data point costs money, so we brought it 
>>>>>> down to 1m intervals.
>>>>>>
>>>>>> I'm not sure I understand the relationship between scrape interval 
>>>>>> and counter resets, especially considering there doesn't appear to be a 
>>>>>> counter reset in the raw data of the time series in question.
>>>>>>
>>>>>> You mentioned "true counter reset", does prometheus have some 
>>>>>> internal distinction between types of counter reset?
>>>>>>
>>>>>> On Wednesday, April 6, 2022 at 2:03:40 PM UTC+1 [email protected] 
>>>>>> wrote:
>>>>>>
>>>>>>> I would recommend using the `$__rate_interval` magic variable in 
>>>>>>> Grafana. Note that Grafana assumes a default interval of 15s in the 
>>>>>>> datasource settings.
>>>>>>>
>>>>>>> If your data is mostly 60s scrape intervals, you can configure this 
>>>>>>> setting in the Grafana datasource settings.
>>>>>>>
>>>>>>> If you want to be able to view 1m resolution rates, I recommend 
>>>>>>> increasing your scrape interval to 15s. This makes sure you have 
>>>>>>> several 
>>>>>>> samples in the rate window. This helps Prometheus better handle true 
>>>>>>> counter resets and lost scrapes.
>>>>>>>
>>>>>>> On Wed, Apr 6, 2022 at 2:56 PM Sam Rose <[email protected]> wrote:
>>>>>>>
>>>>>>>> Thanks for the heads up! We've flip flopped a bit between using 1m 
>>>>>>>> or 2m. 1m seems to work reliably enough to be useful in most 
>>>>>>>> situations, 
>>>>>>>> but I'll probably end up going back to 2m after this discussion.
>>>>>>>>
>>>>>>>> I don't believe that helps with the reset problem though, right? I 
>>>>>>>> retried the queries using 2m instead of 1m and they still exhibit the 
>>>>>>>> same 
>>>>>>>> problem.
>>>>>>>>
>>>>>>>> Is there any more data I can get you to help debug the problem? We 
>>>>>>>> see this happen multiple times per day, and it's making it difficult 
>>>>>>>> to 
>>>>>>>> monitor our systems in production.
>>>>>>>>
>>>>>>>> On Wednesday, April 6, 2022 at 1:53:26 PM UTC+1 [email protected] 
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Yup, PromQL thinks there's a small dip in the data. I'm not sure 
>>>>>>>>> why tho. I took your raw values:
>>>>>>>>>
>>>>>>>>> 225201
>>>>>>>>> 225226
>>>>>>>>> 225249
>>>>>>>>> 225262
>>>>>>>>> 225278
>>>>>>>>> 225310
>>>>>>>>> 225329
>>>>>>>>> 225363
>>>>>>>>> 225402
>>>>>>>>> 225437
>>>>>>>>> 225466
>>>>>>>>> 225492
>>>>>>>>> 225529
>>>>>>>>> 225555
>>>>>>>>> 225595
>>>>>>>>>
>>>>>>>>> $ awk '{print $1-225201}' values
>>>>>>>>> 0
>>>>>>>>> 25
>>>>>>>>> 48
>>>>>>>>> 61
>>>>>>>>> 77
>>>>>>>>> 109
>>>>>>>>> 128
>>>>>>>>> 162
>>>>>>>>> 201
>>>>>>>>> 236
>>>>>>>>> 265
>>>>>>>>> 291
>>>>>>>>> 328
>>>>>>>>> 354
>>>>>>>>> 394
>>>>>>>>>
>>>>>>>>> I'm not seeing the reset there.
>>>>>>>>>
>>>>>>>>> One thing I noticed, your data interval is 60 seconds and you are 
>>>>>>>>> doing a rate(counter[1m]). This is not going to work reliably, 
>>>>>>>>> because you 
>>>>>>>>> are likely to not have two samples in the same step window. This is 
>>>>>>>>> because 
>>>>>>>>> Prometheus uses millisecond timestamps, so if you have timestamps at 
>>>>>>>>> these 
>>>>>>>>> times:
>>>>>>>>>
>>>>>>>>> 5.335
>>>>>>>>> 65.335
>>>>>>>>> 125.335
>>>>>>>>>
>>>>>>>>> Then you do a rate(counter[1m]) at time 120 (Grafana attempts to 
>>>>>>>>> align queries to even minutes for consistency), the only sample 
>>>>>>>>> you'll get 
>>>>>>>>> back is 65.335.
>>>>>>>>>
>>>>>>>>> You need to do rate(counter[2m]) in order to avoid problems.
>>>>>>>>>
>>>>>>>>> On Wed, Apr 6, 2022 at 2:45 PM Sam Rose <[email protected]> wrote:
>>>>>>>>>
>>>>>>>>>> I just learned about the resets() function and applying it does 
>>>>>>>>>> seem to show that a reset occurred:
>>>>>>>>>>
>>>>>>>>>> {
>>>>>>>>>>   "request": {
>>>>>>>>>>     "url": 
>>>>>>>>>> "api/datasources/proxy/1/api/v1/query_range?query=resets(counter[1m])&start=1649239200&end=1649240100&step=60",
>>>>>>>>>>     "method": "GET",
>>>>>>>>>>     "hideFromInspector": false
>>>>>>>>>>   },
>>>>>>>>>>   "response": {
>>>>>>>>>>     "status": "success",
>>>>>>>>>>     "data": {
>>>>>>>>>>       "resultType": "matrix",
>>>>>>>>>>       "result": [
>>>>>>>>>>         {
>>>>>>>>>>           "metric": {/* redacted */},
>>>>>>>>>>           "values": [
>>>>>>>>>>             [
>>>>>>>>>>               1649239200,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239260,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239320,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239380,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239440,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239500,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239560,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239620,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239680,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239740,
>>>>>>>>>>               "1"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239800,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239860,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239920,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649239980,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649240040,
>>>>>>>>>>               "0"
>>>>>>>>>>             ],
>>>>>>>>>>             [
>>>>>>>>>>               1649240100,
>>>>>>>>>>               "0"
>>>>>>>>>>             ]
>>>>>>>>>>           ]
>>>>>>>>>>         }
>>>>>>>>>>       ]
>>>>>>>>>>     }
>>>>>>>>>>   }
>>>>>>>>>> }
>>>>>>>>>>
>>>>>>>>>> I don't quite understand how, though.
>>>>>>>>>> On Wednesday, April 6, 2022 at 1:40:12 PM UTC+1 Sam Rose wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi there,
>>>>>>>>>>>
>>>>>>>>>>> We're seeing really large spikes when using the `rate()` 
>>>>>>>>>>> function on some of our metrics. I've been able to isolate a single 
>>>>>>>>>>> time 
>>>>>>>>>>> series that displays this problem, which I'm going to call 
>>>>>>>>>>> `counter`. I 
>>>>>>>>>>> haven't attached the actual metric labels here, but all of the data 
>>>>>>>>>>> you see 
>>>>>>>>>>> here is from `counter` over the same time period.
>>>>>>>>>>>
>>>>>>>>>>> This is the raw data, as obtained through a request to 
>>>>>>>>>>> /api/v1/query:
>>>>>>>>>>>
>>>>>>>>>>> {
>>>>>>>>>>>     "data": {
>>>>>>>>>>>         "result": [
>>>>>>>>>>>             {
>>>>>>>>>>>                 "metric": {/* redacted */},
>>>>>>>>>>>                 "values": [
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239253.4,
>>>>>>>>>>>                         "225201"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239313.4,
>>>>>>>>>>>                         "225226"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239373.4,
>>>>>>>>>>>                         "225249"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239433.4,
>>>>>>>>>>>                         "225262"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239493.4,
>>>>>>>>>>>                         "225278"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239553.4,
>>>>>>>>>>>                         "225310"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239613.4,
>>>>>>>>>>>                         "225329"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239673.4,
>>>>>>>>>>>                         "225363"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239733.4,
>>>>>>>>>>>                         "225402"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239793.4,
>>>>>>>>>>>                         "225437"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239853.4,
>>>>>>>>>>>                         "225466"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239913.4,
>>>>>>>>>>>                         "225492"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649239973.4,
>>>>>>>>>>>                         "225529"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649240033.4,
>>>>>>>>>>>                         "225555"
>>>>>>>>>>>                     ],
>>>>>>>>>>>                     [
>>>>>>>>>>>                         1649240093.4,
>>>>>>>>>>>                         "225595"
>>>>>>>>>>>                     ]
>>>>>>>>>>>                 ]
>>>>>>>>>>>             }
>>>>>>>>>>>         ],
>>>>>>>>>>>         "resultType": "matrix"
>>>>>>>>>>>     },
>>>>>>>>>>>     "status": "success"
>>>>>>>>>>> }
>>>>>>>>>>>
>>>>>>>>>>> The next query is taken from the Grafana query inspector, 
>>>>>>>>>>> because for reasons I don't understand I can't get Prometheus to 
>>>>>>>>>>> give me 
>>>>>>>>>>> any data when I issue the same query to /api/v1/query_range. The 
>>>>>>>>>>> query is 
>>>>>>>>>>> the same as the above query, but wrapped in a rate([1m]):
>>>>>>>>>>>
>>>>>>>>>>>     "request": {
>>>>>>>>>>>         "url": 
>>>>>>>>>>> "api/datasources/proxy/1/api/v1/query_range?query=rate(counter[1m])&start=1649239200&end=1649240100&step=60",
>>>>>>>>>>>         "method": "GET",
>>>>>>>>>>>         "hideFromInspector": false
>>>>>>>>>>>     },
>>>>>>>>>>>     "response": {
>>>>>>>>>>>         "status": "success",
>>>>>>>>>>>         "data": {
>>>>>>>>>>>             "resultType": "matrix",
>>>>>>>>>>>             "result": [
>>>>>>>>>>>                 {
>>>>>>>>>>>                     "metric": {/* redacted */},
>>>>>>>>>>>                     "values": [
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239200,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239260,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239320,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239380,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239440,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239500,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239560,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239620,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239680,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239740,
>>>>>>>>>>>                             "9391.766666666665"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239800,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239860,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239920,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649239980,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649240040,
>>>>>>>>>>>                             "0.03333333333333333"
>>>>>>>>>>>                         ],
>>>>>>>>>>>                         [
>>>>>>>>>>>                             1649240100,
>>>>>>>>>>>                             "0"
>>>>>>>>>>>                         ]
>>>>>>>>>>>                     ]
>>>>>>>>>>>                 }
>>>>>>>>>>>             ]
>>>>>>>>>>>         }
>>>>>>>>>>>     }
>>>>>>>>>>> }
>>>>>>>>>>>
>>>>>>>>>>> Given the gradual increase in the underlying counter, I have two 
>>>>>>>>>>> questions:
>>>>>>>>>>>
>>>>>>>>>>> 1. How come the rate is 0 for all except 2 datapoints?
>>>>>>>>>>> 2. How come there is one enormous datapoint in the rate query, 
>>>>>>>>>>> that is seemingly unexplained in the raw data?
>>>>>>>>>>>
>>>>>>>>>>> For 2 I've seen in other threads that the explanation is an 
>>>>>>>>>>> unintentional counter reset, caused by scrapes a millisecond apart 
>>>>>>>>>>> that 
>>>>>>>>>>> make the counter appear to go down for a single scrape interval. I 
>>>>>>>>>>> don't 
>>>>>>>>>>> think I see this in our raw data, though.
>>>>>>>>>>>
>>>>>>>>>>> We're using Prometheus version 2.26.0, revision 
>>>>>>>>>>> 3cafc58827d1ebd1a67749f88be4218f0bab3d8d, go version go1.16.2.
>>>>>>>>>>>
>>>>>>>>>> -- 
>>>>>>>>>> You received this message because you are subscribed to the 
>>>>>>>>>> Google Groups "Prometheus Users" group.
>>>>>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>>>>>> send an email to [email protected].
>>>>>>>>>> To view this discussion on the web visit 
>>>>>>>>>> https://groups.google.com/d/msgid/prometheus-users/c1b7568b-f7f9-4edc-943a-22412658975fn%40googlegroups.com
>>>>>>>>>>  
>>>>>>>>>> <https://groups.google.com/d/msgid/prometheus-users/c1b7568b-f7f9-4edc-943a-22412658975fn%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>>>>> .
>>>>>>>>>>
>>>>>>>>> -- 
>>>>>>>> You received this message because you are subscribed to the Google 
>>>>>>>> Groups "Prometheus Users" group.
>>>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>>>> send an email to [email protected].
>>>>>>>>
>>>>>>> To view this discussion on the web visit 
>>>>>>>> https://groups.google.com/d/msgid/prometheus-users/888affc4-e9ba-4ea8-8a40-c7b7a17affe4n%40googlegroups.com
>>>>>>>>  
>>>>>>>> <https://groups.google.com/d/msgid/prometheus-users/888affc4-e9ba-4ea8-8a40-c7b7a17affe4n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>>>> .
>>>>>>>>
>>>>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/5b7f1f53-1ea4-4c60-b05e-76396a370a46n%40googlegroups.com.

Reply via email to