There are two general threat vectors for privacy/security with "shared
state" (cache, connections, etc).

1 - Explicit user tracking across site boundaries. Either through unique
responses or fingerprinting. This is effectively moot since the tracking
can be done directly by cookies when 3PCs are allowed. This proposal isn't
aimed at this use case where a cooperating resource can already explicitly
track you (and is why the feature is linked to the state of 3PC access).
Additional bits of entropy aren't a concern since it is assumed you have
direct access to sharing ID's.

2 - XS leaks (history sniffing). This is where I can tell that you bank at
BofA because you have the JS loaded that is only used when logged-in, I can
tell that you have a gmail account because of similar resource exposure and
I can tell that you likely live in or visited Groom Lake, Nevada because
the map tiles from Google maps for that location are in your cache.

I think keeping the partitioned cache is still important to solve #2 and
the proposal is all about how to find the sweet spot where we can share a
cache for the resources that a given user is actually likely to benefit
from being shared while limiting the exposure on the information that can
be gathered about the browser. It is essentially reduced to "this browser
likely visited a page that had ads and may have visited a storefront that
was hosted on the shopify platform".

I say "may" because even though the resources are picked to limit the
usefulness of the information that can be extrapolated, we still add a
bunch of protections to make probing for that state destructive and
difficult:
- Any attempt to check a cache entry will write the cache entry (so you
can't tell if it was there naturally or from someone else probing).
- Any attempt to probe for multiple versions of a resource that match the
same pattern will fall back to using the partitioned cache.

I'm happy to explore if you think there are other options to solve for #2
while unlocking more of the shared cache but this is my best attempt at
clawing back the real-world benefits of the single-keyed cache while
keeping the privacy and security benefits that come with the
fully-partitioned cache.

On Wed, Oct 29, 2025 at 10:08 PM Erik Anderson <[email protected]>
wrote:

> My understanding was that there was believed to be a meaningful security
> benefit with partitioning the cache. That’s because it would limit a party
> from being able to inferr that you’ve visited some other site by measuring
> a side effect tied to how quickly a resource loads. That observation could
> potentially be made even if that specific adversary doesn’t have any of
> their own content loaded on the other site.
>
>
>
> Of course, if there is an entity with a resource loaded across both sites
> with a 3p cookie *and* they’re willing to share that info/collude,
> there’s not much benefit. And even when partitioned, if 3p cookies are
> enabled, there are potentially measurable side effects that differ based on
> if the resource request had some specific state in a 3p cookie.
>
>
>
> Does that incremental security benefit of partitioning the cache justify
> the performance costs when 3p cookies are still enabled? I’m not sure.
>
>
>
> Even if partitioning was eliminated, a site could protect themselves a bit
> by specifying Vary: Origin, but that probably doesn’t sufficiently cover
> iframe scenarios (nor would I expect most sites to hold it right).
>
>
>
> *From:* Rick Byers <[email protected]>
> *Sent:* Wednesday, October 29, 2025 11:56 AM
> *To:* Patrick Meenan <[email protected]>
> *Cc:* Mike Taylor <[email protected]>; blink-dev <
> [email protected]>
> *Subject:* [EXTERNAL] Re: [blink-dev] Intent to ship: Cache sharing for
> extremely-pervasive resources
>
>
>
> If this is enabled only when 3PCs are enabled, then what are the tradeoffs
> of going through all this complexity and governance vs. just broadly
> coupling HTTP cache keying behavior to 3PC status in some way? What can a
> tracker credibly do with a single-keyed HTTP cache that they cannot do with
> 3PCs? Are there also concerns about accidental cross-site resource sharing
> which could be mitigated more simply by other means, eg. by scoping to just
> to ETag-based caching?
>
>
>
> I remember the controversy and some real evidence of harm to users and
> businesses in 2020 when we partitioned the HTTP cache, but I was convinced
> that we had to accept that harm in order to credibly achieve 3PCD. At the
> time I was personally a fan of a proposal like this (even for users without
> 3PCs) in order to mitigate the harm. But now it seems to me that if we're
> going to start talking about poking holes in that decision, perhaps we
> should be doing a larger review of the options in that space with the
> knowledge that most Chrome users are likely to continue to have 3PCs
> enabled. WDYT?
>
>
>
> Thanks,
>
>    Rick
>
>
>
> On Mon, Oct 27, 2025 at 10:27 AM Patrick Meenan <[email protected]>
> wrote:
>
> I don't believe the security/privacy protections actually rely on the
> assertions (and it's unlikely those would be public). It's more for
> awareness and to make sure they don't accidentally break something with
> their app if they were relying on the responses being partitioned by site.
>
>
>
> As far as query params go, the browser code already only filters for
> requests with no query params so any that do rely on query params won't get
> included anyway.
>
>
>
> The same goes for cookies. Since the feature is only enabled when
> third-party cookies are enabled, adding cookies to these responses or
> putting unique content in them won't actually pierce any new boundaries but
> it goes against the intent of only using it for public/static resources and
> they'd lose the benefit of the shared cache when it gets updated. Same goes
> for the fingerprinting risks if the pattern was abused.
>
>
>
> On Mon, Oct 27, 2025 at 9:39 AM Mike Taylor <[email protected]>
> wrote:
>
> On 10/22/25 5:48 p.m., Patrick Meenan wrote:
>
> The candidate list goes down to 20k occurrences in order to catch
> resources that were updated mid-crawl and may have multiple entries with
> different hashes that add up to 100k+ occurrences. In the candidate list,
> without any filtering, the 100k cutoff is around 600, I'd estimate that
> well less than 25% of the candidates make it through the filtering for
> stable pattern, correct resource type and reliable pattern. First release
> will likely be 100-200 and I don't expect it will ever grow above 500.
>
> Thanks - I see the living document has been updated to mention 500 as a
> ceiling.
>
>
>
> As far as cadence goes, I expect there will be a lot of activity for the
> next few releases as individual patterns are coordinated with the origin
> owners but then it will settle down to a much more bursty pattern of
> updates every few Chrome releases (likely linked with an origin changing
> their application and adding more/different resources). And yes, it is
> manual.
>
> As far as the process goes, resource owners need to actively assert that
> their resource is appropriate for the single-keyed cache and that they
> would like it included (usually in response to active outreach from us but
> we have the external-facing list for owner-initiated contact as well).  The
> design doc has the documentation for what it means to be appropriate (and
> the doc will be moved to a readme page in the repository next to the actual
> list so it's not a hard-to-find Google doc):
>
> Will there be any kind of public record of this assertion? What happens if
> a site starts using query params or sending cookies? Does the person in
> charge of manual list curation discover that in the next release? Does that
> require a new release (I don't know if this lives in component updater, or
> in the binary itself)?
>
>
>
> *5. Require resource owner opt-in*
> For each URL to be included, reach out to the team/company responsible for
> the resource to validate the URL pattern and get assurances that the
> pattern will always serve the same content to all sites and not be abused
> for tracking (by using unique URLs within the pattern mask as a bit-mask
> for fingerprinting). They will also need to validate that the URLs covered
> by the pattern will not rely on being able to set cookies over HTTP using a
> Set-Cookie HTTP response header because they will not be re-applied
> across cache boundaries (the set-cookie is not cached with the resource).
>
>
>
> On Wed, Oct 22, 2025 at 5:31 PM Mike Taylor <[email protected]>
> wrote:
>
> On 10/18/25 8:34 a.m., Patrick Meenan wrote:
>
> Sorry, I missed a step in making the candidate resource list public. I
> have moved it to my chromium account and made it public here
> <https://docs.google.com/spreadsheets/d/1TgWhdeqKbGm6hLM9WqnnXLn-iiO4Y9HTjDXjVO2aBqI/edit?usp=sharing>.
>
>
>
>
> Not everything in that list meets all of the criteria - it's just the
> first step in the manual curation (same URL served the same content across
> > 20k sites in the HTTP Archive dataset).
>
>
>
> The manual steps frome there for meeting the criteria are basically:
>
>
>
> - Cull the list for scripts, stylesheets and compression dictionaries.
>
> - Remove any URLs that use query parameters.
>
> - Exclude any responses that set cookies.
>
> - Identify URLs that are not manually versioned by site embedders (i.e.
> the embedded resource can not get stale). This is either in-place updating
> resources or automatically versioned resources.
>
> - Only include URLs that can reliably target a single resource by pattern
> (i.e. ..../<hash>-common.js but not ..../<hash>.js)
>
> - Get confirmation from the resource owner that the given URL Pattern is
> and will continue to be appropriate for the single-keyed cache
>
> A few questions on list curation:
>
> Can you clarify how big the list will be? The privacy review at
> https://chromestatus.com/feature/5202380930678784?gate=5174931459145728 
> mentions
> ~500, while the design doc mentions 1000. I see the candidate resource list
> starts at ~5000, then presumably manual curation begins to get to one of
> those numbers.
>
> What is the expected list curation/update cadence? Is it actually manual?
>
> Is there any recourse process for owners of resources that don't want to
> be included? Do we have documentation on what it mean to be appropriate for
> the single-keyed cache?
>
> thanks,
> Mike
>
> --
> You received this message because you are subscribed to the Google Groups
> "blink-dev" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion visit
> https://groups.google.com/a/chromium.org/d/msgid/blink-dev/CAPq58w6UFSnxxzhGKBnY1BJKiZZeH7BUm7PmcjQm_%2BLjGyrtYg%40mail.gmail.com
> <https://groups.google.com/a/chromium.org/d/msgid/blink-dev/CAPq58w6UFSnxxzhGKBnY1BJKiZZeH7BUm7PmcjQm_%2BLjGyrtYg%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>
> --
> You received this message because you are subscribed to the Google Groups
> "blink-dev" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion visit
> https://groups.google.com/a/chromium.org/d/msgid/blink-dev/CAFUtAY9Nffq00r-xbiu2BO00y%2B_2knAi-zheMs9hrE-dB%2BTZ3w%40mail.gmail.com
> <https://groups.google.com/a/chromium.org/d/msgid/blink-dev/CAFUtAY9Nffq00r-xbiu2BO00y%2B_2knAi-zheMs9hrE-dB%2BTZ3w%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"blink-dev" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/a/chromium.org/d/msgid/blink-dev/CAPq58w6jhL6Rx4hpLDU3n-p_%3DDb_wLdHAM%3DGTUTYk5iJterFow%40mail.gmail.com.

Reply via email to