Docs
Launch GraphOS Studio

Advanced topics on caching in Apollo Client


This article describes special cases and considerations when using the

.

Bypassing the cache

Sometimes you shouldn't use the cache for a particular . For example, a 's response might be a token that's only used once. In cases like this, use

:

const { loading, error, data } = useQuery(GET_DOGS, {
fetchPolicy: "no-cache"
});

that use this fetch policy don't write their result to the cache, and they also don't check the cache for data before sending a request to your server.

Persisting the cache

You can persist and rehydrate the InMemoryCache from a storage provider like AsyncStorage or localStorage. To do so, use the

library. This library works with a variety of
storage providers
.

To get started, pass your cache and a storage provider to persistCache. By default, the contents of your cache are immediately restored asynchronously, and they're persisted on every write to the cache with a short configurable debounce interval.

NOTE

The persistCache method is async and returns a Promise.

import AsyncStorage from '@react-native-async-storage/async-storage';
import { InMemoryCache } from '@apollo/client';
import { persistCache } from 'apollo3-cache-persist';
const cache = new InMemoryCache();
persistCache({
cache,
storage: AsyncStorage,
}).then(() => {
// Continue setting up Apollo Client as usual.
})

For advanced usage and additional configuration options, see the

.

Resetting the cache

Sometimes, you might want to reset the cache entirely, such as

. To accomplish this, call client.resetStore. This method is asynchronous, because it also refetches any of your active queries.

import { useQuery } from '@apollo/client';
function Profile() {
const { data, client } = useQuery(PROFILE_QUERY);
return (
<Fragment>
<p>Current user: {data?.currentUser}</p>
<button onClick={async ()=>client.resetStore()}>
Log out
</button>
</Fragment>
);
}

To reset the cache without refetching active queries, use client.clearStore() instead of client.resetStore().

Responding to cache resets

You can register callback functions that execute whenever client.resetStore is called. To do so, call client.onResetStore and pass in your callback. To register multiple callbacks, call client.onResetStore multiple times. All of your callbacks are added to an array and are executed concurrently whenever the cache is reset.

In this example, we use client.onResetStore to write default values to the cache. This is useful when using 's

features and calling client.resetStore anywhere in your application.

import { ApolloClient, InMemoryCache } from '@apollo/client';
import { withClientState } from 'apollo-link-state';
import { resolvers, defaults } from './resolvers';
const cache = new InMemoryCache();
const stateLink = withClientState({ cache, resolvers, defaults });
const client = new ApolloClient({
cache,
link: stateLink,
});
client.onResetStore(stateLink.writeDefaults);

You can also call client.onResetStore from your React components. This can be useful if you want to force your UI to rerender after the cache is reset.

The client.onResetStore method's return value is a function you can call to unregister your callback:

import { useApolloClient } from '@apollo/client';
function Foo (){
const [reset, setReset] = useState(0);
const client = useApolloClient();
useEffect(() => {
const unsubscribe = client.onResetStore(() =>
new Promise(()=>setReset(reset + 1))
);
return () => {
unsubscribe();
};
});
return reset ? <div /> : <span />
}
export default Foo;

TypePolicy inheritence

JavaScript developers will be familiar with the idea of

from the extends clause of class declarations, or possibly from dealing with prototype chains created by Object.create.

Inheritance is a powerful code-sharing tool, and it works well with Apollo Client for several reasons:

  • InMemoryCache already knows about the supertype-subtype relationships (interfaces and unions) in your schema, thanks to possibleTypes, so no additional configuration is necessary to provide that information.

  • Inheritance allows a supertype to provide default configuration values to all its subtypes, including keyFields and individual policies, which can be selectively overridden by subtypes that want something different.

  • A single subtype can have multiple supertypes in a , which is difficult to model using the single inheritance model of classes or prototypes. In other words, supporting multiple inheritance in JavaScript requires building a system something like this one, rather than just reusing built-in language features.

  • Developers can add their own client-only supertypes to the possibleTypes map, as a way of reusing behavior across types, even if their schema knows nothing about those supertypes.

  • The possibleTypes map is currently used only for matching purposes, which is an important but fairly small part of what the client does. Inheritance adds another compelling use for possibleTypes, and should drastically reduce repetition of typePolicies when used effectively.

Here's how type policy inheritance works for InMemoryCache, considering the example below:

const cache = new InMemoryCache({
possibleTypes: {
Reptile: ["Snake", "Turtle"],
Snake: ["Python", "Viper", "Cobra"],
Viper: ["Cottonmouth", "DeathAdder"],
},
typePolicies: {
Reptile: {
// Suppose all our reptiles are captive, and have a tag with an ID.
keyFields: ["tagId"],
fields: {
// Scientific name-related logic can be shared among Reptile subtypes.
scientificName: {
merge(_, incoming) {
// Normalize all scientific names to lower case.
return incoming.toLowerCase();
},
},
},
},
Snake: {
fields: {
// Default to a truthy non-boolean value if we don't know
// whether this snake is venomous.
venomous(status = "unknown") {
return status;
},
},
},
},
});

Refetching queries after a mutation

In certain cases, writing an update function to

can be complex, or even impossible if the doesn't return modified .

In these cases, you can provide a refetchQueries option to the useMutation hook to automatically rerun certain queries after the mutation completes.

For details, see

.

NOTE

Although refetchQueries can be faster to implement than an update function, it also requires additional network requests that are usually undesirable. For more information, see

.

Cache redirects

In some cases, a query requests data that already exists in the cache under a different reference. For example, your UI might have a list view and a detail view that both use the same data.

The list view might run the following query:

query Books {
books {
id
title
abstract
}
}

When a specific book is selected, the detail view might display an individual item using this query:

query Book($id: ID!) {
book(id: $id) {
id
title
abstract
}
}

In a case like this, we know that the second query's data might already be in the cache, but because that data was fetched by a different query, Apollo Client doesn't know that. To tell Apollo Client where to look for the cached Book object, we can define a field policy read function for the book field:

import { ApolloClient, InMemoryCache } from '@apollo/client';
const client = new ApolloClient({
cache: new InMemoryCache({
typePolicies: {
Query: {
fields: {
book: {
read(_, { args, toReference }) {
return toReference({
__typename: 'Book',
id: args.id,
});
}
}
}
}
}
})
});

This read function uses the toReference helper utility to generate and return a cache reference for a Book object, based on its __typename and id.

Now whenever a query includes the book field, the read function above executes and returns a reference to a Book object. Apollo Client uses this reference to look up the object in its cache and return it if it's present. If it isn't present, Apollo Client knows it needs to execute the query over the network.

NOTE

To avoid a network request, all of a query's requested fields must already be present in the cache. If the detail view's query fetches any Book field that the list view's query didn't, Apollo Client considers the cache hit to be incomplete, and it executes the full query over the network.

Pagination utilities

Pagination is a best practice in GraphQL

. Apollo Client enables fetching and caching paginated results using the
Core pagination API
. The API includes a few important utilities, including the
fetchMore
function and the @connection .

Incremental loading: fetchMore

You can use the fetchMore function to update a query's cached result with data returned by a follow-up query. Most often, fetchMore is used to handle infinite-scroll pagination and other situations where you're loading more data when you already have some.

For details, see

.

The @connection directive

The @connection directive solves the problem of multiple copies of the same field in the cache. This can happen with paginated queries because the fetchMore function sends follow-up queries to fetch additional pages of results using like

and
limit
. These arguments inadvertently fragment data from different pagination requests across the cache.

The @connection directive lets you unify paginated results by specifying a custom, stable cache key for a field. It also lets you intentionally separate paginated results in the cache by fields that you specify.

💡 TIP

Starting in Apollo Client v3, setting the

is the most straightforward way to resolve pagination results in the cache. For example, setting
keyArgs to false
indicates that no arguments should be included in cache keys, causing all pagination results to be cached together. Additionally, you only have to set your keyArgs configuration once, rather than using @connection in multiple queries. Refer to the
usage instructions
below to compare @connection and keyArgs usage.

The @connection directive is useful when you want to store distinct data in the cache on a query-by-query, field-by-field basis. See the

for more details.

@connection directive usage

💡 TIP

For the standard @connection directive usage described in this section, it's best to configure a

. For example, you can use the following
keyArgs
configuration for the same effect as the @connection example below.

const cache = new InMemoryCache({
typePolicies: {
Query: {
fields: {
feed: {
keyArgs: ["type"]
}
}
}
}
})

With this centralized keyArgs configuration, you don't need to include the @connection directive in your queries because the type is adequate for keeping feeds of different types separate in the cache. For an example of storing distinct data on a query-by-query basis, see the

.

To use the @connection directive, add it to the field you want a custom cache key for. The directive requires a key parameter to specify the custom cache key. You can optionally include the filter parameter, which takes an array of query argument names to include in the generated custom cache key.

const query = gql`
query Feed($type: FeedType!, $offset: Int, $limit: Int) {
feed(type: $type, offset: $offset, limit: $limit) @connection(key: "feed", filter: ["type"]) {
...FeedEntry
}
}
`

With the above query, even when multiple fetchMores queries are performed, each feed update always results in an update to the cache's feed key with the latest accumulated values. The example also uses the @connection directive's optional filter argument to include the type query argument in the cache key. This creates multiple cache values that accumulate queries from each type of feed.

With a stable cache key, you can use

to perform a cache update that clears out the feed.

client.writeQuery({
query: gql`
query Feed($type: FeedType!) {
feed(type: $type) @connection(key: "feed", filter: ["type"]) {
id
}
}
`,
variables: {
type: "top",
},
data: {
feed: [],
},
});

NOTE

Because this example uses the type argument in the cache key, it doesn't need to provide offset or limit arguments.

Advanced @connection directive usage

The @connection directive is useful when using the same field in multiple queries, with no distinguishing arguments (for example, type) that keyArgs can use, and you want to keep that field's data separate in the cache.

For example, Apollo's

uses @connection to independently cache lists of playlists. One list is in the left sidebar, where you navigate between playlists. The other appears when you right-click a song to add it to a playlist.

Separately cached playlists in Apollo's Spotify Showcase

Without caching the playlists separately, loading the next page of data from one list affects the other, negatively impacting the UX.

For code examples, see:

Previous
Memory Management
Next
Overview
Edit on GitHubEditForumsDiscord

© 2024 Apollo Graph Inc.

Privacy Policy

Company