Your Web News in One Place

Help Webnuz

Referal links:

Sign up for GreenGeeks web hosting
December 11, 2022 06:50 pm GMT

Seeding the Query Cache

A new RFC about first class support for Promises has been released last week, and it got some talk going about how this would introduce fetch waterfalls if used incorrectly. So what are fetch waterfalls exactly?

Fetch waterfalls

A waterfall describes a situation where one request is made, and we wait for it to complete before firing another request.

Sometimes, this is unavoidable, because the first request contains information that is needed to make the second request. We also refer to these as dependent queries:

component mounts, /user endpoint is being fetched, and after that, we fire off two requests for the user's projects and todos

In many cases though, we can actually fetch all the data we need in parallel, because it is independent data:

component mounts, both /issues and /labels are fetched at the same time

In React Query, we can do that in two different ways:

// 1. Use useQuery twiceconst issues = useQuery({ queryKey: ['issues'], queryFn: fetchIssues })const labels = useQuery({ queryKey: ['labels'], queryFn: fetchLabels })// 2. Use the useQueries hookconst [issues, labels] = useQueries([  { queryKey: ['issues'], queryFn: fetchIssues },  { queryKey: ['labels'], queryFn: fetchLabels },])

In both variants, React Query will kick off data fetching in parallel. So where do waterfalls come in?

Suspense

Note: Please keep in mind that at the time of this writing, suspense for data fetching is still experimental! In the following examples, I will not be using the proposal from the RFC, but rather the suspense implementation that is already available in React Query (which is also experimental).

As described in the above linked RFC, suspense a way to unwrap promises with React. A defining trait of promises is that they can be in three different states: pending, fulfilled or rejected.

When rendering components, we are mostly interested in the success scenario. Handling loading and error states in each and every component can be tedious, and suspense is aimed at solving this problem.

When a promise is pending, React will unmount the component tree and render a fallback defined by a Suspense boundary component. In case of errors, the error is bubbled up to the nearest ErrorBoundary.

This will decouple our components from handling those states, and we can focus on the happy path. It almost acts like synchronous code that just reads a value from a cache:

function Issues() {  //  read data from cache  const { data } = useQuery({    queryKey: ['issues'],    queryFn: fetchIssues,    //  this enables experimental suspense mode    suspense: true,  })  //  no need to handle loading or error states  return (    <div>      {data.map((issue) => (        <div>{issue.title}</div>      ))}    </div>  )}function App() {  //  Boundaries handle loading and error states  return (    <Suspense fallback={<div>Loading...</div>}>      <ErrorBoundary fallback={<div>On no!</div>}>        <Issues />      </ErrorBoundary>    </Suspense>  )}

A note on TypeScript

Sadly, when using TypeScript, data will still be potentially undefined in the above example, as suspense is just a flag on useQuery that can be turned on and off at will. It can also be combined with the enabled option, which will make the query not fire and thus make the component not suspend at all.

We might change this in the future with a dedicated useSuspenseQuery hook.

Suspense waterfalls

So this is nice and all, but it can backfire when you use multiple queries in the same component that have suspense turned on. Here is what happens:

component mounts, /issues are fetched and component suspends. After the request is finished, component mounts and suspends again, triggering a fetch for /labels

  • Component renders, tries to read the first query
  • Sees that there is no data in the cache yet, so it suspends
  • This unmounts the component tree, and renders the fallback
  • When the fetch is finished, the component tree is remounted
  • First query is now read successfully from the cache
  • Component sees the second query, and tries to read it
  • Second query has no data in the cache, so it suspends (again)
  • Second query is fetched
  • Component finally renders successfully

This will have pretty impactful implications on your application's performance, because you'll see that fallback for waaay longer than necessary.

The best way to circumvent this problem is to make sure that there is already data in the cache when the component tries to read it.

Prefetching

The earlier you initiate a fetch, the better, because the sooner it starts, the sooner it can finish.

But even if that's not the case, you can still use prefetchQuery to initiate a fetch before the component is rendered:

const issuesQuery = { queryKey: ['issues'], queryFn: fetchIssues }//  initiate a fetch before the component rendersqueryClient.prefetchQuery(issuesQuery)function Issues() {  const issues = useQuery(issuesQuery)}

The call to prefetchQueries is executed as soon as your JavaScript bundle is evaluated. This works very well if you do route base code splitting, because it means the code for a certain page will be lazily loaded and evaluated as soon as the user navigates to that page.

This means it will still be kicked off before the component renders. If you do this for both queries in our example, you will get those parallel queries back even when using suspense.

prefetch for both queries triggers before the component mounts. component suspends for a short time until both queries re finished.

As we can see, the query will still suspend until both are done fetching, but because we've triggered them in parallel, the waiting time is now drastically reduced.

Note: useQueries doesn't support suspense right now, but it might do in the future. If we add support, the goal is to trigger all fetches in parallel to avoid those waterfalls.

Update: useQueries supports suspense as of v4.15.0.

The use RFC

I don't know enough about the RFC yet to properly comment on it. A big part is still missing, namely how the cache API will work. I do think it is a bit problematic that the default behaviour will lead to waterfalls unless developers explicitly seed the cache early on. I'm still pretty excited about it because it will likely make internals of React Query easier to understand and maintain. It remains to be seen if it is something that will be used in userland a lot.

Seeding details from lists

Another nice way to make sure that your cache is filled by the time it is read is to seed it from other parts of the cache. Oftentimes, if you render a detail view of an item, you will have data for that item readily available if you've previously been on a list view that shows a list of items.

There are two common approaches to fill a detail cache with data from a list cache:

Pull approach

This is the one also described in the docs: When you try to render the detail view, you look up the list cache for the item you want to render. If it is there, you use it as initial data for the detail query.

const useTodo = (id: number) => {  const queryClient = useQueryClient()  return useQuery({    queryKey: ['todos', 'detail', id],    queryFn: () => fetchTodo(id),    initialData: () => {      //  look up the list cache for the item      return queryClient        .getQueryData(['todos', 'list'])        ?.find((todo) => todo.id === id)    },  })}

If the initialData function returns undefined, the query will proceed as normal and fetch the data from the server. And if something is found, it will be put into the cache directly.

Be advised that if you have staleTime set, no further background refetch will occur, as initialData is seen as fresh. This might not be what you want if your list was last fetched twenty minutes ago.

As shown in the docs, we can additionally specify initialDataUpdatedAt on our detail query. It will tell React Query when the data we are passing in as initialData was originally fetched, so it can determine staleness correctly. Conveniently, React Query also knows when the list was last fetched, so we can just pass that in:

const useTodo = (id: number) => {  const queryClient = useQueryClient()  return useQuery({    queryKey: ['todos', 'detail', id],    queryFn: () => fetchTodo(id),    initialData: () => {      return queryClient        .getQueryData(['todos', 'list'])        ?.find((todo) => todo.id === id)    },    initialDataUpdatedAt: () =>      //  get the last fetch time of the list      queryClient.getQueryState(['todos', 'list'])?.dataUpdatedAt,  })}

seeds the cache "just in time"
needs more work to account for staleness

Push approach

Alternatively, you can create detail caches whenever you fetch the list query. This has the advantage that staleness is automatically measured from when the list was fetched, because, well, that's when we create the detail entry.

However, there is no good callback to hook into when a query is fetched. onSuccess on useQuery would technically work, but it would be executed for every instance of useQuery. If we have multiple observers, this would lead to the same data being written to the cache multiple times. The global onSuccess callback on the cache itself might also work, but it would be executed for every query, so we'd have to narrow it down to the right query key.

The best way I've found to execute the push approach is to do it directly in the queryFn, after data has been fetched:

const useTodos = () => {  const queryClient = useQueryClient()  return useQuery({    queryKey: ['todos', 'list'],    queryFn: async () => {      const todos = await fetchTodos()      todos.forEach((todo) => {        //  create a detail cache for each item        queryClient.setQueryData(['todos', 'detail', todo.id], todo)      })      return todos    },  })}

This would create a detail entry for each item in the list immediately. Since there is no one interested in those queries at the moment, those would be seen as inactive, which means they might be garbage collected after cacheTime has elapsed (default: 15 minutes).

So if you use the push approach, the detail entries you've created here might no longer be available once the user actually navigates to the detail view. Also, if your list is long, you might be creating way too many entries that will never be needed.

staleTime is automatically respected
there is no good callback
might create unnecessary cache entries
pushed data might be garbage collected too early

Keep in mind that both approaches only work well if the structure of your detail query is exactly the same (or at least assignable to) the structure of the list query. If the detail view has a mandatory field that doesn't exist in the list, seeding via initialData is not a good idea. This is where placeholderData comes in, and I've written a comparison about the two in #9: Placeholder and Initial Data in React Query.

That's it for today. Feel free to reach out to me on twitter
if you have any questions, or just leave a comment below.


Original Link: https://dev.to/tkdodo/seeding-the-query-cache-5g6g

Share this article:    Share on Facebook
View Full Article

Dev To

An online community for sharing and discovering great ideas, having debates, and making friends

More About this Source Visit Dev To