Single-record type merging
This example explores the core techniques for merging typed objects across subschemas using single-record queries, covering most of the topics discussed in the documented basic example.
Using single-record queries means that every record accessed requires a dedicated subschema delegation, versus a single delegation for an entire set of records. This 1:1 delegation strategy has far greater overhead than the array-batched technique discussed in the merged arrays chapter. However—it’s still useful when interacting with services beyond our control, and it’s an excellent way to observe the behavior of execution batching.
This example demonstrates:
- One-way type merging using single-record queries.
- Multi-directional type merging using single-record queries.
- Query/execution batching.
Related examples:
- See array-batched type merging to improve the performance of these basic patterns.
- See stitching directives SDL to write this merge configuration as schema directives.
Sandbox
⬇️ Click ☰ to see the files
You can also see the project on GitHub here.
The following services are available for interactive queries:
- Stitched gateway: listening on 4000/graphql
- Manufacturers subservice: listening on 4001/graphql
- Products subservice: listening on 4002/graphql
- Storefronts subservice: listening on 4003/graphql
Summary
Visit the stitched gateway and try running the following query:
query {
storefront(id: "2") {
id
name
products {
upc
name
manufacturer {
products {
upc
name
}
name
}
}
}
}
If you study the results of this query, the final composition traverses across the service graph:
Storefront
(Storefronts schema)Storefront.products -> Product
(Products schema)Product.manufacturer -> Manufacturer
(Products + Manufacturers schemas)Manufacturer.products
(Products schema)Manufacturer.name
(Manufacturers schema)
That means the gateway performed three rounds of resolution for each service’s data
(Services -> Products -> Manufacturers
).
Batching
While the gateway performed three rounds of resolution, it actually had to perform a single subschema delegation (or, proxy) per record in each round because we’re only fetching one record at a time. This is both expensive for the gateway to process, and for the subservice to fulfill. Thankfully, stitching has a built-in solution for the later inefficiency. Clear your gateway terminal log and run the following query in GraphiQL:
query {
storefront(id: "2") {
products {
upc
name
}
}
}
The Products executor is setup to log all of its operations, so you can watch what’s being requested
in the gateway terminal window. Find these lines in the root index.js
file and try adjusting the
batch
setting:
{
schema: await schemaFromExecutor(productsExec),
executor: buildHTTPExecutor({
url: 'http://localhost:4002/graphql',
}),
batch: true, // << try turning this on and off
}
With batch: false
, the above query logs the following:
# -- OPERATION 2020-12-03T15:32:41.900Z:
query ($_v0_upc: ID!) {
product(upc: $_v0_upc) {
__typename
name
upc
}
}
# -- OPERATION 2020-12-03T15:32:41.901Z:
query ($_v0_upc: ID!) {
product(upc: $_v0_upc) {
__typename
name
upc
}
}
# -- OPERATION 2020-12-03T15:32:41.903Z:
query ($_v0_upc: ID!) {
product(upc: $_v0_upc) {
__typename
name
upc
}
}
Notice that we’re sending three seperate requests to the Products service to resolve each record
in the product set. That’s extremely inefficient. Switching to batch: true
changes the output to:
# -- OPERATION 2020-12-03T15:36:22.889Z:
query ($graphqlTools0__v0_upc: ID!, $graphqlTools1__v0_upc: ID!, $graphqlTools2__v0_upc: ID!) {
graphqlTools0_product: product(upc: $graphqlTools0__v0_upc) {
__typename
name
upc
}
graphqlTools1_product: product(upc: $graphqlTools1__v0_upc) {
__typename
name
upc
}
graphqlTools2_product: product(upc: $graphqlTools2__v0_upc) {
__typename
name
upc
}
}
Now we’re sending a single request that resolves all three single-record queries at once, courtesy of batched execution. There are very few reasons NOT to enable this free batching optimization (it will be enabled by default in the future).
Batch execution is superb for optimizing the exchange with the subservices. However, there are still overhead processing costs on the gateway for delegating each record individually, so batch execution alone is not a perfect solution. The best optimization strategy is to pair batch execution with delegating arrays of records at a time.