book/faqs: Copyedit

This commit is contained in:
Zhaofeng Li 2023-01-01 21:10:03 -07:00
parent 49f2ccbff8
commit 2afecd29ad

View file

@ -52,12 +52,12 @@ We use the hash of the _uncompressed NAR_ to perform global deduplication.
└───────►File Size └───────►File Size
``` ```
At first glance, performing compression on the client and deduplicating the result may sound appealing, but has problems: At first glance, performing compression on the client and deduplicating the result may sound appealing, but this approach isn't without problems:
1. Different compression algorithms and levels naturally lead to different results which can't be deduplicated 1. Different compression algorithms and levels naturally lead to different results which can't be deduplicated
2. Even with the same compression algorithm, the results are often non-deterministic (number of compression threads, library version, etc.) 2. Even with the same compression algorithm, the results are often non-deterministic (number of compression threads, library version, etc.)
When we do the compression on the server and use the hashes of uncompressed NARs for lookups, the problem of non-determinism is no longer a problem since we only compress once. When we perform compression on the server and use the hashes of uncompressed NARs for lookups, non-determinism of compression is no longer a problem since we only compress once.
On the other hand, performing compression on the server leads to additional CPU usage, increasing compute costs and the need to scale. On the other hand, performing compression on the server leads to additional CPU usage, increasing compute costs and the need to scale.
Such design decisions are to be revisited later. Such design decisions are to be revisited later.