book/faqs: Copyedit
This commit is contained in:
parent
49f2ccbff8
commit
2afecd29ad
|
@ -52,12 +52,12 @@ We use the hash of the _uncompressed NAR_ to perform global deduplication.
|
||||||
└───────►File Size
|
└───────►File Size
|
||||||
```
|
```
|
||||||
|
|
||||||
At first glance, performing compression on the client and deduplicating the result may sound appealing, but has problems:
|
At first glance, performing compression on the client and deduplicating the result may sound appealing, but this approach isn't without problems:
|
||||||
|
|
||||||
1. Different compression algorithms and levels naturally lead to different results which can't be deduplicated
|
1. Different compression algorithms and levels naturally lead to different results which can't be deduplicated
|
||||||
2. Even with the same compression algorithm, the results are often non-deterministic (number of compression threads, library version, etc.)
|
2. Even with the same compression algorithm, the results are often non-deterministic (number of compression threads, library version, etc.)
|
||||||
|
|
||||||
When we do the compression on the server and use the hashes of uncompressed NARs for lookups, the problem of non-determinism is no longer a problem since we only compress once.
|
When we perform compression on the server and use the hashes of uncompressed NARs for lookups, non-determinism of compression is no longer a problem since we only compress once.
|
||||||
|
|
||||||
On the other hand, performing compression on the server leads to additional CPU usage, increasing compute costs and the need to scale.
|
On the other hand, performing compression on the server leads to additional CPU usage, increasing compute costs and the need to scale.
|
||||||
Such design decisions are to be revisited later.
|
Such design decisions are to be revisited later.
|
||||||
|
|
Loading…
Reference in a new issue