I have a situation with a relatively large git repository located in a virtual machine on an elderly, slow host on my local network where it takes quite a while to do the initia
PS. Fair warning:
git
is generally considered blazingly fast. You should try cloning a full repo from darcs, bazaar, hg (god forbid: TFS or subversion...). Also, if you routinely clone full repos from scratch, you'd be doing something wrong anyway. You can always justgit remote update
and get incremental changes.For various other ways to keep full repos in synch see, e.g.
- "fetch --all" in a git bare repository doesn't synchronize local branches to the remote ones
- How to update a git clone --mirror?
(The contain links to other relevant SO posts)
As mentioned you could just copy a repository with 'dumb' file transfer.
This will certainly not waste time compressing, repacking, deltifying and/or filtering.
Plus, you will get
This may or may not be what you require, but it is nice to be aware of the fact
Git clone by default optimizes for bandwidth. Since git clone, by default, does not mirror all branches (see --mirror
) it would not make sense to just dump the pack-files as-is (because that will send possibly way more than required).
When distributing to a truly big number of clients, consider using bundles.
If you want a fast clone without the server-side cost, the git way is bundle create
. You can now distribute the bundle, without the server even being involved. If you mean that bundle... --all
includes more than simple git clone
, consider e.g. bundle ... master
to reduce the volume.
git bundle create snapshot.bundle --all # (or mention specific ref names instead of --all)
and distribute the snapshot bundle instead. That's the best of both worlds, while of course you won't get the items from the bullet list above. On the receiving end, just
git clone snapshot.bundle myclonedir/
You can look at lowering server load by reducing/removing compression.
Have a look at these config settings (I assume pack.compression
may help you lower the server load)
core.compression
An integer -1..9, indicating a default compression level. -1 is the zlib default. 0 means no compression, and 1..9 are various speed/size tradeoffs, 9 being slowest. If set, this provides a default to other compression variables, such as core.loosecompression and pack.compression.
core.loosecompression
An integer -1..9, indicating the compression level for objects that are not in a pack file. -1 is the zlib default. 0 means no compression, and 1..9 are various speed/size tradeoffs, 9 being slowest. If not set, defaults to core.compression. If that is not set, defaults to 1 (best speed).
pack.compression
An integer -1..9, indicating the compression level for objects in a pack file. -1 is the zlib default. 0 means no compression, and 1..9 are various speed/size tradeoffs, 9 being slowest. If not set, defaults to core.compression. If that is not set, defaults to -1, the zlib default, which is "a default compromise between speed and compression (currently equivalent to level 6)."
Note that changing the compression level will not automatically recompress all existing objects. You can force recompression by passing the -F option to git-repack(1).
Given ample network bandwidth, this will in fact result in faster clones. Don't forget about git-repack -F
when you decide to benchmark that!