Given following scenario.
- Create new branch
- Commit a 10MB file
- Git push (uploads the 10MB file)
- Create a new branch (orphan)
- Commit the same 10mb file (no changes made, same object sha hash)
- Git push uploads the 10MB file AGAIN
My expectations are, that the already uploaded files won't be uploaded again using git push. But what actual happens is that when a new branch is made all files (even when thousands of smaller source files, instead of one 10MB file) will be uploaded again and again.
My question: How can I make it that Git detects that the 10mb file is already uploaded? Do you know a workaround/fix to make Git detecting already existing objects on the server when pushing commits? Git detects files by its sha, so it should be able to detect that some files in the tree of the commit are already present on the server.
Possible use-case: I have two completely different branches, but some common files are shared within those two. When I push one branch, I don't want to upload the common files again when I push the second branch.
Actual use-case: I do a lot of machine learning experiments using Python scripts and some smaller datasets (1MB - 10MB). Every time I start an experiment, I add all necessary experiment files to a new Git tree, and use that tree in a new commit without branching. That commits hangs completely free in the air and gets then referenced with a new Git reference (e.g. refs/jobs/my-experiment-name). When I now have two experiments with almost the same files (and thus two references), Git pushes all objects again when I push those references. I have low bandwidth and this really slows down my work.
$ mkdir git-test && cd git-test $ git init $ git remote add origin [email protected]:username/projectname.git # create dummy 10MB file $ head -c 10000000 /dev/urandom > dummy $ git add dummy $ git commit -m 'init' # first push, uploads everything - makes sense $ git push origin master Counting objects: 3, done. Delta compression using up to 6 threads. Compressing objects: 100% (2/2), done. Writing objects: 100% (3/3), 9.54 MiB | 1.13 MiB/s, done. Total 3 (delta 0), reused 0 (delta 0) # create new empty branch, not based from master $ git checkout --orphan branch2 # add same files again $ git add dummy $ git commit -m 'init on branch2' # this uploads now again the dummy file (10MB), although the server # has that object alread $ git push origin branch3 Counting objects: 3, done. Delta compression using up to 6 threads. Compressing objects: 100% (2/2), done. Writing objects: 100% (3/3), 9.54 MiB | 838.00 KiB/s, done. On the technical side we have:
- Two commits that do not share the same parents (have completely different history)
- Those two commits have the exact same tree sha id (and thus reference the same object files)
- Pushing both commits results in transferring all the objects in the same tree twice. Although I expect either that Git detects that the tree in the second commit is already present OR that file objects within that tree are already on the server.
Answer (I can't answer anymore, since someone marked this as duplicate).
The solution is unfortunately not that simple.
Every time Git wants to sync two repositories it builds a pack file, that contains all objects necessary (like files, commits, trees). When you execute a git push, the remote sends all existing references (branches) and its head commit SHA to the client. This is the problem: The pack protocol is not meant to be used per-object, but per-commit. So, according to the protocol itself, the explained behaviour above is correct. To work around that, I built a simple script every one can use to do a git push based on objects, instead of commits.
You find it here: https://github.com/marcj/git-objects-sync
What it does:
- Takes one commit (only one, you need to execute it on every unsynced parent commit as well) and builds a list of object SHAs (files, trees, commits) that belong to that commit (except parent commit).
- Sends this list to the server, servers answers back SHAs of objects it does not have yet
- Client builds a pack file based on the missing object SHAs and sends it to the server with the information which ref needs to be updated to which commit.
- Server receives pack file, unpacks it and updates the ref with given commit SHA.
Of course this has some drawbacks, but I described them in the linked Github repository.
With my script above you get now following:
marc@osx ~/git-test (branch11*) $ # added new branch11 as explained at the very top marc@osx ~/git-test (branch11*) $ python git-sync.py refs/heads/branch11 Counting objects: 1, done. Writing objects: 100% (1/1), 158 bytes | 158.00 KiB/s, done. Total 1 (delta 0), reused 0 (delta 0) marc@osx ~/git-test (branch11*) $ git push origin branch11 Everything up-to-date So as you see, it only syncs one object (the commit object), and not the dummy file and its tree object again.
--orphan, no no wonder this file is not in git :D The question is why does he need this flag