Showing results for 
Search instead for 
Did you mean: 
Sign up Log in

Earn badges and make progress

You're on your way to the next level! Join the Kudos program to earn points and save your progress.

Deleted user Avatar
Deleted user

Level 1: Seed

25 / 150 points

Next: Root


1 badge earned


Participate in fun challenges

Challenges come and go, but your rewards stay with you. Do more to earn more!


Gift kudos to your peers

What goes around comes around! Share the love by gifting kudos to your peers.


Rise up in the ranks

Keep earning points to reach the top of the leaderboard. It resets every quarter so you always have a chance!


Come for the products,
stay for the community

The Atlassian Community can help you and your team get more value out of Atlassian products and practices.

Atlassian Community about banner
Community Members
Community Events
Community Groups

Pushing a large ZIP file to Bitbucket using git.

I have enabled LFS with git and Bitbucket. I am now able to push a ZIP file with more than 100MB which is fantastic, but I am a little worried I will blot Bitbucket over time. The ZIP file is an export of an application component that is deployed to a workflow platform similar to Flowable.


The questions I have are the following:

  1. If I push a zip file, then make a small change, and push it again, what will happen? Will it push the difference, or the entire copy of the update ZIP file?
  2. If I merge branch B1 into B2 with the same ZIP file I mentioned above, what will happen? Will it increase the size in Bitbucket by double the size of the ZIP file or it will reuse the ZIP file in B1 to be present in B2?
  3. The process will repeat multiple times, that is push an updated version of the ZIP file to B1 multiple times, and then merge the changes from B1 to B2 and sometimes to B3. Do I have to be worried about bloating Bitbucket?


If the process flow above is problematic for Bitbucket, I am thinking to write a script that will use REST API to export the application component model from the workflow system, and unzip the file (it's all text in JSON and XML), and push the model to Bitbucket. Each ZIP file may contain literally hundreds of files (JSON and XML text files) that describe the application parts. The only problem with this approach is that it will take a long time, and there is a large number of changes that are not related to the programmatic features of the model. Such changes are to flag the metadata of each part within the component which is really not relevant at all to the programmatic change (application implementation details). I once did an experiment to make the application change to the unzipped files, zipped them back, and deployed the model to the workflow, and all was fine.


I appreciate your feedback.


0 answers

Suggest an answer

Log in or Sign up to answer

Atlassian Community Events