Buk Git Clone for BitBucket

Bulk git clone for BitBucket repositories of your team or user

Spread the love


Recently I have decided to make a backup of all my repositories on BitBucket and it is OK if you have up to 10, but if more than 100 then it is a pain.


I used my Synology NAS as a backup storage, which is on Linux but still does not have all the capabilities and freedom as Ubuntu or Debian, so I tried to use what I have (from the software perspective) and do not install anything else (for one time use). Anyway, I have had to add Syno Community as a package source, and then I installed SynoCli File Tools and Git, this is the minimum requirement, the other tools (such as jq) are coming from SynoCli File Tools.

Bash Script

So I created a bash script which was basically retrieving the list of the repositories of the given team or user in JSON format, stored to the local folder, and then parsing the list and go 1 by 1 and clone the repos to the local folder. BitBucket API has a limitation of the number of the repositories rendered on one (single) page, it is max 100, but it gives you an idea that there is another (next) page, so you can parse it as well and loop up to max pages you need.




# you can uncomment this line if you need to create a folder for your team/user
# rm -rf "$TEAM" && mkdir "$TEAM" && cd $TEAM


while [ ! -z $NEXT_URL ]
  echo "Processing ${NEXT_URL}..."
  curl -u "${USER}:${SECRET}" -H "Content-Type: application/json" $NEXT_URL > repoinfo.json
  jq -r '.values[] | .links.clone[0].href' repoinfo.json > repos.txt
  NEXT_URL=`jq -r '.next' repoinfo.json`

  for repo in `cat repos.txt`
    echo "Cloning ${repo}..."
    # I'm using HTTP links
    git clone $repo

You need to replace all in <…> by your values, hope you understand it 🙂

That’s basically it. Don’t forget to grant the right execution permission to your bash script.