summaryrefslogtreecommitdiffstats
path: root/testing/docker/desktop-test/tc-vcs-config.yml
blob: 25e13ee4008b2d56b04c595af155c535dfd1a172 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
# Default configuration used by the tc-vs tools these can be overridden by
# passing the config you wish to use over the command line...
git: git
hg: hg

repoCache:
  # Repo url to clone when running repo init..
  repoUrl: https://gerrit.googlesource.com/git-repo.git
  # Version of repo to utilize...
  repoRevision: master
  # The root where all downloaded cache files are stored on the local machine...
  cacheDir: '{{env.HOME}}/.tc-vcs-repo/'
  # Name/prefixed used as part of the base url.
  cacheName: sources/{{name}}.tar.gz
  # Command used to upload the tarball
  uploadTar: "curl --header 'Content-Type: application/x-tar' --header 'Content-Encoding: gzip' -X PUT --data-binary @'{{source}}' '{{url}}'"
  # Large http get requests are often slower using nodes built in http layer so
  # we utilize a subprocess which is responsible for fetching...
  get: curl --connect-timeout 30 --speed-limit 500000 -L -o {{dest}} {{url}}
  # Used to create clone tarball
  compress: tar -czf {{dest}} {{source}}
  # All cache urls use tar + gz this is the command used to extract those files
  # downloaded by the "get" command.
  extract: tar -x -z -C {{dest}} -f {{source}}

cloneCache:
  # The root where all downloaded cache files are stored on the local machine...
  cacheDir: '{{env.HOME}}/.tc-vcs/'
  # Command used to upload the tarball
  uploadTar: "curl --header 'Content-Type: application/x-tar' --header 'Content-Encoding: gzip' -X PUT --data-binary @'{{source}}' '{{url}}'"
  # Large http get requests are often slower using nodes built in http layer so
  # we utilize a subprocess which is responsible for fetching...
  get: curl --connect-timeout 30 --speed-limit 500000 -L -o {{dest}} {{url}}
  # Used to create clone tarball
  compress: tar -czf {{dest}} {{source}}
  # All cache urls use tar + gz this is the command used to extract those files
  # downloaded by the "get" command.
  extract: tar -x -z --strip-components 1 -C {{dest}} -f {{source}}
  # Name/prefixed used as part of the base url.
  cacheName: clones/{{name}}.tar.gz