diff options
Diffstat (limited to 'subcmds/sync.py')
-rw-r--r-- | subcmds/sync.py | 1054 |
1 files changed, 559 insertions, 495 deletions
diff --git a/subcmds/sync.py b/subcmds/sync.py index 2973a16e..3211cbb1 100644 --- a/subcmds/sync.py +++ b/subcmds/sync.py | |||
@@ -1,5 +1,3 @@ | |||
1 | # -*- coding:utf-8 -*- | ||
2 | # | ||
3 | # Copyright (C) 2008 The Android Open Source Project | 1 | # Copyright (C) 2008 The Android Open Source Project |
4 | # | 2 | # |
5 | # Licensed under the Apache License, Version 2.0 (the "License"); | 3 | # Licensed under the Apache License, Version 2.0 (the "License"); |
@@ -14,37 +12,23 @@ | |||
14 | # See the License for the specific language governing permissions and | 12 | # See the License for the specific language governing permissions and |
15 | # limitations under the License. | 13 | # limitations under the License. |
16 | 14 | ||
17 | from __future__ import print_function | 15 | import errno |
16 | import functools | ||
17 | import http.cookiejar as cookielib | ||
18 | import io | ||
18 | import json | 19 | import json |
20 | import multiprocessing | ||
19 | import netrc | 21 | import netrc |
20 | from optparse import SUPPRESS_HELP | 22 | from optparse import SUPPRESS_HELP |
21 | import os | 23 | import os |
22 | import re | ||
23 | import socket | 24 | import socket |
24 | import subprocess | ||
25 | import sys | 25 | import sys |
26 | import tempfile | 26 | import tempfile |
27 | import time | 27 | import time |
28 | 28 | import urllib.error | |
29 | from pyversion import is_python3 | 29 | import urllib.parse |
30 | if is_python3(): | 30 | import urllib.request |
31 | import http.cookiejar as cookielib | 31 | import xmlrpc.client |
32 | import urllib.error | ||
33 | import urllib.parse | ||
34 | import urllib.request | ||
35 | import xmlrpc.client | ||
36 | else: | ||
37 | import cookielib | ||
38 | import imp | ||
39 | import urllib2 | ||
40 | import urlparse | ||
41 | import xmlrpclib | ||
42 | urllib = imp.new_module('urllib') | ||
43 | urllib.error = urllib2 | ||
44 | urllib.parse = urlparse | ||
45 | urllib.request = urllib2 | ||
46 | xmlrpc = imp.new_module('xmlrpc') | ||
47 | xmlrpc.client = xmlrpclib | ||
48 | 32 | ||
49 | try: | 33 | try: |
50 | import threading as _threading | 34 | import threading as _threading |
@@ -53,44 +37,36 @@ except ImportError: | |||
53 | 37 | ||
54 | try: | 38 | try: |
55 | import resource | 39 | import resource |
40 | |||
56 | def _rlimit_nofile(): | 41 | def _rlimit_nofile(): |
57 | return resource.getrlimit(resource.RLIMIT_NOFILE) | 42 | return resource.getrlimit(resource.RLIMIT_NOFILE) |
58 | except ImportError: | 43 | except ImportError: |
59 | def _rlimit_nofile(): | 44 | def _rlimit_nofile(): |
60 | return (256, 256) | 45 | return (256, 256) |
61 | 46 | ||
62 | try: | ||
63 | import multiprocessing | ||
64 | except ImportError: | ||
65 | multiprocessing = None | ||
66 | |||
67 | import event_log | 47 | import event_log |
68 | from git_command import GIT, git_require | 48 | from git_command import git_require |
69 | from git_config import GetUrlCookieFile | 49 | from git_config import GetUrlCookieFile |
70 | from git_refs import R_HEADS, HEAD | 50 | from git_refs import R_HEADS, HEAD |
51 | import git_superproject | ||
71 | import gitc_utils | 52 | import gitc_utils |
72 | from project import Project | 53 | from project import Project |
73 | from project import RemoteSpec | 54 | from project import RemoteSpec |
74 | from command import Command, MirrorSafeCommand | 55 | from command import Command, MirrorSafeCommand, WORKER_BATCH_SIZE |
75 | from error import RepoChangedException, GitError, ManifestParseError | 56 | from error import RepoChangedException, GitError, ManifestParseError |
76 | import platform_utils | 57 | import platform_utils |
77 | from project import SyncBuffer | 58 | from project import SyncBuffer |
78 | from progress import Progress | 59 | from progress import Progress |
60 | import ssh | ||
79 | from wrapper import Wrapper | 61 | from wrapper import Wrapper |
80 | from manifest_xml import GitcManifest | 62 | from manifest_xml import GitcManifest |
81 | 63 | ||
82 | _ONE_DAY_S = 24 * 60 * 60 | 64 | _ONE_DAY_S = 24 * 60 * 60 |
83 | 65 | ||
84 | class _FetchError(Exception): | ||
85 | """Internal error thrown in _FetchHelper() when we don't want stack trace.""" | ||
86 | pass | ||
87 | |||
88 | class _CheckoutError(Exception): | ||
89 | """Internal error thrown in _CheckoutOne() when we don't want stack trace.""" | ||
90 | 66 | ||
91 | class Sync(Command, MirrorSafeCommand): | 67 | class Sync(Command, MirrorSafeCommand): |
92 | jobs = 1 | 68 | jobs = 1 |
93 | common = True | 69 | COMMON = True |
94 | helpSummary = "Update working tree to the latest revision" | 70 | helpSummary = "Update working tree to the latest revision" |
95 | helpUsage = """ | 71 | helpUsage = """ |
96 | %prog [<project>...] | 72 | %prog [<project>...] |
@@ -133,11 +109,11 @@ if the manifest server specified in the manifest file already includes | |||
133 | credentials. | 109 | credentials. |
134 | 110 | ||
135 | By default, all projects will be synced. The --fail-fast option can be used | 111 | By default, all projects will be synced. The --fail-fast option can be used |
136 | to halt syncing as soon as possible when the the first project fails to sync. | 112 | to halt syncing as soon as possible when the first project fails to sync. |
137 | 113 | ||
138 | The --force-sync option can be used to overwrite existing git | 114 | The --force-sync option can be used to overwrite existing git |
139 | directories if they have previously been linked to a different | 115 | directories if they have previously been linked to a different |
140 | object direcotry. WARNING: This may cause data to be lost since | 116 | object directory. WARNING: This may cause data to be lost since |
141 | refs may be removed when overwriting. | 117 | refs may be removed when overwriting. |
142 | 118 | ||
143 | The --force-remove-dirty option can be used to remove previously used | 119 | The --force-remove-dirty option can be used to remove previously used |
@@ -191,12 +167,21 @@ If the remote SSH daemon is Gerrit Code Review, version 2.0.10 or | |||
191 | later is required to fix a server side protocol bug. | 167 | later is required to fix a server side protocol bug. |
192 | 168 | ||
193 | """ | 169 | """ |
170 | PARALLEL_JOBS = 1 | ||
171 | |||
172 | def _CommonOptions(self, p): | ||
173 | if self.manifest: | ||
174 | try: | ||
175 | self.PARALLEL_JOBS = self.manifest.default.sync_j | ||
176 | except ManifestParseError: | ||
177 | pass | ||
178 | super()._CommonOptions(p) | ||
194 | 179 | ||
195 | def _Options(self, p, show_smart=True): | 180 | def _Options(self, p, show_smart=True): |
196 | try: | 181 | p.add_option('--jobs-network', default=None, type=int, metavar='JOBS', |
197 | self.jobs = self.manifest.default.sync_j | 182 | help='number of network jobs to run in parallel (defaults to --jobs)') |
198 | except ManifestParseError: | 183 | p.add_option('--jobs-checkout', default=None, type=int, metavar='JOBS', |
199 | self.jobs = 1 | 184 | help='number of local checkout jobs to run in parallel (defaults to --jobs)') |
200 | 185 | ||
201 | p.add_option('-f', '--force-broken', | 186 | p.add_option('-f', '--force-broken', |
202 | dest='force_broken', action='store_true', | 187 | dest='force_broken', action='store_true', |
@@ -217,6 +202,10 @@ later is required to fix a server side protocol bug. | |||
217 | p.add_option('-l', '--local-only', | 202 | p.add_option('-l', '--local-only', |
218 | dest='local_only', action='store_true', | 203 | dest='local_only', action='store_true', |
219 | help="only update working tree, don't fetch") | 204 | help="only update working tree, don't fetch") |
205 | p.add_option('--no-manifest-update', '--nmu', | ||
206 | dest='mp_update', action='store_false', default='true', | ||
207 | help='use the existing manifest checkout as-is. ' | ||
208 | '(do not update to the latest revision)') | ||
220 | p.add_option('-n', '--network-only', | 209 | p.add_option('-n', '--network-only', |
221 | dest='network_only', action='store_true', | 210 | dest='network_only', action='store_true', |
222 | help="fetch only, don't update working tree") | 211 | help="fetch only, don't update working tree") |
@@ -226,17 +215,15 @@ later is required to fix a server side protocol bug. | |||
226 | p.add_option('-c', '--current-branch', | 215 | p.add_option('-c', '--current-branch', |
227 | dest='current_branch_only', action='store_true', | 216 | dest='current_branch_only', action='store_true', |
228 | help='fetch only current branch from server') | 217 | help='fetch only current branch from server') |
229 | p.add_option('-q', '--quiet', | 218 | p.add_option('--no-current-branch', |
230 | dest='quiet', action='store_true', | 219 | dest='current_branch_only', action='store_false', |
231 | help='be more quiet') | 220 | help='fetch all branches from server') |
232 | p.add_option('-j', '--jobs', | ||
233 | dest='jobs', action='store', type='int', | ||
234 | help="projects to fetch simultaneously (default %d)" % self.jobs) | ||
235 | p.add_option('-m', '--manifest-name', | 221 | p.add_option('-m', '--manifest-name', |
236 | dest='manifest_name', | 222 | dest='manifest_name', |
237 | help='temporary manifest to use for this sync', metavar='NAME.xml') | 223 | help='temporary manifest to use for this sync', metavar='NAME.xml') |
238 | p.add_option('--no-clone-bundle', | 224 | p.add_option('--clone-bundle', action='store_true', |
239 | dest='no_clone_bundle', action='store_true', | 225 | help='enable use of /clone.bundle on HTTP/HTTPS') |
226 | p.add_option('--no-clone-bundle', dest='clone_bundle', action='store_false', | ||
240 | help='disable use of /clone.bundle on HTTP/HTTPS') | 227 | help='disable use of /clone.bundle on HTTP/HTTPS') |
241 | p.add_option('-u', '--manifest-server-username', action='store', | 228 | p.add_option('-u', '--manifest-server-username', action='store', |
242 | dest='manifest_server_username', | 229 | dest='manifest_server_username', |
@@ -247,12 +234,23 @@ later is required to fix a server side protocol bug. | |||
247 | p.add_option('--fetch-submodules', | 234 | p.add_option('--fetch-submodules', |
248 | dest='fetch_submodules', action='store_true', | 235 | dest='fetch_submodules', action='store_true', |
249 | help='fetch submodules from server') | 236 | help='fetch submodules from server') |
237 | p.add_option('--use-superproject', action='store_true', | ||
238 | help='use the manifest superproject to sync projects') | ||
239 | p.add_option('--no-use-superproject', action='store_false', | ||
240 | dest='use_superproject', | ||
241 | help='disable use of manifest superprojects') | ||
242 | p.add_option('--tags', | ||
243 | action='store_false', | ||
244 | help='fetch tags') | ||
250 | p.add_option('--no-tags', | 245 | p.add_option('--no-tags', |
251 | dest='no_tags', action='store_true', | 246 | dest='tags', action='store_false', |
252 | help="don't fetch tags") | 247 | help="don't fetch tags") |
253 | p.add_option('--optimized-fetch', | 248 | p.add_option('--optimized-fetch', |
254 | dest='optimized_fetch', action='store_true', | 249 | dest='optimized_fetch', action='store_true', |
255 | help='only fetch projects fixed to sha1 if revision does not exist locally') | 250 | help='only fetch projects fixed to sha1 if revision does not exist locally') |
251 | p.add_option('--retry-fetches', | ||
252 | default=0, action='store', type='int', | ||
253 | help='number of times to retry fetches on transient errors') | ||
256 | p.add_option('--prune', dest='prune', action='store_true', | 254 | p.add_option('--prune', dest='prune', action='store_true', |
257 | help='delete refs that no longer exist on the remote') | 255 | help='delete refs that no longer exist on the remote') |
258 | if show_smart: | 256 | if show_smart: |
@@ -265,345 +263,400 @@ later is required to fix a server side protocol bug. | |||
265 | 263 | ||
266 | g = p.add_option_group('repo Version options') | 264 | g = p.add_option_group('repo Version options') |
267 | g.add_option('--no-repo-verify', | 265 | g.add_option('--no-repo-verify', |
268 | dest='no_repo_verify', action='store_true', | 266 | dest='repo_verify', default=True, action='store_false', |
269 | help='do not verify repo source code') | 267 | help='do not verify repo source code') |
270 | g.add_option('--repo-upgraded', | 268 | g.add_option('--repo-upgraded', |
271 | dest='repo_upgraded', action='store_true', | 269 | dest='repo_upgraded', action='store_true', |
272 | help=SUPPRESS_HELP) | 270 | help=SUPPRESS_HELP) |
273 | 271 | ||
274 | def _FetchProjectList(self, opt, projects, sem, *args, **kwargs): | 272 | def _GetBranch(self): |
275 | """Main function of the fetch threads. | 273 | """Returns the branch name for getting the approved manifest.""" |
274 | p = self.manifest.manifestProject | ||
275 | b = p.GetBranch(p.CurrentBranch) | ||
276 | branch = b.merge | ||
277 | if branch.startswith(R_HEADS): | ||
278 | branch = branch[len(R_HEADS):] | ||
279 | return branch | ||
280 | |||
281 | def _GetCurrentBranchOnly(self, opt): | ||
282 | """Returns True if current-branch or use-superproject options are enabled.""" | ||
283 | return opt.current_branch_only or git_superproject.UseSuperproject(opt, self.manifest) | ||
284 | |||
285 | def _UpdateProjectsRevisionId(self, opt, args, load_local_manifests, superproject_logging_data): | ||
286 | """Update revisionId of every project with the SHA from superproject. | ||
287 | |||
288 | This function updates each project's revisionId with SHA from superproject. | ||
289 | It writes the updated manifest into a file and reloads the manifest from it. | ||
290 | |||
291 | Args: | ||
292 | opt: Program options returned from optparse. See _Options(). | ||
293 | args: Arguments to pass to GetProjects. See the GetProjects | ||
294 | docstring for details. | ||
295 | load_local_manifests: Whether to load local manifests. | ||
296 | superproject_logging_data: A dictionary of superproject data that is to be logged. | ||
297 | |||
298 | Returns: | ||
299 | Returns path to the overriding manifest file instead of None. | ||
300 | """ | ||
301 | print_messages = git_superproject.PrintMessages(opt, self.manifest) | ||
302 | superproject = git_superproject.Superproject(self.manifest, | ||
303 | self.repodir, | ||
304 | self.git_event_log, | ||
305 | quiet=opt.quiet, | ||
306 | print_messages=print_messages) | ||
307 | if opt.local_only: | ||
308 | manifest_path = superproject.manifest_path | ||
309 | if manifest_path: | ||
310 | self._ReloadManifest(manifest_path, load_local_manifests) | ||
311 | return manifest_path | ||
312 | |||
313 | all_projects = self.GetProjects(args, | ||
314 | missing_ok=True, | ||
315 | submodules_ok=opt.fetch_submodules) | ||
316 | update_result = superproject.UpdateProjectsRevisionId(all_projects) | ||
317 | manifest_path = update_result.manifest_path | ||
318 | superproject_logging_data['updatedrevisionid'] = bool(manifest_path) | ||
319 | if manifest_path: | ||
320 | self._ReloadManifest(manifest_path, load_local_manifests) | ||
321 | else: | ||
322 | if print_messages: | ||
323 | print('warning: Update of revisionId from superproject has failed, ' | ||
324 | 'repo sync will not use superproject to fetch the source. ', | ||
325 | 'Please resync with the --no-use-superproject option to avoid this repo warning.', | ||
326 | file=sys.stderr) | ||
327 | if update_result.fatal and opt.use_superproject is not None: | ||
328 | sys.exit(1) | ||
329 | return manifest_path | ||
330 | |||
331 | def _FetchProjectList(self, opt, projects): | ||
332 | """Main function of the fetch worker. | ||
333 | |||
334 | The projects we're given share the same underlying git object store, so we | ||
335 | have to fetch them in serial. | ||
276 | 336 | ||
277 | Delegates most of the work to _FetchHelper. | 337 | Delegates most of the work to _FetchHelper. |
278 | 338 | ||
279 | Args: | 339 | Args: |
280 | opt: Program options returned from optparse. See _Options(). | 340 | opt: Program options returned from optparse. See _Options(). |
281 | projects: Projects to fetch. | 341 | projects: Projects to fetch. |
282 | sem: We'll release() this semaphore when we exit so that another thread | ||
283 | can be started up. | ||
284 | *args, **kwargs: Remaining arguments to pass to _FetchHelper. See the | ||
285 | _FetchHelper docstring for details. | ||
286 | """ | 342 | """ |
287 | try: | 343 | return [self._FetchOne(opt, x) for x in projects] |
288 | for project in projects: | ||
289 | success = self._FetchHelper(opt, project, *args, **kwargs) | ||
290 | if not success and opt.fail_fast: | ||
291 | break | ||
292 | finally: | ||
293 | sem.release() | ||
294 | 344 | ||
295 | def _FetchHelper(self, opt, project, lock, fetched, pm, err_event, | 345 | def _FetchOne(self, opt, project): |
296 | clone_filter): | ||
297 | """Fetch git objects for a single project. | 346 | """Fetch git objects for a single project. |
298 | 347 | ||
299 | Args: | 348 | Args: |
300 | opt: Program options returned from optparse. See _Options(). | 349 | opt: Program options returned from optparse. See _Options(). |
301 | project: Project object for the project to fetch. | 350 | project: Project object for the project to fetch. |
302 | lock: Lock for accessing objects that are shared amongst multiple | ||
303 | _FetchHelper() threads. | ||
304 | fetched: set object that we will add project.gitdir to when we're done | ||
305 | (with our lock held). | ||
306 | pm: Instance of a Project object. We will call pm.update() (with our | ||
307 | lock held). | ||
308 | err_event: We'll set this event in the case of an error (after printing | ||
309 | out info about the error). | ||
310 | clone_filter: Filter for use in a partial clone. | ||
311 | 351 | ||
312 | Returns: | 352 | Returns: |
313 | Whether the fetch was successful. | 353 | Whether the fetch was successful. |
314 | """ | 354 | """ |
315 | # We'll set to true once we've locked the lock. | ||
316 | did_lock = False | ||
317 | |||
318 | # Encapsulate everything in a try/except/finally so that: | ||
319 | # - We always set err_event in the case of an exception. | ||
320 | # - We always make sure we unlock the lock if we locked it. | ||
321 | start = time.time() | 355 | start = time.time() |
322 | success = False | 356 | success = False |
357 | buf = io.StringIO() | ||
323 | try: | 358 | try: |
324 | try: | 359 | success = project.Sync_NetworkHalf( |
325 | success = project.Sync_NetworkHalf( | ||
326 | quiet=opt.quiet, | 360 | quiet=opt.quiet, |
327 | current_branch_only=opt.current_branch_only, | 361 | verbose=opt.verbose, |
362 | output_redir=buf, | ||
363 | current_branch_only=self._GetCurrentBranchOnly(opt), | ||
328 | force_sync=opt.force_sync, | 364 | force_sync=opt.force_sync, |
329 | clone_bundle=not opt.no_clone_bundle, | 365 | clone_bundle=opt.clone_bundle, |
330 | no_tags=opt.no_tags, archive=self.manifest.IsArchive, | 366 | tags=opt.tags, archive=self.manifest.IsArchive, |
331 | optimized_fetch=opt.optimized_fetch, | 367 | optimized_fetch=opt.optimized_fetch, |
368 | retry_fetches=opt.retry_fetches, | ||
332 | prune=opt.prune, | 369 | prune=opt.prune, |
333 | clone_filter=clone_filter) | 370 | ssh_proxy=self.ssh_proxy, |
334 | self._fetch_times.Set(project, time.time() - start) | 371 | clone_filter=self.manifest.CloneFilter, |
372 | partial_clone_exclude=self.manifest.PartialCloneExclude) | ||
335 | 373 | ||
336 | # Lock around all the rest of the code, since printing, updating a set | 374 | output = buf.getvalue() |
337 | # and Progress.update() are not thread safe. | 375 | if (opt.verbose or not success) and output: |
338 | lock.acquire() | 376 | print('\n' + output.rstrip()) |
339 | did_lock = True | ||
340 | 377 | ||
341 | if not success: | 378 | if not success: |
342 | err_event.set() | 379 | print('error: Cannot fetch %s from %s' |
343 | print('error: Cannot fetch %s from %s' | 380 | % (project.name, project.remote.url), |
344 | % (project.name, project.remote.url), | 381 | file=sys.stderr) |
345 | file=sys.stderr) | 382 | except GitError as e: |
346 | if opt.fail_fast: | 383 | print('error.GitError: Cannot fetch %s' % str(e), file=sys.stderr) |
347 | raise _FetchError() | 384 | except Exception as e: |
348 | 385 | print('error: Cannot fetch %s (%s: %s)' | |
349 | fetched.add(project.gitdir) | ||
350 | pm.update(msg=project.name) | ||
351 | except _FetchError: | ||
352 | pass | ||
353 | except Exception as e: | ||
354 | print('error: Cannot fetch %s (%s: %s)' \ | ||
355 | % (project.name, type(e).__name__, str(e)), file=sys.stderr) | 386 | % (project.name, type(e).__name__, str(e)), file=sys.stderr) |
356 | err_event.set() | 387 | raise |
357 | raise | 388 | |
358 | finally: | 389 | finish = time.time() |
359 | if did_lock: | 390 | return (success, project, start, finish) |
360 | lock.release() | ||
361 | finish = time.time() | ||
362 | self.event_log.AddSync(project, event_log.TASK_SYNC_NETWORK, | ||
363 | start, finish, success) | ||
364 | 391 | ||
365 | return success | 392 | @classmethod |
393 | def _FetchInitChild(cls, ssh_proxy): | ||
394 | cls.ssh_proxy = ssh_proxy | ||
366 | 395 | ||
367 | def _Fetch(self, projects, opt): | 396 | def _Fetch(self, projects, opt, err_event, ssh_proxy): |
397 | ret = True | ||
398 | |||
399 | jobs = opt.jobs_network if opt.jobs_network else self.jobs | ||
368 | fetched = set() | 400 | fetched = set() |
369 | lock = _threading.Lock() | 401 | pm = Progress('Fetching', len(projects), delay=False, quiet=opt.quiet) |
370 | pm = Progress('Fetching projects', len(projects), | ||
371 | always_print_percentage=opt.quiet) | ||
372 | 402 | ||
373 | objdir_project_map = dict() | 403 | objdir_project_map = dict() |
374 | for project in projects: | 404 | for project in projects: |
375 | objdir_project_map.setdefault(project.objdir, []).append(project) | 405 | objdir_project_map.setdefault(project.objdir, []).append(project) |
376 | 406 | projects_list = list(objdir_project_map.values()) | |
377 | threads = set() | 407 | |
378 | sem = _threading.Semaphore(self.jobs) | 408 | def _ProcessResults(results_sets): |
379 | err_event = _threading.Event() | 409 | ret = True |
380 | for project_list in objdir_project_map.values(): | 410 | for results in results_sets: |
381 | # Check for any errors before running any more tasks. | 411 | for (success, project, start, finish) in results: |
382 | # ...we'll let existing threads finish, though. | 412 | self._fetch_times.Set(project, finish - start) |
383 | if err_event.isSet() and opt.fail_fast: | 413 | self.event_log.AddSync(project, event_log.TASK_SYNC_NETWORK, |
384 | break | 414 | start, finish, success) |
385 | 415 | # Check for any errors before running any more tasks. | |
386 | sem.acquire() | 416 | # ...we'll let existing jobs finish, though. |
387 | kwargs = dict(opt=opt, | 417 | if not success: |
388 | projects=project_list, | 418 | ret = False |
389 | sem=sem, | 419 | else: |
390 | lock=lock, | 420 | fetched.add(project.gitdir) |
391 | fetched=fetched, | 421 | pm.update(msg=project.name) |
392 | pm=pm, | 422 | if not ret and opt.fail_fast: |
393 | err_event=err_event, | 423 | break |
394 | clone_filter=self.manifest.CloneFilter) | 424 | return ret |
395 | if self.jobs > 1: | 425 | |
396 | t = _threading.Thread(target = self._FetchProjectList, | 426 | # We pass the ssh proxy settings via the class. This allows multiprocessing |
397 | kwargs = kwargs) | 427 | # to pickle it up when spawning children. We can't pass it as an argument |
398 | # Ensure that Ctrl-C will not freeze the repo process. | 428 | # to _FetchProjectList below as multiprocessing is unable to pickle those. |
399 | t.daemon = True | 429 | Sync.ssh_proxy = None |
400 | threads.add(t) | 430 | |
401 | t.start() | 431 | # NB: Multiprocessing is heavy, so don't spin it up for one job. |
432 | if len(projects_list) == 1 or jobs == 1: | ||
433 | self._FetchInitChild(ssh_proxy) | ||
434 | if not _ProcessResults(self._FetchProjectList(opt, x) for x in projects_list): | ||
435 | ret = False | ||
436 | else: | ||
437 | # Favor throughput over responsiveness when quiet. It seems that imap() | ||
438 | # will yield results in batches relative to chunksize, so even as the | ||
439 | # children finish a sync, we won't see the result until one child finishes | ||
440 | # ~chunksize jobs. When using a large --jobs with large chunksize, this | ||
441 | # can be jarring as there will be a large initial delay where repo looks | ||
442 | # like it isn't doing anything and sits at 0%, but then suddenly completes | ||
443 | # a lot of jobs all at once. Since this code is more network bound, we | ||
444 | # can accept a bit more CPU overhead with a smaller chunksize so that the | ||
445 | # user sees more immediate & continuous feedback. | ||
446 | if opt.quiet: | ||
447 | chunksize = WORKER_BATCH_SIZE | ||
402 | else: | 448 | else: |
403 | self._FetchProjectList(**kwargs) | 449 | pm.update(inc=0, msg='warming up') |
404 | 450 | chunksize = 4 | |
405 | for t in threads: | 451 | with multiprocessing.Pool( |
406 | t.join() | 452 | jobs, initializer=self._FetchInitChild, initargs=(ssh_proxy,)) as pool: |
407 | 453 | results = pool.imap_unordered( | |
408 | # If we saw an error, exit with code 1 so that other scripts can check. | 454 | functools.partial(self._FetchProjectList, opt), |
409 | if err_event.isSet() and opt.fail_fast: | 455 | projects_list, |
410 | print('\nerror: Exited sync due to fetch errors', file=sys.stderr) | 456 | chunksize=chunksize) |
411 | sys.exit(1) | 457 | if not _ProcessResults(results): |
458 | ret = False | ||
459 | pool.close() | ||
460 | |||
461 | # Cleanup the reference now that we're done with it, and we're going to | ||
462 | # release any resources it points to. If we don't, later multiprocessing | ||
463 | # usage (e.g. checkouts) will try to pickle and then crash. | ||
464 | del Sync.ssh_proxy | ||
412 | 465 | ||
413 | pm.end() | 466 | pm.end() |
414 | self._fetch_times.Save() | 467 | self._fetch_times.Save() |
415 | 468 | ||
416 | if not self.manifest.IsArchive: | 469 | if not self.manifest.IsArchive: |
417 | self._GCProjects(projects) | 470 | self._GCProjects(projects, opt, err_event) |
418 | 471 | ||
419 | return fetched | 472 | return (ret, fetched) |
420 | 473 | ||
421 | def _CheckoutWorker(self, opt, sem, project, *args, **kwargs): | 474 | def _FetchMain(self, opt, args, all_projects, err_event, manifest_name, |
422 | """Main function of the fetch threads. | 475 | load_local_manifests, ssh_proxy): |
423 | 476 | """The main network fetch loop. | |
424 | Delegates most of the work to _CheckoutOne. | ||
425 | 477 | ||
426 | Args: | 478 | Args: |
427 | opt: Program options returned from optparse. See _Options(). | 479 | opt: Program options returned from optparse. See _Options(). |
428 | projects: Projects to fetch. | 480 | args: Command line args used to filter out projects. |
429 | sem: We'll release() this semaphore when we exit so that another thread | 481 | all_projects: List of all projects that should be fetched. |
430 | can be started up. | 482 | err_event: Whether an error was hit while processing. |
431 | *args, **kwargs: Remaining arguments to pass to _CheckoutOne. See the | 483 | manifest_name: Manifest file to be reloaded. |
432 | _CheckoutOne docstring for details. | 484 | load_local_manifests: Whether to load local manifests. |
485 | ssh_proxy: SSH manager for clients & masters. | ||
486 | |||
487 | Returns: | ||
488 | List of all projects that should be checked out. | ||
433 | """ | 489 | """ |
434 | try: | 490 | rp = self.manifest.repoProject |
435 | return self._CheckoutOne(opt, project, *args, **kwargs) | ||
436 | finally: | ||
437 | sem.release() | ||
438 | 491 | ||
439 | def _CheckoutOne(self, opt, project, lock, pm, err_event, err_results): | 492 | to_fetch = [] |
493 | now = time.time() | ||
494 | if _ONE_DAY_S <= (now - rp.LastFetch): | ||
495 | to_fetch.append(rp) | ||
496 | to_fetch.extend(all_projects) | ||
497 | to_fetch.sort(key=self._fetch_times.Get, reverse=True) | ||
498 | |||
499 | success, fetched = self._Fetch(to_fetch, opt, err_event, ssh_proxy) | ||
500 | if not success: | ||
501 | err_event.set() | ||
502 | |||
503 | _PostRepoFetch(rp, opt.repo_verify) | ||
504 | if opt.network_only: | ||
505 | # bail out now; the rest touches the working tree | ||
506 | if err_event.is_set(): | ||
507 | print('\nerror: Exited sync due to fetch errors.\n', file=sys.stderr) | ||
508 | sys.exit(1) | ||
509 | return | ||
510 | |||
511 | # Iteratively fetch missing and/or nested unregistered submodules | ||
512 | previously_missing_set = set() | ||
513 | while True: | ||
514 | self._ReloadManifest(manifest_name, load_local_manifests) | ||
515 | all_projects = self.GetProjects(args, | ||
516 | missing_ok=True, | ||
517 | submodules_ok=opt.fetch_submodules) | ||
518 | missing = [] | ||
519 | for project in all_projects: | ||
520 | if project.gitdir not in fetched: | ||
521 | missing.append(project) | ||
522 | if not missing: | ||
523 | break | ||
524 | # Stop us from non-stopped fetching actually-missing repos: If set of | ||
525 | # missing repos has not been changed from last fetch, we break. | ||
526 | missing_set = set(p.name for p in missing) | ||
527 | if previously_missing_set == missing_set: | ||
528 | break | ||
529 | previously_missing_set = missing_set | ||
530 | success, new_fetched = self._Fetch(missing, opt, err_event, ssh_proxy) | ||
531 | if not success: | ||
532 | err_event.set() | ||
533 | fetched.update(new_fetched) | ||
534 | |||
535 | return all_projects | ||
536 | |||
537 | def _CheckoutOne(self, detach_head, force_sync, project): | ||
440 | """Checkout work tree for one project | 538 | """Checkout work tree for one project |
441 | 539 | ||
442 | Args: | 540 | Args: |
443 | opt: Program options returned from optparse. See _Options(). | 541 | detach_head: Whether to leave a detached HEAD. |
542 | force_sync: Force checking out of the repo. | ||
444 | project: Project object for the project to checkout. | 543 | project: Project object for the project to checkout. |
445 | lock: Lock for accessing objects that are shared amongst multiple | ||
446 | _CheckoutWorker() threads. | ||
447 | pm: Instance of a Project object. We will call pm.update() (with our | ||
448 | lock held). | ||
449 | err_event: We'll set this event in the case of an error (after printing | ||
450 | out info about the error). | ||
451 | err_results: A list of strings, paths to git repos where checkout | ||
452 | failed. | ||
453 | 544 | ||
454 | Returns: | 545 | Returns: |
455 | Whether the fetch was successful. | 546 | Whether the fetch was successful. |
456 | """ | 547 | """ |
457 | # We'll set to true once we've locked the lock. | ||
458 | did_lock = False | ||
459 | |||
460 | # Encapsulate everything in a try/except/finally so that: | ||
461 | # - We always set err_event in the case of an exception. | ||
462 | # - We always make sure we unlock the lock if we locked it. | ||
463 | start = time.time() | 548 | start = time.time() |
464 | syncbuf = SyncBuffer(self.manifest.manifestProject.config, | 549 | syncbuf = SyncBuffer(self.manifest.manifestProject.config, |
465 | detach_head=opt.detach_head) | 550 | detach_head=detach_head) |
466 | success = False | 551 | success = False |
467 | try: | 552 | try: |
468 | try: | 553 | project.Sync_LocalHalf(syncbuf, force_sync=force_sync) |
469 | project.Sync_LocalHalf(syncbuf, force_sync=opt.force_sync) | 554 | success = syncbuf.Finish() |
470 | 555 | except GitError as e: | |
471 | # Lock around all the rest of the code, since printing, updating a set | 556 | print('error.GitError: Cannot checkout %s: %s' % |
472 | # and Progress.update() are not thread safe. | 557 | (project.name, str(e)), file=sys.stderr) |
473 | lock.acquire() | 558 | except Exception as e: |
474 | success = syncbuf.Finish() | 559 | print('error: Cannot checkout %s: %s: %s' % |
475 | did_lock = True | 560 | (project.name, type(e).__name__, str(e)), |
476 | 561 | file=sys.stderr) | |
477 | if not success: | 562 | raise |
478 | err_event.set() | ||
479 | print('error: Cannot checkout %s' % (project.name), | ||
480 | file=sys.stderr) | ||
481 | raise _CheckoutError() | ||
482 | |||
483 | pm.update(msg=project.name) | ||
484 | except _CheckoutError: | ||
485 | pass | ||
486 | except Exception as e: | ||
487 | print('error: Cannot checkout %s: %s: %s' % | ||
488 | (project.name, type(e).__name__, str(e)), | ||
489 | file=sys.stderr) | ||
490 | err_event.set() | ||
491 | raise | ||
492 | finally: | ||
493 | if did_lock: | ||
494 | if not success: | ||
495 | err_results.append(project.relpath) | ||
496 | lock.release() | ||
497 | finish = time.time() | ||
498 | self.event_log.AddSync(project, event_log.TASK_SYNC_LOCAL, | ||
499 | start, finish, success) | ||
500 | 563 | ||
501 | return success | 564 | if not success: |
565 | print('error: Cannot checkout %s' % (project.name), file=sys.stderr) | ||
566 | finish = time.time() | ||
567 | return (success, project, start, finish) | ||
502 | 568 | ||
503 | def _Checkout(self, all_projects, opt): | 569 | def _Checkout(self, all_projects, opt, err_results): |
504 | """Checkout projects listed in all_projects | 570 | """Checkout projects listed in all_projects |
505 | 571 | ||
506 | Args: | 572 | Args: |
507 | all_projects: List of all projects that should be checked out. | 573 | all_projects: List of all projects that should be checked out. |
508 | opt: Program options returned from optparse. See _Options(). | 574 | opt: Program options returned from optparse. See _Options(). |
575 | err_results: A list of strings, paths to git repos where checkout failed. | ||
509 | """ | 576 | """ |
577 | # Only checkout projects with worktrees. | ||
578 | all_projects = [x for x in all_projects if x.worktree] | ||
579 | |||
580 | def _ProcessResults(pool, pm, results): | ||
581 | ret = True | ||
582 | for (success, project, start, finish) in results: | ||
583 | self.event_log.AddSync(project, event_log.TASK_SYNC_LOCAL, | ||
584 | start, finish, success) | ||
585 | # Check for any errors before running any more tasks. | ||
586 | # ...we'll let existing jobs finish, though. | ||
587 | if not success: | ||
588 | ret = False | ||
589 | err_results.append(project.relpath) | ||
590 | if opt.fail_fast: | ||
591 | if pool: | ||
592 | pool.close() | ||
593 | return ret | ||
594 | pm.update(msg=project.name) | ||
595 | return ret | ||
510 | 596 | ||
511 | # Perform checkouts in multiple threads when we are using partial clone. | 597 | return self.ExecuteInParallel( |
512 | # Without partial clone, all needed git objects are already downloaded, | 598 | opt.jobs_checkout if opt.jobs_checkout else self.jobs, |
513 | # in this situation it's better to use only one process because the checkout | 599 | functools.partial(self._CheckoutOne, opt.detach_head, opt.force_sync), |
514 | # would be mostly disk I/O; with partial clone, the objects are only | 600 | all_projects, |
515 | # downloaded when demanded (at checkout time), which is similar to the | 601 | callback=_ProcessResults, |
516 | # Sync_NetworkHalf case and parallelism would be helpful. | 602 | output=Progress('Checking out', len(all_projects), quiet=opt.quiet)) and not err_results |
517 | if self.manifest.CloneFilter: | ||
518 | syncjobs = self.jobs | ||
519 | else: | ||
520 | syncjobs = 1 | ||
521 | |||
522 | lock = _threading.Lock() | ||
523 | pm = Progress('Checking out projects', len(all_projects)) | ||
524 | |||
525 | threads = set() | ||
526 | sem = _threading.Semaphore(syncjobs) | ||
527 | err_event = _threading.Event() | ||
528 | err_results = [] | ||
529 | |||
530 | for project in all_projects: | ||
531 | # Check for any errors before running any more tasks. | ||
532 | # ...we'll let existing threads finish, though. | ||
533 | if err_event.isSet() and opt.fail_fast: | ||
534 | break | ||
535 | |||
536 | sem.acquire() | ||
537 | if project.worktree: | ||
538 | kwargs = dict(opt=opt, | ||
539 | sem=sem, | ||
540 | project=project, | ||
541 | lock=lock, | ||
542 | pm=pm, | ||
543 | err_event=err_event, | ||
544 | err_results=err_results) | ||
545 | if syncjobs > 1: | ||
546 | t = _threading.Thread(target=self._CheckoutWorker, | ||
547 | kwargs=kwargs) | ||
548 | # Ensure that Ctrl-C will not freeze the repo process. | ||
549 | t.daemon = True | ||
550 | threads.add(t) | ||
551 | t.start() | ||
552 | else: | ||
553 | self._CheckoutWorker(**kwargs) | ||
554 | |||
555 | for t in threads: | ||
556 | t.join() | ||
557 | 603 | ||
558 | pm.end() | 604 | def _GCProjects(self, projects, opt, err_event): |
559 | # If we saw an error, exit with code 1 so that other scripts can check. | 605 | pm = Progress('Garbage collecting', len(projects), delay=False, quiet=opt.quiet) |
560 | if err_event.isSet(): | 606 | pm.update(inc=0, msg='prescan') |
561 | print('\nerror: Exited sync due to checkout errors', file=sys.stderr) | ||
562 | if err_results: | ||
563 | print('Failing repos:\n%s' % '\n'.join(err_results), | ||
564 | file=sys.stderr) | ||
565 | sys.exit(1) | ||
566 | 607 | ||
567 | def _GCProjects(self, projects): | ||
568 | gc_gitdirs = {} | 608 | gc_gitdirs = {} |
569 | for project in projects: | 609 | for project in projects: |
570 | if len(project.manifest.GetProjectsWithName(project.name)) > 1: | 610 | # Make sure pruning never kicks in with shared projects. |
571 | print('Shared project %s found, disabling pruning.' % project.name) | 611 | if (not project.use_git_worktrees and |
572 | project.bare_git.config('--replace-all', 'gc.pruneExpire', 'never') | 612 | len(project.manifest.GetProjectsWithName(project.name)) > 1): |
613 | if not opt.quiet: | ||
614 | print('\r%s: Shared project %s found, disabling pruning.' % | ||
615 | (project.relpath, project.name)) | ||
616 | if git_require((2, 7, 0)): | ||
617 | project.EnableRepositoryExtension('preciousObjects') | ||
618 | else: | ||
619 | # This isn't perfect, but it's the best we can do with old git. | ||
620 | print('\r%s: WARNING: shared projects are unreliable when using old ' | ||
621 | 'versions of git; please upgrade to git-2.7.0+.' | ||
622 | % (project.relpath,), | ||
623 | file=sys.stderr) | ||
624 | project.config.SetString('gc.pruneExpire', 'never') | ||
573 | gc_gitdirs[project.gitdir] = project.bare_git | 625 | gc_gitdirs[project.gitdir] = project.bare_git |
574 | 626 | ||
575 | has_dash_c = git_require((1, 7, 2)) | 627 | pm.update(inc=len(projects) - len(gc_gitdirs), msg='warming up') |
576 | if multiprocessing and has_dash_c: | 628 | |
577 | cpu_count = multiprocessing.cpu_count() | 629 | cpu_count = os.cpu_count() |
578 | else: | ||
579 | cpu_count = 1 | ||
580 | jobs = min(self.jobs, cpu_count) | 630 | jobs = min(self.jobs, cpu_count) |
581 | 631 | ||
582 | if jobs < 2: | 632 | if jobs < 2: |
583 | for bare_git in gc_gitdirs.values(): | 633 | for bare_git in gc_gitdirs.values(): |
634 | pm.update(msg=bare_git._project.name) | ||
584 | bare_git.gc('--auto') | 635 | bare_git.gc('--auto') |
636 | pm.end() | ||
585 | return | 637 | return |
586 | 638 | ||
587 | config = {'pack.threads': cpu_count // jobs if cpu_count > jobs else 1} | 639 | config = {'pack.threads': cpu_count // jobs if cpu_count > jobs else 1} |
588 | 640 | ||
589 | threads = set() | 641 | threads = set() |
590 | sem = _threading.Semaphore(jobs) | 642 | sem = _threading.Semaphore(jobs) |
591 | err_event = _threading.Event() | ||
592 | 643 | ||
593 | def GC(bare_git): | 644 | def GC(bare_git): |
645 | pm.start(bare_git._project.name) | ||
594 | try: | 646 | try: |
595 | try: | 647 | try: |
596 | bare_git.gc('--auto', config=config) | 648 | bare_git.gc('--auto', config=config) |
597 | except GitError: | 649 | except GitError: |
598 | err_event.set() | 650 | err_event.set() |
599 | except: | 651 | except Exception: |
600 | err_event.set() | 652 | err_event.set() |
601 | raise | 653 | raise |
602 | finally: | 654 | finally: |
655 | pm.finish(bare_git._project.name) | ||
603 | sem.release() | 656 | sem.release() |
604 | 657 | ||
605 | for bare_git in gc_gitdirs.values(): | 658 | for bare_git in gc_gitdirs.values(): |
606 | if err_event.isSet(): | 659 | if err_event.is_set() and opt.fail_fast: |
607 | break | 660 | break |
608 | sem.acquire() | 661 | sem.acquire() |
609 | t = _threading.Thread(target=GC, args=(bare_git,)) | 662 | t = _threading.Thread(target=GC, args=(bare_git,)) |
@@ -613,84 +666,30 @@ later is required to fix a server side protocol bug. | |||
613 | 666 | ||
614 | for t in threads: | 667 | for t in threads: |
615 | t.join() | 668 | t.join() |
669 | pm.end() | ||
616 | 670 | ||
617 | if err_event.isSet(): | 671 | def _ReloadManifest(self, manifest_name=None, load_local_manifests=True): |
618 | print('\nerror: Exited sync due to gc errors', file=sys.stderr) | 672 | """Reload the manfiest from the file specified by the |manifest_name|. |
619 | sys.exit(1) | 673 | |
674 | It unloads the manifest if |manifest_name| is None. | ||
620 | 675 | ||
621 | def _ReloadManifest(self, manifest_name=None): | 676 | Args: |
677 | manifest_name: Manifest file to be reloaded. | ||
678 | load_local_manifests: Whether to load local manifests. | ||
679 | """ | ||
622 | if manifest_name: | 680 | if manifest_name: |
623 | # Override calls _Unload already | 681 | # Override calls _Unload already |
624 | self.manifest.Override(manifest_name) | 682 | self.manifest.Override(manifest_name, load_local_manifests=load_local_manifests) |
625 | else: | 683 | else: |
626 | self.manifest._Unload() | 684 | self.manifest._Unload() |
627 | 685 | ||
628 | def _DeleteProject(self, path): | ||
629 | print('Deleting obsolete path %s' % path, file=sys.stderr) | ||
630 | |||
631 | # Delete the .git directory first, so we're less likely to have a partially | ||
632 | # working git repository around. There shouldn't be any git projects here, | ||
633 | # so rmtree works. | ||
634 | try: | ||
635 | platform_utils.rmtree(os.path.join(path, '.git')) | ||
636 | except OSError as e: | ||
637 | print('Failed to remove %s (%s)' % (os.path.join(path, '.git'), str(e)), file=sys.stderr) | ||
638 | print('error: Failed to delete obsolete path %s' % path, file=sys.stderr) | ||
639 | print(' remove manually, then run sync again', file=sys.stderr) | ||
640 | return 1 | ||
641 | |||
642 | # Delete everything under the worktree, except for directories that contain | ||
643 | # another git project | ||
644 | dirs_to_remove = [] | ||
645 | failed = False | ||
646 | for root, dirs, files in platform_utils.walk(path): | ||
647 | for f in files: | ||
648 | try: | ||
649 | platform_utils.remove(os.path.join(root, f)) | ||
650 | except OSError as e: | ||
651 | print('Failed to remove %s (%s)' % (os.path.join(root, f), str(e)), file=sys.stderr) | ||
652 | failed = True | ||
653 | dirs[:] = [d for d in dirs | ||
654 | if not os.path.lexists(os.path.join(root, d, '.git'))] | ||
655 | dirs_to_remove += [os.path.join(root, d) for d in dirs | ||
656 | if os.path.join(root, d) not in dirs_to_remove] | ||
657 | for d in reversed(dirs_to_remove): | ||
658 | if platform_utils.islink(d): | ||
659 | try: | ||
660 | platform_utils.remove(d) | ||
661 | except OSError as e: | ||
662 | print('Failed to remove %s (%s)' % (os.path.join(root, d), str(e)), file=sys.stderr) | ||
663 | failed = True | ||
664 | elif len(platform_utils.listdir(d)) == 0: | ||
665 | try: | ||
666 | platform_utils.rmdir(d) | ||
667 | except OSError as e: | ||
668 | print('Failed to remove %s (%s)' % (os.path.join(root, d), str(e)), file=sys.stderr) | ||
669 | failed = True | ||
670 | continue | ||
671 | if failed: | ||
672 | print('error: Failed to delete obsolete path %s' % path, file=sys.stderr) | ||
673 | print(' remove manually, then run sync again', file=sys.stderr) | ||
674 | return 1 | ||
675 | |||
676 | # Try deleting parent dirs if they are empty | ||
677 | project_dir = path | ||
678 | while project_dir != self.manifest.topdir: | ||
679 | if len(platform_utils.listdir(project_dir)) == 0: | ||
680 | platform_utils.rmdir(project_dir) | ||
681 | else: | ||
682 | break | ||
683 | project_dir = os.path.dirname(project_dir) | ||
684 | |||
685 | return 0 | ||
686 | |||
687 | def UpdateProjectList(self, opt): | 686 | def UpdateProjectList(self, opt): |
688 | new_project_paths = [] | 687 | new_project_paths = [] |
689 | for project in self.GetProjects(None, missing_ok=True): | 688 | for project in self.GetProjects(None, missing_ok=True): |
690 | if project.relpath: | 689 | if project.relpath: |
691 | new_project_paths.append(project.relpath) | 690 | new_project_paths.append(project.relpath) |
692 | file_name = 'project.list' | 691 | file_name = 'project.list' |
693 | file_path = os.path.join(self.manifest.repodir, file_name) | 692 | file_path = os.path.join(self.repodir, file_name) |
694 | old_project_paths = [] | 693 | old_project_paths = [] |
695 | 694 | ||
696 | if os.path.exists(file_path): | 695 | if os.path.exists(file_path): |
@@ -705,28 +704,20 @@ later is required to fix a server side protocol bug. | |||
705 | gitdir = os.path.join(self.manifest.topdir, path, '.git') | 704 | gitdir = os.path.join(self.manifest.topdir, path, '.git') |
706 | if os.path.exists(gitdir): | 705 | if os.path.exists(gitdir): |
707 | project = Project( | 706 | project = Project( |
708 | manifest = self.manifest, | 707 | manifest=self.manifest, |
709 | name = path, | 708 | name=path, |
710 | remote = RemoteSpec('origin'), | 709 | remote=RemoteSpec('origin'), |
711 | gitdir = gitdir, | 710 | gitdir=gitdir, |
712 | objdir = gitdir, | 711 | objdir=gitdir, |
713 | worktree = os.path.join(self.manifest.topdir, path), | 712 | use_git_worktrees=os.path.isfile(gitdir), |
714 | relpath = path, | 713 | worktree=os.path.join(self.manifest.topdir, path), |
715 | revisionExpr = 'HEAD', | 714 | relpath=path, |
716 | revisionId = None, | 715 | revisionExpr='HEAD', |
717 | groups = None) | 716 | revisionId=None, |
718 | 717 | groups=None) | |
719 | if project.IsDirty() and opt.force_remove_dirty: | 718 | if not project.DeleteWorktree( |
720 | print('WARNING: Removing dirty project "%s": uncommitted changes ' | 719 | quiet=opt.quiet, |
721 | 'erased' % project.relpath, file=sys.stderr) | 720 | force=opt.force_remove_dirty): |
722 | self._DeleteProject(project.worktree) | ||
723 | elif project.IsDirty(): | ||
724 | print('error: Cannot remove project "%s": uncommitted changes ' | ||
725 | 'are present' % project.relpath, file=sys.stderr) | ||
726 | print(' commit changes, then run sync again', | ||
727 | file=sys.stderr) | ||
728 | return 1 | ||
729 | elif self._DeleteProject(project.worktree): | ||
730 | return 1 | 721 | return 1 |
731 | 722 | ||
732 | new_project_paths.sort() | 723 | new_project_paths.sort() |
@@ -735,6 +726,56 @@ later is required to fix a server side protocol bug. | |||
735 | fd.write('\n') | 726 | fd.write('\n') |
736 | return 0 | 727 | return 0 |
737 | 728 | ||
729 | def UpdateCopyLinkfileList(self): | ||
730 | """Save all dests of copyfile and linkfile, and update them if needed. | ||
731 | |||
732 | Returns: | ||
733 | Whether update was successful. | ||
734 | """ | ||
735 | new_paths = {} | ||
736 | new_linkfile_paths = [] | ||
737 | new_copyfile_paths = [] | ||
738 | for project in self.GetProjects(None, missing_ok=True): | ||
739 | new_linkfile_paths.extend(x.dest for x in project.linkfiles) | ||
740 | new_copyfile_paths.extend(x.dest for x in project.copyfiles) | ||
741 | |||
742 | new_paths = { | ||
743 | 'linkfile': new_linkfile_paths, | ||
744 | 'copyfile': new_copyfile_paths, | ||
745 | } | ||
746 | |||
747 | copylinkfile_name = 'copy-link-files.json' | ||
748 | copylinkfile_path = os.path.join(self.manifest.repodir, copylinkfile_name) | ||
749 | old_copylinkfile_paths = {} | ||
750 | |||
751 | if os.path.exists(copylinkfile_path): | ||
752 | with open(copylinkfile_path, 'rb') as fp: | ||
753 | try: | ||
754 | old_copylinkfile_paths = json.load(fp) | ||
755 | except: | ||
756 | print('error: %s is not a json formatted file.' % | ||
757 | copylinkfile_path, file=sys.stderr) | ||
758 | platform_utils.remove(copylinkfile_path) | ||
759 | return False | ||
760 | |||
761 | need_remove_files = [] | ||
762 | need_remove_files.extend( | ||
763 | set(old_copylinkfile_paths.get('linkfile', [])) - | ||
764 | set(new_linkfile_paths)) | ||
765 | need_remove_files.extend( | ||
766 | set(old_copylinkfile_paths.get('copyfile', [])) - | ||
767 | set(new_copyfile_paths)) | ||
768 | |||
769 | for need_remove_file in need_remove_files: | ||
770 | # Try to remove the updated copyfile or linkfile. | ||
771 | # So, if the file is not exist, nothing need to do. | ||
772 | platform_utils.remove(need_remove_file, missing_ok=True) | ||
773 | |||
774 | # Create copy-link-files.json, save dest path of "copyfile" and "linkfile". | ||
775 | with open(copylinkfile_path, 'w', encoding='utf-8') as fp: | ||
776 | json.dump(new_paths, fp) | ||
777 | return True | ||
778 | |||
738 | def _SmartSyncSetup(self, opt, smart_sync_manifest_path): | 779 | def _SmartSyncSetup(self, opt, smart_sync_manifest_path): |
739 | if not self.manifest.manifest_server: | 780 | if not self.manifest.manifest_server: |
740 | print('error: cannot smart sync: no manifest server defined in ' | 781 | print('error: cannot smart sync: no manifest server defined in ' |
@@ -745,7 +786,7 @@ later is required to fix a server side protocol bug. | |||
745 | if not opt.quiet: | 786 | if not opt.quiet: |
746 | print('Using manifest server %s' % manifest_server) | 787 | print('Using manifest server %s' % manifest_server) |
747 | 788 | ||
748 | if not '@' in manifest_server: | 789 | if '@' not in manifest_server: |
749 | username = None | 790 | username = None |
750 | password = None | 791 | password = None |
751 | if opt.manifest_server_username and opt.manifest_server_password: | 792 | if opt.manifest_server_username and opt.manifest_server_password: |
@@ -782,19 +823,15 @@ later is required to fix a server side protocol bug. | |||
782 | try: | 823 | try: |
783 | server = xmlrpc.client.Server(manifest_server, transport=transport) | 824 | server = xmlrpc.client.Server(manifest_server, transport=transport) |
784 | if opt.smart_sync: | 825 | if opt.smart_sync: |
785 | p = self.manifest.manifestProject | 826 | branch = self._GetBranch() |
786 | b = p.GetBranch(p.CurrentBranch) | 827 | |
787 | branch = b.merge | 828 | if 'SYNC_TARGET' in os.environ: |
788 | if branch.startswith(R_HEADS): | 829 | target = os.environ['SYNC_TARGET'] |
789 | branch = branch[len(R_HEADS):] | ||
790 | |||
791 | env = os.environ.copy() | ||
792 | if 'SYNC_TARGET' in env: | ||
793 | target = env['SYNC_TARGET'] | ||
794 | [success, manifest_str] = server.GetApprovedManifest(branch, target) | 830 | [success, manifest_str] = server.GetApprovedManifest(branch, target) |
795 | elif 'TARGET_PRODUCT' in env and 'TARGET_BUILD_VARIANT' in env: | 831 | elif ('TARGET_PRODUCT' in os.environ and |
796 | target = '%s-%s' % (env['TARGET_PRODUCT'], | 832 | 'TARGET_BUILD_VARIANT' in os.environ): |
797 | env['TARGET_BUILD_VARIANT']) | 833 | target = '%s-%s' % (os.environ['TARGET_PRODUCT'], |
834 | os.environ['TARGET_BUILD_VARIANT']) | ||
798 | [success, manifest_str] = server.GetApprovedManifest(branch, target) | 835 | [success, manifest_str] = server.GetApprovedManifest(branch, target) |
799 | else: | 836 | else: |
800 | [success, manifest_str] = server.GetApprovedManifest(branch) | 837 | [success, manifest_str] = server.GetApprovedManifest(branch) |
@@ -833,12 +870,15 @@ later is required to fix a server side protocol bug. | |||
833 | """Fetch & update the local manifest project.""" | 870 | """Fetch & update the local manifest project.""" |
834 | if not opt.local_only: | 871 | if not opt.local_only: |
835 | start = time.time() | 872 | start = time.time() |
836 | success = mp.Sync_NetworkHalf(quiet=opt.quiet, | 873 | success = mp.Sync_NetworkHalf(quiet=opt.quiet, verbose=opt.verbose, |
837 | current_branch_only=opt.current_branch_only, | 874 | current_branch_only=self._GetCurrentBranchOnly(opt), |
838 | no_tags=opt.no_tags, | 875 | force_sync=opt.force_sync, |
876 | tags=opt.tags, | ||
839 | optimized_fetch=opt.optimized_fetch, | 877 | optimized_fetch=opt.optimized_fetch, |
878 | retry_fetches=opt.retry_fetches, | ||
840 | submodules=self.manifest.HasSubmodules, | 879 | submodules=self.manifest.HasSubmodules, |
841 | clone_filter=self.manifest.CloneFilter) | 880 | clone_filter=self.manifest.CloneFilter, |
881 | partial_clone_exclude=self.manifest.PartialCloneExclude) | ||
842 | finish = time.time() | 882 | finish = time.time() |
843 | self.event_log.AddSync(mp, event_log.TASK_SYNC_NETWORK, | 883 | self.event_log.AddSync(mp, event_log.TASK_SYNC_NETWORK, |
844 | start, finish, success) | 884 | start, finish, success) |
@@ -852,7 +892,7 @@ later is required to fix a server side protocol bug. | |||
852 | start, time.time(), clean) | 892 | start, time.time(), clean) |
853 | if not clean: | 893 | if not clean: |
854 | sys.exit(1) | 894 | sys.exit(1) |
855 | self._ReloadManifest(opt.manifest_name) | 895 | self._ReloadManifest(manifest_name) |
856 | if opt.jobs is None: | 896 | if opt.jobs is None: |
857 | self.jobs = self.manifest.default.sync_j | 897 | self.jobs = self.manifest.default.sync_j |
858 | 898 | ||
@@ -886,7 +926,10 @@ later is required to fix a server side protocol bug. | |||
886 | 926 | ||
887 | manifest_name = opt.manifest_name | 927 | manifest_name = opt.manifest_name |
888 | smart_sync_manifest_path = os.path.join( | 928 | smart_sync_manifest_path = os.path.join( |
889 | self.manifest.manifestProject.worktree, 'smart_sync_override.xml') | 929 | self.manifest.manifestProject.worktree, 'smart_sync_override.xml') |
930 | |||
931 | if opt.clone_bundle is None: | ||
932 | opt.clone_bundle = self.manifest.CloneBundle | ||
890 | 933 | ||
891 | if opt.smart_sync or opt.smart_tag: | 934 | if opt.smart_sync or opt.smart_tag: |
892 | manifest_name = self._SmartSyncSetup(opt, smart_sync_manifest_path) | 935 | manifest_name = self._SmartSyncSetup(opt, smart_sync_manifest_path) |
@@ -898,8 +941,17 @@ later is required to fix a server side protocol bug. | |||
898 | print('error: failed to remove existing smart sync override manifest: %s' % | 941 | print('error: failed to remove existing smart sync override manifest: %s' % |
899 | e, file=sys.stderr) | 942 | e, file=sys.stderr) |
900 | 943 | ||
944 | err_event = multiprocessing.Event() | ||
945 | |||
901 | rp = self.manifest.repoProject | 946 | rp = self.manifest.repoProject |
902 | rp.PreSync() | 947 | rp.PreSync() |
948 | cb = rp.CurrentBranch | ||
949 | if cb: | ||
950 | base = rp.GetBranch(cb).merge | ||
951 | if not base or not base.startswith('refs/heads/'): | ||
952 | print('warning: repo is not tracking a remote branch, so it will not ' | ||
953 | 'receive updates; run `repo init --repo-rev=stable` to fix.', | ||
954 | file=sys.stderr) | ||
903 | 955 | ||
904 | mp = self.manifest.manifestProject | 956 | mp = self.manifest.manifestProject |
905 | mp.PreSync() | 957 | mp.PreSync() |
@@ -907,7 +959,21 @@ later is required to fix a server side protocol bug. | |||
907 | if opt.repo_upgraded: | 959 | if opt.repo_upgraded: |
908 | _PostRepoUpgrade(self.manifest, quiet=opt.quiet) | 960 | _PostRepoUpgrade(self.manifest, quiet=opt.quiet) |
909 | 961 | ||
910 | self._UpdateManifestProject(opt, mp, manifest_name) | 962 | if not opt.mp_update: |
963 | print('Skipping update of local manifest project.') | ||
964 | else: | ||
965 | self._UpdateManifestProject(opt, mp, manifest_name) | ||
966 | |||
967 | load_local_manifests = not self.manifest.HasLocalManifests | ||
968 | use_superproject = git_superproject.UseSuperproject(opt, self.manifest) | ||
969 | superproject_logging_data = { | ||
970 | 'superproject': use_superproject, | ||
971 | 'haslocalmanifests': bool(self.manifest.HasLocalManifests), | ||
972 | 'hassuperprojecttag': bool(self.manifest.superproject), | ||
973 | } | ||
974 | if use_superproject: | ||
975 | manifest_name = self._UpdateProjectsRevisionId( | ||
976 | opt, args, load_local_manifests, superproject_logging_data) or opt.manifest_name | ||
911 | 977 | ||
912 | if self.gitc_manifest: | 978 | if self.gitc_manifest: |
913 | gitc_manifest_projects = self.GetProjects(args, | 979 | gitc_manifest_projects = self.GetProjects(args, |
@@ -948,56 +1014,92 @@ later is required to fix a server side protocol bug. | |||
948 | missing_ok=True, | 1014 | missing_ok=True, |
949 | submodules_ok=opt.fetch_submodules) | 1015 | submodules_ok=opt.fetch_submodules) |
950 | 1016 | ||
1017 | err_network_sync = False | ||
1018 | err_update_projects = False | ||
1019 | |||
951 | self._fetch_times = _FetchTimes(self.manifest) | 1020 | self._fetch_times = _FetchTimes(self.manifest) |
952 | if not opt.local_only: | 1021 | if not opt.local_only: |
953 | to_fetch = [] | 1022 | with multiprocessing.Manager() as manager: |
954 | now = time.time() | 1023 | with ssh.ProxyManager(manager) as ssh_proxy: |
955 | if _ONE_DAY_S <= (now - rp.LastFetch): | 1024 | # Initialize the socket dir once in the parent. |
956 | to_fetch.append(rp) | 1025 | ssh_proxy.sock() |
957 | to_fetch.extend(all_projects) | 1026 | all_projects = self._FetchMain(opt, args, all_projects, err_event, |
958 | to_fetch.sort(key=self._fetch_times.Get, reverse=True) | 1027 | manifest_name, load_local_manifests, |
959 | 1028 | ssh_proxy) | |
960 | fetched = self._Fetch(to_fetch, opt) | 1029 | |
961 | _PostRepoFetch(rp, opt.no_repo_verify) | ||
962 | if opt.network_only: | 1030 | if opt.network_only: |
963 | # bail out now; the rest touches the working tree | ||
964 | return | 1031 | return |
965 | 1032 | ||
966 | # Iteratively fetch missing and/or nested unregistered submodules | 1033 | # If we saw an error, exit with code 1 so that other scripts can check. |
967 | previously_missing_set = set() | 1034 | if err_event.is_set(): |
968 | while True: | 1035 | err_network_sync = True |
969 | self._ReloadManifest(manifest_name) | 1036 | if opt.fail_fast: |
970 | all_projects = self.GetProjects(args, | 1037 | print('\nerror: Exited sync due to fetch errors.\n' |
971 | missing_ok=True, | 1038 | 'Local checkouts *not* updated. Resolve network issues & ' |
972 | submodules_ok=opt.fetch_submodules) | 1039 | 'retry.\n' |
973 | missing = [] | 1040 | '`repo sync -l` will update some local checkouts.', |
974 | for project in all_projects: | 1041 | file=sys.stderr) |
975 | if project.gitdir not in fetched: | 1042 | sys.exit(1) |
976 | missing.append(project) | ||
977 | if not missing: | ||
978 | break | ||
979 | # Stop us from non-stopped fetching actually-missing repos: If set of | ||
980 | # missing repos has not been changed from last fetch, we break. | ||
981 | missing_set = set(p.name for p in missing) | ||
982 | if previously_missing_set == missing_set: | ||
983 | break | ||
984 | previously_missing_set = missing_set | ||
985 | fetched.update(self._Fetch(missing, opt)) | ||
986 | 1043 | ||
987 | if self.manifest.IsMirror or self.manifest.IsArchive: | 1044 | if self.manifest.IsMirror or self.manifest.IsArchive: |
988 | # bail out now, we have no working tree | 1045 | # bail out now, we have no working tree |
989 | return | 1046 | return |
990 | 1047 | ||
991 | if self.UpdateProjectList(opt): | 1048 | if self.UpdateProjectList(opt): |
992 | sys.exit(1) | 1049 | err_event.set() |
1050 | err_update_projects = True | ||
1051 | if opt.fail_fast: | ||
1052 | print('\nerror: Local checkouts *not* updated.', file=sys.stderr) | ||
1053 | sys.exit(1) | ||
993 | 1054 | ||
994 | self._Checkout(all_projects, opt) | 1055 | err_update_linkfiles = not self.UpdateCopyLinkfileList() |
1056 | if err_update_linkfiles: | ||
1057 | err_event.set() | ||
1058 | if opt.fail_fast: | ||
1059 | print('\nerror: Local update copyfile or linkfile failed.', file=sys.stderr) | ||
1060 | sys.exit(1) | ||
1061 | |||
1062 | err_results = [] | ||
1063 | # NB: We don't exit here because this is the last step. | ||
1064 | err_checkout = not self._Checkout(all_projects, opt, err_results) | ||
1065 | if err_checkout: | ||
1066 | err_event.set() | ||
995 | 1067 | ||
996 | # If there's a notice that's supposed to print at the end of the sync, print | 1068 | # If there's a notice that's supposed to print at the end of the sync, print |
997 | # it now... | 1069 | # it now... |
998 | if self.manifest.notice: | 1070 | if self.manifest.notice: |
999 | print(self.manifest.notice) | 1071 | print(self.manifest.notice) |
1000 | 1072 | ||
1073 | # If we saw an error, exit with code 1 so that other scripts can check. | ||
1074 | if err_event.is_set(): | ||
1075 | print('\nerror: Unable to fully sync the tree.', file=sys.stderr) | ||
1076 | if err_network_sync: | ||
1077 | print('error: Downloading network changes failed.', file=sys.stderr) | ||
1078 | if err_update_projects: | ||
1079 | print('error: Updating local project lists failed.', file=sys.stderr) | ||
1080 | if err_update_linkfiles: | ||
1081 | print('error: Updating copyfiles or linkfiles failed.', file=sys.stderr) | ||
1082 | if err_checkout: | ||
1083 | print('error: Checking out local projects failed.', file=sys.stderr) | ||
1084 | if err_results: | ||
1085 | print('Failing repos:\n%s' % '\n'.join(err_results), file=sys.stderr) | ||
1086 | print('Try re-running with "-j1 --fail-fast" to exit at the first error.', | ||
1087 | file=sys.stderr) | ||
1088 | sys.exit(1) | ||
1089 | |||
1090 | # Log the previous sync analysis state from the config. | ||
1091 | self.git_event_log.LogDataConfigEvents(mp.config.GetSyncAnalysisStateData(), | ||
1092 | 'previous_sync_state') | ||
1093 | |||
1094 | # Update and log with the new sync analysis state. | ||
1095 | mp.config.UpdateSyncAnalysisState(opt, superproject_logging_data) | ||
1096 | self.git_event_log.LogDataConfigEvents(mp.config.GetSyncAnalysisStateData(), | ||
1097 | 'current_sync_state') | ||
1098 | |||
1099 | if not opt.quiet: | ||
1100 | print('repo sync has finished successfully.') | ||
1101 | |||
1102 | |||
1001 | def _PostRepoUpgrade(manifest, quiet=False): | 1103 | def _PostRepoUpgrade(manifest, quiet=False): |
1002 | wrapper = Wrapper() | 1104 | wrapper = Wrapper() |
1003 | if wrapper.NeedSetupGnuPG(): | 1105 | if wrapper.NeedSetupGnuPG(): |
@@ -1006,15 +1108,29 @@ def _PostRepoUpgrade(manifest, quiet=False): | |||
1006 | if project.Exists: | 1108 | if project.Exists: |
1007 | project.PostRepoUpgrade() | 1109 | project.PostRepoUpgrade() |
1008 | 1110 | ||
1009 | def _PostRepoFetch(rp, no_repo_verify=False, verbose=False): | 1111 | |
1112 | def _PostRepoFetch(rp, repo_verify=True, verbose=False): | ||
1010 | if rp.HasChanges: | 1113 | if rp.HasChanges: |
1011 | print('info: A new version of repo is available', file=sys.stderr) | 1114 | print('info: A new version of repo is available', file=sys.stderr) |
1012 | print(file=sys.stderr) | 1115 | wrapper = Wrapper() |
1013 | if no_repo_verify or _VerifyTag(rp): | 1116 | try: |
1014 | syncbuf = SyncBuffer(rp.config) | 1117 | rev = rp.bare_git.describe(rp.GetRevisionId()) |
1015 | rp.Sync_LocalHalf(syncbuf) | 1118 | except GitError: |
1016 | if not syncbuf.Finish(): | 1119 | rev = None |
1017 | sys.exit(1) | 1120 | _, new_rev = wrapper.check_repo_rev(rp.gitdir, rev, repo_verify=repo_verify) |
1121 | # See if we're held back due to missing signed tag. | ||
1122 | current_revid = rp.bare_git.rev_parse('HEAD') | ||
1123 | new_revid = rp.bare_git.rev_parse('--verify', new_rev) | ||
1124 | if current_revid != new_revid: | ||
1125 | # We want to switch to the new rev, but also not trash any uncommitted | ||
1126 | # changes. This helps with local testing/hacking. | ||
1127 | # If a local change has been made, we will throw that away. | ||
1128 | # We also have to make sure this will switch to an older commit if that's | ||
1129 | # the latest tag in order to support release rollback. | ||
1130 | try: | ||
1131 | rp.work_git.reset('--keep', new_rev) | ||
1132 | except GitError as e: | ||
1133 | sys.exit(str(e)) | ||
1018 | print('info: Restarting repo with latest version', file=sys.stderr) | 1134 | print('info: Restarting repo with latest version', file=sys.stderr) |
1019 | raise RepoChangedException(['--repo-upgraded']) | 1135 | raise RepoChangedException(['--repo-upgraded']) |
1020 | else: | 1136 | else: |
@@ -1024,53 +1140,6 @@ def _PostRepoFetch(rp, no_repo_verify=False, verbose=False): | |||
1024 | print('repo version %s is current' % rp.work_git.describe(HEAD), | 1140 | print('repo version %s is current' % rp.work_git.describe(HEAD), |
1025 | file=sys.stderr) | 1141 | file=sys.stderr) |
1026 | 1142 | ||
1027 | def _VerifyTag(project): | ||
1028 | gpg_dir = os.path.expanduser('~/.repoconfig/gnupg') | ||
1029 | if not os.path.exists(gpg_dir): | ||
1030 | print('warning: GnuPG was not available during last "repo init"\n' | ||
1031 | 'warning: Cannot automatically authenticate repo."""', | ||
1032 | file=sys.stderr) | ||
1033 | return True | ||
1034 | |||
1035 | try: | ||
1036 | cur = project.bare_git.describe(project.GetRevisionId()) | ||
1037 | except GitError: | ||
1038 | cur = None | ||
1039 | |||
1040 | if not cur \ | ||
1041 | or re.compile(r'^.*-[0-9]{1,}-g[0-9a-f]{1,}$').match(cur): | ||
1042 | rev = project.revisionExpr | ||
1043 | if rev.startswith(R_HEADS): | ||
1044 | rev = rev[len(R_HEADS):] | ||
1045 | |||
1046 | print(file=sys.stderr) | ||
1047 | print("warning: project '%s' branch '%s' is not signed" | ||
1048 | % (project.name, rev), file=sys.stderr) | ||
1049 | return False | ||
1050 | |||
1051 | env = os.environ.copy() | ||
1052 | env['GIT_DIR'] = project.gitdir.encode() | ||
1053 | env['GNUPGHOME'] = gpg_dir.encode() | ||
1054 | |||
1055 | cmd = [GIT, 'tag', '-v', cur] | ||
1056 | proc = subprocess.Popen(cmd, | ||
1057 | stdout = subprocess.PIPE, | ||
1058 | stderr = subprocess.PIPE, | ||
1059 | env = env) | ||
1060 | out = proc.stdout.read() | ||
1061 | proc.stdout.close() | ||
1062 | |||
1063 | err = proc.stderr.read() | ||
1064 | proc.stderr.close() | ||
1065 | |||
1066 | if proc.wait() != 0: | ||
1067 | print(file=sys.stderr) | ||
1068 | print(out, file=sys.stderr) | ||
1069 | print(err, file=sys.stderr) | ||
1070 | print(file=sys.stderr) | ||
1071 | return False | ||
1072 | return True | ||
1073 | |||
1074 | 1143 | ||
1075 | class _FetchTimes(object): | 1144 | class _FetchTimes(object): |
1076 | _ALPHA = 0.5 | 1145 | _ALPHA = 0.5 |
@@ -1090,7 +1159,7 @@ class _FetchTimes(object): | |||
1090 | old = self._times.get(name, t) | 1159 | old = self._times.get(name, t) |
1091 | self._seen.add(name) | 1160 | self._seen.add(name) |
1092 | a = self._ALPHA | 1161 | a = self._ALPHA |
1093 | self._times[name] = (a*t) + ((1-a) * old) | 1162 | self._times[name] = (a * t) + ((1 - a) * old) |
1094 | 1163 | ||
1095 | def _Load(self): | 1164 | def _Load(self): |
1096 | if self._times is None: | 1165 | if self._times is None: |
@@ -1098,10 +1167,7 @@ class _FetchTimes(object): | |||
1098 | with open(self._path) as f: | 1167 | with open(self._path) as f: |
1099 | self._times = json.load(f) | 1168 | self._times = json.load(f) |
1100 | except (IOError, ValueError): | 1169 | except (IOError, ValueError): |
1101 | try: | 1170 | platform_utils.remove(self._path, missing_ok=True) |
1102 | platform_utils.remove(self._path) | ||
1103 | except OSError: | ||
1104 | pass | ||
1105 | self._times = {} | 1171 | self._times = {} |
1106 | 1172 | ||
1107 | def Save(self): | 1173 | def Save(self): |
@@ -1119,15 +1185,14 @@ class _FetchTimes(object): | |||
1119 | with open(self._path, 'w') as f: | 1185 | with open(self._path, 'w') as f: |
1120 | json.dump(self._times, f, indent=2) | 1186 | json.dump(self._times, f, indent=2) |
1121 | except (IOError, TypeError): | 1187 | except (IOError, TypeError): |
1122 | try: | 1188 | platform_utils.remove(self._path, missing_ok=True) |
1123 | platform_utils.remove(self._path) | ||
1124 | except OSError: | ||
1125 | pass | ||
1126 | 1189 | ||
1127 | # This is a replacement for xmlrpc.client.Transport using urllib2 | 1190 | # This is a replacement for xmlrpc.client.Transport using urllib2 |
1128 | # and supporting persistent-http[s]. It cannot change hosts from | 1191 | # and supporting persistent-http[s]. It cannot change hosts from |
1129 | # request to request like the normal transport, the real url | 1192 | # request to request like the normal transport, the real url |
1130 | # is passed during initialization. | 1193 | # is passed during initialization. |
1194 | |||
1195 | |||
1131 | class PersistentTransport(xmlrpc.client.Transport): | 1196 | class PersistentTransport(xmlrpc.client.Transport): |
1132 | def __init__(self, orig_host): | 1197 | def __init__(self, orig_host): |
1133 | self.orig_host = orig_host | 1198 | self.orig_host = orig_host |
@@ -1138,7 +1203,7 @@ class PersistentTransport(xmlrpc.client.Transport): | |||
1138 | # Since we're only using them for HTTP, copy the file temporarily, | 1203 | # Since we're only using them for HTTP, copy the file temporarily, |
1139 | # stripping those prefixes away. | 1204 | # stripping those prefixes away. |
1140 | if cookiefile: | 1205 | if cookiefile: |
1141 | tmpcookiefile = tempfile.NamedTemporaryFile() | 1206 | tmpcookiefile = tempfile.NamedTemporaryFile(mode='w') |
1142 | tmpcookiefile.write("# HTTP Cookie File") | 1207 | tmpcookiefile.write("# HTTP Cookie File") |
1143 | try: | 1208 | try: |
1144 | with open(cookiefile) as f: | 1209 | with open(cookiefile) as f: |
@@ -1162,7 +1227,7 @@ class PersistentTransport(xmlrpc.client.Transport): | |||
1162 | if proxy: | 1227 | if proxy: |
1163 | proxyhandler = urllib.request.ProxyHandler({ | 1228 | proxyhandler = urllib.request.ProxyHandler({ |
1164 | "http": proxy, | 1229 | "http": proxy, |
1165 | "https": proxy }) | 1230 | "https": proxy}) |
1166 | 1231 | ||
1167 | opener = urllib.request.build_opener( | 1232 | opener = urllib.request.build_opener( |
1168 | urllib.request.HTTPCookieProcessor(cookiejar), | 1233 | urllib.request.HTTPCookieProcessor(cookiejar), |
@@ -1219,4 +1284,3 @@ class PersistentTransport(xmlrpc.client.Transport): | |||
1219 | 1284 | ||
1220 | def close(self): | 1285 | def close(self): |
1221 | pass | 1286 | pass |
1222 | |||