largefiles: update help
authorMads Kiilerich <madski@unity3d.com>
Mon, 15 Apr 2013 01:46:10 +0200
changeset 18975 aa8205a9f51a
parent 18974 d78a136a8036
child 18976 6734951e2d24
largefiles: update help Some clarifications, and some clean-up after --cache-largefiles was introduced.
hgext/largefiles/__init__.py
--- a/hgext/largefiles/__init__.py	Mon Apr 15 01:43:31 2013 +0200
+++ b/hgext/largefiles/__init__.py	Mon Apr 15 01:46:10 2013 +0200
@@ -41,24 +41,23 @@
 enabled for this to work.
 
 When you pull a changeset that affects largefiles from a remote
-repository, the largefiles for the changeset won't be pulled down.
-Instead, when you later update to such a revision, any largefiles
-needed by that revision are downloaded and cached (if they have
-never been downloaded before).  This means that network access may
-be required to update to changesets you have previously updated to.
+repository, the largefiles for the changeset will by default not be
+pulled down. However, when you update to such a revision, any
+largefiles needed by that revision are downloaded and cached (if
+they have never been downloaded before). One way to pull largefiles
+when pulling is thus to use --update, which will update your working
+copy to the latest pulled revision (and thereby downloading any new
+largefiles).
 
-If you know you are pulling from a non-default location and want to
-ensure that you will have the largefiles needed to merge or rebase
-with new heads that you are pulling, then you can pull with the
---cache-largefiles flag to pre-emptively download any largefiles
+If you just want to ensure that you will have the largefiles needed to
+merge or rebase with new heads that you are pulling, then you can pull
+with `--cache-largefiles` flag to pre-emptively download any largefiles
 that are new in the heads you are pulling.
 
-The one exception to the "largefiles won't be pulled until you update
-to a revision that changes them" rule is when you pull new heads.
-Because you could be pulling new heads (that you may later want to
-merge with) from a non-default location (that Mercurial won't know
-about later), when you pull new heads, largefiles revisions for those
-heads are downloaded and cached locally.
+Keep in mind that network access may now be required to update to
+changesets that you have not previously updated to. The nature of the
+largefiles extension means that updating is no longer guaranteed to
+be a local-only operation.
 
 If you already have large files tracked by Mercurial without the
 largefiles extension, you will need to convert your repository in