interfaces: create a new folder for interfaces and move repository.py in it
authorPulkit Goyal <pulkit@yandex-team.ru>
Sun, 18 Aug 2019 00:45:33 +0300
changeset 42813 268662aac075
parent 42807 383fdfa6bba9
child 42814 2c4f656c8e9f
interfaces: create a new folder for interfaces and move repository.py in it I was trying to understand current interfaces and write new ones and I realized we need to improve how current interfaces are organised. This creates a dedicated folder for defining interfaces and move `repository.py` which defines all the current interfaces inside it. Differential Revision: https://phab.mercurial-scm.org/D6741
contrib/import-checker.py
hgext/lfs/__init__.py
hgext/lfs/wrapper.py
hgext/narrow/__init__.py
hgext/narrow/narrowbundle2.py
hgext/narrow/narrowcommands.py
hgext/sqlitestore.py
mercurial/changegroup.py
mercurial/exchange.py
mercurial/exchangev2.py
mercurial/filelog.py
mercurial/hg.py
mercurial/httppeer.py
mercurial/interfaces/__init__.py
mercurial/interfaces/repository.py
mercurial/localrepo.py
mercurial/manifest.py
mercurial/narrowspec.py
mercurial/repository.py
mercurial/revlog.py
mercurial/revlogutils/constants.py
mercurial/streamclone.py
mercurial/testing/storage.py
mercurial/utils/storageutil.py
mercurial/wireprotov1peer.py
setup.py
tests/notcapable
tests/pullext.py
tests/simplestorerepo.py
tests/test-check-interfaces.py
tests/wireprotosimplecache.py
--- a/contrib/import-checker.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/contrib/import-checker.py	Sun Aug 18 00:45:33 2019 +0300
@@ -28,6 +28,7 @@
     'mercurial.hgweb.common',
     'mercurial.hgweb.request',
     'mercurial.i18n',
+    'mercurial.interfaces',
     'mercurial.node',
     # for revlog to re-export constant to extensions
     'mercurial.revlogutils.constants',
--- a/hgext/lfs/__init__.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/lfs/__init__.py	Sun Aug 18 00:45:33 2019 +0300
@@ -141,13 +141,16 @@
     minifileset,
     node,
     pycompat,
-    repository,
     revlog,
     scmutil,
     templateutil,
     util,
 )
 
+from mercurial.interfaces import (
+    repository,
+)
+
 from . import (
     blobstore,
     wireprotolfsserver,
--- a/hgext/lfs/wrapper.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/lfs/wrapper.py	Sun Aug 18 00:45:33 2019 +0300
@@ -21,7 +21,6 @@
     exchange,
     exthelper,
     localrepo,
-    repository,
     revlog,
     scmutil,
     upgrade,
@@ -30,6 +29,10 @@
     wireprotov1server,
 )
 
+from mercurial.interfaces import (
+    repository,
+)
+
 from mercurial.utils import (
     storageutil,
     stringutil,
--- a/hgext/narrow/__init__.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/narrow/__init__.py	Sun Aug 18 00:45:33 2019 +0300
@@ -17,6 +17,9 @@
 from mercurial import (
     localrepo,
     registrar,
+)
+
+from mercurial.interfaces import (
     repository,
 )
 
--- a/hgext/narrow/narrowbundle2.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/narrow/narrowbundle2.py	Sun Aug 18 00:45:33 2019 +0300
@@ -23,10 +23,12 @@
     localrepo,
     narrowspec,
     repair,
-    repository,
     util,
     wireprototypes,
 )
+from mercurial.interfaces import (
+    repository,
+)
 from mercurial.utils import (
     stringutil,
 )
--- a/hgext/narrow/narrowcommands.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/narrow/narrowcommands.py	Sun Aug 18 00:45:33 2019 +0300
@@ -25,12 +25,14 @@
     pycompat,
     registrar,
     repair,
-    repository,
     repoview,
     sparse,
     util,
     wireprototypes,
 )
+from mercurial.interfaces import (
+    repository,
+)
 
 table = {}
 command = registrar.command(table)
--- a/hgext/sqlitestore.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/hgext/sqlitestore.py	Sun Aug 18 00:45:33 2019 +0300
@@ -70,10 +70,12 @@
     mdiff,
     pycompat,
     registrar,
-    repository,
     util,
     verify,
 )
+from mercurial.interfaces import (
+    repository,
+)
 from mercurial.utils import (
     interfaceutil,
     storageutil,
--- a/mercurial/changegroup.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/changegroup.py	Sun Aug 18 00:45:33 2019 +0300
@@ -25,8 +25,11 @@
     mdiff,
     phases,
     pycompat,
+    util,
+)
+
+from .interfaces import (
     repository,
-    util,
 )
 
 _CHANGEGROUPV1_DELTA_HEADER = struct.Struct("20s20s20s20s")
--- a/mercurial/exchange.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/exchange.py	Sun Aug 18 00:45:33 2019 +0300
@@ -34,7 +34,6 @@
     phases,
     pushkey,
     pycompat,
-    repository,
     scmutil,
     sslutil,
     streamclone,
@@ -42,6 +41,9 @@
     util,
     wireprototypes,
 )
+from .interfaces import (
+    repository,
+)
 from .utils import (
     stringutil,
 )
--- a/mercurial/exchangev2.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/exchangev2.py	Sun Aug 18 00:45:33 2019 +0300
@@ -22,8 +22,10 @@
     narrowspec,
     phases,
     pycompat,
+    setdiscovery,
+)
+from .interfaces import (
     repository,
-    setdiscovery,
 )
 
 def pull(pullop):
--- a/mercurial/filelog.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/filelog.py	Sun Aug 18 00:45:33 2019 +0300
@@ -14,8 +14,10 @@
 )
 from . import (
     error,
+    revlog,
+)
+from .interfaces import (
     repository,
-    revlog,
 )
 from .utils import (
     interfaceutil,
--- a/mercurial/hg.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/hg.py	Sun Aug 18 00:45:33 2019 +0300
@@ -39,7 +39,6 @@
     node,
     phases,
     pycompat,
-    repository as repositorymod,
     scmutil,
     sshpeer,
     statichttprepo,
@@ -51,6 +50,10 @@
     vfs as vfsmod,
 )
 
+from .interfaces import (
+    repository as repositorymod,
+)
+
 release = lock.release
 
 # shared features
--- a/mercurial/httppeer.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/httppeer.py	Sun Aug 18 00:45:33 2019 +0300
@@ -16,12 +16,14 @@
 import weakref
 
 from .i18n import _
+from .interfaces import (
+    repository,
+)
 from . import (
     bundle2,
     error,
     httpconnection,
     pycompat,
-    repository,
     statichttprepo,
     url as urlmod,
     util,
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/mercurial/interfaces/repository.py	Sun Aug 18 00:45:33 2019 +0300
@@ -0,0 +1,1877 @@
+# repository.py - Interfaces and base classes for repositories and peers.
+#
+# Copyright 2017 Gregory Szorc <gregory.szorc@gmail.com>
+#
+# This software may be used and distributed according to the terms of the
+# GNU General Public License version 2 or any later version.
+
+from __future__ import absolute_import
+
+from ..i18n import _
+from .. import (
+    error,
+)
+from ..utils import (
+    interfaceutil,
+)
+
+# When narrowing is finalized and no longer subject to format changes,
+# we should move this to just "narrow" or similar.
+NARROW_REQUIREMENT = 'narrowhg-experimental'
+
+# Local repository feature string.
+
+# Revlogs are being used for file storage.
+REPO_FEATURE_REVLOG_FILE_STORAGE = b'revlogfilestorage'
+# The storage part of the repository is shared from an external source.
+REPO_FEATURE_SHARED_STORAGE = b'sharedstore'
+# LFS supported for backing file storage.
+REPO_FEATURE_LFS = b'lfs'
+# Repository supports being stream cloned.
+REPO_FEATURE_STREAM_CLONE = b'streamclone'
+# Files storage may lack data for all ancestors.
+REPO_FEATURE_SHALLOW_FILE_STORAGE = b'shallowfilestorage'
+
+REVISION_FLAG_CENSORED = 1 << 15
+REVISION_FLAG_ELLIPSIS = 1 << 14
+REVISION_FLAG_EXTSTORED = 1 << 13
+
+REVISION_FLAGS_KNOWN = (
+    REVISION_FLAG_CENSORED | REVISION_FLAG_ELLIPSIS | REVISION_FLAG_EXTSTORED)
+
+CG_DELTAMODE_STD = b'default'
+CG_DELTAMODE_PREV = b'previous'
+CG_DELTAMODE_FULL = b'fulltext'
+CG_DELTAMODE_P1 = b'p1'
+
+class ipeerconnection(interfaceutil.Interface):
+    """Represents a "connection" to a repository.
+
+    This is the base interface for representing a connection to a repository.
+    It holds basic properties and methods applicable to all peer types.
+
+    This is not a complete interface definition and should not be used
+    outside of this module.
+    """
+    ui = interfaceutil.Attribute("""ui.ui instance""")
+
+    def url():
+        """Returns a URL string representing this peer.
+
+        Currently, implementations expose the raw URL used to construct the
+        instance. It may contain credentials as part of the URL. The
+        expectations of the value aren't well-defined and this could lead to
+        data leakage.
+
+        TODO audit/clean consumers and more clearly define the contents of this
+        value.
+        """
+
+    def local():
+        """Returns a local repository instance.
+
+        If the peer represents a local repository, returns an object that
+        can be used to interface with it. Otherwise returns ``None``.
+        """
+
+    def peer():
+        """Returns an object conforming to this interface.
+
+        Most implementations will ``return self``.
+        """
+
+    def canpush():
+        """Returns a boolean indicating if this peer can be pushed to."""
+
+    def close():
+        """Close the connection to this peer.
+
+        This is called when the peer will no longer be used. Resources
+        associated with the peer should be cleaned up.
+        """
+
+class ipeercapabilities(interfaceutil.Interface):
+    """Peer sub-interface related to capabilities."""
+
+    def capable(name):
+        """Determine support for a named capability.
+
+        Returns ``False`` if capability not supported.
+
+        Returns ``True`` if boolean capability is supported. Returns a string
+        if capability support is non-boolean.
+
+        Capability strings may or may not map to wire protocol capabilities.
+        """
+
+    def requirecap(name, purpose):
+        """Require a capability to be present.
+
+        Raises a ``CapabilityError`` if the capability isn't present.
+        """
+
+class ipeercommands(interfaceutil.Interface):
+    """Client-side interface for communicating over the wire protocol.
+
+    This interface is used as a gateway to the Mercurial wire protocol.
+    methods commonly call wire protocol commands of the same name.
+    """
+
+    def branchmap():
+        """Obtain heads in named branches.
+
+        Returns a dict mapping branch name to an iterable of nodes that are
+        heads on that branch.
+        """
+
+    def capabilities():
+        """Obtain capabilities of the peer.
+
+        Returns a set of string capabilities.
+        """
+
+    def clonebundles():
+        """Obtains the clone bundles manifest for the repo.
+
+        Returns the manifest as unparsed bytes.
+        """
+
+    def debugwireargs(one, two, three=None, four=None, five=None):
+        """Used to facilitate debugging of arguments passed over the wire."""
+
+    def getbundle(source, **kwargs):
+        """Obtain remote repository data as a bundle.
+
+        This command is how the bulk of repository data is transferred from
+        the peer to the local repository
+
+        Returns a generator of bundle data.
+        """
+
+    def heads():
+        """Determine all known head revisions in the peer.
+
+        Returns an iterable of binary nodes.
+        """
+
+    def known(nodes):
+        """Determine whether multiple nodes are known.
+
+        Accepts an iterable of nodes whose presence to check for.
+
+        Returns an iterable of booleans indicating of the corresponding node
+        at that index is known to the peer.
+        """
+
+    def listkeys(namespace):
+        """Obtain all keys in a pushkey namespace.
+
+        Returns an iterable of key names.
+        """
+
+    def lookup(key):
+        """Resolve a value to a known revision.
+
+        Returns a binary node of the resolved revision on success.
+        """
+
+    def pushkey(namespace, key, old, new):
+        """Set a value using the ``pushkey`` protocol.
+
+        Arguments correspond to the pushkey namespace and key to operate on and
+        the old and new values for that key.
+
+        Returns a string with the peer result. The value inside varies by the
+        namespace.
+        """
+
+    def stream_out():
+        """Obtain streaming clone data.
+
+        Successful result should be a generator of data chunks.
+        """
+
+    def unbundle(bundle, heads, url):
+        """Transfer repository data to the peer.
+
+        This is how the bulk of data during a push is transferred.
+
+        Returns the integer number of heads added to the peer.
+        """
+
+class ipeerlegacycommands(interfaceutil.Interface):
+    """Interface for implementing support for legacy wire protocol commands.
+
+    Wire protocol commands transition to legacy status when they are no longer
+    used by modern clients. To facilitate identifying which commands are
+    legacy, the interfaces are split.
+    """
+
+    def between(pairs):
+        """Obtain nodes between pairs of nodes.
+
+        ``pairs`` is an iterable of node pairs.
+
+        Returns an iterable of iterables of nodes corresponding to each
+        requested pair.
+        """
+
+    def branches(nodes):
+        """Obtain ancestor changesets of specific nodes back to a branch point.
+
+        For each requested node, the peer finds the first ancestor node that is
+        a DAG root or is a merge.
+
+        Returns an iterable of iterables with the resolved values for each node.
+        """
+
+    def changegroup(nodes, source):
+        """Obtain a changegroup with data for descendants of specified nodes."""
+
+    def changegroupsubset(bases, heads, source):
+        pass
+
+class ipeercommandexecutor(interfaceutil.Interface):
+    """Represents a mechanism to execute remote commands.
+
+    This is the primary interface for requesting that wire protocol commands
+    be executed. Instances of this interface are active in a context manager
+    and have a well-defined lifetime. When the context manager exits, all
+    outstanding requests are waited on.
+    """
+
+    def callcommand(name, args):
+        """Request that a named command be executed.
+
+        Receives the command name and a dictionary of command arguments.
+
+        Returns a ``concurrent.futures.Future`` that will resolve to the
+        result of that command request. That exact value is left up to
+        the implementation and possibly varies by command.
+
+        Not all commands can coexist with other commands in an executor
+        instance: it depends on the underlying wire protocol transport being
+        used and the command itself.
+
+        Implementations MAY call ``sendcommands()`` automatically if the
+        requested command can not coexist with other commands in this executor.
+
+        Implementations MAY call ``sendcommands()`` automatically when the
+        future's ``result()`` is called. So, consumers using multiple
+        commands with an executor MUST ensure that ``result()`` is not called
+        until all command requests have been issued.
+        """
+
+    def sendcommands():
+        """Trigger submission of queued command requests.
+
+        Not all transports submit commands as soon as they are requested to
+        run. When called, this method forces queued command requests to be
+        issued. It will no-op if all commands have already been sent.
+
+        When called, no more new commands may be issued with this executor.
+        """
+
+    def close():
+        """Signal that this command request is finished.
+
+        When called, no more new commands may be issued. All outstanding
+        commands that have previously been issued are waited on before
+        returning. This not only includes waiting for the futures to resolve,
+        but also waiting for all response data to arrive. In other words,
+        calling this waits for all on-wire state for issued command requests
+        to finish.
+
+        When used as a context manager, this method is called when exiting the
+        context manager.
+
+        This method may call ``sendcommands()`` if there are buffered commands.
+        """
+
+class ipeerrequests(interfaceutil.Interface):
+    """Interface for executing commands on a peer."""
+
+    limitedarguments = interfaceutil.Attribute(
+        """True if the peer cannot receive large argument value for commands."""
+    )
+
+    def commandexecutor():
+        """A context manager that resolves to an ipeercommandexecutor.
+
+        The object this resolves to can be used to issue command requests
+        to the peer.
+
+        Callers should call its ``callcommand`` method to issue command
+        requests.
+
+        A new executor should be obtained for each distinct set of commands
+        (possibly just a single command) that the consumer wants to execute
+        as part of a single operation or round trip. This is because some
+        peers are half-duplex and/or don't support persistent connections.
+        e.g. in the case of HTTP peers, commands sent to an executor represent
+        a single HTTP request. While some peers may support multiple command
+        sends over the wire per executor, consumers need to code to the least
+        capable peer. So it should be assumed that command executors buffer
+        called commands until they are told to send them and that each
+        command executor could result in a new connection or wire-level request
+        being issued.
+        """
+
+class ipeerbase(ipeerconnection, ipeercapabilities, ipeerrequests):
+    """Unified interface for peer repositories.
+
+    All peer instances must conform to this interface.
+    """
+
+class ipeerv2(ipeerconnection, ipeercapabilities, ipeerrequests):
+    """Unified peer interface for wire protocol version 2 peers."""
+
+    apidescriptor = interfaceutil.Attribute(
+        """Data structure holding description of server API.""")
+
+@interfaceutil.implementer(ipeerbase)
+class peer(object):
+    """Base class for peer repositories."""
+
+    limitedarguments = False
+
+    def capable(self, name):
+        caps = self.capabilities()
+        if name in caps:
+            return True
+
+        name = '%s=' % name
+        for cap in caps:
+            if cap.startswith(name):
+                return cap[len(name):]
+
+        return False
+
+    def requirecap(self, name, purpose):
+        if self.capable(name):
+            return
+
+        raise error.CapabilityError(
+            _('cannot %s; remote repository does not support the '
+              '\'%s\' capability') % (purpose, name))
+
+class iverifyproblem(interfaceutil.Interface):
+    """Represents a problem with the integrity of the repository.
+
+    Instances of this interface are emitted to describe an integrity issue
+    with a repository (e.g. corrupt storage, missing data, etc).
+
+    Instances are essentially messages associated with severity.
+    """
+    warning = interfaceutil.Attribute(
+        """Message indicating a non-fatal problem.""")
+
+    error = interfaceutil.Attribute(
+        """Message indicating a fatal problem.""")
+
+    node = interfaceutil.Attribute(
+        """Revision encountering the problem.
+
+        ``None`` means the problem doesn't apply to a single revision.
+        """)
+
+class irevisiondelta(interfaceutil.Interface):
+    """Represents a delta between one revision and another.
+
+    Instances convey enough information to allow a revision to be exchanged
+    with another repository.
+
+    Instances represent the fulltext revision data or a delta against
+    another revision. Therefore the ``revision`` and ``delta`` attributes
+    are mutually exclusive.
+
+    Typically used for changegroup generation.
+    """
+
+    node = interfaceutil.Attribute(
+        """20 byte node of this revision.""")
+
+    p1node = interfaceutil.Attribute(
+        """20 byte node of 1st parent of this revision.""")
+
+    p2node = interfaceutil.Attribute(
+        """20 byte node of 2nd parent of this revision.""")
+
+    linknode = interfaceutil.Attribute(
+        """20 byte node of the changelog revision this node is linked to.""")
+
+    flags = interfaceutil.Attribute(
+        """2 bytes of integer flags that apply to this revision.
+
+        This is a bitwise composition of the ``REVISION_FLAG_*`` constants.
+        """)
+
+    basenode = interfaceutil.Attribute(
+        """20 byte node of the revision this data is a delta against.
+
+        ``nullid`` indicates that the revision is a full revision and not
+        a delta.
+        """)
+
+    baserevisionsize = interfaceutil.Attribute(
+        """Size of base revision this delta is against.
+
+        May be ``None`` if ``basenode`` is ``nullid``.
+        """)
+
+    revision = interfaceutil.Attribute(
+        """Raw fulltext of revision data for this node.""")
+
+    delta = interfaceutil.Attribute(
+        """Delta between ``basenode`` and ``node``.
+
+        Stored in the bdiff delta format.
+        """)
+
+class ifilerevisionssequence(interfaceutil.Interface):
+    """Contains index data for all revisions of a file.
+
+    Types implementing this behave like lists of tuples. The index
+    in the list corresponds to the revision number. The values contain
+    index metadata.
+
+    The *null* revision (revision number -1) is always the last item
+    in the index.
+    """
+
+    def __len__():
+        """The total number of revisions."""
+
+    def __getitem__(rev):
+        """Returns the object having a specific revision number.
+
+        Returns an 8-tuple with the following fields:
+
+        offset+flags
+           Contains the offset and flags for the revision. 64-bit unsigned
+           integer where first 6 bytes are the offset and the next 2 bytes
+           are flags. The offset can be 0 if it is not used by the store.
+        compressed size
+            Size of the revision data in the store. It can be 0 if it isn't
+            needed by the store.
+        uncompressed size
+            Fulltext size. It can be 0 if it isn't needed by the store.
+        base revision
+            Revision number of revision the delta for storage is encoded
+            against. -1 indicates not encoded against a base revision.
+        link revision
+            Revision number of changelog revision this entry is related to.
+        p1 revision
+            Revision number of 1st parent. -1 if no 1st parent.
+        p2 revision
+            Revision number of 2nd parent. -1 if no 1st parent.
+        node
+            Binary node value for this revision number.
+
+        Negative values should index off the end of the sequence. ``-1``
+        should return the null revision. ``-2`` should return the most
+        recent revision.
+        """
+
+    def __contains__(rev):
+        """Whether a revision number exists."""
+
+    def insert(self, i, entry):
+        """Add an item to the index at specific revision."""
+
+class ifileindex(interfaceutil.Interface):
+    """Storage interface for index data of a single file.
+
+    File storage data is divided into index metadata and data storage.
+    This interface defines the index portion of the interface.
+
+    The index logically consists of:
+
+    * A mapping between revision numbers and nodes.
+    * DAG data (storing and querying the relationship between nodes).
+    * Metadata to facilitate storage.
+    """
+    def __len__():
+        """Obtain the number of revisions stored for this file."""
+
+    def __iter__():
+        """Iterate over revision numbers for this file."""
+
+    def hasnode(node):
+        """Returns a bool indicating if a node is known to this store.
+
+        Implementations must only return True for full, binary node values:
+        hex nodes, revision numbers, and partial node matches must be
+        rejected.
+
+        The null node is never present.
+        """
+
+    def revs(start=0, stop=None):
+        """Iterate over revision numbers for this file, with control."""
+
+    def parents(node):
+        """Returns a 2-tuple of parent nodes for a revision.
+
+        Values will be ``nullid`` if the parent is empty.
+        """
+
+    def parentrevs(rev):
+        """Like parents() but operates on revision numbers."""
+
+    def rev(node):
+        """Obtain the revision number given a node.
+
+        Raises ``error.LookupError`` if the node is not known.
+        """
+
+    def node(rev):
+        """Obtain the node value given a revision number.
+
+        Raises ``IndexError`` if the node is not known.
+        """
+
+    def lookup(node):
+        """Attempt to resolve a value to a node.
+
+        Value can be a binary node, hex node, revision number, or a string
+        that can be converted to an integer.
+
+        Raises ``error.LookupError`` if a node could not be resolved.
+        """
+
+    def linkrev(rev):
+        """Obtain the changeset revision number a revision is linked to."""
+
+    def iscensored(rev):
+        """Return whether a revision's content has been censored."""
+
+    def commonancestorsheads(node1, node2):
+        """Obtain an iterable of nodes containing heads of common ancestors.
+
+        See ``ancestor.commonancestorsheads()``.
+        """
+
+    def descendants(revs):
+        """Obtain descendant revision numbers for a set of revision numbers.
+
+        If ``nullrev`` is in the set, this is equivalent to ``revs()``.
+        """
+
+    def heads(start=None, stop=None):
+        """Obtain a list of nodes that are DAG heads, with control.
+
+        The set of revisions examined can be limited by specifying
+        ``start`` and ``stop``. ``start`` is a node. ``stop`` is an
+        iterable of nodes. DAG traversal starts at earlier revision
+        ``start`` and iterates forward until any node in ``stop`` is
+        encountered.
+        """
+
+    def children(node):
+        """Obtain nodes that are children of a node.
+
+        Returns a list of nodes.
+        """
+
+class ifiledata(interfaceutil.Interface):
+    """Storage interface for data storage of a specific file.
+
+    This complements ``ifileindex`` and provides an interface for accessing
+    data for a tracked file.
+    """
+    def size(rev):
+        """Obtain the fulltext size of file data.
+
+        Any metadata is excluded from size measurements.
+        """
+
+    def revision(node, raw=False):
+        """"Obtain fulltext data for a node.
+
+        By default, any storage transformations are applied before the data
+        is returned. If ``raw`` is True, non-raw storage transformations
+        are not applied.
+
+        The fulltext data may contain a header containing metadata. Most
+        consumers should use ``read()`` to obtain the actual file data.
+        """
+
+    def rawdata(node):
+        """Obtain raw data for a node.
+        """
+
+    def read(node):
+        """Resolve file fulltext data.
+
+        This is similar to ``revision()`` except any metadata in the data
+        headers is stripped.
+        """
+
+    def renamed(node):
+        """Obtain copy metadata for a node.
+
+        Returns ``False`` if no copy metadata is stored or a 2-tuple of
+        (path, node) from which this revision was copied.
+        """
+
+    def cmp(node, fulltext):
+        """Compare fulltext to another revision.
+
+        Returns True if the fulltext is different from what is stored.
+
+        This takes copy metadata into account.
+
+        TODO better document the copy metadata and censoring logic.
+        """
+
+    def emitrevisions(nodes,
+                      nodesorder=None,
+                      revisiondata=False,
+                      assumehaveparentrevisions=False,
+                      deltamode=CG_DELTAMODE_STD):
+        """Produce ``irevisiondelta`` for revisions.
+
+        Given an iterable of nodes, emits objects conforming to the
+        ``irevisiondelta`` interface that describe revisions in storage.
+
+        This method is a generator.
+
+        The input nodes may be unordered. Implementations must ensure that a
+        node's parents are emitted before the node itself. Transitively, this
+        means that a node may only be emitted once all its ancestors in
+        ``nodes`` have also been emitted.
+
+        By default, emits "index" data (the ``node``, ``p1node``, and
+        ``p2node`` attributes). If ``revisiondata`` is set, revision data
+        will also be present on the emitted objects.
+
+        With default argument values, implementations can choose to emit
+        either fulltext revision data or a delta. When emitting deltas,
+        implementations must consider whether the delta's base revision
+        fulltext is available to the receiver.
+
+        The base revision fulltext is guaranteed to be available if any of
+        the following are met:
+
+        * Its fulltext revision was emitted by this method call.
+        * A delta for that revision was emitted by this method call.
+        * ``assumehaveparentrevisions`` is True and the base revision is a
+          parent of the node.
+
+        ``nodesorder`` can be used to control the order that revisions are
+        emitted. By default, revisions can be reordered as long as they are
+        in DAG topological order (see above). If the value is ``nodes``,
+        the iteration order from ``nodes`` should be used. If the value is
+        ``storage``, then the native order from the backing storage layer
+        is used. (Not all storage layers will have strong ordering and behavior
+        of this mode is storage-dependent.) ``nodes`` ordering can force
+        revisions to be emitted before their ancestors, so consumers should
+        use it with care.
+
+        The ``linknode`` attribute on the returned ``irevisiondelta`` may not
+        be set and it is the caller's responsibility to resolve it, if needed.
+
+        If ``deltamode`` is CG_DELTAMODE_PREV and revision data is requested,
+        all revision data should be emitted as deltas against the revision
+        emitted just prior. The initial revision should be a delta against its
+        1st parent.
+        """
+
+class ifilemutation(interfaceutil.Interface):
+    """Storage interface for mutation events of a tracked file."""
+
+    def add(filedata, meta, transaction, linkrev, p1, p2):
+        """Add a new revision to the store.
+
+        Takes file data, dictionary of metadata, a transaction, linkrev,
+        and parent nodes.
+
+        Returns the node that was added.
+
+        May no-op if a revision matching the supplied data is already stored.
+        """
+
+    def addrevision(revisiondata, transaction, linkrev, p1, p2, node=None,
+                    flags=0, cachedelta=None):
+        """Add a new revision to the store.
+
+        This is similar to ``add()`` except it operates at a lower level.
+
+        The data passed in already contains a metadata header, if any.
+
+        ``node`` and ``flags`` can be used to define the expected node and
+        the flags to use with storage. ``flags`` is a bitwise value composed
+        of the various ``REVISION_FLAG_*`` constants.
+
+        ``add()`` is usually called when adding files from e.g. the working
+        directory. ``addrevision()`` is often called by ``add()`` and for
+        scenarios where revision data has already been computed, such as when
+        applying raw data from a peer repo.
+        """
+
+    def addgroup(deltas, linkmapper, transaction, addrevisioncb=None,
+                 maybemissingparents=False):
+        """Process a series of deltas for storage.
+
+        ``deltas`` is an iterable of 7-tuples of
+        (node, p1, p2, linknode, deltabase, delta, flags) defining revisions
+        to add.
+
+        The ``delta`` field contains ``mpatch`` data to apply to a base
+        revision, identified by ``deltabase``. The base node can be
+        ``nullid``, in which case the header from the delta can be ignored
+        and the delta used as the fulltext.
+
+        ``addrevisioncb`` should be called for each node as it is committed.
+
+        ``maybemissingparents`` is a bool indicating whether the incoming
+        data may reference parents/ancestor revisions that aren't present.
+        This flag is set when receiving data into a "shallow" store that
+        doesn't hold all history.
+
+        Returns a list of nodes that were processed. A node will be in the list
+        even if it existed in the store previously.
+        """
+
+    def censorrevision(tr, node, tombstone=b''):
+        """Remove the content of a single revision.
+
+        The specified ``node`` will have its content purged from storage.
+        Future attempts to access the revision data for this node will
+        result in failure.
+
+        A ``tombstone`` message can optionally be stored. This message may be
+        displayed to users when they attempt to access the missing revision
+        data.
+
+        Storage backends may have stored deltas against the previous content
+        in this revision. As part of censoring a revision, these storage
+        backends are expected to rewrite any internally stored deltas such
+        that they no longer reference the deleted content.
+        """
+
+    def getstrippoint(minlink):
+        """Find the minimum revision that must be stripped to strip a linkrev.
+
+        Returns a 2-tuple containing the minimum revision number and a set
+        of all revisions numbers that would be broken by this strip.
+
+        TODO this is highly revlog centric and should be abstracted into
+        a higher-level deletion API. ``repair.strip()`` relies on this.
+        """
+
+    def strip(minlink, transaction):
+        """Remove storage of items starting at a linkrev.
+
+        This uses ``getstrippoint()`` to determine the first node to remove.
+        Then it effectively truncates storage for all revisions after that.
+
+        TODO this is highly revlog centric and should be abstracted into a
+        higher-level deletion API.
+        """
+
+class ifilestorage(ifileindex, ifiledata, ifilemutation):
+    """Complete storage interface for a single tracked file."""
+
+    def files():
+        """Obtain paths that are backing storage for this file.
+
+        TODO this is used heavily by verify code and there should probably
+        be a better API for that.
+        """
+
+    def storageinfo(exclusivefiles=False, sharedfiles=False,
+                    revisionscount=False, trackedsize=False,
+                    storedsize=False):
+        """Obtain information about storage for this file's data.
+
+        Returns a dict describing storage for this tracked path. The keys
+        in the dict map to arguments of the same. The arguments are bools
+        indicating whether to calculate and obtain that data.
+
+        exclusivefiles
+           Iterable of (vfs, path) describing files that are exclusively
+           used to back storage for this tracked path.
+
+        sharedfiles
+           Iterable of (vfs, path) describing files that are used to back
+           storage for this tracked path. Those files may also provide storage
+           for other stored entities.
+
+        revisionscount
+           Number of revisions available for retrieval.
+
+        trackedsize
+           Total size in bytes of all tracked revisions. This is a sum of the
+           length of the fulltext of all revisions.
+
+        storedsize
+           Total size in bytes used to store data for all tracked revisions.
+           This is commonly less than ``trackedsize`` due to internal usage
+           of deltas rather than fulltext revisions.
+
+        Not all storage backends may support all queries are have a reasonable
+        value to use. In that case, the value should be set to ``None`` and
+        callers are expected to handle this special value.
+        """
+
+    def verifyintegrity(state):
+        """Verifies the integrity of file storage.
+
+        ``state`` is a dict holding state of the verifier process. It can be
+        used to communicate data between invocations of multiple storage
+        primitives.
+
+        If individual revisions cannot have their revision content resolved,
+        the method is expected to set the ``skipread`` key to a set of nodes
+        that encountered problems.
+
+        The method yields objects conforming to the ``iverifyproblem``
+        interface.
+        """
+
+class idirs(interfaceutil.Interface):
+    """Interface representing a collection of directories from paths.
+
+    This interface is essentially a derived data structure representing
+    directories from a collection of paths.
+    """
+
+    def addpath(path):
+        """Add a path to the collection.
+
+        All directories in the path will be added to the collection.
+        """
+
+    def delpath(path):
+        """Remove a path from the collection.
+
+        If the removal was the last path in a particular directory, the
+        directory is removed from the collection.
+        """
+
+    def __iter__():
+        """Iterate over the directories in this collection of paths."""
+
+    def __contains__(path):
+        """Whether a specific directory is in this collection."""
+
+class imanifestdict(interfaceutil.Interface):
+    """Interface representing a manifest data structure.
+
+    A manifest is effectively a dict mapping paths to entries. Each entry
+    consists of a binary node and extra flags affecting that entry.
+    """
+
+    def __getitem__(path):
+        """Returns the binary node value for a path in the manifest.
+
+        Raises ``KeyError`` if the path does not exist in the manifest.
+
+        Equivalent to ``self.find(path)[0]``.
+        """
+
+    def find(path):
+        """Returns the entry for a path in the manifest.
+
+        Returns a 2-tuple of (node, flags).
+
+        Raises ``KeyError`` if the path does not exist in the manifest.
+        """
+
+    def __len__():
+        """Return the number of entries in the manifest."""
+
+    def __nonzero__():
+        """Returns True if the manifest has entries, False otherwise."""
+
+    __bool__ = __nonzero__
+
+    def __setitem__(path, node):
+        """Define the node value for a path in the manifest.
+
+        If the path is already in the manifest, its flags will be copied to
+        the new entry.
+        """
+
+    def __contains__(path):
+        """Whether a path exists in the manifest."""
+
+    def __delitem__(path):
+        """Remove a path from the manifest.
+
+        Raises ``KeyError`` if the path is not in the manifest.
+        """
+
+    def __iter__():
+        """Iterate over paths in the manifest."""
+
+    def iterkeys():
+        """Iterate over paths in the manifest."""
+
+    def keys():
+        """Obtain a list of paths in the manifest."""
+
+    def filesnotin(other, match=None):
+        """Obtain the set of paths in this manifest but not in another.
+
+        ``match`` is an optional matcher function to be applied to both
+        manifests.
+
+        Returns a set of paths.
+        """
+
+    def dirs():
+        """Returns an object implementing the ``idirs`` interface."""
+
+    def hasdir(dir):
+        """Returns a bool indicating if a directory is in this manifest."""
+
+    def matches(match):
+        """Generate a new manifest filtered through a matcher.
+
+        Returns an object conforming to the ``imanifestdict`` interface.
+        """
+
+    def walk(match):
+        """Generator of paths in manifest satisfying a matcher.
+
+        This is equivalent to ``self.matches(match).iterkeys()`` except a new
+        manifest object is not created.
+
+        If the matcher has explicit files listed and they don't exist in
+        the manifest, ``match.bad()`` is called for each missing file.
+        """
+
+    def diff(other, match=None, clean=False):
+        """Find differences between this manifest and another.
+
+        This manifest is compared to ``other``.
+
+        If ``match`` is provided, the two manifests are filtered against this
+        matcher and only entries satisfying the matcher are compared.
+
+        If ``clean`` is True, unchanged files are included in the returned
+        object.
+
+        Returns a dict with paths as keys and values of 2-tuples of 2-tuples of
+        the form ``((node1, flag1), (node2, flag2))`` where ``(node1, flag1)``
+        represents the node and flags for this manifest and ``(node2, flag2)``
+        are the same for the other manifest.
+        """
+
+    def setflag(path, flag):
+        """Set the flag value for a given path.
+
+        Raises ``KeyError`` if the path is not already in the manifest.
+        """
+
+    def get(path, default=None):
+        """Obtain the node value for a path or a default value if missing."""
+
+    def flags(path, default=''):
+        """Return the flags value for a path or a default value if missing."""
+
+    def copy():
+        """Return a copy of this manifest."""
+
+    def items():
+        """Returns an iterable of (path, node) for items in this manifest."""
+
+    def iteritems():
+        """Identical to items()."""
+
+    def iterentries():
+        """Returns an iterable of (path, node, flags) for this manifest.
+
+        Similar to ``iteritems()`` except items are a 3-tuple and include
+        flags.
+        """
+
+    def text():
+        """Obtain the raw data representation for this manifest.
+
+        Result is used to create a manifest revision.
+        """
+
+    def fastdelta(base, changes):
+        """Obtain a delta between this manifest and another given changes.
+
+        ``base`` in the raw data representation for another manifest.
+
+        ``changes`` is an iterable of ``(path, to_delete)``.
+
+        Returns a 2-tuple containing ``bytearray(self.text())`` and the
+        delta between ``base`` and this manifest.
+        """
+
+class imanifestrevisionbase(interfaceutil.Interface):
+    """Base interface representing a single revision of a manifest.
+
+    Should not be used as a primary interface: should always be inherited
+    as part of a larger interface.
+    """
+
+    def new():
+        """Obtain a new manifest instance.
+
+        Returns an object conforming to the ``imanifestrevisionwritable``
+        interface. The instance will be associated with the same
+        ``imanifestlog`` collection as this instance.
+        """
+
+    def copy():
+        """Obtain a copy of this manifest instance.
+
+        Returns an object conforming to the ``imanifestrevisionwritable``
+        interface. The instance will be associated with the same
+        ``imanifestlog`` collection as this instance.
+        """
+
+    def read():
+        """Obtain the parsed manifest data structure.
+
+        The returned object conforms to the ``imanifestdict`` interface.
+        """
+
+class imanifestrevisionstored(imanifestrevisionbase):
+    """Interface representing a manifest revision committed to storage."""
+
+    def node():
+        """The binary node for this manifest."""
+
+    parents = interfaceutil.Attribute(
+        """List of binary nodes that are parents for this manifest revision."""
+    )
+
+    def readdelta(shallow=False):
+        """Obtain the manifest data structure representing changes from parent.
+
+        This manifest is compared to its 1st parent. A new manifest representing
+        those differences is constructed.
+
+        The returned object conforms to the ``imanifestdict`` interface.
+        """
+
+    def readfast(shallow=False):
+        """Calls either ``read()`` or ``readdelta()``.
+
+        The faster of the two options is called.
+        """
+
+    def find(key):
+        """Calls self.read().find(key)``.
+
+        Returns a 2-tuple of ``(node, flags)`` or raises ``KeyError``.
+        """
+
+class imanifestrevisionwritable(imanifestrevisionbase):
+    """Interface representing a manifest revision that can be committed."""
+
+    def write(transaction, linkrev, p1node, p2node, added, removed, match=None):
+        """Add this revision to storage.
+
+        Takes a transaction object, the changeset revision number it will
+        be associated with, its parent nodes, and lists of added and
+        removed paths.
+
+        If match is provided, storage can choose not to inspect or write out
+        items that do not match. Storage is still required to be able to provide
+        the full manifest in the future for any directories written (these
+        manifests should not be "narrowed on disk").
+
+        Returns the binary node of the created revision.
+        """
+
+class imanifeststorage(interfaceutil.Interface):
+    """Storage interface for manifest data."""
+
+    tree = interfaceutil.Attribute(
+        """The path to the directory this manifest tracks.
+
+        The empty bytestring represents the root manifest.
+        """)
+
+    index = interfaceutil.Attribute(
+        """An ``ifilerevisionssequence`` instance.""")
+
+    indexfile = interfaceutil.Attribute(
+        """Path of revlog index file.
+
+        TODO this is revlog specific and should not be exposed.
+        """)
+
+    opener = interfaceutil.Attribute(
+        """VFS opener to use to access underlying files used for storage.
+
+        TODO this is revlog specific and should not be exposed.
+        """)
+
+    version = interfaceutil.Attribute(
+        """Revlog version number.
+
+        TODO this is revlog specific and should not be exposed.
+        """)
+
+    _generaldelta = interfaceutil.Attribute(
+        """Whether generaldelta storage is being used.
+
+        TODO this is revlog specific and should not be exposed.
+        """)
+
+    fulltextcache = interfaceutil.Attribute(
+        """Dict with cache of fulltexts.
+
+        TODO this doesn't feel appropriate for the storage interface.
+        """)
+
+    def __len__():
+        """Obtain the number of revisions stored for this manifest."""
+
+    def __iter__():
+        """Iterate over revision numbers for this manifest."""
+
+    def rev(node):
+        """Obtain the revision number given a binary node.
+
+        Raises ``error.LookupError`` if the node is not known.
+        """
+
+    def node(rev):
+        """Obtain the node value given a revision number.
+
+        Raises ``error.LookupError`` if the revision is not known.
+        """
+
+    def lookup(value):
+        """Attempt to resolve a value to a node.
+
+        Value can be a binary node, hex node, revision number, or a bytes
+        that can be converted to an integer.
+
+        Raises ``error.LookupError`` if a ndoe could not be resolved.
+        """
+
+    def parents(node):
+        """Returns a 2-tuple of parent nodes for a node.
+
+        Values will be ``nullid`` if the parent is empty.
+        """
+
+    def parentrevs(rev):
+        """Like parents() but operates on revision numbers."""
+
+    def linkrev(rev):
+        """Obtain the changeset revision number a revision is linked to."""
+
+    def revision(node, _df=None, raw=False):
+        """Obtain fulltext data for a node."""
+
+    def rawdata(node, _df=None):
+        """Obtain raw data for a node."""
+
+    def revdiff(rev1, rev2):
+        """Obtain a delta between two revision numbers.
+
+        The returned data is the result of ``bdiff.bdiff()`` on the raw
+        revision data.
+        """
+
+    def cmp(node, fulltext):
+        """Compare fulltext to another revision.
+
+        Returns True if the fulltext is different from what is stored.
+        """
+
+    def emitrevisions(nodes,
+                      nodesorder=None,
+                      revisiondata=False,
+                      assumehaveparentrevisions=False):
+        """Produce ``irevisiondelta`` describing revisions.
+
+        See the documentation for ``ifiledata`` for more.
+        """
+
+    def addgroup(deltas, linkmapper, transaction, addrevisioncb=None):
+        """Process a series of deltas for storage.
+
+        See the documentation in ``ifilemutation`` for more.
+        """
+
+    def rawsize(rev):
+        """Obtain the size of tracked data.
+
+        Is equivalent to ``len(m.rawdata(node))``.
+
+        TODO this method is only used by upgrade code and may be removed.
+        """
+
+    def getstrippoint(minlink):
+        """Find minimum revision that must be stripped to strip a linkrev.
+
+        See the documentation in ``ifilemutation`` for more.
+        """
+
+    def strip(minlink, transaction):
+        """Remove storage of items starting at a linkrev.
+
+        See the documentation in ``ifilemutation`` for more.
+        """
+
+    def checksize():
+        """Obtain the expected sizes of backing files.
+
+        TODO this is used by verify and it should not be part of the interface.
+        """
+
+    def files():
+        """Obtain paths that are backing storage for this manifest.
+
+        TODO this is used by verify and there should probably be a better API
+        for this functionality.
+        """
+
+    def deltaparent(rev):
+        """Obtain the revision that a revision is delta'd against.
+
+        TODO delta encoding is an implementation detail of storage and should
+        not be exposed to the storage interface.
+        """
+
+    def clone(tr, dest, **kwargs):
+        """Clone this instance to another."""
+
+    def clearcaches(clear_persisted_data=False):
+        """Clear any caches associated with this instance."""
+
+    def dirlog(d):
+        """Obtain a manifest storage instance for a tree."""
+
+    def add(m, transaction, link, p1, p2, added, removed, readtree=None,
+            match=None):
+        """Add a revision to storage.
+
+        ``m`` is an object conforming to ``imanifestdict``.
+
+        ``link`` is the linkrev revision number.
+
+        ``p1`` and ``p2`` are the parent revision numbers.
+
+        ``added`` and ``removed`` are iterables of added and removed paths,
+        respectively.
+
+        ``readtree`` is a function that can be used to read the child tree(s)
+        when recursively writing the full tree structure when using
+        treemanifets.
+
+        ``match`` is a matcher that can be used to hint to storage that not all
+        paths must be inspected; this is an optimization and can be safely
+        ignored. Note that the storage must still be able to reproduce a full
+        manifest including files that did not match.
+        """
+
+    def storageinfo(exclusivefiles=False, sharedfiles=False,
+                    revisionscount=False, trackedsize=False,
+                    storedsize=False):
+        """Obtain information about storage for this manifest's data.
+
+        See ``ifilestorage.storageinfo()`` for a description of this method.
+        This one behaves the same way, except for manifest data.
+        """
+
+class imanifestlog(interfaceutil.Interface):
+    """Interface representing a collection of manifest snapshots.
+
+    Represents the root manifest in a repository.
+
+    Also serves as a means to access nested tree manifests and to cache
+    tree manifests.
+    """
+
+    def __getitem__(node):
+        """Obtain a manifest instance for a given binary node.
+
+        Equivalent to calling ``self.get('', node)``.
+
+        The returned object conforms to the ``imanifestrevisionstored``
+        interface.
+        """
+
+    def get(tree, node, verify=True):
+        """Retrieve the manifest instance for a given directory and binary node.
+
+        ``node`` always refers to the node of the root manifest (which will be
+        the only manifest if flat manifests are being used).
+
+        If ``tree`` is the empty string, the root manifest is returned.
+        Otherwise the manifest for the specified directory will be returned
+        (requires tree manifests).
+
+        If ``verify`` is True, ``LookupError`` is raised if the node is not
+        known.
+
+        The returned object conforms to the ``imanifestrevisionstored``
+        interface.
+        """
+
+    def getstorage(tree):
+        """Retrieve an interface to storage for a particular tree.
+
+        If ``tree`` is the empty bytestring, storage for the root manifest will
+        be returned. Otherwise storage for a tree manifest is returned.
+
+        TODO formalize interface for returned object.
+        """
+
+    def clearcaches():
+        """Clear caches associated with this collection."""
+
+    def rev(node):
+        """Obtain the revision number for a binary node.
+
+        Raises ``error.LookupError`` if the node is not known.
+        """
+
+class ilocalrepositoryfilestorage(interfaceutil.Interface):
+    """Local repository sub-interface providing access to tracked file storage.
+
+    This interface defines how a repository accesses storage for a single
+    tracked file path.
+    """
+
+    def file(f):
+        """Obtain a filelog for a tracked path.
+
+        The returned type conforms to the ``ifilestorage`` interface.
+        """
+
+class ilocalrepositorymain(interfaceutil.Interface):
+    """Main interface for local repositories.
+
+    This currently captures the reality of things - not how things should be.
+    """
+
+    supportedformats = interfaceutil.Attribute(
+        """Set of requirements that apply to stream clone.
+
+        This is actually a class attribute and is shared among all instances.
+        """)
+
+    supported = interfaceutil.Attribute(
+        """Set of requirements that this repo is capable of opening.""")
+
+    requirements = interfaceutil.Attribute(
+        """Set of requirements this repo uses.""")
+
+    features = interfaceutil.Attribute(
+        """Set of "features" this repository supports.
+
+        A "feature" is a loosely-defined term. It can refer to a feature
+        in the classical sense or can describe an implementation detail
+        of the repository. For example, a ``readonly`` feature may denote
+        the repository as read-only. Or a ``revlogfilestore`` feature may
+        denote that the repository is using revlogs for file storage.
+
+        The intent of features is to provide a machine-queryable mechanism
+        for repo consumers to test for various repository characteristics.
+
+        Features are similar to ``requirements``. The main difference is that
+        requirements are stored on-disk and represent requirements to open the
+        repository. Features are more run-time capabilities of the repository
+        and more granular capabilities (which may be derived from requirements).
+        """)
+
+    filtername = interfaceutil.Attribute(
+        """Name of the repoview that is active on this repo.""")
+
+    wvfs = interfaceutil.Attribute(
+        """VFS used to access the working directory.""")
+
+    vfs = interfaceutil.Attribute(
+        """VFS rooted at the .hg directory.
+
+        Used to access repository data not in the store.
+        """)
+
+    svfs = interfaceutil.Attribute(
+        """VFS rooted at the store.
+
+        Used to access repository data in the store. Typically .hg/store.
+        But can point elsewhere if the store is shared.
+        """)
+
+    root = interfaceutil.Attribute(
+        """Path to the root of the working directory.""")
+
+    path = interfaceutil.Attribute(
+        """Path to the .hg directory.""")
+
+    origroot = interfaceutil.Attribute(
+        """The filesystem path that was used to construct the repo.""")
+
+    auditor = interfaceutil.Attribute(
+        """A pathauditor for the working directory.
+
+        This checks if a path refers to a nested repository.
+
+        Operates on the filesystem.
+        """)
+
+    nofsauditor = interfaceutil.Attribute(
+        """A pathauditor for the working directory.
+
+        This is like ``auditor`` except it doesn't do filesystem checks.
+        """)
+
+    baseui = interfaceutil.Attribute(
+        """Original ui instance passed into constructor.""")
+
+    ui = interfaceutil.Attribute(
+        """Main ui instance for this instance.""")
+
+    sharedpath = interfaceutil.Attribute(
+        """Path to the .hg directory of the repo this repo was shared from.""")
+
+    store = interfaceutil.Attribute(
+        """A store instance.""")
+
+    spath = interfaceutil.Attribute(
+        """Path to the store.""")
+
+    sjoin = interfaceutil.Attribute(
+        """Alias to self.store.join.""")
+
+    cachevfs = interfaceutil.Attribute(
+        """A VFS used to access the cache directory.
+
+        Typically .hg/cache.
+        """)
+
+    wcachevfs = interfaceutil.Attribute(
+        """A VFS used to access the cache directory dedicated to working copy
+
+        Typically .hg/wcache.
+        """)
+
+    filteredrevcache = interfaceutil.Attribute(
+        """Holds sets of revisions to be filtered.""")
+
+    names = interfaceutil.Attribute(
+        """A ``namespaces`` instance.""")
+
+    def close():
+        """Close the handle on this repository."""
+
+    def peer():
+        """Obtain an object conforming to the ``peer`` interface."""
+
+    def unfiltered():
+        """Obtain an unfiltered/raw view of this repo."""
+
+    def filtered(name, visibilityexceptions=None):
+        """Obtain a named view of this repository."""
+
+    obsstore = interfaceutil.Attribute(
+        """A store of obsolescence data.""")
+
+    changelog = interfaceutil.Attribute(
+        """A handle on the changelog revlog.""")
+
+    manifestlog = interfaceutil.Attribute(
+        """An instance conforming to the ``imanifestlog`` interface.
+
+        Provides access to manifests for the repository.
+        """)
+
+    dirstate = interfaceutil.Attribute(
+        """Working directory state.""")
+
+    narrowpats = interfaceutil.Attribute(
+        """Matcher patterns for this repository's narrowspec.""")
+
+    def narrowmatch(match=None, includeexact=False):
+        """Obtain a matcher for the narrowspec."""
+
+    def setnarrowpats(newincludes, newexcludes):
+        """Define the narrowspec for this repository."""
+
+    def __getitem__(changeid):
+        """Try to resolve a changectx."""
+
+    def __contains__(changeid):
+        """Whether a changeset exists."""
+
+    def __nonzero__():
+        """Always returns True."""
+        return True
+
+    __bool__ = __nonzero__
+
+    def __len__():
+        """Returns the number of changesets in the repo."""
+
+    def __iter__():
+        """Iterate over revisions in the changelog."""
+
+    def revs(expr, *args):
+        """Evaluate a revset.
+
+        Emits revisions.
+        """
+
+    def set(expr, *args):
+        """Evaluate a revset.
+
+        Emits changectx instances.
+        """
+
+    def anyrevs(specs, user=False, localalias=None):
+        """Find revisions matching one of the given revsets."""
+
+    def url():
+        """Returns a string representing the location of this repo."""
+
+    def hook(name, throw=False, **args):
+        """Call a hook."""
+
+    def tags():
+        """Return a mapping of tag to node."""
+
+    def tagtype(tagname):
+        """Return the type of a given tag."""
+
+    def tagslist():
+        """Return a list of tags ordered by revision."""
+
+    def nodetags(node):
+        """Return the tags associated with a node."""
+
+    def nodebookmarks(node):
+        """Return the list of bookmarks pointing to the specified node."""
+
+    def branchmap():
+        """Return a mapping of branch to heads in that branch."""
+
+    def revbranchcache():
+        pass
+
+    def branchtip(branchtip, ignoremissing=False):
+        """Return the tip node for a given branch."""
+
+    def lookup(key):
+        """Resolve the node for a revision."""
+
+    def lookupbranch(key):
+        """Look up the branch name of the given revision or branch name."""
+
+    def known(nodes):
+        """Determine whether a series of nodes is known.
+
+        Returns a list of bools.
+        """
+
+    def local():
+        """Whether the repository is local."""
+        return True
+
+    def publishing():
+        """Whether the repository is a publishing repository."""
+
+    def cancopy():
+        pass
+
+    def shared():
+        """The type of shared repository or None."""
+
+    def wjoin(f, *insidef):
+        """Calls self.vfs.reljoin(self.root, f, *insidef)"""
+
+    def setparents(p1, p2):
+        """Set the parent nodes of the working directory."""
+
+    def filectx(path, changeid=None, fileid=None):
+        """Obtain a filectx for the given file revision."""
+
+    def getcwd():
+        """Obtain the current working directory from the dirstate."""
+
+    def pathto(f, cwd=None):
+        """Obtain the relative path to a file."""
+
+    def adddatafilter(name, fltr):
+        pass
+
+    def wread(filename):
+        """Read a file from wvfs, using data filters."""
+
+    def wwrite(filename, data, flags, backgroundclose=False, **kwargs):
+        """Write data to a file in the wvfs, using data filters."""
+
+    def wwritedata(filename, data):
+        """Resolve data for writing to the wvfs, using data filters."""
+
+    def currenttransaction():
+        """Obtain the current transaction instance or None."""
+
+    def transaction(desc, report=None):
+        """Open a new transaction to write to the repository."""
+
+    def undofiles():
+        """Returns a list of (vfs, path) for files to undo transactions."""
+
+    def recover():
+        """Roll back an interrupted transaction."""
+
+    def rollback(dryrun=False, force=False):
+        """Undo the last transaction.
+
+        DANGEROUS.
+        """
+
+    def updatecaches(tr=None, full=False):
+        """Warm repo caches."""
+
+    def invalidatecaches():
+        """Invalidate cached data due to the repository mutating."""
+
+    def invalidatevolatilesets():
+        pass
+
+    def invalidatedirstate():
+        """Invalidate the dirstate."""
+
+    def invalidate(clearfilecache=False):
+        pass
+
+    def invalidateall():
+        pass
+
+    def lock(wait=True):
+        """Lock the repository store and return a lock instance."""
+
+    def wlock(wait=True):
+        """Lock the non-store parts of the repository."""
+
+    def currentwlock():
+        """Return the wlock if it's held or None."""
+
+    def checkcommitpatterns(wctx, vdirs, match, status, fail):
+        pass
+
+    def commit(text='', user=None, date=None, match=None, force=False,
+               editor=False, extra=None):
+        """Add a new revision to the repository."""
+
+    def commitctx(ctx, error=False, origctx=None):
+        """Commit a commitctx instance to the repository."""
+
+    def destroying():
+        """Inform the repository that nodes are about to be destroyed."""
+
+    def destroyed():
+        """Inform the repository that nodes have been destroyed."""
+
+    def status(node1='.', node2=None, match=None, ignored=False,
+               clean=False, unknown=False, listsubrepos=False):
+        """Convenience method to call repo[x].status()."""
+
+    def addpostdsstatus(ps):
+        pass
+
+    def postdsstatus():
+        pass
+
+    def clearpostdsstatus():
+        pass
+
+    def heads(start=None):
+        """Obtain list of nodes that are DAG heads."""
+
+    def branchheads(branch=None, start=None, closed=False):
+        pass
+
+    def branches(nodes):
+        pass
+
+    def between(pairs):
+        pass
+
+    def checkpush(pushop):
+        pass
+
+    prepushoutgoinghooks = interfaceutil.Attribute(
+        """util.hooks instance.""")
+
+    def pushkey(namespace, key, old, new):
+        pass
+
+    def listkeys(namespace):
+        pass
+
+    def debugwireargs(one, two, three=None, four=None, five=None):
+        pass
+
+    def savecommitmessage(text):
+        pass
+
+class completelocalrepository(ilocalrepositorymain,
+                              ilocalrepositoryfilestorage):
+    """Complete interface for a local repository."""
+
+class iwireprotocolcommandcacher(interfaceutil.Interface):
+    """Represents a caching backend for wire protocol commands.
+
+    Wire protocol version 2 supports transparent caching of many commands.
+    To leverage this caching, servers can activate objects that cache
+    command responses. Objects handle both cache writing and reading.
+    This interface defines how that response caching mechanism works.
+
+    Wire protocol version 2 commands emit a series of objects that are
+    serialized and sent to the client. The caching layer exists between
+    the invocation of the command function and the sending of its output
+    objects to an output layer.
+
+    Instances of this interface represent a binding to a cache that
+    can serve a response (in place of calling a command function) and/or
+    write responses to a cache for subsequent use.
+
+    When a command request arrives, the following happens with regards
+    to this interface:
+
+    1. The server determines whether the command request is cacheable.
+    2. If it is, an instance of this interface is spawned.
+    3. The cacher is activated in a context manager (``__enter__`` is called).
+    4. A cache *key* for that request is derived. This will call the
+       instance's ``adjustcachekeystate()`` method so the derivation
+       can be influenced.
+    5. The cacher is informed of the derived cache key via a call to
+       ``setcachekey()``.
+    6. The cacher's ``lookup()`` method is called to test for presence of
+       the derived key in the cache.
+    7. If ``lookup()`` returns a hit, that cached result is used in place
+       of invoking the command function. ``__exit__`` is called and the instance
+       is discarded.
+    8. The command function is invoked.
+    9. ``onobject()`` is called for each object emitted by the command
+       function.
+    10. After the final object is seen, ``onfinished()`` is called.
+    11. ``__exit__`` is called to signal the end of use of the instance.
+
+    Cache *key* derivation can be influenced by the instance.
+
+    Cache keys are initially derived by a deterministic representation of
+    the command request. This includes the command name, arguments, protocol
+    version, etc. This initial key derivation is performed by CBOR-encoding a
+    data structure and feeding that output into a hasher.
+
+    Instances of this interface can influence this initial key derivation
+    via ``adjustcachekeystate()``.
+
+    The instance is informed of the derived cache key via a call to
+    ``setcachekey()``. The instance must store the key locally so it can
+    be consulted on subsequent operations that may require it.
+
+    When constructed, the instance has access to a callable that can be used
+    for encoding response objects. This callable receives as its single
+    argument an object emitted by a command function. It returns an iterable
+    of bytes chunks representing the encoded object. Unless the cacher is
+    caching native Python objects in memory or has a way of reconstructing
+    the original Python objects, implementations typically call this function
+    to produce bytes from the output objects and then store those bytes in
+    the cache. When it comes time to re-emit those bytes, they are wrapped
+    in a ``wireprototypes.encodedresponse`` instance to tell the output
+    layer that they are pre-encoded.
+
+    When receiving the objects emitted by the command function, instances
+    can choose what to do with those objects. The simplest thing to do is
+    re-emit the original objects. They will be forwarded to the output
+    layer and will be processed as if the cacher did not exist.
+
+    Implementations could also choose to not emit objects - instead locally
+    buffering objects or their encoded representation. They could then emit
+    a single "coalesced" object when ``onfinished()`` is called. In
+    this way, the implementation would function as a filtering layer of
+    sorts.
+
+    When caching objects, typically the encoded form of the object will
+    be stored. Keep in mind that if the original object is forwarded to
+    the output layer, it will need to be encoded there as well. For large
+    output, this redundant encoding could add overhead. Implementations
+    could wrap the encoded object data in ``wireprototypes.encodedresponse``
+    instances to avoid this overhead.
+    """
+    def __enter__():
+        """Marks the instance as active.
+
+        Should return self.
+        """
+
+    def __exit__(exctype, excvalue, exctb):
+        """Called when cacher is no longer used.
+
+        This can be used by implementations to perform cleanup actions (e.g.
+        disconnecting network sockets, aborting a partially cached response.
+        """
+
+    def adjustcachekeystate(state):
+        """Influences cache key derivation by adjusting state to derive key.
+
+        A dict defining the state used to derive the cache key is passed.
+
+        Implementations can modify this dict to record additional state that
+        is wanted to influence key derivation.
+
+        Implementations are *highly* encouraged to not modify or delete
+        existing keys.
+        """
+
+    def setcachekey(key):
+        """Record the derived cache key for this request.
+
+        Instances may mutate the key for internal usage, as desired. e.g.
+        instances may wish to prepend the repo name, introduce path
+        components for filesystem or URL addressing, etc. Behavior is up to
+        the cache.
+
+        Returns a bool indicating if the request is cacheable by this
+        instance.
+        """
+
+    def lookup():
+        """Attempt to resolve an entry in the cache.
+
+        The instance is instructed to look for the cache key that it was
+        informed about via the call to ``setcachekey()``.
+
+        If there's no cache hit or the cacher doesn't wish to use the cached
+        entry, ``None`` should be returned.
+
+        Else, a dict defining the cached result should be returned. The
+        dict may have the following keys:
+
+        objs
+           An iterable of objects that should be sent to the client. That
+           iterable of objects is expected to be what the command function
+           would return if invoked or an equivalent representation thereof.
+        """
+
+    def onobject(obj):
+        """Called when a new object is emitted from the command function.
+
+        Receives as its argument the object that was emitted from the
+        command function.
+
+        This method returns an iterator of objects to forward to the output
+        layer. The easiest implementation is a generator that just
+        ``yield obj``.
+        """
+
+    def onfinished():
+        """Called after all objects have been emitted from the command function.
+
+        Implementations should return an iterator of objects to forward to
+        the output layer.
+
+        This method can be a generator.
+        """
--- a/mercurial/localrepo.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/localrepo.py	Sun Aug 18 00:45:33 2019 +0300
@@ -52,7 +52,6 @@
     phases,
     pushkey,
     pycompat,
-    repository,
     repoview,
     revset,
     revsetlang,
@@ -66,6 +65,11 @@
     util,
     vfs as vfsmod,
 )
+
+from .interfaces import (
+    repository,
+)
+
 from .utils import (
     interfaceutil,
     procutil,
--- a/mercurial/manifest.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/manifest.py	Sun Aug 18 00:45:33 2019 +0300
@@ -24,10 +24,12 @@
     mdiff,
     policy,
     pycompat,
-    repository,
     revlog,
     util,
 )
+from .interfaces import (
+    repository,
+)
 from .utils import (
     interfaceutil,
 )
--- a/mercurial/narrowspec.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/narrowspec.py	Sun Aug 18 00:45:33 2019 +0300
@@ -8,11 +8,13 @@
 from __future__ import absolute_import
 
 from .i18n import _
+from .interfaces import (
+    repository,
+)
 from . import (
     error,
     match as matchmod,
     merge,
-    repository,
     scmutil,
     sparse,
     util,
--- a/mercurial/repository.py	Thu Aug 22 16:47:31 2019 -0700
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1877 +0,0 @@
-# repository.py - Interfaces and base classes for repositories and peers.
-#
-# Copyright 2017 Gregory Szorc <gregory.szorc@gmail.com>
-#
-# This software may be used and distributed according to the terms of the
-# GNU General Public License version 2 or any later version.
-
-from __future__ import absolute_import
-
-from .i18n import _
-from . import (
-    error,
-)
-from .utils import (
-    interfaceutil,
-)
-
-# When narrowing is finalized and no longer subject to format changes,
-# we should move this to just "narrow" or similar.
-NARROW_REQUIREMENT = 'narrowhg-experimental'
-
-# Local repository feature string.
-
-# Revlogs are being used for file storage.
-REPO_FEATURE_REVLOG_FILE_STORAGE = b'revlogfilestorage'
-# The storage part of the repository is shared from an external source.
-REPO_FEATURE_SHARED_STORAGE = b'sharedstore'
-# LFS supported for backing file storage.
-REPO_FEATURE_LFS = b'lfs'
-# Repository supports being stream cloned.
-REPO_FEATURE_STREAM_CLONE = b'streamclone'
-# Files storage may lack data for all ancestors.
-REPO_FEATURE_SHALLOW_FILE_STORAGE = b'shallowfilestorage'
-
-REVISION_FLAG_CENSORED = 1 << 15
-REVISION_FLAG_ELLIPSIS = 1 << 14
-REVISION_FLAG_EXTSTORED = 1 << 13
-
-REVISION_FLAGS_KNOWN = (
-    REVISION_FLAG_CENSORED | REVISION_FLAG_ELLIPSIS | REVISION_FLAG_EXTSTORED)
-
-CG_DELTAMODE_STD = b'default'
-CG_DELTAMODE_PREV = b'previous'
-CG_DELTAMODE_FULL = b'fulltext'
-CG_DELTAMODE_P1 = b'p1'
-
-class ipeerconnection(interfaceutil.Interface):
-    """Represents a "connection" to a repository.
-
-    This is the base interface for representing a connection to a repository.
-    It holds basic properties and methods applicable to all peer types.
-
-    This is not a complete interface definition and should not be used
-    outside of this module.
-    """
-    ui = interfaceutil.Attribute("""ui.ui instance""")
-
-    def url():
-        """Returns a URL string representing this peer.
-
-        Currently, implementations expose the raw URL used to construct the
-        instance. It may contain credentials as part of the URL. The
-        expectations of the value aren't well-defined and this could lead to
-        data leakage.
-
-        TODO audit/clean consumers and more clearly define the contents of this
-        value.
-        """
-
-    def local():
-        """Returns a local repository instance.
-
-        If the peer represents a local repository, returns an object that
-        can be used to interface with it. Otherwise returns ``None``.
-        """
-
-    def peer():
-        """Returns an object conforming to this interface.
-
-        Most implementations will ``return self``.
-        """
-
-    def canpush():
-        """Returns a boolean indicating if this peer can be pushed to."""
-
-    def close():
-        """Close the connection to this peer.
-
-        This is called when the peer will no longer be used. Resources
-        associated with the peer should be cleaned up.
-        """
-
-class ipeercapabilities(interfaceutil.Interface):
-    """Peer sub-interface related to capabilities."""
-
-    def capable(name):
-        """Determine support for a named capability.
-
-        Returns ``False`` if capability not supported.
-
-        Returns ``True`` if boolean capability is supported. Returns a string
-        if capability support is non-boolean.
-
-        Capability strings may or may not map to wire protocol capabilities.
-        """
-
-    def requirecap(name, purpose):
-        """Require a capability to be present.
-
-        Raises a ``CapabilityError`` if the capability isn't present.
-        """
-
-class ipeercommands(interfaceutil.Interface):
-    """Client-side interface for communicating over the wire protocol.
-
-    This interface is used as a gateway to the Mercurial wire protocol.
-    methods commonly call wire protocol commands of the same name.
-    """
-
-    def branchmap():
-        """Obtain heads in named branches.
-
-        Returns a dict mapping branch name to an iterable of nodes that are
-        heads on that branch.
-        """
-
-    def capabilities():
-        """Obtain capabilities of the peer.
-
-        Returns a set of string capabilities.
-        """
-
-    def clonebundles():
-        """Obtains the clone bundles manifest for the repo.
-
-        Returns the manifest as unparsed bytes.
-        """
-
-    def debugwireargs(one, two, three=None, four=None, five=None):
-        """Used to facilitate debugging of arguments passed over the wire."""
-
-    def getbundle(source, **kwargs):
-        """Obtain remote repository data as a bundle.
-
-        This command is how the bulk of repository data is transferred from
-        the peer to the local repository
-
-        Returns a generator of bundle data.
-        """
-
-    def heads():
-        """Determine all known head revisions in the peer.
-
-        Returns an iterable of binary nodes.
-        """
-
-    def known(nodes):
-        """Determine whether multiple nodes are known.
-
-        Accepts an iterable of nodes whose presence to check for.
-
-        Returns an iterable of booleans indicating of the corresponding node
-        at that index is known to the peer.
-        """
-
-    def listkeys(namespace):
-        """Obtain all keys in a pushkey namespace.
-
-        Returns an iterable of key names.
-        """
-
-    def lookup(key):
-        """Resolve a value to a known revision.
-
-        Returns a binary node of the resolved revision on success.
-        """
-
-    def pushkey(namespace, key, old, new):
-        """Set a value using the ``pushkey`` protocol.
-
-        Arguments correspond to the pushkey namespace and key to operate on and
-        the old and new values for that key.
-
-        Returns a string with the peer result. The value inside varies by the
-        namespace.
-        """
-
-    def stream_out():
-        """Obtain streaming clone data.
-
-        Successful result should be a generator of data chunks.
-        """
-
-    def unbundle(bundle, heads, url):
-        """Transfer repository data to the peer.
-
-        This is how the bulk of data during a push is transferred.
-
-        Returns the integer number of heads added to the peer.
-        """
-
-class ipeerlegacycommands(interfaceutil.Interface):
-    """Interface for implementing support for legacy wire protocol commands.
-
-    Wire protocol commands transition to legacy status when they are no longer
-    used by modern clients. To facilitate identifying which commands are
-    legacy, the interfaces are split.
-    """
-
-    def between(pairs):
-        """Obtain nodes between pairs of nodes.
-
-        ``pairs`` is an iterable of node pairs.
-
-        Returns an iterable of iterables of nodes corresponding to each
-        requested pair.
-        """
-
-    def branches(nodes):
-        """Obtain ancestor changesets of specific nodes back to a branch point.
-
-        For each requested node, the peer finds the first ancestor node that is
-        a DAG root or is a merge.
-
-        Returns an iterable of iterables with the resolved values for each node.
-        """
-
-    def changegroup(nodes, source):
-        """Obtain a changegroup with data for descendants of specified nodes."""
-
-    def changegroupsubset(bases, heads, source):
-        pass
-
-class ipeercommandexecutor(interfaceutil.Interface):
-    """Represents a mechanism to execute remote commands.
-
-    This is the primary interface for requesting that wire protocol commands
-    be executed. Instances of this interface are active in a context manager
-    and have a well-defined lifetime. When the context manager exits, all
-    outstanding requests are waited on.
-    """
-
-    def callcommand(name, args):
-        """Request that a named command be executed.
-
-        Receives the command name and a dictionary of command arguments.
-
-        Returns a ``concurrent.futures.Future`` that will resolve to the
-        result of that command request. That exact value is left up to
-        the implementation and possibly varies by command.
-
-        Not all commands can coexist with other commands in an executor
-        instance: it depends on the underlying wire protocol transport being
-        used and the command itself.
-
-        Implementations MAY call ``sendcommands()`` automatically if the
-        requested command can not coexist with other commands in this executor.
-
-        Implementations MAY call ``sendcommands()`` automatically when the
-        future's ``result()`` is called. So, consumers using multiple
-        commands with an executor MUST ensure that ``result()`` is not called
-        until all command requests have been issued.
-        """
-
-    def sendcommands():
-        """Trigger submission of queued command requests.
-
-        Not all transports submit commands as soon as they are requested to
-        run. When called, this method forces queued command requests to be
-        issued. It will no-op if all commands have already been sent.
-
-        When called, no more new commands may be issued with this executor.
-        """
-
-    def close():
-        """Signal that this command request is finished.
-
-        When called, no more new commands may be issued. All outstanding
-        commands that have previously been issued are waited on before
-        returning. This not only includes waiting for the futures to resolve,
-        but also waiting for all response data to arrive. In other words,
-        calling this waits for all on-wire state for issued command requests
-        to finish.
-
-        When used as a context manager, this method is called when exiting the
-        context manager.
-
-        This method may call ``sendcommands()`` if there are buffered commands.
-        """
-
-class ipeerrequests(interfaceutil.Interface):
-    """Interface for executing commands on a peer."""
-
-    limitedarguments = interfaceutil.Attribute(
-        """True if the peer cannot receive large argument value for commands."""
-    )
-
-    def commandexecutor():
-        """A context manager that resolves to an ipeercommandexecutor.
-
-        The object this resolves to can be used to issue command requests
-        to the peer.
-
-        Callers should call its ``callcommand`` method to issue command
-        requests.
-
-        A new executor should be obtained for each distinct set of commands
-        (possibly just a single command) that the consumer wants to execute
-        as part of a single operation or round trip. This is because some
-        peers are half-duplex and/or don't support persistent connections.
-        e.g. in the case of HTTP peers, commands sent to an executor represent
-        a single HTTP request. While some peers may support multiple command
-        sends over the wire per executor, consumers need to code to the least
-        capable peer. So it should be assumed that command executors buffer
-        called commands until they are told to send them and that each
-        command executor could result in a new connection or wire-level request
-        being issued.
-        """
-
-class ipeerbase(ipeerconnection, ipeercapabilities, ipeerrequests):
-    """Unified interface for peer repositories.
-
-    All peer instances must conform to this interface.
-    """
-
-class ipeerv2(ipeerconnection, ipeercapabilities, ipeerrequests):
-    """Unified peer interface for wire protocol version 2 peers."""
-
-    apidescriptor = interfaceutil.Attribute(
-        """Data structure holding description of server API.""")
-
-@interfaceutil.implementer(ipeerbase)
-class peer(object):
-    """Base class for peer repositories."""
-
-    limitedarguments = False
-
-    def capable(self, name):
-        caps = self.capabilities()
-        if name in caps:
-            return True
-
-        name = '%s=' % name
-        for cap in caps:
-            if cap.startswith(name):
-                return cap[len(name):]
-
-        return False
-
-    def requirecap(self, name, purpose):
-        if self.capable(name):
-            return
-
-        raise error.CapabilityError(
-            _('cannot %s; remote repository does not support the '
-              '\'%s\' capability') % (purpose, name))
-
-class iverifyproblem(interfaceutil.Interface):
-    """Represents a problem with the integrity of the repository.
-
-    Instances of this interface are emitted to describe an integrity issue
-    with a repository (e.g. corrupt storage, missing data, etc).
-
-    Instances are essentially messages associated with severity.
-    """
-    warning = interfaceutil.Attribute(
-        """Message indicating a non-fatal problem.""")
-
-    error = interfaceutil.Attribute(
-        """Message indicating a fatal problem.""")
-
-    node = interfaceutil.Attribute(
-        """Revision encountering the problem.
-
-        ``None`` means the problem doesn't apply to a single revision.
-        """)
-
-class irevisiondelta(interfaceutil.Interface):
-    """Represents a delta between one revision and another.
-
-    Instances convey enough information to allow a revision to be exchanged
-    with another repository.
-
-    Instances represent the fulltext revision data or a delta against
-    another revision. Therefore the ``revision`` and ``delta`` attributes
-    are mutually exclusive.
-
-    Typically used for changegroup generation.
-    """
-
-    node = interfaceutil.Attribute(
-        """20 byte node of this revision.""")
-
-    p1node = interfaceutil.Attribute(
-        """20 byte node of 1st parent of this revision.""")
-
-    p2node = interfaceutil.Attribute(
-        """20 byte node of 2nd parent of this revision.""")
-
-    linknode = interfaceutil.Attribute(
-        """20 byte node of the changelog revision this node is linked to.""")
-
-    flags = interfaceutil.Attribute(
-        """2 bytes of integer flags that apply to this revision.
-
-        This is a bitwise composition of the ``REVISION_FLAG_*`` constants.
-        """)
-
-    basenode = interfaceutil.Attribute(
-        """20 byte node of the revision this data is a delta against.
-
-        ``nullid`` indicates that the revision is a full revision and not
-        a delta.
-        """)
-
-    baserevisionsize = interfaceutil.Attribute(
-        """Size of base revision this delta is against.
-
-        May be ``None`` if ``basenode`` is ``nullid``.
-        """)
-
-    revision = interfaceutil.Attribute(
-        """Raw fulltext of revision data for this node.""")
-
-    delta = interfaceutil.Attribute(
-        """Delta between ``basenode`` and ``node``.
-
-        Stored in the bdiff delta format.
-        """)
-
-class ifilerevisionssequence(interfaceutil.Interface):
-    """Contains index data for all revisions of a file.
-
-    Types implementing this behave like lists of tuples. The index
-    in the list corresponds to the revision number. The values contain
-    index metadata.
-
-    The *null* revision (revision number -1) is always the last item
-    in the index.
-    """
-
-    def __len__():
-        """The total number of revisions."""
-
-    def __getitem__(rev):
-        """Returns the object having a specific revision number.
-
-        Returns an 8-tuple with the following fields:
-
-        offset+flags
-           Contains the offset and flags for the revision. 64-bit unsigned
-           integer where first 6 bytes are the offset and the next 2 bytes
-           are flags. The offset can be 0 if it is not used by the store.
-        compressed size
-            Size of the revision data in the store. It can be 0 if it isn't
-            needed by the store.
-        uncompressed size
-            Fulltext size. It can be 0 if it isn't needed by the store.
-        base revision
-            Revision number of revision the delta for storage is encoded
-            against. -1 indicates not encoded against a base revision.
-        link revision
-            Revision number of changelog revision this entry is related to.
-        p1 revision
-            Revision number of 1st parent. -1 if no 1st parent.
-        p2 revision
-            Revision number of 2nd parent. -1 if no 1st parent.
-        node
-            Binary node value for this revision number.
-
-        Negative values should index off the end of the sequence. ``-1``
-        should return the null revision. ``-2`` should return the most
-        recent revision.
-        """
-
-    def __contains__(rev):
-        """Whether a revision number exists."""
-
-    def insert(self, i, entry):
-        """Add an item to the index at specific revision."""
-
-class ifileindex(interfaceutil.Interface):
-    """Storage interface for index data of a single file.
-
-    File storage data is divided into index metadata and data storage.
-    This interface defines the index portion of the interface.
-
-    The index logically consists of:
-
-    * A mapping between revision numbers and nodes.
-    * DAG data (storing and querying the relationship between nodes).
-    * Metadata to facilitate storage.
-    """
-    def __len__():
-        """Obtain the number of revisions stored for this file."""
-
-    def __iter__():
-        """Iterate over revision numbers for this file."""
-
-    def hasnode(node):
-        """Returns a bool indicating if a node is known to this store.
-
-        Implementations must only return True for full, binary node values:
-        hex nodes, revision numbers, and partial node matches must be
-        rejected.
-
-        The null node is never present.
-        """
-
-    def revs(start=0, stop=None):
-        """Iterate over revision numbers for this file, with control."""
-
-    def parents(node):
-        """Returns a 2-tuple of parent nodes for a revision.
-
-        Values will be ``nullid`` if the parent is empty.
-        """
-
-    def parentrevs(rev):
-        """Like parents() but operates on revision numbers."""
-
-    def rev(node):
-        """Obtain the revision number given a node.
-
-        Raises ``error.LookupError`` if the node is not known.
-        """
-
-    def node(rev):
-        """Obtain the node value given a revision number.
-
-        Raises ``IndexError`` if the node is not known.
-        """
-
-    def lookup(node):
-        """Attempt to resolve a value to a node.
-
-        Value can be a binary node, hex node, revision number, or a string
-        that can be converted to an integer.
-
-        Raises ``error.LookupError`` if a node could not be resolved.
-        """
-
-    def linkrev(rev):
-        """Obtain the changeset revision number a revision is linked to."""
-
-    def iscensored(rev):
-        """Return whether a revision's content has been censored."""
-
-    def commonancestorsheads(node1, node2):
-        """Obtain an iterable of nodes containing heads of common ancestors.
-
-        See ``ancestor.commonancestorsheads()``.
-        """
-
-    def descendants(revs):
-        """Obtain descendant revision numbers for a set of revision numbers.
-
-        If ``nullrev`` is in the set, this is equivalent to ``revs()``.
-        """
-
-    def heads(start=None, stop=None):
-        """Obtain a list of nodes that are DAG heads, with control.
-
-        The set of revisions examined can be limited by specifying
-        ``start`` and ``stop``. ``start`` is a node. ``stop`` is an
-        iterable of nodes. DAG traversal starts at earlier revision
-        ``start`` and iterates forward until any node in ``stop`` is
-        encountered.
-        """
-
-    def children(node):
-        """Obtain nodes that are children of a node.
-
-        Returns a list of nodes.
-        """
-
-class ifiledata(interfaceutil.Interface):
-    """Storage interface for data storage of a specific file.
-
-    This complements ``ifileindex`` and provides an interface for accessing
-    data for a tracked file.
-    """
-    def size(rev):
-        """Obtain the fulltext size of file data.
-
-        Any metadata is excluded from size measurements.
-        """
-
-    def revision(node, raw=False):
-        """"Obtain fulltext data for a node.
-
-        By default, any storage transformations are applied before the data
-        is returned. If ``raw`` is True, non-raw storage transformations
-        are not applied.
-
-        The fulltext data may contain a header containing metadata. Most
-        consumers should use ``read()`` to obtain the actual file data.
-        """
-
-    def rawdata(node):
-        """Obtain raw data for a node.
-        """
-
-    def read(node):
-        """Resolve file fulltext data.
-
-        This is similar to ``revision()`` except any metadata in the data
-        headers is stripped.
-        """
-
-    def renamed(node):
-        """Obtain copy metadata for a node.
-
-        Returns ``False`` if no copy metadata is stored or a 2-tuple of
-        (path, node) from which this revision was copied.
-        """
-
-    def cmp(node, fulltext):
-        """Compare fulltext to another revision.
-
-        Returns True if the fulltext is different from what is stored.
-
-        This takes copy metadata into account.
-
-        TODO better document the copy metadata and censoring logic.
-        """
-
-    def emitrevisions(nodes,
-                      nodesorder=None,
-                      revisiondata=False,
-                      assumehaveparentrevisions=False,
-                      deltamode=CG_DELTAMODE_STD):
-        """Produce ``irevisiondelta`` for revisions.
-
-        Given an iterable of nodes, emits objects conforming to the
-        ``irevisiondelta`` interface that describe revisions in storage.
-
-        This method is a generator.
-
-        The input nodes may be unordered. Implementations must ensure that a
-        node's parents are emitted before the node itself. Transitively, this
-        means that a node may only be emitted once all its ancestors in
-        ``nodes`` have also been emitted.
-
-        By default, emits "index" data (the ``node``, ``p1node``, and
-        ``p2node`` attributes). If ``revisiondata`` is set, revision data
-        will also be present on the emitted objects.
-
-        With default argument values, implementations can choose to emit
-        either fulltext revision data or a delta. When emitting deltas,
-        implementations must consider whether the delta's base revision
-        fulltext is available to the receiver.
-
-        The base revision fulltext is guaranteed to be available if any of
-        the following are met:
-
-        * Its fulltext revision was emitted by this method call.
-        * A delta for that revision was emitted by this method call.
-        * ``assumehaveparentrevisions`` is True and the base revision is a
-          parent of the node.
-
-        ``nodesorder`` can be used to control the order that revisions are
-        emitted. By default, revisions can be reordered as long as they are
-        in DAG topological order (see above). If the value is ``nodes``,
-        the iteration order from ``nodes`` should be used. If the value is
-        ``storage``, then the native order from the backing storage layer
-        is used. (Not all storage layers will have strong ordering and behavior
-        of this mode is storage-dependent.) ``nodes`` ordering can force
-        revisions to be emitted before their ancestors, so consumers should
-        use it with care.
-
-        The ``linknode`` attribute on the returned ``irevisiondelta`` may not
-        be set and it is the caller's responsibility to resolve it, if needed.
-
-        If ``deltamode`` is CG_DELTAMODE_PREV and revision data is requested,
-        all revision data should be emitted as deltas against the revision
-        emitted just prior. The initial revision should be a delta against its
-        1st parent.
-        """
-
-class ifilemutation(interfaceutil.Interface):
-    """Storage interface for mutation events of a tracked file."""
-
-    def add(filedata, meta, transaction, linkrev, p1, p2):
-        """Add a new revision to the store.
-
-        Takes file data, dictionary of metadata, a transaction, linkrev,
-        and parent nodes.
-
-        Returns the node that was added.
-
-        May no-op if a revision matching the supplied data is already stored.
-        """
-
-    def addrevision(revisiondata, transaction, linkrev, p1, p2, node=None,
-                    flags=0, cachedelta=None):
-        """Add a new revision to the store.
-
-        This is similar to ``add()`` except it operates at a lower level.
-
-        The data passed in already contains a metadata header, if any.
-
-        ``node`` and ``flags`` can be used to define the expected node and
-        the flags to use with storage. ``flags`` is a bitwise value composed
-        of the various ``REVISION_FLAG_*`` constants.
-
-        ``add()`` is usually called when adding files from e.g. the working
-        directory. ``addrevision()`` is often called by ``add()`` and for
-        scenarios where revision data has already been computed, such as when
-        applying raw data from a peer repo.
-        """
-
-    def addgroup(deltas, linkmapper, transaction, addrevisioncb=None,
-                 maybemissingparents=False):
-        """Process a series of deltas for storage.
-
-        ``deltas`` is an iterable of 7-tuples of
-        (node, p1, p2, linknode, deltabase, delta, flags) defining revisions
-        to add.
-
-        The ``delta`` field contains ``mpatch`` data to apply to a base
-        revision, identified by ``deltabase``. The base node can be
-        ``nullid``, in which case the header from the delta can be ignored
-        and the delta used as the fulltext.
-
-        ``addrevisioncb`` should be called for each node as it is committed.
-
-        ``maybemissingparents`` is a bool indicating whether the incoming
-        data may reference parents/ancestor revisions that aren't present.
-        This flag is set when receiving data into a "shallow" store that
-        doesn't hold all history.
-
-        Returns a list of nodes that were processed. A node will be in the list
-        even if it existed in the store previously.
-        """
-
-    def censorrevision(tr, node, tombstone=b''):
-        """Remove the content of a single revision.
-
-        The specified ``node`` will have its content purged from storage.
-        Future attempts to access the revision data for this node will
-        result in failure.
-
-        A ``tombstone`` message can optionally be stored. This message may be
-        displayed to users when they attempt to access the missing revision
-        data.
-
-        Storage backends may have stored deltas against the previous content
-        in this revision. As part of censoring a revision, these storage
-        backends are expected to rewrite any internally stored deltas such
-        that they no longer reference the deleted content.
-        """
-
-    def getstrippoint(minlink):
-        """Find the minimum revision that must be stripped to strip a linkrev.
-
-        Returns a 2-tuple containing the minimum revision number and a set
-        of all revisions numbers that would be broken by this strip.
-
-        TODO this is highly revlog centric and should be abstracted into
-        a higher-level deletion API. ``repair.strip()`` relies on this.
-        """
-
-    def strip(minlink, transaction):
-        """Remove storage of items starting at a linkrev.
-
-        This uses ``getstrippoint()`` to determine the first node to remove.
-        Then it effectively truncates storage for all revisions after that.
-
-        TODO this is highly revlog centric and should be abstracted into a
-        higher-level deletion API.
-        """
-
-class ifilestorage(ifileindex, ifiledata, ifilemutation):
-    """Complete storage interface for a single tracked file."""
-
-    def files():
-        """Obtain paths that are backing storage for this file.
-
-        TODO this is used heavily by verify code and there should probably
-        be a better API for that.
-        """
-
-    def storageinfo(exclusivefiles=False, sharedfiles=False,
-                    revisionscount=False, trackedsize=False,
-                    storedsize=False):
-        """Obtain information about storage for this file's data.
-
-        Returns a dict describing storage for this tracked path. The keys
-        in the dict map to arguments of the same. The arguments are bools
-        indicating whether to calculate and obtain that data.
-
-        exclusivefiles
-           Iterable of (vfs, path) describing files that are exclusively
-           used to back storage for this tracked path.
-
-        sharedfiles
-           Iterable of (vfs, path) describing files that are used to back
-           storage for this tracked path. Those files may also provide storage
-           for other stored entities.
-
-        revisionscount
-           Number of revisions available for retrieval.
-
-        trackedsize
-           Total size in bytes of all tracked revisions. This is a sum of the
-           length of the fulltext of all revisions.
-
-        storedsize
-           Total size in bytes used to store data for all tracked revisions.
-           This is commonly less than ``trackedsize`` due to internal usage
-           of deltas rather than fulltext revisions.
-
-        Not all storage backends may support all queries are have a reasonable
-        value to use. In that case, the value should be set to ``None`` and
-        callers are expected to handle this special value.
-        """
-
-    def verifyintegrity(state):
-        """Verifies the integrity of file storage.
-
-        ``state`` is a dict holding state of the verifier process. It can be
-        used to communicate data between invocations of multiple storage
-        primitives.
-
-        If individual revisions cannot have their revision content resolved,
-        the method is expected to set the ``skipread`` key to a set of nodes
-        that encountered problems.
-
-        The method yields objects conforming to the ``iverifyproblem``
-        interface.
-        """
-
-class idirs(interfaceutil.Interface):
-    """Interface representing a collection of directories from paths.
-
-    This interface is essentially a derived data structure representing
-    directories from a collection of paths.
-    """
-
-    def addpath(path):
-        """Add a path to the collection.
-
-        All directories in the path will be added to the collection.
-        """
-
-    def delpath(path):
-        """Remove a path from the collection.
-
-        If the removal was the last path in a particular directory, the
-        directory is removed from the collection.
-        """
-
-    def __iter__():
-        """Iterate over the directories in this collection of paths."""
-
-    def __contains__(path):
-        """Whether a specific directory is in this collection."""
-
-class imanifestdict(interfaceutil.Interface):
-    """Interface representing a manifest data structure.
-
-    A manifest is effectively a dict mapping paths to entries. Each entry
-    consists of a binary node and extra flags affecting that entry.
-    """
-
-    def __getitem__(path):
-        """Returns the binary node value for a path in the manifest.
-
-        Raises ``KeyError`` if the path does not exist in the manifest.
-
-        Equivalent to ``self.find(path)[0]``.
-        """
-
-    def find(path):
-        """Returns the entry for a path in the manifest.
-
-        Returns a 2-tuple of (node, flags).
-
-        Raises ``KeyError`` if the path does not exist in the manifest.
-        """
-
-    def __len__():
-        """Return the number of entries in the manifest."""
-
-    def __nonzero__():
-        """Returns True if the manifest has entries, False otherwise."""
-
-    __bool__ = __nonzero__
-
-    def __setitem__(path, node):
-        """Define the node value for a path in the manifest.
-
-        If the path is already in the manifest, its flags will be copied to
-        the new entry.
-        """
-
-    def __contains__(path):
-        """Whether a path exists in the manifest."""
-
-    def __delitem__(path):
-        """Remove a path from the manifest.
-
-        Raises ``KeyError`` if the path is not in the manifest.
-        """
-
-    def __iter__():
-        """Iterate over paths in the manifest."""
-
-    def iterkeys():
-        """Iterate over paths in the manifest."""
-
-    def keys():
-        """Obtain a list of paths in the manifest."""
-
-    def filesnotin(other, match=None):
-        """Obtain the set of paths in this manifest but not in another.
-
-        ``match`` is an optional matcher function to be applied to both
-        manifests.
-
-        Returns a set of paths.
-        """
-
-    def dirs():
-        """Returns an object implementing the ``idirs`` interface."""
-
-    def hasdir(dir):
-        """Returns a bool indicating if a directory is in this manifest."""
-
-    def matches(match):
-        """Generate a new manifest filtered through a matcher.
-
-        Returns an object conforming to the ``imanifestdict`` interface.
-        """
-
-    def walk(match):
-        """Generator of paths in manifest satisfying a matcher.
-
-        This is equivalent to ``self.matches(match).iterkeys()`` except a new
-        manifest object is not created.
-
-        If the matcher has explicit files listed and they don't exist in
-        the manifest, ``match.bad()`` is called for each missing file.
-        """
-
-    def diff(other, match=None, clean=False):
-        """Find differences between this manifest and another.
-
-        This manifest is compared to ``other``.
-
-        If ``match`` is provided, the two manifests are filtered against this
-        matcher and only entries satisfying the matcher are compared.
-
-        If ``clean`` is True, unchanged files are included in the returned
-        object.
-
-        Returns a dict with paths as keys and values of 2-tuples of 2-tuples of
-        the form ``((node1, flag1), (node2, flag2))`` where ``(node1, flag1)``
-        represents the node and flags for this manifest and ``(node2, flag2)``
-        are the same for the other manifest.
-        """
-
-    def setflag(path, flag):
-        """Set the flag value for a given path.
-
-        Raises ``KeyError`` if the path is not already in the manifest.
-        """
-
-    def get(path, default=None):
-        """Obtain the node value for a path or a default value if missing."""
-
-    def flags(path, default=''):
-        """Return the flags value for a path or a default value if missing."""
-
-    def copy():
-        """Return a copy of this manifest."""
-
-    def items():
-        """Returns an iterable of (path, node) for items in this manifest."""
-
-    def iteritems():
-        """Identical to items()."""
-
-    def iterentries():
-        """Returns an iterable of (path, node, flags) for this manifest.
-
-        Similar to ``iteritems()`` except items are a 3-tuple and include
-        flags.
-        """
-
-    def text():
-        """Obtain the raw data representation for this manifest.
-
-        Result is used to create a manifest revision.
-        """
-
-    def fastdelta(base, changes):
-        """Obtain a delta between this manifest and another given changes.
-
-        ``base`` in the raw data representation for another manifest.
-
-        ``changes`` is an iterable of ``(path, to_delete)``.
-
-        Returns a 2-tuple containing ``bytearray(self.text())`` and the
-        delta between ``base`` and this manifest.
-        """
-
-class imanifestrevisionbase(interfaceutil.Interface):
-    """Base interface representing a single revision of a manifest.
-
-    Should not be used as a primary interface: should always be inherited
-    as part of a larger interface.
-    """
-
-    def new():
-        """Obtain a new manifest instance.
-
-        Returns an object conforming to the ``imanifestrevisionwritable``
-        interface. The instance will be associated with the same
-        ``imanifestlog`` collection as this instance.
-        """
-
-    def copy():
-        """Obtain a copy of this manifest instance.
-
-        Returns an object conforming to the ``imanifestrevisionwritable``
-        interface. The instance will be associated with the same
-        ``imanifestlog`` collection as this instance.
-        """
-
-    def read():
-        """Obtain the parsed manifest data structure.
-
-        The returned object conforms to the ``imanifestdict`` interface.
-        """
-
-class imanifestrevisionstored(imanifestrevisionbase):
-    """Interface representing a manifest revision committed to storage."""
-
-    def node():
-        """The binary node for this manifest."""
-
-    parents = interfaceutil.Attribute(
-        """List of binary nodes that are parents for this manifest revision."""
-    )
-
-    def readdelta(shallow=False):
-        """Obtain the manifest data structure representing changes from parent.
-
-        This manifest is compared to its 1st parent. A new manifest representing
-        those differences is constructed.
-
-        The returned object conforms to the ``imanifestdict`` interface.
-        """
-
-    def readfast(shallow=False):
-        """Calls either ``read()`` or ``readdelta()``.
-
-        The faster of the two options is called.
-        """
-
-    def find(key):
-        """Calls self.read().find(key)``.
-
-        Returns a 2-tuple of ``(node, flags)`` or raises ``KeyError``.
-        """
-
-class imanifestrevisionwritable(imanifestrevisionbase):
-    """Interface representing a manifest revision that can be committed."""
-
-    def write(transaction, linkrev, p1node, p2node, added, removed, match=None):
-        """Add this revision to storage.
-
-        Takes a transaction object, the changeset revision number it will
-        be associated with, its parent nodes, and lists of added and
-        removed paths.
-
-        If match is provided, storage can choose not to inspect or write out
-        items that do not match. Storage is still required to be able to provide
-        the full manifest in the future for any directories written (these
-        manifests should not be "narrowed on disk").
-
-        Returns the binary node of the created revision.
-        """
-
-class imanifeststorage(interfaceutil.Interface):
-    """Storage interface for manifest data."""
-
-    tree = interfaceutil.Attribute(
-        """The path to the directory this manifest tracks.
-
-        The empty bytestring represents the root manifest.
-        """)
-
-    index = interfaceutil.Attribute(
-        """An ``ifilerevisionssequence`` instance.""")
-
-    indexfile = interfaceutil.Attribute(
-        """Path of revlog index file.
-
-        TODO this is revlog specific and should not be exposed.
-        """)
-
-    opener = interfaceutil.Attribute(
-        """VFS opener to use to access underlying files used for storage.
-
-        TODO this is revlog specific and should not be exposed.
-        """)
-
-    version = interfaceutil.Attribute(
-        """Revlog version number.
-
-        TODO this is revlog specific and should not be exposed.
-        """)
-
-    _generaldelta = interfaceutil.Attribute(
-        """Whether generaldelta storage is being used.
-
-        TODO this is revlog specific and should not be exposed.
-        """)
-
-    fulltextcache = interfaceutil.Attribute(
-        """Dict with cache of fulltexts.
-
-        TODO this doesn't feel appropriate for the storage interface.
-        """)
-
-    def __len__():
-        """Obtain the number of revisions stored for this manifest."""
-
-    def __iter__():
-        """Iterate over revision numbers for this manifest."""
-
-    def rev(node):
-        """Obtain the revision number given a binary node.
-
-        Raises ``error.LookupError`` if the node is not known.
-        """
-
-    def node(rev):
-        """Obtain the node value given a revision number.
-
-        Raises ``error.LookupError`` if the revision is not known.
-        """
-
-    def lookup(value):
-        """Attempt to resolve a value to a node.
-
-        Value can be a binary node, hex node, revision number, or a bytes
-        that can be converted to an integer.
-
-        Raises ``error.LookupError`` if a ndoe could not be resolved.
-        """
-
-    def parents(node):
-        """Returns a 2-tuple of parent nodes for a node.
-
-        Values will be ``nullid`` if the parent is empty.
-        """
-
-    def parentrevs(rev):
-        """Like parents() but operates on revision numbers."""
-
-    def linkrev(rev):
-        """Obtain the changeset revision number a revision is linked to."""
-
-    def revision(node, _df=None, raw=False):
-        """Obtain fulltext data for a node."""
-
-    def rawdata(node, _df=None):
-        """Obtain raw data for a node."""
-
-    def revdiff(rev1, rev2):
-        """Obtain a delta between two revision numbers.
-
-        The returned data is the result of ``bdiff.bdiff()`` on the raw
-        revision data.
-        """
-
-    def cmp(node, fulltext):
-        """Compare fulltext to another revision.
-
-        Returns True if the fulltext is different from what is stored.
-        """
-
-    def emitrevisions(nodes,
-                      nodesorder=None,
-                      revisiondata=False,
-                      assumehaveparentrevisions=False):
-        """Produce ``irevisiondelta`` describing revisions.
-
-        See the documentation for ``ifiledata`` for more.
-        """
-
-    def addgroup(deltas, linkmapper, transaction, addrevisioncb=None):
-        """Process a series of deltas for storage.
-
-        See the documentation in ``ifilemutation`` for more.
-        """
-
-    def rawsize(rev):
-        """Obtain the size of tracked data.
-
-        Is equivalent to ``len(m.rawdata(node))``.
-
-        TODO this method is only used by upgrade code and may be removed.
-        """
-
-    def getstrippoint(minlink):
-        """Find minimum revision that must be stripped to strip a linkrev.
-
-        See the documentation in ``ifilemutation`` for more.
-        """
-
-    def strip(minlink, transaction):
-        """Remove storage of items starting at a linkrev.
-
-        See the documentation in ``ifilemutation`` for more.
-        """
-
-    def checksize():
-        """Obtain the expected sizes of backing files.
-
-        TODO this is used by verify and it should not be part of the interface.
-        """
-
-    def files():
-        """Obtain paths that are backing storage for this manifest.
-
-        TODO this is used by verify and there should probably be a better API
-        for this functionality.
-        """
-
-    def deltaparent(rev):
-        """Obtain the revision that a revision is delta'd against.
-
-        TODO delta encoding is an implementation detail of storage and should
-        not be exposed to the storage interface.
-        """
-
-    def clone(tr, dest, **kwargs):
-        """Clone this instance to another."""
-
-    def clearcaches(clear_persisted_data=False):
-        """Clear any caches associated with this instance."""
-
-    def dirlog(d):
-        """Obtain a manifest storage instance for a tree."""
-
-    def add(m, transaction, link, p1, p2, added, removed, readtree=None,
-            match=None):
-        """Add a revision to storage.
-
-        ``m`` is an object conforming to ``imanifestdict``.
-
-        ``link`` is the linkrev revision number.
-
-        ``p1`` and ``p2`` are the parent revision numbers.
-
-        ``added`` and ``removed`` are iterables of added and removed paths,
-        respectively.
-
-        ``readtree`` is a function that can be used to read the child tree(s)
-        when recursively writing the full tree structure when using
-        treemanifets.
-
-        ``match`` is a matcher that can be used to hint to storage that not all
-        paths must be inspected; this is an optimization and can be safely
-        ignored. Note that the storage must still be able to reproduce a full
-        manifest including files that did not match.
-        """
-
-    def storageinfo(exclusivefiles=False, sharedfiles=False,
-                    revisionscount=False, trackedsize=False,
-                    storedsize=False):
-        """Obtain information about storage for this manifest's data.
-
-        See ``ifilestorage.storageinfo()`` for a description of this method.
-        This one behaves the same way, except for manifest data.
-        """
-
-class imanifestlog(interfaceutil.Interface):
-    """Interface representing a collection of manifest snapshots.
-
-    Represents the root manifest in a repository.
-
-    Also serves as a means to access nested tree manifests and to cache
-    tree manifests.
-    """
-
-    def __getitem__(node):
-        """Obtain a manifest instance for a given binary node.
-
-        Equivalent to calling ``self.get('', node)``.
-
-        The returned object conforms to the ``imanifestrevisionstored``
-        interface.
-        """
-
-    def get(tree, node, verify=True):
-        """Retrieve the manifest instance for a given directory and binary node.
-
-        ``node`` always refers to the node of the root manifest (which will be
-        the only manifest if flat manifests are being used).
-
-        If ``tree`` is the empty string, the root manifest is returned.
-        Otherwise the manifest for the specified directory will be returned
-        (requires tree manifests).
-
-        If ``verify`` is True, ``LookupError`` is raised if the node is not
-        known.
-
-        The returned object conforms to the ``imanifestrevisionstored``
-        interface.
-        """
-
-    def getstorage(tree):
-        """Retrieve an interface to storage for a particular tree.
-
-        If ``tree`` is the empty bytestring, storage for the root manifest will
-        be returned. Otherwise storage for a tree manifest is returned.
-
-        TODO formalize interface for returned object.
-        """
-
-    def clearcaches():
-        """Clear caches associated with this collection."""
-
-    def rev(node):
-        """Obtain the revision number for a binary node.
-
-        Raises ``error.LookupError`` if the node is not known.
-        """
-
-class ilocalrepositoryfilestorage(interfaceutil.Interface):
-    """Local repository sub-interface providing access to tracked file storage.
-
-    This interface defines how a repository accesses storage for a single
-    tracked file path.
-    """
-
-    def file(f):
-        """Obtain a filelog for a tracked path.
-
-        The returned type conforms to the ``ifilestorage`` interface.
-        """
-
-class ilocalrepositorymain(interfaceutil.Interface):
-    """Main interface for local repositories.
-
-    This currently captures the reality of things - not how things should be.
-    """
-
-    supportedformats = interfaceutil.Attribute(
-        """Set of requirements that apply to stream clone.
-
-        This is actually a class attribute and is shared among all instances.
-        """)
-
-    supported = interfaceutil.Attribute(
-        """Set of requirements that this repo is capable of opening.""")
-
-    requirements = interfaceutil.Attribute(
-        """Set of requirements this repo uses.""")
-
-    features = interfaceutil.Attribute(
-        """Set of "features" this repository supports.
-
-        A "feature" is a loosely-defined term. It can refer to a feature
-        in the classical sense or can describe an implementation detail
-        of the repository. For example, a ``readonly`` feature may denote
-        the repository as read-only. Or a ``revlogfilestore`` feature may
-        denote that the repository is using revlogs for file storage.
-
-        The intent of features is to provide a machine-queryable mechanism
-        for repo consumers to test for various repository characteristics.
-
-        Features are similar to ``requirements``. The main difference is that
-        requirements are stored on-disk and represent requirements to open the
-        repository. Features are more run-time capabilities of the repository
-        and more granular capabilities (which may be derived from requirements).
-        """)
-
-    filtername = interfaceutil.Attribute(
-        """Name of the repoview that is active on this repo.""")
-
-    wvfs = interfaceutil.Attribute(
-        """VFS used to access the working directory.""")
-
-    vfs = interfaceutil.Attribute(
-        """VFS rooted at the .hg directory.
-
-        Used to access repository data not in the store.
-        """)
-
-    svfs = interfaceutil.Attribute(
-        """VFS rooted at the store.
-
-        Used to access repository data in the store. Typically .hg/store.
-        But can point elsewhere if the store is shared.
-        """)
-
-    root = interfaceutil.Attribute(
-        """Path to the root of the working directory.""")
-
-    path = interfaceutil.Attribute(
-        """Path to the .hg directory.""")
-
-    origroot = interfaceutil.Attribute(
-        """The filesystem path that was used to construct the repo.""")
-
-    auditor = interfaceutil.Attribute(
-        """A pathauditor for the working directory.
-
-        This checks if a path refers to a nested repository.
-
-        Operates on the filesystem.
-        """)
-
-    nofsauditor = interfaceutil.Attribute(
-        """A pathauditor for the working directory.
-
-        This is like ``auditor`` except it doesn't do filesystem checks.
-        """)
-
-    baseui = interfaceutil.Attribute(
-        """Original ui instance passed into constructor.""")
-
-    ui = interfaceutil.Attribute(
-        """Main ui instance for this instance.""")
-
-    sharedpath = interfaceutil.Attribute(
-        """Path to the .hg directory of the repo this repo was shared from.""")
-
-    store = interfaceutil.Attribute(
-        """A store instance.""")
-
-    spath = interfaceutil.Attribute(
-        """Path to the store.""")
-
-    sjoin = interfaceutil.Attribute(
-        """Alias to self.store.join.""")
-
-    cachevfs = interfaceutil.Attribute(
-        """A VFS used to access the cache directory.
-
-        Typically .hg/cache.
-        """)
-
-    wcachevfs = interfaceutil.Attribute(
-        """A VFS used to access the cache directory dedicated to working copy
-
-        Typically .hg/wcache.
-        """)
-
-    filteredrevcache = interfaceutil.Attribute(
-        """Holds sets of revisions to be filtered.""")
-
-    names = interfaceutil.Attribute(
-        """A ``namespaces`` instance.""")
-
-    def close():
-        """Close the handle on this repository."""
-
-    def peer():
-        """Obtain an object conforming to the ``peer`` interface."""
-
-    def unfiltered():
-        """Obtain an unfiltered/raw view of this repo."""
-
-    def filtered(name, visibilityexceptions=None):
-        """Obtain a named view of this repository."""
-
-    obsstore = interfaceutil.Attribute(
-        """A store of obsolescence data.""")
-
-    changelog = interfaceutil.Attribute(
-        """A handle on the changelog revlog.""")
-
-    manifestlog = interfaceutil.Attribute(
-        """An instance conforming to the ``imanifestlog`` interface.
-
-        Provides access to manifests for the repository.
-        """)
-
-    dirstate = interfaceutil.Attribute(
-        """Working directory state.""")
-
-    narrowpats = interfaceutil.Attribute(
-        """Matcher patterns for this repository's narrowspec.""")
-
-    def narrowmatch(match=None, includeexact=False):
-        """Obtain a matcher for the narrowspec."""
-
-    def setnarrowpats(newincludes, newexcludes):
-        """Define the narrowspec for this repository."""
-
-    def __getitem__(changeid):
-        """Try to resolve a changectx."""
-
-    def __contains__(changeid):
-        """Whether a changeset exists."""
-
-    def __nonzero__():
-        """Always returns True."""
-        return True
-
-    __bool__ = __nonzero__
-
-    def __len__():
-        """Returns the number of changesets in the repo."""
-
-    def __iter__():
-        """Iterate over revisions in the changelog."""
-
-    def revs(expr, *args):
-        """Evaluate a revset.
-
-        Emits revisions.
-        """
-
-    def set(expr, *args):
-        """Evaluate a revset.
-
-        Emits changectx instances.
-        """
-
-    def anyrevs(specs, user=False, localalias=None):
-        """Find revisions matching one of the given revsets."""
-
-    def url():
-        """Returns a string representing the location of this repo."""
-
-    def hook(name, throw=False, **args):
-        """Call a hook."""
-
-    def tags():
-        """Return a mapping of tag to node."""
-
-    def tagtype(tagname):
-        """Return the type of a given tag."""
-
-    def tagslist():
-        """Return a list of tags ordered by revision."""
-
-    def nodetags(node):
-        """Return the tags associated with a node."""
-
-    def nodebookmarks(node):
-        """Return the list of bookmarks pointing to the specified node."""
-
-    def branchmap():
-        """Return a mapping of branch to heads in that branch."""
-
-    def revbranchcache():
-        pass
-
-    def branchtip(branchtip, ignoremissing=False):
-        """Return the tip node for a given branch."""
-
-    def lookup(key):
-        """Resolve the node for a revision."""
-
-    def lookupbranch(key):
-        """Look up the branch name of the given revision or branch name."""
-
-    def known(nodes):
-        """Determine whether a series of nodes is known.
-
-        Returns a list of bools.
-        """
-
-    def local():
-        """Whether the repository is local."""
-        return True
-
-    def publishing():
-        """Whether the repository is a publishing repository."""
-
-    def cancopy():
-        pass
-
-    def shared():
-        """The type of shared repository or None."""
-
-    def wjoin(f, *insidef):
-        """Calls self.vfs.reljoin(self.root, f, *insidef)"""
-
-    def setparents(p1, p2):
-        """Set the parent nodes of the working directory."""
-
-    def filectx(path, changeid=None, fileid=None):
-        """Obtain a filectx for the given file revision."""
-
-    def getcwd():
-        """Obtain the current working directory from the dirstate."""
-
-    def pathto(f, cwd=None):
-        """Obtain the relative path to a file."""
-
-    def adddatafilter(name, fltr):
-        pass
-
-    def wread(filename):
-        """Read a file from wvfs, using data filters."""
-
-    def wwrite(filename, data, flags, backgroundclose=False, **kwargs):
-        """Write data to a file in the wvfs, using data filters."""
-
-    def wwritedata(filename, data):
-        """Resolve data for writing to the wvfs, using data filters."""
-
-    def currenttransaction():
-        """Obtain the current transaction instance or None."""
-
-    def transaction(desc, report=None):
-        """Open a new transaction to write to the repository."""
-
-    def undofiles():
-        """Returns a list of (vfs, path) for files to undo transactions."""
-
-    def recover():
-        """Roll back an interrupted transaction."""
-
-    def rollback(dryrun=False, force=False):
-        """Undo the last transaction.
-
-        DANGEROUS.
-        """
-
-    def updatecaches(tr=None, full=False):
-        """Warm repo caches."""
-
-    def invalidatecaches():
-        """Invalidate cached data due to the repository mutating."""
-
-    def invalidatevolatilesets():
-        pass
-
-    def invalidatedirstate():
-        """Invalidate the dirstate."""
-
-    def invalidate(clearfilecache=False):
-        pass
-
-    def invalidateall():
-        pass
-
-    def lock(wait=True):
-        """Lock the repository store and return a lock instance."""
-
-    def wlock(wait=True):
-        """Lock the non-store parts of the repository."""
-
-    def currentwlock():
-        """Return the wlock if it's held or None."""
-
-    def checkcommitpatterns(wctx, vdirs, match, status, fail):
-        pass
-
-    def commit(text='', user=None, date=None, match=None, force=False,
-               editor=False, extra=None):
-        """Add a new revision to the repository."""
-
-    def commitctx(ctx, error=False, origctx=None):
-        """Commit a commitctx instance to the repository."""
-
-    def destroying():
-        """Inform the repository that nodes are about to be destroyed."""
-
-    def destroyed():
-        """Inform the repository that nodes have been destroyed."""
-
-    def status(node1='.', node2=None, match=None, ignored=False,
-               clean=False, unknown=False, listsubrepos=False):
-        """Convenience method to call repo[x].status()."""
-
-    def addpostdsstatus(ps):
-        pass
-
-    def postdsstatus():
-        pass
-
-    def clearpostdsstatus():
-        pass
-
-    def heads(start=None):
-        """Obtain list of nodes that are DAG heads."""
-
-    def branchheads(branch=None, start=None, closed=False):
-        pass
-
-    def branches(nodes):
-        pass
-
-    def between(pairs):
-        pass
-
-    def checkpush(pushop):
-        pass
-
-    prepushoutgoinghooks = interfaceutil.Attribute(
-        """util.hooks instance.""")
-
-    def pushkey(namespace, key, old, new):
-        pass
-
-    def listkeys(namespace):
-        pass
-
-    def debugwireargs(one, two, three=None, four=None, five=None):
-        pass
-
-    def savecommitmessage(text):
-        pass
-
-class completelocalrepository(ilocalrepositorymain,
-                              ilocalrepositoryfilestorage):
-    """Complete interface for a local repository."""
-
-class iwireprotocolcommandcacher(interfaceutil.Interface):
-    """Represents a caching backend for wire protocol commands.
-
-    Wire protocol version 2 supports transparent caching of many commands.
-    To leverage this caching, servers can activate objects that cache
-    command responses. Objects handle both cache writing and reading.
-    This interface defines how that response caching mechanism works.
-
-    Wire protocol version 2 commands emit a series of objects that are
-    serialized and sent to the client. The caching layer exists between
-    the invocation of the command function and the sending of its output
-    objects to an output layer.
-
-    Instances of this interface represent a binding to a cache that
-    can serve a response (in place of calling a command function) and/or
-    write responses to a cache for subsequent use.
-
-    When a command request arrives, the following happens with regards
-    to this interface:
-
-    1. The server determines whether the command request is cacheable.
-    2. If it is, an instance of this interface is spawned.
-    3. The cacher is activated in a context manager (``__enter__`` is called).
-    4. A cache *key* for that request is derived. This will call the
-       instance's ``adjustcachekeystate()`` method so the derivation
-       can be influenced.
-    5. The cacher is informed of the derived cache key via a call to
-       ``setcachekey()``.
-    6. The cacher's ``lookup()`` method is called to test for presence of
-       the derived key in the cache.
-    7. If ``lookup()`` returns a hit, that cached result is used in place
-       of invoking the command function. ``__exit__`` is called and the instance
-       is discarded.
-    8. The command function is invoked.
-    9. ``onobject()`` is called for each object emitted by the command
-       function.
-    10. After the final object is seen, ``onfinished()`` is called.
-    11. ``__exit__`` is called to signal the end of use of the instance.
-
-    Cache *key* derivation can be influenced by the instance.
-
-    Cache keys are initially derived by a deterministic representation of
-    the command request. This includes the command name, arguments, protocol
-    version, etc. This initial key derivation is performed by CBOR-encoding a
-    data structure and feeding that output into a hasher.
-
-    Instances of this interface can influence this initial key derivation
-    via ``adjustcachekeystate()``.
-
-    The instance is informed of the derived cache key via a call to
-    ``setcachekey()``. The instance must store the key locally so it can
-    be consulted on subsequent operations that may require it.
-
-    When constructed, the instance has access to a callable that can be used
-    for encoding response objects. This callable receives as its single
-    argument an object emitted by a command function. It returns an iterable
-    of bytes chunks representing the encoded object. Unless the cacher is
-    caching native Python objects in memory or has a way of reconstructing
-    the original Python objects, implementations typically call this function
-    to produce bytes from the output objects and then store those bytes in
-    the cache. When it comes time to re-emit those bytes, they are wrapped
-    in a ``wireprototypes.encodedresponse`` instance to tell the output
-    layer that they are pre-encoded.
-
-    When receiving the objects emitted by the command function, instances
-    can choose what to do with those objects. The simplest thing to do is
-    re-emit the original objects. They will be forwarded to the output
-    layer and will be processed as if the cacher did not exist.
-
-    Implementations could also choose to not emit objects - instead locally
-    buffering objects or their encoded representation. They could then emit
-    a single "coalesced" object when ``onfinished()`` is called. In
-    this way, the implementation would function as a filtering layer of
-    sorts.
-
-    When caching objects, typically the encoded form of the object will
-    be stored. Keep in mind that if the original object is forwarded to
-    the output layer, it will need to be encoded there as well. For large
-    output, this redundant encoding could add overhead. Implementations
-    could wrap the encoded object data in ``wireprototypes.encodedresponse``
-    instances to avoid this overhead.
-    """
-    def __enter__():
-        """Marks the instance as active.
-
-        Should return self.
-        """
-
-    def __exit__(exctype, excvalue, exctb):
-        """Called when cacher is no longer used.
-
-        This can be used by implementations to perform cleanup actions (e.g.
-        disconnecting network sockets, aborting a partially cached response.
-        """
-
-    def adjustcachekeystate(state):
-        """Influences cache key derivation by adjusting state to derive key.
-
-        A dict defining the state used to derive the cache key is passed.
-
-        Implementations can modify this dict to record additional state that
-        is wanted to influence key derivation.
-
-        Implementations are *highly* encouraged to not modify or delete
-        existing keys.
-        """
-
-    def setcachekey(key):
-        """Record the derived cache key for this request.
-
-        Instances may mutate the key for internal usage, as desired. e.g.
-        instances may wish to prepend the repo name, introduce path
-        components for filesystem or URL addressing, etc. Behavior is up to
-        the cache.
-
-        Returns a bool indicating if the request is cacheable by this
-        instance.
-        """
-
-    def lookup():
-        """Attempt to resolve an entry in the cache.
-
-        The instance is instructed to look for the cache key that it was
-        informed about via the call to ``setcachekey()``.
-
-        If there's no cache hit or the cacher doesn't wish to use the cached
-        entry, ``None`` should be returned.
-
-        Else, a dict defining the cached result should be returned. The
-        dict may have the following keys:
-
-        objs
-           An iterable of objects that should be sent to the client. That
-           iterable of objects is expected to be what the command function
-           would return if invoked or an equivalent representation thereof.
-        """
-
-    def onobject(obj):
-        """Called when a new object is emitted from the command function.
-
-        Receives as its argument the object that was emitted from the
-        command function.
-
-        This method returns an iterator of objects to forward to the output
-        layer. The easiest implementation is a generator that just
-        ``yield obj``.
-        """
-
-    def onfinished():
-        """Called after all objects have been emitted from the command function.
-
-        Implementations should return an iterator of objects to forward to
-        the output layer.
-
-        This method can be a generator.
-        """
--- a/mercurial/revlog.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/revlog.py	Sun Aug 18 00:45:33 2019 +0300
@@ -65,10 +65,12 @@
     mdiff,
     policy,
     pycompat,
-    repository,
     templatefilters,
     util,
 )
+from .interfaces import (
+    repository,
+)
 from .revlogutils import (
     deltas as deltautil,
     flagutil,
--- a/mercurial/revlogutils/constants.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/revlogutils/constants.py	Sun Aug 18 00:45:33 2019 +0300
@@ -9,7 +9,7 @@
 
 from __future__ import absolute_import
 
-from .. import (
+from ..interfaces import (
     repository,
 )
 
--- a/mercurial/streamclone.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/streamclone.py	Sun Aug 18 00:45:33 2019 +0300
@@ -12,13 +12,15 @@
 import struct
 
 from .i18n import _
+from .interfaces import (
+    repository,
+)
 from . import (
     cacheutil,
     error,
     narrowspec,
     phases,
     pycompat,
-    repository,
     store,
     util,
 )
--- a/mercurial/testing/storage.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/testing/storage.py	Sun Aug 18 00:45:33 2019 +0300
@@ -17,6 +17,8 @@
 from .. import (
     error,
     mdiff,
+)
+from ..interfaces import (
     repository,
 )
 from ..utils import (
--- a/mercurial/utils/storageutil.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/utils/storageutil.py	Sun Aug 18 00:45:33 2019 +0300
@@ -22,8 +22,8 @@
     error,
     mdiff,
     pycompat,
-    repository,
 )
+from ..interfaces import repository
 
 _nullhash = hashlib.sha1(nullid)
 
--- a/mercurial/wireprotov1peer.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/mercurial/wireprotov1peer.py	Sun Aug 18 00:45:33 2019 +0300
@@ -22,10 +22,12 @@
     error,
     pushkey as pushkeymod,
     pycompat,
-    repository,
     util,
     wireprototypes,
 )
+from .interfaces import (
+    repository,
+)
 from .utils import (
     interfaceutil,
 )
--- a/setup.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/setup.py	Sun Aug 18 00:45:33 2019 +0300
@@ -1067,6 +1067,7 @@
             'mercurial.cext',
             'mercurial.cffi',
             'mercurial.hgweb',
+            'mercurial.interfaces',
             'mercurial.pure',
             'mercurial.thirdparty',
             'mercurial.thirdparty.attr',
--- a/tests/notcapable	Thu Aug 22 16:47:31 2019 -0700
+++ b/tests/notcapable	Sun Aug 18 00:45:33 2019 +0300
@@ -6,7 +6,8 @@
 fi
 
 cat > notcapable-$CAP.py << EOF
-from mercurial import extensions, localrepo, repository
+from mercurial import extensions, localrepo
+from mercurial.interfaces import repository
 def extsetup(ui):
     extensions.wrapfunction(repository.peer, 'capable', wrapcapable)
     extensions.wrapfunction(localrepo.localrepository, 'peer', wrappeer)
--- a/tests/pullext.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/tests/pullext.py	Sun Aug 18 00:45:33 2019 +0300
@@ -13,6 +13,8 @@
     error,
     extensions,
     localrepo,
+)
+from mercurial.interfaces import (
     repository,
 )
 
--- a/tests/simplestorerepo.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/tests/simplestorerepo.py	Sun Aug 18 00:45:33 2019 +0300
@@ -32,11 +32,13 @@
     localrepo,
     mdiff,
     pycompat,
-    repository,
     revlog,
     store,
     verify,
 )
+from mercurial.interfaces import (
+    repository,
+)
 from mercurial.utils import (
     cborutil,
     interfaceutil,
--- a/tests/test-check-interfaces.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/tests/test-check-interfaces.py	Sun Aug 18 00:45:33 2019 +0300
@@ -14,6 +14,9 @@
                     'test-repo']):
     sys.exit(80)
 
+from mercurial.interfaces import (
+    repository,
+)
 from mercurial.thirdparty.zope import (
     interface as zi,
 )
@@ -27,7 +30,6 @@
     localrepo,
     manifest,
     pycompat,
-    repository,
     revlog,
     sshpeer,
     statichttprepo,
--- a/tests/wireprotosimplecache.py	Thu Aug 22 16:47:31 2019 -0700
+++ b/tests/wireprotosimplecache.py	Sun Aug 18 00:45:33 2019 +0300
@@ -10,12 +10,14 @@
 from mercurial import (
     extensions,
     registrar,
-    repository,
     util,
     wireprotoserver,
     wireprototypes,
     wireprotov2server,
 )
+from mercurial.interfaces import (
+    repository,
+)
 from mercurial.utils import (
     interfaceutil,
     stringutil,