Mercurial > public > mercurial-scm > hg-stable
diff tests/test-remotefilelog-histpack.py @ 40545:3a333a582d7b
remotefilelog: import pruned-down remotefilelog extension from hg-experimental
This is remotefilelog as of my recent patches for compatibility with
current tip of hg, minus support for old versions of Mercurial and
some FB-specific features like their treemanifest extension and
fetching linkrev data from a patched phabricator. The file extutil.py
moved from hgext3rd to remotefilelog.
This is not yet ready to be landed, consider it a preview for
now. Planned changes include:
* replace lz4 with zstd
* rename some capabilities, requirements and wireproto commands to mark
them as experimental
* consolidate bits of shallowutil with related functions (eg readfile)
I'm certainly open to other (small) changes, but my rough mission is
to land this largely as-is so we can use it as a model of the
functionality we need going forward for lazy-fetching of file contents
from a server.
# no-check-commit because of a few foo_bar functions
Differential Revision: https://phab.mercurial-scm.org/D4782
author | Augie Fackler <augie@google.com> |
---|---|
date | Thu, 27 Sep 2018 13:03:19 -0400 |
parents | |
children | 10c10da14c5d |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/tests/test-remotefilelog-histpack.py Thu Sep 27 13:03:19 2018 -0400 @@ -0,0 +1,274 @@ +#!/usr/bin/env python +from __future__ import absolute_import + +import hashlib +import os +import random +import shutil +import stat +import struct +import sys +import tempfile +import unittest + +import silenttestrunner + +from mercurial.node import nullid +from mercurial import ( + ui as uimod, +) +# Load the local remotefilelog, not the system one +sys.path[0:0] = [os.path.join(os.path.dirname(__file__), '..')] +from hgext.remotefilelog import ( + basepack, + historypack, +) + +class histpacktests(unittest.TestCase): + def setUp(self): + self.tempdirs = [] + + def tearDown(self): + for d in self.tempdirs: + shutil.rmtree(d) + + def makeTempDir(self): + tempdir = tempfile.mkdtemp() + self.tempdirs.append(tempdir) + return tempdir + + def getHash(self, content): + return hashlib.sha1(content).digest() + + def getFakeHash(self): + return ''.join(chr(random.randint(0, 255)) for _ in range(20)) + + def createPack(self, revisions=None): + """Creates and returns a historypack containing the specified revisions. + + `revisions` is a list of tuples, where each tuple contains a filanem, + node, p1node, p2node, and linknode. + """ + if revisions is None: + revisions = [("filename", self.getFakeHash(), nullid, nullid, + self.getFakeHash(), None)] + + packdir = self.makeTempDir() + packer = historypack.mutablehistorypack(uimod.ui(), packdir, + version=1) + + for filename, node, p1, p2, linknode, copyfrom in revisions: + packer.add(filename, node, p1, p2, linknode, copyfrom) + + path = packer.close() + return historypack.historypack(path) + + def testAddSingle(self): + """Test putting a single entry into a pack and reading it out. + """ + filename = "foo" + node = self.getFakeHash() + p1 = self.getFakeHash() + p2 = self.getFakeHash() + linknode = self.getFakeHash() + + revisions = [(filename, node, p1, p2, linknode, None)] + pack = self.createPack(revisions) + + actual = pack.getancestors(filename, node)[node] + self.assertEquals(p1, actual[0]) + self.assertEquals(p2, actual[1]) + self.assertEquals(linknode, actual[2]) + + def testAddMultiple(self): + """Test putting multiple unrelated revisions into a pack and reading + them out. + """ + revisions = [] + for i in range(10): + filename = "foo-%s" % i + node = self.getFakeHash() + p1 = self.getFakeHash() + p2 = self.getFakeHash() + linknode = self.getFakeHash() + revisions.append((filename, node, p1, p2, linknode, None)) + + pack = self.createPack(revisions) + + for filename, node, p1, p2, linknode, copyfrom in revisions: + actual = pack.getancestors(filename, node)[node] + self.assertEquals(p1, actual[0]) + self.assertEquals(p2, actual[1]) + self.assertEquals(linknode, actual[2]) + self.assertEquals(copyfrom, actual[3]) + + def testAddAncestorChain(self): + """Test putting multiple revisions in into a pack and read the ancestor + chain. + """ + revisions = [] + filename = "foo" + lastnode = nullid + for i in range(10): + node = self.getFakeHash() + revisions.append((filename, node, lastnode, nullid, nullid, None)) + lastnode = node + + # revisions must be added in topological order, newest first + revisions = list(reversed(revisions)) + pack = self.createPack(revisions) + + # Test that the chain has all the entries + ancestors = pack.getancestors(revisions[0][0], revisions[0][1]) + for filename, node, p1, p2, linknode, copyfrom in revisions: + ap1, ap2, alinknode, acopyfrom = ancestors[node] + self.assertEquals(ap1, p1) + self.assertEquals(ap2, p2) + self.assertEquals(alinknode, linknode) + self.assertEquals(acopyfrom, copyfrom) + + def testPackMany(self): + """Pack many related and unrelated ancestors. + """ + # Build a random pack file + allentries = {} + ancestorcounts = {} + revisions = [] + random.seed(0) + for i in range(100): + filename = "filename-%s" % i + entries = [] + p2 = nullid + linknode = nullid + for j in range(random.randint(1, 100)): + node = self.getFakeHash() + p1 = nullid + if len(entries) > 0: + p1 = entries[random.randint(0, len(entries) - 1)] + entries.append(node) + revisions.append((filename, node, p1, p2, linknode, None)) + allentries[(filename, node)] = (p1, p2, linknode) + if p1 == nullid: + ancestorcounts[(filename, node)] = 1 + else: + newcount = ancestorcounts[(filename, p1)] + 1 + ancestorcounts[(filename, node)] = newcount + + # Must add file entries in reverse topological order + revisions = list(reversed(revisions)) + pack = self.createPack(revisions) + + # Verify the pack contents + for (filename, node), (p1, p2, lastnode) in allentries.iteritems(): + ancestors = pack.getancestors(filename, node) + self.assertEquals(ancestorcounts[(filename, node)], + len(ancestors)) + for anode, (ap1, ap2, alinknode, copyfrom) in ancestors.iteritems(): + ep1, ep2, elinknode = allentries[(filename, anode)] + self.assertEquals(ap1, ep1) + self.assertEquals(ap2, ep2) + self.assertEquals(alinknode, elinknode) + self.assertEquals(copyfrom, None) + + def testGetNodeInfo(self): + revisions = [] + filename = "foo" + lastnode = nullid + for i in range(10): + node = self.getFakeHash() + revisions.append((filename, node, lastnode, nullid, nullid, None)) + lastnode = node + + pack = self.createPack(revisions) + + # Test that getnodeinfo returns the expected results + for filename, node, p1, p2, linknode, copyfrom in revisions: + ap1, ap2, alinknode, acopyfrom = pack.getnodeinfo(filename, node) + self.assertEquals(ap1, p1) + self.assertEquals(ap2, p2) + self.assertEquals(alinknode, linknode) + self.assertEquals(acopyfrom, copyfrom) + + def testGetMissing(self): + """Test the getmissing() api. + """ + revisions = [] + filename = "foo" + for i in range(10): + node = self.getFakeHash() + p1 = self.getFakeHash() + p2 = self.getFakeHash() + linknode = self.getFakeHash() + revisions.append((filename, node, p1, p2, linknode, None)) + + pack = self.createPack(revisions) + + missing = pack.getmissing([(filename, revisions[0][1])]) + self.assertFalse(missing) + + missing = pack.getmissing([(filename, revisions[0][1]), + (filename, revisions[1][1])]) + self.assertFalse(missing) + + fakenode = self.getFakeHash() + missing = pack.getmissing([(filename, revisions[0][1]), + (filename, fakenode)]) + self.assertEquals(missing, [(filename, fakenode)]) + + # Test getmissing on a non-existant filename + missing = pack.getmissing([("bar", fakenode)]) + self.assertEquals(missing, [("bar", fakenode)]) + + def testAddThrows(self): + pack = self.createPack() + + try: + pack.add('filename', nullid, nullid, nullid, nullid, None) + self.assertTrue(False, "historypack.add should throw") + except RuntimeError: + pass + + def testBadVersionThrows(self): + pack = self.createPack() + path = pack.path + '.histpack' + with open(path) as f: + raw = f.read() + raw = struct.pack('!B', 255) + raw[1:] + os.chmod(path, os.stat(path).st_mode | stat.S_IWRITE) + with open(path, 'w+') as f: + f.write(raw) + + try: + pack = historypack.historypack(pack.path) + self.assertTrue(False, "bad version number should have thrown") + except RuntimeError: + pass + + def testLargePack(self): + """Test creating and reading from a large pack with over X entries. + This causes it to use a 2^16 fanout table instead.""" + total = basepack.SMALLFANOUTCUTOFF + 1 + revisions = [] + for i in xrange(total): + filename = "foo-%s" % i + node = self.getFakeHash() + p1 = self.getFakeHash() + p2 = self.getFakeHash() + linknode = self.getFakeHash() + revisions.append((filename, node, p1, p2, linknode, None)) + + pack = self.createPack(revisions) + self.assertEquals(pack.params.fanoutprefix, basepack.LARGEFANOUTPREFIX) + + for filename, node, p1, p2, linknode, copyfrom in revisions: + actual = pack.getancestors(filename, node)[node] + self.assertEquals(p1, actual[0]) + self.assertEquals(p2, actual[1]) + self.assertEquals(linknode, actual[2]) + self.assertEquals(copyfrom, actual[3]) +# TODO: +# histpack store: +# - repack two packs into one + +if __name__ == '__main__': + silenttestrunner.main(__name__)