Thu, 12 Jun 2014 03:20:28 +0530 run-tests: fixes the number of tests ran when '--retest' is enabled
anuraggoel <anurag.dsps@gmail.com> [Thu, 12 Jun 2014 03:20:28 +0530] rev 21740
run-tests: fixes the number of tests ran when '--retest' is enabled This patch fixes a regression recently introduced by a refactoring. (see 92a6b16c9186 and about 200 previous changesets from Gregory Szorc) While retesting, that is when '--retest' is enabled, only failure tests run and others either skipped or ignored. During retesting, "result.testsRun" holds the count of failure test that has run. But as while printing output, we have subtracted the skipped and ignored count from "result.testsRun". Therefore, to make the count remain the same, we need to add skipped and ignored count before printing.
Thu, 12 Jun 2014 02:36:52 +0530 run-tests: skipped test should not produce 'i' mark while retesting
anuraggoel <anurag.dsps@gmail.com> [Thu, 12 Jun 2014 02:36:52 +0530] rev 21739
run-tests: skipped test should not produce 'i' mark while retesting When '--retest' option is enabled then skipped test should not produce 'i' mark. This fixes a regression introduced by 92a6b16c9186 and about 200 previous changesets from Gregory Szorc.
Fri, 30 May 2014 12:10:56 -0700 test-run-tests.t: test running a failing test
Pierre-Yves David <pierre-yves.david@fb.com> [Fri, 30 May 2014 12:10:56 -0700] rev 21738
test-run-tests.t: test running a failing test
Fri, 13 Jun 2014 16:36:53 -0500 run-tests: filter whitespace at end of error diffs
Matt Mackall <mpm@selenic.com> [Fri, 13 Jun 2014 16:36:53 -0500] rev 21737
run-tests: filter whitespace at end of error diffs Python 2.6 and 2.7 produce slightly different diff headers.
Thu, 12 Jun 2014 01:28:22 +0530 run-tests: produce '!' mark after running a failing test
anuraggoel <anurag.dsps@gmail.com> [Thu, 12 Jun 2014 01:28:22 +0530] rev 21736
run-tests: produce '!' mark after running a failing test This patch fixes a regression recently introduced by a refactoring (see 92a6b16c9186 and about 200 previous changesets from Gregory Szorc). '!' mark denotes that the test gets failed while testing.
Thu, 12 Jun 2014 01:36:57 +0530 run-tests: produce error on running a failing test
anuraggoel <anurag.dsps@gmail.com> [Thu, 12 Jun 2014 01:36:57 +0530] rev 21735
run-tests: produce error on running a failing test This patch fixes a regression recently introduced by a refactoring (see 92a6b16c9186 and about 200 previous changesets from Gregory Szorc). It produce an error message everytime with a test filename which gets fail while testing except at one condition when '--nodiff' option is enabled.
Fri, 30 May 2014 12:10:19 -0700 test-run-tests.t: test running a passing test
Pierre-Yves David <pierre-yves.david@fb.com> [Fri, 30 May 2014 12:10:19 -0700] rev 21734
test-run-tests.t: test running a passing test
Tue, 10 Jun 2014 14:32:52 -0700 run-tests: don't check for the mercurial library used when using --with-hg
Pierre-Yves David <pierre-yves.david@fb.com> [Tue, 10 Jun 2014 14:32:52 -0700] rev 21733
run-tests: don't check for the mercurial library used when using --with-hg In the ``--with-hg`` case, we cannot properly infer the location of the mercurial library, so we skip the warning completely.
Fri, 30 May 2014 12:09:49 -0700 test: introduce test-run-tests.t
Pierre-Yves David <pierre-yves.david@fb.com> [Fri, 30 May 2014 12:09:49 -0700] rev 21732
test: introduce test-run-tests.t This testfile tests the test runner itself. There were multiple regressions in the last few weeks. Having tests would prevent this in the future.
Fri, 30 May 2014 11:40:24 -0700 tests: comment the main purpose of ``test-unified-test.t``
Pierre-Yves David <pierre-yves.david@fb.com> [Fri, 30 May 2014 11:40:24 -0700] rev 21731
tests: comment the main purpose of ``test-unified-test.t``
(0) -10000 -3000 -1000 -300 -100 -10 +10 +100 +300 +1000 +3000 +10000 tip