[py-svn] pytest commit 2e4c9cdd531a: regenerating examples

commits-noreply at bitbucket.org commits-noreply at bitbucket.org
Fri Nov 26 15:43:09 CET 2010


# HG changeset patch -- Bitbucket.org
# Project pytest
# URL http://bitbucket.org/hpk42/pytest/overview
# User holger krekel <holger at merlinux.eu>
# Date 1290774416 -3600
# Node ID 2e4c9cdd531abbc69a90cd8615ea0d72ab479e58
# Parent  66a7de5f7a3e6778cf0424423dbfde94d4297576
regenerating examples

--- a/doc/Makefile
+++ b/doc/Makefile
@@ -15,7 +15,7 @@ ALLSPHINXOPTS   = -d $(BUILDDIR)/doctree
 .PHONY: help clean html dirhtml singlehtml pickle json htmlhelp qthelp devhelp epub latex latexpdf text man changes linkcheck doctest
 
 regen:
-	COLUMNS=76 regendoc --update *.txt */*.txt
+	PYTHONDONTWRITEBYTECODE=1 COLUMNS=76 regendoc --update *.txt */*.txt
     
 help:
 	@echo "Please use \`make <target>' where <target> is one of"

--- a/doc/unittest.txt
+++ b/doc/unittest.txt
@@ -24,8 +24,8 @@ Running it yields::
 
     $ py.test test_unittest.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_unittest.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
     test_unittest.py F
     
@@ -56,7 +56,7 @@ Running it yields::
     /usr/lib/python2.6/unittest.py:350: AssertionError
     ----------------------------- Captured stdout ------------------------------
     hello
-    ========================= 1 failed in 0.02 seconds =========================
+    ========================= 1 failed in 0.03 seconds =========================
 
 .. _`unittest.py style`: http://docs.python.org/library/unittest.html
 

--- a/doc/projects.txt
+++ b/doc/projects.txt
@@ -15,9 +15,12 @@ Here are some examples of projects using
 * `mwlib <http://pypi.python.org/pypi/mwlib>`_ mediawiki parser and utility library
 * `The Translate Toolkit <http://translate.sourceforge.net/wiki/toolkit/index>`_ for localization and conversion
 * `execnet <http://codespeak.net/execnet>`_ rapid multi-Python deployment
+* `Pacha <http://pacha.cafepais.com/>`_ configuration management in five minutes
 * `bbfreeze <http://pypi.python.org/pypi/bbfreeze>`_ create standalone executables from Python scripts
 * `pdb++ <http://bitbucket.org/antocuni/pdb>`_ a fancier version of PDB
 * `py-s3fuse <http://code.google.com/p/py-s3fuse/>`_ Amazon S3 FUSE based filesystem
+* `waskr <http://pacha.cafepais.com/>`_ WSGI Stats Middleware
+* `guachi <http://code.google.com/p/guachi/>`_ global persistent configs for Python modules
 * `Circuits <http://pypi.python.org/pypi/circuits>`_ lightweight Event Driven Framework
 * `pygtk-helpers <http://bitbucket.org/aafshar/pygtkhelpers-main/>`_ easy interaction with PyGTK
 * `QuantumCore <http://quantumcore.org/>`_ statusmessage and repoze openid plugin

--- a/doc/example/simple.txt
+++ b/doc/example/simple.txt
@@ -7,6 +7,8 @@ basic patterns and examples
 pass different values to a test function, depending on command line options
 ----------------------------------------------------------------------------
 
+.. regendoc:wipe
+
 Suppose we want to write a test that depends on a command line option.
 Here is a basic pattern how to achieve this::
 
@@ -32,7 +34,8 @@ provide the ``cmdopt`` through a :ref:`f
 
 Let's run this without supplying our new command line option::
 
-    $ py.test -q
+    $ py.test -q test_sample.py
+    collecting ... collected 1 items
     F
     ================================= FAILURES =================================
     _______________________________ test_answer ________________________________
@@ -55,6 +58,7 @@ Let's run this without supplying our new
 And now with supplying a command line option::
 
     $ py.test -q --cmdopt=type2
+    collecting ... collected 1 items
     F
     ================================= FAILURES =================================
     _______________________________ test_answer ________________________________
@@ -83,6 +87,8 @@ on real-life examples.
 generating parameters combinations, depending on command line
 ----------------------------------------------------------------------------
 
+.. regendoc:wipe
+
 Let's say we want to execute a test with different parameters
 and the parameter range shall be determined by a command
 line argument.  Let's first write a simple computation test::
@@ -112,13 +118,15 @@ Now we add a test configuration like thi
 This means that we only run 2 tests if we do not pass ``--all``::
 
     $ py.test -q test_compute.py
+    collecting ... collected 2 items
     ..
     2 passed in 0.01 seconds
 
 We run only two computations, so we see two dots.
 let's run the full monty::
 
-    $ py.test -q --all test_compute.py
+    $ py.test -q --all
+    collecting ... collected 5 items
     ....F
     ================================= FAILURES =================================
     _____________________________ test_compute[4] ______________________________
@@ -141,6 +149,8 @@ we'll get an error on the last one.
 control skipping of tests according to command line option
 --------------------------------------------------------------
 
+.. regendoc:wipe
+
 Here is a ``conftest.py`` file adding a ``--runslow`` command
 line option to control skipping of ``slow`` marked tests::
 
@@ -171,32 +181,33 @@ We can now write a test module like this
 
 and when running it will see a skipped "slow" test::
 
-    $ py.test test_module.py -rs    # "-rs" means report details on the little 's'
+    $ py.test -rs    # "-rs" means report details on the little 's'
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_module.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 2 items
     
     test_module.py .s
     ========================= short test summary info ==========================
-    SKIP [1] /tmp/doc-exec-104/conftest.py:9: need --runslow option to run
+    SKIP [1] /tmp/doc-exec-479/conftest.py:9: need --runslow option to run
     
     =================== 1 passed, 1 skipped in 0.02 seconds ====================
 
 Or run it including the ``slow`` marked test::
 
-    $ py.test test_module.py --runslow
+    $ py.test --runslow
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_module.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 2 items
     
     test_module.py ..
     
     ========================= 2 passed in 0.01 seconds =========================
 
-
 writing well integrated assertion helpers
 --------------------------------------------------
 
+.. regendoc:wipe
+
 If you have a test helper function called from a test you can
 use the ``pytest.fail`` marker to fail a test with a certain message.
 The test support function will not show up in the traceback if you
@@ -218,7 +229,8 @@ of tracebacks: the ``checkconfig`` funct
 unless the ``--fulltrace`` command line option is specified.
 Let's run our little function::
 
-    $ py.test -q
+    $ py.test -q test_checkconfig.py
+    collecting ... collected 1 items
     F
     ================================= FAILURES =================================
     ______________________________ test_something ______________________________
@@ -230,16 +242,17 @@ Let's run our little function::
     test_checkconfig.py:8: Failed
     1 failed in 0.02 seconds
 
-
 Detect if running from within a py.test run
 --------------------------------------------------------------
 
+.. regendoc:wipe
+
 Usually it is a bad idea to make application code
 behave differently if called from a test.  But if you
 absolutely must find out if your application code is
 running from a test you can do something like this::
 
-    # content of conftest.py in your testing directory
+    # content of conftest.py
 
     def pytest_configure(config):
         import sys

--- a/doc/monkeypatch.txt
+++ b/doc/monkeypatch.txt
@@ -39,8 +39,8 @@ will be undone.
 .. background check:
    $ py.test
    =========================== test session starts ============================
-   platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-   test path 1: /tmp/doc-exec-75
+   platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+   collecting ... collected 0 items
    
    =============================  in 0.00 seconds =============================
 

--- a/doc/doctest.txt
+++ b/doc/doctest.txt
@@ -15,11 +15,11 @@ python test modules)::
     py.test --doctest-modules
 
 You can make these changes permanent in your project by
-putting them into a conftest.py file like this::
+putting them into a pytest.ini file like this::
 
-    # content of conftest.py
-    option_doctestmodules = True
-    option_doctestglob = "*.rst"
+    # content of pytest.ini
+    [pytest]
+    addopts = --doctest-modules
 
 If you then have a text file like this::
 
@@ -35,7 +35,7 @@ and another like this::
     # content of mymodule.py
     def something():
         """ a doctest in a docstring
-        >>> something()  
+        >>> something()
         42
         """
         return 42
@@ -44,7 +44,9 @@ then you can just invoke ``py.test`` wit
 
     $ py.test
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: /tmp/doc-exec-66
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
-    =============================  in 0.00 seconds =============================
+    mymodule.py .
+    
+    ========================= 1 passed in 0.03 seconds =========================

--- a/doc/funcargs.txt
+++ b/doc/funcargs.txt
@@ -45,8 +45,8 @@ Running the test looks like this::
 
     $ py.test test_simplefactory.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_simplefactory.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
     test_simplefactory.py F
     
@@ -150,8 +150,8 @@ Running this::
 
     $ py.test test_example.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_example.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 10 items
     
     test_example.py .........F
     
@@ -188,8 +188,8 @@ If you want to select only the run with 
 
     $ py.test -v -k 7 test_example.py  # or -k test_func[7]
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30 -- /home/hpk/venv/0/bin/python
-    test path 1: test_example.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0 -- /home/hpk/venv/0/bin/python
+    collecting ... collected 10 items
     
     test_example.py:6: test_func[7] PASSED
     

--- a/doc/builtin.txt
+++ b/doc/builtin.txt
@@ -30,25 +30,25 @@ You can ask for available builtin or pro
         captures writes to sys.stdout/sys.stderr and makes
         them available successively via a ``capsys.readouterr()`` method
         which returns a ``(out, err)`` tuple of captured snapshot strings.
-
+        
     capfd
         captures writes to file descriptors 1 and 2 and makes
         snapshotted ``(out, err)`` string tuples available
         via the ``capsys.readouterr()`` method.  If the underlying
         platform does not have ``os.dup`` (e.g. Jython) tests using
         this funcarg will automatically skip.
-
+        
     tmpdir
         return a temporary directory path object
         unique to each test function invocation,
         created as a sub directory of the base temporary
         directory.  The returned object is a `py.path.local`_
         path object.
-
+        
     monkeypatch
         The returned ``monkeypatch`` funcarg provides these
         helper methods to modify objects, dictionaries or os.environ::
-
+        
         monkeypatch.setattr(obj, name, value, raising=True)
         monkeypatch.delattr(obj, name, raising=True)
         monkeypatch.setitem(mapping, name, value)
@@ -56,15 +56,15 @@ You can ask for available builtin or pro
         monkeypatch.setenv(name, value, prepend=False)
         monkeypatch.delenv(name, value, raising=True)
         monkeypatch.syspath_prepend(path)
-
+        
         All modifications will be undone when the requesting
         test function finished its execution.  The ``raising``
         parameter determines if a KeyError or AttributeError
         will be raised if the set/deletion operation has no target.
-
+        
     recwarn
         Return a WarningsRecorder instance that provides these methods:
-
+        
         * ``pop(category=None)``: return last warning matching the category.
         * ``clear()``: clear list of warnings
-
+        

--- a/doc/mark.txt
+++ b/doc/mark.txt
@@ -88,8 +88,8 @@ You can use the ``-k`` command line opti
 
     $ py.test -k webtest  # running with the above defined examples yields
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: /tmp/doc-exec-74
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 4 items
     
     test_mark.py ..
     test_mark_classlevel.py ..
@@ -100,8 +100,8 @@ And you can also run all tests except th
 
     $ py.test -k-webtest
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: /tmp/doc-exec-74
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 4 items
     
     ===================== 4 tests deselected by '-webtest' =====================
     ======================= 4 deselected in 0.01 seconds =======================
@@ -110,8 +110,8 @@ Or to only select the class::
 
     $ py.test -kTestClass
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: /tmp/doc-exec-74
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 4 items
     
     test_mark_classlevel.py ..
     

--- a/doc/example/parametrize.txt
+++ b/doc/example/parametrize.txt
@@ -41,11 +41,12 @@ Running it means we are two tests for ea
 the respective settings::
 
     $ py.test -q
+    collecting ... collected 4 items
     F..F
     ================================= FAILURES =================================
     _________________________ TestClass.test_equals[0] _________________________
     
-    self = <test_parametrize.TestClass instance at 0x128a638>, a = 1, b = 2
+    self = <test_parametrize.TestClass instance at 0x26ef2d8>, a = 1, b = 2
     
         def test_equals(self, a, b):
     >       assert a == b
@@ -54,7 +55,7 @@ the respective settings::
     test_parametrize.py:17: AssertionError
     ______________________ TestClass.test_zerodivision[1] ______________________
     
-    self = <test_parametrize.TestClass instance at 0x1296440>, a = 3, b = 2
+    self = <test_parametrize.TestClass instance at 0x26fa758>, a = 3, b = 2
     
         def test_zerodivision(self, a, b):
     >       pytest.raises(ZeroDivisionError, "a/b")
@@ -97,11 +98,12 @@ for parametrizing test methods::
 Running it gives similar results as before::
 
     $ py.test -q test_parametrize2.py
+    collecting ... collected 4 items
     F..F
     ================================= FAILURES =================================
     _________________________ TestClass.test_equals[0] _________________________
     
-    self = <test_parametrize2.TestClass instance at 0x1dbcc68>, a = 1, b = 2
+    self = <test_parametrize2.TestClass instance at 0x1e5e638>, a = 1, b = 2
     
         @params([dict(a=1, b=2), dict(a=3, b=3), ])
         def test_equals(self, a, b):
@@ -111,7 +113,7 @@ Running it gives similar results as befo
     test_parametrize2.py:19: AssertionError
     ______________________ TestClass.test_zerodivision[1] ______________________
     
-    self = <test_parametrize2.TestClass instance at 0x1dd0488>, a = 3, b = 2
+    self = <test_parametrize2.TestClass instance at 0x1e6f560>, a = 3, b = 2
     
         @params([dict(a=1, b=0), dict(a=3, b=2)])
         def test_zerodivision(self, a, b):
@@ -138,5 +140,6 @@ with different sets of arguments for its
 Running it (with Python-2.4 through to Python2.7 installed)::
 
    . $ py.test -q multipython.py
+   collecting ... collected 75 items
    ....s....s....s....ssssss....s....s....s....ssssss....s....s....s....ssssss
-   48 passed, 27 skipped in 2.55 seconds
+   48 passed, 27 skipped in 2.74 seconds

--- a/doc/getting-started.txt
+++ b/doc/getting-started.txt
@@ -16,7 +16,7 @@ Installation options::
 To check your installation has installed the correct version::
 
     $ py.test --version
-    This is py.test version 2.0.0.dev30, imported from /home/hpk/p/pytest/pytest.py
+    This is py.test version 2.0.0, imported from /home/hpk/p/pytest/pytest.pyc
 
 If you get an error checkout :ref:`installation issues`.
 
@@ -38,19 +38,19 @@ That's it. You can execute the test func
 
     $ py.test
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: /tmp/doc-exec-70
-
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
+    
     test_sample.py F
-
+    
     ================================= FAILURES =================================
     _______________________________ test_answer ________________________________
-
+    
         def test_answer():
     >       assert func(3) == 5
     E       assert 4 == 5
     E        +  where 4 = func(3)
-
+    
     test_sample.py:5: AssertionError
     ========================= 1 failed in 0.02 seconds =========================
 
@@ -58,9 +58,10 @@ py.test found the ``test_answer`` functi
 
 .. note::
 
-    You can simply use the ``assert`` statement for coding expectations because
-    intermediate values will be presented to you. This is arguably easier than
-    learning all the `the JUnit legacy methods`_.
+    You can simply use the ``assert`` statement for asserting
+    expectations because intermediate values will be presented to you.
+    This is arguably easier than learning all the `the JUnit legacy
+    methods`_.
 
     However, there remains one caveat to using simple asserts: your
     assertion expression should better be side-effect free.  Because
@@ -94,6 +95,7 @@ use the ``raises`` helper::
 Running it with, this time in "quiet" reporting mode::
 
     $ py.test -q test_sysexit.py
+    collecting ... collected 1 items
     .
     1 passed in 0.00 seconds
 
@@ -121,17 +123,18 @@ There is no need to subclass anything.  
 run the module by passing its filename::
 
     $ py.test -q test_class.py
+    collecting ... collected 2 items
     .F
     ================================= FAILURES =================================
     ____________________________ TestClass.test_two ____________________________
-
-    self = <test_class.TestClass instance at 0x288fc20>
-
+    
+    self = <test_class.TestClass instance at 0x11fa320>
+    
         def test_two(self):
             x = "hello"
     >       assert hasattr(x, 'check')
     E       assert hasattr('hello', 'check')
-
+    
     test_class.py:8: AssertionError
     1 failed, 1 passed in 0.02 seconds
 
@@ -157,21 +160,22 @@ py.test will lookup and call a factory t
 before performing the test function call.  Let's just run it::
 
     $ py.test -q test_tmpdir.py
+    collecting ... collected 1 items
     F
     ================================= FAILURES =================================
     _____________________________ test_needsfiles ______________________________
-
-    tmpdir = local('/tmp/pytest-122/test_needsfiles0')
-
+    
+    tmpdir = local('/tmp/pytest-7/test_needsfiles0')
+    
         def test_needsfiles(tmpdir):
             print tmpdir
     >       assert 0
     E       assert 0
-
+    
     test_tmpdir.py:3: AssertionError
     ----------------------------- Captured stdout ------------------------------
-    /tmp/pytest-122/test_needsfiles0
-    1 failed in 0.05 seconds
+    /tmp/pytest-7/test_needsfiles0
+    1 failed in 0.04 seconds
 
 Before the test runs, a unique-per-test-invocation temporary directory
 was created.  More info at :ref:`tmpdir handling`.

--- a/doc/assert.txt
+++ b/doc/assert.txt
@@ -23,21 +23,21 @@ assertion fails you will see the value o
 
     $ py.test test_assert1.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_assert1.py
-
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
+    
     test_assert1.py F
-
+    
     ================================= FAILURES =================================
     ______________________________ test_function _______________________________
-
+    
         def test_function():
     >       assert f() == 4
     E       assert 3 == 4
     E        +  where 3 = f()
-
+    
     test_assert1.py:5: AssertionError
-    ========================= 1 failed in 0.03 seconds =========================
+    ========================= 1 failed in 0.02 seconds =========================
 
 Reporting details about the failing assertion is achieved by re-evaluating
 the assert expression and recording intermediate values.
@@ -105,14 +105,14 @@ if you run this module::
 
     $ py.test test_assert2.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_assert2.py
-
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
+    
     test_assert2.py F
-
+    
     ================================= FAILURES =================================
     ___________________________ test_set_comparison ____________________________
-
+    
         def test_set_comparison():
             set1 = set("1308")
             set2 = set("8035")
@@ -122,7 +122,7 @@ if you run this module::
     E         '1'
     E         Extra items in the right set:
     E         '5'
-
+    
     test_assert2.py:5: AssertionError
     ========================= 1 failed in 0.02 seconds =========================
 

--- a/doc/example/mysetup.txt
+++ b/doc/example/mysetup.txt
@@ -49,15 +49,15 @@ You can now run the test::
 
     $ py.test test_sample.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_sample.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
     test_sample.py F
     
     ================================= FAILURES =================================
     _______________________________ test_answer ________________________________
     
-    mysetup = <conftest.MySetup instance at 0x16f5998>
+    mysetup = <conftest.MySetup instance at 0x2c88128>
     
         def test_answer(mysetup):
             app = mysetup.myapp()
@@ -122,12 +122,12 @@ Running it yields::
 
     $ py.test test_ssh.py -rs
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_ssh.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
     test_ssh.py s
     ========================= short test summary info ==========================
-    SKIP [1] /tmp/doc-exec-107/conftest.py:22: specify ssh host with --ssh
+    SKIP [1] /tmp/doc-exec-474/conftest.py:22: specify ssh host with --ssh
     
     ======================== 1 skipped in 0.02 seconds =========================
 

--- a/doc/skipping.txt
+++ b/doc/skipping.txt
@@ -121,14 +121,14 @@ Running it with the report-on-xfail opti
 
     example $ py.test -rx xfail_demo.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev31
-    test path 1: xfail_demo.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 5 items
     
     xfail_demo.py xxxxx
     ========================= short test summary info ==========================
     XFAIL xfail_demo.py::test_hello
     XFAIL xfail_demo.py::test_hello2
-      reason: [NOTRUN]
+      reason: [NOTRUN] 
     XFAIL xfail_demo.py::test_hello3
       condition: hasattr(os, 'sep')
     XFAIL xfail_demo.py::test_hello4

--- a/doc/example/nonpython.txt
+++ b/doc/example/nonpython.txt
@@ -27,8 +27,8 @@ now execute the test specification::
 
     nonpython $ py.test test_simple.yml
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_simple.yml
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 2 items
     
     test_simple.yml .F
     
@@ -37,9 +37,7 @@ now execute the test specification::
     usecase execution failed
        spec failed: 'some': 'other'
        no further details known at this point.
-    ========================= short test summary info ==========================
-    FAIL test_simple.yml::hello
-    ==================== 1 failed, 1 passed in 0.06 seconds ====================
+    ==================== 1 failed, 1 passed in 0.03 seconds ====================
 
 You get one dot for the passing ``sub1: sub1`` check and one failure.
 Obviously in the above ``conftest.py`` you'll want to implement a more
@@ -58,8 +56,8 @@ reporting in ``verbose`` mode::
 
     nonpython $ py.test -v
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30 -- /home/hpk/venv/0/bin/python
-    test path 1: /home/hpk/p/pytest/doc/example/nonpython
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0 -- /home/hpk/venv/0/bin/python
+    collecting ... collected 2 items
     
     test_simple.yml:1: usecase: ok PASSED
     test_simple.yml:1: usecase: hello FAILED
@@ -69,9 +67,7 @@ reporting in ``verbose`` mode::
     usecase execution failed
        spec failed: 'some': 'other'
        no further details known at this point.
-    ========================= short test summary info ==========================
-    FAIL test_simple.yml::hello
-    ==================== 1 failed, 1 passed in 0.06 seconds ====================
+    ==================== 1 failed, 1 passed in 0.03 seconds ====================
 
 While developing your custom test collection and execution it's also
 interesting to just look at the collection tree::

--- a/doc/example/reportingdemo.txt
+++ b/doc/example/reportingdemo.txt
@@ -13,9 +13,8 @@ get on the terminal - we are working on 
 
     assertion $ py.test failure_demo.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev38
-    collecting ...
-    collected 35 items
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 35 items
     
     failure_demo.py FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF
     
@@ -31,7 +30,7 @@ get on the terminal - we are working on 
     failure_demo.py:15: AssertionError
     _________________________ TestFailing.test_simple __________________________
     
-    self = <failure_demo.TestFailing object at 0x2c9da90>
+    self = <failure_demo.TestFailing object at 0x17d8750>
     
         def test_simple(self):
             def f():
@@ -41,21 +40,21 @@ get on the terminal - we are working on 
         
     >       assert f() == g()
     E       assert 42 == 43
-    E        +  where 42 = <function f at 0x2c447d0>()
-    E        +  and   43 = <function g at 0x2c44cf8>()
+    E        +  where 42 = <function f at 0x17e2488>()
+    E        +  and   43 = <function g at 0x17e2140>()
     
     failure_demo.py:28: AssertionError
     ____________________ TestFailing.test_simple_multiline _____________________
     
-    self = <failure_demo.TestFailing object at 0x2c9dc90>
+    self = <failure_demo.TestFailing object at 0x17d4390>
     
         def test_simple_multiline(self):
             otherfunc_multi(
                       42,
     >                 6*9)
     
-    failure_demo.py:33:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:33: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
     a = 42, b = 54
     
@@ -67,19 +66,19 @@ get on the terminal - we are working on 
     failure_demo.py:12: AssertionError
     ___________________________ TestFailing.test_not ___________________________
     
-    self = <failure_demo.TestFailing object at 0x2c93f10>
+    self = <failure_demo.TestFailing object at 0x17d8cd0>
     
         def test_not(self):
             def f():
                 return 42
     >       assert not f()
     E       assert not 42
-    E        +  where 42 = <function f at 0x2ca1050>()
+    E        +  where 42 = <function f at 0x17e25f0>()
     
     failure_demo.py:38: AssertionError
     _________________ TestSpecialisedExplanations.test_eq_text _________________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2c9d9d0>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17e33d0>
     
         def test_eq_text(self):
     >       assert 'spam' == 'eggs'
@@ -90,7 +89,7 @@ get on the terminal - we are working on 
     failure_demo.py:42: AssertionError
     _____________ TestSpecialisedExplanations.test_eq_similar_text _____________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2a04e90>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17ee990>
     
         def test_eq_similar_text(self):
     >       assert 'foo 1 bar' == 'foo 2 bar'
@@ -103,7 +102,7 @@ get on the terminal - we are working on 
     failure_demo.py:45: AssertionError
     ____________ TestSpecialisedExplanations.test_eq_multiline_text ____________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2c9d710>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17d8d10>
     
         def test_eq_multiline_text(self):
     >       assert 'foo\nspam\nbar' == 'foo\neggs\nbar'
@@ -116,7 +115,7 @@ get on the terminal - we are working on 
     failure_demo.py:48: AssertionError
     ______________ TestSpecialisedExplanations.test_eq_long_text _______________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2c9db10>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17e3990>
     
         def test_eq_long_text(self):
             a = '1'*100 + 'a' + '2'*100
@@ -133,7 +132,7 @@ get on the terminal - we are working on 
     failure_demo.py:53: AssertionError
     _________ TestSpecialisedExplanations.test_eq_long_text_multiline __________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2caf950>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17eee10>
     
         def test_eq_long_text_multiline(self):
             a = '1\n'*100 + 'a' + '2\n'*100
@@ -157,7 +156,7 @@ get on the terminal - we are working on 
     failure_demo.py:58: AssertionError
     _________________ TestSpecialisedExplanations.test_eq_list _________________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2caf590>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17e3490>
     
         def test_eq_list(self):
     >       assert [0, 1, 2] == [0, 1, 3]
@@ -167,7 +166,7 @@ get on the terminal - we are working on 
     failure_demo.py:61: AssertionError
     ______________ TestSpecialisedExplanations.test_eq_list_long _______________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2c9e310>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17e35d0>
     
         def test_eq_list_long(self):
             a = [0]*100 + [1] + [3]*100
@@ -179,7 +178,7 @@ get on the terminal - we are working on 
     failure_demo.py:66: AssertionError
     _________________ TestSpecialisedExplanations.test_eq_dict _________________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2c9dc50>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17eef10>
     
         def test_eq_dict(self):
     >       assert {'a': 0, 'b': 1} == {'a': 0, 'b': 2}
@@ -192,7 +191,7 @@ get on the terminal - we are working on 
     failure_demo.py:69: AssertionError
     _________________ TestSpecialisedExplanations.test_eq_set __________________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2cafc10>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17f4d10>
     
         def test_eq_set(self):
     >       assert set([0, 10, 11, 12]) == set([0, 20, 21])
@@ -208,7 +207,7 @@ get on the terminal - we are working on 
     failure_demo.py:72: AssertionError
     _____________ TestSpecialisedExplanations.test_eq_longer_list ______________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2cba890>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x17f4e10>
     
         def test_eq_longer_list(self):
     >       assert [1,2] == [1,2,3]
@@ -218,7 +217,7 @@ get on the terminal - we are working on 
     failure_demo.py:75: AssertionError
     _________________ TestSpecialisedExplanations.test_in_list _________________
     
-    self = <failure_demo.TestSpecialisedExplanations object at 0x2cba6d0>
+    self = <failure_demo.TestSpecialisedExplanations object at 0x1801690>
     
         def test_in_list(self):
     >       assert 1 in [0, 2, 3, 4, 5]
@@ -233,7 +232,7 @@ get on the terminal - we are working on 
             i = Foo()
     >       assert i.b == 2
     E       assert 1 == 2
-    E        +  where 1 = <failure_demo.Foo object at 0x2c9d750>.b
+    E        +  where 1 = <failure_demo.Foo object at 0x17f48d0>.b
     
     failure_demo.py:85: AssertionError
     _________________________ test_attribute_instance __________________________
@@ -243,8 +242,8 @@ get on the terminal - we are working on 
                 b = 1
     >       assert Foo().b == 2
     E       assert 1 == 2
-    E        +  where 1 = <failure_demo.Foo object at 0x2cafdd0>.b
-    E        +    where <failure_demo.Foo object at 0x2cafdd0> = <class 'failure_demo.Foo'>()
+    E        +  where 1 = <failure_demo.Foo object at 0x17f4390>.b
+    E        +    where <failure_demo.Foo object at 0x17f4390> = <class 'failure_demo.Foo'>()
     
     failure_demo.py:91: AssertionError
     __________________________ test_attribute_failure __________________________
@@ -257,10 +256,10 @@ get on the terminal - we are working on 
             i = Foo()
     >       assert i.b == 2
     
-    failure_demo.py:100:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:100: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
-    self = <failure_demo.Foo object at 0x2cba790>
+    self = <failure_demo.Foo object at 0x17ee790>
     
         def _get_b(self):
     >       raise Exception('Failed to get attrib')
@@ -276,22 +275,22 @@ get on the terminal - we are working on 
                 b = 2
     >       assert Foo().b == Bar().b
     E       assert 1 == 2
-    E        +  where 1 = <failure_demo.Foo object at 0x2cba210>.b
-    E        +    where <failure_demo.Foo object at 0x2cba210> = <class 'failure_demo.Foo'>()
-    E        +  and   2 = <failure_demo.Bar object at 0x2cba850>.b
-    E        +    where <failure_demo.Bar object at 0x2cba850> = <class 'failure_demo.Bar'>()
+    E        +  where 1 = <failure_demo.Foo object at 0x17eed10>.b
+    E        +    where <failure_demo.Foo object at 0x17eed10> = <class 'failure_demo.Foo'>()
+    E        +  and   2 = <failure_demo.Bar object at 0x17eead0>.b
+    E        +    where <failure_demo.Bar object at 0x17eead0> = <class 'failure_demo.Bar'>()
     
     failure_demo.py:108: AssertionError
     __________________________ TestRaises.test_raises __________________________
     
-    self = <failure_demo.TestRaises instance at 0x2cc2560>
+    self = <failure_demo.TestRaises instance at 0x1808170>
     
         def test_raises(self):
             s = 'qwe'
     >       raises(TypeError, "int(s)")
     
-    failure_demo.py:117:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:117: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
     >   int(s)
     E   ValueError: invalid literal for int() with base 10: 'qwe'
@@ -299,7 +298,7 @@ get on the terminal - we are working on 
     <0-codegen /home/hpk/p/pytest/_pytest/python.py:819>:1: ValueError
     ______________________ TestRaises.test_raises_doesnt _______________________
     
-    self = <failure_demo.TestRaises instance at 0x2cb6bd8>
+    self = <failure_demo.TestRaises instance at 0x17fd908>
     
         def test_raises_doesnt(self):
     >       raises(IOError, "int('3')")
@@ -308,7 +307,7 @@ get on the terminal - we are working on 
     failure_demo.py:120: Failed
     __________________________ TestRaises.test_raise ___________________________
     
-    self = <failure_demo.TestRaises instance at 0x2cc4830>
+    self = <failure_demo.TestRaises instance at 0x1809440>
     
         def test_raise(self):
     >       raise ValueError("demo error")
@@ -317,7 +316,7 @@ get on the terminal - we are working on 
     failure_demo.py:123: ValueError
     ________________________ TestRaises.test_tupleerror ________________________
     
-    self = <failure_demo.TestRaises instance at 0x2cc5560>
+    self = <failure_demo.TestRaises instance at 0x1806170>
     
         def test_tupleerror(self):
     >       a,b = [1]
@@ -326,7 +325,7 @@ get on the terminal - we are working on 
     failure_demo.py:126: ValueError
     ______ TestRaises.test_reinterpret_fails_with_print_for_the_fun_of_it ______
     
-    self = <failure_demo.TestRaises instance at 0x2cc6248>
+    self = <failure_demo.TestRaises instance at 0x1806dd0>
     
         def test_reinterpret_fails_with_print_for_the_fun_of_it(self):
             l = [1,2,3]
@@ -339,7 +338,7 @@ get on the terminal - we are working on 
     l is [1, 2, 3]
     ________________________ TestRaises.test_some_error ________________________
     
-    self = <failure_demo.TestRaises instance at 0x2cc6f38>
+    self = <failure_demo.TestRaises instance at 0x180ab48>
     
         def test_some_error(self):
     >       if namenotexi:
@@ -357,8 +356,8 @@ get on the terminal - we are working on 
             py.std.sys.modules[name] = module
     >       module.foo()
     
-    failure_demo.py:149:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:149: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
         def foo():
     >    assert 1 == 0
@@ -367,7 +366,7 @@ get on the terminal - we are working on 
     <2-codegen 'abc-123' /home/hpk/p/pytest/doc/example/assertion/failure_demo.py:146>:2: AssertionError
     ____________________ TestMoreErrors.test_complex_error _____________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2cc4050>
+    self = <failure_demo.TestMoreErrors instance at 0x18081b8>
     
         def test_complex_error(self):
             def f():
@@ -376,16 +375,16 @@ get on the terminal - we are working on 
                 return 43
     >       somefunc(f(), g())
     
-    failure_demo.py:159:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:159: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
     x = 44, y = 43
     
         def somefunc(x,y):
     >       otherfunc(x,y)
     
-    failure_demo.py:8:
-    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
+    failure_demo.py:8: 
+    _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
     
     a = 44, b = 43
     
@@ -396,7 +395,7 @@ get on the terminal - we are working on 
     failure_demo.py:5: AssertionError
     ___________________ TestMoreErrors.test_z1_unpack_error ____________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2cc7ab8>
+    self = <failure_demo.TestMoreErrors instance at 0x180d6c8>
     
         def test_z1_unpack_error(self):
             l = []
@@ -406,7 +405,7 @@ get on the terminal - we are working on 
     failure_demo.py:163: ValueError
     ____________________ TestMoreErrors.test_z2_type_error _____________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2ccb8c0>
+    self = <failure_demo.TestMoreErrors instance at 0x18114d0>
     
         def test_z2_type_error(self):
             l = 3
@@ -416,19 +415,19 @@ get on the terminal - we are working on 
     failure_demo.py:167: TypeError
     ______________________ TestMoreErrors.test_startswith ______________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2ccd5f0>
+    self = <failure_demo.TestMoreErrors instance at 0x18d1200>
     
         def test_startswith(self):
             s = "123"
             g = "456"
     >       assert s.startswith(g)
-    E       assert <built-in method startswith of str object at 0x2c321b0>('456')
-    E        +  where <built-in method startswith of str object at 0x2c321b0> = '123'.startswith
+    E       assert <built-in method startswith of str object at 0x177e240>('456')
+    E        +  where <built-in method startswith of str object at 0x177e240> = '123'.startswith
     
     failure_demo.py:172: AssertionError
     __________________ TestMoreErrors.test_startswith_nested ___________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2ccbc20>
+    self = <failure_demo.TestMoreErrors instance at 0x1811908>
     
         def test_startswith_nested(self):
             def f():
@@ -436,15 +435,15 @@ get on the terminal - we are working on 
             def g():
                 return "456"
     >       assert f().startswith(g())
-    E       assert <built-in method startswith of str object at 0x2c321b0>('456')
-    E        +  where <built-in method startswith of str object at 0x2c321b0> = '123'.startswith
-    E        +    where '123' = <function f at 0x2c2d140>()
-    E        +  and   '456' = <function g at 0x2cb00c8>()
+    E       assert <built-in method startswith of str object at 0x177e240>('456')
+    E        +  where <built-in method startswith of str object at 0x177e240> = '123'.startswith
+    E        +    where '123' = <function f at 0x176f668>()
+    E        +  and   '456' = <function g at 0x1800578>()
     
     failure_demo.py:179: AssertionError
     _____________________ TestMoreErrors.test_global_func ______________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2cb69e0>
+    self = <failure_demo.TestMoreErrors instance at 0x17fd9e0>
     
         def test_global_func(self):
     >       assert isinstance(globf(42), float)
@@ -454,19 +453,19 @@ get on the terminal - we are working on 
     failure_demo.py:182: AssertionError
     _______________________ TestMoreErrors.test_instance _______________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2cc6440>
+    self = <failure_demo.TestMoreErrors instance at 0x1806b90>
     
         def test_instance(self):
             self.x = 6*7
     >       assert self.x != 42
     E       assert 42 != 42
     E        +  where 42 = 42
-    E        +    where 42 = <failure_demo.TestMoreErrors instance at 0x2cc6440>.x
+    E        +    where 42 = <failure_demo.TestMoreErrors instance at 0x1806b90>.x
     
     failure_demo.py:186: AssertionError
     _______________________ TestMoreErrors.test_compare ________________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2dcc200>
+    self = <failure_demo.TestMoreErrors instance at 0x18d1f38>
     
         def test_compare(self):
     >       assert globf(10) < 5
@@ -476,7 +475,7 @@ get on the terminal - we are working on 
     failure_demo.py:189: AssertionError
     _____________________ TestMoreErrors.test_try_finally ______________________
     
-    self = <failure_demo.TestMoreErrors instance at 0x2dce0e0>
+    self = <failure_demo.TestMoreErrors instance at 0x18d4cf8>
     
         def test_try_finally(self):
             x = 1

--- a/doc/tmpdir.txt
+++ b/doc/tmpdir.txt
@@ -28,15 +28,15 @@ Running this would result in a passed te
 
     $ py.test test_tmpdir.py
     =========================== test session starts ============================
-    platform linux2 -- Python 2.6.5 -- pytest-2.0.0.dev30
-    test path 1: test_tmpdir.py
+    platform linux2 -- Python 2.6.5 -- pytest-2.0.0
+    collecting ... collected 1 items
     
     test_tmpdir.py F
     
     ================================= FAILURES =================================
     _____________________________ test_create_file _____________________________
     
-    tmpdir = local('/tmp/pytest-123/test_create_file0')
+    tmpdir = local('/tmp/pytest-8/test_create_file0')
     
         def test_create_file(tmpdir):
             p = tmpdir.mkdir("sub").join("hello.txt")
@@ -47,7 +47,7 @@ Running this would result in a passed te
     E       assert 0
     
     test_tmpdir.py:7: AssertionError
-    ========================= 1 failed in 0.02 seconds =========================
+    ========================= 1 failed in 0.04 seconds =========================
 
 .. _`base temporary directory`:



More information about the pytest-commit mailing list