TDD - Test Driven Development

For this assessment, We'll be using the concept of Test Driven Development (TDD) to created a modified version of the echo command line tool. This tool is built into every shell, and functions very much like print() does in Python. For this assessment we'll be creating our own version of echo which accepts a few additional parameters on the command line.

example output

In doing so, you'll be demonstrating a basic understanding of the following:

Getting Started

TDD starts with setting up a test harness, and writing the tests FIRST, before writing application code. A test harness (or fixture) is a completely separate module designed to test your application code. One analogy to think of is how rocket or jet engines are tested: A new engine is mounted into a test fixture, which supplies inputs like fuel and control signals. It measures the performance of the engine in a safe environment, before the engine gets deployed into the real world.

Think of the test cases that you will need for the echo application that you will build. You will need to check each command line option, in various forms, to make sure they are working. You'll also need to test them together, to make sure order of precedence is kept (e.g. should TitleCase be applied before or after Lowercase if they are both present?)

For a small project like this, it's sufficient to have a single test module named test_echo.py. In the beginning, all tests should fail (of course, because you haven't written anything!) However, you are proving out the basic execution paths and setup of your program.

Keep in Mind:

  • When writing a test, write the assertion statement first, and work backwards.
  • Before writing the application code, run the test to make sure it fails in the way you think it should.
  • Write meaningful, short tests that are self-explanatory. Keep the "Reader" of your code in mind. It may not be you. Is your intention clear?
  • Always keep test code separate from application code.
  • Each test should run indepedently. No cross-dependencies between tests.
  • Make frequent, small commits after each test is passing.

Pay special attention to the name of the test module: test_echo.py. When writing test modules, start each module filename with the prefix test*. Many testing frameworks are set up to auto-discover test modules that adhere to this naming convention. Auto-discovery is important-- in a Continuous Integration (CI) environment, your tests will be discovered and run when you attempt to push changes to the repository.

The two files you will be editing are test_echo.py and echo.py. Don't worry about the files inside the soln directory, those are encrypted and used only by your instructor team. When done, you should have a project directory that looks something like this:

.
├── README.md
├── USAGE
├── echo.py
├── screenshots
│   └── result.gif
├── soln
│   ├── __init__.py
│   ├── .gitattributes
│   ├── echo.py
│   └── test_echo.py
└── tests
    └── test_echo.py

Even before you write a single test, you might find it useful to try out your test harness. Note the use of the rerun helper utility. You can pip install this useful tool that watches a directory for file changes, and re-runs the command each time it detects a newly-saved file. From the command line in terminal:

% pip install rerun
% rerun "python -m unittest discover tests"

NOTE: when running tests from the command line, your current working directory must be the project root directory. If you run the tests from within the tests folder, you will see an error like this

Running Tests from VSCode

If you don't want your tests manually from the command line, VSCode has a nifty plugin feature that runs your tests directly from your code window. Access this feature through the Test Tube icon on the left ribbon bar. This VSCode plugin also supports built-in unit test discovery, but you must manually enable it.

Acceptance Criteria

Step 1: Test the Help/Usage

When the user provides invalid options or supplies the -h/--help flag, the program should print the following usage message:

usage: echo.py [-h] [-u] [-l] [-t] text

Perform transformation on input text.

positional arguments:
    text         text to be manipulated

optional arguments:
    -h, --help   show this help message and exit
    -u, --upper  convert text to uppercase
    -l, --lower  convert text to lowercase
    -t, --title  convert text to titlecase

This text is actually composed and printed by the Argparse module. Argparse will assemble the various description and help messages that you have coded into your parser object, and print them all in a coherent "USAGE" message. The test_echo.py file also contains a helper method that will allow you to capture the printed output of echo.py, by running it as a subprocess.

# Students should use this function in their tests
def run_capture(pyfile, args=()):
    """
    Runs a python program in a separate process,
    returns the output lines as a list.
    """
    cmd = ["python", pyfile]
    cmd.extend(args)
    try:
        result = subprocess.run(
            cmd,
            stdout=subprocess.PIPE,
            stderr=subprocess.STDOUT,
            check=True
        )
        output = result.stdout.decode()
    except subprocess.CalledProcessError as err:
        output = err.stdout.decode()
    assert output, "Nothing was printed!"
    return output.splitlines()

This code is provided for you inside the test_echo.py file. Study this function and understand how it works.

Step 2: Test the -u/--upper option

Write a unit test that asserts that upper gets stored inside of the namespace returned from parser.parse_args() when either "-u" or "--upper" arguments are given on the command line.

It should also test that "hello" gets turned into "HELLO" when the program is run.

Step 3: Test the -l/--lower option

Write a unit test that asserts that lower gets stored inside of the namespace returned from parser.parse_args() when either "-l" or "--lower" arguments are given on the command line.

It should also test that "Hello" gets turned into "hello" when the program is run.

Step 4: Test the -t/--title option

Write a unit test that asserts that title gets stored inside of the namespace returned from parser.parse_args() when either "-t" or "--title" arguments are given on the command line.

It should also test that "hello" gets turned into "Hello" when the program is run.

Step 6: Test for when all option flags are provided

When a user provides all three option flags (-lut), they should be applied in the order listed in the helpful usage message that Argparse constructs from the argument definitions. Here are a few examples:

% python echo.py -tul "heLLo!"
Hello!
% python echo.py -ul "heLLo!"
hello!

Note that the order that the options are provided doesn't matter, e.g. '-tul' and '-utl' and '-lut' are all equivalent inputs to Argparse. Only the final text transform result should be printed.

Step 7: Test for no options flags

Write a unit test that asserts that when no options flags are given, the program prints the unaltered input text.

Step 8: Implement echo.py

This step can be done in several ways.

  • You can try to implement ALL the test cases in advance and then code the application to pass all the tests. You'll probably forget some along the way and come back later to add them.
  • A better approach might be to write a small failing test, then write some application code to make it pass. Then write another failing test, and write more app code to make it pass. You get the idea, to alternate between test dev and app dev as you code.

Structuring your echo.py application

Remember to separate functionality in your echo.py application. Notice that many of the tests above are checking to see if the argument parser has done its job correctly by parsing out an option from the command line, and making it available in parser output (the Namespace, or parsed args dict).
Therefore, it makes sense to have a function in echo.py whose sole purpose is to deliver back a parser object, that can be stored in your TestEcho class and invoked by calling its parse_args() method with various argument lists. Such a function might be named create_parser().

You may also benefit from having a separate main() function in your echo.py appliction. A main() function can be invoked from the command line directly as part of the application, but it can also be directly imported by your test program so you can test it all different ways. The starter code for these two functions is provided in echo.py.

Submitting your work

To submit your solution for grading, you will need to create a github Pull Request (PR). Refer to the PR Workflow article in your course content for details.

Assignment FAQ

  • "Q: How do I write tests for the namespace?" A: A Namespace is the return value type that you get when you call parser.parse_args(). You want to write tests that verify if the parser is doing its job of converting an argument list into a Namespace:
    • Is an instance of a Parser object type actually returned after calling the test module's create_parser function?
    • After running parser.parse_args(), does the returned Namespace contain the attributes you expected?
    • Does it contain extra attributes you were not expecting?
    • What are the values contained in the attributes? Are they correct and expected, according to the argument list you provided as input?