Skip to main content

This module can be used to handle comma-separated values (CSV) files and do lightweight processing of their data with support for row and column filtering

Project description

PyPI version License Python versions supported Format

https://travis-ci.org/pmacosta/pcsv.svg?branch=master Windows continuous integration Continuous integration coverage Documentation status

Description

This module can be used to handle comma-separated values (CSV) files and do lightweight processing of their data with support for row and column filtering. In addition to basic read, write and data replacement, files can be concatenated, merged, and sorted

Examples

Read/write

# pcsv_example_1.py
import pmisc, pcsv

def main():
    with pmisc.TmpFile() as fname:
        ref_data = [
            ['Item', 'Cost'],
            [1, 9.99],
            [2, 10000],
            [3, 0.10]
        ]
        # Write reference data to a file
        pcsv.write(fname, ref_data, append=False)
        # Read the data back
        obj = pcsv.CsvFile(fname)
    # After the object creation the I/O is done,
    # can safely remove file (exit context manager)
    # Check that data read is correct
    assert obj.header() == ref_data[0]
    assert obj.data() == ref_data[1:]
    # Add a simple row filter, only look at rows that have
    # values 1 and 3 in the "Items" column
    obj.rfilter = {'Item':[1, 3]}
    assert obj.data(filtered=True) == [ref_data[1], ref_data[3]]

if __name__ == '__main__':
    main()

Replace data

# pcsv_example_2.py
import pmisc, pcsv

def main():
    ctx = pmisc.TmpFile
    with ctx() as fname1:
        with ctx() as fname2:
            with ctx() as ofname:
                # Create first (input) data file
                input_data = [
                    ['Item', 'Cost'],
                    [1, 9.99],
                    [2, 10000],
                    [3, 0.10]
                ]
                pcsv.write(fname1, input_data, append=False)
                # Create second (replacement) data file
                replacement_data = [
                    ['Staff', 'Rate', 'Days'],
                    ['Joe', 10, 'Sunday'],
                    ['Sue', 20, 'Thursday'],
                    ['Pat', 15, 'Tuesday']
                ]
                pcsv.write(fname2, replacement_data, append=False)
                # Replace "Cost" column of input file with "Rate" column
                # of replacement file for "Items" 2 and 3 with "Staff" data
                # from Joe and Pat. Save resulting data to another file
                pcsv.replace(
                    fname1=fname1,
                    dfilter1=('Cost', {'Item':[1, 3]}),
                    fname2=fname2,
                    dfilter2=('Rate', {'Staff':['Joe', 'Pat']}),
                    ofname=ofname
                )
                # Verify that resulting file is correct
                ref_data = [
                    ['Item', 'Cost'],
                    [1, 10],
                    [2, 10000],
                    [3, 15]
                ]
                obj = pcsv.CsvFile(ofname)
                assert obj.header() == ref_data[0]
                assert obj.data() == ref_data[1:]

if __name__ == '__main__':
    main()

Concatenate two files

# pcsv_example_3.py
import pmisc, pcsv

def main():
    ctx = pmisc.TmpFile
    with ctx() as fname1:
        with ctx() as fname2:
            with ctx() as ofname:
                # Create first data file
                data1 = [
                    [1, 9.99],
                    [2, 10000],
                    [3, 0.10]
                ]
                pcsv.write(fname1, data1, append=False)
                # Create second data file
                data2 = [
                    ['Joe', 10, 'Sunday'],
                    ['Sue', 20, 'Thursday'],
                    ['Pat', 15, 'Tuesday']
                ]
                pcsv.write(fname2, data2, append=False)
                # Concatenate file1 and file2. Filter out
                # second column of file2
                pcsv.concatenate(
                    fname1=fname1,
                    fname2=fname2,
                    has_header1=False,
                    has_header2=False,
                    dfilter2=[0, 2],
                    ofname=ofname,
                    ocols=['D1', 'D2']
                )
                # Verify that resulting file is correct
                ref_data = [
                    ['D1', 'D2'],
                    [1, 9.99],
                    [2, 10000],
                    [3, 0.10],
                    ['Joe', 'Sunday'],
                    ['Sue', 'Thursday'],
                    ['Pat', 'Tuesday']
                ]
                obj = pcsv.CsvFile(ofname)
                assert obj.header() == ref_data[0]
                assert obj.data() == ref_data[1:]

if __name__ == '__main__':
    main()

Merge two files

# pcsv_example_4.py
import pmisc, pcsv

def main():
    ctx = pmisc.TmpFile
    with ctx() as fname1:
        with ctx() as fname2:
            with ctx() as ofname:
                # Create first data file
                data1 = [
                    [1, 9.99],
                    [2, 10000],
                    [3, 0.10]
                ]
                pcsv.write(fname1, data1, append=False)
                # Create second data file
                data2 = [
                    ['Joe', 10, 'Sunday'],
                    ['Sue', 20, 'Thursday'],
                    ['Pat', 15, 'Tuesday']
                ]
                pcsv.write(fname2, data2, append=False)
                # Merge file1 and file2
                pcsv.merge(
                    fname1=fname1,
                    has_header1=False,
                    fname2=fname2,
                    has_header2=False,
                    ofname=ofname
                )
                # Verify that resulting file is correct
                ref_data = [
                    [1, 9.99, 'Joe', 10, 'Sunday'],
                    [2, 10000, 'Sue', 20, 'Thursday'],
                    [3, 0.10, 'Pat', 15, 'Tuesday'],
                ]
                obj = pcsv.CsvFile(ofname, has_header=False)
                assert obj.header() == list(range(0, 5))
                assert obj.data() == ref_data

if __name__ == '__main__':
    main()

Sort a file

# pcsv_example_5.py
import pmisc, pcsv

def main():
    ctx = pmisc.TmpFile
    with ctx() as ifname:
        with ctx() as ofname:
            # Create first data file
            data = [
                ['Ctrl', 'Ref', 'Result'],
                [1, 3, 10],
                [1, 4, 20],
                [2, 4, 30],
                [2, 5, 40],
                [3, 5, 50]
            ]
            pcsv.write(ifname, data, append=False)
            # Sort
            pcsv.dsort(
                fname=ifname,
                order=[{'Ctrl':'D'}, {'Ref':'A'}],
                has_header=True,
                ofname=ofname
            )
            # Verify that resulting file is correct
            ref_data = [
                [3, 5, 50],
                [2, 4, 30],
                [2, 5, 40],
                [1, 3, 10],
                [1, 4, 20]
            ]
            obj = pcsv.CsvFile(ofname, has_header=True)
            assert obj.header() == ['Ctrl', 'Ref', 'Result']
            assert obj.data() == ref_data

if __name__ == '__main__':
    main()

Interpreter

The package has been developed and tested with Python 2.6, 2.7, 3.3, 3.4, 3.5 and 3.6 under Linux (Debian, Ubuntu), Apple macOS and Microsoft Windows

Installing

$ pip install pcsv

Documentation

Available at Read the Docs

Contributing

  1. Abide by the adopted code of conduct

  2. Fork the repository from GitHub and then clone personal copy [1]:

    $ git clone \
          https://github.com/[github-user-name]/pcsv.git
    Cloning into 'pcsv'...
    ...
    $ cd pcsv
    $ export PCSV_DIR=${PWD}
  3. Install the project’s Git hooks and build the documentation. The pre-commit hook does some minor consistency checks, namely trailing whitespace and PEP8 compliance via Pylint. Assuming the directory to which the repository was cloned is in the $PCSV_DIR shell environment variable:

    $ ${PCSV_DIR}/sbin/complete-cloning.sh
    Installing Git hooks
    Building pcsv package documentation
    ...
  4. Ensure that the Python interpreter can find the package modules (update the $PYTHONPATH environment variable, or use sys.paths(), etc.)

    $ export PYTHONPATH=${PYTHONPATH}:${PCSV_DIR}
  5. Install the dependencies (if needed, done automatically by pip):

    • Astroid (Python 2.6: older than 1.4, Python 2.7 or newer: 1.3.8 or newer)

    • Cog (2.4 or newer)

    • Coverage (3.7.1 or newer)

    • Docutils (Python 2.6: 0.12 or newer and older than 0.13, Python 2.7: 0.12 or newer, Python 3.3: 0.12 or newer and older than 0.13, Python 3.4: 0.12 or newer, Python 3.5: 0.12 or newer, Python 3.6: 0.12 or newer)

    • Inline Syntax Highlight Sphinx Extension (0.2 or newer)

    • Mock (Python 2.x only, 1.0.1 or newer)

    • Pexdoc (1.0.9 or newer)

    • Pmisc (1.2.2 or newer)

    • Py.test (2.7.0 or newer)

    • PyContracts (1.7.2 or newer except 1.7.7)

    • Pylint (Python 2.6: 1.3 or newer and older than 1.4, Python 2.7 or newer: 1.3.1 or newer)

    • Pytest-coverage (1.8.0 or newer)

    • Pytest-xdist (optional, 1.8.0 or newer)

    • ReadTheDocs Sphinx theme (0.1.9 or newer)

    • Sphinx (Python 2.6: 1.2.3 or newer and 1.4.9 or older, Python 2.7: 1.5 or newer, Python 3.3: 1.2.3 or newer and 1.4.9 or older, Python 3.4: 1.5 or newer, Python 3.5: 1.5 or newer, Python 3.6: 1.5 or newer)

    • Tox (1.9.0 or newer)

    • Virtualenv (13.1.2 or newer)

  6. Implement a new feature or fix a bug

  7. Write a unit test which shows that the contributed code works as expected. Run the package tests to ensure that the bug fix or new feature does not have adverse side effects. If possible achieve 100% code and branch coverage of the contribution. Thorough package validation can be done via Tox and Py.test:

    $ tox
    GLOB sdist-make: .../pcsv/setup.py
    py26-pkg inst-nodeps: .../pcsv/.tox/dist/pcsv-...zip

    Setuptools can also be used (Tox is configured as its virtual environment manager) [2]:

    $ python setup.py tests
    running tests
    running egg_info
    writing requirements to pcsv.egg-info/requires.txt
    writing pcsv.egg-info/PKG-INFO
    ...

    Tox (or Setuptools via Tox) runs with the following default environments: py26-pkg, py27-pkg, py33-pkg, py34-pkg and py35-pkg [3]. These use the Python 2.6, 2.7, 3.3, 3.4 and 3.5 interpreters, respectively, to test all code in the documentation (both in Sphinx *.rst source files and in docstrings), run all unit tests, measure test coverage and re-build the exceptions documentation. To pass arguments to Py.test (the test runner) use a double dash (--) after all the Tox arguments, for example:

    $ tox -e py27-pkg -- -n 4
    GLOB sdist-make: .../pcsv/setup.py
    py27-pkg inst-nodeps: .../pcsv/.tox/dist/pcsv-...zip
    ...

    Or use the -a Setuptools optional argument followed by a quoted string with the arguments for Py.test. For example:

    $ python setup.py tests -a "-e py27-pkg -- -n 4"
    running tests
    ...

    There are other convenience environments defined for Tox [4]:

    • py26-repl, py27-repl, py33-repl, py34-repl and py35-repl run the Python 2.6, 2.7, 3.3, 3.4 or 3.5 REPL, respectively, in the appropriate virtual environment. The pcsv package is pip-installed by Tox when the environments are created. Arguments to the interpreter can be passed in the command line after a double dash (--)

    • py26-test, py27-test, py33-test, py34-test and py35-test run py.test using the Python 2.6, 2.7, 3.3, 3.4 or Python 3.5 interpreter, respectively, in the appropriate virtual environment. Arguments to py.test can be passed in the command line after a double dash (--) , for example:

      $ tox -e py34-test -- -x test_eng.py
      GLOB sdist-make: [...]/pcsv/setup.py
      py34-test inst-nodeps: [...]/pcsv/.tox/dist/pcsv-[...].zip
      py34-test runtests: PYTHONHASHSEED='680528711'
      py34-test runtests: commands[0] | [...]py.test -x test_eng.py
      ==================== test session starts ====================
      platform linux -- Python 3.4.2 -- py-1.4.30 -- [...]
      ...
    • py26-cov, py27-cov, py33-cov, py34-cov and py35-cov test code and branch coverage using the Python 2.6, 2.7, 3.3, 3.4 or 3.5 interpreter, respectively, in the appropriate virtual environment. Arguments to py.test can be passed in the command line after a double dash (--). The report can be found in ${PCSV_DIR}/.tox/py[PV]/usr/share/pcsv/tests/htmlcov/index.html where [PV] stands for 26, 27, 33, 34 or 35 depending on the interpreter used

  8. Verify that continuous integration tests pass. The package has continuous integration configured for Linux (via Travis) and for Microsoft Windows (via Appveyor). Aggregation/cloud code coverage is configured via Codecov. It is assumed that the Codecov repository upload token in the Travis build is stored in the ${CODECOV_TOKEN} environment variable (securely defined in the Travis repository settings page). Travis build artifacts can be transferred to Dropbox using the Dropbox Uploader script (included for convenience in the ${PCSV_DIR}/sbin directory). For an automatic transfer that does not require manual entering of authentication credentials place the APPKEY, APPSECRET, ACCESS_LEVEL, OAUTH_ACCESS_TOKEN and OAUTH_ACCESS_TOKEN_SECRET values required by Dropbox Uploader in the in the ${DBU_APPKEY}, ${DBU_APPSECRET}, ${DBU_ACCESS_LEVEL}, ${DBU_OAUTH_ACCESS_TOKEN} and ${DBU_OAUTH_ACCESS_TOKEN_SECRET} environment variables, respectively (also securely defined in Travis repository settings page)

  9. Document the new feature or bug fix (if needed). The script ${PCSV_DIR}/sbin/build_docs.py re-builds the whole package documentation (re-generates images, cogs source files, etc.):

    $ ${PUTIL_DIR}/sbin/build_docs.py -h
    usage: build_docs.py [-h] [-d DIRECTORY] [-r]
                         [-n NUM_CPUS] [-t]
    
    Build pcsv package documentation
    
    optional arguments:
      -h, --help            show this help message and exit
      -d DIRECTORY, --directory DIRECTORY
                            specify source file directory
                            (default ../pcsv)
      -r, --rebuild         rebuild exceptions documentation.
                            If no module name is given all
                            modules with auto-generated
                            exceptions documentation are
                            rebuilt
      -n NUM_CPUS, --num-cpus NUM_CPUS
                            number of CPUs to use (default: 1)
      -t, --test            diff original and rebuilt file(s)
                            (exit code 0 indicates file(s) are
                            identical, exit code 1 indicates
                            file(s) are different)

    Output of shell commands can be automatically included in reStructuredText source files with the help of Cog and the docs.support.term_echo module.

    Similarly Python files can be included in docstrings with the help of Cog and the docs.support.incfile module

Footnotes

License

The MIT License (MIT)

Copyright (c) 2013-2017 Pablo Acosta-Serafini

Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the “Software”), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. .. CHANGELOG.rst .. Copyright (c) 2013-2017 Pablo Acosta-Serafini .. See LICENSE for details

Changelog

  • 1.0.5 [2017-02-10]: Package build enhancements and fixes

  • 1.0.4 [2017-02-07]: Python 3.6 support

  • 1.0.3 [2016-06-10]: Minor documentation build bug fix

  • 1.0.2 [2016-05-12]: Minor documentation updates

  • 1.0.1 [2016-05-12]: Minor documentation updates

  • 1.0.0 [2016-05-12]: Final release of 1.0.0 branch

  • 1.0.0rc1 [2016-05-11]: Initial commit, forked a subset from putil PyPI package

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pcsv-1.0.5.tar.gz (101.6 kB view hashes)

Uploaded Source

Built Distributions

pcsv-1.0.5-py36-none-any.whl (36.2 kB view hashes)

Uploaded Python 3.6

pcsv-1.0.5-py35-none-any.whl (36.2 kB view hashes)

Uploaded Python 3.5

pcsv-1.0.5-py34-none-any.whl (36.2 kB view hashes)

Uploaded Python 3.4

pcsv-1.0.5-py33-none-any.whl (36.2 kB view hashes)

Uploaded Python 3.3

pcsv-1.0.5-py27-none-any.whl (36.2 kB view hashes)

Uploaded Python 2.7

pcsv-1.0.5-py26-none-any.whl (36.2 kB view hashes)

Uploaded Python 2.6

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page