Datconv’s documentation

Datconv - Universal Data Converter

Datconv is a program designed to perform configurable comversion of file with data in one format to file with data in another format or database.

Program should be run using Python 2.7 or Python 3.x interpretter. It also requires installation of external packages: lxml, PyYAML. For more information see INSTALL.rst file distributed in source pack.

Both input and output files can be text or binary files. However it is assumed that those files have following structure:

—–
Header
—–
Record 1
Record 2
…..
Record N
—–
Footer
—–

There may be different types of records (i.e. every record has attribute called record type). Each record may contain different number and kind of data (has different internal structure) even among records of the same type.

Program has modular architecture with following swichable compoments:

Reader

Major obligatory component responsible for:

  • reading input data (i.e. every reader class assumes certain input file format)
  • driving entire data conversion process (i.e. main processing loop in implemented in this class)
  • determine internal representation of header, records and footer (this heavily depands on reader and kind of input format).

API of this component: Reader interface

Filter

Optional compoment that is able to:

  • filter data (i.e. do not pass certain records further - i.e. to Writer)
  • change data (i.e. change on the fly contents of certain records)
  • produce data (i.e. cause that certain records, maybe slightly modified, are being sent multiply times to writer)
  • break conversion process (i.e. cause that conversion stop on certain record).

API of this component: Filter interface

Writer

Obligatory component responsible for:

  • re-packing data from element-tree internal format to strings or objects.

API of this component: Writer interface

Output Connector

Obligatory component responsible for:

  • writing data to destination storage.

API of this component: Output Connector interface

Logger
All messages intended to be presented to user are being send (except few very initial error messages) to Logger classes from Python standard package logging. Datconv can use all logging comfiguration power available in this package.

In this version of package following compoments are included:

  • Readers: XML (sax), CSV (sax), JSON (sax).
  • Filters: Few basic/sample filters.
  • Writers: XML, CSV, XPath (helper module), JSON.
  • Output Connectors: File (Text, MS Excel), Databases (SQLite, PostgreSQL, Crate).

So Datconv program can be used to convert files between XML, CVS and JSON formats and saving data in those formats to database. Sax means that used parsers are of event type - i.e. entire data are not being stored in memory (typically only just one record), what means that program is able to process large files without allocating a lot of memory. It may be also usefull in case you have some files in custom program/company specific data format that you want to look up or convert. Then it is enough to write the reader component for your cutom data format compatible with Datconv API and let Datconv do the rest. Actually this is how I’m currently using this program in my work.

If you’d prefer to work in JavaScript environment please look at Pandat Project which has similar design and purpose.

This program was inspired by design of Pandoc Project.

Performance

Main design principle of this tool was generality and flexibility rather than performance and use scenarios with very big data. This version of program runs in one thread (on one CPU core) and does not consume a lot of modern computer resources. So in case of processing of very big data consider dividing data into smaller chunks and run few instances of this program in parallell or use rfrom-rto parameters avaialble in readers. Or if you have to process big files in short time and do not need that much flexibility (espacially filtering possibilities) probaly special dedicated program (which will not translate data to internal XML-like format) would process your data faster.

Measured performance (version 0.6.0):

  • Hardware: Powerfull Laptop (2017 year), CPU Frequency 2.9 GHz, SSD Drive
  • Input: XML File: 942MB (400.000 records)
  • Output: JSON File: 639MB
  • Conversioon time without filter: 4 min 41 s
  • With filter (datconv.filters.delfield, 2 tags to remove): 4 min 41 s (the same; probably smaller record to write to JSON file compensated effort for Filter invocation).
  • Opposite direction (JSON output converted back to XML): 5 min 51 s.

Datconv data flow schema (without filter)


_images/schema1.svg

Datconv data flow schema (with filter)


_images/schema2.svg

Indices and tables