Please tell me which format to choose for storing large-scale text data. It must meet the following criteria (at least 3 of 5). The criteria are in order of priority:
- Convenient view of data structure (data nesting can be of arbitrary depth)
- Python library support (converting, parsing)
- Fast and small (parsing, loading into the database)
- Convenient distribution (download, send over the network)
- Human readable view (ability to read and edit directly by human)
Considering formats - CSV, XML, JSON. I would be grateful for advice on the choice of these formats or your suggestions.
UPD. Few clarifications to the question. Why bothered to choose the format?
Collected a large amount of data for your project (engineering and scientific data).
The task of structuring and storing them has arisen, suddenly information will be useful to someone, and I will be able to pass it on to him. Consequently, a humanoid look would be most welcome.
It is possible that some value will change, and in order not to start parsing on a new one, you must edit the file directly.
In addition, the received data must be imported into the database, in my case PostgresQL, and any person who received my text data can do the same in any convenient database.