On Thu, 13 Nov 2014 11:15:24 GMT, David Monagle wrote:

Apologies for the essay, I'm really interested in where this is going so as to know the best place to work and contribute.

I can't say much about the direction it's taking, but there's a little bit of room for improvement using buffers, mostly for BIG json documents (> 1MB).

  • There's currently no way of having access to part of the schema during the transfer, which is useful to stop transfer if the metadata is wrong (e.g. mixing in a session ID with Json data that contains a base64 image)

  • There's possibilities for a memory attack with Json documents that exceed a certain size. Currently, the entire document piles up in the RAM.

I have a "mix" of solutions that also enable other openings ie. in the database world. This involves using Json as a schema, with the values being light integers to file offsets. The Json object would query a disk file for the string data, and throw an error if the schema has too many keys. If you add a small log writer / reader for the BigJson modifications, you get a nice little embedded database.

Other than that, anything related to the standards seems nicely implemented.