From boris.smilga at gmail.com Wed Jul 23 17:07:48 2008 From: boris.smilga at gmail.com (Boris Smilga) Date: Wed, 23 Jul 2008 21:07:48 +0400 Subject: [cl-json-devel] proposed patch: encoding and unencoding CLOS objects In-Reply-To: <50e8e4f60712270426m13e85ed2uf46fec3b4f5fce3f@mail.gmail.com> References: <50e8e4f60712270426m13e85ed2uf46fec3b4f5fce3f@mail.gmail.com> Message-ID: <4AAE3300-605A-4EE9-9505-1AC3482AA8DB@gmail.com> Please find attached a patch bundle with testcases for CLOS semantics and prototypes. I have added a score of new tests, and fixed some existing tests so that their results do not depend on the global (more precisely, dynamic) settings of the semantics variables, *JSON-SYMBOLS-PACKAGE* and *PROTOTYPE-NAME*. All tests now run OK (tested under both Clozure CL and SBCL on Darwin/PPC). Sincerely, B. Smilga. -------------- next part -------------- A non-text attachment was scrubbed... Name: testcases-for-clos-semantics.dpatch Type: application/octet-stream Size: 23349 bytes Desc: not available URL: From boris.smilga at gmail.com Wed Jul 23 17:32:41 2008 From: boris.smilga at gmail.com (Boris Smilga) Date: Wed, 23 Jul 2008 21:32:41 +0400 Subject: [cl-json-devel] proposed: improvements to decoder customization Message-ID: <30B0F7B1-A050-4433-B2B0-63F56840A60F@gmail.com> [This memo started as a follow-up to the e-mail exchange between Henrik and your humble author (now recorded in the TODO file), but, on reviewing it, I thought that it would perhaps be of some interest to the broader readership of cl-json-devel. This is to explain the rather personal style of discourse, and make apologies to a reader who might ?but who need not!? feel excluded.] ?1. The design of the improved decoder interface laid out below came to be as I was considering an application of CL-JSON that would involve local structured storage of transmitted data in a Berkeley DB. In such a setting one would like to bypass creating Lisp data structures from JSON input, instead packing it directly to FFI C data to be processed by libdb. The question that I asked myself was how to (re)design the decoder so that this task could be dealt with gracefully. I have also checked the scheme against the several probable sets of requirements you had described in your e-mail: > Configurable: I really would want to have it optional exactly how > to decode and encode objects. You are probably right in that your > solution is a good default behaviour, but it will not be perfect in > every situation for every user. So if possible, I would like to > have the decoding and encoding configurable. So backwards > compatibility does not really mean backwards, it means > compatibility between different setups. For example, when doing > testcases or a simple json-bind the old alist setup is good, for a > more advanced setup your code is great. For a secure setup you will > probably want to have access control to what objects you are > allowed to create and so on. One solution will never be sufficient > for everyone. ?2. Hence, the general principles which should be obeyed by the implementation, are: A. Separation of concerns. The implementation should comprise a fixed basic level to handle the parsing, JSON well-formedness, and flow control; and a customizable level to produce Lisp data (or perform some other JSON-driven task, as it might please the user to enjoin). You were very rightfully incensed against my intermingling parsing with looking for prototype in READ-JSON-OBJECT. It will become clear below how this can be done away with. B. Fine grain. Not only the handling of objects, but also that of other JSON types should be customizable. In particular, the handling of arrays and strings should be customizable on elemental level, i.e. the user should have a way to determine how the decoder handles elements of arrays and strings. In handling objects, customization should be available for keys as well as for key-value pairs. ?3. This suggests a design similar in spirit to SAX, with a set of handlers triggered by ?events?. The current implementation partly follows this scheme by providing the *JSON-OBJECT-FACTORY... callbacks, but there are more kinds of events than can be handled by those three; here's a tentatively exhaustive list: 1. An atomic constant (integer, real, or boolean). 2a. Beginning of string. 2e. A string character or escape sequence. 2z. End of string. 3a. Beginning of array. 3e. An array element. 3z. End of array. 4a. Beginning of object. 4k. An object property key. 4v. An object property value. 4z. End of object. Accordingly, we need at least as many handlers. The handlers for (1), (2e), (3e), (4k/v) shall be passed the token that triggered the event; (2a), (3a), and (4a) shall produce some fresh ?accumulator? value that is then piped through successive calls to (2e), (3e), and (4k/v), respectively. E.g., reading the JSON {"f\u0151o": [1, true]} could result in the following flow of calls to handlers (I use mnemonic names of function-valued handler variables, which are certainly not definitive): READ-JSON-OBJECT 4a: *BEGINNING-OF-OBJECT-HANDLER* () produces some value (?) O READ-JSON-STRING 2a: *BEGINNING-OF-STRING-HANDLER* () ? S 2e: *STRING-CHAR-HANDLER* (#x66, S) ? S? 2e: *STRING-CHAR-HANDLER* (#x151, S?) ? S? 2e: *STRING-CHAR-HANDLER* (#x6f, S?) ? S? 2z: *END-OF-STRING-HANDLER* (S?) ? T 4k: *OBJECT-KEY-HANDLER* (T) ? K READ-JSON-ARRAY 3a: *BEGINNING-OF-ARRAY-HANDLER* () ? A 1: *INTEGER-HANDLER* ("1") ? I 3e: *ARRAY-ELEMENT-HANDLER* (I, A) ? A? 1: *BOOLEAN-HANDLER* ("true") ? B 3e: *ARRAY-ELEMENT-HANDLER* (B, A?) ? A? 3z: *END-OF-ARRAY-HANDLER* (A?) ? V 4v: *OBJECT-VALUE-HANDLER* (K, V, O) ? O? 4z: *END-OF-OBJECT-HANDLER* (O?) ? P, which is also the return value of READ-JSON-OBJECT. The nature of the values O, S, S?, S?, S?, T, etc. is of absolutely no concern to the base level of the decoder whose only duty is to faithfully pass them around. For one, a dumb JSON syntax checker can have all handlers set to (CONSTANTLY T). ?4. Now the current list semantics can be expressed straightforwardly in terms of these handlers: *BEGINNING-OF-ARRAY-HANDLER* ? (lambda () (let ((list (cons nil nil))) ; First element is never used (cons list list))) ; First and last pair of the list *ARRAY-ELEMENT-HANDLER* ? (lambda (elt accumulator) (destructuring-bind (head . last) accumulator (cons head (setf (cdr last) (cons elt nil))))) *END-OF-ARRAY-HANDLER* ? (lambda (accumulator) (coerce (cdar accumulator) *json-array-type*)) *BEGINNING-OF-OBJECT-HANDLER* ? same as *BEGINNING-OF-ARRAY-HANDLER* *OBJECT-KEY-HANDLER* ? #'json-intern *OBJECT-VALUE-HANDLER* ? (lambda (key value accumulator) (destructuring-bind (head . last) accumulator (cons head (setf (cdr last) (cons (cons key value) nil))))) *END-OF-OBJECT-HANDLER* ? #'cdar *INTEGER-HANDLER* ? #'parse-integer You can easily reconstruct the rest. ?5. Our CLOS semantics is a more tricky affair. If I may allow myself to reiterate what I previously said about the handling of prototype fields: > READ-JSON-OBJECT works by reading in the key string and the colon > separator, and then recursively calling READ-JSON to consume > characters from the input and construct the object which would be > the corresponding value. With my modifications, when the key > "prototype" (or whatever the name shoud be) is encountered on the > input, the decoder is told that the value we are going to construct > shall be the prototype object. If we were not able to communicate > this beforehand, we'd have to do much post-processing of the > factory object: look up the prototype (note that, at this point, we > would not know the package and could not yet intern the keys, so > that the matching would be done on strings, degrading the > performance), convert it to some internal format (note that we > would not be able to predict accurately enough what it would have > been decoded to, as that may be influenced by user-side > configuration), remove the prototype and key from the factory, and > only then could we create the object. Put rather abstractly, to overcome this difficulty we need a means to pass certain information from outer to inner recursive calls of READ-JSON-OBJECT, and between handlers invoked on the same level. I currently see two options of implementing this: A. ?Accumulator? return values of (3a/e) and (4a/k/v) handlers (what is signified by A, A?, ..., and O, O? in the above example) are passed down as additional arguments to the recursive calls to READ-JSON-OBJECT and READ-JSON-ARRAY, and are then also passed to (3a) and (4a) handlers in these inner functions (and perhaps also to (1) and (2a)). The (4k) handler, like (4v), receives and produces an ?accumulator? value rather than a representation of the key. Thus, the flow in the above example becomes: ... 4k: *OBJECT-KEY-HANDLER* (T, O) ? O? READ-JSON-ARRAY (..., O?) 3a: *BEGINNING-OF-ARRAY-HANDLER* (O?) ? A ... 3z: *END-OF-ARRAY-HANDLER* (A?) ? V 4v: *OBJECT-VALUE-HANDLER* (V, O?) ? O? ... This would allow for an implementation of the CLOS semantics along the following lines: *BEGINNING-OF-OBJECT-HANDLER* ? (lambda (&optional (above-accumulator nil not-toplevel)) (let* ((prototype (if not-toplevel (caar above-accumulator))) (list (cons prototype nil))) (cons list list))) *OBJECT-KEY-HANDLER* ? (lambda (key accumulator) (destructuring-bind (head . last) accumulator (let ((prototype (car head))) (if (and (not prototype) (string= key (symbol-name *prototype-name*))) (cons (cons t (cdr head)) last) (cons head (setf (cdr last) (cons (cons key nil) nil))))))) *OBJECT-VALUE-HANDLER* ? (lambda (value accumulator) (destructuring-bind (head . last) accumulator (if (typep value 'prototype) (cons (cons value (cdr head)) last) (progn (setf (cdar last) value) accumulator)))) *END-OF-OBJECT-HANDLER* ? (lambda (accumulator) (destructuring-bind ((prototype . fields) . last) accumulator (let ((*json-object-prototype* prototype)) (json-factory-make-object fields)))) B. In addition to custom handlers, the user is given the possibility to provide a list of dynamic variables which he wishes to have ?JSON-structure scope?. That is, the bodies of READ-JSON-OBJECT and READ-JSON-ARRAY are wrapped in PROGVs which establish new dynamic bindings for these variables (using their outer-level bindings for respective initial values). If a handler sets a structure-scope variable, the new value is then visible to all subsequent handlers until the current READ-JSON-OBJECT or READ-JSON-ARRAY loop exits. By constrast, the handlers on the outer levels never see the value change. The flow in the example is unaltered, and the CLOS semantics is implemented thus: *JSON-STRUCTURE-SCOPE-VARIABLES* ? '(*json-object-prototype*) *BEGINNING-OF-OBJECT-HANDLER* ? (lambda () (let ((list (cons nil nil))) (cons list list))) *OBJECT-KEY-HANDLER* ? (lambda (key) (if (and (not *json-object-prototype*) (string= key (symbol-name *prototype-name*))) (setq *json-object-prototype* t)) key) *OBJECT-VALUE-HANDLER* ? (lambda (key value accumulator) (destructuring-bind (head . last) accumulator (if (typep value 'prototype) (progn (setq *json-object-prototype* value) accumulator) (cons head (setf (cdr last) (cons (cons key value) nil)))))) *END-OF-OBJECT-HANDLER* ? (lambda (accumulator) (json-factory-make-object (cdar accumulator))) The choice between the options A and B seems to be one between simplicity of interface and efficiency: a lot of nested PROGVs are likely to incur some cost overhead (I do not feel myself qualified enough to predict exactly how much). On the other hand, handlers are rather more readable, as you can see in the above comparison, and the interface also stays more intuitive that way. ?6. Returning to my BDB + CL-JSON application: the (2a), (3a), (4a) handlers coud be customized to call data constructor / initializer functions over the FFI, and (2z), (3z), (4z) to call the DB put method. The uppermost call to READ-JSON would be wrapped up as a transaction. That would be it. Sincerely, - B. Smilga.