HCK-15381: fix possible JSON parse error#326
Merged
VitaliiBedletskyi merged 1 commit intodevelopfrom Mar 19, 2026
Merged
Conversation
|
VitaliiBedletskyi
approved these changes
Mar 19, 2026
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.



Content
Fix JSON parse error in
filterCorruptedDatawhen Databricks response is truncated mid-tableWhen reverse engineering large schemas, Databricks may truncate API responses by inserting a *** WARNING: max output size exceeded, skipping output. *** marker in the middle of the output. The
filterCorruptedDatafunction attempts to recover valid JSON by discarding incomplete objects from both sides of the split.Root cause: Two bugs in the
isTruncatedInMiddlebranch:Wrong join delimiter — filtered chunks were joined with ', ' instead of
JSON_OBJECTS_DELIMITER(}, {). This consumed the object boundary, merging two table objects into one and causing silent data loss (duplicate keys resolved by last-wins semantics).Empty chunk not handled — when the second chunk contains only a single incomplete table (no }, { delimiter),
filterCorruptedStartreturns an empty string. Joining with ', ' produced a trailing comma ("..., ") which causedJSON.parseto throw:SyntaxError: Expected double-quoted property name in JSON at position ...Fix:
JSON_OBJECTS_DELIMITERfor joining to preserve proper JSON object boundaries}]}only when the last chunk contributed no complete tables