-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Event parsing error: String length (...) exceeds the maximum length (20000000) #1316
Comments
I cannot find a way to set |
so I'm not atually sure if that spark config is being used at all, we are using the JsonProtocol to read the json events. that has its own json mapper:
Which pehaps has a default the same max string len? Anyway, I do wonder if we could override that by making our own class |
After discussing this issue offline with @tgravescs , we decided to do the following:
|
Signed-off-by: Ahmed Hussein <[email protected]> Fixes NVIDIA#1316 Allow user-tools to pick the SPARK dependencies based on a runtime env_var. The value format follows the same format of `buildver` in the scala pom file. Currently 333 and 350 (default) are supported. If user specifies an invalid value, there will be a warning message, then the process fails running the java cmd. **Changes** - Add dependency key to the platform config-file - A platform can define its own default dependency versions using `activeBuildVer` key - Add a default `RUNTIME_BUILDVER` in the `__init__.py` to allow upgrades of spark release during official releases - Read an env_var `RAPIDS_USER_TOOLS_RUNTIME_BUILDVER` to pick the correct dependency. - Currently, only `333` and `350` are supported. Default is `350`
Signed-off-by: Ahmed Hussein (amahussein) <[email protected]> Followup on NVIDIA#1326 to set the default spark version to 3.4.2 for onPrem to avoid the bug described in NVIDIA#1316 without need to do something on customer side.
Signed-off-by: Ahmed Hussein (amahussein) <[email protected]> Followup on #1326 to set the default spark version to 3.4.2 for onPrem to avoid the bug described in #1316 without need to do something on customer side.
Describe the bug
Looks like Spark 3.5.0 changed the jackson version and now there is a max json string length configuration. One of our customers hit this issue running the qualification tool, see:
https://issues.apache.org/jira/browse/SPARK-47150
Note that if the event isn't critical it will just ignore it and go on, I assume if it is critical that app may not be qualified properly.
A temporary solution is to go back to 3.4.X. But we may want to look at setting that spark config to up the maximum allowed.
full stack trace:
The text was updated successfully, but these errors were encountered: