Skip to content

[BUG]: Spark has to be restarted if using UDFs and recompiling #426

Open
@borgdylan

Description

@borgdylan

Whenever I use UDFs in an application utilizing .NET for Spark, and recompile the application regardless of whether I change the UDF code or not, the Spark worker no longer manages to load the UDF when it has to be executed as part of a query. The only workaround I found is to restart the Spark process before each run. This is OK when doing data science on a desktop, but not so if using Spark for server side processing where one expects Spark to keep running when updating the binaries that leverage Spark (there my be old and new instances during a roll-out for example). This is consistently reproducible for me when running Spark in debug mode (as a standalone server) regardless of which UDF I use or which dataset I use.

Desktop (please complete the following information):

  • OS: Linux x64 (Ubuntu 19.10)

Metadata

Metadata

Assignees

No one assigned

    Labels

    questionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions