Skip to content

Some issues when combining deep_architect and ray.tune #8

@iacolippo

Description

@iacolippo

Hi, first of all, I'd like to thank you for building and releasing deep_architect.

I am opening this issue because I'd like to use deep_architect together with ray.tuneto get the best of both worlds, but I encountered some issues. Feel free to close this if you think it is out of the scope of the project.

My goal is to use the sampling capabilities of deep_architect and the tools for multiprocessing and logging of ray and ray.tune. Therefore I'm using tune.run and tune.Trainable with the searchers, helpers and modules of deep_architect.

If I write my code with the call to the sampling function inside the _setup method of a tune.Trainable

https://gist.github.com/iacolippo/1262c8afbfd9f5e491add5fbae105afa (line 124)

then I have an issue with ray(tensorboard) logging. I'd say this is not an issue of deep_architect, and it shouldn't be too hard to fix in the source code of ray if need be.

If I write my code as ray wants it (the config["model"] is the model object, in this case, a PytorchModel from deep_architect), then I have a different error.

RuntimeError: Only Tensors created explicitly by the user (graph leaves) support the deepcopy protocol at the moment

https://gist.github.com/iacolippo/3f815fa90c254f7a065bdc446406233a (not that the () disappeared at line 124)

This might be an issue with deep_architect and multiprocessing, or Pytorch itself, I don't know, I didn't dig into it too much for lack of time. Here is the traceback.

traceback.log

I am using

-e [email protected]:negrinho/deep_architect.git@3427c5d45b0cbdc9c2fe1f4e5213f6961ef41749#egg=deep_architect
ray==0.8.4
torch==1.5.0
torchvision==0.6.0a0+82fd1c8

Stay safe!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions