Skip to main content

All Questions

Tagged with
Filter by
Sorted by
Tagged with
0 votes
1 answer
32 views

Pytorch + Ray Tune Reporting ImplicitFunc Is Too Large, No Idea Which Reference Is Large

Similar to this question, Ray Tune is reporting to me: ValueError: The actor ImplicitFunc is too large (421 MiB > FUNCTION_SIZE_ERROR_THRESHOLD=95 MiB). Check that its definition is not implicitly ...
Falcondance's user avatar
0 votes
0 answers
21 views

Policies' directory not present in saved checkpoint

I'm using RayRL Lib, and after switching to the new API version, the checkpoint directory no longer includes the policies folder. Why this might be happening? Currently, the checkpoints contain the ...
Khashayar Ghamati's user avatar
1 vote
0 answers
24 views

Using one Neural Network model inside of ray.pool

I am concerned about using Pool() from ray library. This is simplified structure of my question: import torch import multiprocessing as mp from ray.util.multiprocessing import Pool def acquire_data(...
Vukasin S's user avatar
0 votes
1 answer
364 views

Where should I put reuse_actors=True?

After running the below code, it says INFO trainable.py:172 – Trainable.setup took 2940.989 seconds. if your trainable is slow to initialize, consider setting reuse_actors=True to reduce actor ...
Animesh Kumar Paul's user avatar
1 vote
1 answer
795 views

Select best training results from all history using ray

I'm trying to find optimal hyperparams with ray: tuner = tune.Tuner( train, param_space=hyperparams1, tune_config=tune.TuneConfig( num_samples=200, metric="score",...
ckorzhik's user avatar
  • 788
1 vote
1 answer
596 views

How can I change the ray_results folder when using TuneGridSearchCV?

I am running quite a large parameter search using TuneGridSearchCV on an xgboost model using my university's HPC cluster. The results are being saved to ~/ray_results however I don't have enough space ...
shoopdoop's user avatar
1 vote
0 answers
377 views

Memory usage on node keeps increasing while training a model with Ray Tune

This is the first time I am using Ray Tune to look for the best hyperparameters for an DL model and I am experiencing some problems related to memory usage. The Memory usage on this node keeps ...
Benjamin Cretois's user avatar
0 votes
1 answer
2k views

Use GPU OR CPU on Ray tune

I have 1 GPU and 32 CPUs available in my machine. Is it possible in Ray to use them separatelly? For instance, one task gets allocated with 1 CPU and another task with 1 GPU? If I use tune.run(...
Douglas C Vasconcelos's user avatar
2 votes
0 answers
427 views

Trained well with DQN, but not learning with A2C

I've used Ray RLlib's DQN to train in my custom simulator. It usually produced good results after 15 million steps. After playing around with DQN for a while, I'm now trying to train A2C in the ...
Kai Yun's user avatar
  • 107
0 votes
1 answer
1k views

Does RLlib `rollout.py` work for evaluation?

TL;DR: RLlib's rollout command seems to be training the network, not evaluating. I'm trying to use Ray RLlib's DQN to train, save, and evaluate neural networks on a custom made simulator. To do so, I'...
Kai Yun's user avatar
  • 107
1 vote
2 answers
1k views

Using Ray RLlib with custom simulator

I'm very new to Ray RLlib and have an issue with using a custom simulator my team made. We're trying to integrate a custom Python-based simulator into Ray RLlib to do a single-agent DQN training. ...
Kai Yun's user avatar
  • 107
0 votes
1 answer
373 views

Errors when trying to use DQN algorithm for FrozenLake Openai game

I am trying to make a very simple DQN algorithm work with the FrozenLake-v0 game but I am getting errors. I understand that it could be an overkill using DQN instead of a Q-table, but I nonetheless ...
mikanim's user avatar
  • 439
0 votes
1 answer
1k views

Using Ray-Tune with sklearn's RandomForestClassifier

Putting together different base and documentation examples, I have managed to come up with this: X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2) def objective(config, ...
LeggoMaEggo's user avatar
1 vote
0 answers
333 views

Reinforcement Learning Agent Policy Won't Update

I am currently training a deep reinforcement learning model in a continuous environment using Ray. The environment I am using was coded up in OpenAI Gym using baselines by another person who's ...
Jerry Ginger's user avatar
1 vote
1 answer
421 views

Configuring `ray` for multiple users

Is there a way to set up a ray cluster so that multiple users can access it and submit jobs? I haven't been able to find anything about this in their documentation.
Andrew Sedler's user avatar
4 votes
0 answers
974 views

What is the appropriate conv_filter specifications for custom observation_space?

I'm trying to create a custom openai gym environment for use in a ray reinforcement learning environment. Currently I have an observation space of (6, 94). However, when I attempt to train in my ...
advanced_wombat's user avatar