-
Notifications
You must be signed in to change notification settings - Fork 6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prevent hasher from running out of memory on large files #2451
Conversation
Test PASSed. |
Test FAILed. |
Test FAILed. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, just one comment on the printing.
python/ray/monitor.py
Outdated
ip = self.local_scheduler_id_to_ip_map.get(client_id) | ||
if ip: | ||
self.load_metrics.update(ip, static_resources, dynamic_resources) | ||
else: | ||
print("Warning: could not find ip for client {}." | ||
.format(client_id)) | ||
for keys, values in self.local_scheduler_id_to_ip_map.items(): | ||
print(keys) | ||
print(values) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should this be a single print of the entire map?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That is my bad. This was some later debugging I was doing. Didn't know it would automatically end up here. I will remove it.
Test FAILed. |
@ericl I don't understand why the test is failing |
�[32;1mThe command "cd .." exited with 0.�[0m travis_time:end:047f9cbc:start=1532559737674351176,finish=1532559750107885280,duration=12433534104
travis_time:end:004128c0:start=1532559750112843161,finish=1532559783434014143,duration=33321170982 Done. Your build exited with 1. |
Some lint errors, other tests look unrelated. |
Test FAILed. |
Lint error fixed. |
Looks good, thanks! |
What do these changes do?
When the autoscaler calculates the hash it can run out of memory on large files (3GB in my case). I added support for incremental hashing on files larger than 1GB to fix this. I just picked this threshold arbitrarily and we could turn it down if that is desired.
Related issue number
I did not open an issue.