-
Notifications
You must be signed in to change notification settings - Fork 2
Concurrency #75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Concurrency #75
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, please address comments (esp. requirements.txt one) and merge (assuming that the tests pass then)
|
||
# Find if some ledger files are too large to fit in the system's available memory. | ||
too_large_ledgers = set() | ||
for ledger in get_ledgers(): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't see why this loop is needed, can't it be done within the previous loop? (starting in L623)
@@ -6,7 +6,7 @@ | |||
import os | |||
import datetime | |||
import calendar | |||
import argparse | |||
import psutil |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Needs to be added to requirements.txt
6f9ae63
to
ccb829e
Compare
All Submissions:
Description
Enables concurrency during analyzing. Adds a helper function that computes the max number of parallel processes that can run per ledger (based on the memory consumption of each ledger's largest file) and uses Python's multiprocessing to analyze snapshots with multiple parallel processes.