You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I haven't thought through all the details of how this would be implemented, but I feel like re-fragmenting the entire PDB every time is a waste of power.
It seems to me outputting a text file with parameters such as what the window size is, when the fragments were made (i.e. less than a week ago, otherwise there are new PDBs deposited), etc and then writing all the hadoop sequence files of the fragments to a new directory would be ideal. The pseudo-code would potentially look something like this:
If param file exists
if param file is same window size & less than a week old
pass path to stored files to remainder of fragment-search (i.e. calculations)
elseif less than a week old
just update the last week's pdbs and pass them in + the old fragments (if possible?)
else
start fresh
Some of it may be more complicated than it's worth. But minimally checking to see if the files already exist and are of same window size should be do-able I think.
The text was updated successfully, but these errors were encountered:
I haven't thought through all the details of how this would be implemented, but I feel like re-fragmenting the entire PDB every time is a waste of power.
It seems to me outputting a text file with parameters such as what the window size is, when the fragments were made (i.e. less than a week ago, otherwise there are new PDBs deposited), etc and then writing all the hadoop sequence files of the fragments to a new directory would be ideal. The pseudo-code would potentially look something like this:
Some of it may be more complicated than it's worth. But minimally checking to see if the files already exist and are of same window size should be do-able I think.
The text was updated successfully, but these errors were encountered: