You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Using the flores101 or flores200 tokenizers is resulting in a memory leak.
I am using version 2.4.2 on Windows 11, but the same was also occurring on version 2.4.0.
Running the following results in memory usage increasing linearly until crashing: import sacrebleu while True: sacrebleu.sentence_bleu("Hello world.", ["Hello world."], tokenize="flores101")
This is also the case for corpus_bleu.
I do not think that it is due to caching since I am running it over and over on the same sentence.
The text was updated successfully, but these errors were encountered:
Using the flores101 or flores200 tokenizers is resulting in a memory leak.
I am using version 2.4.2 on Windows 11, but the same was also occurring on version 2.4.0.
Running the following results in memory usage increasing linearly until crashing:
import sacrebleu
while True:
sacrebleu.sentence_bleu("Hello world.", ["Hello world."], tokenize="flores101")
This is also the case for corpus_bleu.
I do not think that it is due to caching since I am running it over and over on the same sentence.
The text was updated successfully, but these errors were encountered: