From now, we tokenize each answers using 'gpt2' tokenizer, and it is not selectable.
For more accuracy, we have to tokenize each generator results using corresponding tokenizer to LLM models.
It can be useful to use tokenlog library.
But have to implement LlamaIndex support at tokenlog library.
Last question is...
Is it worth it?
Pay now to fund the work behind this issue.
Get updates on progress being made.
Maintainer is rewarded once the issue is completed.
You're funding impactful open source efforts
You want to contribute to this effort
You want to get funding like this too