• 0 Posts
  • 2 Comments
Joined 2 years ago
cake
Cake day: June 20th, 2023

help-circle

  • Anecdotally, I use it a lot and I feel like my responses are better when I’m polite. I have a couple of theories as to why.

    1. More tokens in the context window of your question, and a clear separator between ideas in a conversation make it easier for the inference tokenizer to recognize disparate ideas.

    2. Higher quality datasets contain american boomer/millennial notions of “politeness” and when responses are structured in kind, they’re more likely to contain tokens from those higher quality datasets.

    I haven’t mathematically proven any of this within the llama.cpp tokenizer, but I strongly suspect that I could at least prove a correlation between polite token input and dataset representation output tokens