In the discussion, it was mentioned that using all caps, such as capitalizing "do not," seems to affect the performance of large language models (LLMs). At timestamp [00:10:09], the speaker shared an anecdote about how capitalizing "do not" yielded better results, acknowledging that there's a lack of understanding about why this happens. The conversation at [00:10:41] further delves into this, suggesting that if an LLM has case recognition, then "DO NOT" might be interpreted semantically differently than "do not." This implies that using all caps could potentially be seen as a different prompt by the model, affecting the results. However, this is based on observations and experiences, and there's no clear explanation provided as to the mechanics behind why this occurs.
Recommendations