My minimal LLM instructions. These are the current ones for @jetbrains IDEs with their in-IDE AI in particular.
LLMs overengineer so easily. They have verbal diarrhea so readily. They have comments like “set a to 1” for lines like “a = 1” all the time.
This prompt reins them in … somewhat.
Does this work? Given that the LLM doesn’t actually know anything or have feelings of uncertainty, surely it just adds a chance that it will say “I don’t know” purely at random, without making it any more likely that the answers it does give are correct.
I find it kind of hilarious how almost every prompt I’ve seen leaked from various apps almost always has a similar clause, as if it would have any effect at all on the result.
Seeing engineers resort to this level of basically praying and wishful thinking that in reality has no factual value is pretty funny.
“Please, don’t give me wrong results 0_0”