I find it kind of hilarious how almost every prompt I've seen leaked from various apps almost always has a similar clause, as if it would have any effect at all on the result.
Seeing engineers resort to this level of basically praying and wishful thinking that in reality has no factual value is pretty funny.
"Please, don't give me wrong results 0_0"
What's the state of LLM detection algorithms? Is there any with a higher sucess rate and with OK-ish amount of false positives? Is there even a FOSS solution for detecting chatgpt? Would make for a great tool to have, I'm getting tired of this.