Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Fascinating. My thinking was that this is an upgrade over a bash script because you can prompt the AI to check it, clear installs with you, or otherwise investigate safety before installing in a way that isn't natural with *.sh. Does that make any amount of sense or am I just crazy?


Bash scripts give you visibility into what they are going to do by virtue of being machine instructions in a determimistic language. MD files you pipe to matrix multiplication has a much lower chance of being explainable.


Yeah, someone else was pointing that the bash scripts are guaranteed to do the same thing on every system which I think is in the same vein as your feedback. It's for sure a downside of the markdown that I need to explain the docs behind the proposal.


Time and time again, be it "hallucination", prompt injection, or just plain randomness, LLMs have proven themselves woefully insufficient at best when presented with and asked to work with untrusted documents. This simply changes the attack vector rather than solving a real problem


In a computing system, LLMs aren't substituting for code, they're substituting for humans. Treat them accordingly.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: