Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Can this be used for deobfuscation of code? I really hadn’t thought about LLM being a tool during reverse engineering.


Big LLMs like GPT-4 (and even GPT 3.5 Turbo) can be directly used to beautify obfuscated/minified JS, see e.g. https://thejunkland.com/blog/using-llms-to-reverse-javascrip... and https://news.ycombinator.com/item?id=34503233


I have tried feeding some of the foundation models obfuscated code from some of the competitions.

People might think that the answers would be in the training data already, but I didn't find that to be the case. At least in my small experiments.

The model's did try to guess what the code does. They would say things like, "It seems to be trying to print some message to the console". I wasn't able to get full solutions.

It's definitely worth more research, not just as a curiosity, but these kinds of problems are good proxies for other tasks and also excellent benchmarks for LLMs particularly.



Thanks! The model is trained only for O0-3, not support for obfuscation. There's still a long way for llm to go.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: