> LLMs with coding abilities could be employed to create sophisticated malware with unprecedented ease.<p>If that is possible then shouldn't it also be possible to ask the AI to find and code remediation to the vulnerabilities it found?<p>So AI could be used to find all possible code-vulnerabilities and then how to neutralize those? This would advance software security in general.<p>In other words AI could be used like a microscope discovering tiny defects in our software which are not visible to the naked eye. Like a microscope that detects viruses and thus allows us to guard against them. Like a COVID-test.
Does anyone know what happens if you do transfer learning in addition to scaling? It feels like people used to use transfer learning in lieu of scaling and I haven't wrapped my head around how they work together.
One important but that is often left out is that ChatGPT is not the first model to come out using RLHF to train LLMs.<p>As is typical in the AI field, Deepmind was key in the development of the process. Deepmind 's Sparrow came out just before ChatGPT (regarding language modeling with RLHF), and much of the RLHF work was explored in their robotics/agent exploration work just prior to application in language.<p>OpenAI was integral in PPO, but it's important to know and understand it wasn't ChatGPT or OpenAI that is solely leading these advancements.