[Day 71] Backprop, GELU, Tricking ChatGPT, and Stealing part of an LLM
Hello :) Today is Day 71! A quick summary of today: Finished up the manual backprop code to make it more clear Read some research papers Gaussian Error Linear Units (GELUs) Using Hallucinations to Bypass RLHF Filters Stealing Part of a Production Language Model Firstly, the manual backprop code cleanup. To finish it up (for now atleast), I decided to get more data. Original was ~32000 names, on kaggle I found with around ~90000 names. As for the architecture - I settled on the original input > flatten > batch norm > activation > output, I added more even up to 2 hidden layers and batch norm for each, but the training time got longer, and the results were not even better than the simple original one. Even though I stayed with the original model structure, I tried to modify embedding size, context length, hidden layers, just to play around with them and see the result. At the end the generated names look decent. But I was not able to achieve a much better re...