Submitted by floppy_llama t3_1266d02 in MachineLearning
saintshing t1_jeaowjz wrote
Reply to comment by A_Light_Spark in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
I almost missed it too. There are too many new results.
The most crazy thing is it is all done by one person when the big techs all work on transformer models.
Viewing a single comment thread. View all comments