Post AXnLeaMTZ7srba9APo by m8ta@fediscience.org
(DIR) More posts by m8ta@fediscience.org
(DIR) Post #AXnLeaMTZ7srba9APo by m8ta@fediscience.org
2023-07-17T19:25:32Z
0 likes, 1 repeats
Remarkable: LLama-Adapter is about as strong as GPT-4 chain-of-thought on the Science QA test set, following only ~ 1 hour of fine-tuning. The model works by concatenating ~1M trainable (but default zero-gated) tokens to the upper layers of Llama, then fine-tuning on the 52k instructions from Alpaca. (Science QA = Multi-modal elementary and high school curricula)https://arxiv.org/abs/2303.16199