[HN Gopher] LLM in a Flash: Efficient Large Language Model Infer...
___________________________________________________________________
LLM in a Flash: Efficient Large Language Model Inference with
Limited Memory
Author : keep_reading
Score : 6 points
Date : 2023-12-21 22:31 UTC (29 minutes ago)
(HTM) web link (arxiv.org)
(TXT) w3m dump (arxiv.org)
___________________________________________________________________
(page generated 2023-12-21 23:00 UTC)