Following up on our conversation on quantized models on smartphones, Stefano Fiorucci wrote a post about how to run a small language model on a smartphone: .
This involves either Layla Lite App or Temux. One commenter recommended LLM Farm on iPhone.