If you look at the model documentation on HuggingFace, for 'TinyLlama/TinyLlama-1.1B-Chat-v0.6', in the 'Inference Providers' section in the right side, it will be written as 'This model isn't deployed by any Inference Provider.' - meaning that that you cannot use the model through the free, serverless Inference API provided by Hugging Face. In this case, you must use the local Inference, i.e, download the model.
This is confusing because the LangChain's HuggingFaceEndpoint is primarily designed to work with Inference APIs. This is frustrating but we have to get used to these kind of things as beginners or self-learners.