- Load a text generation model from Hugging Face's Transformers library (specifically, the TinyLlama model)
- Tokenize a given text input using the model's tokenizer.
- Convert the loaded model to OpenVINO IR format for inference on Intel hardware.
- Perform inference with the converted OpenVINO model.
- Print the decoded output from the inference.
Hugging Face Model Hub with OpenVINO