LLMs for Everyone: Running the HuggingFace Text Generation Inference in Google Colab
Author:Murphy | View: 25931 | Time: 2025-03-22 23:18:27

In the first part of the story, we used a free Google Colab instance to run a Mistral-7B model and extract information using the FAISS (Facebook AI Similarity Search) database. In the second part of the story, we used a LLaMA-13B model and a LangChain library to make a chat with text summarization and other features. In this part, I will show how to use a HuggingFace