""" This module loads the LLM from the local file system Modify this file if you need to download some other model from Hugging Face or OpenAI/ChatGPT """ # from langchain.llms import CTransformers # from langchain_openai import OpenAI from langchain_groq import ChatGroq from langchain_google_genai import ChatGoogleGenerativeAI from dotenv import load_dotenv import os model_name = 'gemma2-9b-it' def load_llm(model_name=model_name): # llm = ChatGroq( # temperature=0, # model=model_name, # ) llm = ChatGoogleGenerativeAI( model="gemini-1.5-flash", temperature=0, max_tokens=None, timeout=None, max_retries=2, ) return llm if __name__ == '__main__': load_dotenv() llm = load_llm() result = llm.invoke("Provide a short answer: What is machine learning?") print(result.content)