from langchain.chains import LLMChain from langchain.llms import OpenAI import gradio as gr from dotenv import load_dotenv from langchain.prompts.prompt import PromptTemplate load_dotenv() def generate_questions(resume,role='',experience=''): _PROMPT_TEMPLATE = """ this is the resume of user: {resume_details} here is the role he want to join in : {role} Based on the following experience: {experience} What are your interview questions for the given user resume and role he want to join in with that experience? generate no of questions = {questions}! """ PROMPT = PromptTemplate(input_variables=["resume_details", "role", "experience",'questions'], template=_PROMPT_TEMPLATE) llm1 = OpenAI(model_name="text-davinci-003", temperature=0) chain = LLMChain(llm=llm1, prompt=PROMPT) prompt = chain.predict_and_parse(resume_details= resume, role= role, experience= experience, questions=10) return prompt.split('\n') k=gr.Interface( fn=generate_questions, inputs=['textbox','text','text'], outputs=['text'] ) k.launch()