Skip to content

app.py #17

@ghost

Description

app.py loads every time from new the model intro ram, by press Get Answer, best way will be load model then just Get Answer and model stays in ram if project become big like gpt-j6B model must stay loaded in ram then Get Answer
#-----------------------------------------------------------------
#-something like this but model must stay in ram for every answer

import torch
import streamlit as st
from transformers import GPT2Tokenizer, GPT2LMHeadModel

tokenizer = GPT2Tokenizer.from_pretrained("gagan3012/project-code-py-small" , low_ram_method = True, use_cache=True, use_fast=True, low_cpu_mem_usage = True)
model = GPT2LMHeadModel.from_pretrained("gagan3012/project-code-py-small")

#-step to keep tokenizer and model loaded at this point

st.set_page_config(
page_title="AI Leetcode",
layout="wide",
initial_sidebar_state="expanded", )
...

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions