Commit
·
686a471
1
Parent(s):
ffd1a8e
don't need IPython
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ import torch
|
|
5 |
from tqdm.auto import tqdm
|
6 |
from PIL import Image
|
7 |
import gradio as gr
|
8 |
-
from IPython.display import display
|
9 |
|
10 |
tokenizer = CLIPTokenizer.from_pretrained("openai/clip-vit-large-patch14", torch_dtype=torch.float16)
|
11 |
text_encoder = CLIPTextModel.from_pretrained("openai/clip-vit-large-patch14", torch_dtype=torch.float16).to("cuda")
|
|
|
5 |
from tqdm.auto import tqdm
|
6 |
from PIL import Image
|
7 |
import gradio as gr
|
8 |
+
#from IPython.display import display
|
9 |
|
10 |
tokenizer = CLIPTokenizer.from_pretrained("openai/clip-vit-large-patch14", torch_dtype=torch.float16)
|
11 |
text_encoder = CLIPTextModel.from_pretrained("openai/clip-vit-large-patch14", torch_dtype=torch.float16).to("cuda")
|