Skip to content

Commit

Permalink
Upgrade gradio server version.
Browse files Browse the repository at this point in the history
  • Loading branch information
haotian-liu committed Feb 3, 2024
1 parent 5204cb4 commit f04e1cc
Show file tree
Hide file tree
Showing 2 changed files with 29 additions and 29 deletions.
54 changes: 27 additions & 27 deletions llava/serve/gradio_web_server.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,9 +19,9 @@

headers = {"User-Agent": "LLaVA Client"}

no_change_btn = gr.Button.update()
enable_btn = gr.Button.update(interactive=True)
disable_btn = gr.Button.update(interactive=False)
no_change_btn = gr.Button()
enable_btn = gr.Button(interactive=True)
disable_btn = gr.Button(interactive=False)

priority = {
"vicuna-13b": "aaaaaaa",
Expand Down Expand Up @@ -58,12 +58,11 @@ def get_model_list():
def load_demo(url_params, request: gr.Request):
logger.info(f"load_demo. ip: {request.client.host}. params: {url_params}")

dropdown_update = gr.Dropdown.update(visible=True)
dropdown_update = gr.Dropdown(visible=True)
if "model" in url_params:
model = url_params["model"]
if model in models:
dropdown_update = gr.Dropdown.update(
value=model, visible=True)
dropdown_update = gr.Dropdown(value=model, visible=True)

state = default_conversation.copy()
return state, dropdown_update
Expand All @@ -73,7 +72,7 @@ def load_demo_refresh_model_list(request: gr.Request):
logger.info(f"load_demo. ip: {request.client.host}")
models = get_model_list()
state = default_conversation.copy()
dropdown_update = gr.Dropdown.update(
dropdown_update = gr.Dropdown(
choices=models,
value=models[0] if len(models) > 0 else ""
)
Expand Down Expand Up @@ -314,7 +313,7 @@ def http_bot(state, model_selector, temperature, top_p, max_new_tokens, request:
"""

def build_demo(embed_mode, cur_dir=None):
def build_demo(embed_mode, cur_dir=None, concurrency_count=10):
textbox = gr.Textbox(show_label=False, placeholder="Enter text and press ENTER", container=False)
with gr.Blocks(title="LLaVA", theme=gr.themes.Default(), css=block_css) as demo:
state = gr.State()
Expand Down Expand Up @@ -350,7 +349,12 @@ def build_demo(embed_mode, cur_dir=None):
max_output_tokens = gr.Slider(minimum=0, maximum=1024, value=512, step=64, interactive=True, label="Max output tokens",)

with gr.Column(scale=8):
chatbot = gr.Chatbot(elem_id="chatbot", label="LLaVA Chatbot", height=550)
chatbot = gr.Chatbot(
elem_id="chatbot",
label="LLaVA Chatbot",
height=650,
layout="panel",
)
with gr.Row():
with gr.Column(scale=8):
textbox.render()
Expand All @@ -374,31 +378,28 @@ def build_demo(embed_mode, cur_dir=None):
upvote_btn.click(
upvote_last_response,
[state, model_selector],
[textbox, upvote_btn, downvote_btn, flag_btn],
queue=False
[textbox, upvote_btn, downvote_btn, flag_btn]
)
downvote_btn.click(
downvote_last_response,
[state, model_selector],
[textbox, upvote_btn, downvote_btn, flag_btn],
queue=False
[textbox, upvote_btn, downvote_btn, flag_btn]
)
flag_btn.click(
flag_last_response,
[state, model_selector],
[textbox, upvote_btn, downvote_btn, flag_btn],
queue=False
[textbox, upvote_btn, downvote_btn, flag_btn]
)

regenerate_btn.click(
regenerate,
[state, image_process_mode],
[state, chatbot, textbox, imagebox] + btn_list,
queue=False
[state, chatbot, textbox, imagebox] + btn_list
).then(
http_bot,
[state, model_selector, temperature, top_p, max_output_tokens],
[state, chatbot] + btn_list
[state, chatbot] + btn_list,
concurrency_limit=concurrency_count
)

clear_btn.click(
Expand All @@ -416,27 +417,27 @@ def build_demo(embed_mode, cur_dir=None):
).then(
http_bot,
[state, model_selector, temperature, top_p, max_output_tokens],
[state, chatbot] + btn_list
[state, chatbot] + btn_list,
concurrency_limit=concurrency_count
)

submit_btn.click(
add_text,
[state, textbox, imagebox, image_process_mode],
[state, chatbot, textbox, imagebox] + btn_list,
queue=False
[state, chatbot, textbox, imagebox] + btn_list
).then(
http_bot,
[state, model_selector, temperature, top_p, max_output_tokens],
[state, chatbot] + btn_list
[state, chatbot] + btn_list,
concurrency_limit=concurrency_count
)

if args.model_list_mode == "once":
demo.load(
load_demo,
[url_params],
[state, model_selector],
_js=get_window_url_params,
queue=False
_js=get_window_url_params
)
elif args.model_list_mode == "reload":
demo.load(
Expand All @@ -456,7 +457,7 @@ def build_demo(embed_mode, cur_dir=None):
parser.add_argument("--host", type=str, default="0.0.0.0")
parser.add_argument("--port", type=int)
parser.add_argument("--controller-url", type=str, default="http://localhost:21001")
parser.add_argument("--concurrency-count", type=int, default=10)
parser.add_argument("--concurrency-count", type=int, default=16)
parser.add_argument("--model-list-mode", type=str, default="once",
choices=["once", "reload"])
parser.add_argument("--share", action="store_true")
Expand All @@ -468,9 +469,8 @@ def build_demo(embed_mode, cur_dir=None):
models = get_model_list()

logger.info(args)
demo = build_demo(args.embed)
demo = build_demo(args.embed, concurrency_count=args.concurrency_count)
demo.queue(
concurrency_count=args.concurrency_count,
api_open=False
).launch(
server_name=args.host,
Expand Down
4 changes: 2 additions & 2 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"

[project]
name = "llava"
version = "1.2.1"
version = "1.2.1.post2"
description = "Towards GPT-4 like large language and visual assistant."
readme = "README.md"
requires-python = ">=3.8"
Expand All @@ -17,7 +17,7 @@ dependencies = [
"transformers==4.36.2", "tokenizers==0.15.0", "sentencepiece==0.1.99", "shortuuid",
"accelerate==0.21.0", "peft==0.4.0", "bitsandbytes==0.41.0",
"pydantic<2,>=1", "markdown2[all]", "numpy", "scikit-learn==1.2.2",
"gradio==3.35.2", "gradio_client==0.2.9",
"gradio==4.16.0", "gradio_client==0.8.1",
"requests", "httpx==0.24.0", "uvicorn", "fastapi",
"einops==0.6.1", "einops-exts==0.0.4", "timm==0.6.13",
]
Expand Down

0 comments on commit f04e1cc

Please sign in to comment.