vikhyatk commited on
Commit
5442928
·
verified ·
1 Parent(s): e148304

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +21 -85
app.py CHANGED
@@ -243,14 +243,9 @@ with gr.Blocks(title="moondream vl (new)", css=css, js=js) as demo:
243
 
244
  gr.Markdown(
245
  """
246
- # 🌔 test space, pls ignore
247
  """
248
  )
249
- mode_radio = gr.Radio(
250
- ["Caption", "Query", "Detect", "Point", "Localized"],
251
- show_label=False,
252
- value=lambda: "Caption",
253
- )
254
 
255
  input_image = gr.State(None)
256
 
@@ -259,85 +254,26 @@ with gr.Blocks(title="moondream vl (new)", css=css, js=js) as demo:
259
 
260
  @gr.render(inputs=[mode_radio])
261
  def show_inputs(mode):
262
- if mode == "Query":
263
- with gr.Group():
264
- with gr.Row():
265
- prompt = gr.Textbox(
266
- label="Input",
267
- value="How many people are in this image?",
268
- scale=4,
269
- )
270
- submit = gr.Button("Submit")
271
- reasoning = gr.Checkbox(label="Enable reasoning")
272
- img = gr.Image(type="pil", label="Upload an Image")
273
- submit.click(answer_question, [img, prompt, reasoning], [output, thought])
274
- prompt.submit(answer_question, [img, prompt, reasoning], [output, thought])
275
- reasoning.change(answer_question, [img, prompt, reasoning], [output, thought])
276
- img.change(answer_question, [img, prompt, reasoning], [output, thought])
277
- img.change(lambda img: img, [img], [input_image])
278
- elif mode == "Caption":
279
- with gr.Group():
280
- with gr.Row():
281
- caption_mode = gr.Radio(
282
- ["Short", "Normal", "Long"],
283
- label="Caption Length",
284
- value=lambda: "Normal",
285
- scale=4,
286
- )
287
- submit = gr.Button("Submit")
288
- img = gr.Image(type="pil", label="Upload an Image")
289
- submit.click(caption, [img, caption_mode], output)
290
- img.change(caption, [img, caption_mode], output)
291
- elif mode == "Detect":
292
- with gr.Group():
293
- with gr.Row():
294
- prompt = gr.Textbox(
295
- label="Object",
296
- value="Cat",
297
- scale=4,
298
- )
299
- submit = gr.Button("Submit")
300
- img = gr.Image(type="pil", label="Upload an Image")
301
- eos_bias = gr.Textbox(label="EOS Bias", value="0")
302
- submit.click(detect, [img, prompt, eos_bias], [thought, ann])
303
- prompt.submit(detect, [img, prompt, eos_bias], [thought, ann])
304
- img.change(detect, [img, prompt, eos_bias], [thought, ann])
305
- elif mode == "Point":
306
- with gr.Group():
307
- with gr.Row():
308
- prompt = gr.Textbox(
309
- label="Object",
310
- value="Cat",
311
- scale=4,
312
- )
313
- submit = gr.Button("Submit")
314
- img = gr.Image(type="pil", label="Upload an Image")
315
- submit.click(point, [img, prompt], [thought, ann])
316
- prompt.submit(point, [img, prompt], [thought, ann])
317
- img.change(point, [img, prompt], [thought, ann])
318
- elif mode == "Localized":
319
- with gr.Group():
320
- with gr.Row():
321
- prompt = gr.Textbox(
322
- label="Input",
323
- value="What is this?",
324
- scale=4,
325
- )
326
- submit = gr.Button("Submit")
327
- img = gr.Image(type="pil", label="Upload an Image")
328
- x_slider = gr.Slider(label="x", minimum=0, maximum=1)
329
- y_slider = gr.Slider(label="y", minimum=0, maximum=1)
330
- submit.click(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
331
- prompt.submit(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
332
- x_slider.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
333
- y_slider.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
334
- img.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
335
- def select_handler(image, evt: gr.SelectData):
336
- w, h = img.size
337
- return [evt.index[0] / w, evt.index[1] / h]
338
- img.select(select_handler, img, [x_slider, y_slider])
339
- else:
340
- gr.Markdown("Coming soon!")
341
 
342
  with gr.Column():
343
  thought = gr.HighlightedText(
 
243
 
244
  gr.Markdown(
245
  """
246
+ # 🌔 grounded visual question asnwering
247
  """
248
  )
 
 
 
 
 
249
 
250
  input_image = gr.State(None)
251
 
 
254
 
255
  @gr.render(inputs=[mode_radio])
256
  def show_inputs(mode):
257
+ with gr.Group():
258
+ with gr.Row():
259
+ prompt = gr.Textbox(
260
+ label="Input",
261
+ value="What is this?",
262
+ scale=4,
263
+ )
264
+ submit = gr.Button("Submit")
265
+ img = gr.Image(type="pil", label="Upload an Image")
266
+ x_slider = gr.Slider(label="x", minimum=0, maximum=1)
267
+ y_slider = gr.Slider(label="y", minimum=0, maximum=1)
268
+ submit.click(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
269
+ prompt.submit(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
270
+ x_slider.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
271
+ y_slider.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
272
+ img.change(localized_query, [img, x_slider, y_slider, prompt], [output, thought, ann])
273
+ def select_handler(image, evt: gr.SelectData):
274
+ w, h = img.size
275
+ return [evt.index[0] / w, evt.index[1] / h]
276
+ img.select(select_handler, img, [x_slider, y_slider])
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
277
 
278
  with gr.Column():
279
  thought = gr.HighlightedText(