flystem-usls/examples/blip
Jamjamjon 51b75e9a21
Add YOLOPv2 & Face-Parsing model (#3)
* Add YOLOP and face parsing model
2024-04-14 15:15:59 +08:00
..
README.md Add YOLOPv2 & Face-Parsing model (#3) 2024-04-14 15:15:59 +08:00
main.rs Add RTMO model for keypoint detection (#2) 2024-04-08 20:45:38 +08:00

README.md

This demo shows how to use BLIP to do conditional or unconditional image captioning.

Quick Start

cargo run -r --example blip

BLIP ONNX Model

Results

[Unconditional image captioning]: a group of people walking around a bus
[Conditional image captioning]: three man walking in front of a bus

TODO

  • VQA
  • Retrival
  • TensorRT support for textual model