from grid.model.perception.vlm.video_llava import VideoLLaVAcar = AirGenCar()# We will be capturing an image from the AirGen simulator # and run model inference on it.img = car.getImage("front_center","rgb").datamodel = VideoLLaVA(use_local =True)result = model.run(rgbimage=img, prompt=<prompt>)print(result)
The VideoLLaVA class provides a wrapper for the Video-LLaVA model, which answers questions
about visual media (videos).
from grid.model.perception.vlm.video_llava import VideoLLaVAcar = AirGenCar()# We will be capturing an image from the AirGen simulator # and run model inference on it.img = car.getImage("front_center","rgb").datamodel = VideoLLaVA(use_local =True)result = model.run(rgbimage=img, prompt=<prompt>)print(result)
This code is licensed under the Apache 2.0 License.
from grid.model.perception.vlm.video_llava import VideoLLaVAcar = AirGenCar()# We will be capturing an image from the AirGen simulator # and run model inference on it.img = car.getImage("front_center","rgb").datamodel = VideoLLaVA(use_local =True)result = model.run(rgbimage=img, prompt=<prompt>)print(result)