Flash attention for udop large

#8
by Vasanth2002 - opened

I am trying to increase the inference speed of UDOP but nothing works. I want to use flashattention for udop but it is not supported

Feel free to open an issue on Github regarding adding support

Sign up or log in to comment