Flash attention for udop large
#8
by
Vasanth2002
- opened
I am trying to increase the inference speed of UDOP but nothing works. I want to use flashattention for udop but it is not supported
Feel free to open an issue on Github regarding adding support