I found there is no flash attention in the inference kernel,what is the reason,is the flash attention slow in this case?