Commits


Viswanath Boga authored and GitHub committed 6dee9b9d2d2
attention fusion kernel refactoring (#8432) * attention fusion kernel refactored * consider the case of none in add_qk * variabled added to check for pre-pack weights * added a comment to PrePack() * Optimized prepack and try to free the weights * making comment sound better * fixing a bug with optimizer.py * commented out changes to be done * removed comments * make the private fn() private * fix build * making clean up fn static * backed out optimizer tool change, needs more looking into