TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

FlashAttention-2, 2x faster than FlashAttention

82 pointsby machdiamondsalmost 2 years ago

4 comments

whimsicalismalmost 2 years ago
Does anyone have resources for a good way to get started with this sort of modern GPU systems work?
评论 #36763272 未加载
评论 #36764179 未加载
评论 #36763025 未加载
ternausalmost 2 years ago
I would be very greatfull to see how one can leverage it not for LLMs but for Stable Diffusion models
评论 #36763410 未加载
lucidrainsalmost 2 years ago
huge! thank you Tri!
评论 #36762663 未加载
hunkinsalmost 2 years ago
Excellent work by the team on this. Excited to see this applied to the diffusers library.