FlashAttention: Fast Transformer training with long sequences

FlashAttention: Fast Transformer training with long sequences

‍Transformers have grown deeper and wider, but training them on long sequences remains difficult. The attention layer at their heart is the compute and memory bottleneck: doubling the sequence length would quadruple the runtime and memory requirements. FlashAttention is a new algorithm to speed up attention and reduce its memory footprint—without any approximation. Since FlashAttention was released 6 months ago, it has been adopted by many organizations and research labs to speed up their training & inference (see this page for a partial list)....

Published in adept.ai · by Tri Dao · 7 min read · July 13, 2023
Announcing our Series B

Announcing our Series B

We’ve raised $350M in new funding as part of our Series B led by General Catalyst and co-led by Spark Capital, with additional participation from existing investors, new financial partners, and some of the most iconic companies in tech. The recent pace of progress in generative AI has been nothing short of astounding. For years, we’ve heard talk of AI’s enormous potential, but over the past few months, a new wave of advancements has catapulted the field into the mainstream....

Published in adept.ai · by Adept Team · 3 min read · July 12, 2023
Introducing Adept

Introducing Adept

Adept is an ML research and product lab building general intelligence by enabling people and computers to work together creatively. We believe that AI systems should be built with users at the center — our vision is one where machines work together with people in the driver’s seat: discovering new solutions, enabling more informed decisions, and giving us more time for the work we love. Machine learning has seen more progress in the last five years than in the prior 60....

Published in adept.ai · by David Luan · 5 min read · July 8, 2023