mirror of
https://github.com/deepseek-ai/FlashMLA
synced 2025-06-26 18:15:54 +00:00
add missing copyright
This commit is contained in:
parent
3e123bc93c
commit
b31bfe72a8
@ -1,4 +1,7 @@
|
|||||||
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/csrc/flash_attn/flash_api.cpp
|
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/csrc/flash_attn/flash_api.cpp
|
||||||
|
/******************************************************************************
|
||||||
|
* Copyright (c) 2024, Tri Dao.
|
||||||
|
******************************************************************************/
|
||||||
|
|
||||||
#include <torch/python.h>
|
#include <torch/python.h>
|
||||||
#include <torch/nn/functional.h>
|
#include <torch/nn/functional.h>
|
||||||
|
@ -1,4 +1,7 @@
|
|||||||
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/csrc/flash_attn/src/softmax.h
|
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/csrc/flash_attn/src/softmax.h
|
||||||
|
/******************************************************************************
|
||||||
|
* Copyright (c) 2024, Tri Dao.
|
||||||
|
******************************************************************************/
|
||||||
|
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
|
@ -1,4 +1,7 @@
|
|||||||
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/hopper/utils.h
|
// Adapted from https://github.com/Dao-AILab/flash-attention/blob/main/hopper/utils.h
|
||||||
|
/******************************************************************************
|
||||||
|
* Copyright (c) 2024, Tri Dao.
|
||||||
|
******************************************************************************/
|
||||||
|
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user