news 2026/5/9 21:44:34

CANN/sip asdMul复数矩阵乘积算子

作者头像

张小明

前端开发工程师

1.2k 24
文章封面图
CANN/sip asdMul复数矩阵乘积算子

asdMul

【免费下载链接】sip本项目是CANN提供的一款高效、可靠的高性能信号处理算子加速库,基于华为Ascend AI处理器,专门为信号处理领域而设计。项目地址: https://gitcode.com/cann/sip

产品支持情况

产品是否支持
Atlas 200I/500 A2 推理产品×
Atlas 推理系列产品×
Atlas 训练系列产品×
Atlas A3 训练系列产品/Atlas A3 推理系列产品
Atlas A2 训练系列产品/Atlas A2 推理系列产品
Ascend 950PR/Ascend 950DT×

功能说明

  • 接口功能:支持向量逐元素乘积(Hadamard)能力,返回一个和输入同样形状大小的复数矩阵。

  • 计算公式:

    $$ result=A \odot\ B =(A){ij}(B){ij} $$

    示例:

    输入“A”为:
    [ [ 1+1i, 1+1i ],
    [ 2+2i, 2+2i ] ]
    输入“B”为:
    [ [ 1+1i, 1+1i ],
    [ 2+2i, 2+2i ] ]
    调用asdMul算子后,输出“result”为:
    [ [ 0+2i, 0+2i ],
    [ 0+8i, 0+8i ] ]

函数原型

AspbStatus asdMul( int n, const void * x, const void * y, const void * z, void * stream, void * workspace = nullptr)

asdMul

  • 参数说明:

    参数名输入/输出描述
    n(int)输入表示输入的元素个数。
    x(void *)输入
    • 表示输入的矩阵,对应公式中的'A'。
    • 数据类型支持COMPLEX32、COMPLEX64
    • 数据格式支持ND。
    • shape为[n]
    y(void *)输入
    • 表示输入的矩阵,对应公式中的'B'。
    • 数据类型支持COMPLEX32、COMPLEX64
    • 数据格式支持ND。
    • shape为[n]
    z(void *)输出
    • 表示输出的矩阵,对应公式中的'result'。
    • 数据类型支持COMPLEX32、COMPLEX64
    • 数据格式支持ND。
    • shape为[n]
    stream(void *)输入npu执行流。
    workspace(void *)输入asdMul算子所需要的workspace。
  • 返回值

    返回状态码,具体参见SiP返回码。

约束说明

  • 输入的元素个数n理论支持[1,9.22e+18]。

调用示例

示例代码如下,该样例旨在提供快速上手、开发和调试算子的最小化实现,其核心目标是使用最精简的代码展示算子的核心功能,而非提供生产级的安全保障。不推荐用户直接将示例代码作为业务代码,若用户将示例代码应用在自身的真实业务场景中且发生了安全问题,则需用户自行承担。

  • mul_complex32
#include <iostream> #include <vector> #include <complex> #include "asdsip.h" #include "acl/acl.h" #include "acl_meta.h" using namespace AsdSip; #define ASD_STATUS_CHECK(err) \ do { \ AsdSip::AspbStatus err_ = (err); \ if (err_ != AsdSip::ErrorType::ACL_SUCCESS) { \ std::cout << "Execute failed." << std::endl; \ exit(-1); \ } \ } while (0) #define CHECK_RET(cond, return_expr) \ do { \ if (!(cond)) { \ return_expr; \ } \ } while (0) #define LOG_PRINT(message, ...) \ do { \ printf(message, ##__VA_ARGS__); \ } while (0) int64_t GetShapeSize(const std::vector<int64_t> &shape) { int64_t shapeSize = 1; for (auto i : shape) { shapeSize *= i; } return shapeSize; } int Init(int32_t deviceId, aclrtStream *stream) { // 固定写法,acl初始化 auto ret = aclInit(nullptr); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclInit failed. ERROR: %d\n", ret); return ret); ret = aclrtSetDevice(deviceId); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtSetDevice failed. ERROR: %d\n", ret); return ret); ret = aclrtCreateStream(stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtCreateStream failed. ERROR: %d\n", ret); return ret); return 0; } template <typename T> int CreateAclTensor(const std::vector<T> &hostData, const std::vector<int64_t> &shape, void **deviceAddr, aclDataType dataType, aclTensor **tensor) { auto size = GetShapeSize(shape) * sizeof(T); // 调用aclrtMalloc申请device侧内存 auto ret = aclrtMalloc(deviceAddr, size, ACL_MEM_MALLOC_HUGE_FIRST); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtMalloc failed. ERROR: %d\n", ret); return ret); // 调用aclrtMemcpy将host侧数据复制到device侧内存上 ret = aclrtMemcpy(*deviceAddr, size, hostData.data(), size, ACL_MEMCPY_HOST_TO_DEVICE); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtMemcpy failed. ERROR: %d\n", ret); return ret); // 计算连续tensor的strides std::vector<int64_t> strides(shape.size(), 1); for (int64_t i = shape.size() - 2; i >= 0; i--) { strides[i] = shape[i + 1] * strides[i + 1]; } // 调用aclCreateTensor接口创建aclTensor *tensor = aclCreateTensor(shape.data(), shape.size(), dataType, strides.data(), 0, aclFormat::ACL_FORMAT_ND, shape.data(), shape.size(), *deviceAddr); return 0; } void printTensor(const std::complex<op::fp16_t> *tensorData, int64_t nums) { for (int64_t i = 0; i < nums; i++) { std::cout << "(" << (float)tensorData[i].real() << "," << (float)tensorData[i].imag() << ")" << " "; } std::cout << std::endl; } int main(int argc, char **argv) { int deviceId = 0; aclrtStream stream; auto ret = Init(deviceId, &stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("Init acl failed. ERROR: %d\n", ret); return ret); int64_t n = 8; int64_t vecSize = n; std::vector<std::complex<op::fp16_t>> tensorInXData; std::vector<std::complex<op::fp16_t>> tensorInYData; tensorInXData.reserve(vecSize); tensorInYData.reserve(vecSize); for (int64_t i = 0; i < vecSize; i++) { tensorInXData.push_back({(op::fp16_t)(9.0f + i), (op::fp16_t)(100.0f + i)}); } for (int64_t i = 0; i < vecSize; i++) { tensorInYData.push_back({(op::fp16_t)(22.0f + i), (op::fp16_t)(33.0f * (i + 1))}); } std::vector<std::complex<op::fp16_t>> tensorOutZData( vecSize, {(op::fp16_t)0.0f, (op::fp16_t)0.0f}); std::cout << "------- input X -------" << std::endl; printTensor(tensorInXData.data(), vecSize); std::cout << "------- input Y -------" << std::endl; printTensor(tensorInYData.data(), vecSize); std::vector<int64_t> xShape = {vecSize}; std::vector<int64_t> yShape = {vecSize}; std::vector<int64_t> zShape = {vecSize}; aclTensor *inputX = nullptr; aclTensor *inputY = nullptr; aclTensor *outputZ = nullptr; void *inputXDeviceAddr = nullptr; void *inputYDeviceAddr = nullptr; void *outputZDeviceAddr = nullptr; ret = CreateAclTensor(tensorInXData, xShape, &inputXDeviceAddr, aclDataType::ACL_COMPLEX32, &inputX); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ret = CreateAclTensor(tensorInYData, yShape, &inputYDeviceAddr, aclDataType::ACL_COMPLEX32, &inputY); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ret = CreateAclTensor(tensorOutZData, zShape, &outputZDeviceAddr, aclDataType::ACL_COMPLEX32, &outputZ); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ASD_STATUS_CHECK(asdMul(n, inputX, inputY, outputZ, stream)); ret = aclrtSynchronizeStream(stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtSynchronizeStream failed. ERROR: %d\n", ret); return ret); ret = aclrtMemcpy(tensorOutZData.data(), vecSize * sizeof(std::complex<op::fp16_t>), outputZDeviceAddr, vecSize * sizeof(std::complex<op::fp16_t>), ACL_MEMCPY_DEVICE_TO_HOST); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("copy z from device to host failed. ERROR: %d\n", ret); return ret); std::cout << "------- output Z -------" << std::endl; printTensor(tensorOutZData.data(), vecSize); std::cout << "Execute successfully." << std::endl; aclDestroyTensor(inputX); aclDestroyTensor(inputY); aclDestroyTensor(outputZ); aclrtFree(inputXDeviceAddr); aclrtFree(inputYDeviceAddr); aclrtFree(outputZDeviceAddr); aclrtDestroyStream(stream); aclrtResetDevice(deviceId); aclFinalize(); return 0; }
  • mul_complex64
#include <iostream> #include <vector> #include <complex> #include "asdsip.h" #include "acl/acl.h" #include "acl_meta.h" using namespace AsdSip; #define ASD_STATUS_CHECK(err) \ do { \ AsdSip::AspbStatus err_ = (err); \ if (err_ != AsdSip::ErrorType::ACL_SUCCESS) { \ std::cout << "Execute failed." << std::endl; \ exit(-1); \ } \ } while (0) #define CHECK_RET(cond, return_expr) \ do { \ if (!(cond)) { \ return_expr; \ } \ } while (0) #define LOG_PRINT(message, ...) \ do { \ printf(message, ##__VA_ARGS__); \ } while (0) int64_t GetShapeSize(const std::vector<int64_t> &shape) { int64_t shapeSize = 1; for (auto i : shape) { shapeSize *= i; } return shapeSize; } int Init(int32_t deviceId, aclrtStream *stream) { // 固定写法,acl初始化 auto ret = aclInit(nullptr); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclInit failed. ERROR: %d\n", ret); return ret); ret = aclrtSetDevice(deviceId); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtSetDevice failed. ERROR: %d\n", ret); return ret); ret = aclrtCreateStream(stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtCreateStream failed. ERROR: %d\n", ret); return ret); return 0; } template <typename T> int CreateAclTensor(const std::vector<T> &hostData, const std::vector<int64_t> &shape, void **deviceAddr, aclDataType dataType, aclTensor **tensor) { auto size = GetShapeSize(shape) * sizeof(T); // 调用aclrtMalloc申请device侧内存 auto ret = aclrtMalloc(deviceAddr, size, ACL_MEM_MALLOC_HUGE_FIRST); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtMalloc failed. ERROR: %d\n", ret); return ret); // 调用aclrtMemcpy将host侧数据复制到device侧内存上 ret = aclrtMemcpy(*deviceAddr, size, hostData.data(), size, ACL_MEMCPY_HOST_TO_DEVICE); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtMemcpy failed. ERROR: %d\n", ret); return ret); // 计算连续tensor的strides std::vector<int64_t> strides(shape.size(), 1); for (int64_t i = shape.size() - 2; i >= 0; i--) { strides[i] = shape[i + 1] * strides[i + 1]; } // 调用aclCreateTensor接口创建aclTensor *tensor = aclCreateTensor(shape.data(), shape.size(), dataType, strides.data(), 0, aclFormat::ACL_FORMAT_ND, shape.data(), shape.size(), *deviceAddr); return 0; } void printTensor(const std::complex<float> *tensorData, int64_t nums) { for (int64_t i = 0; i < nums; i++) { std::cout << tensorData[i] << " "; } std::cout << std::endl; } int main(int argc, char **argv) { int deviceId = 0; aclrtStream stream; auto ret = Init(deviceId, &stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("Init acl failed. ERROR: %d\n", ret); return ret); int64_t n = 8; int64_t vecSize = n; std::vector<std::complex<float>> tensorInXData; std::vector<std::complex<float>> tensorInYData; tensorInXData.reserve(vecSize); tensorInYData.reserve(vecSize); for (int64_t i = 0; i < vecSize; i++) { tensorInXData[i] = {(float)(1.0 + i), (float)(1.0 + i)}; } for (int64_t i = 0; i < vecSize; i++) { tensorInYData[i] = {(float)(2.0 + i), 3.0}; } std::vector<std::complex<float>> tensorOutZData(vecSize, {0.0f, 0.0f}); std::cout << "------- input X -------" << std::endl; printTensor(tensorInXData.data(), vecSize); std::cout << "------- input Y -------" << std::endl; printTensor(tensorInYData.data(), vecSize); std::vector<int64_t> xShape = {vecSize}; std::vector<int64_t> yShape = {vecSize}; std::vector<int64_t> zShape = {vecSize}; aclTensor *inputX = nullptr; aclTensor *inputY = nullptr; aclTensor *outputZ = nullptr; void *inputXDeviceAddr = nullptr; void *inputYDeviceAddr = nullptr; void *outputZDeviceAddr = nullptr; ret = CreateAclTensor(tensorInXData, xShape, &inputXDeviceAddr, aclDataType::ACL_COMPLEX64, &inputX); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ret = CreateAclTensor(tensorInYData, yShape, &inputYDeviceAddr, aclDataType::ACL_COMPLEX64, &inputY); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ret = CreateAclTensor(tensorOutZData, zShape, &outputZDeviceAddr, aclDataType::ACL_COMPLEX64, &outputZ); CHECK_RET(ret == ::ACL_SUCCESS, return ret); ASD_STATUS_CHECK(asdMul(n, inputX, inputY, outputZ, stream)); ret = aclrtSynchronizeStream(stream); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("aclrtSynchronizeStream failed. ERROR: %d\n", ret); return ret); ret = aclrtMemcpy(tensorOutZData.data(), vecSize * sizeof(std::complex<float>), outputZDeviceAddr, vecSize * sizeof(std::complex<float>), ACL_MEMCPY_DEVICE_TO_HOST); CHECK_RET(ret == ::ACL_SUCCESS, LOG_PRINT("copy z from device to host failed. ERROR: %d\n", ret); return ret); std::cout << "------- Output -------" << std::endl; printTensor(tensorOutZData.data(), vecSize); std::cout << "Execute successfully." << std::endl; aclDestroyTensor(inputX); aclDestroyTensor(inputY); aclDestroyTensor(outputZ); aclrtFree(inputXDeviceAddr); aclrtFree(inputYDeviceAddr); aclrtFree(outputZDeviceAddr); aclrtDestroyStream(stream); aclrtResetDevice(deviceId); aclFinalize(); return 0; }

【免费下载链接】sip本项目是CANN提供的一款高效、可靠的高性能信号处理算子加速库,基于华为Ascend AI处理器,专门为信号处理领域而设计。项目地址: https://gitcode.com/cann/sip

创作声明:本文部分内容由AI辅助生成(AIGC),仅供参考

版权声明: 本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若内容造成侵权/违法违规/事实不符,请联系邮箱:809451989@qq.com进行投诉反馈,一经查实,立即删除!
网站建设 2026/5/9 21:41:32

CANN/pyasc max函数API文档

asc.language.basic.max 【免费下载链接】pyasc 本项目为Python用户提供算子编程接口&#xff0c;支持在昇腾AI处理器上加速计算&#xff0c;接口与Ascend C一一对应并遵守Python原生语法。 项目地址: https://gitcode.com/cann/pyasc asc.language.basic.max(dst: Loca…

作者头像 李华
网站建设 2026/5/9 21:39:44

JSON可视化利器:用图形思维解析复杂数据结构

1. 项目概述&#xff1a;从JSON到可视化图谱的“降维打击”如果你也经常和JSON数据打交道&#xff0c;尤其是那种嵌套了七八层、动辄几千行的配置文件或者API响应&#xff0c;那你一定懂我的痛苦。盯着密密麻麻的括号和引号&#xff0c;想理清一个对象里到底有什么、谁引用了谁…

作者头像 李华
网站建设 2026/5/9 21:36:59

用Pluto SDR和MATLAB复现经典:四种模拟波形传输实测与波形畸变全解析

用Pluto SDR和MATLAB复现经典&#xff1a;四种模拟波形传输实测与波形畸变全解析 在通信工程实验室里&#xff0c;我们常常需要验证教科书上的理论——那些关于信号完整性、采样定理和滤波器效应的数学推导&#xff0c;是否真的能在实际硬件中重现&#xff1f;Pluto SDR作为一…

作者头像 李华
网站建设 2026/5/9 21:36:56

硬件木马检测:跨尺度持久性分析方法解析

1. 硬件木马检测的挑战与现状硬件木马&#xff08;Hardware Trojans, HTs&#xff09;已成为集成电路安全领域最严峻的威胁之一。与软件层面的恶意代码不同&#xff0c;硬件木马直接植入芯片设计或制造环节&#xff0c;具有物理层面的隐蔽性和持久性。其中&#xff0c;始终激活…

作者头像 李华