Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

鲲鹏920环境Benchmark跑量化模型crash #2856

Open
zhenjing opened this issue Apr 30, 2024 · 2 comments
Open

鲲鹏920环境Benchmark跑量化模型crash #2856

zhenjing opened this issue Apr 30, 2024 · 2 comments
Labels
bug Something isn't working

Comments

@zhenjing
Copy link

平台(如果交叉编译请再附上交叉编译目标平台):

-- >>>>>>>>>>>>>
-- MNN BUILD INFO:
-- System: Linux
-- Processor: aarch64
-- Version: 2.8.4
-- Metal: OFF
-- OpenCL: OFF
-- OpenGL: OFF
-- Vulkan: OFF
-- ARM82: ON
-- oneDNN: OFF
-- TensorRT: OFF
-- CoreML: OFF
-- NNAPI: OFF
-- CUDA: OFF
-- OpenMP: ON
-- BF16: OFF
-- ThreadPool: OFF
-- Hidden: TRUE
-- Build Path: /home/zhenjing/MNN-2.8.4/build
-- CUDA PROFILE: OFF
-- WIN_USE_ASM:
-- Enabling AArch64 Assemblies
-- Enable INT8 SDOT
-- [*] Checking OpenMP
-- Found OpenMP_C: -fopenmp
-- Found OpenMP_CXX: -fopenmp
-- Found OpenMP: TRUE
-- Configuring done
-- Generating done

Github版本: 2.8.4

./benchmark.out models/ 50 0 0 1 2 0 1 1
MNN benchmark
Forward type: CPU thread=1 precision=2 sparsity=0 sparseBlockOC=1 testQuantizedModel=1
--------> Benchmarking... loop = 50, warmup = 0
[-INFO-]: precision=2, use fp16 inference if your device supports and open MNN_ARM82=ON.
[-INFO-]: Auto set sparsity=0 when test quantized model in benchmark...
Auto set sparsity=0 when test quantized model in benchmark...
The device support i8sdot:1, support fp16:1, support i8mm: 1
[ - ] MobileNetV2_224.mnn max = 14.354 ms min = 13.765 ms avg = 13.997 ms
Illegal instruction (core dumped)

@zhenjing
Copy link
Author

./benchmark.out models/ 50 0 0 1 0 0 1 1
MNN benchmark
Forward type: CPU thread=1 precision=0 sparsity=0 sparseBlockOC=1 testQuantizedModel=1
--------> Benchmarking... loop = 50, warmup = 0
[-INFO-]: precision!=2, use fp32 inference.
[-INFO-]: Auto set sparsity=0 when test quantized model in benchmark...
Auto set sparsity=0 when test quantized model in benchmark...
The device support i8sdot:1, support fp16:1, support i8mm: 1
[ - ] MobileNetV2_224.mnn max = 29.129 ms min = 28.343 ms avg = 28.640 ms
Illegal instruction (core dumped)

@jxt1234
Copy link
Collaborator

jxt1234 commented May 8, 2024

看上去是判定 i8mm feature 出错了: support i8mm: 1
可以自行修改 cpuruntime.cpp 试试

@jxt1234 jxt1234 added the bug Something isn't working label May 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants