ncnn is a high-performance neural network inference framework optimized for the mobile platform
-
Updated
Jun 12, 2024 - C++
ncnn is a high-performance neural network inference framework optimized for the mobile platform
OpenGL Mathematics (GLM)
Parsing gigabytes of JSON per second : used by Facebook/Meta Velox, the Node.js runtime, ClickHouse, WatermelonDB, Apache Doris, Milvus, StarRocks
The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologies.
C++ image processing and machine learning library with using of SIMD: SSE, AVX, AVX-512, AMX for x86/x64, VMX(Altivec) and VSX(Power7) for PowerPC, NEON for ARM.
Performance-portable, length-agnostic SIMD with runtime dispatch
Open source c++ skeletal animation library and toolset
C++ wrappers for SIMD intrinsics and parallelized, optimized mathematical functions (SSE, AVX, AVX512, NEON, SVE))
Fast, modern C++ DSP framework, FFT, Sample Rate Conversion, FIR/IIR/Biquad Filters (SSE, AVX, AVX-512, ARM NEON)
A translator from Intel SSE intrinsics to Arm/Aarch64 NEON implementation
The Vector Optimized Library of Kernels
SeqAn's official repository.
SIMD Vector Classes for C++
Fast and exact implementation of the C++ from_chars functions for number types: 4x to 10x faster than strtod, part of GCC 12 and WebKit/Safari
Fast Open-Source Search & Clustering engine × for Vectors & 🔜 Strings × in C++, C, Python, JavaScript, Rust, Java, Objective-C, Swift, C#, GoLang, and Wolfram 🔍
C++11 multiplatform utility library
Add a description, image, and links to the simd topic page so that developers can more easily learn about it.
To associate your repository with the simd topic, visit your repo's landing page and select "manage topics."