Tencent/ncnn
ncnn is a high-performance neural network inference framework optimized for the mobile platform
HunyuanVideo, ncnn, and Tencent AI Lab projects
ncnn is a high-performance neural network inference framework optimized for the mobile platform
A full-stack AI Red Teaming platform securing AI ecosystems via OpenClaw Security Scan, Agent Scan, Skills Scan, MCP scan, AI Infra scan and LLM jailbreak evaluation.
[CVPR2026]🚀🚀🚀Official code for the paper "YOLO-Master: MOE-Accelerated with Specialized Transformers for Enhanced Real-time Detection." *(YOLO = You Only Look Once)* 🔥🔥🔥
High Performance LLM Inference Operator Library
Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.
LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.
Many studies have shown that the performance on deep learning is significantly affected by volume of training data. The MedicalNet project provides a series of 3D-ResNet pre-trained models and relative code.
Deep learning-based Video Quality Assessment
TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework.
FeatherCNN is a high performance inference engine for convolutional neural networks.
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
WeDLM: The fastest diffusion language model with standard causal attention and native KV cache compatibility, delivering real speedups over vLLM-optimized baselines.
Largest multi-label image database; ResNet-101 model; 80.73% top-1 acc on ImageNet
An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
A library for high performance deep learning inference on NVIDIA GPUs.
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.