Visualizer for neural network, deep learning, and machine learning models
-
Updated
Jul 25, 2021 - JavaScript
{{ message }}
Visualizer for neural network, deep learning, and machine learning models
YOLOv5
ncnn is a high-performance neural network inference framework optimized for the mobile platform
YOLOv3 in PyTorch > ONNX > CoreML > TFLite
Set up deep learning environment in a single command line.
MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g. model conversion and visualization. Convert models between Caffe, Keras, MXNet, Tensorflow, CNTK, PyTorch Onnx and CoreML.
Describe the bug
when axis has duplicate value , onnxruntime compute result is all same value ,which is different with expect of tensorflow
Urgency
2020.11.18
System information
Linux Ubuntu 16.04
Expected behavior
When there are duplicate values, the duplicate can be removed. j
Deep Learning Visualization Toolkit(『飞桨』深度学习可视化工具 )
A collection of pre-trained, state-of-the-art models in the ONNX format
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
PyTorch ,ONNX and TensorRT implementation of YOLOv4
请问可以直接training tmfile出来吗? 因为tengine-convert-tool covert 会有error
tengine-lite library version: 1.4-dev
Get input tensor failed

或是有例子能training出下面tmfile 呢?
, High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Pretrained EfficientNet, EfficientNet-Lite, MixNet, MobileNetV3 / V2, MNASNet A1 and B1, FBNet, Single-Path NAS
nGraph has moved to OpenVINO
Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning models. Advbox give a command line tool to generate adversarial examples with Zero-Coding.
Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference
Silero Models: pre-trained speech-to-text, text-to-speech models and benchmarks made embarrassingly simple
I am trying to convert a custom pytorch model to tensorflow, I am abe to convert pytorch to onnx but converting onnx to tensorflow gives issue.
The code snippets are as follows-
net = custom pytorch model
net.load_state_dict("pre-trained model")
dummyInput = np.random.uniform(0,1,(1,8,3,256,256))
dummyInput = Variable(torch.FloatTensor(dummyInput))
torch.onnx.export(ne
A repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
'max_request_size' seems to refer to bytes, not mb.
PyTorch to Keras model convertor
Samples and Tools for Windows ML.
Add a description, image, and links to the onnx topic page so that developers can more easily learn about it.
To associate your repository with the onnx topic, visit your repo's landing page and select "manage topics."
New Operator
Describe the operator
Why is this operator necessary? What does it accomplish?
This is a frequently used operator in
tensorflow/kerasCan this operator be constructed using existing onnx operators?
If so, why not add it as a function?
I don't know.
Is this operator used by any model currently? Which one?
Are you willing to contribute it?