YOLOv5
-
Updated
Nov 4, 2021 - Python
{{ message }}
YOLOv5
YOLOv3 in PyTorch > ONNX > CoreML > TFLite
YOLOv4, YOLOv4-tiny, YOLOv3, YOLOv3-tiny Implemented in Tensorflow 2.0, Android. Convert YOLO v4 .weights tensorflow, tensorrt and tflite
Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX
A repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
An awesome list of TensorFlow Lite models, samples, tutorials, tools and learning resources.
Android TensorFlow Lite Machine Learning Example
TFLite are now up to 2.6.1 (stable) and beyond in testing releases: https://github.com/tensorflow/tensorflow/releases
These may offer fixes for newer toolchains (GCC 11+, Clang 12+), in particular on more bleeding edge distros such as Manjaro.
Ticket here is to try these later versions of TFLite against newer toolchains, and ensure backscrub still actually works
See also: #93
虚拟爱抖露(アイドル)共享计划, 是基于单目RGB摄像头的人眼与人脸特征点检测算法, 在实时3D面部捕捉以及模型驱动领域的应用.
Prebuilt binary with Tensorflow Lite enabled. For RaspberryPi / Jetson Nano. Support for custom operations in MediaPipe. XNNPACK, XNNPACK Multi-Threads, FlexDelegate.
A Wide Range of Custom Functions for YOLOv4, YOLOv4-tiny, YOLOv3, and YOLOv3-tiny Implemented in TensorFlow, TFLite, and TensorRT.
GPU accelerated deep learning inference applications for RaspberryPi / JetsonNano / Linux PC using TensorflowLite GPUDelegate / TensorRT
Creating a software for automatic monitoring in online proctoring
React Native library for TensorFlow Lite
hair segmentation in mobile device
Demo on adding virtual background to a live video stream in the browser
Multi-Person Pose Estimation project for Tensorflow 2.0 with a small and fast model based on MobilenetV3
This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and pb. PyTorch (NCHW) -> ONNX (NCHW) -> OpenVINO (NCHW) -> openvino2tensorflow -> Tensorflow/Keras (NHWC/NCHW) -> TFLite (NHWC/NCHW). And the conversion from .pb to saved_model and from saved_model to .pb and from .pb to .tflite and saved_model to .tflite and saved_model to onnx. Support for building environments with Docker. It is possible to directly access the host PC GUI and the camera to verify the operation. NVIDIA GPU (dGPU) support. Intel iHD GPU (iGPU) support.
Generate saved_model, tfjs, tf-trt, EdgeTPU, CoreML, quantized tflite, ONNX, OpenVINO, Myriad Inference Engine blob and .pb from .tflite. Support for building environments with Docker. It is possible to directly access the host PC GUI and the camera to verify the operation. NVIDIA GPU (dGPU) support. Intel iHD GPU (iGPU) support. Supports inverse quantization of INT8 quantization model.
A Dart interface to TensorFlow Lite (tflite) through dart:ffi
Prebuilt binary for TensorFlowLite's standalone installer. For RaspberryPi. A very lightweight installer. I provide a FlexDelegate, MediaPipe Custom OP and XNNPACK enabled binary.
YOLOv3: convert .weights to .tflite format for tensorflow lite. Convert .weights to .pb format for tensorflow serving
TensorFlow Lite models for MIRNet for low-light image enhancement.
Create a project that does not link in any of the TFLM library or models. It would provide a starting point for software for non-TFLM based experimentation with CFUs.
Add a description, image, and links to the tflite topic page so that developers can more easily learn about it.
To associate your repository with the tflite topic, visit your repo's landing page and select "manage topics."
how to use gui in the AidLearning?how to custom the gui?