-
Stability.ai, Eleuther.ai
- Seattle, WA
- http:https://dmarx.github.io
- @DigThatData
Block or Report
Block or report dmarx
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseML Performance
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Distributed Machine Learning Patterns from Manning Publications by Yuan Tang https://bit.ly/2RKv8Zo
Efficient, scalable and enterprise-grade CPU/GPU inference server for 🤗 Hugging Face transformer models 🚀
Prefect is a workflow orchestration framework for building resilient data pipelines in Python.
ncnn is a high-performance neural network inference framework optimized for the mobile platform
A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.
FFCV: Fast Forward Computer Vision (and other ML workloads!)
Neo-AI-DLR is a common runtime for machine learning models compiled by AWS SageMaker Neo, TVM, or TreeLite.
Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals
A retargetable MLIR-based machine learning compiler and runtime toolkit.
Knowledge-Aware machine LEarning (KALE): accessible machine learning from multiple sources for interdisciplinary research, part of the 🔥PyTorch ecosystem. ⭐ Star to support our work!
The net:cal calibration framework is a Python 3 library for measuring and mitigating miscalibration of uncertainty estimates, e.g., by a neural network.
The Tensor Algebra SuperOptimizer for Deep Learning
Making large AI models cheaper, faster and more accessible
Training and serving large-scale neural networks with auto parallelization.
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Neural Network Compression Framework for enhanced OpenVINO™ inference
Module to Automatically maximize the utilization of GPU resources in a Kubernetes cluster through real-time dynamic partitioning and elastic quotas - Effortless optimization at its finest!
Efficiently read embedding in streaming from any filesystem
Hunter is a flexible code tracing toolkit.
Accelerated NLP pipelines for fast inference on CPU and GPU. Built with Transformers, Optimum and ONNX Runtime.
py.test fixture for benchmarking code
Productive, portable, and performant GPU programming in Python.
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
A collection of optimizers, some arcane others well known, for Flax.