- MLCEngine: Universal LLM Deployment to Both Cloud and Local Devices
by crowwork on 6/8/24, 10:45 PM, with comments
- Running Google's Gemma 2B on Android
by crowwork on 2/23/24, 5:46 PM, with comments
- MLC Chat: Chat with Open Language Models Locally on iPad and iPhone
by crowwork on 6/4/23, 10:48 PM, with comments
- WebLLM NPM Package
by crowwork on 5/26/23, 1:48 AM, with comments
- Bringing Hardware Accelerated Language Models to Android Devices
by crowwork on 5/9/23, 2:43 PM, with comments
- Bringing Hardware Accelerated Language Models to Consumer Devices
by crowwork on 5/1/23, 11:12 PM, with comments
- MLC LLM – Large Language Models on iPhone GPU and Many More GPU Platforms
by crowwork on 4/29/23, 3:15 PM, with comments
- MLC LLM: Universal LLM Deployment with GPU Acceleration
by crowwork on 4/29/23, 3:46 AM, with comments
- Web LLM
by crowwork on 4/15/23, 2:26 AM, with comments
- Web Stable Diffusion
by crowwork on 3/17/23, 1:37 AM, with comments
- Running Stable Diffusion fully in browser with WebGPU
by crowwork on 3/8/23, 6:33 PM, with comments
- TinyML – How TVM Is Taming Tiny
by crowwork on 6/9/20, 7:40 PM, with comments
- Compiling Machine Learning to WASM and WebGPU with Apache TVM
by crowwork on 5/14/20, 6:02 PM, with comments
- FFI Navigator: Language Server for Cross Language FFI Calls
by crowwork on 1/19/20, 2:30 AM, with comments
- TVM and Deep Learning Compilation Conference 2019 Videos and Slides
by crowwork on 12/20/19, 8:51 PM, with comments
- Automating Optimization of Quantized Deep Learning Models on CUDA
by crowwork on 4/30/19, 6:21 PM, with comments
- Golang Runtime for Deep Learning Deployment in TVM
by crowwork on 1/19/19, 6:56 PM, with comments