This paper proposes evaluating pruning and quantization techniques to reduce model size and inference cost while maintaining classification accuracy on ecological datasets such as iNaturalist.
-
Updated
Mar 7, 2026
This paper proposes evaluating pruning and quantization techniques to reduce model size and inference cost while maintaining classification accuracy on ecological datasets such as iNaturalist.
Mistral Legal Advisor is a full-stack AI application that connects a fine-tuned Mistral language model running on Google Colab (GPU) with a local Flask-based frontend. The system generates comprehensive legal document requirements for startups based on user inputs, using an ngrok tunnel to securely bridge local and cloud environments.
A complete, menu-driven AI model interface for Windows that simplifies running local GGUF language models with llama.cpp. This tool automatically manages dependencies, provides multiple interaction modes, and prioritizes user privacy through fully offline operation.
A privacy-first, fully offline neural machine translation application supporting 200+ languages. Built on Meta's NLLB-200 model, this tool runs entirely in your browser without requiring server infrastructure or internet connectivity after initial model download.
Add a description, image, and links to the quantized-models topic page so that developers can more easily learn about it.
To associate your repository with the quantized-models topic, visit your repo's landing page and select "manage topics."