π Showcasing Project, in 2024 Google Machine Learning Bootcamp π
KOREAN
Β |Β
ENGLISH
Award-Factory: Awards crafted for you by a hilariously talented generative AI
Check out prototypes in the badge below
Note
- This project aims to develop a service where anyone can effortlessly create a customized certificate in just a few minutes, making it easy to celebrate and appreciate others.
- Award Factory was conceived as a heartwarming project to spread happiness, inspired by the idea of creating special certificates for parents. Built with sustainability in mind, the service integrates front-end components and leverages the fine-tuned Google Gemma:2b model to deliver personalized award texts. While the service is not fully active due to server operation costs, a demo is available on Huggingface.
- Advanced technologies like QLoRA quantization and llama-cpp optimizations were employed to reduce model size and improve performance, ensuring an efficient user experience in the future.
demo.mp4
|
App Design |
Generated Awards |
|
|
|
Google Gemma:2B Finetuning
Implemented prompt engineering and QLoRA-based quantization fine-tuning using the Google/Gemma-2b-it model with PEFT techniques to optimize personalized award text generation tailored to user preferences.llama-cpp Quantization
Applied quantization with the Q5_K_M option in llama-cpp, achieving a 63.3% reduction in model size and an 83.4% decrease in inference time without compromising performance, enabling faster and more efficient service.$ llama.cpp/llama-quantize gguf_model/gemma-2b-it-award-factory-v2.gguf gguf_model/gemma-2b-it-award-factory-v2.gguf-Q5_K_M.gguf Q5_K_M
...
llama_model_quantize_internal: model size = 4780.29 MB
llama_model_quantize_internal: quant size = 1748.67 MB
main: quantize time = 17999.81 ms
main: total time = 17999.81 ms
$ ollama list
NAME ID SIZE MODIFIED
award-factory:q5 8df06172b64b 1.8 GB 19 seconds ago
award-factory:latest ae186115cc83 5.0 GB 28 minutes ago



