My initial aim was to build a document processing model. But the idea was far fetched for my skill at the time. So, I settled down for building a toy version of visual language model for better understanding of VLM. I’m documenting my intuition for the benefit of myself and others. My model will receive image as an input and return its caption as an output. Luckily, I found a dataset with image and it’s caption to train.| Poonai's space
An interactive visualization tool showing you how transformer models work in large language models (LLM) like GPT.| poloclub.github.io
a newsletter from W. W. Norton and Annette Vee| aiandhowweteach.substack.com
Instruction tuning the GPT2 model on the Alpaca dataset using the Hugging Face Transformers library and the SFT Trainer pipine.| DebuggerCafe
We’re on a journey to advance and democratize artificial intelligence through open source and open science.| huggingface.co