InAI SciencebyLuv BansalSpeculative Decoding — Make LLM Inference FasterImprove LLM inference speed by 2–3X without degrading any accuracyApr 8, 20242Apr 8, 20242
InTDS ArchivebyThomas KurbielDrawing the Transformer Network from Scratch (Part 1)Getting a mental model of the Transformer in a playful wayNov 15, 20207Nov 15, 20207
InTDS ArchivebyMaxime LabonneFine-tune a Mistral-7b model with Direct Preference OptimizationBoost the performance of your supervised fine-tuned modelsJan 1, 20249Jan 1, 20249
InIntuitively and Exhaustively ExplainedbyDaniel WarfieldGPT — Intuitively and Exhaustively ExplainedExploring the architecture of OpenAI’s Generative Pre-trained Transformers.Dec 1, 202321Dec 1, 202321
InTDS ArchivebyTies de KokBeyond chat-bots: the power of prompt-based GPT models for downstream NLP tasksThis post shows how can use prompt engineering to use GPT models for downstream NLP tasks. It includes a Python example with the OpenAI…Apr 16, 20222Apr 16, 20222