Home
  • CV
  • Tech Stack
  • Books
  • Projects
  • List 100
Context window management for 200K token large language models
Context Engineering: Mastering the 200K Token Era

Master context window management for 200K+ token models. Optimize packing and eliminate truncation.

PROMPT AND CONTEXT ENGINEERING
LoRA fine-tuning for efficient large language model training and optimization
Fine-Tuning LLMs with LoRA: 2025 Guide

Master LoRA for efficient LLM fine-tuning. Reduce training time 90% and memory by 75%.

LLM MODELS, PROVIDERS AND TRAINING
LLM inference optimization strategies to reduce AI costs
LLM Inference: Cut AI Costs by 80%

Practical strategies to reduce LLM inference costs by 80% while maintaining output quality.

INFERENCE, SERVING AND COST CONTROL

© 2025 Amir Teymoori