LoRA (Low-Rank Adaptation) is a fine-tuning method that adjusts pre-trained models by injecting trainable low-rank matrices into existing layers while keeping the original weights frozen. This approach reduces the number of trainable parameters, enabling efficient adaptation with less computational resources and memory, making it ideal for customizing large models for specific tasks without extensive retraining.