LoRA and the /v1/completions Interface
1. Background: Evolution of Model Fine-Tuning and the Role of LoRA
Key Concept: What is LoRA?
Comparison with Full Fine-Tuning
Feature
LoRA
Full Fine-Tuning
Example: Using LoRA in llama.cpp
llama.cpp./main -m llama.gguf --lora lora.gguf -p "Your prompt"2. Background: Code Completion Requirements and API Interface Differences
Purpose of /v1/completions
/v1/completionsComparison with /v1/chat/completions
/v1/chat/completionsEndpoint
Supported Models
Input Style
Suitable for FIM
3. Background: FIM and Meta-Language Prompting
Example Prompt
4. Background: Why chat/completions Is Not Recommended for FIM
chat/completions Is Not Recommended for FIM5. Background: Industry Practice and Model Interface Compatibility
Model
FIM Support
Prompt Format
Recommended API
Last updated