Architectures

Four retrieval architectures compared under a frozen Mistral-7B-Instruct-v0.2 generator (greedy decoding, chat template applied).

IDArchitectureTypeRoleKey Parameters
(a)BM25Non-neural baselineReference floork1=1.5, b=0.75
(b)BGE-M3Dense retriever (CLS pooling)Primary dense baselinelr=1e-5, batch=32, epochs=3, 1024-subword chunks
(c)Hybrid BM25+BGE-M3+CrossEncoderLexical + Dense + RerankerExpected strongestRRF top-50 → rerank → top-10, max_length=1024, batch_size=4
(d)Legal-BERT Bi-EncoderDomain-specific TransformerOptional domain-referencelr=2e-5, warmup=10%, batch=32, epochs=3, 512-subword chunks

API Placeholder

GET /api/architectures — returns architecture configs and training status (pending)