Yifan Zhu

Results 2 comments of Yifan Zhu

Very helpful discussion. I successfully ran the inference on 4*RTX 4090 by following your modifications. Thx for you all : )

流程是差不多的。简单去理解的话,前面还是搜索引擎这套链路,最后把排序靠前的 top k 文档丢给 LLM 作为 context,让它结合 context 去回答用户的 query