Yusuf Mansur Özer
Yusuf Mansur Özer
I really think we need an API for this. We might not always want to get the markdown from a file.
I've just discovered this repo and it is very promising. It would sooo good to have an NPM alternative of GPT Index. Thanks for the great work! My humble view:...
Happy to help! Now we need to somehow handle embedding requests that exceed 250,000 tokens in total to overcome rate limits. :) I'll DM on Discord now. @cfortuner
Thank you very much, truly appreciated. Please indulge me for asking this but I understand `whitelist` is the `safelist` option of PurgeCSS and that is used to "indicate which selectors...
You mean only get the top k results and not the answer? Here is how I do it: ``` def query_index(question, index_path, top_k): # Load index index = GPTSimpleVectorIndex.load_from_disk(index_path) #...
@jerryjliu Ah, I see and that makes perfect sense. But still, when you get embeddings for large bodies of text, there is some duplication (or am I doing something wrong?)....
@jerryjliu Just wanted to follow up and ask if you have any ideas about this because removing this duplication would drastically reduce index file sizes.
Makes sense. But I am having a similar issue where my users are uploading documents and when they want to query multiple documents at once, I need to construct a...
Fully agree. A JS/TS version would be great. I have to say I am learning Python solely to use GPT Index. :)
You can get the `source_nodes` from the response object and it will include the similarity score. Is this what you're looking for?