Indexing errors on JetBrains
Before submitting your bug report
- [ ] I believe this is a bug. I'll try to join the Continue Discord for questions
- [ ] I'm not able to find an open issue that reports the same bug
- [ ] I've seen the troubleshooting guide on the Continue Docs
Relevant environment info
- OS:
- Continue:
- IDE:
- Model:
- config.json:
Description
Multiple JetBrains users are reporting indexing issues.
Error: SQLITE_CONSTRAINT: UNIQUE constraint failed: tag_catalog.dir, tag_catalog.branch, tag_catalog.artifactId, tag_catalog.path, tag_catalog.cacheKey
- https://discord.com/channels/1108621136150929458/1286327457166790686
- https://github.com/continuedev/continue/issues/2300
Error: Failed to generate embeddings for 13 chunks with provider: OllamaEmbeddingsProvider::nomic-embed-text: Error: Failed to embed chunk:
- https://discord.com/channels/1108621136150929458/1282948472882069534
Error: Failed to generate embeddings for 233 chunks with provider: OpenAIEmbeddingsProvider: BAAI/bge-m3: Error: {"error":{"message":"The parameter is invalid. Please check again. (request id: 2024091810405578781328035296047)","type":"upstream_error","param":"400","code":"bad_response_status_code"}}
- https://github.com/continuedev/continue/issues/2289
Client network socket disconnected before secure TLS connection was established
- https://github.com/continuedev/continue/issues/2338
IntelliJ Ultimate 2024.1 with Continue 0.0.64 or 0.0.69 (EAP) both result in this indexing error on Windows 10 Enterprise:
[2024-09-23T18:35:51] Indexing: 0.0% complete, elapsed time: 2s, 0.00 file/sec
[2024-09-23T18:35:51] Indexing: 25.0% complete, elapsed time: 2s, 48.50 file/sec
[2024-09-23T18:35:52] Indexing: 25.0% complete, elapsed time: 3s, 32.33 file/sec
[2024-09-23T18:36:12] error when indexing: Error: Failed to generate embeddings for 1631 chunks with provider: _TransformersJsEmbeddingsProvider::all-MiniLM-L6-v2: Error: `local_files_only=true` or `env.allowRemoteModels=false` and file was not found locally at "C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json".
[2024-09-23T18:36:12] Indexing failed with error: Error: Failed to generate embeddings for 1631 chunks with provider: _TransformersJsEmbeddingsProvider::all-MiniLM-L6-v2: Error: `local_files_only=true` or `env.allowRemoteModels=false` and file was not found locally at "C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json". LanceDbIndex.getEmbeddings, process.processTicksAndRejections, async LanceDbIndex.computeRows, async LanceDbIndex.update, async CodebaseIndexer.indexFiles, async CodebaseIndexer.refresh, async Core.refreshCodebaseIndex, async C:\snapshot\continue\binary\out\index.js:498256:7, async C:\snapshot\continue\binary\out\index.js:498574:28
This path doesn't exist: C:\snapshot\continue . Maybe it's supposed to be a relative path from somewhere and is reverting to C: as a root instead?
I also encountered this error in PyCharm. What impact does this have? Is there a solution available now? Thank you!
With IntelliJ Ultimate 2024.1 on Windows and Continue 0.0.75, re-indexing doesn't show an error in the plugin panel view any more, but it does log errors to core.log and no indexes turn up in $USERPROFILE/.continue/index/lancedb, and I get this error (see screenshot) when trying to use the @codebase feature. The listed path does not exist (C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json) and for some reason it has both \ and / for path separators.
Searching through the codebase (https://github.com/search?q=repo%3Acontinuedev%2Fcontinue%20all-MiniLM-L6-v2&type=code), it seems like the embedding model is only included in the vscode extension and not the jetbrains one?
With IntelliJ Ultimate 2024.1 on Windows and Continue 0.0.75, re-indexing doesn't show an error in the plugin panel view any more, but it does log errors to core.log and no indexes turn up in $USERPROFILE/.continue/index/lancedb, and I get this error (see screenshot) when trying to use the @codebase feature. The listed path does not exist (C:\snapshot\continue\binary\models/all-MiniLM-L6-v2/tokenizer.json) and for some reason it has both
\and/for path separators.Searching through the codebase (https://github.com/search?q=repo%3Acontinuedev%2Fcontinue%20all-MiniLM-L6-v2&type=code), it seems like the embedding model is only included in the vscode extension and not the jetbrains one?
i have the same qustion , i have no solution
I would recommend changing the plugin logic to check if the embedding model is available first (maybe by calculating a single vector) before starting the compute-intensive crawling part, during which the IntelliJ editor window is unusable.
This issue hasn't been updated in 90 days and will be closed after an additional 10 days without activity. If it's still important, please leave a comment and share any new information that would help us address the issue.
This issue was closed because it wasn't updated for 10 days after being marked stale. If it's still important, please reopen + comment and we'll gladly take another look!
