LitServe icon indicating copy to clipboard operation
LitServe copied to clipboard

Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.

Results 75 LitServe issues
Sort by recently updated
recently updated
newest added

The streaming works really nicely now using the latest litgpt version from main. The only little issue is that it creates a `%` character e.g., ``` ⚑ ~/streaming python streaming_client.py...

bug
help wanted

## πŸš€ Feature Add the ability for developers to customize the FastAPI `lifespan`, `startup`, and 'shutdown` event handlers. ### Motivation Users can use this to setup logging or resource management....

enhancement
help wanted

Before submitting - [x] Was this discussed/agreed via a Github issue? (no need for typos and docs improvements) - [x] Did you read the [contributor guideline](https://github.com/Lightning-AI/pytorch-lightning/blob/main/.github/CONTRIBUTING.md), Pull Request section? -...

fixes #99 * Select streaming or non-streaming loop based on client's `request["stream"]` parameter. ## Server ```py # server.py import logging import litserve as ls logging.basicConfig(level=logging.INFO) # STEP 1: DEFINE YOUR...

has conflicts

## πŸš€ Feature Request: Add Support for Images in Chat Messages ### Description Add support for images to be included in chat messages, similar to [gpt-4o](https://platform.openai.com/docs/guides/vision). ```json { "messages": [...

enhancement
help wanted

Investigate recycling of Pipe connections. Let’s merge it as a P0, but please add an issue to investigate recycling pipes _Originally posted by @lantiga in https://github.com/Lightning-AI/LitServe/pull/108#pullrequestreview-2073769140_

Allow LitServer to send requests to either streaming or non-streaming inference worker dynamically based on the request. cc: @lantiga

enhancement

FastAPI background_task cleanup is not executed if any error is encountered. To reproduce, run this code and add a print in cleanup function. ```py import litserve as ls class SimpleLitAPI(ls.LitAPI):...

bug
help wanted

## We need to add tests for when: **- predict fails during streamed generation** -> Users will receive 200 HTTP status if fails at LitAPI side and 50x if fails...

## πŸ› Bug ### To Reproduce Steps to reproduce the behavior: 1. Go to '...' 1. Run '....' 1. Scroll down to '....' 1. See error #### Code sample ###...

bug
help wanted