Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Connectivity Error #74

Open
1 task done
rosiecakes5 opened this issue Mar 11, 2025 · 3 comments
Open
1 task done

Connectivity Error #74

rosiecakes5 opened this issue Mar 11, 2025 · 3 comments
Assignees
Labels
bug Something isn't working

Comments

@rosiecakes5
Copy link

Description

Chat stopped responding, followed by the following error pop up:

Image

Reproduction

This bug occurred while using the app; there was no identified trigger

Expected behavior

The chat would continue to respond

Additional context

No response

Code of Conduct

  • I agree to follow this project's Code of Conduct and Contributing Guidelines
@rosiecakes5 rosiecakes5 added the bug Something isn't working label Mar 11, 2025
@jdisho jdisho changed the title 🐛 Bug report: Remove this title with a descriptive title. Connectivity Error Mar 12, 2025
@philippzagar philippzagar self-assigned this Mar 12, 2025
@philippzagar
Copy link
Collaborator

Thanks for raising the issue @rosiecakes5! 🚀

This seems like a pretty regular network layer timeout to me, meaning the transport threw an error. Tbh, not that unlikely with all the different inference requests while summarizing individual FHIR resources.
I'll investigate further to see if we have any actual error here. We probably should implement a retry mechanism in case the network request fails / times out.

@PSchmiedmayer
Copy link
Member

I am wondering if this is due to the face that we have a lot of parallel function calls; could be that some dispatching of them is failing due to that parallel nature? And agree, maybe a small retry (one time) on a few HTTP codes that indicate that would make sense; could be configurable in the LLM Schema definition and could have a default value.

@philippzagar
Copy link
Collaborator

@PSchmiedmayer Totally agreed, as already mentioned above, the large amount of individual parallel function calls might lead to that. The LLMOpenAIPlatform however offers an easy way to constrain the maximal amount of parallel LLM inferences, so we might be able to mitigate that by reducing that number, of course at the cost of overall response time from the model.

In general, a small retry policy in the case of a transport error (such as timeouts) would be a great thing in SpeziLLM. Not sure if I would actually extend that to HTTP response codes, but there might be some use cases where that makes sense.
Here's the issue: StanfordSpezi/SpeziLLM#107

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants