Log in
Log into community
Helicone Community Page
View all posts
Related posts
Did this answer your question?
😞
😐
😃
Powered by
Hall
Inactive
Updated 4 months ago
0
Follow
Helicone's impact on latency in LLM calls
Helicone's impact on latency in LLM calls
Inactive
0
Follow
D
David Alonso
4 months ago
·
Am I crazy or does Helicone add a bunch of latency to every single LLM call? I'm using the Vercel AI SDK
J
D
18 comments
Share
Open in Discord
J
Justin
4 months ago
Hi David! It should not, unless you are using Cache or Rate Limiting
D
David Alonso
4 months ago
ah okay, my understanding is that we need to hit your server first and then you forward to whatever LLM provider. But maybe I don't get what the baseUrl param does
D
David Alonso
4 months ago
how can I check whether I'm using cache or rate limiting? feel free to share a link and happy to do some reading
J
Justin
4 months ago
it'd be really explict
https://docs.helicone.ai/features/advanced-usage/caching#llm-caching
J
Justin
4 months ago
If you want to use our async integration and pass traffic through Helicone you can use our OpenLLMetry integration
https://docs.helicone.ai/getting-started/integration-method/openllmetry
D
David Alonso
4 months ago
this link is broken i think
Attachment
D
David Alonso
4 months ago
in the last page you sent
D
David Alonso
4 months ago
does this sdk require a node environment or can it run on edge runtimes?
D
David Alonso
4 months ago
is this intuition correct?
J
Justin
4 months ago
Hye @David Alonso great question! I am not sure, I am double checking with the OpenLLMetry team
J
Justin
4 months ago
correct!
D
David Alonso
4 months ago
nice, but then there is added network latency right? so the async method would lead to faster inference iiuc
D
David Alonso
4 months ago
well faster response time for the user i mean
J
Justin
4 months ago
That's correct, but it should really be marginal.
D
David Alonso
4 months ago
I assume they haven’t replied, but super interested to hear back!
J
Justin
4 months ago
Thanks for following up, their answer was not very helpful lol
J
Justin
4 months ago
What edge time environment are you looking to run this in @David Alonso
D
David Alonso
4 months ago
The Convex JS runtime:
https://docs.convex.dev/functions/runtimes
Add a reply
Sign up and join the conversation on Discord
Join on Discord