Skip to content

Releases: helixml/helix

0.10.7 - fix gpu memory leak

23 Aug 16:58
7a34d5e
Compare
Choose a tag to compare

What's Changed

Fix newly introduced regression where runners leak memory because we kill the parent process before cleaning up the children.

Full Changelog: 0.10.6...0.10.7

0.10.6 - improve logging

23 Aug 06:57
ae3443c
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.10.5...0.10.6

0.10.5 - fixes

22 Aug 01:59
1be7d40
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.10.4...0.10.5

0.10.4 - declarative knowledge, custom rag server example, improve tools reliability

20 Aug 14:49
2115c0d
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.10.3...0.10.4

0.10.3

19 Aug 16:07
031b926
Compare
Choose a tag to compare

What's Changed

Lots of fixes since big refactor in 0.10. Most notably, TOOLS_PROVIDER is no longer an option, it will use the same OpenAI client as INFERENCE_PROVIDER, which defaults to helix. What's more, when INFERENCE_PROVIDER=helix (now the default), API tools queries now work again. Sorry for the regression!


Full Changelog: 0.10.2...0.10.3

0.10.2 - frontend support for external LLM providers and stability fixes

16 Aug 09:21
7cae8d9
Compare
Choose a tag to compare

What's Changed

The frontend and API now supports external LLM providers, configured by:

INFERENCE_PROVIDER=openai
OPENAI_API_KEY=sk-...
OPENAI_BASE_URL=<any OpenAI compatible API>

This works with RAG, API calling, and apps. You still need to attach a GPU runner for image inference, and text and image fine-tuning.

This release also improves stability using our own runners. A bug that caused models to be prematurely unloaded has been fixed. Another bug that caused more memory to be allocated than was available under load has also been fixed. Latency and availability of LLMs when using our runners should be improved.

RAG and system prompts now properly work with the new OpenAI compatible API path (for both external and local LLMs).

This release gives you full optionality over whether to use external LLMs or local HA ollama with our apps stack (RAG, API calling, system prompts - all configured via version controlled yaml), all running in docker or Kubernetes.

Next, we'll add function calling support so that gptscript can run fully locally.


Full Changelog: 0.10.1...0.10.2

0.10.1

14 Aug 13:02
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.10.0...0.10.1

0.10.0

12 Aug 13:24
dd876df
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.9.26...0.10.0

0.9.26

31 Jul 21:53
13c6542
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.9.25...0.9.26

0.9.25

30 Jul 21:52
07d6c3d
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: 0.9.24...0.9.25