Skip to content

feat: use llama.cpp server #2128

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 48 commits into from
Apr 4, 2025
Merged

feat: use llama.cpp server #2128

merged 48 commits into from
Apr 4, 2025

Conversation

vansangpfiev
Copy link
Contributor

@vansangpfiev vansangpfiev commented Mar 17, 2025

Describe Your Changes

  • Ready to review, e2e tests work on VMs, still checking why they do not work on CIs

This pull request includes various updates and improvements across multiple files, focusing on version updates, code refactoring, and adding new functionality. The most important changes include updating version strings, refactoring code to use constants, and adding new engine support.

Version Updates:

  • Updated version strings in docs/static/openapi/cortex.json to reflect new versions and naming conventions. [1] [2] [3] [4] [5]

Code Refactoring:

  • Removed the IsSupported method from EngineI class in docs/docs/engines/engine-extension.mdx and engine/cortex-common/EngineI.h for cleaner interface. [1] [2]
  • Replaced hardcoded organization and repository names with constants in various files for maintainability. [1] [2] [3] [4]

New Functionality:

  • Added support for local-engine in CMakeLists.txt files to include new engine extension. [1] [2]
  • Added a custom command for MSVC builds in engine/CMakeLists.txt to copy necessary files post-build.

Logging and Debugging:

  • Added logging statements in engine/controllers/engines.cc for better traceability of engine versions and variants during installation. [1] [2]

Test Updates:

  • Updated test cases in engine/e2e-test/api/engines to use new version strings and variant names. [1] [2] [3] [4] [5] [6] [7] [8]

Fixes Issues

Self Checklist

  • Added relevant comments, esp in complex areas
  • Updated docs (for bug fixes / features)
  • Created issues for follow-up changes or refactoring needed

@vansangpfiev vansangpfiev force-pushed the s/feat/spawn-llama-cpp branch from 8350e2d to 0968abe Compare March 17, 2025 07:50
Copy link
Contributor

github-actions bot commented Mar 17, 2025

Preview URL: https://be4cfec3.cortex-docs.pages.dev

* chore: release CIs

* fix: cuda urls

---------

Co-authored-by: sangjanai <sang@jan.ai>
@vansangpfiev vansangpfiev merged commit b901e01 into dev Apr 4, 2025
9 checks passed
@vansangpfiev vansangpfiev deleted the s/feat/spawn-llama-cpp branch April 4, 2025 02:08
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants