Jan
An open source alternative to OpenAI that runs on your own computer or server
Popular repositories Loading
-
-
cortex.tensorrt-llm
cortex.tensorrt-llm PublicForked from NVIDIA/TensorRT-LLM
Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU accelerated inference on NVIDIA's GPUs.
-
Repositories
Showing 10 of 39 repositories
- research-utils Public
janhq/research-utils’s past year of commit activity - cortex.llamacpp Public
cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server at runtime.
janhq/cortex.llamacpp’s past year of commit activity - cortex.tensorrt-llm Public Forked from NVIDIA/TensorRT-LLM
Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU accelerated inference on NVIDIA's GPUs.
janhq/cortex.tensorrt-llm’s past year of commit activity - cortex-web Public
janhq/cortex-web’s past year of commit activity