Skip to content
This repository was archived by the owner on Jun 5, 2025. It is now read-only.

Commit 12dd7c4

Browse files
author
Luke Hinds
authored
Merge pull request #158 from stacklok/bug-109
Convert async _close_models in LlamaCppInferenceEngine
2 parents b518f0d + fbabe31 commit 12dd7c4

File tree

1 file changed

+10
-10
lines changed

1 file changed

+10
-10
lines changed

src/codegate/inference/inference_engine.py

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,16 @@ def __init__(self):
2424
self.__models = {}
2525

2626
def __del__(self):
27-
self.__close_models()
27+
self._close_models()
28+
29+
def _close_models(self):
30+
"""
31+
Closes all open models and samplers
32+
"""
33+
for _, model in self.__models.items():
34+
if model._sampler:
35+
model._sampler.close()
36+
model.close()
2837

2938
async def __get_model(self, model_path, embedding=False, n_ctx=512, n_gpu_layers=0):
3039
"""
@@ -70,12 +79,3 @@ async def embed(self, model_path, content):
7079
"""
7180
model = await self.__get_model(model_path=model_path, embedding=True)
7281
return model.embed(content)
73-
74-
async def __close_models(self):
75-
"""
76-
Closes all open models and samplers
77-
"""
78-
for _, model in self.__models:
79-
if model._sampler:
80-
model._sampler.close()
81-
model.close()

0 commit comments

Comments
 (0)