|
3 | 3 | # This file is a part of the ModelEngine Project. |
4 | 4 | # Licensed under the MIT License. See License.txt in the project root for license information. |
5 | 5 | # ====================================================================================================================== |
6 | | -import functools |
7 | 6 | import os |
8 | 7 | import traceback |
9 | 8 | from enum import Enum, unique |
10 | | -from inspect import signature |
11 | 9 | from typing import List, Callable, Any, Tuple |
12 | 10 |
|
13 | 11 | from fitframework import fit_logger |
14 | 12 | from fitframework.core.repo.fitable_register import register_fitable |
15 | | -from llama_index.core import PromptTemplate |
16 | 13 | from llama_index.core.base.base_selector import SingleSelection |
17 | 14 | from llama_index.core.postprocessor import SimilarityPostprocessor, SentenceEmbeddingOptimizer, LLMRerank, \ |
18 | 15 | LongContextReorder, FixedRecencyPostprocessor |
19 | 16 | from llama_index.core.postprocessor.types import BaseNodePostprocessor |
20 | | -from llama_index.core.prompts import PromptType |
| 17 | +from llama_index.core.prompts import PromptType, PromptTemplate |
21 | 18 | from llama_index.core.prompts.default_prompts import DEFAULT_CHOICE_SELECT_PROMPT_TMPL |
22 | 19 | from llama_index.core.selectors import LLMSingleSelector, LLMMultiSelector |
23 | 20 | from llama_index.core.selectors.prompts import DEFAULT_SINGLE_SELECT_PROMPT_TMPL, DEFAULT_MULTI_SELECT_PROMPT_TMPL |
24 | 21 | from llama_index.embeddings.openai import OpenAIEmbedding |
25 | | -from llama_index.legacy.llms import OpenAILike |
| 22 | +from llama_index.llms.openai import OpenAI |
26 | 23 |
|
27 | 24 | from .callable_registers import register_callable_tool |
28 | | -from .types.document import Document |
29 | 25 | from .node_utils import document_to_query_node, query_node_to_document |
| 26 | +from .types.document import Document |
30 | 27 |
|
31 | 28 | os.environ["no_proxy"] = "*" |
32 | 29 |
|
@@ -80,7 +77,7 @@ def llm_rerank(nodes: List[Document], query_str: str, **kwargs) -> List[Document |
80 | 77 | choice_batch_size = int(kwargs.get("choice_batch_size") or 10) |
81 | 78 | top_n = int(kwargs.get("top_n") or 10) |
82 | 79 |
|
83 | | - llm = OpenAILike(model=model_name, api_base=api_base, api_key=api_key, max_tokens=4096) |
| 80 | + llm = OpenAI(model=model_name, api_base=api_base, api_key=api_key, max_tokens=4096) |
84 | 81 | choice_select_prompt = PromptTemplate(prompt, prompt_type=PromptType.CHOICE_SELECT) |
85 | 82 | llm_rerank_obj = LLMRerank(llm=llm, choice_select_prompt=choice_select_prompt, choice_batch_size=choice_batch_size, |
86 | 83 | top_n=top_n) |
@@ -110,7 +107,7 @@ def llm_choice_selector(choice: List[str], query_str: str, **kwargs) -> List[Sin |
110 | 107 | if mode.lower() not in [m.value for m in SelectorMode]: |
111 | 108 | raise ValueError(f"Invalid mode {mode}.") |
112 | 109 |
|
113 | | - llm = OpenAILike(model=model_name, api_base=api_base, api_key=api_key, max_tokens=4096) |
| 110 | + llm = OpenAI(model=model_name, api_base=api_base, api_key=api_key, max_tokens=4096) |
114 | 111 | if mode.lower() == SelectorMode.SINGLE.value: |
115 | 112 | selector_prompt = prompt or DEFAULT_SINGLE_SELECT_PROMPT_TMPL |
116 | 113 | selector = LLMSingleSelector.from_defaults(llm=llm, prompt_template_str=selector_prompt) |
|
0 commit comments