-
Notifications
You must be signed in to change notification settings - Fork 0
/
llm_client.py
50 lines (43 loc) · 1.46 KB
/
llm_client.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
from langchain.llms.base import LLM
from typing import Optional, List, Mapping, Any
import requests
HOST = 'HOST'
URI = f'https://{HOST}/api/v1/generate'
class AlpacaLLM(LLM):
@property
def _llm_type(self) -> str:
return "custom"
def _call(self, prompt: str, stop: Optional[List[str]] = None) -> str:
if isinstance(stop, list):
stop = stop + ["\n###","\nObservation:", "\nObservations:"]
response = requests.post(
URI,
json={
"prompt": prompt,
"temperature": 0.3, #0.7
"max_new_tokens": 256,
"early_stopping": True,
"stopping_strings": stop,
'do_sample': True,
'top_p': 0.1,
'typical_p': 1,
'repetition_penalty': 1.18,
'top_k': 40,
'min_length': 0,
'no_repeat_ngram_size': 0,
'num_beams': 1,
'penalty_alpha': 0,
'length_penalty': 1,
'seed': -1,
'add_bos_token': True,
'truncation_length': 2048,
'ban_eos_token': False,
'skip_special_tokens': True,
},
)
response.raise_for_status()
return response.json()['results'][0]['text']
@property
def _identifying_params(self) -> Mapping[str, Any]:
"""Get the identifying parameters."""
return {}