-
Notifications
You must be signed in to change notification settings - Fork 1
/
prompt_llm.py
129 lines (115 loc) · 4.15 KB
/
prompt_llm.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
import os
from yaml import safe_load
import openai
import ollama
from read import get_df_segments
from datetime import datetime
import json
from tqdm import tqdm
from prompts import get_context
import csv
from copy import deepcopy
def call_api(llm_location, messages, engine, max_tokens):
assert llm_location in ["azure", "local"]
if llm_location == "azure":
return openai.chat.completions.create(
model=engine, messages=messages, max_tokens=500
)
else:
return client.chat.completions.create(
model=engine, messages=messages, max_tokens=500
)
def call_with_context(
llm_location, context: list, sentence: str, engine: str, max_tokens: int, role="user"
) -> str:
current_context = deepcopy(context)
current_context.append({"role": role, "content": sentence})
response = call_api(llm_location, current_context, engine, max_tokens)
message = response.choices[0].message
print(message)
return message.content
def get_llm_response(llm_location, sentence, context, engine, max_tokens):
try:
answer = call_with_context(llm_location, context, sentence, engine, max_tokens)
print(answer)
return answer
except Exception as e:
print(e)
return "ERROR"
if __name__ == "__main__":
# load config
with open("config.json", "r") as f:
config = json.load(f)
if config["llm_location"] == "azure":
# connect to api
# Ollama not applicable here
with open("openai.credential", "r") as stream:
credential_data = safe_load(stream)
openai_config = credential_data["openai"]
openai.api_type = "azure"
openai.azure_endpoint = openai_config["endpoint"]
openai.api_version = "2024-02-15-preview"
openai.api_key = openai_config["key"]
print(openai.version.VERSION)
else:
client = openai.OpenAI(base_url=config["base_url"], api_key=config["api_key"])
# select prompt
context = get_context(config["prompt_id"])
# get df with segments to be queried
if not "remove_HTML" in config:
config["remove_HTML"] = False
df_segments = get_df_segments(
config["dataset_name"],
n_policies=config["n_policies"],
remove_html_tags=config["remove_HTML"],
)
df_segments["prompt_id"] = config["prompt_id"]
df_segments["engine"] = config["engine"]
df_segments["dataset_name"] = config["dataset_name"]
# df_segments = df_segments[0:10]
# get save location ready
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
if config["n_policies"] is None:
mode = "complete"
else:
mode = "partial"
result_path = os.path.join(
"results",
config["dataset_name"],
config["engine"],
config["prompt_id"],
mode,
timestamp,
)
if not os.path.exists(result_path):
os.makedirs(result_path)
# query llm and save progress
for index, row in tqdm(df_segments.iterrows(), total=df_segments.shape[0]):
answer = get_llm_response(
config["llm_location"],
row["segment_text"],
context,
config["engine"],
config["max_tokens"]
)
df_segments.loc[index, "llm_response"] = answer
with open(os.path.join(result_path, "progress.csv"), "a") as file:
writer = csv.writer(file)
writer.writerow([row["complete_segment_ID"], answer])
# tqdm.pandas()
# df_segments["llm_response"] = df_segments["segment_text"].progress_apply(
# get_llm_response, args=(context, config["engine"])
# )
# save final results
with open(os.path.join(result_path, "config.json"), "w") as f:
json.dump(config, f)
if config["n_policies"] is None:
df_segments.to_csv(os.path.join(result_path, "results.csv"))
df_segments.to_excel(os.path.join(result_path, "results.xlsx"))
else:
df_segments.to_csv(
os.path.join(result_path, "results_" + str(config["n_policies"]) + ".csv")
)
df_segments.to_excel(
os.path.join(result_path, "results_" + str(config["n_policies"]) + ".xlsx")
)