You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
INFERENCE_RESPONSE_PER_MINUTE_LIMIT=4# slow down to not exceed token per minute (tpm) limit of 60k
23
-
INFERENCE_API_KEY=os.getenv("NEBULA_API_KEY")
24
-
INFERENCE_MAX_CHARACTERS=100000# max characters in all files provided to the model, approximately 25k tokens (limit is 32k)
18
+
19
+
# Get inference server configuration from environment variables
20
+
INFERENCE_URL=os.getenv("INFERENCE_URL")
21
+
ifnotINFERENCE_URL:
22
+
raiseValueError("INFERENCE_URL environment variable is not set. Please set it before running the script.")
23
+
INFERENCE_MODEL=os.getenv("INFERENCE_MODEL")
24
+
ifnotINFERENCE_MODEL:
25
+
raiseValueError("INFERENCE_MODEL environment variable is not set. Please set it before running the script.")
26
+
INFERENCE_API_KEY=os.getenv("INFERENCE_API_KEY")
27
+
ifnotINFERENCE_API_KEY:
28
+
raiseValueError("INFERENCE_API_KEY environment variable is not set. Please set it before running the script.")
29
+
30
+
INFERENCE_RESPONSE_PER_MINUTE_LIMIT=10# slow down to not exceed token per minute (tpm) limit
31
+
INFERENCE_MAX_CHARACTERS=400000# max characters in all files provided to the model, approximately 100k tokens
25
32
26
33
QUESTIONS= [
27
34
["Is there a EXTENSION_DESCRIPTION variable in the CMakeLists.txt file that describes what the extension does in a few sentences that can be understood by a person knowledgeable in medical image computing?", ["cmake"]],
0 commit comments