mirror of
https://github.com/johndoe6345789/AutoMetabuilder.git
synced 2026-04-24 13:54:59 +00:00
38 lines
923 B
Python
38 lines
923 B
Python
import os
|
|
import requests
|
|
import yaml
|
|
from openai import OpenAI
|
|
|
|
RAW_PROMPT_URL = "https://raw.githubusercontent.com/johndoe6345789/metabuilder/main/getonwithit.prompt.yml"
|
|
|
|
token = os.environ["GITHUB_TOKEN"]
|
|
endpoint = "https://models.github.ai/inference"
|
|
|
|
client = OpenAI(
|
|
base_url=endpoint,
|
|
api_key=token,
|
|
)
|
|
|
|
def load_prompt_yaml(url: str) -> dict:
|
|
headers = {"Authorization": f"Bearer {token}"}
|
|
r = requests.get(url, headers=headers, timeout=30)
|
|
r.raise_for_status()
|
|
return yaml.safe_load(r.text)
|
|
|
|
prompt = load_prompt_yaml(RAW_PROMPT_URL)
|
|
|
|
messages = prompt["messages"]
|
|
model = prompt.get("model", "openai/gpt-4.1")
|
|
|
|
# Add runtime request (optional)
|
|
messages = messages + [{"role": "user", "content": "What should I do next?"}]
|
|
|
|
response = client.chat.completions.create(
|
|
model=model,
|
|
messages=messages,
|
|
temperature=1.0,
|
|
top_p=1.0,
|
|
)
|
|
|
|
print(response.choices[0].message.content)
|