Files
AutoMetabuilder/main.py

38 lines
923 B
Python

import os
import requests
import yaml
from openai import OpenAI
RAW_PROMPT_URL = "https://raw.githubusercontent.com/johndoe6345789/metabuilder/main/getonwithit.prompt.yml"
token = os.environ["GITHUB_TOKEN"]
endpoint = "https://models.github.ai/inference"
client = OpenAI(
base_url=endpoint,
api_key=token,
)
def load_prompt_yaml(url: str) -> dict:
headers = {"Authorization": f"Bearer {token}"}
r = requests.get(url, headers=headers, timeout=30)
r.raise_for_status()
return yaml.safe_load(r.text)
prompt = load_prompt_yaml(RAW_PROMPT_URL)
messages = prompt["messages"]
model = prompt.get("model", "openai/gpt-4.1")
# Add runtime request (optional)
messages = messages + [{"role": "user", "content": "What should I do next?"}]
response = client.chat.completions.create(
model=model,
messages=messages,
temperature=1.0,
top_p=1.0,
)
print(response.choices[0].message.content)