diff --git a/config.toml b/config.toml index 22850d3..f949d22 100644 --- a/config.toml +++ b/config.toml @@ -50,6 +50,7 @@ generate_models = ["llama-3.2-90b-vision-preview","llama-3.3-70b-versatile","lla generate_endpoint = "http://IP:PORT/api/generate" # Can be empty if using groq use_groq = true groq_api_token = "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" # Get one at https://console.groq.com/keys +# Example API key prompt = "Analyze the given code and return an abuse score (0-10) with a brief reason. Example abuses: Crypto Mining, Shell Access, Nezha Proxy (VPN/Proxy usage), Disk Filling, Tor, DDoS, Abusive Resource Usage. Response format: '**5/10** '. No extra messages." @@ -58,4 +59,5 @@ prompt = "Analyze the given code and return an abuse score (0-10) with a brief r enabled = true webhook_url = "https://discord.com/api/webhooks/1353420407511973948/knrSGrfLDvi_60Mese1LAIBmkrK05a_L4PmyyE7R7wvGZEXiWdzrRT8pdicj0aHe88m4" +# Example webhook truncate_text = true # Used only if AI INTEGRATION is enabled, trunclates text if true to maxium allowed characters or when false splits in few webhook messages.