diff --git a/.env b/.env index 65d8dec6..eaa3d763 100644 --- a/.env +++ b/.env @@ -104,14 +104,30 @@ OINKCODE=OPEN # Beelzebub is not part of the standard edition, please follow the README regarding setup. # It is recommended to use the Ollama backend to keep costs at bay. # Remember to rate limit API usage / set budget alerts when using ChatGPT API. -# LLMMODEL: Set to "ollama" or "gpt4-o". -# LLMHOST: When using "ollama" set it to the URL of your Ollama backend. -# OLLAMAMODEL: Set to the model you are serving on your Ollama backend, i.e. "llama3.1". -LLMMODEL: "ollama" -LLMHOST: "http://ollama.local:11434/api/chat" -OLLAMAMODEL: "llama3.1" -#LLMMODEL: "gpt4-o" -#OPENAISECRETKEY: "sk-proj-123456" +# BEELZEBUB_LLM_MODEL: Set to "ollama" or "gpt4-o". +# BEELZEBUB_LLM_HOST: When using "ollama" set it to the URL of your Ollama backend. +# BEELZEBUB_OLLAMA_MODEL: Set to the model you are serving on your Ollama backend, i.e. "openchat". +# BEELZEBUB_LLM_MODEL: "gpt4-o" +# BEELZEBUB_OPENAISECRETKEY: "sk-proj-123456" +BEELZEBUB_LLM_MODEL: "ollama" +BEELZEBUB_LLM_HOST: "http://ollama.local:11434/api/chat" +BEELZEBUB_OLLAMA_MODEL: "openchat" + +# Galah is a LLM-powered web honeypot supporting various LLM backends. +# Galah is not part of the standard edition, please follow the README regarding setup. +# It is recommended to use the Ollama backend to keep costs at bay. +# Remember to rate limit API usage / set budget alerts when using ChatGPT API. +# GALAH_LLM_PROVIDER: Set to "ollama" or "gpt4-o". +# GALAH_LLM_SERVER_URL: When using "ollama" set it to the URL of your Ollama backend. +# GALAH_LLM_MODEL: Set to the model you are serving on your Ollama backend, i.e. "llama3". +# GALAH_LLM_TEMPERATURE: "1" +# GALAH_LLM_API_KEY: "sk-XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX" +# GALAH_LLM_CLOUD_LOCATION: "" +# GALAH_LLM_CLOUD_PROJECT: "" +GALAH_LLM_PROVIDER: "ollama" +GALAH_LLM_SERVER_URL: "http://ollama.local:11434" +GALAH_LLM_MODEL: "llama3" + ################################################################################### # NEVER MAKE CHANGES TO THIS SECTION UNLESS YOU REALLY KNOW WHAT YOU ARE DOING!!! # diff --git a/compose/llm.yml b/compose/llm.yml index 7317ea57..1469816c 100644 --- a/compose/llm.yml +++ b/compose/llm.yml @@ -81,7 +81,7 @@ services: - "80:80" - "443:443" - "8443:8443" - - "8888:8888" + - "8080:8080" image: ${TPOT_REPO}/galah:${TPOT_VERSION} environment: LLM_PROVIDER: ${GALAH_LLM_PROVIDER} diff --git a/docker/beelzebub/docker-compose.yml b/docker/beelzebub/docker-compose.yml index 6fc31a31..8f5f1392 100644 --- a/docker/beelzebub/docker-compose.yml +++ b/docker/beelzebub/docker-compose.yml @@ -14,11 +14,10 @@ services: - beelzebub_local ports: - "22:22" - - "2222:2222" - - "8080:8080" - - "8081:8081" - "80:80" + - "2222:2222" - "3306:3306" + - "8080:8080" environment: LLMMODEL: "ollama" LLMHOST: "http://ollama.local:11434/api/chat" diff --git a/docker/galah/docker-compose.yml b/docker/galah/docker-compose.yml index 3bf660fa..4e502a87 100644 --- a/docker/galah/docker-compose.yml +++ b/docker/galah/docker-compose.yml @@ -16,7 +16,7 @@ services: - "80:80" - "443:443" - "8443:8443" - - "8888:8888" + - "8080:8080" image: dtagdevsec/galah:24.04 environment: LLM_PROVIDER: "ollama" diff --git a/docker/tpotinit/dist/etc/objects/elkbase.tgz b/docker/tpotinit/dist/etc/objects/elkbase.tgz index d28752f6..1147d299 100644 Binary files a/docker/tpotinit/dist/etc/objects/elkbase.tgz and b/docker/tpotinit/dist/etc/objects/elkbase.tgz differ diff --git a/docker/tpotinit/dist/etc/objects/kibana_export.ndjson.zip b/docker/tpotinit/dist/etc/objects/kibana_export.ndjson.zip index 9419af18..b1c3a4e4 100644 Binary files a/docker/tpotinit/dist/etc/objects/kibana_export.ndjson.zip and b/docker/tpotinit/dist/etc/objects/kibana_export.ndjson.zip differ diff --git a/env.example b/env.example index d7329bbc..eaa3d763 100644 --- a/env.example +++ b/env.example @@ -106,7 +106,7 @@ OINKCODE=OPEN # Remember to rate limit API usage / set budget alerts when using ChatGPT API. # BEELZEBUB_LLM_MODEL: Set to "ollama" or "gpt4-o". # BEELZEBUB_LLM_HOST: When using "ollama" set it to the URL of your Ollama backend. -# BEELZEBUB_OLLAMA_MODEL: Set to the model you are serving on your Ollama backend, i.e. "llama3.1". +# BEELZEBUB_OLLAMA_MODEL: Set to the model you are serving on your Ollama backend, i.e. "openchat". # BEELZEBUB_LLM_MODEL: "gpt4-o" # BEELZEBUB_OPENAISECRETKEY: "sk-proj-123456" BEELZEBUB_LLM_MODEL: "ollama" @@ -119,7 +119,7 @@ BEELZEBUB_OLLAMA_MODEL: "openchat" # Remember to rate limit API usage / set budget alerts when using ChatGPT API. # GALAH_LLM_PROVIDER: Set to "ollama" or "gpt4-o". # GALAH_LLM_SERVER_URL: When using "ollama" set it to the URL of your Ollama backend. -# GALAH_LLM_MODEL: Set to the model you are serving on your Ollama backend, i.e. "llama3.1". +# GALAH_LLM_MODEL: Set to the model you are serving on your Ollama backend, i.e. "llama3". # GALAH_LLM_TEMPERATURE: "1" # GALAH_LLM_API_KEY: "sk-XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX" # GALAH_LLM_CLOUD_LOCATION: ""