diff --git a/server/src/handlers/http/llm.rs b/server/src/handlers/http/llm.rs index ef8feccc0..1ed3dc6eb 100644 --- a/server/src/handlers/http/llm.rs +++ b/server/src/handlers/http/llm.rs @@ -139,7 +139,7 @@ pub async fn make_llm_request(body: web::Json) -> Result) -> HttpResponse { +pub async fn is_llm_active() -> HttpResponse { let is_active = matches!(&CONFIG.parseable.open_ai_key, Some(api_key) if api_key.len() > 3); HttpResponse::Ok() .content_type("application/json") diff --git a/server/src/option.rs b/server/src/option.rs index 3da036d0c..610e0c209 100644 --- a/server/src/option.rs +++ b/server/src/option.rs @@ -369,7 +369,7 @@ impl Server { .arg( Arg::new(Self::OPEN_AI_KEY) .long(Self::OPEN_AI_KEY) - .env("OPENAI_API_KEY") + .env("P_OPENAI_API_KEY") .value_name("STRING") .required(false) .help("Set OpenAI key to enable llm feature"),