Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@ This section elaborates available options to run Parseable in production or deve
## Features :rocket:

- [High availability & Cluster mode ↗︎](https://www.parseable.com/docs/concepts/distributed-architecture)
- [Local cache ↗︎](https://www.parseable.com/docs/features/tiering)
- [Hot Tier ↗︎](https://www.parseable.com/docs/features/tiering)
- [Alerts ↗︎](https://www.parseable.com/docs/alerts)
- [Role based access control ↗︎](https://www.parseable.com/docs/rbac)
- [OAuth2 support ↗︎](https://www.parseable.com/docs/oidc)
Expand Down
1 change: 0 additions & 1 deletion src/handlers/http/about.rs
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,6 @@ use std::path::PathBuf;
/// "license": "AGPL-3.0-only",
/// "mode": mode,
/// "staging": staging,
/// "cache": cache_details,
/// "grpcPort": grpc_port,
/// "store": {
/// "type": CONFIG.get_storage_mode_string(),
Expand Down
1 change: 0 additions & 1 deletion src/handlers/http/logstream.rs
Original file line number Diff line number Diff line change
Expand Up @@ -567,7 +567,6 @@ pub async fn get_stream_info(req: HttpRequest) -> Result<impl Responder, StreamE
.time_partition_limit
.map(|limit| limit.to_string()),
custom_partition: stream_meta.custom_partition.clone(),
cache_enabled: stream_meta.cache_enabled,
static_schema_flag: stream_meta.static_schema_flag.clone(),
};

Expand Down
8 changes: 0 additions & 8 deletions src/handlers/http/modal/ingest_server.rs
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,6 @@ use super::OpenIdClient;
use super::ParseableServer;
use crate::analytics;
use crate::handlers::airplane;
use crate::handlers::http::caching_removed;
use crate::handlers::http::ingest;
use crate::handlers::http::logstream;
use crate::handlers::http::middleware::DisAllowRootUser;
Expand Down Expand Up @@ -268,13 +267,6 @@ impl IngestServer {
.authorize_for_stream(Action::GetStats),
),
)
.service(
web::resource("/cache")
// PUT "/logstream/{logstream}/cache" ==> caching has been deprecated
.route(web::put().to(caching_removed))
// GET "/logstream/{logstream}/cache" ==> caching has been deprecated
.route(web::get().to(caching_removed)),
)
.service(
web::scope("/retention").service(
web::resource("/cleanup").route(
Expand Down
10 changes: 1 addition & 9 deletions src/handlers/http/modal/query_server.rs
Original file line number Diff line number Diff line change
Expand Up @@ -18,11 +18,11 @@

use crate::correlation::CORRELATIONS;
use crate::handlers::airplane;
use crate::handlers::http::base_path;
use crate::handlers::http::cluster::{self, init_cluster_metrics_schedular};
use crate::handlers::http::logstream::create_internal_stream_if_not_exists;
use crate::handlers::http::middleware::{DisAllowRootUser, RouteExt};
use crate::handlers::http::{self, role};
use crate::handlers::http::{base_path, caching_removed};
use crate::handlers::http::{logstream, MAX_EVENT_PAYLOAD_SIZE};
use crate::hottier::HotTierManager;
use crate::rbac::role::Action;
Expand Down Expand Up @@ -54,7 +54,6 @@ impl ParseableServer for QueryServer {
.service(Server::get_correlation_webscope())
.service(Server::get_query_factory())
.service(Server::get_trino_factory())
.service(Server::get_cache_webscope())
.service(Server::get_liveness_factory())
.service(Server::get_readiness_factory())
.service(Server::get_about_factory())
Expand Down Expand Up @@ -331,13 +330,6 @@ impl QueryServer {
.authorize_for_stream(Action::GetRetention),
),
)
.service(
web::resource("/cache")
// PUT "/logstream/{logstream}/cache" ==> caching has been deprecated
.route(web::put().to(caching_removed))
// GET "/logstream/{logstream}/cache" ==> caching has been deprecated
.route(web::get().to(caching_removed)),
)
.service(
web::resource("/hottier")
// PUT "/logstream/{logstream}/hottier" ==> Set hottier for given logstream
Expand Down
21 changes: 0 additions & 21 deletions src/handlers/http/modal/server.rs
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@ use crate::correlation::CORRELATIONS;
use crate::handlers;
use crate::handlers::http::about;
use crate::handlers::http::base_path;
use crate::handlers::http::caching_removed;
use crate::handlers::http::health_check;
use crate::handlers::http::query;
use crate::handlers::http::trino;
Expand Down Expand Up @@ -71,7 +70,6 @@ impl ParseableServer for Server {
.service(Self::get_correlation_webscope())
.service(Self::get_query_factory())
.service(Self::get_trino_factory())
.service(Self::get_cache_webscope())
.service(Self::get_ingest_factory())
.service(Self::get_liveness_factory())
.service(Self::get_readiness_factory())
Expand Down Expand Up @@ -282,18 +280,6 @@ impl Server {
web::resource("/query").route(web::post().to(query::query).authorize(Action::Query))
}

pub fn get_cache_webscope() -> Scope {
web::scope("/cache").service(
web::scope("/{user_id}").service(
web::scope("/{stream}").service(
web::resource("")
.route(web::get().to(caching_removed))
.route(web::post().to(caching_removed)),
),
),
)
}

// get the logstream web scope
pub fn get_logstream_webscope() -> Scope {
web::scope("/logstream")
Expand Down Expand Up @@ -391,13 +377,6 @@ impl Server {
.authorize_for_stream(Action::GetRetention),
),
)
.service(
web::resource("/cache")
// PUT "/logstream/{logstream}/cache" ==> Set retention for given logstream
.route(web::put().to(caching_removed))
// GET "/logstream/{logstream}/cache" ==> Get retention for given logstream
.route(web::get().to(caching_removed)),
)
.service(
web::resource("/hottier")
// PUT "/logstream/{logstream}/hottier" ==> Set hottier for given logstream
Expand Down
3 changes: 0 additions & 3 deletions src/metadata.rs
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,6 @@ pub struct LogStreamMetadata {
pub schema: HashMap<String, Arc<Field>>,
pub alerts: Alerts,
pub retention: Option<Retention>,
pub cache_enabled: bool,
pub created_at: String,
pub first_event_at: Option<String>,
pub time_partition: Option<String>,
Expand Down Expand Up @@ -430,7 +429,6 @@ pub async fn load_stream_metadata_on_server_start(
created_at,
first_event_at,
retention,
cache_enabled,
snapshot,
stats,
time_partition,
Expand Down Expand Up @@ -467,7 +465,6 @@ pub async fn load_stream_metadata_on_server_start(
schema,
alerts,
retention,
cache_enabled,
created_at,
first_event_at,
time_partition,
Expand Down
18 changes: 3 additions & 15 deletions src/metrics/prom_utils.rs
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,6 @@ pub struct Metrics {
event_time: NaiveDateTime,
commit: String,
staging: String,
cache: String,
}

#[derive(Debug, Serialize, Default, Clone)]
Expand Down Expand Up @@ -85,7 +84,6 @@ impl Default for Metrics {
event_time: Utc::now().naive_utc(),
commit: "".to_string(),
staging: "".to_string(),
cache: "".to_string(),
}
}
}
Expand All @@ -109,7 +107,6 @@ impl Metrics {
event_time: Utc::now().naive_utc(),
commit: "".to_string(),
staging: "".to_string(),
cache: "".to_string(),
}
}
}
Expand Down Expand Up @@ -206,7 +203,7 @@ impl Metrics {
}
}
}
let (commit_id, staging, cache) = Self::from_about_api_response(ingestor_metadata.clone())
let (commit_id, staging) = Self::from_about_api_response(ingestor_metadata.clone())
.await
.map_err(|err| {
error!("Fatal: failed to get ingestor info: {:?}", err);
Expand All @@ -215,14 +212,13 @@ impl Metrics {

prom_dress.commit = commit_id;
prom_dress.staging = staging;
prom_dress.cache = cache;

Ok(prom_dress)
}

pub async fn from_about_api_response(
ingestor_metadata: IngestorMetadata,
) -> Result<(String, String, String), PostError> {
) -> Result<(String, String), PostError> {
let uri = Url::parse(&format!(
"{}{}/about",
&ingestor_metadata.domain_name,
Expand All @@ -249,15 +245,7 @@ impl Metrics {
.get("staging")
.and_then(|x| x.as_str())
.unwrap_or_default();
let cache = about_api_json
.get("cache")
.and_then(|x| x.as_str())
.unwrap_or_default();
Ok((
commit_id.to_string(),
staging.to_string(),
cache.to_string(),
))
Ok((commit_id.to_string(), staging.to_string()))
} else {
warn!(
"Failed to fetch about API response from ingestor: {}\n",
Expand Down
1 change: 0 additions & 1 deletion src/option.rs
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,6 @@ use serde::{Deserialize, Serialize};
use std::env;
use std::path::PathBuf;
use std::sync::Arc;
pub const MIN_CACHE_SIZE_BYTES: u64 = 1073741824; // 1 GiB

pub const JOIN_COMMUNITY: &str =
"Join us on Parseable Slack community for questions : https://logg.ing/community";
Expand Down
5 changes: 0 additions & 5 deletions src/storage/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -96,8 +96,6 @@ pub struct ObjectStoreFormat {
pub stats: FullStats,
#[serde(default)]
pub snapshot: Snapshot,
#[serde(default)]
pub cache_enabled: bool,
#[serde(skip_serializing_if = "Option::is_none")]
pub retention: Option<Retention>,
#[serde(skip_serializing_if = "Option::is_none")]
Expand All @@ -120,8 +118,6 @@ pub struct StreamInfo {
#[serde(rename = "first-event-at")]
#[serde(skip_serializing_if = "Option::is_none")]
pub first_event_at: Option<String>,
#[serde(default)]
pub cache_enabled: bool,
#[serde(skip_serializing_if = "Option::is_none")]
pub time_partition: Option<String>,
#[serde(skip_serializing_if = "Option::is_none")]
Expand Down Expand Up @@ -191,7 +187,6 @@ impl Default for ObjectStoreFormat {
permissions: vec![Permisssion::new("parseable".to_string())],
stats: FullStats::default(),
snapshot: Snapshot::default(),
cache_enabled: false,
retention: None,
time_partition: None,
time_partition_limit: None,
Expand Down
Loading