mirror of
https://git.deuxfleurs.fr/Deuxfleurs/garage.git
synced 2024-11-25 17:41:01 +00:00
Slightly improved S3 compatibility
- ListBucket does not require any of the parameters (delimiter, prefix, max-keys, etc) - URLs are properly percent_decoded - PutObject and DeleteObject calls now answer correctly (empty body, version id in the x-amz-version-id: header)
This commit is contained in:
parent
3686f100b7
commit
3324971701
6 changed files with 83 additions and 47 deletions
1
Cargo.lock
generated
1
Cargo.lock
generated
|
@ -374,6 +374,7 @@ dependencies = [
|
||||||
"httpdate 0.3.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
"httpdate 0.3.2 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"hyper 0.13.5 (registry+https://github.com/rust-lang/crates.io-index)",
|
"hyper 0.13.5 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"log 0.4.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
"log 0.4.8 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
"percent-encoding 2.1.0 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"sha2 0.8.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
"sha2 0.8.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"tokio 0.2.18 (registry+https://github.com/rust-lang/crates.io-index)",
|
"tokio 0.2.18 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"url 2.1.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
"url 2.1.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
|
|
@ -30,5 +30,5 @@ http = "0.2"
|
||||||
hyper = "0.13"
|
hyper = "0.13"
|
||||||
url = "2.1"
|
url = "2.1"
|
||||||
httpdate = "0.3"
|
httpdate = "0.3"
|
||||||
|
percent-encoding = "2.1.0"
|
||||||
|
|
||||||
|
|
|
@ -73,7 +73,11 @@ async fn handler_inner(
|
||||||
req: Request<Body>,
|
req: Request<Body>,
|
||||||
) -> Result<Response<BodyType>, Error> {
|
) -> Result<Response<BodyType>, Error> {
|
||||||
let path = req.uri().path().to_string();
|
let path = req.uri().path().to_string();
|
||||||
let (bucket, key) = parse_bucket_key(path.as_str())?;
|
let path = percent_encoding::percent_decode_str(&path)
|
||||||
|
.decode_utf8()
|
||||||
|
.map_err(|e| Error::BadRequest(format!("Invalid utf8 key ({})", e)))?;
|
||||||
|
|
||||||
|
let (bucket, key) = parse_bucket_key(&path)?;
|
||||||
if bucket.len() == 0 {
|
if bucket.len() == 0 {
|
||||||
return Err(Error::Forbidden(format!(
|
return Err(Error::Forbidden(format!(
|
||||||
"Operations on buckets not allowed"
|
"Operations on buckets not allowed"
|
||||||
|
@ -120,7 +124,12 @@ async fn handler_inner(
|
||||||
} else if req.headers().contains_key("x-amz-copy-source") {
|
} else if req.headers().contains_key("x-amz-copy-source") {
|
||||||
// CopyObject query
|
// CopyObject query
|
||||||
let copy_source = req.headers().get("x-amz-copy-source").unwrap().to_str()?;
|
let copy_source = req.headers().get("x-amz-copy-source").unwrap().to_str()?;
|
||||||
let (source_bucket, source_key) = parse_bucket_key(copy_source)?;
|
let copy_source = percent_encoding::percent_decode_str(©_source)
|
||||||
|
.decode_utf8()
|
||||||
|
.map_err(|e| {
|
||||||
|
Error::BadRequest(format!("Invalid utf8 copy_source ({})", e))
|
||||||
|
})?;
|
||||||
|
let (source_bucket, source_key) = parse_bucket_key(©_source)?;
|
||||||
if !api_key.allow_read(&source_bucket) {
|
if !api_key.allow_read(&source_bucket) {
|
||||||
return Err(Error::Forbidden(format!(
|
return Err(Error::Forbidden(format!(
|
||||||
"Reading from bucket {} not allowed for this key",
|
"Reading from bucket {} not allowed for this key",
|
||||||
|
@ -145,8 +154,7 @@ async fn handler_inner(
|
||||||
} else {
|
} else {
|
||||||
// DeleteObject query
|
// DeleteObject query
|
||||||
let version_uuid = handle_delete(garage, &bucket, &key).await?;
|
let version_uuid = handle_delete(garage, &bucket, &key).await?;
|
||||||
let response = format!("{}\n", hex::encode(version_uuid));
|
Ok(put_response(version_uuid))
|
||||||
Ok(Response::new(Box::new(BytesBody::from(response))))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
&Method::POST => {
|
&Method::POST => {
|
||||||
|
@ -170,9 +178,14 @@ async fn handler_inner(
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
match req.method() {
|
match req.method() {
|
||||||
&Method::PUT | &Method::HEAD => {
|
&Method::PUT => {
|
||||||
// If PUT: CreateBucket, if HEAD: HeadBucket
|
// CreateBucket
|
||||||
// If we're here, the bucket already exists, so just answer ok
|
// If we're here, the bucket already exists, so just answer ok
|
||||||
|
println!(
|
||||||
|
"Body: {}",
|
||||||
|
std::str::from_utf8(&hyper::body::to_bytes(req.into_body()).await?)
|
||||||
|
.unwrap_or("<invalid utf8>")
|
||||||
|
);
|
||||||
let empty_body: BodyType = Box::new(BytesBody::from(vec![]));
|
let empty_body: BodyType = Box::new(BytesBody::from(vec![]));
|
||||||
let response = Response::builder()
|
let response = Response::builder()
|
||||||
.header("Location", format!("/{}", bucket))
|
.header("Location", format!("/{}", bucket))
|
||||||
|
@ -180,6 +193,12 @@ async fn handler_inner(
|
||||||
.unwrap();
|
.unwrap();
|
||||||
Ok(response)
|
Ok(response)
|
||||||
}
|
}
|
||||||
|
&Method::HEAD => {
|
||||||
|
// HeadBucket
|
||||||
|
let empty_body: BodyType = Box::new(BytesBody::from(vec![]));
|
||||||
|
let response = Response::builder().body(empty_body).unwrap();
|
||||||
|
Ok(response)
|
||||||
|
}
|
||||||
&Method::DELETE => {
|
&Method::DELETE => {
|
||||||
// DeleteBucket query
|
// DeleteBucket query
|
||||||
Err(Error::Forbidden(
|
Err(Error::Forbidden(
|
||||||
|
@ -187,38 +206,32 @@ async fn handler_inner(
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
&Method::GET => {
|
&Method::GET => {
|
||||||
if params.contains_key(&"prefix".to_string()) {
|
// ListObjects query
|
||||||
// ListObjects query
|
let delimiter = params.get("delimiter").map(|x| x.as_str()).unwrap_or(&"");
|
||||||
let delimiter = params.get("delimiter").map(|x| x.as_str()).unwrap_or(&"");
|
let max_keys = params
|
||||||
let max_keys = params
|
.get("max-keys")
|
||||||
.get("max-keys")
|
.map(|x| {
|
||||||
.map(|x| {
|
x.parse::<usize>().map_err(|e| {
|
||||||
x.parse::<usize>().map_err(|e| {
|
Error::BadRequest(format!("Invalid value for max-keys: {}", e))
|
||||||
Error::BadRequest(format!("Invalid value for max-keys: {}", e))
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
.unwrap_or(Ok(1000))?;
|
})
|
||||||
let prefix = params.get("prefix").unwrap();
|
.unwrap_or(Ok(1000))?;
|
||||||
let urlencode_resp = params
|
let prefix = params.get("prefix").map(|x| x.as_str()).unwrap_or(&"");
|
||||||
.get("encoding-type")
|
let urlencode_resp = params
|
||||||
.map(|x| x == "url")
|
.get("encoding-type")
|
||||||
.unwrap_or(false);
|
.map(|x| x == "url")
|
||||||
let marker = params.get("marker").map(String::as_str);
|
.unwrap_or(false);
|
||||||
Ok(handle_list(
|
let marker = params.get("marker").map(String::as_str);
|
||||||
garage,
|
Ok(handle_list(
|
||||||
bucket,
|
garage,
|
||||||
delimiter,
|
bucket,
|
||||||
max_keys,
|
delimiter,
|
||||||
prefix,
|
max_keys,
|
||||||
marker,
|
prefix,
|
||||||
urlencode_resp,
|
marker,
|
||||||
)
|
urlencode_resp,
|
||||||
.await?)
|
)
|
||||||
} else {
|
.await?)
|
||||||
Err(Error::BadRequest(format!(
|
|
||||||
"Not a list call, so what is it?"
|
|
||||||
)))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
_ => Err(Error::BadRequest(format!("Invalid method"))),
|
_ => Err(Error::BadRequest(format!("Invalid method"))),
|
||||||
}
|
}
|
||||||
|
@ -229,7 +242,14 @@ fn parse_bucket_key(path: &str) -> Result<(&str, Option<&str>), Error> {
|
||||||
let path = path.trim_start_matches('/');
|
let path = path.trim_start_matches('/');
|
||||||
|
|
||||||
match path.find('/') {
|
match path.find('/') {
|
||||||
Some(i) => Ok((&path[..i], Some(&path[i + 1..]))),
|
Some(i) => {
|
||||||
|
let key = &path[i + 1..];
|
||||||
|
if key.len() > 0 {
|
||||||
|
Ok((&path[..i], Some(key)))
|
||||||
|
} else {
|
||||||
|
Ok((&path[..i], None))
|
||||||
|
}
|
||||||
|
}
|
||||||
None => Ok((path, None)),
|
None => Ok((path, None)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -82,3 +82,7 @@ impl From<Vec<u8>> for BytesBody {
|
||||||
Self::new(Bytes::from(x))
|
Self::new(Bytes::from(x))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn empty_body() -> BodyType {
|
||||||
|
Box::new(BytesBody::from(vec![]))
|
||||||
|
}
|
||||||
|
|
|
@ -29,6 +29,7 @@ pub async fn handle_list(
|
||||||
) -> Result<Response<BodyType>, Error> {
|
) -> Result<Response<BodyType>, Error> {
|
||||||
let mut result_keys = BTreeMap::<String, ListResultInfo>::new();
|
let mut result_keys = BTreeMap::<String, ListResultInfo>::new();
|
||||||
let mut result_common_prefixes = BTreeSet::<String>::new();
|
let mut result_common_prefixes = BTreeSet::<String>::new();
|
||||||
|
|
||||||
let mut truncated = true;
|
let mut truncated = true;
|
||||||
let mut next_chunk_start = marker.unwrap_or(prefix).to_string();
|
let mut next_chunk_start = marker.unwrap_or(prefix).to_string();
|
||||||
|
|
||||||
|
@ -44,12 +45,19 @@ pub async fn handle_list(
|
||||||
max_keys,
|
max_keys,
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
|
debug!(
|
||||||
|
"List: get range {} (max {}), results: {}",
|
||||||
|
next_chunk_start,
|
||||||
|
max_keys,
|
||||||
|
objects.len()
|
||||||
|
);
|
||||||
|
|
||||||
for object in objects.iter() {
|
for object in objects.iter() {
|
||||||
|
if !object.key.starts_with(prefix) {
|
||||||
|
truncated = false;
|
||||||
|
break;
|
||||||
|
}
|
||||||
if let Some(version) = object.versions().iter().find(|x| x.is_data()) {
|
if let Some(version) = object.versions().iter().find(|x| x.is_data()) {
|
||||||
if !object.key.starts_with(prefix) {
|
|
||||||
truncated = false;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
let common_prefix = if delimiter.len() > 0 {
|
let common_prefix = if delimiter.len() > 0 {
|
||||||
let relative_key = &object.key[prefix.len()..];
|
let relative_key = &object.key[prefix.len()..];
|
||||||
match relative_key.find(delimiter) {
|
match relative_key.find(delimiter) {
|
||||||
|
@ -117,7 +125,7 @@ pub async fn handle_list(
|
||||||
for pfx in result_common_prefixes.iter() {
|
for pfx in result_common_prefixes.iter() {
|
||||||
writeln!(
|
writeln!(
|
||||||
&mut xml,
|
&mut xml,
|
||||||
"\t<Prefix>{}</Prefix>",
|
"\t\t<Prefix>{}</Prefix>",
|
||||||
xml_escape(pfx),
|
xml_escape(pfx),
|
||||||
//xml_encode_key(pfx, urlencode_resp)
|
//xml_encode_key(pfx, urlencode_resp)
|
||||||
)
|
)
|
||||||
|
@ -126,6 +134,7 @@ pub async fn handle_list(
|
||||||
writeln!(&mut xml, "\t</CommonPrefixes>").unwrap();
|
writeln!(&mut xml, "\t</CommonPrefixes>").unwrap();
|
||||||
}
|
}
|
||||||
writeln!(&mut xml, "</ListBucketResult>").unwrap();
|
writeln!(&mut xml, "</ListBucketResult>").unwrap();
|
||||||
|
println!("{}", xml);
|
||||||
|
|
||||||
Ok(Response::new(Box::new(BytesBody::from(xml.into_bytes()))))
|
Ok(Response::new(Box::new(BytesBody::from(xml.into_bytes()))))
|
||||||
}
|
}
|
||||||
|
|
|
@ -195,9 +195,11 @@ impl BodyChunker {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn put_response(version_uuid: UUID) -> Response<BodyType> {
|
pub fn put_response(version_uuid: UUID) -> Response<BodyType> {
|
||||||
let resp_bytes = format!("{}\n", hex::encode(version_uuid));
|
Response::builder()
|
||||||
Response::new(Box::new(BytesBody::from(resp_bytes)))
|
.header("x-amz-version-id", hex::encode(version_uuid))
|
||||||
|
.body(empty_body())
|
||||||
|
.unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn handle_create_multipart_upload(
|
pub async fn handle_create_multipart_upload(
|
||||||
|
|
Loading…
Reference in a new issue