use std::{self, fmt, io, ptr}; use httparse; use http::{Method, Version, Uri, HttpTryFrom}; use bytes::{Bytes, BytesMut, BufMut}; use futures::{Async, AsyncSink, Poll, Sink}; use futures::unsync::mpsc::{channel, Sender}; use tokio_io::AsyncRead; use hyper::header::{Headers, ContentLength}; use {Payload, PayloadItem}; use error::{Error, Result}; use decode::Decoder; use httpmessage::{Message, HttpRequest}; const MAX_HEADERS: usize = 100; const INIT_BUFFER_SIZE: usize = 8192; pub const MAX_BUFFER_SIZE: usize = 8192 + 4096 * 100; struct PayloadInfo { tx: Sender, decoder: Decoder, tmp_item: Option, } pub struct Reader { read_buf: BytesMut, payload: Option, } enum Decoding { Paused, Ready, NotReady, } impl Reader { pub fn new() -> Reader { Reader { read_buf: BytesMut::new(), payload: None, } } #[allow(dead_code)] pub fn consume_leading_lines(&mut self) { if !self.read_buf.is_empty() { let mut i = 0; while i < self.read_buf.len() { match self.read_buf[i] { b'\r' | b'\n' => i += 1, _ => break, } } self.read_buf.split_to(i); } } fn decode(&mut self) -> std::result::Result { if let Some(ref mut payload) = self.payload { loop { if let Some(item) = payload.tmp_item.take() { let eof = item.is_eof(); match payload.tx.start_send(item) { Ok(AsyncSink::NotReady(item)) => { payload.tmp_item = Some(item); return Ok(Decoding::Paused) } Ok(AsyncSink::Ready) => { if eof { return Ok(Decoding::Ready) } }, Err(_) => return Err(Error::Incomplete), } } match payload.decoder.decode(&mut self.read_buf) { Ok(Async::Ready(Some(bytes))) => { match payload.tx.start_send(PayloadItem::Chunk(bytes)) { Ok(AsyncSink::NotReady(item)) => { payload.tmp_item = Some(item); return Ok(Decoding::Paused) } Ok(AsyncSink::Ready) => { continue } Err(_) => return Err(Error::Incomplete), } }, Ok(Async::Ready(None)) => { match payload.tx.start_send(PayloadItem::Eof) { Ok(AsyncSink::NotReady(item)) => { payload.tmp_item = Some(item); return Ok(Decoding::Paused) } Ok(AsyncSink::Ready) => { return Ok(Decoding::Ready) } Err(_) => return Err(Error::Incomplete), } }, Ok(Async::NotReady) => return Ok(Decoding::NotReady), Err(_) => return Err(Error::Incomplete), } } } else { return Ok(Decoding::Ready) } } pub fn parse(&mut self, io: &mut T) -> Poll<(HttpRequest, Option), Error> where T: AsyncRead { loop { match self.decode()? { Decoding::Paused => return Ok(Async::NotReady), Decoding::Ready => { self.payload = None; break }, Decoding::NotReady => { if 0 == try_ready!(self.read_from_io(io)) { return Err(io::Error::new( io::ErrorKind::UnexpectedEof, ParseEof).into()); } } } } loop { match try!(parse(&mut self.read_buf)) { Some((msg, decoder)) => { let payload = if let Some(decoder) = decoder { let (tx, rx) = channel(32); let payload = PayloadInfo { tx: tx, decoder: decoder, tmp_item: None, }; self.payload = Some(payload); loop { match self.decode()? { Decoding::Paused => break, Decoding::Ready => { self.payload = None; break }, Decoding::NotReady => { match self.read_from_io(io) { Ok(Async::Ready(0)) => { trace!("parse eof"); return Err(io::Error::new( io::ErrorKind::UnexpectedEof, ParseEof).into()); } Ok(Async::Ready(_)) => { continue } Ok(Async::NotReady) => break, Err(err) => return Err(err.into()), } } } } Some(rx) } else { None }; return Ok(Async::Ready((msg, payload))); }, None => { if self.read_buf.capacity() >= MAX_BUFFER_SIZE { debug!("MAX_BUFFER_SIZE reached, closing"); return Err(Error::TooLarge); } }, } if 0 == try_ready!(self.read_from_io(io)) { trace!("parse eof"); return Err(io::Error::new(io::ErrorKind::UnexpectedEof, ParseEof).into()); } } } fn read_from_io(&mut self, io: &mut T) -> Poll { if self.read_buf.remaining_mut() < INIT_BUFFER_SIZE { self.read_buf.reserve(INIT_BUFFER_SIZE); unsafe { // Zero out unused memory let buf = self.read_buf.bytes_mut(); let len = buf.len(); ptr::write_bytes(buf.as_mut_ptr(), 0, len); } } unsafe { let n = match io.read(self.read_buf.bytes_mut()) { Ok(n) => n, Err(e) => { if e.kind() == io::ErrorKind::WouldBlock { return Ok(Async::NotReady); } return Err(e) } }; self.read_buf.advance_mut(n); Ok(Async::Ready(n)) } } } #[derive(Debug)] struct ParseEof; impl fmt::Display for ParseEof { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { f.write_str("parse eof") } } impl ::std::error::Error for ParseEof { fn description(&self) -> &str { "parse eof" } } pub fn parse(buf: &mut BytesMut) -> Result)>> { if buf.is_empty() { return Ok(None); } // Parse http message let mut headers_indices = [HeaderIndices { name: (0, 0), value: (0, 0) }; MAX_HEADERS]; let (len, method, path, version, headers_len) = { let mut headers = [httparse::EMPTY_HEADER; MAX_HEADERS]; trace!("Request.parse([Header; {}], [u8; {}])", headers.len(), buf.len()); let mut req = httparse::Request::new(&mut headers); match try!(req.parse(buf)) { httparse::Status::Complete(len) => { trace!("Request.parse Complete({})", len); let method = Method::try_from(req.method.unwrap()).map_err(|_| Error::Method)?; let path = req.path.unwrap(); let bytes_ptr = buf.as_ref().as_ptr() as usize; let path_start = path.as_ptr() as usize - bytes_ptr; let path_end = path_start + path.len(); let path = (path_start, path_end); let version = if req.version.unwrap() == 1 { Version::HTTP_11 } else { Version::HTTP_10 }; record_header_indices(buf.as_ref(), req.headers, &mut headers_indices); let headers_len = req.headers.len(); (len, method, path, version, headers_len) } httparse::Status::Partial => return Ok(None), } }; let mut headers = Headers::with_capacity(headers_len); let slice = buf.split_to(len).freeze(); let path = slice.slice(path.0, path.1); // path was found to be utf8 by httparse let uri = Uri::from_shared(path).map_err(|_| Error::Uri)?; headers.extend(HeadersAsBytesIter { headers: headers_indices[..headers_len].iter(), slice: slice, }); let msg = HttpRequest::new(method, uri, version, headers); let _upgrade = msg.is_upgrade(); let chunked = msg.is_chunked()?; // Content-Length if let Some(&ContentLength(len)) = msg.headers().get() { if chunked { return Err(Error::Header) } Ok(Some((msg, Some(Decoder::length(len))))) } else if msg.headers().has::() { debug!("illegal Content-Length: {:?}", msg.headers().get_raw("Content-Length")); Err(Error::Header) } else if chunked { Ok(Some((msg, Some(Decoder::chunked())))) } else { Ok(Some((msg, None))) } } #[derive(Clone, Copy)] struct HeaderIndices { name: (usize, usize), value: (usize, usize), } fn record_header_indices(bytes: &[u8], headers: &[httparse::Header], indices: &mut [HeaderIndices]) { let bytes_ptr = bytes.as_ptr() as usize; for (header, indices) in headers.iter().zip(indices.iter_mut()) { let name_start = header.name.as_ptr() as usize - bytes_ptr; let name_end = name_start + header.name.len(); indices.name = (name_start, name_end); let value_start = header.value.as_ptr() as usize - bytes_ptr; let value_end = value_start + header.value.len(); indices.value = (value_start, value_end); } } struct HeadersAsBytesIter<'a> { headers: ::std::slice::Iter<'a, HeaderIndices>, slice: Bytes, } impl<'a> Iterator for HeadersAsBytesIter<'a> { type Item = (&'a str, Bytes); fn next(&mut self) -> Option { self.headers.next().map(|header| { let name = unsafe { let bytes = ::std::slice::from_raw_parts( self.slice.as_ref().as_ptr().offset(header.name.0 as isize), header.name.1 - header.name.0 ); ::std::str::from_utf8_unchecked(bytes) }; (name, self.slice.slice(header.value.0, header.value.1)) }) } }