gstreamer-rs/gstreamer/src/toc_serde.rs

548 lines
22 KiB
Rust

// Take a look at the license at the top of the repository in the LICENSE file.
use serde::{
de::{Deserialize, Deserializer},
ser::{Serialize, SerializeStruct, Serializer},
};
use crate::{toc::*, TagList, TocEntryType, TocLoopType, TocScope};
impl Serialize for TocRef {
fn serialize<S: Serializer>(&self, serializer: S) -> Result<S::Ok, S::Error> {
let mut toc = serializer.serialize_struct("Toc", 3)?;
toc.serialize_field("scope", &self.scope())?;
toc.serialize_field("tags", &self.tags())?;
toc.serialize_field("entries", &self.entries())?;
toc.end()
}
}
impl Serialize for Toc {
fn serialize<S: Serializer>(&self, serializer: S) -> Result<S::Ok, S::Error> {
self.as_ref().serialize(serializer)
}
}
impl Serialize for TocEntryRef {
fn serialize<S: Serializer>(&self, serializer: S) -> Result<S::Ok, S::Error> {
let mut toc_entry = serializer.serialize_struct("TocEntry", 6)?;
toc_entry.serialize_field("entry_type", &self.entry_type())?;
toc_entry.serialize_field("uid", &self.uid())?;
toc_entry.serialize_field("start_stop", &self.start_stop_times())?;
toc_entry.serialize_field("tags", &self.tags())?;
toc_entry.serialize_field("loop", &self.loop_())?;
toc_entry.serialize_field("sub_entries", &self.sub_entries())?;
toc_entry.end()
}
}
impl Serialize for TocEntry {
fn serialize<S: Serializer>(&self, serializer: S) -> Result<S::Ok, S::Error> {
self.as_ref().serialize(serializer)
}
}
#[derive(serde::Deserialize)]
struct TocDe {
scope: TocScope,
tags: Option<TagList>,
entries: Vec<TocEntry>,
}
impl From<TocDe> for Toc {
fn from(mut toc_de: TocDe) -> Self {
skip_assert_initialized!();
let mut toc = Toc::new(toc_de.scope);
{
let toc = toc.get_mut().unwrap();
toc.set_tags(toc_de.tags.take());
let entry_iter = toc_de.entries.drain(..);
for entry in entry_iter {
toc.append_entry(entry);
}
}
toc
}
}
impl<'de> Deserialize<'de> for Toc {
fn deserialize<D: Deserializer<'de>>(deserializer: D) -> Result<Self, D::Error> {
skip_assert_initialized!();
TocDe::deserialize(deserializer).map(|toc_de| toc_de.into())
}
}
#[derive(serde::Deserialize)]
struct TocEntryDe {
entry_type: TocEntryType,
uid: String,
start_stop: Option<(i64, i64)>,
tags: Option<TagList>,
#[serde(rename = "loop")]
loop_: Option<(TocLoopType, i32)>,
sub_entries: Vec<TocEntry>,
}
impl From<TocEntryDe> for TocEntry {
fn from(mut toc_entry_de: TocEntryDe) -> Self {
skip_assert_initialized!();
let mut toc_entry = TocEntry::new(toc_entry_de.entry_type, toc_entry_de.uid.as_str());
{
let toc_entry = toc_entry.get_mut().unwrap();
if let Some(start_stop) = toc_entry_de.start_stop.take() {
toc_entry.set_start_stop_times(start_stop.0, start_stop.1);
}
toc_entry.set_tags(toc_entry_de.tags.take());
if let Some(loop_) = toc_entry_de.loop_.take() {
toc_entry.set_loop(loop_.0, loop_.1);
}
let entry_iter = toc_entry_de.sub_entries.drain(..);
for sub_entries in entry_iter {
toc_entry.append_sub_entry(sub_entries);
}
}
toc_entry
}
}
impl<'de> Deserialize<'de> for TocEntry {
fn deserialize<D: Deserializer<'de>>(deserializer: D) -> Result<Self, D::Error> {
skip_assert_initialized!();
TocEntryDe::deserialize(deserializer).map(|toc_entry_de| toc_entry_de.into())
}
}
#[cfg(test)]
mod tests {
use crate::{tags::Title, toc::*, TagList, TagMergeMode, TocEntryType, TocScope};
#[test]
fn test_serialize() {
crate::init().unwrap();
let mut toc = Toc::new(TocScope::Global);
{
let toc = toc.get_mut().unwrap();
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"toc", TagMergeMode::Append);
toc.set_tags(tags);
let mut toc_edition = TocEntry::new(TocEntryType::Edition, "edition");
{
let toc_edition = toc_edition.get_mut().unwrap();
toc_edition.set_start_stop_times(0, 15);
let mut toc_chap_1 = TocEntry::new(TocEntryType::Chapter, "chapter1");
{
let toc_chap_1 = toc_chap_1.get_mut().unwrap();
toc_chap_1.set_start_stop_times(0, 10);
let mut toc_chap_1_1 = TocEntry::new(TocEntryType::Chapter, "chapter1.1");
{
let toc_chap_1_1 = toc_chap_1_1.get_mut().unwrap();
toc_chap_1_1.set_start_stop_times(0, 4);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 1.1", TagMergeMode::Append);
toc_chap_1_1.set_tags(tags);
}
toc_chap_1.append_sub_entry(toc_chap_1_1);
let mut toc_chap_1_2 = TocEntry::new(TocEntryType::Chapter, "chapter1.2");
{
let toc_chap_1_2 = toc_chap_1_2.get_mut().unwrap();
toc_chap_1_2.set_start_stop_times(4, 10);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 1.2", TagMergeMode::Append);
toc_chap_1_2.set_tags(tags);
}
toc_chap_1.append_sub_entry(toc_chap_1_2);
}
toc_edition.append_sub_entry(toc_chap_1);
let mut toc_chap_2 = TocEntry::new(TocEntryType::Chapter, "chapter2");
{
let toc_chap_2 = toc_chap_2.get_mut().unwrap();
toc_chap_2.set_start_stop_times(10, 15);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 2", TagMergeMode::Append);
toc_chap_2.set_tags(tags);
}
toc_edition.append_sub_entry(toc_chap_2);
}
toc.append_entry(toc_edition);
}
let pretty_config = ron::ser::PrettyConfig::new().new_line("".to_string());
let res = ron::ser::to_string_pretty(&toc, pretty_config);
assert_eq!(
Ok(concat!(
"(",
" scope: Global,",
" tags: Some((",
" scope: Stream,",
" tags: [",
" (\"title\", [",
" \"toc\",",
" ]),",
" ],",
" )),",
" entries: [",
" (",
" entry_type: Edition,",
" uid: \"edition\",",
" start_stop: Some((0, 15)),",
" tags: None,",
" loop: Some((None, 0)),",
" sub_entries: [",
" (",
" entry_type: Chapter,",
" uid: \"chapter1\",",
" start_stop: Some((0, 10)),",
" tags: None,",
" loop: Some((None, 0)),",
" sub_entries: [",
" (",
" entry_type: Chapter,",
" uid: \"chapter1.1\",",
" start_stop: Some((0, 4)),",
" tags: Some((",
" scope: Stream,",
" tags: [",
" (\"title\", [",
" \"chapter 1.1\",",
" ]),",
" ],",
" )),",
" loop: Some((None, 0)),",
" sub_entries: [],",
" ),",
" (",
" entry_type: Chapter,",
" uid: \"chapter1.2\",",
" start_stop: Some((4, 10)),",
" tags: Some((",
" scope: Stream,",
" tags: [",
" (\"title\", [",
" \"chapter 1.2\",",
" ]),",
" ],",
" )),",
" loop: Some((None, 0)),",
" sub_entries: [],",
" ),",
" ],",
" ),",
" (",
" entry_type: Chapter,",
" uid: \"chapter2\",",
" start_stop: Some((10, 15)),",
" tags: Some((",
" scope: Stream,",
" tags: [",
" (\"title\", [",
" \"chapter 2\",",
" ]),",
" ],",
" )),",
" loop: Some((None, 0)),",
" sub_entries: [],",
" ),",
" ],",
" ),",
" ],",
")",
)
.to_owned()),
res,
);
}
#[allow(clippy::cognitive_complexity)]
#[test]
fn test_deserialize() {
use crate::tags::Title;
crate::init().unwrap();
let toc_ron = r#"
(
scope: Global,
tags: Some((
scope: Stream,
tags: [
("title", ["toc"]),
],
)),
entries: [
(
entry_type: Edition,
uid: "edition",
start_stop: Some((0, 15)),
tags: None,
loop: Some((None, 0)),
sub_entries: [
(
entry_type: Chapter,
uid: "chapter1",
start_stop: Some((0, 10)),
tags: None,
loop: Some((None, 0)),
sub_entries: [
(
entry_type: Chapter,
uid: "chapter1.1",
start_stop: Some((0, 4)),
tags: Some((
scope: Stream,
tags: [
("title", ["chapter 1.1"]),
],
)),
loop: Some((None, 0)),
sub_entries: [
],
),
(
entry_type: Chapter,
uid: "chapter1.2",
start_stop: Some((4, 10)),
tags: Some((
scope: Stream,
tags: [
("title", ["chapter 1.2"]),
],
)),
loop: Some((None, 0)),
sub_entries: [
],
),
],
),
(
entry_type: Chapter,
uid: "chapter2",
start_stop: Some((10, 15)),
tags: Some((
scope: Stream,
tags: [
("title", ["chapter 2"]),
],
)),
loop: Some((None, 0)),
sub_entries: [
],
),
],
),
],
)
"#;
let toc: Toc = ron::de::from_str(toc_ron).unwrap();
assert_eq!(toc.scope(), TocScope::Global);
let entries = toc.entries();
assert_eq!(1, entries.len());
let edition = &entries[0];
assert_eq!(TocEntryType::Edition, edition.entry_type());
assert_eq!("edition", edition.uid());
assert!(edition.tags().is_none());
assert_eq!(Some((0, 15)), edition.start_stop_times());
let sub_entries = edition.sub_entries();
assert_eq!(2, sub_entries.len());
let chapter1 = &sub_entries[0];
assert_eq!(TocEntryType::Chapter, chapter1.entry_type());
assert_eq!("chapter1", chapter1.uid());
assert!(chapter1.tags().is_none());
assert_eq!(Some((0, 10)), chapter1.start_stop_times());
let chap1_sub_entries = chapter1.sub_entries();
assert_eq!(2, sub_entries.len());
let chapter1_1 = &chap1_sub_entries[0];
assert_eq!(TocEntryType::Chapter, chapter1_1.entry_type());
assert_eq!("chapter1.1", chapter1_1.uid());
assert_eq!(Some((0, 4)), chapter1_1.start_stop_times());
let tags = chapter1_1.tags().unwrap();
assert_eq!("chapter 1.1", tags.index::<Title>(0).unwrap().get());
assert_eq!(0, chapter1_1.sub_entries().len());
let chapter1_2 = &chap1_sub_entries[1];
assert_eq!(TocEntryType::Chapter, chapter1_2.entry_type());
assert_eq!("chapter1.2", chapter1_2.uid());
assert_eq!(Some((4, 10)), chapter1_2.start_stop_times());
let tags = chapter1_2.tags().unwrap();
assert_eq!("chapter 1.2", tags.index::<Title>(0).unwrap().get());
assert_eq!(0, chapter1_2.sub_entries().len());
let chapter2 = &sub_entries[1];
assert_eq!(TocEntryType::Chapter, chapter2.entry_type());
assert_eq!("chapter2", chapter2.uid());
let tags = chapter2.tags().unwrap();
assert_eq!("chapter 2", tags.index::<Title>(0).unwrap().get());
assert_eq!(Some((10, 15)), chapter2.start_stop_times());
assert_eq!(0, chapter2.sub_entries().len());
}
#[allow(clippy::cognitive_complexity)]
#[test]
fn test_serde_roundtrip() {
crate::init().unwrap();
let mut toc = Toc::new(TocScope::Global);
{
let toc = toc.get_mut().unwrap();
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"toc", TagMergeMode::Append);
toc.set_tags(tags);
let mut toc_edition = TocEntry::new(TocEntryType::Edition, "edition");
{
let toc_edition = toc_edition.get_mut().unwrap();
toc_edition.set_start_stop_times(0, 15);
let mut toc_chap_1 = TocEntry::new(TocEntryType::Chapter, "chapter1");
{
let toc_chap_1 = toc_chap_1.get_mut().unwrap();
toc_chap_1.set_start_stop_times(0, 10);
let mut toc_chap_1_1 = TocEntry::new(TocEntryType::Chapter, "chapter1.1");
{
let toc_chap_1_1 = toc_chap_1_1.get_mut().unwrap();
toc_chap_1_1.set_start_stop_times(0, 4);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 1.1", TagMergeMode::Append);
toc_chap_1_1.set_tags(tags);
}
toc_chap_1.append_sub_entry(toc_chap_1_1);
let mut toc_chap_1_2 = TocEntry::new(TocEntryType::Chapter, "chapter1.2");
{
let toc_chap_1_2 = toc_chap_1_2.get_mut().unwrap();
toc_chap_1_2.set_start_stop_times(4, 10);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 1.2", TagMergeMode::Append);
toc_chap_1_2.set_tags(tags);
}
toc_chap_1.append_sub_entry(toc_chap_1_2);
}
toc_edition.append_sub_entry(toc_chap_1);
let mut toc_chap_2 = TocEntry::new(TocEntryType::Chapter, "chapter2");
{
let toc_chap_2 = toc_chap_2.get_mut().unwrap();
toc_chap_2.set_start_stop_times(10, 15);
let mut tags = TagList::new();
tags.get_mut()
.unwrap()
.add::<Title>(&"chapter 2", TagMergeMode::Append);
toc_chap_2.set_tags(tags);
}
toc_edition.append_sub_entry(toc_chap_2);
}
toc.append_entry(toc_edition);
}
let toc_ser = ron::ser::to_string(&toc).unwrap();
let toc_de: Toc = ron::de::from_str(toc_ser.as_str()).unwrap();
assert_eq!(toc_de.scope(), toc.scope());
let entries_de = toc_de.entries();
let entries = toc.entries();
assert_eq!(entries_de.len(), entries.len());
let edition_de = &entries_de[0];
let edition = &entries[0];
assert_eq!(edition_de.entry_type(), edition.entry_type());
assert_eq!(edition_de.uid(), edition.uid());
assert_eq!(edition_de.tags(), edition.tags());
assert_eq!(edition_de.start_stop_times(), edition.start_stop_times());
let sub_entries_de = edition_de.sub_entries();
let sub_entries = edition.sub_entries();
assert_eq!(sub_entries_de.len(), sub_entries.len());
let chapter1_de = &sub_entries_de[0];
let chapter1 = &sub_entries[0];
assert_eq!(chapter1_de.entry_type(), chapter1.entry_type());
assert_eq!(chapter1_de.uid(), chapter1.uid());
assert_eq!(chapter1_de.tags().is_none(), chapter1.tags().is_none());
assert_eq!(chapter1_de.start_stop_times(), chapter1.start_stop_times());
let chap1_sub_entries_de = chapter1_de.sub_entries();
let chap1_sub_entries = chapter1.sub_entries();
assert_eq!(sub_entries_de.len(), sub_entries.len());
let chapter1_1_de = &chap1_sub_entries_de[0];
let chapter1_1 = &chap1_sub_entries[0];
assert_eq!(chapter1_1_de.entry_type(), chapter1_1.entry_type());
assert_eq!(chapter1_1_de.uid(), chapter1_1.uid());
assert_eq!(
chapter1_1_de.start_stop_times(),
chapter1_1.start_stop_times()
);
let tags_de = chapter1_1_de.tags().unwrap();
let tags = chapter1_1.tags().unwrap();
assert_eq!(
tags_de.index::<Title>(0).unwrap().get(),
tags.index::<Title>(0).unwrap().get()
);
assert_eq!(
chapter1_1_de.sub_entries().len(),
chapter1_1.sub_entries().len()
);
let chapter1_2_de = &chap1_sub_entries_de[1];
let chapter1_2 = &chap1_sub_entries[1];
assert_eq!(chapter1_2_de.entry_type(), chapter1_2.entry_type());
assert_eq!(chapter1_2_de.uid(), chapter1_2.uid());
assert_eq!(
chapter1_2_de.start_stop_times(),
chapter1_2.start_stop_times()
);
let tags_de = chapter1_2_de.tags().unwrap();
let tags = chapter1_2.tags().unwrap();
assert_eq!(
tags_de.index::<Title>(0).unwrap().get(),
tags.index::<Title>(0).unwrap().get()
);
assert_eq!(
chapter1_2_de.sub_entries().len(),
chapter1_2.sub_entries().len()
);
let chapter2_de = &sub_entries_de[1];
let chapter2 = &sub_entries[1];
assert_eq!(chapter2_de.entry_type(), chapter2.entry_type());
assert_eq!(chapter2_de.uid(), chapter2.uid());
let tags_de = chapter2_de.tags().unwrap();
let tags = chapter2.tags().unwrap();
assert_eq!(
tags_de.index::<Title>(0).unwrap().get(),
tags.index::<Title>(0).unwrap().get()
);
assert_eq!(chapter2_de.start_stop_times(), chapter2.start_stop_times());
assert_eq!(
chapter2_de.sub_entries().len(),
chapter2.sub_entries().len()
);
}
}