Video Samples for "D2MNet for Music Generation Joint Driven by Facial Expressions and Dance Movements"

It is worth noting that synthesizing high quality audio itself remains a challenging and computational demanding research problem. At present, shorter music samples are used for training and standard testing in the main experiments. As the length of the generated sample increases, the performance of the model will be affected. Specifically, evaluation metrics of beat correspondence and style-consistency will achieve lower scores. However, our model can also be effectively trained and tested with a longer music sequence length via a relatively larger network with more parameters. And this is also the direction of our future work.

THIS DATASET IS ARCHIVED AT DANS/EASY, BUT NOT ACCESSIBLE HERE. TO VIEW A LIST OF FILES AND ACCESS THE FILES IN THIS DATASET CLICK ON THE DOI-LINK ABOVE

Identifier
DOI https://doi.org/10.17632/cdjfthjmm6.1
PID https://nbn-resolving.org/urn:nbn:nl:ui:13-kn-uqho
Metadata Access https://easy.dans.knaw.nl/oai?verb=GetRecord&metadataPrefix=oai_datacite&identifier=oai:easy.dans.knaw.nl:easy-dataset:339676
Provenance
Creator Huang, J
Publisher Data Archiving and Networked Services (DANS)
Contributor Jiang Huang
Publication Year 2024
Rights info:eu-repo/semantics/openAccess; License: http://creativecommons.org/licenses/by/4.0; http://creativecommons.org/licenses/by/4.0
OpenAccess true
Representation
Resource Type Dataset
Discipline Other