Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/107849
Full metadata record
DC FieldValueLanguage
dc.date.accessioned2023-03-28T12:05:44Z-
dc.date.available2023-03-28T12:05:44Z-
dc.date.issued2022-
dc.identifier.citationRiahi, Y. (2022). Art generation through sound (Bachelor's dissertation).en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar/handle/123456789/107849-
dc.descriptionB.Sc. IT (Hons)(Melit.)en_GB
dc.description.abstractThis work pursuits the prospect of cross-modal generation in computational creativity. The method employed was that of developing a specific type of AI model that is capable of generating many illustrations in the form of a moving visual based on the music provided. This allows for the creation of unique and singular pieces of the audio-visual artform. The AI model developed is based on a Generative Adversarial Network (GAN). This model pits two networks against each other with one network trying to generate data that can pass off as real, while the other network tries to discriminate whether that data is real or not. This is integral for the reason that in this way the AI model is able to generate realistic content similar to what a real person can produce or possibly better. Furthermore, this model strays from conventional GAN models by having the ability to take sound as an input and then generate a moving art piece based on the elements of the sound provided. The model primarily syncs pitch, volume and tempo with the image such that these features control all of the textures, shapes and objects and also control the movement between the frames. The network mainly competes till convergence is reached, this meaning that at a point the generative network improves so much that the discriminative network is not able to tell the difference between real and fake. Hence if the model continues to compete past this point the quality of the content may drop. The system developed was tested for its ability, veracity and ingenuity with satisfactory results. The main issue is the time it takes to generate, this can be attributed to computational power and the size of the training data.en_GB
dc.language.isoenen_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectArtificial intelligenceen_GB
dc.subjectNeural networks (Computer science)en_GB
dc.subjectCreative abilityen_GB
dc.titleArt generation through sounden_GB
dc.typebachelorThesisen_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.en_GB
dc.publisher.institutionUniversity of Maltaen_GB
dc.publisher.departmentFaculty of Information and Communication Technology. Department of Artificial Intelligenceen_GB
dc.description.reviewedN/Aen_GB
dc.contributor.creatorRiahi, Yran (2022)-
Appears in Collections:Dissertations - FacICT - 2022
Dissertations - FacICTAI - 2022

Files in This Item:
File Description SizeFormat 
2208ICTICT390900004128_2.PDF
  Restricted Access
2.09 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.