Controlled Face Generation System using StyleGAN2 Neural Network
Main Article Content
Abstract
A novel approach to supervised face generation using open-source generative models including StyleGAN2 and Ridge Regression is presented. A methodology that extends StyleGAN2 to control facial characteristics such as age, race, gender, facial expression, and hair attributes is developed, and an extensive dataset of human faces with attribute annotations is utilized. The faces were encoded in 256-dimensional latent space using the StyleGAN2 encoder, resulting in a set of characteristic latent codes. We applied the t-SNE algorithm to cluster these feature-based codes, demonstrated the ability to control face generation, and subsequently trained Ridge regression models for each dimension of the latent codes using the labeled features. When decoded using StyleGAN2, the resulting codes successfully reconstructed face images while maintaining the association with the input features. The developed approach provides an easy and efficient way to supervised face generation using existing generative models such as StyleGAN2, and opens up new possibilities for different application areas.
Keywords:
Article Details
References
2. Bishop C. Pattern Recognition and Machine Learning. Information Science and Statistics. 2006. URL: https://link.springer.com/book/9780387310732
3. Karras T., Laine S., Aila T. A Style-Based Generator Architecture for Generative Adversarial Networks. ArXiv181204948 Cs Stat. 2019. URL: http://arxiv.org/abs/1812.04948
4. Karras T., Hellsten J et al. Analyzing and Improving the Image Quality of StyleGAN. arXiv:1912.04958 Cs. 2019. URL: https://arxiv.org/pdf/1912.04958.pdf
5. Kryuchkov M., Khanzhina N., Osmakov I., Ulyanov P. CT images GAN-based augmentation with AdaIN for lung nodules detection // Proceedings of SPIE – The International Society for Optical Engineering: 13, Rome, 02–06 November 2020. Rome, 2020. P. 1160526. https://doi.org/10.1117/12.2587940–EDN JYZOEO.
6. Huang G., Ramesh M., Berg T., Learned-Miller E. Labeled Faces in the Wild: A Database for Studying Face Recognition in Unconstrained Environments. University of Massachusetts, Amherst, Technical Report 07-49. 2018. URL: http://vis-www.cs.umass.edu/lfw/
7. Viola P., Jones M. Robust Real-time Object Detection. Second international workshop on statistical and computational theories of vision. 2001. URL: https://www.cs.cmu.edu/~efros/courses/LBMV07/Papers
8. Ledig C., Theis L. et al. Photo-realistic single image super-resolution using a generative adversarial network. ArXiv:1609.04802v5 Cs. 2016. URL: https://arxiv.org/pdf/1609.04802v5.pdf
9. Goar V., Kuri M., Kumar R., Senjyu T. Data Compression and Visualization Using PCA and T-SNE. Advances in Information Communication Technology and Computing. 2019. URL: https://www.researchgate.net/publication/344000619_Data_Compression_ and_Visualization_Using_PCA_and_T-SNE
10. Kolo B. Binary and Multiclass Classification. Weatherford Press. 2010. URL: https://www.amazon.com/Binary-Multiclass-Classification-Brian-Kolo/dp/1615800131
11. Rawlings J., Pantula S., Dickey D. Polynomial Regression. Applied Regression Analysis. 1998. URL: https://link.springer.com/chapter/10.1007/0-387-22753-9_8
12. Козина Н. И., Шиян Н. В., Чалченко М. Р. Современные достижения в области генерации изображений на примере нейронной сети MIDJOURNEY // Сборник материалов XVI-ой международной очно-заочной научно-практической конференции. М.: Научно-издательский центр «Империя», 2023. С. 121–125.
This work is licensed under a Creative Commons Attribution 4.0 International License.
Presenting an article for publication in the Russian Digital Libraries Journal (RDLJ), the authors automatically give consent to grant a limited license to use the materials of the Kazan (Volga) Federal University (KFU) (of course, only if the article is accepted for publication). This means that KFU has the right to publish an article in the next issue of the journal (on the website or in printed form), as well as to reprint this article in the archives of RDLJ CDs or to include in a particular information system or database, produced by KFU.
All copyrighted materials are placed in RDLJ with the consent of the authors. In the event that any of the authors have objected to its publication of materials on this site, the material can be removed, subject to notification to the Editor in writing.
Documents published in RDLJ are protected by copyright and all rights are reserved by the authors. Authors independently monitor compliance with their rights to reproduce or translate their papers published in the journal. If the material is published in RDLJ, reprinted with permission by another publisher or translated into another language, a reference to the original publication.
By submitting an article for publication in RDLJ, authors should take into account that the publication on the Internet, on the one hand, provide unique opportunities for access to their content, but on the other hand, are a new form of information exchange in the global information society where authors and publishers is not always provided with protection against unauthorized copying or other use of materials protected by copyright.
RDLJ is copyrighted. When using materials from the log must indicate the URL: index.phtml page = elbib / rus / journal?. Any change, addition or editing of the author's text are not allowed. Copying individual fragments of articles from the journal is allowed for distribute, remix, adapt, and build upon article, even commercially, as long as they credit that article for the original creation.
Request for the right to reproduce or use any of the materials published in RDLJ should be addressed to the Editor-in-Chief A.M. Elizarov at the following address: amelizarov@gmail.com.
The publishers of RDLJ is not responsible for the view, set out in the published opinion articles.
We suggest the authors of articles downloaded from this page, sign it and send it to the journal publisher's address by e-mail scan copyright agreements on the transfer of non-exclusive rights to use the work.