Professional Documents
Culture Documents
Achyut Raghavan
Akhilesh Harkude Udit Brahmadevara
Computer Science Engineering
Computer Science Engineering Computer Science Engineering
PES University
PES University PES University
Bangalore, India
Bangalore, India Bangalore, India
raghavanachyut@gmail.com
akhileshharkude1@gmail.com udit.brahmadevara@gmail.com
Ashrita D
Computer Science Engineering
PES University
Bangalore, India
daraashritha0703@gmail.com
H. Model Evaluation:
Quantitative evaluation metrics are employed to assess
the performance of DF-GAN and SA-GAN in text-to-face
generation. These metrics provide insights into the
effectiveness of each model in capturing contextual and
structural nuances. Fig. 1. Cultural Infusion into CelebAHQ
I. Integration of Stages:
The final generation process integrates both stages, with
Stage-I trained on 64 × 64 faces. The output is then passed to
Stage-II for the production of 128 × 128 faces. This
sequential process aims to refine and enhance the generated
facial images.
J. Comparative Analysis:
Generated facial images from DF-GAN and SA-GAN are
compared, highlighting distinctions in their outputs. The
inclusion of AttGAN variations and head rotations further
enriches the comparative analysis.
Training commenced with Stage-I responsible for [1] Tao Xu, Pengchuan Zhang, Qiuyuan Huang, Han Zhang, Zhe Gan,
Xiaolei Huang, Xiaodong He “AttnGAN: Fine-Grained Text to
generating low-resolution outputs. Stage-II was Image Generation with Attentional Generative Adversarial
subsequently trained using the outputs from Stage-I to Networks”, 2017
generate high-resolution facial images. During the training
process, efforts were made to maintain a uniform [2] Ming Tao, Hao Tang, Fei Wu, Xiaoyuan Jing, Bing-Kun Bao,
distribution of captions from each class to prevent biased Changsheng Xu “DF-GAN: A Simple and Effective Baseline for
image generation towards specific classes. Text-to-Image Synthesis”, 2022
[3] Minfeng Zhu, Pingbo Pan, Wei Chen, Yi Yang “DM-GAN: Dynamic
VI. CONCLUSION AND FUTURE WORK Memory Generative Adversarial Networks for Text-to-Image
Synthesis”, 2020
A. Findings and Conclusion:
[4] Jianxin Sun1,2*, Qiyao Deng1,2*, Qi Li1,2 †, Muyi Sun1, Min
Ren1,2, Zhenan Sun1,2 “AnyFace : Free-style Text-to-Face Synthesis
The experimental findings demonstrated the model's and Manipulation”, 2020
potential in generating facial images from textual
descriptions. The dataset fusion of CelebHQ and Indian [5] Harsh Jaykumar Jalan, Gautam Maurya, Canute Corda, Suspect Face
faces enabled diverse and realistic image synthesis based on Generation, 2020
provided textual cues.
[6] Osaid Rehman Nasir+∗, Shailesh Kumar Jha+∗, Manraj Singh
B. Implications and Future Work: Grover∗, Yi Yu†, Ajit Kumar‡ and Rajiv Ratn Shah
“Text2FaceGAN: Face Generation from Fine Grained Textual
Descriptions”, 2019
The successful generation of facial images from textual
descriptions holds promise in various applications, including [7] Tingting Qiao, Jing Zhang*, Duanqing Xu*, Dacheng Tao, College of
creative content generation and facial recognition. Future Computer Science and Technology, Zhejiang University, China,
research directions may include further exploration of School of Automation, Hangzhou Dianzi University, China,
augmentation techniques, architecture refinements, and UBTECH Sydney AI Centre, School of Computer Science, FEIT, The
University of Sydney, Australia: “MirrorGAN: Learning Text-to-
dataset expansion for improved diversity and model image Generation by Redescription”, 2021
performance.
[8] Bowen Li, Xiaojuan Qi, Thomas Lukasiewicz, Philip H. S. Torr,
ACKNOWLEDGMENT University of Oxford: “Controllable Text-to-Image Generation”, 2019
We would like to express our gratitude to Dr. Vinodha K,
Department of Computer Science and Engineering, PES [9] MUHAMMAD ZEESHAN KHAN SAIRA JABEENMUHAMMAD
University, for her continuous guidance, assistance, and USMAN GHANI KHANTANZILA SABA (Senior Member, IEEE),
ASIM REHMATAMJAD REHMAN , (Senior Member, IEEE),AND
encouragement throughout the development of this USMAN TARIQ “A Realistic Image Generation of Face From Text
UE20CS461A - Capstone Project Phase – 2. Description Using the Fully Trained Generative Adversarial
Network”, 2020
We are very grateful to the Capstone Project Coordinator,
Dr. Sarasvathi V, Professor and Dr. Sudeepa Roy Dey, [10] Jupiter Tamrakar , Bal Krishna Nyaupane “Synthesizing Human Face
Image from Textual Description of Facial Attributes Using
Associate Professor, for organizing, managing, and helping Attentional Generative Adversarial Network”, 2021
with the entire process.
[11] YUMING JIANG, S-Lab, Nanyang Technological University,
We take this opportunity to thank Dr. Sandesh B J, SingaporeSHUAI YANG, S-Lab, Nanyang Technological University,
Chairperson, Department of Computer Science and SingaporeHAONAN QIU, S-Lab, Nanyang Technological University,
Engineering, PES University, for all the knowledge and Singapore: “Text2Human: Text-Driven Controllable Human Image
Generation”, 2022
support we have received from the department. We
would like to thank Dr. B.K. Keshavan, Dean of
[12] Zhenliang He, Wangmeng Zuo, Senior Member, IEEE, Meina Kan,
Faculty, PES University for his help. Member, IEEE, Shiguang Shan, Senior Member, IEEE, and Xilin
Chen, Fellow, IEEE “AttGAN: Facial Attribute Editing by Only
We are deeply grateful to Dr.M. R.Doreswamy, Chancellor, Changing What You Want”, 2019
PES University, Prof. Jawahar Doreswamy, Pro Chancellor
– PES University, Dr. Suryaprasad J, Vice- Chancellor, PES [13] Andrea Asperti, Gabriele Colasuonno∗ , and Antonio Guerra∗
∗Department of Informatics, University of Bologna, Italy “Head
University and Prof. Nagarjuna Sadineni, Pro-Vice Rotation in Denoising Diffusion Models”, 2023
Chancellor - PES University, for providing to us various
opportunities and enlightenment every step of the way. [14] Han Zhang, Ian Goodfellow, Dimitris Metaxas, Augustus Odena,
Finally, this project could not have been completed without “Self-Attention Generative Adversarial Networks”, 2018
[15] Amit Kushwaha; Chanakya P; Krishna Pratap Singh, “Text to Face Gustav Reichert, and Helge Ritter, “Face Generation and Editing with
generation using Wasserstein stackGAN”, 2022 StyleGAN: A Survey”, 2022
[16] Hang Zhou,1 Jihao Liu, Ziwei Liu, Yu Liu, Xiaogang Wang, “Rotate- [20] SHU-YU CHEN, FENG-LIN LIU, YU-KUN LAI, PAUL L. ROSIN,
and-Render: Unsupervised Photorealistic Face Rotation from Single- CHUNPENG LI, HONGBO FU, LIN GAO, “DeepFaceEditing: Deep
View Images”, 2020 Face Generation and Editing with Disentangled Geometry and
Appearance Control”, 2021\
[17] Rui Huang, Shu Zhang, Tianyu Li1, Ran He, “Beyond Face Rotation:
Global and Local Perception GAN for Photorealistic and Identity [21] Xianxu Hou, Linlin Shen, Or Patashnik, Daniel Cohen-Or, Hui
Preserving Frontal View Synthesis”, 2017 Huang, “FEAT: Face Editing with Attention”, 2022