Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Generative Artificial Intelligence: A Systematic Review and Applications (2405.11029v1)

Published 17 May 2024 in cs.LG, cs.AI, cs.CL, and cs.CV
Generative Artificial Intelligence: A Systematic Review and Applications

Abstract: In recent years, the study of AI has undergone a paradigm shift. This has been propelled by the groundbreaking capabilities of generative models both in supervised and unsupervised learning scenarios. Generative AI has shown state-of-the-art performance in solving perplexing real-world conundrums in fields such as image translation, medical diagnostics, textual imagery fusion, natural language processing, and beyond. This paper documents the systematic review and analysis of recent advancements and techniques in Generative AI with a detailed discussion of their applications including application-specific models. Indeed, the major impact that generative AI has made to date, has been in language generation with the development of LLMs, in the field of image translation and several other interdisciplinary applications of generative AI. Moreover, the primary contribution of this paper lies in its coherent synthesis of the latest advancements in these areas, seamlessly weaving together contemporary breakthroughs in the field. Particularly, how it shares an exploration of the future trajectory for generative AI. In conclusion, the paper ends with a discussion of Responsible AI principles, and the necessary ethical considerations for the sustainability and growth of these generative models.

Generative Artificial Intelligence: A Systematic Review and Applications

Overview

The paper "Generative Artificial Intelligence: A Systematic Review and Applications" co-authored by Sandeep Singh Sengar, Affan Bin Hasan, Sanjay Kumar, and Fiona Carroll presents a meticulous exploration of the advancements and application-specific models within the domain of Generative AI (GenAI). This comprehensive review is scoped between the years 2018 and 2023, offering a detailed historical context that traces the evolution from foundational models developed between 2012 and 2018. The review is systematically conducted and meticulously documents the impacts, challenges, opportunities, and ethical considerations surrounding Generative AI.

Key Contributions

1. Paradigm Shift in AI and Historical Context:

The paper discusses the shift in AI propelled by generative models capable of unsupervised and supervised learning. A historical progression is provided, elucidating the development of key AI models from 2012 to 2018, which laid the foundational groundwork for contemporary GenAI techniques.

2. Applications in Image Translation:

Among the pivotal applications of Generative AI, image translation stands out prominently. For example, the usage of GAN-based models has revolutionized medical diagnostics and image synthesis. Methods like SwinTransformer, CycleGAN, and Pix2Pix GAN demonstrate substantial advancements in translating medical imaging modalities, improving diagnostic precision through high-quality image synthesis. Pairing traditional GANs with novel architectures like Swin Transformers has yielded state-of-the-art performance in translating MRI images from T1 to T2 modes, as exemplified by results on the BraTs2018 dataset.

3. Video Synthesis and Generation:

Generative AI also significantly impacts video synthesis and generation techniques, including the creation of realistic talking head videos using Depth-Aware GANs and DaGAN++. Innovations such as StyleTalker facilitate highly accurate lip synchronization and realistic head poses in video generation from audio inputs, underpinning advancements in video conferencing, virtual reality, and entertainment industries.

4. Natural Language Processing:

The paper provides a detailed review of transformative progress in NLP driven by models such as BERT, T5, and StyleTalker. Particularly, BERT's performance in tasks like Named Entity Recognition (NER) surpasses previous methodologies. Multilingual model evaluations underscore BERT's robustness across languages, showcasing significant advancements in machine translation, language generation, and educational applications, where models like ChatGPT exhibit impressive question-answering capabilities.

5. Knowledge Graph Generation:

The synthesis and enhancement of knowledge graphs are significantly boosted by Generative AI. Techniques like KBGAN and K-BERT integrate commonsense and domain-specific knowledge to improve the quality and depth of knowledge graphs. Models like TuckER and ComplexGCN further advance link prediction tasks, demonstrating superior performance in structuring and querying rich datasets.

6. Interdisciplinary Applications and Ethical Considerations:

The paper explores the interdisciplinary applications of GenAI in fields such as mechanical fault detection and traffic scenario generation, emphasizing its potential for substantial practical impact. It highlights the ethical implications and need for responsible AI development, discussing frameworks for mitigating biases, ensuring data privacy, and implementing transparency in AI-generated content.

Implications of the Research

Practical Implications:

The advancements in Generative AI facilitate diverse practical applications across medical imaging, video generation, automated content creation, and personalized learning systems, among others. These innovations promise to enhance productivity, improve diagnostic accuracy, and provide more effective educational tools.

Theoretical Implications and Future Directions:

Theoretically, the development of robust and versatile generative models drives forward the understanding of both deep learning architectures and AI application frameworks. Future research directions should focus on further refining model architectures to address current limitations in interpretability, robustness against adversarial attacks, and ethical governance. Additionally, interdisciplinary collaborations and real-world implementations will be pivotal in uncovering new applications and improving existing ones.

Conclusion

The paper delivers a thorough and insightful review of Generative AI, showcasing the extensive impacts and diverse applications of generative models across multiple domains. By documenting the advancements, challenges, and opportunities, the paper paves the way for future research and development in this rapidly evolving field. Emphasizing the necessity for responsible AI principles, ethical considerations, and continuous innovation, the paper reflects a balanced approach towards harnessing the transformative potential of Generative AI.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (162)
  1. Brain tumor classification using a combination of variational autoencoders and generative adversarial networks. Biomedicines, 10(2):223.
  2. Mega: Multilingual evaluation of generative ai.
  3. Contextual string embeddings for sequence labeling. In Proceedings of the 27th international conference on computational linguistics, pages 1638–1649.
  4. A hierarchical structured self-attentive model for extractive document summarization (hssas). IEEE Access, 6:24205–24212.
  5. The role of generative adversarial networks in brain mri: a scoping review. Insights into imaging, 13(1):98.
  6. Wasserstein gan.
  7. Vivit: A video vision transformer. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6836–6846.
  8. Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. Information fusion, 58:82–115.
  9. Real-time monocular depth estimation using synthetic data with domain adaptation via image style transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  10. TuckER: Tensor factorization for knowledge graph completion. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics.
  11. Hp-gan: Probabilistic 3d human motion prediction via gan. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.
  12. Is space-time attention all you need for video understanding? In ICML, volume 2, page 4.
  13. Demystifying mmd gans. arXiv preprint arXiv:1801.01401.
  14. Bozkurt, A. (2023). Generative artificial intelligence (ai) powered conversational educational agents: The inevitable paradigm shift. Asian Journal of Distance Education, 18(1).
  15. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  16. The cora dataset: validation and diagnostics of ocean temperature and salinity in situ measurements. Ocean Science Discussions, 9(2):1273–1312.
  17. Anomaly detection for alzheimer’s disease in brain mris via unsupervised generative adversarial learning. In 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), pages 1–5.
  18. Kbgan: Adversarial learning for knowledge graph embeddings.
  19. A comprehensive survey of ai-generated content (aigc): A history of generative ai from gan to chatgpt.
  20. A comparison of word2vec, hmm2vec, and pca2vec for malware classification.
  21. Infogan: Interpretable representation learning by information maximizing generative adversarial nets.
  22. Upgpt: Universal diffusion model for person image generation, editing and pose transfer.
  23. Semi-supervised sequence modeling with cross-view training. arXiv preprint arXiv:1809.08370.
  24. Xnli: Evaluating cross-lingual sentence representations. arXiv preprint arXiv:1809.05053.
  25. Support-vector networks. Machine learning, 20:273–297.
  26. Courant, R. (1943). Variational methods for the solution of problems of equilibrium and vibrations.
  27. Generative adversarial networks: An overview. IEEE Signal Processing Magazine, 35(1):53–65.
  28. Image synthesis in multi-contrast mri with conditional generative adversarial networks.
  29. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  30. Enhancing cancer differentiation with synthetic mri examinations via generative models: a systematic review. Insights into Imaging, 13(1):188.
  31. Nice: Non-linear independent components estimation.
  32. Density estimation using real nvp.
  33. Adversarial feature learning. arXiv preprint arXiv:1605.09782.
  34. Opinion paper: “so what if chatgpt wrote it?” multidisciplinary perspectives on opportunities, challenges and implications of generative conversational ai for research, practice and policy. International Journal of Information Management, 71:102642.
  35. Christoph feichtenhofer. multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6824–6835.
  36. Feichtenhofer, C. (2020). X3d: Expanding architectures for efficient video recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 203–213.
  37. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6202–6211.
  38. Trafficgen: Learning to generate diverse and realistic traffic scenarios. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 3567–3575.
  39. Automatic generation of semantic parts for face image synthesis.
  40. Synthetic data augmentation using gan for improved liver lesion classification.
  41. Higan+: Handwriting imitation gan with disentangled representations. ACM Trans. Graph., 42(1).
  42. Fem simulation-based generative adversarial networks to detect bearing faults. IEEE Transactions on Industrial Informatics, 16(7):4961–4971.
  43. SimGANs: Simulator-based generative adversarial networks for ECG synthesis to improve deep ECG classification. In III, H. D. and Singh, A., editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 3597–3606. PMLR.
  44. Generative adversarial networks.
  45. Ffjord: Free-form continuous dynamics for scalable reversible generative models. arXiv preprint arXiv:1810.01367.
  46. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  47. Lstm can solve hard long time lag problems. Advances in neural information processing systems, 9.
  48. Dagan++: Depth-aware generative adversarial network for talking head video generation.
  49. Depth-aware generative adversarial network for talking head video generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 3397–3406.
  50. Unsupervised image-to-image translation: A review. Sensors, 22(21).
  51. Extreme learning machine: a new learning scheme of feedforward neural networks. In 2004 IEEE international joint conference on neural networks (IEEE Cat. No. 04CH37541), volume 2, pages 985–990. Ieee.
  52. Image-to-image translation with conditional adversarial networks.
  53. Age-specific diagnostic classification of asd using deep learning approaches. Studies in Health Technology and Informatics, 309:267–271.
  54. Mimic-iii, a freely accessible critical care database. Scientific data, 3(1):1–9.
  55. Extending a parser to distant domains using a few dozen partially annotated examples.
  56. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589.
  57. Malware classification with word2vec, hmm2vec, bert, and elmo. Journal of Computer Virology and Hacking Techniques, 19(1):1–16.
  58. Scaling up gans for text-to-image synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10124–10134.
  59. The kinetics human action video dataset.
  60. Attentional networks for music generation. Multimedia Tools and Applications, 81(4):5179–5189.
  61. Ctrl: A conditional transformer language model for controllable generation.
  62. Dcavn: Cervical cancer prediction and classification using deep convolutional and variational autoencoder network. Multimedia Tools and Applications, 80:30399–30415.
  63. Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems, 31.
  64. Improved variational inference with inverse autoregressive flow. In Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc.
  65. Auto-encoding variational bayes.
  66. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114.
  67. Semi-supervised classification with graph convolutional networks.
  68. A novel diffusivity function-based image denoising for mri medical images. Multimedia Tools and Applications, 82(21):32057–32089.
  69. Movinets: Mobile video networks for efficient video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16020–16030.
  70. Textcontrolgan: Text-to-image synthesis with controllable generative adversarial networks. Applied Sciences, 13(8):5098.
  71. A comprehensive survey on generative adversarial networks used for synthesizing multimedia content. Multimedia Tools and Applications, 82(26):40585–40624.
  72. Community detection in complex networks using stacked autoencoders and crow search algorithm. The Journal of Supercomputing, 79(3):3329–3356.
  73. Classification of autism spectrum disorder based on brain image data using deep neural networks. In International Conference on Frontiers of Intelligent Computing: Theory and Applications, pages 209–218. Springer.
  74. Dbpedia–a large-scale, multilingual knowledge base extracted from wikipedia. Semantic web, 6(2):167–195.
  75. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension.
  76. Mvitv2: Improved multiscale vision transformers for classification and detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4804–4814.
  77. Let’s verify step by step.
  78. Lin, C.-Y. (2004). ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
  79. Attention-based spatial guidance for image-to-image translation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 816–825.
  80. Psgan: A generative adversarial network for remote sensing image pan-sharpening. IEEE Transactions on Geoscience and Remote Sensing, 59(12):10227–10242.
  81. K-bert: Enabling language representation with knowledge graph.
  82. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV).
  83. Video swin transformer. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3202–3211.
  84. Stacked convolutional auto-encoders for hierarchical feature extraction. In Artificial Neural Networks and Machine Learning–ICANN 2011: 21st International Conference on Artificial Neural Networks, Espoo, Finland, June 14-17, 2011, Proceedings, Part I 21, pages 52–59. Springer.
  85. The columbia multi-document summarizer for duc 2002. In Workshop on Automatic Summarization, pages 1–8.
  86. The multimodal brain tumor image segmentation benchmark (brats). IEEE transactions on medical imaging, 34(10):1993–2024.
  87. Which training methods for gans do actually converge?
  88. The numerics of gans.
  89. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.
  90. Styletalker: One-shot style-based audio-driven talking head video generation.
  91. Conditional generative adversarial nets.
  92. Github copilot ai pair programmer: Asset or liability? Journal of Systems and Software, 203:111734.
  93. Gradient descent gan optimization is locally stable. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  94. VoxCeleb: A large-scale speaker identification dataset. In Interspeech 2017. ISCA.
  95. Webgpt: Browser-assisted question-answering with human feedback.
  96. Video transformer network.
  97. Odena, A. (2016). Semi-supervised learning with generative adversarial networks.
  98. Conditional image synthesis with auxiliary classifier gans.
  99. OpenAI (2023). Gpt-4 technical report.
  100. Librispeech: An asr corpus based on public domain audio books. In 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5206–5210.
  101. Correction of banding errors in satellite images with generative adversarial networks (gan). IEEE Access.
  102. Deep contextualized word representations.
  103. A lip sync expert is all you need for speech to lip generation in the wild. In Proceedings of the 28th ACM international conference on multimedia, pages 484–492.
  104. From copilot to pilot: Towards ai supported software development.
  105. Qi, G.-J. (2018). Loss-sensitive generative adversarial networks on lipschitz densities.
  106. Pre-trained models for natural language processing: A survey. Science China Technological Sciences, 63(10):1872–1897.
  107. Quinlan, J. R. (1986). Induction of decision trees. Machine learning, 1:81–106.
  108. Learning transferable visual models from natural language supervision.
  109. Improving language understanding by generative pre-training.
  110. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  111. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551.
  112. An extractive text summarization approach using tagged-lda based topic modeling. Multimedia tools and applications, 80:3275–3305.
  113. Dall-e: Creating images from text. UGC Care Group I Journal, 8(14):71–75.
  114. Reg-gan: Semi-supervised learning based on generative adversarial networks for regression. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 2806–2810. IEEE.
  115. Improved techniques for training gans. In Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc.
  116. Introduction to the conll-2003 shared task: Language-independent named entity recognition. arXiv preprint cs/0306050.
  117. Content-based secure image retrieval in an untrusted third-party environment. In International Conference on Frontiers of Intelligent Computing: Theory and Applications, pages 287–297. Springer.
  118. Multi-planar 3d knee mri segmentation via unet inspired architectures. International Journal of Imaging Systems and Technology, 33(3):985–998.
  119. Moving object tracking using laplacian-dct based perceptual hash. In 2016 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), pages 2345–2349. IEEE.
  120. Motion segmentation-based surveillance video compression using adaptive particle swarm optimization. Neural Computing and Applications, 32(15):11443–11457.
  121. Visualizing and understanding graph convolutional network. Multimedia Tools and Applications, 80:8355–8375.
  122. Singhal, A. (2012). Introducing the knowledge graph: Things, not strings,.
  123. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402.
  124. Adding realtime coverage to the google knowledge graph. In 11th International Semantic Web Conference (ISWC 2012), volume 914, pages 65–68. Citeseer.
  125. Sequence to sequence learning with neural networks. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc.
  126. A diverse domain generative adversarial network for style transfer on face photographs.
  127. Scenegen: Learning to generate realistic traffic scenes. In Proceedings - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 892–901. IEEE Computer Society. Funding Information: Work done at Uber ATG. Publisher Copyright: © 2021 IEEE; 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021 ; Conference date: 19-06-2021 Through 25-06-2021.
  128. Tanchenko, A. (2014). Visual-psnr measure of image quality. Journal of Visual Communication and Image Representation, 25(5):874–878.
  129. Fastmri prostate: A publicly available, biparametric mri dataset to advance machine learning for prostate cancer imaging.
  130. What if the devil is my guardian angel: Chatgpt as a case study of using chatbots in education. Smart Learning Environments, 10(1):15.
  131. Uvcgan: Unet vision transformer cycle-consistent gan for unpaired image-to-image translation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 702–712.
  132. Wavenet: A generative model for raw audio.
  133. Multi-head-self-attention based yolov5x-transformer for multi-scale object detection. Multimedia Tools and Applications, pages 1–27.
  134. Attention is all you need. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  135. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. Journal of machine learning research, 11(12).
  136. Pgnet: Real-time arbitrarily-shaped text spotting with point gathering network. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 2782–2790.
  137. Learning from synthetic data for crowd counting in the wild.
  138. Audio2head: Audio-driven one-shot talking-head generation with natural head motion. arXiv preprint arXiv:2107.09293.
  139. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2097–2106.
  140. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600–612.
  141. Cfrwd-gan for sar-to-optical image translation. Remote Sensing, 15(10):2547.
  142. Reenactgan: Learning to reenact faces via boundary transfer.
  143. Hmnet: Hybrid matching network for few-shot link prediction. In International Conference on Database Systems for Advanced Applications, pages 307–322. Springer.
  144. Generative adversarial networks can create high quality artificial prostate cancer magnetic resonance images. Journal of Personalized Medicine, 13(3):547.
  145. Layoutlm: Pre-training of text and layout for document image understanding. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1192–1200.
  146. Swin transformer-based gan for multi-modal medical image translation. Frontiers in Oncology, 12:942511.
  147. A study of face obfuscation in imagenet. In International Conference on Machine Learning (ICML).
  148. Exploring the limits of chatgpt for query or aspect-based text summarization.
  149. Semantic facial expression editing using autoencoded flow.
  150. Seqgan: Sequence generative adversarial nets with policy gradient.
  151. Deep generative molecular design reshapes drug discovery. Cell Reports Medicine.
  152. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 586–595.
  153. Ernie: Enhanced language representation with informative entities. arXiv preprint arXiv:1905.07129.
  154. Flow-guided one-shot talking face generation with a high-resolution audio-visual dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3661–3670.
  155. A comparative analysis of gan-based methods for sar-to-optical image translation. IEEE Geoscience and Remote Sensing Letters, 19:1–5.
  156. QMSum: A new benchmark for query-based multi-domain meeting summarization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5905–5921, Online. Association for Computational Linguistics.
  157. Makelttalk: speaker-aware talking-head animation. ACM Transactions On Graphics (TOG), 39(6):1–15.
  158. A hierarchical network for abstractive meeting summarization with cross-domain pretraining. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 194–203, Online. Association for Computational Linguistics.
  159. Unpaired image-to-image translation using cycle-consistent adversarial networks. In 2017 IEEE International Conference on Computer Vision (ICCV), pages 2242–2251.
  160. Toward multimodal image-to-image translation. Advances in neural information processing systems, 30.
  161. Toward multimodal image-to-image translation. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  162. Style fader generative adversarial networks for style degree controllable artistic style transfer. In Proc. Int. Joint Conf. on Artif. Intell.(IJCAI), pages 5002–5009.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Sandeep Singh Sengar (9 papers)
  2. Affan Bin Hasan (1 paper)
  3. Sanjay Kumar (82 papers)
  4. Fiona Carroll (1 paper)
Citations (15)
X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com