USING NOISES FOR STABLE IMAGE GENERATION IN DIFFUSION MODELS

2025;
: 113-123
Received: February 18, 2025
Revised: February 26, 2025
Accepted: March 14, 2025
1
Lviv Polytechnic National University
2
Lviv Polytechnic National University
3
Lviv Polytechnic National University

The foundation of this work lies in the study of the image generation process using diffusion models. The potential for increasing generation stability is demonstrated through an approach that involves using constant noise as a regularizer, with the addition of a mask of random noise constructed around each pixel of the input image within a random radius. The models were implemented using Python and TensorFlow, NumPy, and Keras libraries. The final results are presented, showing that stability was achieved while maintaining a certain level of variability in the generated output.

[1] Zhang, M., & Li, J. (2021). A commentary of GPT-3 in MIT Technology Review 2021. Fundamental Research, 1(6), 831-833. https://doi.org/10.1016/j.fmre.2021.11.011

[2] Puteikis, K., & Mameniškienė, R. (2024). Artificial intelligence: Can it help us better grasp the idea of epilepsy? An exploratory dialogue with ChatGPT and DALL· E 2. Epilepsy & Behavior, 156, 109822. https://doi.org/10.1016/j.yebeh.2024.109822

[3] Borji, A. (2022). Generated faces in the wild: Quantitative comparison of stable diffusion, midjourney and dall-e 2. arXiv preprint arXiv:2210.00586.

[4] Sun, Z., Fang, H., Cao, J., Zhao, X., & Wang, D. (2024, October). Rethinking Image Editing Detection in the Era of Generative AI Revolution. In Proceedings of the 32nd ACM International Conference on Multimedia (pp. 3538-3547). https://doi.org/10.1145/3664647.3681445

[5] Archana Balkrishna, Y. (2024). An analysis on the use of image design with generative AI technologies. International Journal of Trend in Scientific Research and Development, 8(1), 596-599.

[6] Van Daele, D., Decleyre, N., Dubois, H., & Meert, W. (2021, May). An automated engineering assistant: Learning parsers for technical drawings. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 35, No. 17, pp. 15195-15203). https://doi.org/10.1609/aaai.v35i17.17783

[7] Banh, L., & Strobel, G. (2023). Generative artificial intelligence. Electronic Markets, 33(1), 63. https://doi.org/10.1007/s12525-023-00680-1

[8] Creswell, A., White, T., Dumoulin, V., Arulkumaran, K., Sengupta, B., & Bharath, A. A. (2018). Generative adversarial networks: An overview. IEEE signal processing magazine, 35(1), 53-65. https://doi.org/10.1109/MSP.2017.2765202

[9] Yang, L., Zhang, Z., Song, Y., Hong, S., Xu, R., Zhao, Y., ... & Yang, M. H. (2023). Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 56(4), 1-39.

[10] Batzolis, G., Stanczuk, J., Schönlieb, C. B., & Etmann, C. (2021). Conditional image generation with score-based diffusion models. arXiv preprint arXiv:2111.13606.

[11] Wang, Z., Zheng, H., He, P., Chen, W., & Zhou, M. (2022). Diffusion-gan: Training gans with diffusion. arXiv preprint arXiv:2206.02262.

[12] Dhariwal, P., & Nichol, A. (2021). Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34, 8780-8794.

[13] Géron, A. (2022). Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow. " O'Reilly Media, Inc.".