Consult an Expert
Trademark
Design Registration
Consult an Expert
Trademark
Copyright
Patent
Infringement
Design Registration
More
Consult an Expert
Consult an Expert
Trademark
Design Registration
Login
A DEEP LEARNING FRAMEWORK FOR DETECTING AI-GENERATED IMAGES
Extensive patent search conducted by a registered patent agent
Patent search done by experts in under 48hrs
₹999
₹399
Abstract
Information
Inventors
Applicants
Specification
Documents
ORDINARY APPLICATION
Published
Filed on 18 November 2024
Abstract
Titie··of Invention: .... A Deep Learning Framework for Detecting AI-Generated Image Field of Invention: Image processor, Artificial Intelligence, Embedded System 7. ABSTRACT The proliferation of advanced generative adversarial networks (GANs) has led to a surge in the creation of highly realistic synthetic images, posing significant challenges to the verification of image authenticity. This research investigates the efficacy of deep learning techniques in distinguishing between genuine and AI-generated images. In the initial phase, the study employed CIF AKE which is a comprehensive dataset of synthetic and real images to finetune various pre-trained convolutional neural networks (CNNs). Through rigorous experimentation, VGG 16 emerged as the superior architecture, demonstrating impressive accuracy in classifying images as. real or synthetic. Grad-CAM, an explainability technique, was employed to provide insights into the model's decision-making process, enhancing both the transparency and reliability of the results. This phase underscored the potential of deep learning in tackling the proliferation of highly realistic deepfakes and AI-generated content.
Patent Information
Application ID | 202441089016 |
Invention Field | COMPUTER SCIENCE |
Date of Application | 18/11/2024 |
Publication Number | 47/2024 |
Inventors
Name | Address | Country | Nationality |
---|---|---|---|
AROKIA NERLING RASONI G | Assistant Professor, Department of ECE, SAVEETHA ENGINEERING COLLEGE SAVEETHA NAGAR, THANDALAM, CHENNAI, TAMIL NADU, INDIA, PIN CODE-602105. | India | India |
FRANCIS J | Assistant Professor, Department of MECH, SAVEETHA ENGINEERING COLLEGE SAVEETHA NAGAR, THANDALAM, CHENNAI, TAMIL NADU, INDIA, PIN CODE-602105. | India | India |
Applicants
Name | Address | Country | Nationality |
---|---|---|---|
SAVEETHA ENGINEERING COLLEGE | SAVEETHA NAGAR, THANDALAM, CHENNAI, TAMIL NADU, INDIA, PIN CODE-602105. | India | India |
Specification
PREAMBLE TO THE DESCRIPTION
In today's digital landscape, the authenticity of visual content is
becoming increasingly difficult to verify, largely due to the rise of
advanced generative technologies like Generative Adversarial Networks
(GANs). These technologies can create synthetic images that are nearly
indistinguishable from real ones, raising significant concerns across
various fields, including journalism, law enforcement, and social media.
The emergence of these highly realistic deepfakes has underscored the
need for effective m ethods to discern genuine images from those
generated by AI, a challenge that this proje~t, "AI vs REAL: A DEEP
LEARNING APPROACH FOR IDENTIFYING AI-GENERATED
IMAGES," aims to address.
The goal of this project is to investigate how deep learning, namely the use
of pre-trained convolutional neural networks (CNNs), m:;ty be used to
differentiate between real and artificial images. We intend to improve a
number of pre-trained CNN models' capacity to correctly identify photos as
either real or artificial intelligence (AI)-generated by utilizing the CIFAKE
dataset.
The objective iUo identify the most effe~tive model for this task, ..
contributing to the development of reliable tools for maintaining the
integrity of digital media. Additionally, this project emphasizes the
importance of model interoperability by incorporating explainability
techniques like Grad-CAM. The significance of this project lies in its
potential to fill the current research gap by developing a more reliable and
interpretable deep fake detection system. By advancing the state of the art
in transfer learning and XAI for deepfake detection, this research will
contribute to safeguarding the integrity of digital media. The outcomes of
this project are timely and relevant, addressing the growing challenges
posed by deepfakes and providing a robust tool for media verification,
cybersecurity, and beyond.
4. DESCRIPTION
4.1 BACKGROUND OF INVENTION
The proliferation of deep~fake technology, driven by rapid
advancements in artificial intelligence, has introduced a significant
challenge in the realm of digital media authenticity. The ability to create
highly realistic synthetic images that can deceive even the most
discerning observers poses a serious threat to individuals, communities,
and society at large. Deep-fakes have the potential to spread
misinformation and disinforrnation, leading to social and economic harm,
psychological distress, and undermining trust ·in digital content. This
problem is exacerbated by the role of social media platforms in the rapid
dissemination of such content, making it increasingly difficult to contain
the damage caused by these fabricated images. The urgent need to
reliably distinguish between real and Al-generated images is at the core
2
~PAT~~1 QfflCE CHE~NAI "f""" . ... i ~ l ~ ~ t 1. t;l c ' 4 ~ 4 : .-} .. 3 +
....
-Q)
C)
Ill
D..
Q)
------------
of this project. Current solutions for detecting deep-fakes rely on various
tech.. n. .i ques, including man' .u. a_. l verification, watermarking, traditional . '• ,• ' ...
image processing techniques and basic neural network models. Manual
verification is time-consuming and prone to human error, while
watermarking is easily circumvented by sophisticated deep-fake
generation methods. Some methods, such as the basic CNN architectures,
have shown promise but still struggle with generalization across different
datasets and types of deep-fakes. Furthermore, the majority of these
models offer limited interchangeability, making it difficult to understand
their decision-making processes and assess their reliability in critical
applications. These limitations highlight the need for more robust,
accurate, and transparent solutions in the realm of deep-fake detection.
This project seeks to address these limitations by exploring
transfer learning techniques using the CIF AKE data-set, a carefully
curated collection of both real and AI-generated images. Transfer
learning allows us to leverage the knowledge embedded in pr-trained
CNN models, and apply it to the task of deep-fake detection. By fmetuning
these models on the CIF AKE data-set, we aim to enhance their
ability to detect subtle patterns and anomalies that distinguish real images
from synthetic ones. Additionally, the project incorporates Explainable
AI (XAI) techniques, such as Grad-CAM, to provide transparency into
the decision-making process of the models. This intractability is crucial
for building trust in the system's outputs and ensuring that the models are
not only accurate but also understandable.
4.2 FIELD OF INVENTION
This approach will provide valuable insights into the decisionmaking
processes of the selected model, helping to ensure that the
classification results are not only accurate but also transparent and
-Q)
C)
Ill
D..
Q)
··1-- N
E....
I 0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
"1".1'"')
1.1')
-.M".".'." . -::1'
N
0
~
> 0 z I
·.
coP A.T E i'!T "f""" ' .....•.
trustworthy. Through this initiative, we aim to advance the capability to
detect d..e..e p-fake images, supp.o..r..t ing broader efforts ··t··o-· preserve the
authenticity of visual content in an increasingly Al-driven world.
This research can be fine-tuned for the following potential usecases such
as detecting deepfake images in social media and online platforms to
combat misinformation and disinformation, assisting law enforcement
agencies in identifying and preventing fraudulent activities involving
manipulated images and safeguarding the authenticity of digital media in
fields such as journalism, media production, and intellectual property
protection.
4.3 DISCUSSION OF THE RELATED ART
The field of this project lies at the intersection of computer vision,
deep learning, and digital forensics. It focuses on developing methods to
distinguish between real and AT-generated images, a critical task given
the growing prevalence and sophistication of deepfake technology. This
project leverages the power of fine-tuned pre-trained convolutional neural
networks to tackle this challenge, contributing to the development of
robust tools for maintaining the integrity of visual content. The model
achieved a co=endable accuracy of 92.98%, utilizing explainable AI
(XAI) techniques like Grad-CAM to provide insights· into the model's
decision-making process. However, while their approach focuses on highquality
synthetic images generated through latent diffusion, the study has .
explored only basic architectures of CNNs. Al-Dulaimi and Kurnaz [2]
present a hybrid CNN-LSTM model combined with machine learning
classifiers such as Random Forest and SVM. Their system demonstrates
high accuracy on datasets like iFakeFaceDB but faces challenges with
lower performance on others, indicating potential issues with model
generalization across diverse deepfake manipulations.The existing
4
ftFFl C.~ :'. . . . .
f· H F r..i:~-! AI :--. . . . . - ·. ~ ·. . .
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
"1".1'"')
1.1')
-.M".".'." . -::1'
N
0
~
> 0 z
, ...
I .
CO n·;.~T~f..iT
"f""" r-~·.·.\,-·_~,
-------- -------
literature on deepfake detection has explored various deep learning models
and methodo..l.o. gies to tackle the cha...l.l. enge of identifying A..l .-.g enerated
images. Each approach offers unique insights and solutions but also comes
with limitations that our project seeks to address. Bird and Lotti [1]
introduce the CIF AKE dataset and leverage CNN models for binary
classification between real and Al-generated images.
Bhargava et a!. [3] employ a CNN-based method for pixel-level
fake image detection, which is effective for identifying tampered regions
but may struggle with more sophisticated deepfake alterations that do not
leave obvious pixel-based artifacts. Their reliance on pixel-level analysis
limits the model's applicability to more nuanced forms of image
manipulation. Meepaganithage et a!. [4] focus on using ResNet-101 for
image forgery detection, achieving a high accuracy rate of 93.46%. While
their approach shows promising results, it relies on a single model
architecture, which may not capture all the variations and complexities of
deepfake images. Smelyakov et al. [5] evaluated several CNN-based
models for deepfake detection, trained on the Deepfake Detection
Challenge Dataset. Their work highlights the relevance of CNNs in this
domain but lacks exploration into newer architectures.
Thomas et al. [6] introduce a Deepfake Image Classifier System
using the Meso Net architecture, achieving an accuracy of 88.81%. While
their system addresses the rapid growth of deepfake technology, it is
constrained by its accuracy and the use of a single model architecture.
Shahin and Deriche [7] repose a novel framework combining Vision
Transformers (ViT) with Convolutional Autoencoders (CAE). Their
approach achieves approximately 87% accuracy and demonstrates the
potential of hybrid models. However, their methods face challenges in
robustness and generalization across diverse datasets.
5
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
1"".1'"')
1.1')
M""'" -..... -::1'
N o·
~
> 0 z I
....
.C..IO.. p _i-\ \ (: ~~ In a nutshell even though previous studies have made great progress in
identifying deepfak.e. . s :, our initiative intends ..t . o . fill in some gaps in th.e. ... literature .
The majority of research has concentrated on certain architectures or datasets,
which has limited its capacity to generalize to a variety of deepfake
manipulations.In addition to that, while some approaches incorporate
explainability techniques, there is a need for more comprehensive methods that
combine state-of-the-art architectures with interchangeability tools to enhance the
reliability and transparency of deep-fake detection systems. Our project seeks to
fill this gap by developing a robust system that leverages multiple pre trained
models, along with Grad-CAM visualization techniques. By doing so, we aim to
create a more generalization and interpretation deep-fake detection system that
can effectively address the challenges posed by AI-generated images across
various domains.
References
[1] Bird, Jordan & Lotti, Ahmad. (2024). CIFAKE: Image Classification
and Explainable Identification of AI-Generated Synthetic Images.
IEEE Access. PP. 1-1. 10.1109/ACCESS.2024.3356122.
[2] 0. A. H. H. Al-Dulaimi and S. Kurnaz, "Deep fake Image Detection
Based on Deep Learning Using a Hybrid CNN-LSTM with Machine
Learning Architectures as Classifier," 2024 International Congress
on Human-Computer Interaction, Optimization and Robotic
Applications (HORA), Istanbul, Turkiye, 2024, pp. 1-7, doi:
I 0.11 09/HORA61326.2024.1 0550728.
[3] D. Bhargava, S. Rani, M. Singh, N. Tripathi, A. Bhargava and G .
Panwar, "Deep-Fake Finder: Uncovering Forgery Image Through
Neural Network Analysis," 2024 International Conference on
Communication, Computer Sciences and Engineering (IC3SE),
Gautam Buddha Nagar, India, 2024, pp. 1-5, doi:
6
fl i= F "f. (· E ·:-: •. '· -: ...
. ...
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
1"".1'"')
1.1')
M""'" -..... -::1'
N
0
~
> 0 z I
CIO p ATE\'tT "f""" ~ ~ ·. ·- . . - .
10.11 09/IC3SE62002.2024.1 0592889 .
. )4] A. Meepaganith~&~· S. Rath, M. Nicole.s~_u, M. Nicolescu and:~:
Sengupta, "Image Forgery Detection Using Convolutional Neural
Networks," 2024 12th International Symposium on Digital Forensics
and Security (ISDFS), San Antonio, TX, USA, 2024, pp. 1-6, doi:
10.11 09/ISDFS60797.2024.1 0527268.
[5] K. Smelyakov, Y. Kitsenko and A. Chupryna, "Deepfake Detection
Models Based on Machine Learning Technologies," 2024 IEEE
Open Conference of Electrical, Electronic and Information Sciences
( eStream), Vilnius, Lithuania, 2024, pp. 1-6, doi:
10.11 09/eStream61684.2024.1 0542582.
[6] J. M. Thomas, V. Ebenezer and R. P. Richard, "A Deepfake Image
Classifier System for real and Doctored Image Differentiation," 2024
International Conference on Inventive Computation Technologies
(ICICT), Lalitpur, Nepal, 2024, pp. 1247-1251, doi:
10.1109/ICICT60155.2024.10544898.
[7] M. Shahin and M. Deriche, "A Novel Framework based on a Hybrid
Vision Transformer and Deep Neural Network for Deepfake
Detection," 2024 21st International Multi-Conference on Systems,
Signals & Devices (SSD), Erbil, Iraq, 2024, pp. 329-333, doi:
10.11 09/SSD61670.2024.1 0548578 .
4.4 SUMMARY OF INVENTION
The primary objective of our project is to develop a sophisticated
deep learning-based system for accurately distinguishing between genuine
and AI-generated images. In response to the growing challenges posed by
advanced image generation techniques, the project aims to leverage stateof-
the-art deep learning methodologies to enhance image authenticity
verification across various applications. The focus is on addressing the
7
- . -~. 4.· I ::...} -~~- ·..,. i:-~. -f. l." i/~·Q:t~-
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
"1".1'"')
1.1')
-.M".".'." . -::1'
N
0
~
> 0 z I
CO PATEf'.L1 "f""" '· '·. · .. -- .
need for reliable detection mechanisms in the face of increasingly realistic
syn...t.h etic images. To achiev..e... . t his objective, the proj·e··c-·t employs a range of
advanced deep learning approaches, particularly those involving pre
trained models. By utilizing transfer learning and fine-tuning techniques,
the project aims to optimize these models for high performance in
classifying images.
Additionally, the incorporation of explainability techniques such as
Grad-CAM will be used to provide insights into the decision-making
process of the model, thus enhancing the interpretability of the results and
improving the overall reliability of the detection system. The key findings
of this project will include an evaluation of the performance of the
different pre trained model architectures in detecting deepfakes and their
ability to generalize across various datasets. The use of Grad-CAM is
expected to yield valuable. insights into the visual features that are
indicative of deepfake images, which can inform future improvements in
detection methodologies. The integration of advanced deep learning
techniques with interpretability methods promises to provide a
· comprehensive solution to the challenges posed by deepfakes.
4.4 DETAILED DESCRIPTION OF THE INVENTION
Our project focuses on developing an advanced system for
detecting AI-generated images, addressing the growing concerns about
deepfakes and their potential misuse. The core objective is to create a
robust deep learning model that can reliably distinguish between genuine
and synthetic images. To achieve this, we will levt:rage transfer learning
and fmc-tuning techniques with several pre-trained models, tailoring them
specifically for the task at hand. The CIF AKE dataset, which includes a
range of Al-generated and authentic images, will serve as the foundation
8
ft f f I t~'. ~ C. H. ~ N t,\ A I ;-....
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
1"".1'"')
1.1')
M""'" -..... -::1'
N
0
~
> 0 z I
.c..o.. p .f.'-T t= N' T
-----------------------------
for training and evaluation. The methodology for this project involves
several ...k. ey steps. Initially, we. ..:.w ill preprocess the CI·F··- A· KE dataset to ....
ensure that it is suitable for training the models. This preprocessing will
include normalization and data augmentation techniques to enhance the
model's ability to generalize across different image variations. Following
preprocessing, the dataset will be split into training, and validation sets.
The transfer learning approach will be employed where we select several
high-performing pre-trained models. These models will be fine-tuned by
adjusting their hyperparameters to optimize performance for our specific
task. The focus will be on achieving high accuracy, which will be our
primary evaluation metric.
Once the models are trained, they will be evaluated using various
me tries such as classification reports, confusion matrices, accuracy, and
loss. This evaluation will help in selecting the best-performing model.
Additionally, the Grad-CAM (Gradient-weighted Class Activation
Mapping) technique will be implemented to provide interpretability for the
model's predictions. This will offer insights into which features the model
is focusing on when making its classification decisions. To facilitate
practical use and integration, a Flask web application will be developed.
This application will interface with the trained model to provide real-time
predictions.
The architecture of the system is designed to be both efficient and
user-friendly. It starts with the dataset acquisition and preprocessing phase,
followed by dataset splitting. The model training phase involves finetuning
pre-trained models, and once trained, these models are evaluated for
performance. The selected model is then integrated into a user interface via
the Flask app. Users will be able to access this interface to input images,
receive predictions, and interact with the system effectively.
9
f·,·. H•. ·. F';- f.-. .t i.' l. _.A . .. I.
.1.· 4.: :
-Q)
C)
Ill
D..
Q)
-1- N
E....
I 0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
1"".1'"')
1.1')
M""'" -..... -::1'
N
0
~
> 0 z I
CO P·A;TEf-.tT "f""" '· . '. · ...•.
~----~~~----------~-----~--~--~~----~~··
Prediction
UX/UI
1 Modtl
f
~
~ ~ ~~t;l~ lA
+-~--~--~--LJ
Dal2 Dataset
Paprocwing
,____·l==::li.J
Algorithm
t ' User Test hnages
~•
'"· "' ...
ARCIDTECTURE DIAGRAM
The zero-level Data Flow Diagram (DFD) of the system simplifies the
process flow, showing the dataset as the input, the entire system as the
processing unit, and the classification results as the output. This high-level
view highlights the essential function of the system: to process input
images and output accurate classifications
10
fl F F I ( \': ' . . . .
•
'
'• .
....
. -----·--
. .. .... ....
----- -- ----------
··. . .. .
The project encompasses several modules, each playing a crucial
role in the system's functionality. The Data Acquisition and Preprocessing
module ensures that the input data ·is well-prepared for training. The Model
Implementation module focuses on setting up and configuring the deep
learning models. Model Training involves fine-tuning and optimizing these
models. Model Evaluation and Selection are critical. for assessing the
performance of the models and choosing the best one. Finally, Grad'-CAM
for Model Interpretability provides valuable insights into the decisionmaking
process of the modeL Overall, the project aims to offer a
comprehensive solution for detecting Al-generated images by combining
........ ~-·-·---
--------------------------
-Q)
C)
Ill
D..
Q)
'I-N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
-::1'
N
0
~
1"".1'"')
1.1')
M""'" -..... -::1'
N
0
~
> 0 z I
~ PJ-\1~N1
advanced deep learning techniques with practical implementation and
.... interpretability featur..e.s. . This will not only. ....e nhance the accuracy. .. . o. f
de.epfake detection but also provide a user-friendly interface for practical
applications.
5. CLAIMS:
• This research presents a comprchensi ve investigation into the detection
of AI-generated images, commonly known as deep-fakes. By employing
transfer learning and fme tuning we present a comparative analysis of
state-of-the-art convolutional neural networks (CNNs), demonstrating
the feasibility of accurately distinguishing between genuine and
synthetic images.
• The integration of Grad-CAM provides valuable insights into the
model's decision-making process,· enhancing the interpretability and
reliability of our results.
• The successful application of deep learning techniques in this study
underscores the potential of artificial intelligence in combating the
spread of misinformation and protecting the integrity of digital media.
• There are several avenues for future work. Expanding the dataset to
include a broader variety of images and deepfake techniques could
enhance the model's generalization capabilities.
• Exploring more advanced hyperparameter tuning methods and
incorporating newer architectures, such as Vision Transformers, might
yield even better performance. Investigating the integration of pre
trained models with other image authenticity verification techniques
could also strengthen its robustness.
12
f1FF1ft ·,-; '· •. -7 •••
1.,c.._. H'· ·. F.,.... ~·.· i. i.'i. ..A.. I.
-Q)
C)
Ill
D..
Q)
-1- N
E....
0
-LL. (0 .....
0 en
CIO
0 .....
-::1'
·-::~"
N
0
~
"1".1'"')
1.1')
-.M".".'." . -::1'
N
0
~
> 0 z
• Developing reaktime detection systems and deploying th~. model in
practical scenarios would be valuable next steps for practical application
and impact. Overall, this study lays the groundwork for significant
advancements in deep fake detection and sets the stage for continued
research and development in this critical area.
Documents
Name | Date |
---|---|
202441089016-Form 1-181124.pdf | 20/11/2024 |
202441089016-Form 2(Title Page)-181124.pdf | 20/11/2024 |
202441089016-Form 3-181124.pdf | 20/11/2024 |
202441089016-Form 5-181124.pdf | 20/11/2024 |
202441089016-Form 9-181124.pdf | 20/11/2024 |
Talk To Experts
Calculators
Downloads
By continuing past this page, you agree to our Terms of Service,, Cookie Policy, Privacy Policy and Refund Policy © - Uber9 Business Process Services Private Limited. All rights reserved.
Uber9 Business Process Services Private Limited, CIN - U74900TN2014PTC098414, GSTIN - 33AABCU7650C1ZM, Registered Office Address - F-97, Newry Shreya Apartments Anna Nagar East, Chennai, Tamil Nadu 600102, India.
Please note that we are a facilitating platform enabling access to reliable professionals. We are not a law firm and do not provide legal services ourselves. The information on this website is for the purpose of knowledge only and should not be relied upon as legal advice or opinion.