Project Synopsis22
Project Synopsis22
Project Synopsis22
Project
ON
Submitted by
B5520 : Shubhankar Madhukar Patil
B5521 : Vedant Manohar Patil
B5452 : Sarthiki Hegade
B5524 : Vishal Nityanand Pawar
Guided by
Prof. S. S. Mane
Research on Splicing Image Detection This study developed a new splicing image
Algorithms Based on Natural Image Statistical detection algorithm combining (DCT),
Characteristics (DWT), and the robust capabilities of (SVM)
classifier
CIFAKE: Image Classification and This study has proposed a method to improve
Explainable Identification of AI-Generated our waning ability to recognise AI-generated
Synthetic Images images through the use of Computer Vision
and to provide insight into predictions with
visual cues.
Methods of improving the quality of speech- A step-by-step algorithm for creating a speech-
to-text conversion to-text system, as well as possible ways to
improve mathematical, linguistic, and
engineering models to reduce conversion
errors, is proposed.
How to create and use a national cross-domain The experiences reported in this paper indicate
ontology and data infrastructure on the that creating and using a national semantic web
Semantic Web infrastructure is useful from the data
producers’ and data user’s point of view
Amalur: Data Integration Meets Machine This work explored the possibilities of
Learning bringing data integration and ML together.
Toward this direction, we have proposed a data
integration-aware ML system Amalur, which
supports machine learning training and
inference over silos
MobileCLIP: Fast Image-Text Models through In this work introduced MobileCLIP aligned
Multi-Modal Reinforced Training image-text backbones, designed for on-device
CLIP inference (low latency and size).
Introduction:
The rapid advancements in hardware and software, along with the increasing
accessibility of powerful AI models, have opened up new opportunities for
integrating intelligent systems into everyday tools and devices. This project aims to
contribute to this growing field by focusing on solutions that assist individuals,
particularly those who may benefit from enhanced sensory and cognitive aids.
The project aims to create practical AI-driven systems that demonstrate the
potential of technology and address real-world challenges. By utilizing advanced
algorithms and data processing techniques, the systems developed will highlight the
ability of AI to assist in complex tasks that require the interpretation and
understanding of vast amounts of sensory data. The primary motivation behind this
work is to design systems that can enhance everyday life, either by increasing
accessibility for those with impairments or offering innovative solutions to problems
that require advanced data interpretation.
Methodology:
Both CLIP and BLIP are employed within the broader framework of this
project to demonstrate how AI models can be used to process sensory inputs, such as
images, and generate outputs that are accessible and useful to a wide range of users.
The integration of these models is done with scalability in mind, allowing the systems
to handle diverse use cases, ranging from personal data management to accessibility
solutions for individuals with disabilities. The systems developed under this
methodology are designed to be modular, ensuring that they can be adapted for
different contexts without extensive modification. By focusing on the generalized
application of AI for sensory augmentation, this project aims to showcase the
potential of these models to create intuitive, responsive systems that enhance human
experiences with minimal intervention. Through the careful use of AI, the project
ensures that each solution is both efficient and resource-conscious, maintaining
optimal performance across different hardware environments.
Conclusion:
The systems developed are designed with adaptability and scalability in mind,
ensuring their usability across various contexts and environments. By focusing on
real-time processing and user-centric design, the project ensures that these AI
solutions are not only technically robust but also offer meaningful improvements in
quality of life and workflow efficiency. Moreover, this work highlights the broader
potential of AI in creating intelligent systems that can augment human capabilities
and extend the possibilities for interaction with the world around us.
As technology continues to evolve, projects like this pave the way for future
innovations that will further close the gap between human and machine intelligence.
The results of this project serve as a testament to the growing role AI plays in shaping
a more inclusive and efficient future, with applications that go beyond traditional
boundaries to redefine how we perceive and engage with data.
References:
Vision: A Review.” EAI Endorsed Scal Inf Syst [Internet]. 2021 Apr. 21
arXiv:2404.16296
5. Li, M., Zhu, Z., Xu, R., Feng, Y., & Xiao, L. (2024). “Research on Image
Date: Time: