Show and Tell Using Computer; A Deep learning approach.

The Artificial intelligence is making its way in all fields of the world by improving every type of work. The field of Artificial Intelligence (AI) is based on understanding cognition, learning and teaching like humans. It's one of the interesting topic that everybody wants to learn how it works. Artificial Neural Networks are forming an important part in Artificial intelligence. I would like to share my experience on Artificial Neural Network, Artificial Neural Networks are used to generate Artificial Intelligence. Artificial Neural Networks are spreading everywhere, the Artificial Neural Network is an attempt to simulate the functions of biological neurons in human brain.

This deep learning application can benefit people with visual impairments. Well-trained models can be deployed within screen readers to help visually challenged people to understand the context and the contents that an image may be trying to communicate. A screen reader is essentially a type of assistive technology that manifests itself as software liaisons or through specially designed hardware devices used by visually challenged people to navigate the web. It is also efficacious in general utilisation of electronic devices such as mobile phones, computers, and tablets. The main goal of a screen reading assistive technology is to encode and transcribe the visual data into synthesised speech. If the screen reader contains a braille output, it generates relevant braille patterns that the user can recognize by tracing their fingers over as guided by the braille system.

Our final year Master’s project is a Captioning System for Image, which produces output in form of text and speech. Our web application system can identify the majority of images and caption them correctly with decent accuracy, But there are some problems with this web application, where a new image may not be identified accurately amongst this model. The reason for this issue is the lack of training image data. However, we have proposed a solution whereas the future model of our application will train a broad range of data sets processed through the model. We have implemented the Transformer deep learning model in our web application, which is the primary model in training and captioning huge sets of images from the data sets. We used the Django framework and Python for back-end. We train our data set using high performance GPU using software as a service provided by PaperSpace where we used the GPU RTX5000 with 30GB RAM to train our model. We created a prototype by using JavaScript for client-side dynamic interactions and form submissions with the Django framework for web-server handling. Finally, we deployed the Web application online using the Heroku Deployment service by SalesForce and synchronised the code to UbiOps inference system, which will process the user input image and leverage the training from the data set to produce accurate and descriptive captions.

Creating an Image caption generation application using deep learning and the front-end application using respective proprietary technologies has been a challenging yet fulfilling endeavour that has allowed the members of this team to not only learn and implement new things, but also helped us build character and learn to work in a team. Coming up with a well-trained deep learning model that is capable of discerning image features while at the same time, using said features for generating grammatically correct semantic sentences is a remarkable achievement. This project serves as a stepping stone for future developments and advancements in web technologies and the field of deep learning as there is always scope for improvement be it the image feature extractions, generation speed of the model or the caption generated by the model. Hopefully, This project marks a major milestone in the years to come and will be remembered with fondness and the memories of hard work and creative ingenuity will be long remembered.