m7mdal7aj commited on
Commit
25fe977
1 Parent(s): b5dba24

Update my_model/utilities/ui_manager.py

Browse files
Files changed (1) hide show
  1. my_model/utilities/ui_manager.py +9 -11
my_model/utilities/ui_manager.py CHANGED
@@ -53,12 +53,18 @@ class UIManager():
53
  st.header('(Knowledge-Based Visual Question Answering)')
54
  with open("Files/Model Arch.html", 'r', encoding='utf-8') as f:
55
  model_arch_html = f.read()
56
- col1, col2, col3 = st.columns([2, 1, 1])
57
  with col1:
58
  st.text('')
59
  st.text('')
60
  st.text('')
61
  st.write("""\n\n\n\n\n\n""")
 
 
 
 
 
 
62
  st.header("Abstract")
63
  st.write("""Navigating the frontier of the Visual Turing Test, this research delves into multimodal learning to bridge the gap between visual perception and linguistic interpretation, a foundational challenge in artificial intelligence. It scrutinizes the integration of visual cognition and external knowledge, emphasizing the pivotal role of the Transformer model in enhancing language processing and supporting complex multimodal tasks.
64
  This research explores the task of Knowledge-Based Visual Question Answering (KB-VQA), it examines the influence of Pre-Trained Large Language Models (PT-LLMs) and Pre-Trained Multimodal Models (PT-LMMs), which have transformed the machine learning landscape by utilizing expansive, pre-trained knowledge repositories to tackle complex tasks, thereby enhancing KB-VQA systems.
@@ -67,16 +73,8 @@ class UIManager():
67
  \nThe evaluation results underscore the developed model’s competent and competitive performance. It achieves a VQA score of 63.57% under syntactic evaluation and excels with an Exact Match (EM) score of 68.36%. Further, semantic evaluations yield even more impressive outcomes, with VQA and EM scores of 71.09% and 72.55%, respectively. These results demonstrate that the model effectively applies reasoning over the visual context and successfully retrieves the necessary knowledge to answer visual questions.""")
68
  st.write("""\n\n\nThis is an interactive application built to demonstrate the project developed and allow for interaction with the KB-VQA model as part of the dissertation for Masters degree in Artificial Intelligence at the [University of Bath](https://www.bath.ac.uk/).
69
  \n\n\nDeveloped by: [Mohammed H AlHaj](https://www.linkedin.com/in/m7mdal7aj)""")
70
- with col2:
71
- # Display the HTML content in Streamlit
72
- #st.markdown(model_arch_html, unsafe_allow_html=True)
73
- #st.image("Files/Model Arch.png")
74
- # Display the HTML content in Streamlit
75
- st.header("Model Architecture")
76
- components.html(model_arch_html, height=1500)
77
-
78
- with col3:
79
- st.image("Files/mm.jpeg")
80
  st.write("""I am profoundly grateful for the support and guidance I have received throughout the course of my dissertation. I would like to extend my deepest appreciation to the following individuals:
81
  \nTo my supervisor, [Dr. Andreas Theophilou](https://researchportal.bath.ac.uk/en/persons/andreas-theophilou), whose expertise, and insightful guidance have been instrumental in the completion of this research. Your mentorship has not only profoundly shaped my work but also my future endeavours in the field of computer science.
82
  Special mention must be made of my mentors at the University of Bath— [Dr. Ben Ralph](https://researchportal.bath.ac.uk/en/persons/ben-ralph), [Dr. Hongping Cai](https://researchportal.bath.ac.uk/en/persons/hongping-cai), and [Dr. Nadejda Roubtsova](https://researchportal.bath.ac.uk/en/persons/nadejda-roubtsova). The wealth of knowledge and insights I have gained from you has been indispensable. Your unwavering dedication to academic excellence and steadfast support have been crucial in navigating my academic journey.
 
53
  st.header('(Knowledge-Based Visual Question Answering)')
54
  with open("Files/Model Arch.html", 'r', encoding='utf-8') as f:
55
  model_arch_html = f.read()
56
+ col1, col2 = st.columns([1,1])
57
  with col1:
58
  st.text('')
59
  st.text('')
60
  st.text('')
61
  st.write("""\n\n\n\n\n\n""")
62
+ st.header("Model Architecture")
63
+ components.html(model_arch_html, height=1500)
64
+
65
+
66
+ with col2:
67
+ st.image("Files/mm.jpeg")
68
  st.header("Abstract")
69
  st.write("""Navigating the frontier of the Visual Turing Test, this research delves into multimodal learning to bridge the gap between visual perception and linguistic interpretation, a foundational challenge in artificial intelligence. It scrutinizes the integration of visual cognition and external knowledge, emphasizing the pivotal role of the Transformer model in enhancing language processing and supporting complex multimodal tasks.
70
  This research explores the task of Knowledge-Based Visual Question Answering (KB-VQA), it examines the influence of Pre-Trained Large Language Models (PT-LLMs) and Pre-Trained Multimodal Models (PT-LMMs), which have transformed the machine learning landscape by utilizing expansive, pre-trained knowledge repositories to tackle complex tasks, thereby enhancing KB-VQA systems.
 
73
  \nThe evaluation results underscore the developed model’s competent and competitive performance. It achieves a VQA score of 63.57% under syntactic evaluation and excels with an Exact Match (EM) score of 68.36%. Further, semantic evaluations yield even more impressive outcomes, with VQA and EM scores of 71.09% and 72.55%, respectively. These results demonstrate that the model effectively applies reasoning over the visual context and successfully retrieves the necessary knowledge to answer visual questions.""")
74
  st.write("""\n\n\nThis is an interactive application built to demonstrate the project developed and allow for interaction with the KB-VQA model as part of the dissertation for Masters degree in Artificial Intelligence at the [University of Bath](https://www.bath.ac.uk/).
75
  \n\n\nDeveloped by: [Mohammed H AlHaj](https://www.linkedin.com/in/m7mdal7aj)""")
76
+ st.write("\n\n")
77
+ st.header("Acknowledgement")
 
 
 
 
 
 
 
 
78
  st.write("""I am profoundly grateful for the support and guidance I have received throughout the course of my dissertation. I would like to extend my deepest appreciation to the following individuals:
79
  \nTo my supervisor, [Dr. Andreas Theophilou](https://researchportal.bath.ac.uk/en/persons/andreas-theophilou), whose expertise, and insightful guidance have been instrumental in the completion of this research. Your mentorship has not only profoundly shaped my work but also my future endeavours in the field of computer science.
80
  Special mention must be made of my mentors at the University of Bath— [Dr. Ben Ralph](https://researchportal.bath.ac.uk/en/persons/ben-ralph), [Dr. Hongping Cai](https://researchportal.bath.ac.uk/en/persons/hongping-cai), and [Dr. Nadejda Roubtsova](https://researchportal.bath.ac.uk/en/persons/nadejda-roubtsova). The wealth of knowledge and insights I have gained from you has been indispensable. Your unwavering dedication to academic excellence and steadfast support have been crucial in navigating my academic journey.