diff --git a/docs/_posts/prabod/2025-04-11-smolvlm_instruct_int4_en.md b/docs/_posts/prabod/2025-04-11-smolvlm_instruct_int4_en.md
new file mode 100644
index 00000000000000..b963b9331659a5
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-11-smolvlm_instruct_int4_en.md
@@ -0,0 +1,115 @@
+---
+layout: model
+title: SmolVLM by HUggingface
+author: John Snow Labs
+name: smolvlm_instruct_int4
+date: 2025-04-11
+tags: [en, openvino, vlm, open_source]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: SmolVLMTransformer
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+SmolVLM is a compact open multimodal model that accepts arbitrary sequences of image and text inputs to produce text outputs. Designed for efficiency, SmolVLM can answer questions about images, describe visual content, create stories grounded on multiple images, or function as a pure language model without visual inputs. Its lightweight architecture makes it suitable for on-device applications while maintaining strong performance on multimodal tasks.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/smolvlm_instruct_int4_en_5.5.1_3.0_1744355673028.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/smolvlm_instruct_int4_en_5.5.1_3.0_1744355673028.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|smolvlm_instruct_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|1.8 GB|
+
+## References
+
+https://huggingface.co/HuggingFaceTB/SmolVLM-Instruct
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-14-paligemma_3b_pt_224_int4_en.md b/docs/_posts/prabod/2025-04-14-paligemma_3b_pt_224_int4_en.md
new file mode 100644
index 00000000000000..218f46d77d7fb7
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-14-paligemma_3b_pt_224_int4_en.md
@@ -0,0 +1,118 @@
+---
+layout: model
+title: PaliGemma 3b 224
+author: John Snow Labs
+name: paligemma_3b_pt_224_int4
+date: 2025-04-14
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: PaliGemmaForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+PaliGemma is a versatile and lightweight vision-language model (VLM) inspired by PaLI-3 and based on open components such as the SigLIP vision model and the Gemma language model. It takes both image and text as input and generates text as output, supporting multiple languages. It is designed for class-leading fine-tune performance on a wide range of vision-language tasks such as image and short video caption, visual question answering, text reading, object detection and object segmentation.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/paligemma_3b_pt_224_int4_en_5.5.1_3.0_1744628366633.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/paligemma_3b_pt_224_int4_en_5.5.1_3.0_1744628366633.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|paligemma_3b_pt_224_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.1 GB|
+
+## References
+
+https://huggingface.co/google/paligemma-3b-pt-224
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-15-paligemma2_3b_mix_224_int4_en.md b/docs/_posts/prabod/2025-04-15-paligemma2_3b_mix_224_int4_en.md
new file mode 100644
index 00000000000000..3787849880dc8c
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-15-paligemma2_3b_mix_224_int4_en.md
@@ -0,0 +1,118 @@
+---
+layout: model
+title: PaliGemma2 3B mix 224 int4
+author: John Snow Labs
+name: paligemma2_3b_mix_224_int4
+date: 2025-04-15
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: PaliGemmaForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+PaliGemma 2 is an update of the PaliGemma vision-language model (VLM) which incorporates the capabilities of the Gemma 2 models. The PaliGemma family of models is inspired by PaLI-3 and based on open components such as the SigLIP vision model and Gemma 2 language models. It takes both image and text as input and generates text as output, supporting multiple languages. It is designed for class-leading fine-tune performance on a wide range of vision-language tasks such as image and short video caption, visual question answering, text reading, object detection and object segmentation.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/paligemma2_3b_mix_224_int4_en_5.5.1_3.0_1744695280010.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/paligemma2_3b_mix_224_int4_en_5.5.1_3.0_1744695280010.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|paligemma2_3b_mix_224_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.0 GB|
+
+## References
+
+https://huggingface.co/google/paligemma2-3b-mix-224
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-15-paligemma2_3b_pt_448_int4_en.md b/docs/_posts/prabod/2025-04-15-paligemma2_3b_pt_448_int4_en.md
new file mode 100644
index 00000000000000..dbb0dc2d97fe92
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-15-paligemma2_3b_pt_448_int4_en.md
@@ -0,0 +1,118 @@
+---
+layout: model
+title: PaliGemma2 3B pt 448 int4
+author: John Snow Labs
+name: paligemma2_3b_pt_448_int4
+date: 2025-04-15
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: PaliGemmaForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+PaliGemma 2 is an update of the PaliGemma vision-language model (VLM) which incorporates the capabilities of the Gemma 2 models. The PaliGemma family of models is inspired by PaLI-3 and based on open components such as the SigLIP vision model and Gemma 2 language models. It takes both image and text as input and generates text as output, supporting multiple languages. It is designed for class-leading fine-tune performance on a wide range of vision-language tasks such as image and short video caption, visual question answering, text reading, object detection and object segmentation.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/paligemma2_3b_pt_448_int4_en_5.5.1_3.0_1744693861187.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/paligemma2_3b_pt_448_int4_en_5.5.1_3.0_1744693861187.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|paligemma2_3b_pt_448_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.0 GB|
+
+## References
+
+https://huggingface.co/google/paligemma2-3b-mix-224
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-15-paligemma_3b_ft_vqav2_448_int4_en.md b/docs/_posts/prabod/2025-04-15-paligemma_3b_ft_vqav2_448_int4_en.md
new file mode 100644
index 00000000000000..4340b430e5423a
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-15-paligemma_3b_ft_vqav2_448_int4_en.md
@@ -0,0 +1,118 @@
+---
+layout: model
+title: PaliGemma 3B ft vqav2 int4
+author: John Snow Labs
+name: paligemma_3b_ft_vqav2_448_int4
+date: 2025-04-15
+tags: [en, openvino, open_source]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: PaliGemmaForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+PaliGemma is a versatile and lightweight vision-language model (VLM) inspired by PaLI-3 and based on open components such as the SigLIP vision model and the Gemma language model. It takes both image and text as input and generates text as output, supporting multiple languages. It is designed for class-leading fine-tune performance on a wide range of vision-language tasks such as image and short video caption, visual question answering, text reading, object detection and object segmentation.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/paligemma_3b_ft_vqav2_448_int4_en_5.5.1_3.0_1744684675991.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/paligemma_3b_ft_vqav2_448_int4_en_5.5.1_3.0_1744684675991.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|paligemma_3b_ft_vqav2_448_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.1 GB|
+
+## References
+
+https://huggingface.co/google/paligemma-3b-pt-224
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-15-paligemma_3b_pt_224_int4_en.md b/docs/_posts/prabod/2025-04-15-paligemma_3b_pt_224_int4_en.md
new file mode 100644
index 00000000000000..00fc7681434281
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-15-paligemma_3b_pt_224_int4_en.md
@@ -0,0 +1,114 @@
+---
+layout: model
+title: PaliGemma 3b pt 224 int4
+author: John Snow Labs
+name: paligemma_3b_pt_224_int4
+date: 2025-04-15
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: PaliGemmaForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+PaliGemma 2 is an update of the PaliGemma vision-language model (VLM) which incorporates the capabilities of the Gemma 2 models. The PaliGemma family of models is inspired by PaLI-3 and based on open components such as the SigLIP vision model and Gemma 2 language models. It takes both image and text as input and generates text as output, supporting multiple languages. It is designed for class-leading fine-tune performance on a wide range of vision-language tasks such as image and short video caption, visual question answering, text reading, object detection and object segmentation.
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/paligemma_3b_pt_224_int4_en_5.5.1_3.0_1744687182569.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/paligemma_3b_pt_224_int4_en_5.5.1_3.0_1744687182569.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|paligemma_3b_pt_224_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.1 GB|
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-28-gemma_3_4b_it_int4_en.md b/docs/_posts/prabod/2025-04-28-gemma_3_4b_it_int4_en.md
new file mode 100644
index 00000000000000..a333cf6e6d0ab9
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-28-gemma_3_4b_it_int4_en.md
@@ -0,0 +1,124 @@
+---
+layout: model
+title: Gemma 3 4B IT INT4
+author: John Snow Labs
+name: gemma_3_4b_it_int4
+date: 2025-04-28
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: Gemma3ForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+Gemma3ForMultiModal can load Gemma 3 Vision models for visual question answering.
+The model consists of a vision encoder, a text encoder, a text decoder and a model merger.
+The vision encoder will encode the input image, the text encoder will encode the input text,
+the model merger will merge the image and text embeddings, and the text decoder will output the answer.
+
+Gemma 3 is a family of lightweight, state-of-the-art open models from Google, built from the same
+research and technology used to create the Gemini models. It features:
+- Large 128K context window
+- Multilingual support in over 140 languages
+- Multimodal capabilities handling both text and image inputs
+- Optimized for deployment on limited resources (laptops, desktops, cloud)
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/gemma_3_4b_it_int4_en_5.5.1_3.0_1745821980536.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/gemma_3_4b_it_int4_en_5.5.1_3.0_1745821980536.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|gemma_3_4b_it_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.1 GB|
\ No newline at end of file
diff --git a/docs/_posts/prabod/2025-04-28-gemma_3_4b_pt_int4_en.md b/docs/_posts/prabod/2025-04-28-gemma_3_4b_pt_int4_en.md
new file mode 100644
index 00000000000000..c5ef93260005e5
--- /dev/null
+++ b/docs/_posts/prabod/2025-04-28-gemma_3_4b_pt_int4_en.md
@@ -0,0 +1,124 @@
+---
+layout: model
+title: Gemma 3 4B PT Int4
+author: John Snow Labs
+name: gemma_3_4b_pt_int4
+date: 2025-04-28
+tags: [en, open_source, openvino]
+task: Image Captioning
+language: en
+edition: Spark NLP 5.5.1
+spark_version: 3.0
+supported: true
+engine: openvino
+annotator: Gemma3ForMultiModal
+article_header:
+ type: cover
+use_language_switcher: "Python-Scala-Java"
+---
+
+## Description
+
+Gemma3ForMultiModal can load Gemma 3 Vision models for visual question answering.
+The model consists of a vision encoder, a text encoder, a text decoder and a model merger.
+The vision encoder will encode the input image, the text encoder will encode the input text,
+the model merger will merge the image and text embeddings, and the text decoder will output the answer.
+
+Gemma 3 is a family of lightweight, state-of-the-art open models from Google, built from the same
+research and technology used to create the Gemini models. It features:
+- Large 128K context window
+- Multilingual support in over 140 languages
+- Multimodal capabilities handling both text and image inputs
+- Optimized for deployment on limited resources (laptops, desktops, cloud)
+
+## Predicted Entities
+
+
+
+{:.btn-box}
+
+
+[Download](https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/models/gemma_3_4b_pt_int4_en_5.5.1_3.0_1745822550666.zip){:.button.button-orange.button-orange-trans.arr.button-icon}
+[Copy S3 URI](s3://auxdata.johnsnowlabs.com/public/models/gemma_3_4b_pt_int4_en_5.5.1_3.0_1745822550666.zip){:.button.button-orange.button-orange-trans.button-icon.button-copy-s3}
+
+## How to use
+
+
+
+
+
+{:.model-param}
+## Model Information
+
+{:.table-model}
+|---|---|
+|Model Name:|gemma_3_4b_pt_int4|
+|Compatibility:|Spark NLP 5.5.1+|
+|License:|Open Source|
+|Edition:|Official|
+|Input Labels:|[image_assembler]|
+|Output Labels:|[answer]|
+|Language:|en|
+|Size:|3.1 GB|
\ No newline at end of file