Skip to content

mohit2b/EcoSage

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

35 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

An EcoSage Assistant: Towards Building A Multimodal Plant Care Dialogue Assistant

This reporsitory contains the code for our ECIR 2024 paper "An EcoSage Assistant: Towards Building A Multimodal Plant Care Dialogue Assistant", Advances in Information Retrieval, 46th European Conference on Information Retrieval, ECIR 2024.

Arxiv Paper - Link

An example dialogue from our dataset is illustrated in the figure below:

Dataset Example

In recent times, there has been an increasing awareness about imminent environmental challenges, resulting in people showing a stronger dedication to taking care of the environment and nurturing green life. The current $19.6 billion indoor gardening industry, reflective of this growing sentiment, not only signifies a monetary value but also speaks of a profound human desire to reconnect with the natural world. However, several recent surveys cast a revealing light on the fate of plants within our care, with more than half succumbing primarily due to the silent menace of improper care. Thus, the need for accessible expertise capable of assisting and guiding individuals through the intricacies of plant care has become paramount more than ever. In this work, we make the very first attempt at building a plant care assistant, which aims to assist people with plant(-ing) concerns through conversations. We propose a plant care conversational dataset named Plantational, which contains around 1K dialogues between users and plant care experts. Our end-to-end proposed approach is two-fold : (i) We first benchmark the dataset with the help of various large language models (LLMs) and visual language model (VLM) by studying the impact of instruction tuning (zero-shot and few-shot prompting) and fine-tuning techniques on this task; (ii) finally, we build EcoSage, a multi-modal plant care assisting dialogue generation framework, incorporating an adapter-based modality infusion using a gated mechanism. We performed an extensive examination (both automated and manual evaluation) of the performance exhibited by various LLMs and VLM in the generation of the domain-specific dialogue responses to underscore the respective strengths and weaknesses of these diverse models.

  • Authors: Mohit Tomar, Abhisek Tiwari, Tulika Saha, Prince Jha, Sriparna Saha

Dataset Description

  • Our (Plantational) dataset is curated by converting posts from Reddit and Houzz into dialogues.
  • We create dialogues between the user and the assistant revolving around queries related to plant care.
  • This dataset has the corresponding post link, dialogue between user and assistant, image link (if present), intent, and dialogue acts.
  • The intent categories are “Suggestion”, “Conformation”, “Feedback”, and “Awareness”.
  • The Dialogue Acts categories used are “Greeting” (g), “Question” (q), “Answer” (ans), “Statement-Opinion” (o), “Statement Non-Opinion” (s), “Agreement” (ag), “Disagreement” (dag), “Acknowledge” (a) and “Others” (oth).

Here is the link to our dataset - https://docs.google.com/spreadsheets/d/1wJjnk4WKmtW6nqo8xcblFH-8zmFrhwCY2QkExoNgXjM/edit?usp=sharing

If you consider this work to be useful, please cite it as

@inproceedings{tomar2024ecosage,
  title={An EcoSage Assistant: Towards Building A Multimodal Plant Care Dialogue Assistant},
  author={Tomar, Mohit and Tiwari, Abhisek and Saha, Tulika and Jha, Prince and Saha, Sriparna},
  booktitle={European Conference on Information Retrieval},
  pages={318--332},
  year={2024},
  organization={Springer}
}

This code has been adapted from the "MiniGPT-4" paper. In MiniGPT-4 we have made changes in minigpt_base.py to get our code (ecosage).

Contact

For any queries, feel free to contact Mohit Tomar ([email protected])

About

Multimodal Plant Care Dialogue Assistant

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages