This is the poster presentation for CaGIS + UCGIS Symposium 2024 .
Abstract:
The recent advancements in spatial analytics and geospatial artificial intelligence (GeoAI) have unveiled new horizons for leveraging social media data in realm of natural disaster
management. Focusing on Twitter (now X), a widely recognized platform, this research highlights how user-generated textual contents and associated images can offer invaluable
spatiotemporal insights, aiding in disaster response efforts and bolstering community resilience. Despite the acknowledged potentials of social media data in emergency management,
there is an evident gap in comprehensive studies that integrate GeoAI and cartographic visualizations to swiftly mine and utilize this data for enhancing situational awareness.
The advent of cutting-edge machine learning technologies, especially those facilitating the processing of multimodal data, suggest significant prospects for enriching disaster
management applications. A prime example of such innovation is the Contrastive Language–Image Pre-training (CLIP) model developed by OpenAI, which synergizes text and image analysis
in a unified framework, thus enabling emergency responders and decision makers to gain a deep and more nuanced understanding of crises for timely and adaptive interventions.
This study employs multiple hurricane incidents to evaluation and compare the effectiveness of various machine learning models in processing and integrating Twitter’s text and image
data. The findings affirm that the multimodal approach, particularly through the use of the CLIP model, excels in performing crucial tasks relevant to disaster management such as
informativeness evaluation, humanitarian category classification, and damage severity assessment. After the validation process, the proposed multimodal method is then implemented to
extract disaster relevant information from a distinct Twitter dataset collected during Hurricane Harvey in 2017. The enumerated results of tweets, integrated with spatiotemporal
information, are utilized to generate a suite of static and dynamic cartographic visualizations, which serve to reveal a more comprehensive understanding of the complex dynamics in
this disaster event.
In conclusion, this research demonstrates that the application of CLIP multimodal data mining method stands out as a powerful tool for integrating the user-generated texts and images
in social media, enhancing the disaster management capabilities by offering crucial information extractions and cartographic visualizations, paving the way for future advancements in
the field of natural disaster management.