Skip to main content
PyCon US is happening May 14th-22nd in Pittsburgh, PA USA.  Learn more

Multimodal Large Language Model (MLLM) for urban environment evaluation.

Project description

image PyPI Downloads PyPI Downloads Docs image

logo

Urban-Worm

Introduction

Urban-Worm is a Python library that integrates remote sensing imagery, street view data, and vision-language models (VLMs) to assess urban units. Using APIs for data collection and VLMs for inference, Urban-Worm is designed to support the automation of the evaluation for urban environments, including roof integrity, structural condition, landscape quality, and urban perception.

workflow

Features

  • Run VLMs locally with local datasets and ensure information privacy
  • Download building footprints from OSM and global building data released by Bing Maps, with options to filter building footprints by area
  • Search and clip aerial and street view images (via APIs) based on urban units such as parcel and building footprint data
  • Automatically calibrate the orientation of the panorama street view and the extent of the aerial image
  • Visualize results on maps and in tables
  • Interact with LLMs through a streaming chat interface to analyze and interpret results

Installation

install Ollama client

Please make sure Ollama is installed before installing urban-worm

For Linux, users can also install ollama by running in the terminal:

curl -fsSL https://ollama.com/install.sh | sh

For MacOS, users can also install ollama using brew:

brew install ollama

To install brew, run in the terminal:

/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"

Windows users should directly install the Ollama client

install GDAL first

For macOS, Linux, and Windows users, gdal may need to be installed at very begining using conda. Please download and install Anaconda to use conda.

If the installation method above does not work, try to install with conda:

 conda install -c conda-forge gdal

Mac users may install gdal (if the installation method below does not work, try to install with conda):

 brew install gdal

install the package

The package urabnworm can be installed with pip:

pip install urban-worm 

To install the development version from this repo:

pip install -e git+https://github.com/billbillbilly/urbanworm.git#egg=urban-worm

Usage

single-image inference

from urbanworm import UrbanDataSet

data = UrbanDataSet(image = '../docs/data/test1.jpg')
system = '''
    Given a top view image, you are going to roughly estimate house conditions. Your answer should be based only on your observation. 
    The format of your response must include question, answer (yes or no), explanation (within 50 words)
'''
prompt = '''
    Is there any damage on the roof?
'''
data.oneImgChat(system=system, prompt=prompt)
# output:
# {'question': 'Is there any damage on the roof?',
#  'answer': 'no',
#  'explanation': 'No visible signs of damage or wear on the roof',
#  'img': '/9j/4AAQSkZ...'}

multiple (aerial & street view) images inference using OSM data

bbox = (-83.235572,42.348092,-83.235154,42.348806)
data = UrbanDataSet()
data.bbox2Buildings(bbox)

system = '''
    Given a top view image or street view images, you are going to roughly estimate house conditions. 
    Your answer should be based only on your observation. 
    The format of your response must include question, answer (yes or no), explanation (within 50 words) for each question.
'''

prompt = {
    'top': '''
        Is there any damage on the roof?
    ''',
    'street': '''
        Is the wall missing or damaged?
        Is the yard maintained well?
    '''
}

# add the Mapillary key
data.mapillary_key = 'MLY|......'
# use both the aerial and street view images (with type='both')
data.loopUnitChat(system=system, prompt=prompt, type='both', epsg=2253)
# convert results into GeoDataframe
data.to_gdf()

More examples can be found here.

To do

  • One-shot learning in each chat method to help the model get familiar with the questions and expected answers
  • Multiple images inference for pairwise comparison and more
  • Basic plot method in UrbanDataSet class
  • Improve the method dataAnalyst in UrbanDataSet class by adding functionality of feeding a more meaningful introduction of data to LLMs
  • A web UI providing interactive operation and data visualization

The next version (v0.2.0) will have:

  • agent-based city walk simulation
  • search for a unit with an address (using Google APIs)
  • find historical images (using Google APIs)

Legal Notice

This repository and its content are provided for educational purposes only. By using the information and code provided, users acknowledge that they are using the APIs and models at their own risk and agree to comply with any applicable laws and regulations. Users who intend to download a large number of image tiles from any basemap are advised to contact the basemap provider to obtain permission before doing so. Unauthorized use of the basemap or any of its components may be a violation of copyright laws or other applicable laws and regulations.

Acknowledgements

The package is heavily built on the Ollama client and Ollama-python. Credit goes to the developers of these projects.

The functionality about sourcing and processing GIS data (satellite & street view imagery) and 360-degree street view image processing is built on the following open projects. Credit goes to the developers of these projects.

The development of this package is supported and inspired by the city of Detroit.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page