Git Product home page Git Product logo

google-landmark's Introduction

Google Landmarks Dataset v2

NEW: Explore the dataset visually here.

This is the second version of the Google Landmarks dataset (GLDv2), which contains images annotated with labels representing human-made and natural landmarks. The dataset can be used for landmark recognition and retrieval experiments. This version of the dataset contains approximately 5 million images, split into 3 sets of images: train, index and test. The dataset was presented in our CVPR'20 paper and Google AI blog post. A hierarchical extension of the dataset is presented in an under-submission paper to the IEEE Transactions on Pattern Analysis and Machine Intelligence. In this repository, we present download links for all dataset files, baseline models and code for metric computation.

This dataset was associated to two Kaggle challenges, on landmark recognition and landmark retrieval. Results were discussed as part of a CVPR'19 workshop. In this repository, we also provide scores for the top 10 teams in the challenges, based on the latest ground-truth version. Please visit the challenge and workshop webpages for more details on the data, tasks and technical solutions from top teams.

As a reference, the previous version of the Google Landmarks dataset (referred to as Google Landmarks dataset v1, GLDv1) was available here. It is no longer available.

If you make use of this dataset, please consider citing the following papers:

Original GLDv2 CVPR'20 paper: Paper

"Google Landmarks Dataset v2 - A Large-Scale Benchmark for Instance-Level Recognition and Retrieval"
T. Weyand*, A. Araujo*, B. Cao, J. Sim
Proc. CVPR'20

For the hierarchical labels:

"Optimization of Rank Losses for Image Retrieval"
E. Ramzi, N. Audebert, C. Rambour, A. Araujo, X. Bitot, N. Thome
In submission to: IEEE Transactions on Pattern Analysis and Machine Intelligence.

Dataset webpage

Explore the dataset visually here.

Current version

The current dataset version is 2.1. See the release history for details, including re-scored challenge submissions based on the latest ground-truth version.

Download train set

There are 4,132,914 images in the train set.

Download the labels and metadata

Downloading the data

The train set is split into 500 TAR files (each of size ~1GB) containing JPG-encoded images. The files are located in the train/ directory, and are named images_000.tar, images_001.tar, ..., images_499.tar. To download them, access the following link:

https://s3.amazonaws.com/google-landmark/train/images_000.tar

And similarly for the other files.

Using the provided script

mkdir train && cd train
bash ../download-dataset.sh train 499

This will automatically download, verify and extract the images to the train directory.

Note: This script downloads files in parallel. To adjust the number of parallel downloads, modify NUM_PROC in the script.

train image licenses

All images in the train set have CC-BY licenses without the NonDerivs (ND) restriction. To verify the license for a particular image, please refer to train_attribution.csv.

Download index set

There are 761,757 images in the index set.

Download the list of images and metadata

IMPORTANT: Note that the integer landmark id's mentioned here are different from the ones in the train set above.

Downloading the data

The index set is split into 100 TAR files (each of size ~850MB) containing JPG-encoded images. The files are located in the index/ directory, and are named images_000.tar, images_001.tar, ..., images_099.tar. To download them, access the following link:

https://s3.amazonaws.com/google-landmark/index/images_000.tar

And similarly for the other files.

Using the provided script

mkdir index && cd index
bash ../download-dataset.sh index 99

This will automatically download, verify and extract the images to the index directory.

Note: This script downloads files in parallel. To adjust the number of parallel downloads, modify NUM_PROC in the script.

index image licenses

All images in the index set have CC-0 or Public Domain licenses.

Download test set

There are 117,577 images in the test set.

Download the list of images and ground-truth

Downloading the data

The test set is split into 20 TAR files (each of size ~500MB) containing JPG-encoded images. The files are located in the test/ directory, and are named images_000.tar, images_001.tar, ..., images_019.tar. To download them, access the following link:

https://s3.amazonaws.com/google-landmark/test/images_000.tar

And similarly for the other files.

Using the provided script

mkdir test && cd test
bash ../download-dataset.sh test 19

This will automatically download, verify and extract the images to the test directory.

Note: This script downloads files in parallel. To adjust the number of parallel downloads, modify NUM_PROC in the script.

test image licenses

All images in the test set have CC-0 or Public Domain licenses.

Checking the download

We also make available md5sum files for checking the integrity of the downloaded files. Each md5sum file corresponds to one of the TAR files mentioned above; they are located in the md5sum/index/, md5sum/test/ and md5sum/train/ directories, with file names md5.images_000.txt, md5.images_001.txt, etc. For example, the md5sum file corresponding to the images_000.tar file in the index set can be found via the following link:

https://s3.amazonaws.com/google-landmark/md5sum/index/md5.images_000.txt

And similarly for the other files.

If you use the provided download-dataset.sh script, the integrity of the files is already checked right after download.

Extracting the data

We recommend that the set of TAR files corresponding to each dataset split be extracted into a directory per split; ie, the index TARs extracted into an index directory; train TARs extracted into a train directory; test TARs extracted into a test directory. This is done automatically if you use the above download instructions/script.

The directory structure of the image data is as follows: Each image is stored in a directory ${a}/${b}/${c}/${id}.jpg, where ${a}, ${b} and ${c} are the first three letters of the image id, and ${id} is the image id found in the csv files. For example, an image with the id 0123456789abcdef would be stored in 0/1/2/0123456789abcdef.jpg.

Baseline models

We make available the ResNet101-ArcFace baseline model from the paper, see instructions here.

Metric computation code

The metric computation scripts have been made available, via the DELF github repository, see the python scripts compute_recognition_metrics.py and compute_retrieval_metrics.py. These scripts accept as input the ground-truth files, along with predictions in the format submitted to Kaggle.

Dataset licenses

The annotations are licensed by Google under CC BY 4.0 license. The images listed in this dataset are publicly available on the web, and may have different licenses. Google does not own their copyright. Note: while we tried to identify images that are licensed under a Creative Commons Attribution license, we make no representations or warranties regarding the license status of each image and you should verify the license for each image yourself.

Release history

May 2023 (version 2.1)

As an addition to the original dataset, we added hierarchical labels for landmarks.

Sept 2019 (version 2.1)

Ground-truth and labelmaps released. Note that the ground-truth has been substantially updated since the end of the 2019 Kaggle challenges; it is not the one that was used for scoring in the challenge.

We have re-computed metrics for the top 10 teams in the 2019 challenges (see the Kaggle challenge webpages for precise definitions of the metrics):

Recognition metrics

Team Private GAP (%) Public GAP (%)
JL 66.53 61.86
GLRunner 53.08 52.07
smlyaka 69.39 65.85
Chundi Liu 60.86 56.77
Cookpad 33.66 31.12
bestfitting 54.53 52.46
Himanshu Rai 60.32 56.28
Eduardo 46.88 44.07
ods.ai 24.02 22.28
ZFTurbo & Weimin & David 38.99 39.83

Retrieval metrics

Team Private mAP@100 (%) Public mAP@100 (%)
smlyaka 37.14 35.63
imagesearch 34.38 32.04
Layer 6 AI 32.10 29.92
bestfitting 32.12 29.09
ods.ai 29.82 27.82
learner 28.98 27.33
CVSSP 28.07 26.59
Clova Vision, NAVER/LINE Corp. 27.77 25.85
VRG Prague 25.48 23.71
JL 24.98 22.73

May 2019 (version 2.0)

Included data for test and index sets.

Apr 2019 (version 2.0)

Initial version, including only train set.

Contact

For any questions/suggestions/comments/corrections, please open an issue in this github repository, and tag @andrefaraujo. In particular, we plan to maintain and release new versions of the ground-truth as corrections are found.

Paper references

Original GLDv2 paper:

@inproceedings{weyand2020GLDv2,
  author = {Weyand, T. and Araujo, A. and Cao, B. and Sim, J.},
  title = {{Google Landmarks Dataset v2 - A Large-Scale Benchmark for Instance-Level Recognition and Retrieval}},
  year = {2020},
  booktitle = {Proc. CVPR},
}

Hierarchical extension:

@inproceedings{ramzi2023optimization,
  author = {Ramzi, E. and Audebert, N. and Rambour, C. and Araujo, A. and Bitot, X. and Thome, N.},
  title = {{Optimization of Rank Losses for Image Retrieval}},
  year = {2023},
  booktitle = {In submission to: IEEE Transactions on Pattern Analysis and Machine Intelligence},
}

Dataset Metadata

The following table is necessary for this dataset to be indexed by search engines such as Google Dataset Search.

property value
name Google Landmarks Dataset v2
url
description This is the second version of the Google Landmarks dataset (GLDv2), which contains images annotated with labels representing human-made and natural landmarks. The dataset can be used for landmark recognition and retrieval experiments. This version of the dataset contains approximately 5 million images, split into 3 sets of images: train, index and test. The dataset was presented in our CVPR'20 paper. In this repository, we present download links for all dataset files and relevant code for metric computation.

This dataset was associated to two Kaggle challenges, on landmark recognition and landmark retrieval. Results were discussed as part of a CVPR'19 workshop. In this repository, we also provide scores for the top 10 teams in the challenges, based on the latest ground-truth version. Please visit the challenge and workshop webpages for more details on the data, tasks and technical solutions from top teams.

provider
property value
name Google
sameAs https://en.wikipedia.org/wiki/Google
license
The annotations are licensed by Google under CC BY 4.0 license. The images listed in this dataset are publicly available on the web, and may have different licenses. Google does not own their copyright. Note: while we tried to identify images that are licensed under a Creative Commons Attribution license, we make no representations or warranties regarding the license status of each image and you should verify the license for each image yourself.
citation Weyand, T. and Araujo, A. and Cao, B. and Sim, J., "Google Landmarks Dataset v2 - A Large-Scale Benchmark for Instance-Level Recognition and Retrieval", Proc. CVPR 2020

google-landmark's People

Contributors

andrefaraujo avatar dbishai avatar natashafn avatar rm-wu avatar tobias-fischer avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

google-landmark's Issues

geo-location labels

Hi @andrefaraujo,
I am working on a research project with the GLD v2, which requires the geo-location of each image, such as the lat and lon. For example, I want to plot the figure in the paper.
image

Could you please provide me the geo-location label for each image or tell me where I can get this information for plotting such a figure. Thank you so much!

Best,
Chenning

Obtaining location information

@andrefaraujo I would like to be able to get location information (country, region, state, city, etc.) for the landmarks in the data set much like what is available in the dataset web page.

How might someone go about obtaining this kind of information?

Memory Requirements

Can someone tell me, what was the minimum requirement of the disk space for the storage of this entire data?

Few annotated labels in test set

Hi! I am working on the research based on this valuable dataset. However, I found few annotated labels in the test set as for recognition. In summary, 1973 samples have only 849 landmarks compared to the total 200 000 ones. For the rest, 115605 samples' labels are missing. Thus,

  • I wonder if a large number of samples are of non-landmarks.
  • Anyway, I consider there are too few landmarks for testing. Which public test set did your baseline paper use?

Thanks.

duplicate file across each download batch

Hi, thanks for sharing this! Below is the log displayed when I tried to run this script. But it seems for each batch, one file that belongs to the previous batch will be re-downloaded. Any idea what's happening about this duplicate?

Downloading images_000.tar...
Downloading images_001.tar...
Downloading images_002.tar...
Downloading images_003.tar...
Downloading images_004.tar...
Downloading images_005.tar...
Downloading images_006.tar...
images_004.tar extracted!
images_000.tar extracted!
images_002.tar extracted!
images_001.tar extracted!
images_006.tar extracted!
images_003.tar extracted!
images_005.tar extracted!
Downloading images_006.tar...
Downloading images_007.tar...
Downloading images_008.tar...
Downloading images_009.tar...
Downloading images_010.tar...
Downloading images_011.tar...
Downloading images_012.tar...
images_012.tar extracted!
images_011.tar extracted!
images_008.tar extracted!
images_009.tar extracted!
images_006.tar extracted!
images_007.tar extracted!
images_010.tar extracted!

TensorFlow Hub Model - Country Names

Hi!

I'm using the TensorFlow Hub Pre-Trained Model for landmarks in Europe. The labelmap it uses does not have the country name.

In issue #20 there is a reference to a JSON file I could use to list landmarks per country. However, the ID's do not match.

Is there an easy way to get the country name then? Or should I join the CSV and JSON's by name? Are the names exactly equal?

Thank you in advance!

Sort landmarks based on country

Hello @andrefaraujo , is there a way that i can get landmarks that are only in one country from the dataset?

I can see in the visualization here - https://storage.googleapis.com/gld-v2/web/index.html that i can select a country and i get a list of the landmarks in it, is there a variable for the country in which the landmark is located in?

Or is there a way to get a list of the landmarks in a certain country, in my case Bulgaria, that are in the dataset.

Thank you in advance!: )

No such file or directory

Downloading images_000.tar...
Downloading images_001.tar...
Downloading images_002.tar...
Downloading images_003.tar...
Downloading images_004.tar...
Downloading images_006.tar...
Downloading images_005.tar...
md5sum: images_001.tar: No such file or directory
MD5 checksum for images_001.tar did not match checksum in md5.images_001.txt
MD5 checksum for images_006.tar did not match checksum in md5.images_006.txt
md5sum: images_005.tar: No such file or directory
MD5 checksum for images_005.tar did not match checksum in md5.images_005.txt
md5sum: images_004.tar: No such file or directory
cut: md5.images_004.txt: No such file or directory
MD5 checksum for images_004.tar did not match checksum in md5.images_004.txt
cut: md5.images_003.txt: No such file or directory
MD5 checksum for images_003.tar did not match checksum in md5.images_003.txt
MD5 checksum for images_002.tar did not match checksum in md5.images_002.txt
MD5 checksum for images_000.tar did not match checksum in md5.images_000.txt

Adding extra metadata

Would it be possible to add:
• The actual landmark name and it's corresponding id not just the category of it
• Country, Continent, etc.. where each landmark is located

I believe this would help those who are interested in using only subsets of the dataset (e.g. Egyptian/Mexican/Asian/.. landmarks recognition)

3 computed checksum did NOT match in train

cd train

images_172.tar: FAILED
md5sum: WARNING: 1 computed checksum did NOT match
images_206.tar: FAILED
md5sum: WARNING: 1 computed checksum did NOT match
images_225.tar: FAILED
md5sum: WARNING: 1 computed checksum did NOT match

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.