Skip to content

Latest commit

 

History

History
32 lines (19 loc) · 1.51 KB

README.md

File metadata and controls

32 lines (19 loc) · 1.51 KB

scrape-a-grave

Scrape and Retrieve FindAGrave memorial page data and save them to an SQL database.

Scraping

FindAGrave is an index of gravemarkers from cemeteries around the world. Often when doing genealogy research, you don't want to rely on a webpage's future and so you want to download the information to your local file. This python script takes a list of Grave Marker numbers, or FindAGrave urls, scrapes the site for data and prints out a citation of the information. It is currently setup to also save the data in an SQL database.

Requirements

You are expected to have Python3. It also requires the BeautifulSoup package, downloadable through pip:

$ pip3 install bs4

Usage

Download these files and change the contents of input text to be a list of FindAGrave ids, or FindAGrave urls. Then run

$ python3 getgraveids.py

The citations will be printed to the console and saved in an SQL database named graves.db.

It is also possible to read links from a GEDCOM by un-highlighting the "read from gedcom" section. This assumes your GEDCOM source citations have a LINK field with the FindAGrave site.

License

This is intended as a convenient tool for personal genealogy research. Please be aware of FindAGrave's Terms of Service.

MIT © Robert Pirtle