Skip to content

Scrape and Retrieve FindAGrave memorial page data and save them to an SQL database

License

Notifications You must be signed in to change notification settings

pirtleshell/scrape-a-grave

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

scrape-a-grave

Scrape and Retrieve FindAGrave memorial page data and save them to an SQL database.

Scraping

FindAGrave is an index of gravemarkers from cemeteries around the world. Often when doing genealogy research, you don't want to rely on a webpage's future and so you want to download the information to your local file. This python script takes a list of Grave Marker numbers, or FindAGrave urls, scrapes the site for data and prints out a citation of the information. It is currently setup to also save the data in an SQL database.

Requirements

You are expected to have Python3. It also requires the BeautifulSoup package, downloadable through pip:

$ pip3 install bs4

Usage

Download these files and change the contents of input text to be a list of FindAGrave ids, or FindAGrave urls. Then run

$ python3 getgraveids.py

The citations will be printed to the console and saved in an SQL database named graves.db.

It is also possible to read links from a GEDCOM by un-highlighting the "read from gedcom" section. This assumes your GEDCOM source citations have a LINK field with the FindAGrave site.

License

This is intended as a convenient tool for personal genealogy research. Please be aware of FindAGrave's Terms of Service.

MIT © Robert Pirtle

About

Scrape and Retrieve FindAGrave memorial page data and save them to an SQL database

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages