Photo by Sharon McCutcheon on Unsplash
Every election cycle brings its own brand of excitement – and lots of money. Political donations are of particular interest to political scientists and other researchers studying politics and voting patterns. They are also of interest to citizens who want to stay informed of how much money their candidates raise and where that money comes from.
In the United States, “only American citizens (and immigrants with green cards) can contribute to federal politics, but the American divisions of foreign companies can form political action committees (PACs) and collect contributions from their American employees.”1
In this assignment we will scrape and work with data foreign connected PACs that donate to US political campaigns. First, we will get data foreign connected PAC contributions in the 2020 election cycle. Then, you will use a similar approach to get data such contributions from previous years so that we can examine trends over time.
In order to complete this assignment you will need a Chrome browser with the Selector Gadget extension installed.
Click on the assignment link, clone it in RStudio and open the R Markdown document. Knit the document to make sure it compiles without errors.
Before we introduce the data, let’s warm up with some simple exercises. Update the YAML of your R Markdown file with your information, knit, commit, and push your changes. Make sure to commit with a meaningful commit message. Then, go to your repo on GitHub and confirm that your changes are visible in your Rmd and md files. If anything is missing, commit and push again.
We’ll use the tidyverse package for much of the data wrangling and visualisation, the robotstxt package to check if we’re allowed to scrape the data, the rvest package for data scraping, and the scales package for better formatting of labels on visualisations. These packages are already installed for you. You can load them by running the following in your Console:
library(tidyverse)
library(robotstxt)
library(rvest)
library(scales)
This assignment does not come with any prepared datasets. Instead you’ll be scraping the data!
The data come from OpenSecrets.org, a “website tracking the influence of money on U.S. politics, and how that money affects policy and citizens’ lives”. This website is hosted by The Center for Responsive Politics, which is a nonpartisan, independent nonprofit that “tracks money in U.S. politics and its effect on elections and public policy.”2
Before getting started, let’s check that a bot has permissions to access pages on this domain.
library(robotstxt)
paths_allowed("https://www.opensecrets.org")
## [1] TRUE
Our goal is to scrape data for contributions in all election years Open Secrets has data for. Since that means repeating a task many times, let’s first write a function that works on the first page. Confirm it works on a few others. We will be scraping from the page https://www.opensecrets.org/political-action-committees-pacs/foreign-connected-pacs/2022. Note that you can select the year from a drop down.
Then iterate it over pages for all years.
Complete the following set of steps in the scrape-pac.R
file in the scripts
folder of your repository. This file
already contains some starter code to help you out.
Write a function called scrape_pac()
that scrapes
information from the Open Secrets webpage for foreign-connected PAC
contributions in a given year. This function should
snake_case
naming.Country of Origin/Parent Company
variable
with str_squish()
.year
. We will
want this information when we ultimately have data from all years, so
this is a good time to keep track of it. Our function doesn’t take a
year argument, but the year is embedded in the URL, so we can extract it
out of there, and add it as a new column. Use the str_sub()
function to extract the last 4 characters from the URL. You will
probably want to look at the help for this function to figure out how to
specify “last 4 characters”.Define the URLs for 2022. 2020, 2018, through 2000 contributions and more! This dataset is constantly updated and the dates may shift. Use the drop-down menu to make sure you cover all dates and keep these updated. Then, test your function using these URLs as inputs. Does the function seem to do what you expected it to do?
Construct a vector called urls
that contains the
URLs for each webpage that contains information on foreign-connected PAC
contributions for a given year.
Map the scrape_pac()
function over urls
in a way that will result in a data frame called
pac_all
.
Write the data frame to a csv file called
pac-all.csv
in the data
folder.
✅⬆️ If you haven’t yet done so, now is definitely a good time to commit and push your changes to GitHub with an appropriate commit message (e.g. “Data scraping complete”). Make sure to commit and push all changed files so that your Git pane is cleared up afterwards.
Complete the following set of steps in the hw-06.Rmd
file in your repository.
pac-all.csv
and report
its number of observations and variables using inline code.In this section we clean the pac_all
data frame to
prepare it for analysis and visualization. We have two goals in data
cleaning:
Separate the country_parent
into two such that
country and parent company appear in different columns for country-level
analysis.
Convert contribution amounts in total
,
dems
, and repubs
from character strings to
numeric values.
The following exercises walk you through how to make these fixes to the data.
Use the separate()
function to separate
country_parent
into country
and
parent
columns. Note that country and parent company names
are separated by \
(which will need to be specified in your
function) and also note that there are some entries where the
\
sign appears twice and in these cases we want to only
split the value at the first occurrence of \
. This can be
accomplished by setting the extra
argument in to
"merge"
so that the cell is split into only 2 segments,
e.g. we want "Denmark/Novo Nordisk A/S"
to be split into
"Denmark"
and "Novo Nordisk A/S"
. (See help
for separate()
for more on this.) End your code chunk by
printing out the top 10 rows of your data frame (if you just type the
data frame name it should automatically do this for you).
Remove the character strings including $
and
,
signs in the total
, dems
,and
repubs
columns and convert these columns to numeric. End
your code chunk by printing out the top 10 rows of your data frame (if
you just type the data frame name it should automatically do this for
you). A couple hints to help you out:
The $
character is a special character so it will
need to be escaped.
Some contribution amounts are in the millions
(e.g. Anheuser-Busch contributed a total of $1,510,897 in 2008). In this
case we need to remove all occurrences of ,
, which we can
do by using str_remove_all()
instead of
str_remove()
.
🧶 ✅ ⬆️ Now is a good time to knit your document, commit, and push your changes to GitHub with an appropriate commit message. Make sure to commit and push all changed files so that your Git pane is cleared up afterwards.
Create a line plot of total contributions from all foreign-connected PACs in the Canada and Mexico over the years. Once you have made the plot, write a brief interpretation of what the graph reveals. Few hints to help you out:
Canada
and Mexico
.group_by()
then
summarise()
.Note: The figure you create might look slightly different than this one if the data on the website has been updated recently.
Hint:
You will need to use pivot_longer
to create this
graphic.
🧶 ✅ ⬆️ Knit, commit, and push your changes to GitHub with an appropriate commit message. Make sure to commit and push all changed files so that your Git pane is cleared up afterwards and review the md document on GitHub to make sure you’re happy with the final state of your work.
Source: Open Secrets - About.↩︎