USITC

From edegan.com
Revision as of 15:34, 26 October 2017 by Hbrown512 (talk | contribs)
Jump to navigation Jump to search


McNair Project
USITC
Project logo 02.png
Project Information
Project Title USITC Data
Owner Harrison Brown
Start Date 9/11/2017
Deadline
Primary Billing
Notes In Progress
Has project status Active
Copyright © 2016 edegan.com. All Rights Reserved.


Files

My files are in 2 different places: Did development work here

E:\McNair\Projects\USITC

I had to use to the Postgres SQL server to download the PDFS The Postgres SQL Server:

128.42.44.182/bulk/USITC

Old Work

This is the work that was done before I knew about the XML files provided by the USITC (see below)

Here is a csv of the data that I have been able to scrape from the HTML of https://www.usitc.gov/secretary/fed_reg_notices/337.htm The results.csv file found here,

E:\McNair\Projects\USITC

For every notice paper, there is a line in the CSV file that contains the Investigation Title, Investigation No., link to the PDF on the website, Notice description, and date the notice was issued

I have also downloaded the PDFS from the website. These are the pdfs from the links that are in the csv file. Some of the PDFS were not able to be downloaded. The PDFs are here

E:\McNair\Projects\USITC\pdfs_copy

These files were downloaded using the script on the Postgres Server. There are issues downloading PDFs onto the remote windows machine for some reason. You must download the PDFs on the Postgres Server and transfer them to the RDP. The script to download the PDFS is

128.42.44.182\bulk\USITC\download

Using the pdf scraper from previous project found here

E:\McNair\software\utilities\PDF_RIPPER

You can scrape the PDFs. This file was modified to scrape all of the pdfs in the pdfs folder. The modified code is in the McNair/Project/USITC/ directory and it is called pdf_to_text_bulk.py

An example of PDF parsing that works is parsing this pdf: https://www.usitc.gov/secretary/fed_reg_notices/337/337_959_notice02062017sgl.pdf

E:\McNair\Projects\USITC\Parsed_Texts\337_959_notice02062017sgl.txt

There are PDFs where the parsing does not work completely and the text is scrambled.

I have started using NLTK to gather information about the filings. See code in Projects/USITC/ProcessingTexts I am working on extracting the respondents from these documents.

Alternative Solutions

337Info - Unfair Import Investigations Information System

https://pubapps2.usitc.gov/337external/

This link has links to Complaints. Most of the data that can be gathered is there and it is in a nice format that could be easily web scraped if we needed to. They have a query builder (link below). There is also a link that download all of the raw data in JSON or XML.

Here are links to various statistics we could use: Section 337 Statistics: Settlement Rate Data

https://www.usitc.gov/intellectual_property/337_statistics_settlement_rate_data.htm

Section 337 Statistics

https://www.usitc.gov/press_room/337_stats.htm

Section 337 Statistics: Number of New, Completed, and Active Investigations by Fiscal Year (Updated Quarterly)

https://www.usitc.gov/intellectual_property/337_statistics_number_new_completed_and_active.htm

Section 337 Statistics This contains links to various other pages with statistics

https://www.usitc.gov/intellectual_property/337_statistics.htm

Here is a dictionary of terms used in these Documents

https://www.usitc.gov/documents/337Info_ext_DataDictionary.pdf


There are FAQs listed here

https://www.usitc.gov/documents/337Info_FAQ.pdf

There is a query builder for Section 337 Notices Here

https://pubapps2.usitc.gov/337external/advanced

To use you must select fields from the GUI at the bottom of the page


Status

Currently the web scraper is able to gather all of the data that I can gather from the HTML. There are a few cases where the Investigation Number is not listed and I need to test for those and fix that in the code.

Downloaded most of the PDFs. There were errors download some of the files. I need to calculate what PDFs were not able to be downloaded and why. Investigating what other ways we can gather the information.