The goal of this project is to leverage data mining with Selenium and Machine Learning to get good candidate web pages for Demo Days for accelerators. Relevant information on the project can be found on the [http://mcnair.bakerinstitute.org/wiki/Accelerator_Data Accelerator Data] page.
COMPLETE FILES
============-----------------
These files hold data for all the accelerators: not just the test set.
ListOfAccs.txt
The full list of potential keywords (used for throwing out irrelevant results)search terms to match with the text versions of news articles: KeywordsCohortAndAcceleratorsFullList.txt
A list of accelerators, queries, and urls:
A file with the name of the results that passed keyword matching:
DemoDayHitsFull.txt
A file with an analysis of the most frequent matched words in each text file:
topWordsFull.txt
==Faulty Results==
The first pass through the data revealed articles that had thousands of hits for keyword matches. This seemed highly suspicious, so we dug in deeper to investigate the cause of this issue.
The following script in the same directory analyzes the keyword matches to determine the words with the highest number of hits.
DemoDayAnalysis.py
After investigation, it was found that many company names were taken after common english words. Here are some of the companies causing issues along with their associated accelerator:
Rather than removing these companies from the list of search terms, we opted to not include as search terms any words that were considered among the top 10000 most common English words. For reference, we used the top 10000 most common English words according to a Google research study. The github documentation of the study can be found [https://github.com/first20hours/google-10000-english here].
The file containing the 10000 most common English words can be found: