Git Product home page Git Product logo

pyresearchinsights's Introduction

pyResearchInsights: An open-source Python package for scientific text analysis

Contents

1.0 Introduction

2.0 Installation

3.0 How it works

4.0 Results

Citation

  • Academic publishing has risen 2-fold in the past ten years, making it nearly impossible to sift through a large number of papers and identify broad areas of research within disciplines.

  • In order to understand such vast volumes of research, there is a need for automated content analysis (ACA) tools.

  • However, existing ACA tools such are expensive and lack in-depth analysis of publications.

  • To address these issues, we developed pyResearchInsights an end-to-end, open-source, automated content analysis tool that:

    • Scrapes abstracts from scientific repositories,
    • Cleans the abstracts collected,
    • Analyses temporal frequency of keywords,
    • Visualizes themes of discussions using natural language processing.

This project is a collaboration between Sarthak J. Shetty, from the Center for Ecological Sciences, Indian Institute of Science and Vijay Ramesh, from the Department of Ecology, Evolution & Environmental Biology, Columbia University.

⚠️ Note: pyLDAvis one of the dependencies of pyResearchInsights, depends on sklearn. sklearn has now been renamed to scikit-learn. To install the pyResearchInsights, set this environment variable and continue with the installation ⚠️

export SKLEARN_ALLOW_DEPRECATED_SKLEARN_PACKAGE_INSTALL=True

To install the package using pip, use the command:

pip install pyResearchInsights

Note: pyResearchInsights requires python 3.7 to run. If you're unsure about the local dependencies on your system, please use package on Google Colab or create a python 3.7 Conda environment and install pyResearchInsights inside that environment.

Since pyResearchInsights is available on pip, it can be run on Google Colab as well where users can leverage Google's powerful CPU and GPU hardware.

Run on Google Colab

Cite This!

Bias Pipeline

Figure 3.1 Diagrammatic representation of the pipeline.

pyResearchInsights is modular in nature. Each part of the package can be run independently or part of a larger pipeline.

This is an example pipeline, where we scrape abstracts from Springer pertaining to the conservation efforts in the Western Ghats.

from pyResearchInsights.common_functions import pre_processing
from pyResearchInsights.Scraper import scraper_main
from pyResearchInsights.Cleaner import cleaner_main
from pyResearchInsights.Analyzer import analyzer_main
from pyResearchInsights.NLP_Engine import nlp_engine_main

'''Abstracts containing these keywords will be queried from Springer'''
keywords_to_search = "Western Ghats Conservation"

'''Calling the pre_processing functions here so that abstracts_log_name and status_logger_name is available across the code.'''
abstracts_log_name, status_logger_name = pre_processing(keywords_to_search)

'''Runs the scraper here to scrape the details from the scientific repository'''
scraper_main(keywords_to_search, abstracts_log_name, status_logger_name)

'''Cleaning the corpus here before any of the other modules use it for analysis'''
cleaner_main(abstracts_log_name, status_logger_name)

'''Calling the Analyzer Function here'''
analyzer_main(abstracts_log_name, status_logger_name)

'''Calling the visualizer code below this portion'''
nlp_engine_main(abstracts_log_name, status_logger_name)

Each module of the pacakage can be run independtely, as described in the following sections:

'''Importing pre_processing() which generates LOG files during the code run'''
from pyResearchInsights.common_functions import pre_processing

'''Importing the scraper_main() which initiates the scraping process'''
from pyResearchInsights.Scraper import scraper_main

'''Abstracts containing these keywords will be scraped from Springer'''
keywords_to_search = "Valdivian Forests Conservation"

'''The refernce to the LOG folder and the status_logger are returned by pre_processing() here'''
abstracts_log_name, status_logger_name = pre_processing(keywords_to_search)

'''Calling the scraper_main() to start the scraping processing'''
scraper_main(keywords_to_search, abstracts_log_name, status_logger_name)

Here,

  • keywords - Abstracts queried from Springer will contain these keywords.

  • abstracts_log_name - The .txt file containing the abstracts downloaded.

  • status_logger_name - File that contains logs the sequence of functions executed for later debugging.

  • This script downloads abstracts from Springer containing the keywords "Valdivian Forests Conservation".

'''Importing the cleaner_main() to clean the txt file of abstracts'''
from pyResearchInsights.Cleaner import cleaner_main

'''The location of the file to be cleaned is mentioned here'''
abstracts_log_name = "/location/to/txt/file/to/be/cleaned"

'''status_logger() logs the seequence of functions executed during the code run'''
status_logger_name = "Status_Logger_Name"

'''Calling the cleaner_main() here to clean the text file provided'''
cleaner_main(abstracts_log_name, status_logger_name)

Here,

  • abstracts_log_name - The .txt file containing the abstracts to be cleaned before generating research themes.

  • status_logger_name - File that contains logs the sequence of functions executed for later debugging.

  • This script cleans the file_name.txt and generates a file_name_CLEANED.txt file. Abstracts available online are often riddled with poor formatting and special characters.

Figure 3.2.1a The text collected by the Scraper consists of special characters (second last line in figure above, '30\xc2\xa0cm'), which has to be cleaned before performing topic-modelling

Figure 3.2.1b The Cleaner gets rid of the special characters seen throughout the corpus as in Figure 3.2.1a, and thereby ensures legible topic-modelling results
'''Importing the analyzer_main() to analyze the frequency of keywords encountered in the text file'''
from pyResearchInsights.Analyzer import analyzer_main

'''The location of the file to be analyzed is mentioned here'''
abstracts_log_name = "/location/to/txt/file/to/be/analyzed"

'''status_logger() logs the seequence of functions executed during the code run'''
status_logger_name = "Status_Logger_Name"

'''Calling the cleaner_main() here to analyze the text file provided'''
analyzer_main(abstracts_log_name, status_logger_name)

Here,

  • abstracts_log_name - The .txt file containing the abstracts to be analyzed for temporal frequency of various keywords.

  • status_logger_name - File that contains logs the sequence of functions executed for later debugging.

  • This script analyzes the frequency of different keywords occuring in texts contained in file_name.txt, and generates a file_name_FREQUENCY_CSV.csv file.

'''Importing the nlp_engine_main() to generate the interactive topic modelling charts'''
from pyResearchInsights.NLP_Engine import nlp_engine_main

'''The location of the abstracts which will be used to train the language models'''
abstracts_log_name = "/location/to/txt/file/to/be/analyzed"

'''status_logger() logs the seequence of functions executed during the code run'''
status_logger_name = "Status_Logger_Name"

'''Calling the nlp_engine_main() here to train the language models on the texts provided'''
nlp_engine_main(abstracts_log_name, status_logger_name)

Note: The Visualizer is integrated within the NLP_Engine function.

Here,

  • abstracts_log_name - The .txt file containing the abstracts from which research themes are to be generated.

  • status_logger_name - File that contains logs the sequence of functions executed for later debugging.

  • This script generates the topic modelling and the frequency/weight charts for the abstracts in the abstracts_log_name file.

Topic Modelling Chart

Figure 4.1 Distribution of topics presented as pyLDAvis charts

  • Circles indicate topics generated from the .txt file supplied to the NLP_Engine.py. The number of topics here can be varied usine the --num_topics flag of the NLP_Engine.
  • Each topic is made of a number of keywords, seen on the right.
  • More details regarding the visualizations and the udnerlying mechanics can be checked out here.

Weights and Frequncy

Figure 4.2 Here, we plot the variation in the weights and frequency of topic keywords.

  • The weight of a keyword is calculated by its: i) frequency of occurance in the corpus and, ii) its frequency of co-occurance with other keywords in the same topic.

If you use pyReasearchInsights for your research, please cite us! Here is the BibTex entry for our 2021 Ecology & Evolution paper.

	@article{shetty2021pyresearchinsights,
	title={pyResearchInsights—An open-source Python package for scientific text analysis},
	author={Shetty, Sarthak J and Ramesh, Vijay},
	journal={Ecology and Evolution},
	volume={11},
	number={20},
	pages={13920--13929},
	year={2021},
	publisher={Wiley Online Library}
	}

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.