Git Product home page Git Product logo

evaluation-ai's Introduction

Evaluation-AI: A LLM for Assessing Government Programs

The Evaluation-AI repository is dedicated to building a comprehensive collection of program evaluation reports and leveraging state-of-the-art techniques in large language models (LLMs) to create a powerful tool for assessing government programs. Our mission is to develop a model that can effectively understand and analyze policy frameworks using retrieval-augmented generation (RAG) and a hybrid search strategy. We are committed to transparency and open-source fairness measurement to ensure the reliability and equity of our models.

To run a Streamlit app of a local LLM specialized in program evaluation critiques, go to our page to run the app codes: https://github.com/casualcomputer/evaluation-ai-pro. Our backend design is as follows: Backend

Objectives

Repository of Program Evaluation Reports:

  • Curate and maintain a rich collection of program evaluation reports, serving as a valuable resource for researchers and practitioners.
  • Advanced LLM Training: Utilize the latest advancements in LLMs to train a model that comprehensively understands policy frameworks and program evaluation contexts.
  • Retrieval-Augmented Generation (RAG): Employ a hybrid search strategy combining traditional retrieval methods with generative capabilities to enhance information extraction and comprehension.
  • Contextual Application: Develop and tailor the tool for various tasks within the program evaluation context, ensuring its relevance and effectiveness in real-world applications.

Design:

Backend pipelines:

  1. Scrape and clean reports
  2. RAG for document retrival and summarization a. retrival b. summarization c. LLM validation/critique
  3. Human-in-the-loop for user feedback
  4. Model evaluation and online/batch learning

Frontend:

  1. Chat interface

Usage

Clone the repository

git clone https://github.com/casualcomputer/evaluation-ai.git

Navigate to the script location

you can use -h with the scripts to see the help messages.

cd evaluation-ai/src/data/
python 00_load_raw_data.py
python 01_extract_text.py

Data Sources

Source Name Source Link Number of Extracted Reports
ESDC Link 177
CRA Link 196
Health Canada Link 129
Natural Resources Canada Link 119

Naming Convention

  • Reports are named as <department acronym>_<id>_<title>.<extension>

evaluation-ai's People

Contributors

aminkln avatar casualcomputer avatar jpegmpeg avatar

Watchers

 avatar  avatar

evaluation-ai's Issues

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.