A React + AWS Serverless full stack implementation of the 30 example applications found in the official OpenAI API documentation. This repository is used as an instructional tool for the YouTube channel "Full Stack With Lawrence" as well as for University of British Columbia course, "Artificial Intelligence Cloud Technology Implementation" taught by Lawrence McDaniel.
IMPORTANT DISCLAIMER: AWS' Lambda service has a hard 29-second timeout. OpenAI API calls often take longer than this, in which case the AWS API Gateway endpoint will return a 504 "Gateway timeout error" response to the React client. This happens frequently with apps created using chatgpt-4. Each of the 30 OpenAI API example applications are nonetheless implemented exactly as they are specified in the official documentation.
Code composition as of Dec-2023:
-------------------------------------------------------------------------------
Language files blank comment code
-------------------------------------------------------------------------------
Markdown 50 739 6 2357
HCL 23 296 574 1948
Python 14 352 369 1306
YAML 19 109 101 1077
JavaScript 39 112 126 1074
JSX 5 37 41 771
CSS 5 31 14 172
make 1 22 32 82
INI 2 15 0 69
HTML 2 1 0 65
Text 3 3 0 56
Jupyter Notebook 1 0 186 48
Bourne Shell 4 10 44 32
TOML 1 3 0 32
Dockerfile 1 4 4 7
-------------------------------------------------------------------------------
SUM: 170 1734 1497 9096
-------------------------------------------------------------------------------
Complete documentation is located here.
React app that leverages Vite.js, @chatscope/chat-ui-kit-react, and react-pro-sidebar.
- robust, highly customizable chat features
- A component model for implementing your own highly personalized OpenAI apps
- Skinnable UI for each app
- Includes default assets for each app
- Small compact code base
- Robust error handling for non-200 response codes from the custom REST API
- Handles direct text input as well as file attachments
- Info link to the OpenAI API official code sample
- Build-deploy managed with Vite
Complete documentation is located here.
A REST API implementing each of the 30 example applications from the official OpenAI API Documentation using a modularized Terraform approach. Leverages OpenAI's suite of AI models, including GPT-3.5, GPT-4, DALL·E, Whisper, Embeddings, and Moderation.
- OpenAI API library for Python. LangChain enabled API endpoints where designated.
- Pydantic based CI-CD friendly Settings configuration class that consistently and automatically manages Python Lambda initializations from multiple sources including bash environment variables,
.env
andterraform.tfvars
files. - CloudWatch logging
- Terraform fully automated and parameterized build. Usually builds your infrastructure in less than a minute.
- Secure: uses AWS role-based security and custom IAM policies. Best practice handling of secrets and sensitive data in all environments (dev, test, CI-CD, prod). Proxy-based API that hides your OpenAI API calls and credentials. Runs on https with AWS-managed SSL/TLS certificate.
- Excellent documentation
- AWS serverless implementation. Free or nearly free in most cases
- Deploy to a custom domain name
- git. pre-installed on Linux and macOS
- make. pre-installed on Linux and macOS.
- AWS account
- AWS Command Line Interface
- Terraform. If you're new to Terraform then see Getting Started With AWS and Terraform
- OpenAI platform API key. If you're new to OpenAI API then see How to Get an OpenAI API Key
- Python 3.11: for creating virtual environment used for building AWS Lambda Layer, and locally by pre-commit linters and code formatters.
- NodeJS: used with NPM for local ReactJS developer environment, and for configuring/testing Semantic Release.
- Docker Compose: used by an automated Terraform process to create the AWS Lambda Layer for OpenAI and LangChain.
Detailed documentation for each endpoint is available here: Documentation
To get community support, go to the official Issues Page for this project.
This project demonstrates a wide variety of good coding best practices for managing mission-critical cloud-based micro services in a team environment, namely its adherence to 12-Factor Methodology. Please see this Code Management Best Practices for additional details.
We want to make this project more accessible to students and learners as an instructional tool while not adding undue code review workloads to anyone with merge authority for the project. To this end we've also added several pre-commit code linting and code style enforcement tools, as well as automated procedures for version maintenance of package dependencies, pull request evaluations, and semantic releases.
We welcome contributions! There are a variety of ways for you to get involved, regardless of your background. In addition to Pull requests, this project would benefit from contributors focused on documentation and how-to video content creation, testing, community engagement, and stewards to help us to ensure that we comply with evolving standards for the ethical use of AI.
For developers, please see:
- the Developer Setup Guide
- and these commit comment guidelines 😬😬😬 for managing CI rules for automated semantic releases.
You can also contact Lawrence McDaniel directly.