Git Product home page Git Product logo

aistudentsh's Projects

adabound icon adabound

An optimizer that trains as fast as Adam and as good as SGD.

awesome-visual-question-answering icon awesome-visual-question-answering

A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.

bert-as-service icon bert-as-service

Mapping a variable-length sentence to a fixed-length vector using BERT model

bottom-up-attention icon bottom-up-attention

Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome

credit-card-recognition icon credit-card-recognition

A python program to extract the vital information from credit card image such as type of card, card number, card holder's name and validity date using OCR-A font recognition in openCV

lxmert icon lxmert

PyTorch code for EMNLP 2019 paper "LXMERT: Learning Cross-Modality Encoder Representations from Transformers".

mcan-vqa icon mcan-vqa

Deep Modular Co-Attention Networks for Visual Question Answering

n2nmn icon n2nmn

Code release for Hu et al. Learning to Reason: End-to-End Module Networks for Visual Question Answering. in ICCV, 2017

pythia icon pythia

A modular framework for Visual Question Answering research by the FAIR A-STAR team

self_critical_vqa icon self_critical_vqa

Code for NeurIPS 2019 paper ``Self-Critical Reasoning for Robust Visual Question Answering''

spanbert icon spanbert

Code for using and evaluating SpanBERT.

svm-w-smo icon svm-w-smo

Simple implementation of a Support Vector Machine using the Sequential Minimal Optimization (SMO) algorithm for training.

vctree-visual-question-answering icon vctree-visual-question-answering

Code for the Visual Question Answering (VQA) part of CVPR 2019 oral paper: "Learning to Compose Dynamic Tree Structures for Visual Contexts"

vl-bert icon vl-bert

Code for the paper "VL-BERT: Pre-training of Generic Visual-Linguistic Representations".

vqa-project icon vqa-project

Code for our paper: Learning Conditioned Graph Structures for Interpretable Visual Question Answering

vqa-sva icon vqa-sva

Structured Attentions for Visual Question Answering

vqa2.0-recent-approachs-2018.pytorch icon vqa2.0-recent-approachs-2018.pytorch

A pytroch reimplementation of "Bilinear Attention Network", "Intra- and Inter-modality Attention", "Learning Conditioned Graph Structures", "Learning to count object", "Bottom-up top-down" for Visual Question Answering 2.0

vqa2019 icon vqa2019

Our approach to tackle the task of Visual Question Answering

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.