Comments (1)
Hi @jhj7905,
Thanks for your interest in our work and for your question about our data generation methods.
You're correct in your estimation. The dataset is indeed comprised of 30% human-annotated data and 70% semi-automatic annotations. As you correctly pointed out, creating a fully human-annotated dataset can be costly and time-consuming, so we complemented it with semi-automatic methods.
For the semi-automatic portion of our dataset, we developed a comprehensive method that combined predictions from SOTA models to extract relevant cues. We used specific models to eliminate any noisy or irrelevant context from the data. This rigorous process ensured that the data maintained its accuracy and relevance despite not being fully human-annotated.
I hope this clarifies the methodology we used for data generation. Please feel free to reach out if you have any more questions.
from video-chatgpt.
Related Issues (20)
- Question about VideoInstruct_Dataset.json HOT 1
- Question of training utilizing A6000 HOT 6
- Inho jake park inho jake park where are you?
- ActivityNet few shot videos can't be downloaded HOT 7
- Issue running offline demo HOT 1
- Can you provide the original video caption generating the instruction qa? HOT 1
- The download of dataset HOT 1
- Is Zeroshot evaluation on validation or test datasets HOT 1
- Fail to run Video-ChatGPT Demo Offline HOT 2
- The dataset all generated from ChatGPT, it doesn't looked video actually, how to make sure the dataset is right? HOT 3
- Llava Weights not available for running demo offline HOT 2
- Can I deal with Chinese video? HOT 1
- Clarification Needed on the Role of **QA_GT_caption_based_noisy** in the Dataset HOT 1
- Why is OpenAI used for Activitynet-QA Evaluation? HOT 3
- About zero-shot test on TGIF-QA HOT 1
- Fail to execute Video-ChatGPT offline demo (when click "Upload Video") HOT 7
- Code for Dataset Generation HOT 1
- Train errors: some videos in train_video_ids.txt do not appear in activity_clip-14L_spatio_temporal_356 HOT 2
- What is the formal name of the benchmark? HOT 3
- GPT Judge does not work for the Prompt
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from video-chatgpt.