radekd91 / inferno Goto Github PK
View Code? Open in Web Editor NEW๐ฅ๐ฅ๐ฅ Set the world of 3D faces on fire with INFERNO ๐ฅ๐ฅ๐ฅ
License: Other
๐ฅ๐ฅ๐ฅ Set the world of 3D faces on fire with INFERNO ๐ฅ๐ฅ๐ฅ
License: Other
Could not import SPECTRE. Make sure you pull the repository with submodules to enable SPECTRE.
Traceback (most recent call last):
File "/mnt/workspace/inferno/inferno/models/temporal/external/SpectrePreprocessor.py", line 16, in
from spectre.src.spectre import SPECTRE
ModuleNotFoundError: No module named 'spectre.src'
Could not import EmoSwinModule. SWIN models will not be available. Make sure you pull the repository with submodules to enable SWIN.
Could not import EmoSwinModule. SWIN models will not be available. Make sure you pull the repository with submodules to enable SWIN.
SWIN not found, will not be able to use SWIN models
Looking for checkpoint in '/mnt/workspace/inferno/assets/TalkingHead/models/EMOTE_v2/checkpoints'
Found 1 checkpoints
the link https://download.is.tue.mpg.de/emote/EMOTE_test_example_data.zip is broken
Hi there,
First of all, thank you for your incredible work on EMOTE!
I've been experimenting with training EMOTE and encountered some issues during the second stage. Here's a summary of the problem:
Issue Description:
In the first stage, which involves only the vertex level loss, everything seemed to work smoothly. The loss values descended as expected and converged to some stable values. However, when I moved on to the second stage, which includes both disentangle loss and lip reading loss, I noticed that the loss values for vertex level, lip reading, and disentangle started behaving erratically. They don't seem to descend well, and instead, they vibrate or fluctuate.
My Question:
I'm wondering if you, or anyone else using EMOTE, have encountered similar issues during the second stage of training.
Maybe I have messed up with implementing a custom renderer using pytorch3d,,,I'm not sure hence the issue.
Thanks in advance for any insights or assistance you can provide!
Great work!
If I just want to use FLINT, how do i use it?
Marvelous project! Looking forward to the release of FaceReconstruction model.
Hi, thanks for releasing the code! I want to use the video emotion recognition network, and I found a question in its used module TransformerEncoder. It seems that the newly computed encoded_feature have overwritten the encoded_feature previously calculated using the alibi mask. This does not correspond to the description in the paper.
I also wanted to ask, how long do you usually set the sequence length T when using it?
Followed all the instructions to get the environment running, I've also ran the submodules (optional) script at the start of the instructions. Whenever I try to run the demo, I get the following:
Could not import SPECTRE. Make sure you pull the repository with submodules to enable SPECTRE.
Traceback (most recent call last):
File "/home/ubuntu/inferno/inferno/models/temporal/external/SpectrePreprocessor.py", line 16, in
from spectre.src.spectre import SPECTRE
ModuleNotFoundError: No module named 'spectre'Could not import EmoSwinModule. SWIN models will not be available. Make sure you pull the repository with submodules to enable SWIN.
Could not import EmoSwinModule. SWIN models will not be available. Make sure you pull the repository with submodules to enable SWIN.
SWIN not found, will not be able to use SWIN models
Traceback (most recent call last):
File "demos/demo_eval_talking_head_on_audio.py", line 21, in
from inferno_apps.TalkingHead.evaluation.evaluation_functions import *
File "/home/ubuntu/inferno/inferno_apps/TalkingHead/evaluation/evaluation_functions.py", line 35, in
from psbody.mesh import Mesh
File "/home/ubuntu/miniconda3/envs/work38/lib/python3.8/site-packages/psbody/mesh/init.py", line 10, in
from .meshviewer import MeshViewer, MeshViewers
File "/home/ubuntu/miniconda3/envs/work38/lib/python3.8/site-packages/psbody/mesh/meshviewer.py", line 49, in
from OpenGL import GL, GLU, GLUT
File "/home/ubuntu/miniconda3/envs/work38/lib/python3.8/site-packages/OpenGL/GLUT/init.py", line 5, in
from OpenGL.GLUT.fonts import *
File "/home/ubuntu/miniconda3/envs/work38/lib/python3.8/site-packages/OpenGL/GLUT/fonts.py", line 20, in
p = platform.getGLUTFontPointer( name )
File "/home/ubuntu/miniconda3/envs/work38/lib/python3.8/site-packages/OpenGL/platform/baseplatform.py", line 350, in getGLUTFontPointer
raise NotImplementedError(
NotImplementedError: Platform does not define a GLUT font retrieval function
Any suggestions where I may be going wrong?
Thank you!
I downloaded the processed MEAD dataset using download_processed_mead.sh you provided. However, it seems that the landmarks from the processed/landmarks_original/.../landmarks.pkl do not align with the emica's reconstruction after flame lbs and orthogonal projection using the predicted 'cam', 'shape', 'exp' and 'pose'. Could you kindly elaborate on what the input images (how to warp it) of the EMICA is?
The first image is the processed 478 landmarks (from processed/landmarks_original/.../landmarks.pkl) drawn on the warped image using the landmarks_original.pkl. The second image is the projected 2d mediapipe lmks onto the same image using the reconstruction from the processed/reconstructions/.../shape_pose_cam.hdf5.
Hello, thanks for your nice work at first! However, when I ran wget https://download.is.tue.mpg.de/emote/mead_25fps/processed/metadata.pkl -O metadata.pkl to download metadata.pkl, it couldn't find the file, could you give me some help!
Thank you for this great work!
I downloaded a part of the dataset of mead and followed the data processing method of readme, modified the file path of input, and output, and set detect landmark to true in order, but I got this error.
I located VideoFaceDetectionDataset here, self.index_for_frame_map are all 0, resulting in detection_in_ frame_index always 0, so there is an error reading the next frame. But I don't know how to fix it.
Is there something wrong with my steps๏ผ More details are below.
Thanks for sharing the code.
I wonder when can we expect the update for EMICA.
There are detailed documentation for the use of instructions
Hi ๐ I noticed that the video emotion recognition is still under construction. Is there any plan to release this module soon? Thanks!
Hi @radekd91, thanks again for releasing the code for EMOTE!
I am training EMOTE stage 1, and I noticed that while the training loss is converging, the validation loss goes up - is this something you noticed in your runs?
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.