MAFW is a large-scale, multi-modal, compound affective database for dynamic facial expression recognition in the wild. Clips in this database come from China, Japan, Korea, Europe, America and India, and cover various themes, e.g., variety, family, science fiction, suspense, love, comedy, and interviews, encompassing a wide range of human emotions. Each clip has been independently labeled 11 times by 11 well-trained annotators. MAFW database has enormous diversities, large quantities, and rich annotations, including:
- 10,045 number of video clips from movies, TV dramas, and short videos,
- a 11-dimensional expression distribution vector for each video clip,
- three kinds of annotations: (1) single expression label; (2) multiple expression label; (3) bilingual emotional descriptive text,
- two subsets: single-expression set, including 11 classes of single emotions; multiple-expression set, including 32 classes of multiple emotions,
- three automatic annotations: the frame-level 68 facial landmarks, bounding boxes of face regions, and gender,
- four benchmarks : uni-modal single expression classification, multi-modal single expression classification, uni-modal compound expression classification, and multi-modal compound expression classification.
- MAFW database is available for non-commercial research purposes only.
- You agree not to reproduce, duplicate, copy, sell, trade, resell or exploit for commercial purposes, any portion of the clips, and any derived data.
- You agree not to further copy, publish, or distribute any portion of the MAFW database. Except for internal use at a single site within the same organization, it is allowed to make copies of the dataset.
This database is publicly available and free for professors and research scientists affiliated to a university. For students interested in accessing the dataset, please note that the application requires formal endorsement by a faculty member from your institution.
Permission to use (but not reproduce or distribute) the MAFW database is granted only if the following steps are properly followed:
- Download the MAFW-academics -final.pdf document, which serves as the End-User License Agreement (EULA).
- Carefully review the terms and conditions to confirm acceptance. The required information at the end of the document must be completed and signed—for student applicants, this signature must be from a professor at their affiliated university to validate the request.
- Send the fully completed and signed document to: 1202411179@cug.edu.cn.
- After review and approval, you will receive download links via email, including two options: Baidu Netdisk and Google Drive.
Please cite our paper if you find our work useful for your research:
- Yuanyuan Liu, Wei Dai, Chuanxu Feng, Wenbin Wang, Guanghao Yin, Jiabei Zeng, and Shiguang Shan. 2022. MAFW: A Large-scale, Multi-modal, Compound Affective Database for Dynamic Facial Expression Recognition in the Wild. In Proceedings of the 30th ACM International Conference on Multimedia (MM ’22), October 10–14, 2022, Lisboa, Portugal. ACM, New York, NY, USA, 9 pages. https://doi.org/10.1145/3503161.3548190
@inbook{liu_mafw_2022,
author = {Liu, Yuanyuan and Dai, Wei and Feng, Chuanxu and Wang, Wenbin and Yin, Guanghao and Zeng, Jiabei and Shan, Shiguang},
title = {MAFW: A Large-scale, Multi-modal, Compound Affective Database for Dynamic Facial Expression Recognition in the Wild},
year = {2022}
isbn = {978-1-4503-9203-7},
publisher = {ACM},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3503161.3548190},
booktitle = {Proceedings of the 30th ACM International Conference on Multimedia (MM’22)},
numpages = {9}
}
- Data
- Labels
- Labels (auto)
- Train & Test Set
For more details of the dataset, please refer to the paper: MAFW: A Large-scale, Multi-modal, Compound Affective Database for Dynamic Facial Expression Recognition in the Wild.
For more details of emotional descriptive texts, please refer to supplementary materials for MAFW.
After your application is approved, you will receive two download options:
- Baidu Netdisk: Contains the complete dataset, including frame data.
- Google Drive: Does not include frame data (only video clips and label files).
If you require frame data but cannot access Baidu Netdisk, you can extract and process frames from the video clips obtained via Google Drive using the following steps (consistent with the preprocessing pipeline in our paper):
- Extract frame pictures from video clips using tools like OpenCV.
- Face detection and landmark extraction: Use face recognition tools (e.g., face-alignment-master referenced in our paper, or other convenient face detection libraries) to identify face regions and 68 facial landmarks.
- Face alignment and resizing: Perform affine transformation and matrix rotation (via OpenCV or similar libraries) to align faces, then resize the aligned face regions to 224×224 pixels (consistent with the dataset's standard format).
You may use any familiar face detection, landmark extraction, or alignment tool that suits your workflow— the key is to ensure the final output is 224×224 aligned face frames for consistency with the dataset's benchmark settings.
Due to the large size of the clips and frames directories, they are split into multiple compressed files:
- Clips:
clips.7z.001,clips.7z.002, ... (sequential numbering) - Frames:
frames.7z.001,frames.7z.002, ...,frames.7z.010(up to 10 parts)
Extraction instructions:
- Windows: Use 7-Zip or WinRAR. Right-click the first file (e.g.,
clips.7z.001orframes.7z.001) and select "Extract here" – the software will automatically merge all parts into a single folder. - Linux/macOS: Use the
7zcommand in the terminal. Run7z x clips.7z.001(orframes.7z.001) – the tool will detect and process all related parts sequentially.
Critical notes:
- Ensure all split files are in the same folder (do not separate them into subdirectories).
- Do not rename any split files (e.g., avoid changing
clips.7z.001toclips_part1.7z), as this will break the extraction sequence. - Verify that all files are fully downloaded (no corruption or missing parts) – incomplete downloads will cause extraction failures.
- The final extracted folder will be named
clipsorframes(no need to manually merge folders).
The source code of our proposed T-ESFL model can be downloaded in https://github.com/MAFW-database/MAFW.
Please contact us for any questions about MAFW.
| Yuanyuan Liu | Professor, China University of Geosciences | liuyy@cug.edu.cn |
| Shuyang Liu | Master, China University of Geosciences | 20171003670@cug.edu.cn |
| Ying Qian | Master, China University of Geosciences | 1202411179@cug.edu.cn |
For more information, welcome to visit our team's homepage: https://cvlab-liuyuanyuan.github.io/


















