Recognition: 2 theorem links
· Lean TheoremFaceLiVTv2: An Improved Hybrid Architecture for Efficient Mobile Face Recognition
Pith reviewed 2026-05-10 17:57 UTC · model grok-4.3
The pith
FaceLiVTv2 refines a hybrid CNN-Transformer design to deliver better accuracy at lower latency for mobile face recognition.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
FaceLiVTv2 achieves improved accuracy-efficiency trade-offs by introducing Lite MHLA, which reduces redundancy in global token interactions through multi-head linear projections and affine rescale transformations, and embedding it within a RepMix block that unifies local-global feature coordination with global depthwise convolution.
What carries the argument
Lite MHLA, a lightweight global token interaction module that replaces multi-layer attention with multi-head linear token projections and affine rescale transformations while preserving head diversity.
Load-bearing premise
The measured gains in latency and accuracy come primarily from the Lite MHLA and RepMix architectural changes rather than from unreported differences in training data, optimization settings, or hardware tuning.
What would settle it
Re-training FaceLiVTv2 and the compared models from scratch with identical data, optimizer settings, and hardware, then observing no latency reduction or accuracy gain, would show that the architectural claims do not hold.
Figures
read the original abstract
Lightweight face recognition is increasingly important for deployment on edge and mobile devices, where strict constraints on latency, memory, and energy consumption must be met alongside reliable accuracy. Although recent hybrid CNN-Transformer architectures have advanced global context modeling, striking an effective balance between recognition performance and computational efficiency remains an open challenge. In this work, we present FaceLiVTv2, an improved version of our FaceLiVT hybrid architecture designed for efficient global--local feature interaction in mobile face recognition. At its core is Lite MHLA, a lightweight global token interaction module that replaces the original multi-layer attention design with multi-head linear token projections and affine rescale transformations, reducing redundancy while preserving representational diversity across heads. We further integrate Lite MHLA into a unified RepMix block that coordinates local and global feature interactions and adopts global depthwise convolution for adaptive spatial aggregation in the embedding stage. Under our experimental setup, results on LFW, CA-LFW, CP-LFW, CFP-FP, AgeDB-30, and IJB show that FaceLiVTv2 consistently improves the accuracy-efficiency trade-off over existing lightweight methods. Notably, FaceLiVTv2 reduces mobile inference latency by 22% relative to FaceLiVTv1, achieves speedups of up to 30.8% over GhostFaceNets on mobile devices, and delivers 20-41% latency improvements over EdgeFace and KANFace across platforms while maintaining higher recognition accuracy. These results demonstrate that FaceLiVTv2 offers a practical and deployable solution for real-time face recognition. Code is available at https://github.com/novendrastywn/FaceLiVT.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces FaceLiVTv2, an improved hybrid CNN-Transformer architecture for efficient mobile face recognition. It replaces the original multi-layer attention with Lite MHLA (multi-head linear token projections plus affine rescale) and integrates it into a RepMix block that adds global depthwise convolution for spatial aggregation. Under the authors' experimental setup, FaceLiVTv2 is reported to improve the accuracy-efficiency trade-off on LFW, CA-LFW, CP-LFW, CFP-FP, AgeDB-30, and IJB benchmarks, with concrete gains of 22% lower mobile latency versus FaceLiVTv1, up to 30.8% speedup versus GhostFaceNets, and 20-41% latency reductions versus EdgeFace and KANFace while preserving higher accuracy. Code is released.
Significance. If the accuracy and latency deltas are shown to arise from the Lite MHLA and RepMix substitutions under matched training and measurement conditions, the work would provide a practical, deployable advance for real-time face recognition on edge devices. The public code release is a positive factor for reproducibility.
major comments (1)
- [Abstract] Abstract: the central claim that FaceLiVTv2 improves the accuracy-efficiency trade-off 'under our experimental setup' rests on direct comparisons to FaceLiVTv1, GhostFaceNets, EdgeFace, and KANFace, yet no information is supplied on whether those baselines were retrained with identical data, augmentations, loss, optimizer schedule, input resolution, or hardware. Face-recognition accuracy is known to vary several percent from such factors alone; without this control the attribution of the reported 22%, 30.8%, and 20-41% deltas specifically to Lite MHLA and RepMix cannot be verified and is load-bearing for the paper's conclusion.
minor comments (1)
- The abstract refers to 'multi-layer attention design' in the original FaceLiVT without a concise recap of the differences; a short comparison table or paragraph in the introduction would help readers assess the incremental contribution of Lite MHLA.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback on our manuscript. We address the concern regarding baseline comparisons point by point below and will revise the manuscript to improve clarity on the experimental controls.
read point-by-point responses
-
Referee: [Abstract] Abstract: the central claim that FaceLiVTv2 improves the accuracy-efficiency trade-off 'under our experimental setup' rests on direct comparisons to FaceLiVTv1, GhostFaceNets, EdgeFace, and KANFace, yet no information is supplied on whether those baselines were retrained with identical data, augmentations, loss, optimizer schedule, input resolution, or hardware. Face-recognition accuracy is known to vary several percent from such factors alone; without this control the attribution of the reported 22%, 30.8%, and 20-41% deltas specifically to Lite MHLA and RepMix cannot be verified and is load-bearing for the paper's conclusion.
Authors: We agree that the manuscript should explicitly document the training and measurement protocols to allow readers to verify the source of the reported deltas. In the revised manuscript we will add a new subsection (4.1.1) titled 'Baseline Implementation and Training Protocol' that details: (i) which baselines were re-implemented and trained from scratch using the exact same dataset splits, augmentations, loss (ArcFace with identical margin and scale), optimizer (SGD with the same momentum and weight decay), learning-rate schedule, and input resolution (112×112) as FaceLiVTv2; (ii) which baselines used official pre-trained weights (with citations to their original papers) and the hardware/platform used for latency measurement (same mobile device and batch size=1); and (iii) a table summarizing these controls for every compared method. Latency numbers were all obtained on the same device under identical conditions. These additions will make the attribution to Lite MHLA and RepMix verifiable while preserving the original experimental results. revision: yes
Circularity Check
No circularity: architecture design and empirical claims are independent of self-referential inputs
full rationale
The paper introduces Lite MHLA (multi-head linear projections + affine rescale) and RepMix (with global depthwise conv) as explicit design choices for the FaceLiVTv2 hybrid block, then reports measured accuracy and latency on standard face-recognition benchmarks against external baselines (GhostFaceNets, EdgeFace, KANFace, prior FaceLiVTv1). No equations, fitted parameters, or predictions appear in the provided text; performance deltas are direct experimental outcomes rather than quantities forced by construction from the same inputs. Self-reference to FaceLiVTv1 is versioning only and not invoked as a uniqueness theorem or load-bearing premise. The derivation chain consists of engineering substitutions validated externally, satisfying the self-contained benchmark criterion for score 0.
Axiom & Free-Parameter Ledger
free parameters (1)
- model hyperparameters and training settings
axioms (1)
- domain assumption Standard assumptions of i.i.d. training data and convergence of gradient-based optimization
invented entities (2)
-
Lite MHLA
no independent evidence
-
RepMix block
no independent evidence
Lean theorems connected to this paper
-
IndisputableMonolith/Cost/FunctionalEquation.leanwashburn_uniqueness_aczel unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
Lite MHLA replaces the original multi-layer attention design with multi-head linear token projections and affine rescale transformations
-
IndisputableMonolith/Foundation/AlexanderDuality.leanalexander_duality_circle_linking unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
FaceLiVTv2 reduces mobile inference latency by 22% relative to FaceLiVTv1... 30.8% over GhostFaceNets
What do these tags mean?
- matches
- The paper's claim is directly supported by a theorem in the formal canon.
- supports
- The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
- extends
- The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
- uses
- The paper appears to rely on the theorem as machinery.
- contradicts
- The paper's claim conflicts with a theorem or certificate in the canon.
- unclear
- Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.
Reference graph
Works this paper leans on
-
[1]
A comprehensive review of face recognition techniques, trends, and challenges,
H. L. Gururaj, B. C. Soundarya, S. Priya, J. Shreyas, and F. Flam- mini, “A comprehensive review of face recognition techniques, trends, and challenges,” IEEE Access, vol. 12, pp. 107903–107926, 2024
2024
-
[2]
Detect faces efficiently: A survey and evaluations,
Y. Feng, S. Yu, H. Peng, Y.-R. Li, and J. Zhang, “Detect faces efficiently: A survey and evaluations,” IEEE Trans. Biom. Behav. Identity Sci., vol. 4, no. 1, pp. 1–18, 2022
2022
-
[3]
Edgeface: Efficient face recognition model for edge devices,
A. George, C. Ecabert, H. O. Shahreza, K. Kotwal, and S. Marcel, “Edgeface: Efficient face recognition model for edge devices,” IEEE Trans. Biom. Behav. Identity Sci., 2024
2024
-
[4]
Swiftfaceformer: An efficient and lightweight hybrid architecture for accurate face recognition applications,
L. S. Luevano, Y. Mart ´ınez-D´ıaz, H. M ´endez-V´azquez, M. Gonz ´alez-Mendoza, and D. Frey, “Swiftfaceformer: An efficient and lightweight hybrid architecture for accurate face recognition applications,” in Int. Conf. Pattern Recognit. (ICPR) , pp. 244–258, Springer, 2024
2024
-
[5]
Kanface: A novel approach to face recognition using kolmogorov-arnold net- works,
H. Pham, B. Dao, P . N. Tran, T. Nguyen, and C. Do, “Kanface: A novel approach to face recognition using kolmogorov-arnold net- works,” in Int. Symp. Neural Netw. (ISNN) , pp. 149–160, Springer, 2025
2025
-
[6]
Mobilefacenets: Efficient cnns for accurate real-time face verification on mobile devices,
S. Chen, Y. Liu, X. Gao, and Z. Han, “Mobilefacenets: Efficient cnns for accurate real-time face verification on mobile devices,” in Chin. Conf. Biometric Recognit. (CCBR), pp. 428–438, Springer, 2018
2018
-
[7]
MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications
A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861, 2017
work page internal anchor Pith review arXiv 2017
-
[8]
Ghostfacenets: Lightweight face recognition model from cheap operations,
M. Alansari, O. A. Hay, S. Javed, A. Shoufan, Y. Zweiri, and N. Werghi, “Ghostfacenets: Lightweight face recognition model from cheap operations,” IEEE Access , vol. 11, pp. 35429–35446, 2023
2023
-
[9]
Ghostnet: More features from cheap operations,
K. Han, Y. Wang, Q. Tian, J. Guo, C. Xu, and C. Xu, “Ghostnet: More features from cheap operations,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1580–1589, 2020
2020
-
[10]
Transface: Calibrating transformer training for face recognition from a data- centric perspective,
J. Dan, Y. Liu, H. Xie, J. Deng, H. Xie, X. Xie, and B. Sun, “Transface: Calibrating transformer training for face recognition from a data- centric perspective,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 20642–20653, 2023
2023
-
[11]
An image is worth 16x16 words: Transformers for image recognition at scale,
A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. , “An image is worth 16x16 words: Transformers for image recognition at scale,” Int. Conf. Learn. Represent. (ICLR), 2021
2021
-
[12]
Edgenext: efficiently amalgamated cnn-transformer architecture for mobile vision applications,
M. Maaz, A. Shaker, H. Cholakkal, S. Khan, S. W. Zamir, R. M. Anwer, and F. Shahbaz Khan, “Edgenext: efficiently amalgamated cnn-transformer architecture for mobile vision applications,” in Eur. Conf. Comput. Vis. (ECCV), pp. 3–20, Springer, 2022
2022
-
[13]
Kan: Kolmogorov-arnold networks,
Z. Liu, Y. Wang, S. Vaidya, F. Ruehle, J. Halverson, M. Soljaˇci´c, T. Y. Hou, and M. Tegmark, “Kan: Kolmogorov-arnold networks,” Int. Conf. Learn. Represent. (ICLR), vol. 1, no. 2, p. 3, 2025
2025
-
[14]
Swiftformer: Efficient additive attention for transformer- based real-time mobile vision applications,
A. Shaker, M. Maaz, H. Rasheed, S. Khan, M.-H. Yang, and F. S. Khan, “Swiftformer: Efficient additive attention for transformer- based real-time mobile vision applications,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 17425–17436, 2023
2023
-
[15]
Facelivt: Face recognition using linear vision transformer with structural reparameterization for mobile device,
N. Setyawan, C.-C. Sun, M.-H. Hsu, W.-K. Kuo, and J.-W. Hsieh, “Facelivt: Face recognition using linear vision transformer with structural reparameterization for mobile device,” in Proc. IEEE Int. Conf. Image Process. (ICIP), pp. 1720–1725, 2025
2025
-
[16]
Facelivt: Energy efficient face recognition with linear vision transformer for limited resource device,
N. Setyawan, J.-X. Gu, C.-C. Sun, M.-H. Hsu, W.-K. Kuo, C.-A. Shen, and J.-W. Hsieh, “Facelivt: Energy efficient face recognition with linear vision transformer for limited resource device,” inProc. IEEE Asia Pacific Conf. Circuits Syst. (APCCAS), pp. 1–5, IEEE, 2025
2025
-
[17]
Arcface: Additive angular margin loss for deep face recognition,
J. Deng, J. Guo, N. Xue, and S. Zafeiriou, “Arcface: Additive angular margin loss for deep face recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 4690–4699, 2019
2019
-
[18]
Texture-guided transfer learning for low-quality face recognition,
M. Zhang, R. Liu, D. Deguchi, and H. Murase, “Texture-guided transfer learning for low-quality face recognition,” IEEE Trans. Image Process., vol. 33, pp. 95–107, 2023
2023
-
[19]
Benchmarking lightweight face archi- tectures on specific face recognition scenarios,
Y. Martinez-Diaz et al. , “Benchmarking lightweight face archi- tectures on specific face recognition scenarios,” Artif. Intell. Rev. , vol. 54, no. 8, pp. 6201–6244, 2021
2021
-
[20]
Vargfacenet: An efficient variable group convolutional neural network for lightweight face recognition,
M. Yan, M. Zhao, Z. Xu, Q. Zhang, G. Wang, and Z. Su, “Vargfacenet: An efficient variable group convolutional neural network for lightweight face recognition,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. Workshops (ICCVW), pp. 0–0, 2019
2019
-
[21]
Mixfacenets: Extremely efficient face recognition networks,
F. Boutros, N. Damer, M. Fang, F. Kirchbuchner, and A. Kuijper, “Mixfacenets: Extremely efficient face recognition networks,” in IEEE Int. Joint Conf. Biometrics (IJCB) , pp. 1–8, IEEE, 2021
2021
-
[22]
Face transformer for recognition,
Y. Zhong and W. Deng, “Face transformer for recognition,” arXiv preprint arXiv:2103.14803, 2021
-
[23]
Cosface: Large margin cosine loss for deep face recognition,
H. Wang, Y. Wang, Z. Zhou, X. Ji, D. Gong, J. Zhou, Z. Li, and W. Liu, “Cosface: Large margin cosine loss for deep face recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 5265–5274, 2018
2018
-
[24]
Magface: A univer- sal representation for face recognition and quality assessment,
Q. Meng, S. Zhao, Z. Huang, and F. Zhou, “Magface: A univer- sal representation for face recognition and quality assessment,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR) , pp. 14225–14234, 2021
2021
-
[25]
Elasticface: Elastic margin loss for deep face recognition,
F. Boutros, N. Damer, F. Kirchbuchner, and A. Kuijper, “Elasticface: Elastic margin loss for deep face recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1578–1587, 2022
2022
-
[26]
Adaface: Quality adaptive margin for face recognition,
M. Kim, A. K. Jain, and X. Liu, “Adaface: Quality adaptive margin for face recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 18750–18759, 2022
2022
-
[27]
Topofr: A closer look at topology alignment on face recognition,
J. Dan, Y. Liu, J. Deng, H. Xie, S. Li, B. Sun, and S. Luo, “Topofr: A closer look at topology alignment on face recognition,”Adv. Neural Inf. Process. Syst. (NeurIPS), vol. 37, pp. 37213–37240, 2024
2024
-
[28]
Arface: Attention-aware and regularization for face recognition with reinforcement learning,
L. Zhang, L. Sun, L. Yu, X. Dong, J. Chen, W. Cai, C. Wang, and X. Ning, “Arface: Attention-aware and regularization for face recognition with reinforcement learning,” IEEE Trans. Biom. Behav. Identity Sci., vol. 4, no. 1, pp. 30–42, 2022
2022
-
[29]
Catface: Cross-attribute-guided transformer with self-attention distillation for low-quality face recognition,
N. Alipour Talemi, H. Kashiani, and N. M. Nasrabadi, “Catface: Cross-attribute-guided transformer with self-attention distillation for low-quality face recognition,” IEEE Trans. Biom. Behav. Identity Sci., vol. 6, no. 1, pp. 132–146, 2024
2024
-
[30]
Resolution invariant face recognition using a distillation approach,
S. S. Khalid, M. Awais, Z.-H. Feng, C.-H. Chan, A. Farooq, A. Ak- bari, and J. Kittler, “Resolution invariant face recognition using a distillation approach,” IEEE Trans. Biom. Behav. Identity Sci., vol. 2, no. 4, pp. 410–420, 2020
2020
-
[31]
Labeled faces in the wild: A database for studying face recognition in unconstrained environments,
G. B. Huang, M. Mattar, T. Berg, and E. Learned-Miller, “Labeled faces in the wild: A database for studying face recognition in unconstrained environments,” in Workshop Faces Real-Life Images: Detection, Alignment, Recognit., 2008
2008
-
[32]
Agedb: the first manually collected, in-the-wild age database,
S. Moschoglou, A. Papaioannou, C. Sagonas, J. Deng, I. Kotsia, and S. Zafeiriou, “Agedb: the first manually collected, in-the-wild age database,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 51–59, 2017
2017
-
[33]
Shufflefacenet: A lightweight face architecture for efficient and highly-accurate face recognition,
Y. M.-D., L. S. Luevano, H. Me., M. N.-D., L. Chang, and M. G.- M., “Shufflefacenet: A lightweight face architecture for efficient and highly-accurate face recognition,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. Workshops (ICCVW), pp. 0–0, 2019
2019
-
[34]
Transface++: Re- thinking the face recognition paradigm with a focus on accuracy, efficiency, and security,
J. Dan, Y. Liu, B. Sun, J. Deng, and S. Luo, “Transface++: Re- thinking the face recognition paradigm with a focus on accuracy, efficiency, and security,” IEEE Trans. Pattern Anal. Mach. Intell. , 2025
2025
-
[35]
Mobilefaceformer: a lightweight face recognition model against face variations,
J. Li, L. Zhou, and J. Chen, “Mobilefaceformer: a lightweight face recognition model against face variations,” Multimed. Tools Appl., vol. 83, no. 5, pp. 12669–12685, 2024
2024
-
[36]
Effi- cientvit: Memory efficient vision transformer with cascaded group attention,
X. Liu, H. Peng, N. Zheng, Y. Yang, H. Hu, and Y. Yuan, “Effi- cientvit: Memory efficient vision transformer with cascaded group attention,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 14420–14430, 2023
2023
-
[37]
Edgevits: Competing light-weight cnns on mobile devices with vision transformers,
J. Pan, A. Bulat, F. Tan, X. Zhu, L. Dudziak, H. Li, G. Tzimiropou- los, and B. Martinez, “Edgevits: Competing light-weight cnns on mobile devices with vision transformers,” in Eur. Conf. Comput. Vis. (ECCV), pp. 294–311, Springer, 2022
2022
-
[38]
Resmlp: Feedforward networks for image classification with data-efficient training,
H. Touvron, P . Bojanowski, M. Caron, M. Cord, A. El-Nouby, E. Grave, G. Izacard, A. Joulin, G. Synnaeve, J. Verbeek, et al. , “Resmlp: Feedforward networks for image classification with data-efficient training,” IEEE Trans. Pattern Anal. Mach. Intell. , vol. 45, no. 4, pp. 5314–5321, 2022
2022
-
[39]
Pocketnet: Extreme lightweight face recognition net- work using neural architecture search and multistep knowledge distillation,
F. Boutros, P . Siebke, M. Klemt, N. Damer, F. Kirchbuchner, and A. Kuijper, “Pocketnet: Extreme lightweight face recognition net- work using neural architecture search and multistep knowledge distillation,” IEEE Access, vol. 10, pp. 46823–46833, 2022. IEEE TRANSACTIONS ON BIOMETRICS, BEHAVIOR, AND IDENTITY SCIENCE 14
2022
-
[40]
Partial fc: Training 10 million identities on a sin- gle machine,
X. An et al. , “Partial fc: Training 10 million identities on a sin- gle machine,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV) , pp. 1445–1449, 2021
2021
-
[41]
Low-resolution face recognition,
Z. Cheng, X. Zhu, and S. Gong, “Low-resolution face recognition,” in Asian Conf. Comput. Vis. (ACCV), pp. 605–621, Springer, 2018
2018
-
[42]
Cross-Age LFW: A Database for Studying Cross-Age Face Recognition in Unconstrained Environments
T. Zheng, W. Deng, and J. Hu, “Cross-age lfw: A database for studying cross-age face recognition in unconstrained environ- ments,” arXiv preprint arXiv:1708.08197, 2017
work page Pith review arXiv 2017
-
[43]
Cross-pose lfw: A database for study- ing cross-pose face recognition in unconstrained environments,
T. Zheng and W. Deng, “Cross-pose lfw: A database for study- ing cross-pose face recognition in unconstrained environments,” Beijing Univ. Posts Telecommun., Tech. Rep, vol. 5, no. 7, p. 5, 2018
2018
-
[44]
Frontal to profile face verification in the wild,
S. Sengupta, J.-C. Chen, C. Castillo, V . M. Patel, R. Chellappa, and D. W. Jacobs, “Frontal to profile face verification in the wild,” in Proc. IEEE Winter Conf. Appl. Comput. Vis. (WACV) , pp. 1–9, IEEE, 2016
2016
-
[45]
Iarpa janus benchmark-b face dataset,
C. Whitelam et al., “Iarpa janus benchmark-b face dataset,” inProc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 90–98, 2017
2017
-
[46]
Iarpa janus benchmark-c: Face dataset and protocol,
B. Maze, J. Adams, J. A. Duncan, N. Kalka, T. Miller, C. Otto, A. K. Jain, W. T. Niggel, J. Anderson, J. Cheney, et al., “Iarpa janus benchmark-c: Face dataset and protocol,” in Int. Conf. Biometrics (ICB), pp. 158–165, IEEE, 2018
2018
-
[47]
Transport-based single frame super resolution of very low resolution face images,
S. Kolouri and G. K. Rohde, “Transport-based single frame super resolution of very low resolution face images,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 4876–4884, 2015. Novendra Setyawan (Student Member, IEEE) received the B.Eng. degree in Electrical Engi- neering from the University of Muhammadiyah Malang, Indonesia, in 2015, a...
2015
-
[48]
He and his students received the Silver Medal of 2019 National College Software Creation Competition, the Silver Medal of 2018 Na- tional Microcomputer Competition, the Best Paper awards of Information Technology and Applications in Outlying Islands Conference, in 2013, 2014, 2016, 2017, 2018, 2021, and 2022, respectively, the Best Paper Award of Tanet 20...
2019
-
[49]
neuralnetwork
on Glint360K [40]. The AdamW optimizer and a polyno- mial decay learning rate schedule are utilized. The starting learning rate is 6 × 10−3 with a minimum of 1 × 10−5. The batch size is 342 on each 3× NVIDIA RTX-A6000 GPU, and the weight decay is 1 × 10−4. The model is trained for 50 epochs with a pre-processing resolution of (112 × 112). TABLE 14: FaceLi...
-
[50]
ACCURACY GAP FaceLiVTv2-L demonstrates the most compelling accuracy- efficiency trade-off among the evaluated lightweight mod- els
— alongside the FaceLiVTv2 variants. ACCURACY GAP FaceLiVTv2-L demonstrates the most compelling accuracy- efficiency trade-off among the evaluated lightweight mod- els. Against ResNet200-TopoFR, FaceLiVTv2-L narrows the mean accuracy gap to only 0.69% while requiring 13.9× fewer parameters and 76.1× fewer FLOPs. Remark- ably, FaceLiVTv2-L achieves 96.59% ...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.