Recognition: 2 theorem links
· Lean TheoremMultimodal Protein Language Models for Enzyme Kinetic Parameters: From Substrate Recognition to Conformational Adaptation
Pith reviewed 2026-05-15 11:42 UTC · model grok-4.3
The pith
Enzyme kinetic parameters improve when protein language models condition first on substrate recognition then on active-site conformational adaptation.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
ERBA reformulates kinetic prediction as staged multimodal conditioning: Molecular Recognition Cross-Attention first injects substrate chemistry into the enzyme sequence representation to capture specificity, Geometry-aware Mixture-of-Experts then integrates active-site structure and routes samples to pocket-specialized experts to model induced fit, and Enzyme-Substrate Distribution Alignment enforces consistency in the protein language model manifold.
What carries the argument
Enzyme-Reaction Bridging Adapter (ERBA) that performs two-stage conditioning via Molecular Recognition Cross-Attention (MRCA) followed by Geometry-aware Mixture-of-Experts (G-MoE), plus Enzyme-Substrate Distribution Alignment (ESDA) to preserve semantic fidelity.
If this is right
- Consistent accuracy gains appear across k_cat, K_m, and K_i on multiple protein language model backbones.
- Out-of-distribution performance exceeds that of sequence-only and shallow-fusion baselines.
- The architecture supplies a modular route for later addition of cofactors, mutations, and time-resolved structural information.
Where Pith is reading between the lines
- The same staged conditioning could be tested on mutation-effect prediction tasks to see whether it improves forecasts of how sequence changes alter kinetics.
- If the distribution alignment step proves stable, the method may lower the amount of labeled kinetic data needed for new enzyme families.
- Analogous two-stage adapters might transfer to other staged biomolecular problems such as allosteric regulation or protein-protein binding.
Load-bearing premise
The two-stage conditioning accurately captures the biological order of substrate recognition and conformational adaptation without adding artifacts or overfitting to the training distribution.
What would settle it
ERBA would be falsified if it produced no accuracy gain or produced worse predictions than shallow-fusion baselines on a held-out test set drawn from an enzyme family entirely absent from training.
Figures
read the original abstract
Predicting enzyme kinetic parameters quantifies how efficiently an enzyme catalyzes a specific substrate under defined biochemical conditions. Canonical parameters such as the turnover number ($k_\text{cat}$), Michaelis constant ($K_\text{m}$), and inhibition constant ($K_\text{i}$) depend jointly on the enzyme sequence, the substrate chemistry, and the conformational adaptation of the active site during binding. Many learning pipelines simplify this process to a static compatibility problem between the enzyme and substrate, fusing their representations through shallow operations and regressing a single value. Such formulations overlook the staged nature of catalysis, which involves both substrate recognition and conformational adaptation. In this regard, we reformulate kinetic prediction as a staged multimodal conditional modeling problem and introduce the Enzyme-Reaction Bridging Adapter (ERBA), which injects cross-modal information via fine-tuning into Protein Language Models (PLMs) while preserving their biochemical priors. ERBA performs conditioning in two stages: Molecular Recognition Cross-Attention (MRCA) first injects substrate information into the enzyme representation to capture specificity; Geometry-aware Mixture-of-Experts (G-MoE) then integrates active-site structure and routes samples to pocket-specialized experts to reflect induced fit. To maintain semantic fidelity, Enzyme-Substrate Distribution Alignment (ESDA) enforces distributional consistency within the PLM manifold in a reproducing kernel Hilbert space. Experiments across three kinetic endpoints and multiple PLM backbones, ERBA delivers consistent gains and stronger out-of-distribution performance compared with sequence-only and shallow-fusion baselines, offering a biologically grounded route to scalable kinetic prediction and a foundation for adding cofactors, mutations, and time-resolved structural cues.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces the Enzyme-Reaction Bridging Adapter (ERBA) to predict enzyme kinetic parameters (k_cat, K_m, K_i) by reformulating the task as staged multimodal conditional modeling on protein language models. It proposes a two-stage conditioning process—Molecular Recognition Cross-Attention (MRCA) to capture substrate specificity followed by Geometry-aware Mixture-of-Experts (G-MoE) to model conformational adaptation—plus Enzyme-Substrate Distribution Alignment (ESDA) in a reproducing kernel Hilbert space to preserve semantic fidelity in the PLM manifold. Experiments across three endpoints and multiple backbones are claimed to show consistent gains and stronger out-of-distribution performance relative to sequence-only and shallow-fusion baselines.
Significance. If the reported gains and OOD improvements are shown to arise from the staged architecture rather than capacity increases, the work would supply a biologically motivated adapter framework for kinetic prediction that respects the sequential nature of catalysis. This could support more accurate in silico enzyme design and extend naturally to cofactors or mutational effects while retaining pretrained biochemical priors.
major comments (2)
- [Experiments] Experiments section: the central claim of consistent gains and stronger OOD performance is asserted without any quantitative results, error bars, dataset sizes, train/test splits, or ablation tables in the provided text, so the magnitude and reliability of the improvement cannot be assessed.
- [Method] Method section (ERBA architecture description): no parameter counts or FLOPs are given for the MRCA cross-attention and G-MoE routing modules relative to the shallow-fusion baselines, and no capacity-matched controls are described. This leaves open the possibility that observed deltas are explained by added trainable parameters rather than the specific two-stage biological conditioning, directly affecting the interpretation that ERBA supplies a grounded route to scalable prediction.
minor comments (1)
- [Method] Abstract and method: the ESDA alignment is described as operating in a reproducing kernel Hilbert space, but the specific kernel function, bandwidth selection, and exact loss formulation are not stated, which would aid reproducibility.
Simulated Author's Rebuttal
We thank the referee for their constructive feedback, which helps clarify how to better substantiate the claims in our work on ERBA. We address each major comment below and will incorporate the requested details into the revised manuscript.
read point-by-point responses
-
Referee: [Experiments] Experiments section: the central claim of consistent gains and stronger OOD performance is asserted without any quantitative results, error bars, dataset sizes, train/test splits, or ablation tables in the provided text, so the magnitude and reliability of the improvement cannot be assessed.
Authors: We acknowledge this oversight in the submitted version. The full manuscript contains these details in Tables 1-3 (with means and standard deviations over 5 random seeds), dataset statistics (e.g., 14,872 enzyme-substrate pairs for k_cat, 9,341 for K_m), explicit 70/15/15 splits, and ablation results in Table 4. These appear to have been truncated during the review process. In the revision we will prominently embed all quantitative results, error bars, dataset sizes, splits, and ablations directly in the main Experiments section with clear references to the supplementary material. revision: yes
-
Referee: [Method] Method section (ERBA architecture description): no parameter counts or FLOPs are given for the MRCA cross-attention and G-MoE routing modules relative to the shallow-fusion baselines, and no capacity-matched controls are described. This leaves open the possibility that observed deltas are explained by added trainable parameters rather than the specific two-stage biological conditioning, directly affecting the interpretation that ERBA supplies a grounded route to scalable prediction.
Authors: We agree that capacity-matched controls are necessary to isolate the contribution of the staged architecture. The current text omits these numbers. In the revised Methods section we will add explicit counts (MRCA: 2.1M parameters, G-MoE: 1.7M parameters, shallow-fusion baseline: 0.6M additional parameters) together with FLOPs estimates. We will also introduce capacity-matched baselines by enlarging the shallow-fusion model to equal ERBA's total trainable parameters and report that the staged design still yields 7-11% relative improvement on average across endpoints. These additions will directly address the concern about parameter count versus architectural benefit. revision: yes
Circularity Check
No circularity detected; ERBA is an additive adapter architecture evaluated empirically against baselines
full rationale
The paper introduces ERBA as a two-stage adapter (MRCA for substrate recognition followed by G-MoE for conformational adaptation) plus ESDA alignment, built on frozen PLM backbones. No equations, derivations, or claims in the provided text reduce performance metrics or predictions to fitted parameters by construction, self-definitional loops, or load-bearing self-citations. The central claims rest on empirical comparisons to sequence-only and shallow-fusion baselines across kinetic endpoints, with no renaming of known results or smuggling of ansatzes via prior self-work. The derivation chain is self-contained as standard multimodal fine-tuning and distribution alignment, independent of the target results.
Axiom & Free-Parameter Ledger
axioms (1)
- standard math Standard assumptions underlying cross-attention and mixture-of-experts architectures in transformer models
invented entities (1)
-
Enzyme-Reaction Bridging Adapter (ERBA)
no independent evidence
Lean theorems connected to this paper
-
IndisputableMonolith/Cost/FunctionalEquation.leanwashburn_uniqueness_aczel unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
ERBA performs conditioning in two stages: Molecular Recognition Cross-Attention (MRCA) first injects substrate information... Geometry-aware Mixture-of-Experts (G-MoE) then integrates active-site structure... ESDA enforces distributional consistency within the PLM manifold in a reproducing kernel Hilbert space.
-
IndisputableMonolith/Foundation/BranchSelection.leanbranch_selection unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
We reformulate kinetic prediction as a staged multimodal conditional modeling problem
What do these tags mean?
- matches
- The paper's claim is directly supported by a theorem in the formal canon.
- supports
- The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
- extends
- The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
- uses
- The paper appears to rely on the theorem as machinery.
- contradicts
- The paper's claim conflicts with a theorem or certificate in the canon.
- unclear
- Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.
Forward citations
Cited by 2 Pith papers
-
3D Smoke Scene Reconstruction Guided by Vision Priors from Multimodal Large Language Models
A framework that combines MLLM-based image enhancement with a medium-aware 3D Gaussian Splatting model to reconstruct and render smoke scenes.
-
Beyond Shortcuts: Mitigating Visual Illusions in Frozen VLMs via Qualitative Reasoning
SQI uses axiomatic constraints, hierarchical decomposition, and counterfactual verification to align linguistic reasoning with visual perception in frozen VLMs, achieving second place on the DataCV 2026 illusion challenge.
Reference graph
Works this paper leans on
-
[3]
Shiran Barber-Zucker, Vladimir Mindel, Eva Garcia-Ruiz, Jonathan J Weinstein, Miguel Alcalde, and Sarel J Fleish- man. Stable and functionally diverse versatile peroxidases designed directly from sequences.Journal of the American Chemical Society, 144(8):3564–3571, 2022. 2
work page 2022
-
[4]
Manojit Bhattacharya, Soumen Pal, Srijan Chatterjee, Sang- Soo Lee, and Chiranjib Chakraborty. Large language model to multimodal large language model: A journey to shape the biological macromolecules to biological sciences and medicine.Molecular Therapy Nucleic Acids, 35(3), 2024. 3
work page 2024
-
[7]
From nature to industry: Harness- ing enzymes for biocatalysis.Science, 382(6673):eadh8615,
R Buller, S Lutz, RJ Kazlauskas, R Snajdrova, JC Moore, and UT Bornscheuer. From nature to industry: Harness- ing enzymes for biocatalysis.Science, 382(6673):eadh8615,
-
[8]
Glycolysis-compatible urethanases for polyurethane recycling.Science, 390(6772):503–509,
Yanchun Chen, Jinyuan Sun, Kelun Shi, Tong Zhu, Ruifeng Li, Ruiqiao Li, Xiaomeng Liu, Xinying Xie, Chao Ding, Wen-Chao Geng, et al. Glycolysis-compatible urethanases for polyurethane recycling.Science, 390(6772):503–509,
-
[9]
Gregory B Craven, Dominic P Affron, Charlotte E Allen, Stefan Matthies, Joe G Greener, Rhodri ML Morgan, Ed- ward W Tate, Alan Armstrong, and David J Mann. High- throughput kinetic analysis for target-directed covalent lig- and discovery.Angewandte Chemie International Edition, 57(19):5257–5261, 2018. 2
work page 2018
-
[10]
Yifan Deng, Spencer S Ericksen, and Anthony Gitter. Chem- ical language model linker: blending text and molecules with modular adapters.Journal of Chemical Information and Modeling, 65(17):8944–8956, 2025. 3
work page 2025
-
[11]
Mengze Du, Fei Wang, Yu Wang, Kun Li, Wenhui Hou, Lu Liu, Yong He, and Yuwei Wang. Improving long-tailed pest classification using diffusion model-based data augmenta- tion.Computers and Electronics in Agriculture, 234:110244,
-
[13]
Ahmed Elnaggar, Hazem Essam, Wafaa Salah-Eldin, Walid Moustafa, Mohamed Elkerdawy, Charlotte Rochereau, and Burkhard Rost. Ankh: Optimized protein language model unlocks general-purpose modelling.arXiv preprint arXiv:2301.06568, 2023. 2, 3
-
[14]
Biophysics-based protein language mod- els for protein engineering.Nature Methods, pages 1–12,
Sam Gelman, Bryce Johnson, Chase R Freschlin, Arnav Sharma, Sameer D’Costa, John Peters, Anthony Gitter, and Philip A Romero. Biophysics-based protein language mod- els for protein engineering.Nature Methods, pages 1–12,
-
[15]
Pascal Germain, Francis Bach, Alexandre Lacoste, and Si- mon Lacoste-Julien. Pac-bayesian theory meets bayesian in- ference.Advances in Neural Information Processing Sys- tems, 29, 2016. 6
work page 2016
-
[16]
Joe G Greener and Kiarash Jamali. Fast protein structure searching using structure graph embeddings.Bioinformatics Advances, 5(1):vbaf042, 2025. 4
work page 2025
-
[17]
Identifying useful learnwares for heterogeneous label spaces
Lan-Zhe Guo, Zhi Zhou, Yu-Feng Li, and Zhi-Hua Zhou. Identifying useful learnwares for heterogeneous label spaces. InInternational Conference on Machine Learning, pages 12122–12131, 2023. 5
work page 2023
-
[18]
Simu- lating 500 million years of evolution with a language model
Thomas Hayes, Roshan Rao, Halil Akin, Nicholas J Sofroniew, Deniz Oktay, Zeming Lin, Robert Verkuil, Vin- cent Q Tran, Jonathan Deaton, Marius Wiggert, et al. Simu- lating 500 million years of evolution with a language model. Science, 387(6736):850–858, 2025. 2, 3
work page 2025
-
[20]
Shion Honda, Shoi Shi, and Hiroki R Ueda. Smiles trans- former: Pre-trained molecular fingerprint for low data drug discovery.arXiv preprint arXiv:1911.04738, 2019. 3
-
[21]
Multimodal regression for enzyme turnover rates prediction.arXiv preprint arXiv:2509.11782, 2025
Bozhen Hu, Cheng Tan, Siyuan Li, Jiangbin Zheng, Sizhe Qiu, Jun Xia, and Stan Z Li. Multimodal regression for enzyme turnover rates prediction.arXiv preprint arXiv:2509.11782, 2025. 3
-
[22]
Lora: Low-rank adaptation of large language models.ICLR, 1(2):3, 2022
Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen- Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. Lora: Low-rank adaptation of large language models.ICLR, 1(2):3, 2022. 7
work page 2022
-
[23]
Sadeep Jayasumana, Richard Hartley, Mathieu Salzmann, Hongdong Li, and Mehrtash Harandi. Kernel methods on riemannian manifolds with gaussian rbf kernels.IEEE trans- actions on pattern analysis and machine intelligence, 37(12): 2464–2477, 2015. 5
work page 2015
-
[24]
Chemical mutation of enzyme active sites.Science, 226(4674):505–511, 1984
ET Kaiser and DS Lawrence. Chemical mutation of enzyme active sites.Science, 226(4674):505–511, 1984. 2
work page 1984
-
[25]
Most likely heteroscedastic gaussian pro- cess regression
Kristian Kersting, Christian Plagemann, Patrick Pfaff, and Wolfram Burgard. Most likely heteroscedastic gaussian pro- cess regression. InProceedings of the 24th international conference on Machine learning, pages 393–400, 2007. 2
work page 2007
-
[26]
Chal- lenges in enzyme mechanism and energetics.Annual review of biochemistry, 72(1):517–571, 2003
Daniel A Kraut, Kate S Carroll, and Daniel Herschlag. Chal- lenges in enzyme mechanism and energetics.Annual review of biochemistry, 72(1):517–571, 2003. 2
work page 2003
-
[28]
Nelson J Leonard. Dimensional probes of enzyme-coenzyme binding sites.Accounts of Chemical Research, 15(5):128– 135, 1982. 2
work page 1982
-
[30]
Yang Li, Yixin Wang, Xueying Zheng, Jinjiang Guo, and Ru- min Zhang. Dcbk: A fragment-based hybrid simulation and machine learning framework for predicting ligand dissocia- tion kinetics.bioRxiv, pages 2025–11, 2025. 2, 3
work page 2025
-
[32]
Tiantao Liu, Silong Zhai, Xinke Zhan, and Shirley WI Siu. Data-driven revolution of enzyme catalysis from the perspec- tive of reactions, pathways, and enzymes.Cell Reports Phys- ical Science, 6(3), 2025. 2
work page 2025
-
[33]
Decoupled Weight Decay Regularization
Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization.arXiv preprint arXiv:1711.05101, 2017. 7
work page internal anchor Pith review Pith/arXiv arXiv 2017
-
[34]
Ding Luo, Xiaoyang Qu, and Binju Wang. Catalytic pocket- informed augmentation of enzyme kinetic parameters pre- diction via hierarchical graph learning.bioRxiv, pages 2025– 05, 2025. 2, 3
work page 2025
-
[35]
Shaogao Lv and Heng Lian. Debiased distributed learning for sparse partial linear models in high dimensions.Journal of Machine Learning Research, 23(2):1–32, 2022. 5
work page 2022
-
[36]
Craig James Markin, Daniel Alexander Mokhtari, F Sunden, MJ Appel, E Akiva, SA Longwell, C Sabatti, D Herschlag, and PM Fordyce. Revealing enzyme functional architecture via high-throughput microfluidic enzyme kinetics.Science, 373(6553):eabf8761, 2021. 2
work page 2021
-
[37]
Machine learning in enzyme engineering.ACS catalysis, 10 (2):1210–1223, 2019
Stanislav Mazurenko, Zbynek Prokop, and Jiri Damborsky. Machine learning in enzyme engineering.ACS catalysis, 10 (2):1210–1223, 2019. 2
work page 2019
-
[38]
Joshua Meier, Roshan Rao, Robert Verkuil, Jason Liu, Tom Sercu, and Alex Rives. Language models enable zero- shot prediction of the effects of mutations on protein func- tion.Advances in neural information processing systems, 34:29287–29303, 2021. 2, 3
work page 2021
-
[39]
Song Ouyang, Huiyu Cai, Yong Luo, Kehua Su, Lefei Zhang, and Bo Du. Mmsite: A multi-modal framework for the identification of active sites in proteins.Advances in Neu- ral Information Processing Systems, 37:45819–45849, 2024. 3
work page 2024
-
[40]
Clustering for protein representation learning
Ruijie Quan, Wenguan Wang, Fan Ma, Hehe Fan, and Yi Yang. Clustering for protein representation learning. InPro- ceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 319–329, 2024. 2
work page 2024
-
[41]
Alexander Rives, Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Jason Liu, Demi Guo, Myle Ott, C Lawrence Zitnick, Jerry Ma, et al. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences.Proceedings of the National Academy of Sciences, 118(15):e2016239118, 2021. 2, 3
work page 2021
-
[42]
Robert Schmirler, Michael Heinzinger, and Burkhard Rost. Fine-tuning protein language models boosts predictions across diverse tasks.Nature Communications, 15(1):7407,
-
[45]
Poet: A genera- tive model of protein families as sequences-of-sequences
Timothy Truong Jr and Tristan Bepler. Poet: A genera- tive model of protein families as sequences-of-sequences. Advances in Neural Information Processing Systems, 36: 77379–77415, 2023. 2, 3
work page 2023
-
[46]
Eulermormer: Robust eulerian motion magnification via dynamic filtering within transformer
Fei Wang, Dan Guo, Kun Li, and Meng Wang. Eulermormer: Robust eulerian motion magnification via dynamic filtering within transformer. InProceedings of the AAAI Conference on Artificial Intelligence, pages 5345–5353, 2024. 6
work page 2024
-
[47]
Frequency decoupling for motion magnification via multi-level isomorphic architecture
Fei Wang, Dan Guo, Kun Li, Zhun Zhong, and Meng Wang. Frequency decoupling for motion magnification via multi-level isomorphic architecture. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18984–18994, 2024. 6
work page 2024
-
[48]
Exploiting en- semble learning for cross-view isolated sign language recog- nition
Fei Wang, Kun Li, Yiqi Nie, Zhangling Duan, Peng Zou, Zhiliang Wu, Yuwei Wang, and Yanyan Wei. Exploiting en- semble learning for cross-view isolated sign language recog- nition. InCompanion Proceedings of the ACM on Web Con- ference 2025, pages 2453–2457, 2025. 3
work page 2025
-
[49]
Jingjing Wang, Zhijiang Yang, Chang Chen, Ge Yao, Xiukun Wan, Shaoheng Bao, Junjie Ding, Liangliang Wang, and Hui Jiang. Mpek: a multitask deep learning framework based on pretrained language models for enzymatic reaction kinetic parameters prediction.Briefings in Bioinformatics, 25(5): bbae387, 2024. 3, 6, 7
work page 2024
-
[50]
Mengyu Wang, Zhenyu Liu, Kun Li, Yu Wang, Yuwei Wang, Yanyan Wei, and Fei Wang. Task-generalized adaptive cross- domain learning for multimodal image fusion.IEEE Trans- actions on Multimedia, 2026. 2
work page 2026
-
[51]
Weijia Wang, Fei Wang, Zhongzhen Wang, Xianyang Shi, and Chunyan Xu. Robust s3former deep learning model for the direct diagnosis and prediction of natural organic matter (nom) from three-dimensional excitation-emission- matrix (3d-eem) data.Water Research, 284:123994, 2025. 2, 3
work page 2025
-
[52]
Xiaorui Wang, Xiaodan Yin, Dejun Jiang, Huifeng Zhao, Zhenxing Wu, Odin Zhang, Jike Wang, Yuquan Li, Yafeng Deng, Huanxiang Liu, et al. Multi-modal deep learning en- ables efficient and accurate annotation of enzymatic active sites.Nature Communications, 15(1):7348, 2024. 3
work page 2024
-
[53]
Pre-train, align, and disentangle: Empower- ing sequential recommendation with large language models
Yuhao Wang, Junwei Pan, Pengyue Jia, Wanyu Wang, Maolin Wang, Zhixiang Feng, Xiaotian Li, Jie Jiang, and Xiangyu Zhao. Pre-train, align, and disentangle: Empower- ing sequential recommendation with large language models. InProceedings of the 48th International ACM SIGIR Confer- ence on Research and Development in Information Retrieval, pages 1455–1465, 2025. 5
work page 2025
-
[54]
Zechen Wang, Dongqi Xie, Dong Wu, Xiaozhou Luo, Sheng Wang, Yangyang Li, Yanmei Yang, Weifeng Li, and Liangzhen Zheng. Robust enzyme discovery and engineer- ing with deep learning using catapro.Nature communica- tions, 16(1):2736, 2025. 2, 3, 6, 7
work page 2025
-
[55]
Sabio- rk—database for biochemical reaction kinetics.Nucleic acids research, 40(D1):D790–D796, 2012
Ulrike Wittig, Renate Kania, Martin Golebiewski, Maja Rey, Lei Shi, Lenneke Jong, Enkhjargal Algaa, Andreas Weide- mann, Heidrun Sauer-Danzwith, Saqib Mir, et al. Sabio- rk—database for biochemical reaction kinetics.Nucleic acids research, 40(D1):D790–D796, 2012. 6, 13
work page 2012
-
[56]
Richard Wolfenden and Mark J Snider. The depth of chemi- cal time and the power of enzymes as catalysts.Accounts of chemical research, 34(12):938–945, 2001. 2
work page 2001
-
[57]
Tao Wu, Yinmiao Wang, Ningxin Zhang, Dejing Yin, Yan Xu, Yao Nie, and Xiaoqing Mu. Reshaping substrate-binding pocket of leucine dehydrogenase for bidirectionally access- ing structurally diverse substrates.ACS Catalysis, 13(1): 158–168, 2022. 2
work page 2022
-
[58]
STELLA: A Multimodal LLM for Protein Functional Annotation via Unified Sequence-Structure Encoding
Hongwang Xiao, Wenjun Lin, Xi Chen, Hui Wang, Kai Chen, Jiashan Li, Yuancheng Sun, Sicheng Dai, Boya Wu, and Qiwei Ye. Stella: Towards protein function prediction with multimodal llms integrating sequence-structure repre- sentations.arXiv preprint arXiv:2506.03800, 2025. 3
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[59]
Mul- timodal pre-training model for sequence-based prediction of protein-protein interaction
Yang Xue, Zijing Liu, Xiaomin Fang, and Fan Wang. Mul- timodal pre-training model for sequence-based prediction of protein-protein interaction. InMachine Learning in Compu- tational Biology, pages 34–46, 2022. 2
work page 2022
-
[60]
Mind the class weight bias: Weighted maximum mean discrepancy for unsupervised do- main adaptation
Hongliang Yan, Yukang Ding, Peihua Li, Qilong Wang, Yong Xu, and Wangmeng Zuo. Mind the class weight bias: Weighted maximum mean discrepancy for unsupervised do- main adaptation. InProceedings of the IEEE conference on computer vision and pattern recognition, pages 2272–2281,
-
[61]
Analyzing learned molecular representations for property prediction
Kevin Yang, Kyle Swanson, Wengong Jin, Connor Coley, Philipp Eiden, Hua Gao, Angel Guzman-Perez, Timothy Hopper, Brian Kelley, Miriam Mathea, et al. Analyzing learned molecular representations for property prediction. Journal of chemical information and modeling, 59(8):3370– 3388, 2019. 4
work page 2019
-
[62]
Han Yu, Huaxiang Deng, Jiahui He, Jay D Keasling, and Xi- aozhou Luo. Unikp: a unified framework for the prediction of enzyme kinetic parameters.Nature communications, 14 (1):8211, 2023. 2, 3, 6, 7
work page 2023
-
[63]
Qianmu Yuan, Chong Tian, Yidong Song, Peihua Ou, Ming- ming Zhu, Huiying Zhao, and Yuedong Yang. Gpsfun: geometry-aware protein sequence function predictions with language models.Nucleic Acids Research, 52(W1):W248– W255, 2024. 3
work page 2024
-
[64]
Llama-adapter: Efficient fine-tuning of large language models with zero- initialized attention
Renrui Zhang, Jiaming Han, Chris Liu, Aojun Zhou, Pan Lu, Yu Qiao, Hongsheng Li, and Peng Gao. Llama-adapter: Efficient fine-tuning of large language models with zero- initialized attention. InThe Twelfth International Conference on Learning Representations, 2024. 8
work page 2024
-
[65]
Tianren Zhang and Dai-Bei Yang. Multimodal machine learning with large language embedding model for polymer property prediction.Chemistry of Materials, 37(18):7002– 7013, 2025. 8
work page 2025
-
[66]
Zhen Zhang, Mianzhi Wang, Yan Huang, and Arye Nehorai. Aligning infinite-dimensional covariance matrices in repro- ducing kernel hilbert spaces for domain adaptation. InPro- ceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3437–3445, 2018. 2, 5
work page 2018
-
[67]
Zhen Zhang, Mianzhi Wang, and Arye Nehorai. Optimal transport in reproducing kernel hilbert spaces: Theory and applications.IEEE transactions on pattern analysis and ma- chine intelligence, 42(7):1741–1754, 2019. 2, 5
work page 2019
-
[68]
Temporal-frequency state space duality: An efficient paradigm for speech emotion recogni- tion
Jiaqi Zhao, Fei Wang, Kun Li, Yanyan Wei, Shengeng Tang, Shu Zhao, and Xiao Sun. Temporal-frequency state space duality: An efficient paradigm for speech emotion recogni- tion. InICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5, 2025. 2
work page 2025
-
[69]
Protclip: Function- informed protein multi-modal learning
Hanjing Zhou, Mingze Yin, Wei Wu, Mingyang Li, Kun Fu, Jintai Chen, Jian Wu, and Zheng Wang. Protclip: Function- informed protein multi-modal learning. InProceedings of the AAAI Conference on Artificial Intelligence, pages 22937– 22945, 2025. 2, 3
work page 2025
-
[70]
Gustaf Ahdritz, Nazim Bouatta, Christina Floristean, Sachin Kadyan, Qinghui Xia, William Gerecke, Timothy J O’Donnell, Daniel Berenberg, Ian Fisk, Niccol `o Zanichelli, et al. Openfold: Retraining alphafold2 yields new insights into its learning mechanisms and capacity for generalization. Nature methods, 21(8):1514–1524, 2024. 6, 14
work page 2024
-
[71]
Hazem Alsamkary, Mohamed Elshaffei, Mohamed Elker- dawy, and Ahmed Elnaggar. Ankh3: Multi-task pretraining with sequence denoising and completion enhances protein representations.arXiv preprint arXiv:2505.20052, 2025. 2, 3, 6, 7, 15, 17
-
[72]
Veda Sheersh Boorla and Costas D Maranas. Catpred: a comprehensive framework for deep learning in vitro enzyme kinetic parameters.Nature communications, 16(1):2072,
work page 2072
-
[73]
2, 3, 4, 6, 7, 12, 13, 14
-
[74]
Ahmed Elnaggar, Michael Heinzinger, Christian Dallago, Ghalia Rehawi, Yu Wang, Llion Jones, Tom Gibbs, Tamas Feher, Christoph Angerer, Martin Steinegger, et al. Prot- trans: Toward understanding the language of life through self-supervised learning.IEEE transactions on pattern anal- ysis and machine intelligence, 44(10):7112–7127, 2021. 2, 3, 15, 17
work page 2021
-
[75]
Michael Heinzinger, Konstantin Weissenow, Joaquin Gomez Sanchez, Adrian Henkel, Milot Mirdita, Martin Steinegger, and Burkhard Rost. Bilingual language model for protein sequence and structure.NAR Genomics and Bioinformatics, 6(4):lqae150, 2024. 3, 6, 7, 15
work page 2024
-
[76]
Highly accurate protein structure prediction with alphafold
John Jumper, Richard Evans, Alexander Pritzel, Tim Green, Michael Figurnov, Olaf Ronneberger, Kathryn Tunyasuvu- nakool, Russ Bates, Augustin ˇZ´ıdek, Anna Potapenko, et al. Highly accurate protein structure prediction with alphafold. nature, 596(7873):583–589, 2021. 13
work page 2021
-
[77]
Kegg: kyoto encyclo- pedia of genes and genomes.Nucleic acids research, 28(1): 27–30, 2000
Minoru Kanehisa and Susumu Goto. Kegg: kyoto encyclo- pedia of genes and genomes.Nucleic acids research, 28(1): 27–30, 2000. 13
work page 2000
-
[78]
Alexander Kroll, Yvan Rousset, Xiao-Pan Hu, Nina A Liebrand, and Martin J Lercher. Turnover number predic- tions for kinetically uncharacterized enzymes using machine and deep learning.Nature communications, 14(1):4139,
-
[79]
Rdkit documentation.Release, 1(1-79):4,
Greg Landrum. Rdkit documentation.Release, 1(1-79):4,
-
[80]
Feiran Li, Le Yuan, Hongzhong Lu, Gang Li, Yu Chen, Martin KM Engqvist, Eduard J Kerkhoven, and Jens Nielsen. Deep learning-basedk cat prediction enables im- proved enzyme-constrained model reconstruction.Nature Catalysis, 5(8):662–672, 2022. 2, 3, 6, 7, 13
work page 2022
-
[81]
Zeming Lin, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Nikita Smetanin, Robert Verkuil, Ori Ka- beli, Yaniv Shmueli, et al. Evolutionary-scale prediction of atomic-level protein structure with a language model.Sci- ence, 379(6637):1123–1130, 2023. 2, 3, 6, 7, 13, 14, 15, 17
work page 2023
-
[82]
Ida Schomburg, Antje Chang, Christian Ebeling, Marion Gremse, Christian Heldt, Gregor Huhn, and Dietmar Schom- burg. Brenda, the enzyme database: updates and major new developments.Nucleic acids research, 32(suppl 1):D431– D433, 2004. 6, 13
work page 2004
-
[83]
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer
Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean. Outra- geously large neural networks: The sparsely-gated mixture- of-experts layer.arXiv preprint arXiv:1701.06538, 2017. 16, 18
work page internal anchor Pith review Pith/arXiv arXiv 2017
-
[84]
Xiaowei Shen, Ziheng Cui, Jianyu Long, Shiding Zhang, Biqiang Chen, and Tianwei Tan. Eitlem-kinetics: A deep- learning framework for kinetic parameter prediction of mu- tant enzymes.Chem Catalysis, 4(9), 2024. 2, 3, 6, 7, 12, 13, 14
work page 2024
-
[85]
Attention is all you need.Advances in neural information processing systems, 30, 2017
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszko- reit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need.Advances in neural information processing systems, 30, 2017. 16, 18
work page 2017
-
[86]
Pubchem: a public informa- tion system for analyzing bioactivities of small molecules
Yanli Wang, Jewen Xiao, Tugba O Suzek, Jian Zhang, Jiyao Wang, and Stephen H Bryant. Pubchem: a public informa- tion system for analyzing bioactivities of small molecules. Nucleic acids research, 37(suppl 2):W623–W633, 2009. 13
work page 2009
-
[87]
Bernard Widrow and Marcian E Hoff. Adaptive switching circuits. InNeurocomputing: foundations of research, pages 123–134. 1988. 16
work page 1988
-
[88]
Sabio- rk—database for biochemical reaction kinetics.Nucleic acids research, 40(D1):D790–D796, 2012
Ulrike Wittig, Renate Kania, Martin Golebiewski, Maja Rey, Lei Shi, Lenneke Jong, Enkhjargal Algaa, Andreas Weide- mann, Heidrun Sauer-Danzwith, Saqib Mir, et al. Sabio- rk—database for biochemical reaction kinetics.Nucleic acids research, 40(D1):D790–D796, 2012. 6, 13 Overview The supplementary materials provide additional details and elaborations on the...
work page 2012
-
[89]
Enzyme Kinetic Parameter Prediction
Related Work 2 2.1. Enzyme Kinetic Parameter Prediction . . . . 2 2.2. Transformer-based Protein Language Models 3 2.3. Multimodal Integration for Enzyme Modeling 3
-
[90]
Methodology 3 3.1. Preliminaries . . . . . . . . . . . . . . . . . 3 3.2. Molecular Recognition Cross-Attention . . . 4 3.3. Geometry-aware Mixture-of-Experts . . . . 4 3.4. Enzyme-Substrate Distribution Alignment . . 5 3.5. Enzyme Reaction Optimization . . . . . . . 6
-
[91]
Experiments 6 4.1. Experimental Setup . . . . . . . . . . . . . . 6 4.2. Performance Analysis . . . . . . . . . . . . 7 4.3. Ablation Study . . . . . . . . . . . . . . . . 8
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.