LiteSeqCNN + MULTI-ATTENTION BASED FRAMEWORK FOR PROTEIN FUNCTION PREDICTION | IJCT Volume 13 – Issue 2 | IJCT-V13I2P16

International Journal of Computer Techniques
ISSN 2394-2231
Volume 13, Issue 2  |  Published: March – April 2026

Author

SHAMSUDDEEN MOHAMMED MAMMAN, KISLAY SINGH RAJPUT

Abstract

Protein function prediction is a vital task in computational biology, with applications in understanding biological processes, drug discovery, and disease research. Lightweight architectures like Lite-SeqCNN have demonstrated the ability to balance efficiency and accuracy for this task. However, these models lack mechanisms to focus dynamically on the most biologically informative sequence regions, which limits their predictive performance. In this study, we enhance Lite-SeqCNN by integrating an attention mechanism, allowing the model to identify and prioritize key functional motifs within protein sequences. Evaluated on the Data2017 dataset, the proposed attention-enhanced Lite-SeqCNN demonstrates significant improvements over the baseline model, achieving 0.669 precision, 0.532 F1-score for BP dataset and 0.809 precision, 0.654 F1-score for MF dataset. Additionally, the attention mechanism enhances interpretability, providing insights into the sequence regions most critical for function prediction. Our findings highlight the potential of lightweight architectures augmented with attention mechanisms to advance protein function prediction and uncover biologically meaningful patterns.

Keywords

^KEYWORDS^

Conclusion

This study establishes the significant benefits of integrating an attention mechanism with the Lite-SeqCNN architecture for protein function prediction, addressing key limitations of the original model. By leveraging attention, the enhanced model selectively focuses on biologically relevant regions within protein sequences, yielding substantial improvements in both predictive accuracy and interpretability. This advancement not only supports more precise annotations of protein functions but also provides insights into the underlying molecular mechanisms, fostering a deeper understanding of sequence-function relationships. The attention-enhanced Lite-SeqCNN demonstrates the importance of emphasizing functional motifs, which are often sparse but critical, enabling researchers to identify sequence regions that contribute most to the predictive outcomes. The interpretability offered by attention mechanisms bridges computational models and biological reasoning, making the predictions more transparent and actionable in bioinformatics applications. Despite the computational cost introduced by the attention mechanism, the improvements justify the trade-off, particularly when applied to large datasets such as Data2017. The approach sets a precedent for designing lightweight yet interpretable models in protein bioinformatics, paving the way for their use in high- throughput genomic and proteomic studies. Future directions include exploring advanced attention mechanisms, such as multi- head attention or scaled dot-product attention, to further enhance the model’s performance. Expanding the training dataset to include more diverse protein families could improve generalizability across various biological contexts. Additionally, applying the attention-enhanced Lite-SeqCNN to related bioinformatics challenges, such as predicting protein-protein interactions or identifying post-translational modification sites, could demonstrate the versatility and broader applicability of the model. Overall, this work provides a robust foundation for integrating attention mechanisms into deep learning frameworks for biological sequence analysis, addressing challenges posed by large-scale data while offering interpretable and accurate predictions that can drive future discoveries in the field.

References

1Ranjan, A., Tiwari, A. and Deepak, A., 2021. A sub-sequence based approach to protein function prediction via multi-attention based multi-aspect network. IEEE/ACM transactions on computational biology and bioinformatics, 20(1), pp.94-105. 2Ranjan, A., Fahad, M.S. and Deepak, A., 2022. λ-Scaled-attention: A novel fast attention mechanism for efficient modeling of protein sequences. Information Sciences, 609, pp.1098-1112. 3Kumar, V., Deepak, A., Ranjan, A. and Prakash, A., 2023. Lite-SeqCNN: A light-weight deep CNN architecture for protein function prediction. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 20(3), pp.2242-2253. 4Ranjan, A., Fahad, M.S., Fernández-Baca, D., Tripathi, S. and Deepak, A., 2022. MCWS-transformers: towards an efficient modeling of protein sequences via multi context-window based scaled self-attention. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 20(2), pp.1188-1199. 5Dhanuka, R., Singh, J.P. and Tripathi, A., 2023. A comprehensive survey of deep learning techniques in protein function prediction. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 20(3), pp.2291-2301. 6Wu, Z., Guo, M., Jin, X., Chen, J. and Liu, B., 2023. CFAGO: cross-fusion of network and attributes based on attention mechanism for protein function prediction. Bioinformatics, 39(3), p.btad123.

How to Cite This Paper

SHAMSUDDEEN MOHAMMED MAMMAN, KISLAY SINGH RAJPUT (2026). LiteSeqCNN + MULTI-ATTENTION BASED FRAMEWORK FOR PROTEIN FUNCTION PREDICTION. International Journal of Computer Techniques, 13(2). ISSN: 2394-2231.

© 2026 International Journal of Computer Techniques (IJCT). All rights reserved.

Submit Your Paper