JOURNAL ARTICLE

Compressing Pre-trained Models of Code into 3 MB

Abstract

Although large pre-trained models of code have delivered significant advancements in various code processing tasks, there is an impediment to the wide and fluent adoption of these powerful models in software developers' daily workflow: these large models consume hundreds of megabytes of memory and run slowly on personal devices, which causes problems in model deployment and greatly degrades the user experience.

Keywords:
Computer science Software deployment Workflow Code (set theory) Software Software engineering Programming language Database

Metrics

39
Cited By
8.14
FWCI (Field Weighted Citation Impact)
38
Refs
0.96
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Software System Performance and Reliability
Physical Sciences →  Computer Science →  Computer Networks and Communications
Parallel Computing and Optimization Techniques
Physical Sciences →  Computer Science →  Hardware and Architecture
Software Testing and Debugging Techniques
Physical Sciences →  Computer Science →  Software
© 2026 ScienceGate Book Chapters — All rights reserved.