| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - Novora/CodeClassifier_v1 |
| | pipeline_tag: text-classification |
| | --- |
| | |
| | # Introduction |
| |
|
| | Novora Code Classifier v1 Tiny, is a tiny `Text Classification` model, which classifies given code text input under 1 of `31` different classes (programming languages). |
| |
|
| | This model is designed to be able to run on CPU, but optimally runs on GPUs. |
| |
|
| | # Info |
| | - 1 of 31 classes output |
| | - 512 token input dimension |
| | - 64 hidden dimensions |
| | - 2 linear layers |
| | - The `snowflake-arctic-embed-xs` model is used as the embeddings model. |
| | - Dataset split into 80% training set, 20% testing set. |
| | - The combined test and training data is 100 chunks per programming language, the data is 3,100 chunks (entries) as 512 tokens per chunk, being a snippet of the code. |
| |
|
| | # Architecture |
| |
|
| | The `CodeClassifier-v1-Tiny` model employs a neural network architecture optimized for text classification tasks, specifically for classifying programming languages from code snippets. This model includes: |
| |
|
| | - **Bidirectional LSTM Feature Extractor**: This bidirectional LSTM layer processes input embeddings, effectively capturing contextual relationships in both forward and reverse directions within the code snippets. |
| |
|
| | - **Adaptive Pooling**: Following the LSTM, adaptive average pooling reduces the feature dimension to a fixed size, accommodating variable-length inputs. |
| |
|
| | - **Fully Connected Layers**: The network includes two linear layers. The first projects the pooled features into a hidden feature space, and the second linear layer maps these to the output classes, which correspond to different programming languages. A dropout layer with a rate of 0.5 between these layers helps mitigate overfitting. |
| |
|
| | The model's bidirectional nature and architectural components make it adept at understanding the syntax and structure crucial for code classification. |
| |
|
| |
|