| | """Continue-1-OSS Model Implementation""" |
| |
|
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaAttention as _BaseAttention |
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaDecoderLayer as _BaseDecoderLayer |
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaForCausalLM as _BaseModel |
| | from transformers.models.llama.modeling_llama import LlamaMLP as _BaseMLP |
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaModel as _BaseTransformer |
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaRMSNorm as _BaseRMSNorm |
| | from transformers.models.llama.modeling_llama import \ |
| | LlamaRotaryEmbedding as _BaseRotaryEmbedding |
| |
|
| | from .configuration_continue_oss import Continue1Config |
| |
|
| |
|
| | |
| | class Continue1RMSNorm(_BaseRMSNorm): |
| | """Continue-1-OSS Root Mean Square Layer Normalization""" |
| | pass |
| |
|
| |
|
| | class Continue1RotaryEmbedding(_BaseRotaryEmbedding): |
| | """Continue-1-OSS Rotary Position Embeddings""" |
| | pass |
| |
|
| |
|
| | class Continue1MLP(_BaseMLP): |
| | """Continue-1-OSS MLP (Feed-Forward Network)""" |
| | pass |
| |
|
| |
|
| | class Continue1Attention(_BaseAttention): |
| | """Continue-1-OSS Multi-Head Attention""" |
| | pass |
| |
|
| |
|
| | class Continue1DecoderLayer(_BaseDecoderLayer): |
| | """Continue-1-OSS Transformer Decoder Layer""" |
| | pass |
| |
|
| |
|
| | class Continue1Model(_BaseTransformer): |
| | """ |
| | Continue-1-OSS Transformer Model |
| | |
| | Core transformer model without the language modeling head. |
| | """ |
| | config_class = Continue1Config |
| | |
| | def __init__(self, config: Continue1Config): |
| | super().__init__(config) |
| |
|
| |
|
| | class Continue1ForCausalLM(_BaseModel): |
| | """ |
| | Continue-1-OSS Model for Causal Language Modeling |
| | |
| | Designed by SVECTOR Corporation for high-quality text generation, |
| | instruction following, and long-context understanding. |
| | |
| | Example: |
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | |
| | model = AutoModelForCausalLM.from_pretrained( |
| | "SVECTOR-CORPORATION/Continue-1-OSS", |
| | trust_remote_code=True |
| | ) |
| | tokenizer = AutoTokenizer.from_pretrained("SVECTOR-CORPORATION/Continue-1-OSS") |
| | |
| | messages = [{"role": "user", "content": "Hello There!"}] |
| | inputs = tokenizer.apply_chat_template(messages, return_tensors="pt") |
| | outputs = model.generate(inputs, max_new_tokens=100) |
| | ``` |
| | """ |
| | config_class = Continue1Config |
| | |
| | def __init__(self, config: Continue1Config): |
| | super().__init__(config) |
| |
|