| | --- |
| | library_name: transformers |
| | pipeline_tag: text-generation |
| | inference: true |
| | widget: |
| | - text: Hello! |
| | example_title: Hello world |
| | group: Python |
| | --- |
| | |
| | This tiny model is for debugging. It is randomly initialized with the config adapted from [microsoft/Phi-4-multimodal-instruct](https://huggingface.co/microsoft/Phi-4-multimodal-instruct). |
| |
|
| | ### Example usage: |
| |
|
| | ```python |
| | import io |
| | import os |
| | from urllib.request import urlopen |
| | |
| | import torch |
| | |
| | import requests |
| | import soundfile as sf |
| | from PIL import Image |
| | from transformers import AutoModelForCausalLM, AutoProcessor, GenerationConfig |
| | |
| | # Define model path |
| | model_id = "tiny-random/phi-4-multimodal" |
| | |
| | # Load model and processor |
| | processor = AutoProcessor.from_pretrained(model_id, trust_remote_code=True) |
| | model = AutoModelForCausalLM.from_pretrained( |
| | model_id, |
| | device_map="cuda", |
| | torch_dtype="auto", |
| | trust_remote_code=True, |
| | attn_implementation='flash_attention_2', |
| | ).cuda() |
| | |
| | # Load generation config |
| | generation_config = GenerationConfig.from_pretrained(model_id) |
| | |
| | # Define prompt structure |
| | user_prompt = '<|user|>' |
| | assistant_prompt = '<|assistant|>' |
| | prompt_suffix = '<|end|>' |
| | |
| | # Part 1: Image Processing |
| | print("\n--- IMAGE PROCESSING ---") |
| | image_url = 'https://www.ilankelman.org/stopsigns/australia.jpg' |
| | prompt = f'{user_prompt}<|image_1|>What is shown in this image?{prompt_suffix}{assistant_prompt}' |
| | print(f'>>> Prompt\n{prompt}') |
| | |
| | # Download and open image |
| | image = Image.open(requests.get(image_url, stream=True).raw) |
| | inputs = processor(text=prompt, images=image, return_tensors='pt').to('cuda:0') |
| | |
| | # Generate response |
| | generate_ids = model.generate( |
| | **inputs, |
| | max_new_tokens=8, |
| | generation_config=generation_config, |
| | ) |
| | generate_ids = generate_ids[:, inputs['input_ids'].shape[1]:] |
| | response = processor.batch_decode( |
| | generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False |
| | )[0] |
| | print(f'>>> Response\n{response}') |
| | |
| | # Part 2: Audio Processing |
| | print("\n--- AUDIO PROCESSING ---") |
| | audio_url = "https://upload.wikimedia.org/wikipedia/commons/b/b0/Barbara_Sahakian_BBC_Radio4_The_Life_Scientific_29_May_2012_b01j5j24.flac" |
| | speech_prompt = "Transcribe the audio to text, and then translate the audio to French. Use <sep> as a separator between the original transcript and the translation." |
| | prompt = f'{user_prompt}<|audio_1|>{speech_prompt}{prompt_suffix}{assistant_prompt}' |
| | print(f'>>> Prompt\n{prompt}') |
| | |
| | # Downlowd and open audio file |
| | audio, samplerate = sf.read(io.BytesIO(urlopen(audio_url).read())) |
| | |
| | # Process with the model |
| | inputs = processor(text=prompt, audios=[(audio, samplerate)], return_tensors='pt').to('cuda:0') |
| | |
| | generate_ids = model.generate( |
| | **inputs, |
| | max_new_tokens=8, |
| | generation_config=generation_config, |
| | ) |
| | generate_ids = generate_ids[:, inputs['input_ids'].shape[1]:] |
| | response = processor.batch_decode( |
| | generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False |
| | )[0] |
| | print(f'>>> Response\n{response}') |
| | ``` |
| |
|
| | ### Codes to create this repo: |
| |
|
| | ```python |
| | import json |
| | import shutil |
| | import sys |
| | from pathlib import Path |
| | |
| | import torch |
| | |
| | from huggingface_hub import hf_hub_download |
| | from transformers import ( |
| | AutoConfig, |
| | AutoModelForCausalLM, |
| | AutoProcessor, |
| | AutoTokenizer, |
| | GenerationConfig, |
| | pipeline, |
| | set_seed, |
| | ) |
| | |
| | source_model_id = "microsoft/Phi-4-multimodal-instruct" |
| | save_folder = "/tmp/tiny-random/phi-4-multimodal" |
| | Path(save_folder).mkdir(exist_ok=True) |
| | AutoTokenizer.from_pretrained(source_model_id).save_pretrained(save_folder) |
| | |
| | # preprocessor config |
| | for json_file in ['preprocessor_config.json', 'processor_config.json', 'config.json']: |
| | with open(hf_hub_download(source_model_id, json_file), 'r') as f: |
| | config = json.load(f) |
| | auto_map = config.get('auto_map', {}) |
| | for key, value in auto_map.items(): |
| | if '.' in value: |
| | auto_map[key] = f'{source_model_id}--{value}' |
| | with open(f'{save_folder}/{json_file}', 'w') as f: |
| | json.dump(config, f, indent=2) |
| | |
| | # model config |
| | with open(f'{save_folder}/config.json', 'r') as f: |
| | config = json.load(f) |
| | |
| | config['hidden_size'] = 16 |
| | config['intermediate_size'] = 32 |
| | config['num_attention_heads'] = 2 |
| | config['num_hidden_layers'] = 2 |
| | config['num_key_value_heads'] = 1 |
| | |
| | config['audio_processor']['config']['num_blocks'] = 2 |
| | config['audio_processor']['config']['attention_dim'] = 16 |
| | config['audio_processor']['config']['attention_heads'] = 2 |
| | config['audio_processor']['config']['nemo_conv_settings']['conv_channels'] = 16 |
| | config['audio_processor']['config']['depthwise_seperable_out_channel'] = 16 |
| | config['audio_processor']['config']['ext_pw_out_channel'] = 16 |
| | config['audio_processor']['config']['linear_units'] = 24 |
| | |
| | config['vision_lora']['r'] = 8 |
| | config['vision_lora']['lora_alpha'] = 16 |
| | config['speech_lora']['r'] = 8 |
| | config['speech_lora']['lora_alpha'] = 16 |
| | |
| | config['rope_scaling']['long_factor'] = [1.0] * 3 |
| | config['rope_scaling']['short_factor'] = [1.0] * 3 |
| | |
| | with open(f'{save_folder}/config.json', 'w') as f: |
| | json.dump(config, f, indent=2) |
| | |
| | config = AutoConfig.from_pretrained( |
| | save_folder, |
| | trust_remote_code=True, |
| | ) |
| | |
| | Path(save_folder, 'phi4mm').mkdir(exist_ok=True) |
| | for python_files in ['modeling_phi4mm.py', 'configuration_phi4mm.py', 'speech_conformer_encoder.py', 'vision_siglip_navit.py', 'processing_phi4mm.py']: |
| | with open(hf_hub_download(source_model_id, python_files), 'r') as f: |
| | codes = f.read() |
| | with open(f'{save_folder}/phi4mm/{python_files}', 'w') as f: |
| | f.write(codes) |
| | with open(Path(save_folder, 'phi4mm/vision_siglip_navit.py'), 'r') as f: |
| | codes = f.read() |
| | codes = codes.replace('def get_siglip_vision_model', '# modified for tiny-random\ndef get_siglip_vision_model') |
| | codes = codes.replace('"hidden_size": 1152,', '"hidden_size": 16,') |
| | codes = codes.replace('"intermediate_size": 4304,', '"intermediate_size": 32,') |
| | codes = codes.replace('"num_attention_heads": 16,', '"num_attention_heads": 2,') |
| | codes = codes.replace('"num_hidden_layers": 27,', '"num_hidden_layers": 2,') |
| | with open(Path(save_folder, 'phi4mm/vision_siglip_navit.py'), 'w') as f: |
| | f.write(codes) |
| | |
| | sys.path.append(str(Path(save_folder))) |
| | from phi4mm.modeling_phi4mm import Phi4MMForCausalLM |
| | print(Phi4MMForCausalLM) # ensure imported |
| | model = Phi4MMForCausalLM(config).to(torch.bfloat16) |
| | |
| | set_seed(42) |
| | with torch.no_grad(): |
| | for name, p in sorted(model.named_parameters()): |
| | torch.nn.init.normal_(p, 0, 0.5) |
| | print(name, p.shape) |
| | |
| | model.save_pretrained(Path(save_folder)) |
| | shutil.rmtree(Path(save_folder, 'phi4mm')) |
| | generation_config = GenerationConfig.from_pretrained( |
| | source_model_id, trust_remote_code=True, |
| | ) |
| | generation_config.save_pretrained(save_folder) |
| | ``` |