Skip to main content

Python module

lora

AttentionWithRopeAndLoRA

class max.nn.legacy.lora.AttentionWithRopeAndLoRA(*, rope, num_attention_heads, num_key_value_heads, hidden_size, kv_params, max_lora_rank, max_num_loras, devices=None, dtype=float32, linear_cls=<class 'max.nn.legacy.linear.Linear'>, stacked_qkv=False, scale=None, has_bias=False, float8_config=None, clip_qkv=None)

Parameters:

rope

rope: RotaryEmbedding

LinearLoRA

class max.nn.legacy.lora.LinearLoRA(in_dim, out_dim, max_num_loras, max_lora_rank, dtype, device, has_lora_bias=False, name=None, quantization_encoding=None)

Parameters:

set_lora_batch_info()

set_lora_batch_info(lora_ids, lora_ranks, lora_grouped_offsets, num_active_loras, lora_end_idx, batch_seq_len, lora_ids_kv, lora_grouped_offsets_kv)

Parameters:

Return type:

None

SupportsLoRA

class max.nn.legacy.lora.SupportsLoRA(*args, **kwargs)

Base class for supporting LoRA functionality in Modules

set_lora_batch_info()

set_lora_batch_info(lora_ids, lora_ranks, lora_grouped_offsets, num_active_loras, lora_end_idx, batch_seq_len, lora_ids_kv, lora_grouped_offsets_kv)

Parameters:

Return type:

None

Was this page helpful?