vllm.model_executor.models.lightonocr ¶
LightOnOCRForConditionalGeneration ¶
Bases: Mistral3ForConditionalGeneration
Source code in vllm/model_executor/models/lightonocr.py
hf_to_vllm_mapper class-attribute
instance-attribute
¶
hf_to_vllm_mapper = WeightsMapper(
orig_to_new_prefix={
"model.vision_encoder.": "vision_tower.",
"model.vision_projection.": "multi_modal_projector.",
"lm_head.": "language_model.lm_head.",
"model.language_model.": "language_model.model.",
}
)
language_model instance-attribute
¶
language_model = init_vllm_registered_model(
vllm_config=vllm_config,
hf_config=text_config,
prefix=maybe_prefix(prefix, "language_model"),
)
make_empty_intermediate_tensors instance-attribute
¶
multi_modal_projector instance-attribute
¶
multi_modal_projector = Mistral3MultiModalProjector(
vision_hidden_size=hidden_size,
text_hidden_size=hidden_size,
projector_hidden_act=projector_hidden_act,
spatial_merge_size=spatial_merge_size,
patch_size=patch_size,
multimodal_projector_bias=multimodal_projector_bias,
quant_config=quant_config,
prefix=maybe_prefix(prefix, "multi_modal_projector"),
)
vision_tower instance-attribute
¶
vision_tower = init_vision_tower_for_llava(
config,
quant_config,
require_post_norm=False,
prefix=maybe_prefix(prefix, "vision_tower"),
)
__init__ ¶
__init__(
*, vllm_config: VllmConfig, prefix: str = ""
) -> None
Source code in vllm/model_executor/models/lightonocr.py
load_weights ¶
LightOnOCRMultiModalProcessor ¶
Bases: BaseMultiModalProcessor[Mistral3ProcessingInfo]
Source code in vllm/model_executor/models/lightonocr.py
_call_hf_processor ¶
_call_hf_processor(
prompt: str,
mm_data: Mapping[str, object],
mm_kwargs: Mapping[str, object],
tok_kwargs: Mapping[str, object],
) -> BatchFeature
Source code in vllm/model_executor/models/lightonocr.py
_get_mm_fields_config ¶
_get_mm_fields_config(
hf_inputs: BatchFeature,
hf_processor_mm_kwargs: Mapping[str, object],
) -> Mapping[str, MultiModalFieldConfig]
Source code in vllm/model_executor/models/lightonocr.py
_get_prompt_updates ¶
_get_prompt_updates(
mm_items: MultiModalDataItems,
hf_processor_mm_kwargs: Mapping[str, object],
out_mm_kwargs: MultiModalKwargs,
) -> Sequence[PromptUpdate]
Source code in vllm/model_executor/models/lightonocr.py
_build_LightOnOCR_processor ¶
_build_LightOnOCR_processor(
info: _I,
dummy_inputs: BaseDummyInputsBuilder[_I],
*,
cache: BaseMultiModalProcessorCache | None = None,
)