Skip to content

Failing to reproduce the paper result on videomme #33

@joslefaure

Description

@joslefaure

I use the same script as in the Readme:

accelerate launch --num_processes 8 --main_process_port 12345 -m lmms_eval \
    --model longva \
    --model_args pretrained=lmms-lab/LongVA-7B,conv_template=qwen_1_5,video_decode_backend=decord,max_frames_num=32,model_name=llava_qwen \
    --tasks videomme \
    --batch_size 1 \
    --log_samples \
    --log_samples_suffix videomme_longva \
    --output_path ./logs/ 

With the latest in the latest commit of lmms_eval (main branch): bcbdc493

I get the following results:

Tasks Version Filter n-shot Metric Value Stderr
videomme Yaml none 0 videomme_perception_score 23.5185 ± N/A

Could you please advise on what I am doing wrong? Thanks

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions