Skip to content
/ ASMR Public

Augmenting Life Scenario using Large Generative Models for Robotic Action Reflection

Notifications You must be signed in to change notification settings

MiuLab/ASMR

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 

Repository files navigation

ASMR: Augmenting Life Scenario using Large Generative Models for Robotic Action Reflection

[IWSDS 2024 paper]

Overview

Framework

Requirements

  • Python >= 3.6
  • Transformers
  • torch

Datasets

Our generated dataset can be found in the google drive

Augmentation Scenario Cases

We demonstrate some scenario cases generated from our framework

How to run

Pretrained LLM

Corpus

Run the generation script

  • Modify the paths OPENAI_API_KEY and DATA_JSONL in generate.py to your local paths.
  • Simple run
    python3 generate.py

Finetune a multimoal model

Please refer to the original LLAVA repo for training a multimodal model given the generated dataset. Below is a reference command we used with LLaVA to train the model:

CUDA_VISIBLE_DEVICES=0,1,2,3 python3 train.py \
    --max_grad_norm 2.0 \
    --encoder_model_type hf_bert \
    --pretrained_model_cfg bert-base-uncased \
    --seed 12345 \
    --sequence_length 384 \
    --warmup_steps 1237 \
    --batch_size 64 \
    --dev_batch_size 16 \
    --do_lower_case \
    --train_file ${GENERATED_DATASET} \
    --dev_file ../ConvDR/datasets/or-quac/dev_dpr.json \
    --output_dir ${MODEL_DIR} \
    --learning_rate 2e-05 \
    --num_train_epochs 30 \
    --val_av_rank_start_epoch 0 \
    --fp16 \
    --grad_cache \
    --q_chunk_size 8 \
    --ctx_chunk_size 8 \
    --global_loss_buf_sz 2097152 \
    --val_av_rank_max_qs 10000

Do-I-Demand Benchmark

Utterance-Only

Model SBERT GPT3
LLaVA-13B 20.3 24.5
+ place-based augmentation 29.0 34.3
+ action-based augmentation 31.5 31.5
+ both 36.3 35.3
LLaVA-7B 19.5 22.5
+ place-based augmentation 30.3 36.1
+ action-based augmentation 32.3 32.5
+ both 34.0 33.8

Utterance + Description

Model SBERT GPT3
LLaVA-13B 28.3 34.8
+ place-based augmentation 33.3 39.0
+ action-based augmentation 45.5 45.5
+ both 48.5 47.8
LLaVA-7B 27.8 36.3
+ place-based augmentation 36.0 42.3
+ action-based augmentation 41.8 41.5
+ both 48.8 47.5

Reference

Please cite the following paper

About

Augmenting Life Scenario using Large Generative Models for Robotic Action Reflection

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published