Files
hristudio/docs/roman-2025-talk.md

9.1 KiB
Executable File
Raw Blame History

A Web-Based Wizard-of-Oz Platform for Collaborative and Reproducible Human-Robot Interaction Research

1) Introduction

  • HRI needs rigorous methods for studying robot communication, collaboration, and coexistence with people.
  • WoZ: a wizard remotely operates a robot to simulate autonomous behavior, enabling rapid prototyping and iterative refinement.
  • Challenges with WoZ:
    • Wizard must execute scripted sequences consistently across participants.
    • Deviations and technical barriers reduce methodological rigor and reproducibility.
    • Many available tools require specialized technical expertise.
  • Goal: a platform that lowers barriers to entry, supports rigorous, reproducible WoZ experiments, and provides integrated capabilities.

2) Assessment of the State-of-the-Art

  • Technical infrastructure and architectures:
    • Polonius: ROS-based, finite-state machine scripting, integrated logging for real-time event recording; designed for non-programming collaborators.
    • OpenWoZ: runtime-configurable, multi-client, supports distributed operation and dynamic evaluator interventions (requires programming for behavior creation).
  • Interface design and user experience:
    • NottReal: interface for voice UI studies; tabbed pre-scripted messages, customization slots, message queuing, comprehensive logging, familiar listening/processing feedback.
    • WoZ4U: GUI designed for non-programmers; specialized to Aldebaran Pepper (limited generalizability).
  • Domain specialization vs. generalizability:
    • System longevity is often short (23 years for general-purpose tools).
    • Ozlabs longevity due to: general-purpose design, curricular integration, flexible wizard UI that adapts to experiments.
  • Standardization and methodological approaches:
    • Interaction Specification Language (ISL) and ADEs (Porfirio et al.): hierarchical modularity, formal representations, platform independence for reproducibility.
    • Riek: methodological transparency deficiencies in WoZ literature (insufficient reporting of protocols/training/constraints).
    • Steinfeld et al.: “Oz of Wizard” complements WoZ; structured permutations of real vs. simulated components; both approaches serve valid objectives.
    • Belhassein et al.: recurring HRI study challenges (limited participants, inadequate protocol reporting, weak replication); need for validated measures and comprehensive documentation.
    • Fraune et al.: practical guidance (pilot testing, ensuring intended perception of robot behaviors, managing novelty effects, cross-field collaboration).
  • Remaining challenges:
    • Accessibility for interdisciplinary teams.
    • Methodological standardization and comprehensive data capture/sharing.
    • Balance of structure (for reproducibility) and flexibility (for diverse research questions).

3) Reproducibility Challenges in WoZ Studies

  • Inconsistent wizard behavior across trials undermines reproducibility.
  • Publications often omit critical procedural details, making replication difficult.
  • Custom, ad-hoc setups are hard to recreate; unrecorded changes hinder transparency.
  • HRIStudios reproducibility requirements (five areas):
    • Standardized terminology and structure.
    • Wizard behavior formalization (clear, consistent execution with controlled flexibility).
    • Comprehensive, time-synchronized data capture.
    • Experiment specification sharing (package and distribute complete designs).
    • Procedural documentation (automatic logging of parameters and methodological details).

4) The Design and Architecture of HRIStudio

  • Guiding design principles:
    • Accessibility for researchers without deep robot programming expertise.
    • Abstraction to focus on experimental design over platform details.
    • Comprehensive data management (logs, audio, video, study materials).
    • Collaboration through multi-user accounts, role-based access control, and data sharing.
    • Embedded methodological guidance to encourage scientifically sound practices.
  • Conceptual separation aligned to research needs:
    • User-facing tools for design, execution, and analysis; stewarded data and access control; and standardized interfaces to connect experiments with robots and sensors.
  • Three-layer architecture [Screenshot Placeholder: Architecture Overview]:
    • User Interface Layer:
      • Experiment Designer (visual programming for specifying experiments).
      • Wizard Interface (real-time control for trials).
      • Playback & Analysis (data exploration and visualization).
    • Data Management Layer:
      • Structured storage of experiment definitions, metadata, and media.
      • Role-based access aligned with study responsibilities.
      • Collaboration with secure, compartmentalized access for teams.
    • Robot Integration Layer:
      • Translates standardized abstractions to robot behaviors through plugins.
      • Standardized plugin interfaces support diverse platforms without changing study designs.
      • Integrates with external systems (robot hardware, sensors, tools).
  • Sustained reproducibility and sharing:
    • Study definitions and execution environments can be packaged and shared to support faithful reproduction by independent teams.

5) Experimental Workflow Support

  • Directly addresses reproducibility requirements with standardized structures, wizard guidance, and comprehensive capture.

5.1 Hierarchical Structure for WoZ Studies

  • Standard terminology and elements:
    • Study: top-level container with one or more experiments.
    • Experiment: parameterized protocol template composed of steps.
    • Trial: concrete, executable instance of an experiment for a specific participant; all trial data recorded.
    • Step: type-bound container (wizard or robot) comprising a sequence of actions.
    • Action: atomic task for wizard or robot (e.g., input gathering, speech, movement), parameterized per trial.
  • [Screenshot Placeholder: Experiment Hierarchy Diagram].
  • [Screenshot Placeholder: Study Details View]:
    • Overview of execution summaries, trials, participant info and documents (e.g., consent), members, metadata, and audit activity.

5.2 Collaboration and Knowledge Sharing

  • Dashboard for project overview, collaborators, trial schedules, pending tasks.
  • Role-based access control (pre-defined roles; flexible extensions):
    • Administrator: system configuration/management.
    • Researcher: create/configure studies and experiments.
    • Observer: read-only access and real-time monitoring.
    • Wizard: execute experiments.
  • Packaging and dissemination of complete materials for replication and meta-analyses.

5.3 Visual Experiment Design (EDE)

  • Visual programming canvas for sequencing steps and actions (drag-and-drop).
  • Abstract robot actions translated by plugins into platform-specific commands.
  • Contextual help and documentation in the interface.
  • [Screenshot Placeholder: Experiment Designer].
  • Inspiration: Choregraphes flow-based, no-code composition for steps/actions.

5.4 Wizard Interface and Experiment Execution

  • Adaptable, experiment-specific wizard UI (avoids one-size-fits-all trap).
  • Incremental instructions, “View More” for full script, video feed, timestamped event log, and “quick actions.”
  • Observer view mirrors wizard interface without execution controls.
  • Action execution process:
    1. Translate abstract action into robot-specific calls via plugin.
    2. Route calls through appropriate communication channels.
    3. Process robot feedback, log details, update experiment state.
  • [Screenshot Placeholder: Wizard Interface].

5.5 Robot Platform Integration (Plugin Store)

  • Two-tier abstraction/translation of actions:
    • High-level action components (movement, speech, sensors) with parameter schemas and validation rules.
    • Robot plugins implement concrete mappings appropriate to each platform.
  • [Screenshot Placeholder: Plugin Store]:
    • Trust levels: Official, Verified, Community.
    • Source repositories for precise version tracking and reproducibility.

5.6 Comprehensive Data Capture and Analysis

  • Timestamped logs of all executed actions and events.
  • Robot sensor data (position, orientation, sensor readings).
  • Audio/video recordings of interactions.
  • Wizard decisions/interventions (including unplanned deviations).
  • Observer notes and annotations.
  • Structured storage for long-term preservation and analysis integration.
  • Sensitive participant data encrypted at the database level.
  • Playback for step-by-step trial review and annotation.

6) Conclusion and Future Directions

  • HRIStudio supports rigorous, reproducible WoZ experimentation via:
    • Standardized hierarchy and terminology.
    • Visual designer for protocol specification.
    • Configurable wizard interface for consistent execution.
    • Plugin-based, robot-agnostic integration.
    • Comprehensive capture and structured storage of multimodal data.
  • Future directions:
    • Interface-integrated documentation for installation and operation.
    • Enhanced execution and analysis (advanced guidance, dynamic adaptation, real-time feedback).
    • Playback for synchronized streams and expanded hardware integration.
    • Continued community engagement to refine integration with existing research infrastructures and workflows.
    • Preparation for an open beta release.