GithubHelp home page GithubHelp logo

SLED Research Lab @ University of Michigan

Overview

Language use in human communication is shaped by our goals, our shared experiences, and our understanding of each other’s abilities, knowledge, and beliefs. At the Situated Language and Embodied Dialogue (SLED) group (formerly, the LAIR group), we develop computational models for language processing that take into account the rich situational and communicative context and build embodied AI agents that can communicate with humans in language.

Contact Info

  • Location: Bob and Betty Beyster Building 3632, 2260 Hayward Street, Ann Arbor, MI 48109.
  • Email: [email protected]

News

  • [Oct. 2021] Our work MindCraft: Theory of Mind Modeling for Situated Dialogue in Collaborative Tasks is selected as the Outstanding Paper of EMNLP 2021! Congratulations to Paul!!
  • [Oct. 2021] Congratulations to Yichi for successfully passing the prelim exam!!
  • [Aug. 2021] SLED has 3 papers accepted to EMNLP 2021! Congrats to Shane and Paul! 🙌🙌
  • [June 2021] Congratulations to Emily and Shane for successfully passing the prelim exam!!
  • [Apr. 2021] Very excited to have Ziqiao (Martin) Ma and Jianing (Jed) Yang join SLED in Fall 2021!
  • [Oct. 2020] Congratulations to Paul for successfully passing the prelim exam!!

Current Members

Faculty

Ph.D. Students

A list of active assitants and alumni can be found here.

Research

Mission Statement

We develop computational models for language processing that takes into account the rich situational and communicative context and build embodied AI agents that can communicate with humans in language.

Grounded Language Understanding towards Physical Interaction

With the emergence of a new generation of cognitive robots, the capability to communicate with these robots using natural language has become increasingly important. Verbal communication often involves the use of verbs, for example, to ask a robot to perform some tasks or to monitor some physical activities. Concrete action verbs often denote some change of state as a result of an action; for example, “slice a pizza” implies the state of the object pizza will be changed from one piece to several smaller pieces. The change of state can be perceived from the physical world through different sensors. Given a human utterance, if the robot can anticipate the potential change of the state signaled by the verbs, it can then actively sense the environment and better connect language with the perceived physical world such as who performs the action and what objects and locations are involved. This improved connection will benefit many applications relying on human-robot communication.

Situated Human Robot Dialogue

A new generation of interactive robots have emerged in recent years to provide service, care, and companionship for humans. To support natural interaction between a human and these robots, technology enabling situated dialogue becomes increasingly important. Situated dialogue is drastically different from traditional spoken dialogue systems, multimodal conversational interfaces, and tele-operated human robot interaction. In situated human robot dialogue, human partners and robots are situated and co-present in a shared physical environment. The shared surrounding significantly affects how they interact with each other and how the robot interprets human language and performs tasks. In the last couple years, we have started a couple projects on situated human robot dialogue, particularly focusing on how the situated-ness affects human robot language based interaction, and thus techniques for situated language processing and conversation grounding.

Commonsense Reasoning, Semantic and Discourse Processing

Given a large amount of textual data (e.g., news articles, Wikipedia, weblogs, etc.) available online, it has become increasingly important for techniques that can automatically process this data, for example, to extract event information, to answer user questions, and to make inferences. Along these lines, we are particularly interested in the role of discourse and pragmatics in natural language processing and their applications.

Situated and Multimodal Language Processing in Virtual World

Using situated dialogue (in virtual world) and conversational interfaces as our setting, we have investigated the use of non-verbal modalities (e.g., eye gaze and gestures) in language processing and in conversation grounding. The virtual world setting not only has important applications in education, training, and entertainment; but also provides a simplified simulation environment to support studies on situated language processing towards physical world interaction.

Prospective Students

  • PhD students: Please apply to the CSE department.
  • Masters/Undergraduates/Visitors: Thank you for your interest in participating in our research. To apply for a research position, please complete this Google form.
    • For current Michigan students: We expect you to commit at least 15 hours per week to gain meaningful research experience. If you are chosen for the position, you will need to take EECS X99: Independent Studies for this research experience:
      • Credits: 3 credits for 15 hours per week; and 4 credits for 20 hours per week.
      • Registration: EECS 399 for students below senior standing (less than 85 CTP); EECS 499 for senior standing; and EECS 599 for master students.

Responses may be slow due to the volume of applications we received. Positions are offered upon availability and qualification.

Situated Language and Embodied Dialogue (SLED) Research Group's Projects

chat-with-nerf icon chat-with-nerf

Chat with NeRF enables users to interact with a NeRF model by typing in natural language.

collab-plan-acquisition icon collab-plan-acquisition

Official Code for IJCAI 2023 Paper: Towards Collaborative Plan Acquisition through Theory of Mind Modeling in Situated Dialogue

comparative-learning icon comparative-learning

Official Code for ACL 2023 Paper: Human Inspired Progressive Alignment and Comparative Learning for Grounded Word Acquisition

conversation-entailment icon conversation-entailment

Official dataset for Towards Conversation Entailment: An Empirical Investigation. Chen Zhang, Joyce Chai. EMNLP, 2010

cyclenet icon cyclenet

Official Code for NeurIPS 2023 Paper: CycleNet: Rethinking Cycle Consistent in Text‑Guided Diffusion for Image Manipulation

danli icon danli

Code for EMNLP 2022 Paper DANLI: Deliberative Agent for Following Natural Language Instructions

dorothie icon dorothie

Official Code for DOROTHIE: Spoken Dialogue for Handling Unexpected Situations in Interactive Autonomous Driving Agents (Findings of EMNLP 2022)

grounded-semantic-role-labeling icon grounded-semantic-role-labeling

Official dataset for Grounded Semantic Role Labeling. Shaohua Yang, Qiaozi Gao, Changsong Liu, Caiming Xiong, Song-Chun Zhu, Joyce Y. Chai. NAACL HLT, 2016.

heuristic-analytic-reasoning icon heuristic-analytic-reasoning

Repo for the EMNLP 2023 paper "From Heuristic to Analytic: Cognitively-Motivated Reasoning Strategies for Coherent Physical Commonsense in Pre-Trained Language Models."

hitut icon hitut

Official code for the ACL 2021 Findings paper "Yichi Zhang and Joyce Chai. Hierarchical Task Learning from Language Instructions with Unified Transformers and Self-Monitoring"

infedit icon infedit

[CVPR 2024] Official implementation of CVPR 2024 paper: "Inversion-Free Image Editing with Natural Language"

lerf icon lerf

Code for LERF: Language Embedded Radiance Fields

lerf_lite icon lerf_lite

Code for LERF: Language Embedded Radiance Fields

mindcraft icon mindcraft

Official code for our EMNLP2021 Outstanding Paper MindCraft: Theory of Mind Modeling for Situated Dialogue in Collaborative Tasks

navchat icon navchat

Code for ICRA24 paper "Think, Act, and Ask: Open-World Interactive Personalized Robot Navigation" https://arxiv.org/abs/2310.07968

physical-causality-of-action-verbs icon physical-causality-of-action-verbs

Official dataset for Physical Causality of Action Verbs in Grounded Language Understanding. Qiaozi Gao, Malcolm Doering, Shaohua Yang, Joyce Chai. ACL, 2016.

verifiable-coherent-nlu icon verifiable-coherent-nlu

Shared repository for TRIP dataset for verifiable NLU and coherence measurement for text classifiers.

world-to-words icon world-to-words

Official Code for ACL 2023 Outstanding Paper: World-to-Words: Grounded Open Vocabulary Acquisition through Fast Mapping in Vision-Language Models

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.