
We are happy to announce the initial release of our PLAICraft dataset. Access the data here or on the sidebar of our blog.
This is a novel dataset that captures multiplayer Minecraft interactions across five time-aligned modalities: video, game output, audio, microphone input audio, mouse and keyboard actions. This enables the study of embodied behaviour in a rich, sandbox world like Minecraft.
The initial release comprises over 200 hours from 3 different anonymized players – Dante, Dana and Morgan. Below are a few examples of our rich, multimodal dataset from interactive player sessions.
We plan to make our full 10,000 hour dataset, comprising anonymized player data from over 10,000 players from around the world, available in the future. We’re really excited to continue training and testing embodied AI agents on this dataset and aim to unveil them when they are ready. As always, thank you to everyone participating in our research and we encourage you to check back for more updates on PLAICraft!







Leave a comment