Skip to content

This repository contains the Event-based Manipulation Action Dataset, which represents manipulation actions recorded using event cameras. This dataset consists of 30 different manipulation actions using 6 objects.

License

Notifications You must be signed in to change notification settings

DaniDeniz/DavisHandDataset-Events

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Event-based Manipulation Action Dataset

This repository contains the Event-based Manipulation Action Dataset, which represents manipulation actions recorded using event cameras. This dataset consists of 30 different manipulation actions using 6 objects. The 30 classes included in this dataset are the following:

  • cup
    • drink, pound, shake, move, pour
  • stone
    • pound, move, play, grind, carve
  • sponge
    • squeeze, flip, wash, wipe, scratch
  • spoon
    • scoop, stir, hit, eat, sprinkle
  • knife
    • cut, chop, poke a hole, peel, spread
  • spatula
    • flip, lift, cut, squeeze, scrape

Refer to classes.json to check the ID of every action.

Additionally, the examples were recorded by 5 different subjects (S1, S2, S3, S4, and S5). Each subject recorded five instances of every action. Therefore, each class has 25 samples, which results in 750 examples recorded with events with an average duration of 3 seconds.

Dataset retrieval process

To obtain and download the dataset you should:

git clone https://github.com/DaniDeniz/DavisHandDataset-Events.git
cd DavisHandDataset-Events
./setup_data.sh

This script will download a zip with the event data. The zip includes two folders. One folder (AllEvents) includes all the events recorded from the DAVIS camera, the second folder (TrackerEvents) includes only the events that occur around the hand of the subject (we used a hand tracker). With this, we reduce the number of events required to analyze the sequences and identify the activities.

The event sequences are represented using .npy files that can be opened using numpy. The data is organized as follows: 6 folders for each of the objects, and then, each of that folders include 5 folders representing the different activities. Finally, the actions folders have 5 .npy files with events recorded from 5 subjects (25 in total).

Additionally, the above script also splits the data into train, validation, and test. Concretely, the script creates 6 datasets using the event data:

  • all_davis_hand_dataset_npy: data is randomly split into train, validation, and test
  • s1_davis_hand_dataset_npy: data of s1 used exclusively for test
  • s2_davis_hand_dataset_npy: data of s2 used exclusively for test
  • s3_davis_hand_dataset_npy: data of s3 used exclusively for test
  • s4_davis_hand_dataset_npy: data of s4 used exclusively for test
  • s5_davis_hand_dataset_npy: data of s5 used exclusively for test

Read event data

Once the event data is downloaded, we can read it using numpy. To do that, for example, you should:

import numpy as np
events_segment = np.load("s1_segment_001.npy", allow_pickle=True).item()

The events object is organized as a dictionary with the following keys:

  • x: positions of the events in the x axis (horizontal)
  • y: positions of the events in the y axis (vertical)
  • ts: timestamp in microseconds underlining when the asynchronous event occurred
  • p: polarity value (-1 or 1)
  • max_width: higher width position (useful when building time surfaces)
  • max_height: higher height position (useful when building time surfaces)

With this information, it is possible to build time surfaces using the event data. In particular, you can create time surfaces using different algorithms defining a particular tau value to show the sequence of events as a video. (Refer to tutorial_read_events.ipynb - TBD - for further details)

Action: Cup pour

Action: Spatula lift

Citation

[1] D. Deniz, C. Fermuller, E. Ros, M. Rodriguez-Alvarez, & F. Barranco. (2023) "Event-based Vision for Early Prediction of Manipulation Actions". arXiv preprint arXiv:2307.14332.

[2] D. Deniz, E. Ros, C. Fermuller, & F. Barranco. (2023). "When Do Neuromorphic Sensors Outperform cameras? Learning from Dynamic Features". In 2023 57th Annual Conference on Information Sciences and Systems (CISS) (pp. 1-6). IEEE.

Acknowledgments

We acknowledge the Telluride Neuromorphic Cognition Engineering Workshop (http://www.ine-web.org) for the fruitful discussions on neuromorphic cognition and their participants for helping with the recording of the dataset.

About

This repository contains the Event-based Manipulation Action Dataset, which represents manipulation actions recorded using event cameras. This dataset consists of 30 different manipulation actions using 6 objects.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published