|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| """
|
| Tests for the TokenizerProcessorStep class.
|
| """
|
|
|
| import tempfile
|
| from unittest.mock import patch
|
|
|
| import pytest
|
| import torch
|
|
|
| from lerobot.configs.types import FeatureType, PipelineFeatureType, PolicyFeature
|
| from lerobot.processor import DataProcessorPipeline, TokenizerProcessorStep, TransitionKey
|
| from lerobot.processor.converters import create_transition, identity_transition
|
| from lerobot.utils.constants import ACTION, OBS_IMAGE, OBS_LANGUAGE, OBS_STATE
|
| from tests.utils import require_package
|
|
|
|
|
| class MockTokenizer:
|
| """Mock tokenizer for testing that mimics transformers tokenizer interface."""
|
|
|
| def __init__(self, vocab_size: int = 1000):
|
| self.vocab_size = vocab_size
|
|
|
| def __call__(
|
| self,
|
| text: str | list[str],
|
| max_length: int = 512,
|
| truncation: bool = True,
|
| padding: str = "max_length",
|
| padding_side: str = "right",
|
| return_tensors: str = "pt",
|
| **kwargs,
|
| ) -> dict[str, torch.Tensor]:
|
| """Mock tokenization that returns deterministic tokens based on text."""
|
| texts = [text] if isinstance(text, str) else text
|
|
|
| batch_size = len(texts)
|
|
|
|
|
| input_ids = torch.zeros(batch_size, max_length, dtype=torch.long)
|
| attention_mask = torch.zeros(batch_size, max_length, dtype=torch.long)
|
|
|
| for i, txt in enumerate(texts):
|
|
|
| text_hash = hash(txt) % self.vocab_size
|
| seq_len = min(len(txt.split()), max_length)
|
|
|
|
|
| for j in range(seq_len):
|
| input_ids[i, j] = (text_hash + j) % self.vocab_size
|
|
|
|
|
| attention_mask[i, :seq_len] = 1
|
|
|
| result = {
|
| "input_ids": input_ids,
|
| "attention_mask": attention_mask,
|
| }
|
|
|
|
|
| if len(texts) == 1:
|
| result = {k: v.squeeze(0) for k, v in result.items()}
|
|
|
| return result
|
|
|
|
|
| @pytest.fixture
|
| def mock_tokenizer():
|
| """Provide a mock tokenizer for testing."""
|
| return MockTokenizer(vocab_size=100)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_basic_tokenization(mock_auto_tokenizer):
|
| """Test basic string tokenization functionality."""
|
|
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=10)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "pick up the red cube"},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| assert result[TransitionKey.COMPLEMENTARY_DATA]["task"] == "pick up the red cube"
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| assert f"{OBS_LANGUAGE}.attention_mask" in observation
|
|
|
|
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = observation[f"{OBS_LANGUAGE}.attention_mask"]
|
| assert isinstance(tokens, torch.Tensor)
|
| assert isinstance(attention_mask, torch.Tensor)
|
| assert tokens.shape == (10,)
|
| assert attention_mask.shape == (10,)
|
|
|
|
|
| @require_package("transformers")
|
| def test_basic_tokenization_with_tokenizer_object():
|
| """Test basic string tokenization functionality using tokenizer object directly."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
|
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "pick up the red cube"},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| assert result[TransitionKey.COMPLEMENTARY_DATA]["task"] == "pick up the red cube"
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| assert f"{OBS_LANGUAGE}.attention_mask" in observation
|
|
|
|
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = observation[f"{OBS_LANGUAGE}.attention_mask"]
|
| assert isinstance(tokens, torch.Tensor)
|
| assert isinstance(attention_mask, torch.Tensor)
|
| assert tokens.shape == (10,)
|
| assert attention_mask.shape == (10,)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_list_of_strings_tokenization(mock_auto_tokenizer):
|
| """Test tokenization of a list of strings."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=8)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": ["pick up cube", "place on table"]},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| assert result[TransitionKey.COMPLEMENTARY_DATA]["task"] == ["pick up cube", "place on table"]
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = observation[f"{OBS_LANGUAGE}.attention_mask"]
|
| assert tokens.shape == (2, 8)
|
| assert attention_mask.shape == (2, 8)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_custom_keys(mock_auto_tokenizer):
|
| """Test using custom task_key."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", task_key="instruction", max_length=5)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"instruction": "move forward"},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| assert f"{OBS_LANGUAGE}.attention_mask" in observation
|
|
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| assert tokens.shape == (5,)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_none_complementary_data(mock_auto_tokenizer):
|
| """Test handling of None complementary_data."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
| transition = create_transition(observation={}, complementary_data=None)
|
|
|
|
|
| with pytest.raises(KeyError, match="task"):
|
| processor(transition)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_missing_task_key(mock_auto_tokenizer):
|
| """Test handling when task key is missing."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
| transition = create_transition(observation={}, complementary_data={"other_field": "some value"})
|
|
|
| with pytest.raises(KeyError, match="task"):
|
| processor(transition)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_none_task_value(mock_auto_tokenizer):
|
| """Test handling when task value is None."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
| transition = create_transition(observation={}, complementary_data={"task": None})
|
|
|
| with pytest.raises(ValueError, match="Task extracted from Complementary data is None"):
|
| processor(transition)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_unsupported_task_type(mock_auto_tokenizer):
|
| """Test handling of unsupported task types."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
|
|
| transition = create_transition(observation={}, complementary_data={"task": 123})
|
|
|
| with pytest.raises(ValueError, match="Task cannot be None"):
|
| processor(transition)
|
|
|
|
|
| transition = create_transition(observation={}, complementary_data={"task": ["text", 123, "more text"]})
|
|
|
| with pytest.raises(ValueError, match="Task cannot be None"):
|
| processor(transition)
|
|
|
|
|
| @require_package("transformers")
|
| def test_no_tokenizer_error():
|
| """Test that ValueError is raised when neither tokenizer nor tokenizer_name is provided."""
|
| with pytest.raises(ValueError, match="Either 'tokenizer' or 'tokenizer_name' must be provided"):
|
| TokenizerProcessorStep()
|
|
|
|
|
| @require_package("transformers")
|
| def test_invalid_tokenizer_name_error():
|
| """Test that error is raised when invalid tokenizer_name is provided."""
|
| with patch("lerobot.processor.tokenizer_processor.AutoTokenizer") as mock_auto_tokenizer:
|
|
|
| mock_auto_tokenizer.from_pretrained.side_effect = Exception("Model not found")
|
|
|
| with pytest.raises(Exception, match="Model not found"):
|
| TokenizerProcessorStep(tokenizer_name="invalid-tokenizer")
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_get_config_with_tokenizer_name(mock_auto_tokenizer):
|
| """Test configuration serialization when using tokenizer_name."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(
|
| tokenizer_name="test-tokenizer",
|
| max_length=256,
|
| task_key="instruction",
|
| padding="longest",
|
| truncation=False,
|
| )
|
|
|
| config = processor.get_config()
|
|
|
| expected = {
|
| "tokenizer_name": "test-tokenizer",
|
| "max_length": 256,
|
| "task_key": "instruction",
|
| "padding_side": "right",
|
| "padding": "longest",
|
| "truncation": False,
|
| }
|
|
|
| assert config == expected
|
|
|
|
|
| @require_package("transformers")
|
| def test_get_config_with_tokenizer_object():
|
| """Test configuration serialization when using tokenizer object."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
|
|
| processor = TokenizerProcessorStep(
|
| tokenizer=mock_tokenizer,
|
| max_length=256,
|
| task_key="instruction",
|
| padding="longest",
|
| truncation=False,
|
| )
|
|
|
| config = processor.get_config()
|
|
|
|
|
| expected = {
|
| "max_length": 256,
|
| "task_key": "instruction",
|
| "padding_side": "right",
|
| "padding": "longest",
|
| "truncation": False,
|
| }
|
|
|
| assert config == expected
|
| assert "tokenizer_name" not in config
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_state_dict_methods(mock_auto_tokenizer):
|
| """Test state_dict and load_state_dict methods."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
|
|
| state = processor.state_dict()
|
| assert state == {}
|
|
|
|
|
| processor.load_state_dict({})
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_reset_method(mock_auto_tokenizer):
|
| """Test reset method."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
|
|
| processor.reset()
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_integration_with_robot_processor(mock_auto_tokenizer):
|
| """Test integration with RobotProcessor."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| tokenizer_processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=6)
|
| robot_processor = DataProcessorPipeline(
|
| [tokenizer_processor], to_transition=identity_transition, to_output=identity_transition
|
| )
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "test task"},
|
| )
|
|
|
| result = robot_processor(transition)
|
|
|
|
|
| assert TransitionKey.OBSERVATION in result
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| assert f"{OBS_LANGUAGE}.attention_mask" in observation
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = observation[f"{OBS_LANGUAGE}.attention_mask"]
|
| assert tokens.shape == (6,)
|
| assert attention_mask.shape == (6,)
|
|
|
|
|
| assert torch.equal(
|
| result[TransitionKey.OBSERVATION]["state"], transition[TransitionKey.OBSERVATION]["state"]
|
| )
|
| assert torch.equal(result[TransitionKey.ACTION], transition[TransitionKey.ACTION])
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_save_and_load_pretrained_with_tokenizer_name(mock_auto_tokenizer):
|
| """Test saving and loading processor with tokenizer_name."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| original_processor = TokenizerProcessorStep(
|
| tokenizer_name="test-tokenizer", max_length=32, task_key="instruction"
|
| )
|
|
|
| robot_processor = DataProcessorPipeline(
|
| [original_processor], to_transition=identity_transition, to_output=identity_transition
|
| )
|
|
|
| with tempfile.TemporaryDirectory() as temp_dir:
|
|
|
| robot_processor.save_pretrained(temp_dir)
|
|
|
|
|
| loaded_processor = DataProcessorPipeline.from_pretrained(
|
| temp_dir,
|
| config_filename="dataprocessorpipeline.json",
|
| to_transition=identity_transition,
|
| to_output=identity_transition,
|
| )
|
|
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"instruction": "test instruction"},
|
| )
|
|
|
| result = loaded_processor(transition)
|
| assert TransitionKey.OBSERVATION in result
|
| assert f"{OBS_LANGUAGE}.tokens" in result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.attention_mask" in result[TransitionKey.OBSERVATION]
|
|
|
|
|
| @require_package("transformers")
|
| def test_save_and_load_pretrained_with_tokenizer_object():
|
| """Test saving and loading processor with tokenizer object using overrides."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
|
|
| original_processor = TokenizerProcessorStep(
|
| tokenizer=mock_tokenizer, max_length=32, task_key="instruction"
|
| )
|
|
|
| robot_processor = DataProcessorPipeline(
|
| [original_processor], to_transition=identity_transition, to_output=identity_transition
|
| )
|
|
|
| with tempfile.TemporaryDirectory() as temp_dir:
|
|
|
| robot_processor.save_pretrained(temp_dir)
|
|
|
|
|
| loaded_processor = DataProcessorPipeline.from_pretrained(
|
| temp_dir,
|
| config_filename="dataprocessorpipeline.json",
|
| overrides={"tokenizer_processor": {"tokenizer": mock_tokenizer}},
|
| to_transition=identity_transition,
|
| to_output=identity_transition,
|
| )
|
|
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"instruction": "test instruction"},
|
| )
|
|
|
| result = loaded_processor(transition)
|
| assert TransitionKey.OBSERVATION in result
|
| assert f"{OBS_LANGUAGE}.tokens" in result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.attention_mask" in result[TransitionKey.OBSERVATION]
|
|
|
|
|
| @require_package("transformers")
|
| def test_registry_functionality():
|
| """Test that the processor is properly registered."""
|
| from lerobot.processor import ProcessorStepRegistry
|
|
|
|
|
| assert "tokenizer_processor" in ProcessorStepRegistry.list()
|
|
|
|
|
| retrieved_class = ProcessorStepRegistry.get("tokenizer_processor")
|
| assert retrieved_class is TokenizerProcessorStep
|
|
|
|
|
| @require_package("transformers")
|
| def test_features_basic():
|
| """Test basic feature contract functionality."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=128)
|
|
|
| input_features = {
|
| PipelineFeatureType.OBSERVATION: {OBS_STATE: PolicyFeature(type=FeatureType.STATE, shape=(10,))},
|
| PipelineFeatureType.ACTION: {ACTION: PolicyFeature(type=FeatureType.ACTION, shape=(5,))},
|
| }
|
|
|
| output_features = processor.transform_features(input_features)
|
|
|
|
|
| assert OBS_STATE in output_features[PipelineFeatureType.OBSERVATION]
|
| assert ACTION in output_features[PipelineFeatureType.ACTION]
|
|
|
|
|
| assert f"{OBS_LANGUAGE}.tokens" in output_features[PipelineFeatureType.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.attention_mask" in output_features[PipelineFeatureType.OBSERVATION]
|
|
|
|
|
| tokens_feature = output_features[PipelineFeatureType.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask_feature = output_features[PipelineFeatureType.OBSERVATION][
|
| f"{OBS_LANGUAGE}.attention_mask"
|
| ]
|
|
|
| assert tokens_feature.type == FeatureType.LANGUAGE
|
| assert tokens_feature.shape == (128,)
|
| assert attention_mask_feature.type == FeatureType.LANGUAGE
|
| assert attention_mask_feature.shape == (128,)
|
|
|
|
|
| @require_package("transformers")
|
| def test_features_with_custom_max_length():
|
| """Test feature contract with custom max_length."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=64)
|
|
|
| input_features = {PipelineFeatureType.OBSERVATION: {}}
|
| output_features = processor.transform_features(input_features)
|
|
|
|
|
| assert f"{OBS_LANGUAGE}.tokens" in output_features[PipelineFeatureType.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.attention_mask" in output_features[PipelineFeatureType.OBSERVATION]
|
|
|
| tokens_feature = output_features[PipelineFeatureType.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask_feature = output_features[PipelineFeatureType.OBSERVATION][
|
| f"{OBS_LANGUAGE}.attention_mask"
|
| ]
|
|
|
| assert tokens_feature.shape == (64,)
|
| assert attention_mask_feature.shape == (64,)
|
|
|
|
|
| @require_package("transformers")
|
| def test_features_existing_features():
|
| """Test feature contract when tokenized features already exist."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=256)
|
|
|
| input_features = {
|
| PipelineFeatureType.OBSERVATION: {
|
| f"{OBS_LANGUAGE}.tokens": PolicyFeature(type=FeatureType.LANGUAGE, shape=(100,)),
|
| f"{OBS_LANGUAGE}.attention_mask": PolicyFeature(type=FeatureType.LANGUAGE, shape=(100,)),
|
| }
|
| }
|
|
|
| output_features = processor.transform_features(input_features)
|
|
|
|
|
| assert output_features[PipelineFeatureType.OBSERVATION][f"{OBS_LANGUAGE}.tokens"].shape == (
|
| 100,
|
| )
|
| assert output_features[PipelineFeatureType.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"].shape == (100,)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_tokenization_parameters(mock_auto_tokenizer):
|
| """Test that tokenization parameters are correctly passed to tokenizer."""
|
|
|
|
|
| class TrackingMockTokenizer:
|
| def __init__(self):
|
| self.last_call_args = None
|
| self.last_call_kwargs = None
|
|
|
| def __call__(self, *args, **kwargs):
|
| self.last_call_args = args
|
| self.last_call_kwargs = kwargs
|
|
|
| return {
|
| "input_ids": torch.zeros(16, dtype=torch.long),
|
| "attention_mask": torch.ones(16, dtype=torch.long),
|
| }
|
|
|
| tracking_tokenizer = TrackingMockTokenizer()
|
| mock_auto_tokenizer.from_pretrained.return_value = tracking_tokenizer
|
|
|
| processor = TokenizerProcessorStep(
|
| tokenizer_name="test-tokenizer",
|
| max_length=16,
|
| padding="longest",
|
| truncation=False,
|
| padding_side="left",
|
| )
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "test task"},
|
| )
|
|
|
| processor(transition)
|
|
|
|
|
| assert tracking_tokenizer.last_call_args == (["test task"],)
|
| assert tracking_tokenizer.last_call_kwargs["max_length"] == 16
|
| assert tracking_tokenizer.last_call_kwargs["padding"] == "longest"
|
| assert tracking_tokenizer.last_call_kwargs["padding_side"] == "left"
|
| assert tracking_tokenizer.last_call_kwargs["truncation"] is False
|
| assert tracking_tokenizer.last_call_kwargs["return_tensors"] == "pt"
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_preserves_other_complementary_data(mock_auto_tokenizer):
|
| """Test that other complementary data fields are preserved."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer")
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={
|
| "task": "test task",
|
| "episode_id": 123,
|
| "timestamp": 456.789,
|
| "other_field": {"nested": "data"},
|
| },
|
| )
|
|
|
| result = processor(transition)
|
| comp_data = result[TransitionKey.COMPLEMENTARY_DATA]
|
|
|
|
|
| assert comp_data["task"] == "test task"
|
| assert comp_data["episode_id"] == 123
|
| assert comp_data["timestamp"] == 456.789
|
| assert comp_data["other_field"] == {"nested": "data"}
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| assert f"{OBS_LANGUAGE}.attention_mask" in observation
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_deterministic_tokenization(mock_auto_tokenizer):
|
| """Test that tokenization is deterministic for the same input."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=10)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "consistent test"},
|
| )
|
|
|
| result1 = processor(transition)
|
| result2 = processor(transition)
|
|
|
| tokens1 = result1[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask1 = result1[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
| tokens2 = result2[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask2 = result2[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
|
|
| assert torch.equal(tokens1, tokens2)
|
| assert torch.equal(attention_mask1, attention_mask2)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_empty_string_task(mock_auto_tokenizer):
|
| """Test handling of empty string task."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=8)
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": ""},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| assert f"{OBS_LANGUAGE}.tokens" in observation
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| assert tokens.shape == (8,)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_very_long_task(mock_auto_tokenizer):
|
| """Test handling of very long task strings."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
| processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=5, truncation=True)
|
|
|
| long_task = " ".join(["word"] * 100)
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": long_task},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| observation = result[TransitionKey.OBSERVATION]
|
| tokens = observation[f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = observation[f"{OBS_LANGUAGE}.attention_mask"]
|
| assert tokens.shape == (5,)
|
| assert attention_mask.shape == (5,)
|
|
|
|
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_custom_padding_side(mock_auto_tokenizer):
|
| """Test using custom padding_side parameter."""
|
|
|
|
|
| class PaddingSideTrackingTokenizer:
|
| def __init__(self):
|
| self.padding_side_calls = []
|
|
|
| def __call__(
|
| self,
|
| text,
|
| max_length=512,
|
| truncation=True,
|
| padding="max_length",
|
| padding_side="right",
|
| return_tensors="pt",
|
| **kwargs,
|
| ):
|
| self.padding_side_calls.append(padding_side)
|
|
|
| return {
|
| "input_ids": torch.zeros(max_length, dtype=torch.long),
|
| "attention_mask": torch.ones(max_length, dtype=torch.long),
|
| }
|
|
|
| tracking_tokenizer = PaddingSideTrackingTokenizer()
|
| mock_auto_tokenizer.from_pretrained.return_value = tracking_tokenizer
|
|
|
|
|
| processor_left = TokenizerProcessorStep(
|
| tokenizer_name="test-tokenizer", max_length=10, padding_side="left"
|
| )
|
|
|
| transition = create_transition(
|
| observation={"state": torch.tensor([1.0, 2.0])},
|
| action=torch.tensor([0.1, 0.2]),
|
| complementary_data={"task": "test task"},
|
| )
|
| processor_left(transition)
|
|
|
| assert tracking_tokenizer.padding_side_calls[-1] == "left"
|
|
|
|
|
| processor_right = TokenizerProcessorStep(
|
| tokenizer_name="test-tokenizer", max_length=10, padding_side="right"
|
| )
|
|
|
| processor_right(transition)
|
|
|
| assert tracking_tokenizer.padding_side_calls[-1] == "right"
|
|
|
|
|
| @require_package("transformers")
|
| def test_device_detection_cpu():
|
| """Test that tokenized tensors stay on CPU when other tensors are on CPU."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| observation = {OBS_STATE: torch.randn(10)}
|
| action = torch.randn(5)
|
| transition = create_transition(
|
| observation=observation, action=action, complementary_data={"task": "test task"}
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device.type == "cpu"
|
| assert attention_mask.device.type == "cpu"
|
|
|
|
|
| @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available")
|
| @require_package("transformers")
|
| def test_device_detection_cuda():
|
| """Test that tokenized tensors are moved to CUDA when other tensors are on CUDA."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| observation = {OBS_STATE: torch.randn(10).cuda()}
|
| action = torch.randn(5).cuda()
|
| transition = create_transition(
|
| observation=observation, action=action, complementary_data={"task": "test task"}
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device.type == "cuda"
|
| assert attention_mask.device.type == "cuda"
|
| assert tokens.device.index == 0
|
|
|
|
|
| @pytest.mark.skipif(torch.cuda.device_count() < 2, reason="Requires at least 2 GPUs")
|
| @require_package("transformers")
|
| def test_device_detection_multi_gpu():
|
| """Test that tokenized tensors match device in multi-GPU setup."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| device = torch.device("cuda:1")
|
| observation = {OBS_STATE: torch.randn(10).to(device)}
|
| action = torch.randn(5).to(device)
|
| transition = create_transition(
|
| observation=observation, action=action, complementary_data={"task": "multi gpu test"}
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device == device
|
| assert attention_mask.device == device
|
|
|
|
|
| @require_package("transformers")
|
| def test_device_detection_no_tensors():
|
| """Test that tokenized tensors stay on CPU when no other tensors exist."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| transition = create_transition(
|
| observation={"metadata": {"key": "value"}},
|
| complementary_data={"task": "no tensor test"},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device.type == "cpu"
|
| assert attention_mask.device.type == "cpu"
|
|
|
|
|
| @require_package("transformers")
|
| def test_device_detection_mixed_devices():
|
| """Test device detection when tensors are on different devices (uses first found)."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
| if torch.cuda.is_available():
|
|
|
| observation = {
|
| "observation.cpu": torch.randn(10),
|
| "observation.cuda": torch.randn(10).cuda(),
|
| }
|
| transition = create_transition(
|
| observation=observation, complementary_data={"task": "mixed device test"}
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
|
|
| assert tokens.device == attention_mask.device
|
|
|
|
|
| @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available")
|
| @require_package("transformers")
|
| def test_device_detection_from_action():
|
| """Test that device is detected from action tensor when no observation tensors exist."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| observation = {"metadata": {"key": "value"}}
|
| action = torch.randn(5).cuda()
|
| transition = create_transition(
|
| observation=observation, action=action, complementary_data={"task": "action device test"}
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device.type == "cuda"
|
| assert attention_mask.device.type == "cuda"
|
|
|
|
|
| @require_package("transformers")
|
| def test_device_detection_preserves_dtype():
|
| """Test that device detection doesn't affect dtype of tokenized tensors."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| observation = {OBS_STATE: torch.randn(10, dtype=torch.float16)}
|
| transition = create_transition(observation=observation, complementary_data={"task": "dtype test"})
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.dtype == torch.long
|
| assert attention_mask.dtype == torch.bool
|
|
|
|
|
| @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available")
|
| @require_package("transformers")
|
| @patch("lerobot.processor.tokenizer_processor.AutoTokenizer")
|
| def test_integration_with_device_processor(mock_auto_tokenizer):
|
| """Test that TokenizerProcessorStep works correctly with DeviceProcessorStep in pipeline."""
|
| from lerobot.processor import DeviceProcessorStep
|
|
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| mock_auto_tokenizer.from_pretrained.return_value = mock_tokenizer
|
|
|
|
|
| tokenizer_processor = TokenizerProcessorStep(tokenizer_name="test-tokenizer", max_length=6)
|
| device_processor = DeviceProcessorStep(device="cuda:0")
|
| robot_processor = DataProcessorPipeline(
|
| [tokenizer_processor, device_processor],
|
| to_transition=identity_transition,
|
| to_output=identity_transition,
|
| )
|
|
|
|
|
| transition = create_transition(
|
| observation={OBS_STATE: torch.randn(10)},
|
| action=torch.randn(5),
|
| complementary_data={"task": "pipeline test"},
|
| )
|
|
|
| result = robot_processor(transition)
|
|
|
|
|
| assert result[TransitionKey.OBSERVATION][OBS_STATE].device.type == "cuda"
|
| assert result[TransitionKey.ACTION].device.type == "cuda"
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
| assert tokens.device.type == "cuda"
|
| assert attention_mask.device.type == "cuda"
|
|
|
|
|
| @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available")
|
| @require_package("transformers")
|
| def test_simulated_accelerate_scenario():
|
| """Test scenario simulating Accelerate with data already on GPU."""
|
| mock_tokenizer = MockTokenizer(vocab_size=100)
|
| processor = TokenizerProcessorStep(tokenizer=mock_tokenizer, max_length=10)
|
|
|
|
|
| device = torch.device("cuda:0")
|
| observation = {
|
| OBS_STATE: torch.randn(1, 10).to(device),
|
| OBS_IMAGE: torch.randn(1, 3, 224, 224).to(device),
|
| }
|
| action = torch.randn(1, 5).to(device)
|
|
|
| transition = create_transition(
|
| observation=observation,
|
| action=action,
|
| complementary_data={"task": ["accelerate test"]},
|
| )
|
|
|
| result = processor(transition)
|
|
|
|
|
| tokens = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.tokens"]
|
| attention_mask = result[TransitionKey.OBSERVATION][f"{OBS_LANGUAGE}.attention_mask"]
|
|
|
| assert tokens.device == device
|
| assert attention_mask.device == device
|
|
|
| assert tokens.shape == (10,)
|
| assert attention_mask.shape == (10,)
|
|
|