| |
|
|
| import os |
| import re |
| import subprocess |
| import sys |
| from contextlib import redirect_stderr, redirect_stdout |
| from io import StringIO |
| from unittest.mock import ANY, Mock, call |
|
|
| import pytest |
| import torch |
| import yaml |
|
|
| skip_in_ci_on_macos = pytest.mark.skipif( |
| sys.platform == "darwin" and os.getenv("GITHUB_ACTIONS") == "true", |
| reason="Skipped on macOS in CI environment because CI machine does not have enough memory to run this test.", |
| ) |
|
|
|
|
| @skip_in_ci_on_macos |
| @pytest.mark.parametrize("version", ("v1", "v2")) |
| def test_main(fake_checkpoint_dir, monkeypatch, version, tensor_like): |
| if version == "v1": |
| import litgpt.generate.adapter as generate |
| else: |
| import litgpt.generate.adapter_v2 as generate |
|
|
| config_path = fake_checkpoint_dir / "model_config.yaml" |
| config = {"block_size": 128, "vocab_size": 50, "n_layer": 2, "n_head": 4, "n_embd": 8, "rotary_percentage": 1} |
| config_path.write_text(yaml.dump(config)) |
|
|
| monkeypatch.setattr(generate, "lazy_load", Mock()) |
| monkeypatch.setattr(generate.GPT, "load_state_dict", Mock()) |
| tokenizer_mock = Mock() |
| tokenizer_mock.return_value.encode.return_value = torch.tensor([[1, 2, 3]]) |
| tokenizer_mock.return_value.decode.return_value = "### Response:foo bar baz" |
| monkeypatch.setattr(generate, "Tokenizer", tokenizer_mock) |
| generate_mock = Mock() |
| generate_mock.return_value = torch.tensor([[3, 2, 1]]) |
| monkeypatch.setattr(generate, "generate", generate_mock) |
|
|
| num_samples = 1 |
| out, err = StringIO(), StringIO() |
| with redirect_stdout(out), redirect_stderr(err): |
| generate.main(temperature=2.0, top_k=2, top_p=0.9, checkpoint_dir=fake_checkpoint_dir) |
|
|
| assert len(tokenizer_mock.return_value.decode.mock_calls) == num_samples |
| assert torch.allclose(tokenizer_mock.return_value.decode.call_args[0][0], generate_mock.return_value) |
| assert ( |
| generate_mock.mock_calls |
| == [call(ANY, tensor_like, 101, temperature=2.0, top_k=2, top_p=0.9, eos_id=ANY)] * num_samples |
| ) |
|
|
| expected_output = "foo bar baz\n" * num_samples |
| |
| pattern = rf".*^{re.escape(expected_output.strip())}$.*" |
| assert re.match(pattern, out.getvalue().strip(), re.DOTALL | re.MULTILINE) |
|
|
| err_value = err.getvalue() |
| expected_parts = [ |
| "'padded_vocab_size': 512", |
| "'n_layer': 2", |
| "'n_head': 4", |
| "'head_size': 2", |
| "'n_embd': 8", |
| ] |
| assert all(part in err_value for part in expected_parts) |
|
|
|
|
| @pytest.mark.parametrize("version", ("", "_v2")) |
| def test_cli(version): |
| args = ["litgpt", f"generate_adapter{version}", "-h"] |
| output = subprocess.check_output(args) |
| output = str(output.decode()) |
| assert "For models finetuned with" in output |
|
|