Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .github/workflows/import.yml
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ jobs:
- name: Install dependencies
run: |
python -m pip install --upgrade pip
pip install setuptools wheel
pip install setuptools wheel sentencepiece protobuf
pip install pytest
pip install .

Expand Down
3 changes: 2 additions & 1 deletion google/genai/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,9 @@
from . import types
from . import version
from .client import Client
from .local_tokenizer import LocalTokenizer


__version__ = version.__version__

__all__ = ['Client']
__all__ = ["Client", "LocalTokenizer"]
4 changes: 2 additions & 2 deletions google/genai/_local_tokenizer_loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,8 +22,8 @@
import uuid

import requests # type: ignore
import sentencepiece as spm
from sentencepiece import sentencepiece_model_pb2
from google3.third_party.sentencepiece.src.python import sentencepiece_processor as spm
from google3.third_party.sentencepiece.src import sentencepiece_model_pb2


# Source of truth: https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models
Expand Down
2 changes: 1 addition & 1 deletion google/genai/local_tokenizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@
from typing import Any, Iterable
from typing import Optional, Union

from sentencepiece import sentencepiece_model_pb2
from google3.third_party.sentencepiece.src import sentencepiece_model_pb2

from . import _common
from . import _local_tokenizer_loader as loader
Expand Down
4 changes: 2 additions & 2 deletions google/genai/tests/local_tokenizer/test_local_tokenizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,8 @@

from sentencepiece import sentencepiece_model_pb2

from ... import local_tokenizer
from ... import types
from ...private import local_tokenizer
from ...private import types


class TestLocalTokenizer(unittest.TestCase):
Expand Down
Loading