diff --git a/examples/tensorflow/compositing.py b/examples/tensorflow/compositing.py index af2d51f0..eb6d11d9 100755 --- a/examples/tensorflow/compositing.py +++ b/examples/tensorflow/compositing.py @@ -9,7 +9,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from PIL import Image from picamera2 import MappedArray, Picamera2, Preview @@ -52,7 +52,7 @@ def InferenceTensorFlow(image, model, label=None): else: labels = None - interpreter = tflite.Interpreter(model_path=model, num_threads=4) + interpreter = Interpreter(model_path=model, num_threads=4) interpreter.allocate_tensors() input_details = interpreter.get_input_details() diff --git a/examples/tensorflow/real_time.py b/examples/tensorflow/real_time.py index 0495adb9..367e7b7d 100755 --- a/examples/tensorflow/real_time.py +++ b/examples/tensorflow/real_time.py @@ -9,13 +9,10 @@ # Install necessary dependences before starting, # # $ sudo apt update -# $ sudo apt install build-essential -# $ sudo apt install libatlas-base-dev # $ sudo apt install python3-pip -# $ pip3 install tflite-runtime -# $ pip3 install opencv-python==4.4.0.46 +# $ sudo apt install python3-opencv +# $ pip3 install ai-edge-litert # $ pip3 install pillow -# $ pip3 install numpy # # and run from the command line, # @@ -25,7 +22,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from picamera2 import MappedArray, Picamera2, Preview @@ -61,7 +58,7 @@ def InferenceTensorFlow(image, model, output, label=None): else: labels = None - interpreter = tflite.Interpreter(model_path=model, num_threads=4) + interpreter = Interpreter(model_path=model, num_threads=4) interpreter.allocate_tensors() input_details = interpreter.get_input_details() diff --git a/examples/tensorflow/real_time_with_labels.py b/examples/tensorflow/real_time_with_labels.py index e7854ee7..c1938ed2 100755 --- a/examples/tensorflow/real_time_with_labels.py +++ b/examples/tensorflow/real_time_with_labels.py @@ -9,13 +9,10 @@ # Install necessary dependences before starting, # # $ sudo apt update -# $ sudo apt install build-essential -# $ sudo apt install libatlas-base-dev # $ sudo apt install python3-pip -# $ pip3 install tflite-runtime -# $ pip3 install opencv-python==4.4.0.46 +# $ sudo apt install python3-opencv +# $ pip3 install ai-edge-litert # $ pip3 install pillow -# $ pip3 install numpy # # and run from the command line, # @@ -25,7 +22,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from picamera2 import MappedArray, Picamera2, Preview @@ -67,7 +64,7 @@ def InferenceTensorFlow(image, model, output, label=None): else: labels = None - interpreter = tflite.Interpreter(model_path=model, num_threads=4) + interpreter = Interpreter(model_path=model, num_threads=4) interpreter.allocate_tensors() input_details = interpreter.get_input_details() diff --git a/examples/tensorflow/remove_background.py b/examples/tensorflow/remove_background.py index c01eb468..723519f7 100755 --- a/examples/tensorflow/remove_background.py +++ b/examples/tensorflow/remove_background.py @@ -6,7 +6,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from PIL import Image from picamera2 import Picamera2, Preview @@ -21,7 +21,7 @@ def InferenceTensorFlow(image, model): global background_mask - interpreter = tflite.Interpreter(model_path=model, num_threads=4) + interpreter = Interpreter(model_path=model, num_threads=4) interpreter.allocate_tensors() input_details = interpreter.get_input_details() diff --git a/examples/tensorflow/segmentation.py b/examples/tensorflow/segmentation.py index d0ca7268..e783fc2a 100755 --- a/examples/tensorflow/segmentation.py +++ b/examples/tensorflow/segmentation.py @@ -9,7 +9,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from PIL import Image from picamera2 import Picamera2, Preview @@ -40,7 +40,7 @@ def InferenceTensorFlow(image, model, colours, label=None): else: labels = None - interpreter = tflite.Interpreter(model_path=model, num_threads=4) + interpreter = Interpreter(model_path=model, num_threads=4) interpreter.allocate_tensors() input_details = interpreter.get_input_details() diff --git a/examples/tensorflow/yolo_v5_real_time_with_labels.py b/examples/tensorflow/yolo_v5_real_time_with_labels.py index a73264f7..12d2248b 100644 --- a/examples/tensorflow/yolo_v5_real_time_with_labels.py +++ b/examples/tensorflow/yolo_v5_real_time_with_labels.py @@ -15,7 +15,7 @@ # Use system python or if you prefer not to mess with system python: # install a version manager (like pyenv) and use 'pyenv virtualenv --system-site-packages ENV_NAME' -# $ pip3 install tflite-runtime +# $ pip3 install ai-edge-litert # $ pip3 install opencv-python-headless (if using system python: sudo apt install python3-opencv) # # and run from the command line, @@ -26,7 +26,7 @@ import cv2 import numpy as np -import tflite_runtime.interpreter as tflite +from ai_edge_litert.interpreter import Interpreter from picamera2 import MappedArray, Picamera2, Platform, Preview @@ -103,7 +103,7 @@ def main(): picam2.post_callback = DrawRectangles picam2.start() - interpreter = tflite.Interpreter(model_path=args.model, num_threads=4) + interpreter = Interpreter(model_path=args.model, num_threads=4) interpreter.allocate_tensors() while True: