In this workshop, you’ll learn an easy way to incorporate data science and AI/ML into an OpenShift development workflow. As an example, you’ll use an object detection model to detect ‘dog(s)’ in an image. You will: Use Jupyter Notebooks and TensorFlow to explore a pre-trained object detection model Serve the model in a REST API as a Flask App Use Source-to-Image (S2I) to build and deploy the Flask app Explore Kafka streams from Notebooks Deploy a Kafka consumer with the same object detection model You’ll be able to do all of this without having to install anything on your own computer, thanks to Red Hat OpenShift Data Science and Red Hat OpenShift Streams for Apache Kafka. Note: Beginner data handling and Python skills are required for this workshop.