TM Forum Webinar - Telco API-driven digital marketplace opportunities | Post-...ShubaS4
If you missed the live webinar, you can catch all the details here in this presentation. Expert speakers Karthik TS and Dean Ramsay discussed CSP strategies for a new breed of marketplaces in this on-demand webinar. This slide deck provides a comprehensive overview of the LIVE webinar and is a great resource for CSPs looking for out-of-the-box API-driven digital marketplace solutions.
This year, the focus goes beyond technology to mining business insights around how cloud enables strategic industry trends such as Open and Virtual Banking and Insurance, Security and Compliance, Data Analytics and AI/ ML, FinTech and RegTech, Surveillance and more through sharing of best practices and use cases. In sessions led by customers, partners, industry leaders and AWS subject matter experts, you’ll learn how AWS helps financial institutions to focus on the innovation and outcomes that truly drive business forward. Business stakeholders, market makers, and technology owners will all learn something new, valuable and actionable.
Why Do Banks Need A Customer Data Platform?Lemnisk
Banks traditionally have been known to amass customer information across both online and offline data channels. However, a lot of this data resides in silos and marketers have been unable to leverage this data to run targeted marketing campaigns. Here are the top four reasons why a Customer Data Platform would be best suited for Banks.
Feature Store as a Data Foundation for Machine LearningProvectus
Looking to design and build a centralized, scalable Feature Store for your Data Science & Machine Learning teams to take advantage of? Come and learn from experts of Provectus and Amazon Web Services (AWS) how to!
Feature Store is a key component of the ML stack and data infrastructure, which enables feature engineering and management. By having a Feature Store, organizations can save massive amounts of resources, innovate faster, and drive ML processes at scale. In this webinar, you will learn how to build a Feature Store with a data mesh pattern and see how to achieve consistency between real-time and training features, to improve reproducibility with time-traveling for data.
Agenda
- Modern Data Lakes & Modern ML Infrastructure
- Existing and Emerging Architectural Shifts
- Feature Store: Overview and Reference Architecture
- AWS Perspective on Feature Store
Intended Audience
Technology executives & decision makers, manager-level tech roles, data architects & analysts, data engineers & data scientists, ML practitioners & ML engineers, and developers
Presenters
- Stepan Pushkarev, Chief Technology Officer, Provectus
- Gandhi Raketla, Senior Solutions Architect, AWS
- German Osin, Senior Solutions Architect, Provectus
Feel free to share this presentation with your colleagues and don't hesitate to reach out to us at info@provectus.com if you have any questions!
REQUEST WEBINAR: https://provectus.com/webinar-feature-store-as-data-foundation-for-ml-nov-2020/
Scaling and Modernizing Data Platform with DatabricksDatabricks
Today a Data Platform is expected to process and analyze a multitude of sources spanning batch files, streaming sources, backend databases, REST APIs, and more. There is clearly a need for standardizing the platform that scales and be flexible letting data engineers and data scientists focus on the business problems rather than managing the infrastructure and backend services. Another key aspect of the platform is multi-tenancy to isolate the workloads and able to track cost usage per tenant.
In this talk, Richa Singhal and Esha Shah will cover how to build a scalable Data Platform using Databricks and deploy your data pipelines effectively while managing the costs. The following topics will be covered:
Key tenets of a Data Platform
Setup multistage environment on Databricks
Build data pipelines locally and test on Databricks cluster
CI/CD for data pipelines with Databricks
Orchestrating pipelines using Apache Airflow – Change Data Capture using Databricks Delta
Leveraging Databricks Notebooks for Analytics and Data Science teams
TM Forum Webinar - Telco API-driven digital marketplace opportunities | Post-...ShubaS4
If you missed the live webinar, you can catch all the details here in this presentation. Expert speakers Karthik TS and Dean Ramsay discussed CSP strategies for a new breed of marketplaces in this on-demand webinar. This slide deck provides a comprehensive overview of the LIVE webinar and is a great resource for CSPs looking for out-of-the-box API-driven digital marketplace solutions.
This year, the focus goes beyond technology to mining business insights around how cloud enables strategic industry trends such as Open and Virtual Banking and Insurance, Security and Compliance, Data Analytics and AI/ ML, FinTech and RegTech, Surveillance and more through sharing of best practices and use cases. In sessions led by customers, partners, industry leaders and AWS subject matter experts, you’ll learn how AWS helps financial institutions to focus on the innovation and outcomes that truly drive business forward. Business stakeholders, market makers, and technology owners will all learn something new, valuable and actionable.
Why Do Banks Need A Customer Data Platform?Lemnisk
Banks traditionally have been known to amass customer information across both online and offline data channels. However, a lot of this data resides in silos and marketers have been unable to leverage this data to run targeted marketing campaigns. Here are the top four reasons why a Customer Data Platform would be best suited for Banks.
Feature Store as a Data Foundation for Machine LearningProvectus
Looking to design and build a centralized, scalable Feature Store for your Data Science & Machine Learning teams to take advantage of? Come and learn from experts of Provectus and Amazon Web Services (AWS) how to!
Feature Store is a key component of the ML stack and data infrastructure, which enables feature engineering and management. By having a Feature Store, organizations can save massive amounts of resources, innovate faster, and drive ML processes at scale. In this webinar, you will learn how to build a Feature Store with a data mesh pattern and see how to achieve consistency between real-time and training features, to improve reproducibility with time-traveling for data.
Agenda
- Modern Data Lakes & Modern ML Infrastructure
- Existing and Emerging Architectural Shifts
- Feature Store: Overview and Reference Architecture
- AWS Perspective on Feature Store
Intended Audience
Technology executives & decision makers, manager-level tech roles, data architects & analysts, data engineers & data scientists, ML practitioners & ML engineers, and developers
Presenters
- Stepan Pushkarev, Chief Technology Officer, Provectus
- Gandhi Raketla, Senior Solutions Architect, AWS
- German Osin, Senior Solutions Architect, Provectus
Feel free to share this presentation with your colleagues and don't hesitate to reach out to us at info@provectus.com if you have any questions!
REQUEST WEBINAR: https://provectus.com/webinar-feature-store-as-data-foundation-for-ml-nov-2020/
Scaling and Modernizing Data Platform with DatabricksDatabricks
Today a Data Platform is expected to process and analyze a multitude of sources spanning batch files, streaming sources, backend databases, REST APIs, and more. There is clearly a need for standardizing the platform that scales and be flexible letting data engineers and data scientists focus on the business problems rather than managing the infrastructure and backend services. Another key aspect of the platform is multi-tenancy to isolate the workloads and able to track cost usage per tenant.
In this talk, Richa Singhal and Esha Shah will cover how to build a scalable Data Platform using Databricks and deploy your data pipelines effectively while managing the costs. The following topics will be covered:
Key tenets of a Data Platform
Setup multistage environment on Databricks
Build data pipelines locally and test on Databricks cluster
CI/CD for data pipelines with Databricks
Orchestrating pipelines using Apache Airflow – Change Data Capture using Databricks Delta
Leveraging Databricks Notebooks for Analytics and Data Science teams
Toyota Financial Services Digital Transformation - Think 2019Slobodan Sipcic
Toyota Financial Services (TFS) and IBM partnered to develop Data & Integration Platform (D&IP) to be the hub around which all current and future TFS data sources, services, and processes interact. To that end IBM have architected and deployed a FOAK event-based data stream processing and streaming integration platform. The main components of the architecture include: Kubernetes, Apache NiFi, Apache Kafka, Schema Registry, Jenkins, S3 and MongoDB. The platform is essential for realizing the TFS' strategic data stream processing and integration needs.
This presentation shared experiences and lessons learned in managing a global help desk for the Data for Accountability, Transparency and Impact (DATIM) information system of the U.S. President’s Emergency Plan for AIDS Relief (PEPFAR). DATIM is built on the DHIS 2 platform.
Proof of concepts and use cases with IoT technologiesHeikki Ailisto
Set of proof of concept and use cases with internet of things technologies are presented with one sliders. In each case, the IoT challenge, result, benefits and use case example are given.
From Insights to Action, How to build and maintain a Data Driven Organization...Amazon Web Services Korea
데이터는 혁신과 변혁의 토대입니다. 비즈니스 혁신을 이끄는 혁신은 특정 시점의 전략이나 솔루션이 아니라 성장을 위한 반복적이고 집단적인 계획입니다. 혁신에 이러한 접근 방식을 채택하는 기업은 전략과 비즈니스 문화에서 데이터를 기반으로 하는 경우가 많습니다. 이러한 접근 방식을 개발하려면 리더가 데이터를 조직의 자산처럼 취급하고 조직이 더 나은 비즈니스 성과를 위해 데이터를 활용할 수 있도록 권한을 부여해야 합니다. AWS와 Amazon이 어떻게 데이터와 분석을 활용하여 확장 가능한 비즈니스 효율성을 창출하고 고객의 가장 복잡한 문제를 해결하는 메커니즘을 개발했는지 알아보십시오.
Evolution from EDA to Data Mesh: Data in Motionconfluent
Thoughtworks Zhamak Dehghani observations on these traditional approaches’s failure modes, inspired her to develop an alternative big data management architecture that she aptly named the Data Mesh. This represents a paradigm shift that draws from modern distributed architecture and is founded on the principles of domain-driven design, self-serve platform, and product thinking with Data. In the last decade Apache Kafka has established a new category of data management infrastructure for data in motion that has been leveraged in modern distributed data architectures.
Transform your marketing and sales capabilities with Big Data and A.I
1) Why is Customer Data Platform (CDP) ?
Case study: Enhancing the revenue of your restaurant with CDP and mobile app marketing
Question: Why can CDP disrupt business model for restaurant industry (B2C) ?
2) How would CDP work in practice ?
Introducing USPA.tech as logical framework for implementing CDP in practice
How Can a Customer Data Platform Enhance Your Account-Based Marketing Strategy (B2B) ?
3) How can we implement CDP for business?
Introducing the CDP as customer-first marketing platform for all industries (my key idea in this slide)
Snowflake concepts & hands on expertise to help get you started on implementing Data warehouses using Snowflake. Necessary information and skills that will help you master Snowflake essentials.
Flexible and Scalable Integration in the Automation Industry/Industrial IoTconfluent
Speaker: Kai Waehner, Technology Evangelist, Confluent
Kafka-Native, End-to-End IIoT Data Integration and Processing with Kafka Connect, KSQL, and PLC4X
Nubank is using machine learning, analytics and data engineering to disrupt the financial industry in Brazil. With over 5 million customers, it's already the biggest fintech outside Asia.
In this presentation I go over the main learnings of the company in the data space since it was founded 5 years ago. How did the company look like on each of those years? What mistakes did we make? What lessons did we learn?
This deck was presented in the São Paulo Product.io Meetup
What you need to know about Generative AI and Data Management?Denodo
Watch full webinar here: https://buff.ly/3UXy0A2
It should be no surprise that Generative AI will have a profound impact to data management in years to come. Much like other areas of the technology sector, the opportunities presented by GenAI will accelerate our efforts around all aspects of data management, including self-service, automation, data governance and security. On the other hand, it is also becoming clearer that to unleash the true potential of AI assistants powered by GenAI, we need novel implementation strategies and a reimagined data architecture. This presents an exhilarating yet challenging future, demanding innovative thinking and methodologies in data management.
Join us on this webinar to learn about:
- The opportunities and challenges presented by GenAI today.
- Exploiting GenAI to democratize data management.
- How to augment GenAI applications with corporate data and knowledge.
- How to get started.
The Future of Data Science and Machine Learning at Scale: A Look at MLflow, D...Databricks
Many had dubbed 2020 as the decade of data. This is indeed an era of data zeitgeist.
From code-centric software development 1.0, we are entering software development 2.0, a data-centric and data-driven approach, where data plays a central theme in our everyday lives.
As the volume and variety of data garnered from myriad data sources continue to grow at an astronomical scale and as cloud computing offers cheap computing and data storage resources at scale, the data platforms have to match in their abilities to process, analyze, and visualize at scale and speed and with ease — this involves data paradigm shifts in processing and storing and in providing programming frameworks to developers to access and work with these data platforms.
In this talk, we will survey some emerging technologies that address the challenges of data at scale, how these tools help data scientists and machine learning developers with their data tasks, why they scale, and how they facilitate the future data scientists to start quickly.
In particular, we will examine in detail two open-source tools MLflow (for machine learning life cycle development) and Delta Lake (for reliable storage for structured and unstructured data).
Other emerging tools such as Koalas help data scientists to do exploratory data analysis at scale in a language and framework they are familiar with as well as emerging data + AI trends in 2021.
You will understand the challenges of machine learning model development at scale, why you need reliable and scalable storage, and what other open source tools are at your disposal to do data science and machine learning at scale.
Los bots suelen imitar o reemplazar el comportamiento humano del usuario. Debido a que están automatizados, operan mucho más rápido que los usuarios humanos. Llevan a cabo funciones útiles, como el servicio al cliente o los motores de búsqueda de indexación, pero también pueden venir en forma de malware, utilizado para obtener un control total sobre una computadora. Entonces, en este articulo te mostrare paso a paso como construir un bot con QnA Maker service, Azure Bot service y el servicio de mensajería de Telegram.
Toyota Financial Services Digital Transformation - Think 2019Slobodan Sipcic
Toyota Financial Services (TFS) and IBM partnered to develop Data & Integration Platform (D&IP) to be the hub around which all current and future TFS data sources, services, and processes interact. To that end IBM have architected and deployed a FOAK event-based data stream processing and streaming integration platform. The main components of the architecture include: Kubernetes, Apache NiFi, Apache Kafka, Schema Registry, Jenkins, S3 and MongoDB. The platform is essential for realizing the TFS' strategic data stream processing and integration needs.
This presentation shared experiences and lessons learned in managing a global help desk for the Data for Accountability, Transparency and Impact (DATIM) information system of the U.S. President’s Emergency Plan for AIDS Relief (PEPFAR). DATIM is built on the DHIS 2 platform.
Proof of concepts and use cases with IoT technologiesHeikki Ailisto
Set of proof of concept and use cases with internet of things technologies are presented with one sliders. In each case, the IoT challenge, result, benefits and use case example are given.
From Insights to Action, How to build and maintain a Data Driven Organization...Amazon Web Services Korea
데이터는 혁신과 변혁의 토대입니다. 비즈니스 혁신을 이끄는 혁신은 특정 시점의 전략이나 솔루션이 아니라 성장을 위한 반복적이고 집단적인 계획입니다. 혁신에 이러한 접근 방식을 채택하는 기업은 전략과 비즈니스 문화에서 데이터를 기반으로 하는 경우가 많습니다. 이러한 접근 방식을 개발하려면 리더가 데이터를 조직의 자산처럼 취급하고 조직이 더 나은 비즈니스 성과를 위해 데이터를 활용할 수 있도록 권한을 부여해야 합니다. AWS와 Amazon이 어떻게 데이터와 분석을 활용하여 확장 가능한 비즈니스 효율성을 창출하고 고객의 가장 복잡한 문제를 해결하는 메커니즘을 개발했는지 알아보십시오.
Evolution from EDA to Data Mesh: Data in Motionconfluent
Thoughtworks Zhamak Dehghani observations on these traditional approaches’s failure modes, inspired her to develop an alternative big data management architecture that she aptly named the Data Mesh. This represents a paradigm shift that draws from modern distributed architecture and is founded on the principles of domain-driven design, self-serve platform, and product thinking with Data. In the last decade Apache Kafka has established a new category of data management infrastructure for data in motion that has been leveraged in modern distributed data architectures.
Transform your marketing and sales capabilities with Big Data and A.I
1) Why is Customer Data Platform (CDP) ?
Case study: Enhancing the revenue of your restaurant with CDP and mobile app marketing
Question: Why can CDP disrupt business model for restaurant industry (B2C) ?
2) How would CDP work in practice ?
Introducing USPA.tech as logical framework for implementing CDP in practice
How Can a Customer Data Platform Enhance Your Account-Based Marketing Strategy (B2B) ?
3) How can we implement CDP for business?
Introducing the CDP as customer-first marketing platform for all industries (my key idea in this slide)
Snowflake concepts & hands on expertise to help get you started on implementing Data warehouses using Snowflake. Necessary information and skills that will help you master Snowflake essentials.
Flexible and Scalable Integration in the Automation Industry/Industrial IoTconfluent
Speaker: Kai Waehner, Technology Evangelist, Confluent
Kafka-Native, End-to-End IIoT Data Integration and Processing with Kafka Connect, KSQL, and PLC4X
Nubank is using machine learning, analytics and data engineering to disrupt the financial industry in Brazil. With over 5 million customers, it's already the biggest fintech outside Asia.
In this presentation I go over the main learnings of the company in the data space since it was founded 5 years ago. How did the company look like on each of those years? What mistakes did we make? What lessons did we learn?
This deck was presented in the São Paulo Product.io Meetup
What you need to know about Generative AI and Data Management?Denodo
Watch full webinar here: https://buff.ly/3UXy0A2
It should be no surprise that Generative AI will have a profound impact to data management in years to come. Much like other areas of the technology sector, the opportunities presented by GenAI will accelerate our efforts around all aspects of data management, including self-service, automation, data governance and security. On the other hand, it is also becoming clearer that to unleash the true potential of AI assistants powered by GenAI, we need novel implementation strategies and a reimagined data architecture. This presents an exhilarating yet challenging future, demanding innovative thinking and methodologies in data management.
Join us on this webinar to learn about:
- The opportunities and challenges presented by GenAI today.
- Exploiting GenAI to democratize data management.
- How to augment GenAI applications with corporate data and knowledge.
- How to get started.
The Future of Data Science and Machine Learning at Scale: A Look at MLflow, D...Databricks
Many had dubbed 2020 as the decade of data. This is indeed an era of data zeitgeist.
From code-centric software development 1.0, we are entering software development 2.0, a data-centric and data-driven approach, where data plays a central theme in our everyday lives.
As the volume and variety of data garnered from myriad data sources continue to grow at an astronomical scale and as cloud computing offers cheap computing and data storage resources at scale, the data platforms have to match in their abilities to process, analyze, and visualize at scale and speed and with ease — this involves data paradigm shifts in processing and storing and in providing programming frameworks to developers to access and work with these data platforms.
In this talk, we will survey some emerging technologies that address the challenges of data at scale, how these tools help data scientists and machine learning developers with their data tasks, why they scale, and how they facilitate the future data scientists to start quickly.
In particular, we will examine in detail two open-source tools MLflow (for machine learning life cycle development) and Delta Lake (for reliable storage for structured and unstructured data).
Other emerging tools such as Koalas help data scientists to do exploratory data analysis at scale in a language and framework they are familiar with as well as emerging data + AI trends in 2021.
You will understand the challenges of machine learning model development at scale, why you need reliable and scalable storage, and what other open source tools are at your disposal to do data science and machine learning at scale.
Los bots suelen imitar o reemplazar el comportamiento humano del usuario. Debido a que están automatizados, operan mucho más rápido que los usuarios humanos. Llevan a cabo funciones útiles, como el servicio al cliente o los motores de búsqueda de indexación, pero también pueden venir en forma de malware, utilizado para obtener un control total sobre una computadora. Entonces, en este articulo te mostrare paso a paso como construir un bot con QnA Maker service, Azure Bot service y el servicio de mensajería de Telegram.
MuleSoft Meetup #5 de Ciudad de Panamá.
Mejoras y nuevas funcionalidades de la Anypoint Platform en sus releases del año 2019 incluyendo el reciente release de Mayo 2019 presentado en el MuleSoft CONNECT de Atlanta.
También se habla sobre el tema de monetización de APIs y la API Economy según el modelo de MuleSoft.
Presentación Corporativa de Tecnitek IT Solutions. Consultora tecnológica, desarrollo web, desarrollo software, marketing online,telefonia IP, administracion de sistemas
(PROYECTO) Límites entre el Arte, los Medios de Comunicación y la Informáticavazquezgarciajesusma
En este proyecto de investigación nos adentraremos en el fascinante mundo de la intersección entre el arte y los medios de comunicación en el campo de la informática.
La rápida evolución de la tecnología ha llevado a una fusión cada vez más estrecha entre el arte y los medios digitales, generando nuevas formas de expresión y comunicación.
Continuando con el desarrollo de nuestro proyecto haremos uso del método inductivo porque organizamos nuestra investigación a la particular a lo general. El diseño metodológico del trabajo es no experimental y transversal ya que no existe manipulación deliberada de las variables ni de la situación, si no que se observa los fundamental y como se dan en su contestó natural para después analizarlos.
El diseño es transversal porque los datos se recolectan en un solo momento y su propósito es describir variables y analizar su interrelación, solo se desea saber la incidencia y el valor de uno o más variables, el diseño será descriptivo porque se requiere establecer relación entre dos o más de estás.
Mediante una encuesta recopilamos la información de este proyecto los alumnos tengan conocimiento de la evolución del arte y los medios de comunicación en la información y su importancia para la institución.
Actualmente, y debido al desarrollo tecnológico de campos como la informática y la electrónica, la mayoría de las bases de datos están en formato digital, siendo este un componente electrónico, por tanto se ha desarrollado y se ofrece un amplio rango de soluciones al problema del almacenamiento de datos.
3Redu: Responsabilidad, Resiliencia y Respetocdraco
¡Hola! Somos 3Redu, conformados por Juan Camilo y Cristian. Entendemos las dificultades que enfrentan muchos estudiantes al tratar de comprender conceptos matemáticos. Nuestro objetivo es brindar una solución inclusiva y accesible para todos.
6. ¿Qué es un Bot?
Un Bot es un software de
Inteligencia Artificial que
simula una conversación con
una persona por medio del
lenguaje natural.
Es considerado como una
de las expresiones de
interacción entre humanos y
máquinas más avanzadas.
14. Power Plataforms
Una capacidad de IA de código bajo que abarca Power Platform, Dynamics 365
Data
connectors
Common
data service
AI Builder
Dynamics 365
Business applications
Power Automate
Automatización del flujo de trabajo
PowerApps
Desarrollo de aplicaciones
Power Virtual Agent
ChatBots
Power BI
Análisis de negocio