If you’re like most people shifting to a cloud strategy, your learning curve is steep, and even the basics such as migrating data to the cloud are worrisome. Whether megabytes or petabytes, you need your cloud migration to be an efficient process - continuous, requiring no downtime, and seamless without data loss. The long delays inherent with traditional migration methods, the need for coordinated ingestion across multiple environments and various geolocations, stale, vulnerable, or inconsistent data often leave projects stuck in the starting gate. Even worse, when your business’s data is driving competitive advantage, unless you can be sure of the availability of your data, you and your business are at serious risk. Now it is possible to have guaranteed consistent data everywhere, even when data is growing and changing. This presentation will explain ‘LIVE DATA’. And present architecture and solutions that show how a ‘LIVE DATA strategy’ allows you to surmount common and sometimes scary cloud use-cases across a wide variety of industry scenarios.
15+ years Management/Executive leadership. Ran Pre-Sales Systems/Sales Engineering and Professional Service teams for Enterprise Hardware and Software solutions. The last 5 years the focus is on Big Data - modern data warehouse and analytics platforms in the Cloud.
Tuesday March 20, 2018 10:10am - 11:00am PDT
2-Rm 102
Big Data Analytics requires unstructured data (a.k.a. Big Data) but also Master data generally stored in RDBMS tables.
Apache Spark furnishes the Spark SQL interface for querying unstructured data using SQL. This interface provisions the Datasource API with a JDBC interface for accessing relational databases. How to join RDBMS tables with Big data using Spark SQL without moving data over?
This session describes an implementation of the Spark Datasource API and explains the optimizations that (i) allow parallel and direct access to the RDBMS database (with the option of controlling the number of concurrent connections); (ii) introspect the RDBMS table, generate partitions of Spark JDBCRDDs based on the split pattern and rewrite Spark SQL queries into the RDBMS SQL; (iii) use hooks in the JDBC driver for faster type conversions; (iv) push down predicates to the RDBMS, prune partitions based on the where clause to reduce the amount of data returned to the RDBMS.
Kuassi is Director of Product Management at Oracle. He looks after the following product areas (i) Java connectivity to DB (Cloud, on-premises), in-place processing with DB embedded JVM (ii) MicroServices and DB connectivity, and related topics (Data & Tx models, Kubernetes, SAGAs... Read More →
Tuesday March 20, 2018 11:15am - 12:05pm PDT
5-Rm 105
Customer 360 degree views require data from a variety of streaming sources to be integrated with data stored in Big Data systems and databases. Machine learning and analytics applications need access to mobile device feeds, online community logs, social media feeds, and much more, along with business critical data stored in the database. Tools for data movement between streaming data sources, Big Data platforms and Oracle Databases are essential. In this session, learn about tools that move data between these systems, using technology such as Hive storage handlers for Kafka, direct path load, SQL, and custom Hive SerDes. We will step through using these tools to work with Streaming data (Kafka), Big Data, and Database services in Oracle Cloud Infrastructure.
Melli Annamalai is a product manager at Oracle. She has vast experience in multiple technology areas related to unstructured and semi-structured data management. Her current focus areas are Graphs, Apache Kafka, and Big Data. She works closely with customers as they deploy solutions... Read More →
Tuesday March 20, 2018 2:30pm - 3:20pm PDT
4-Rm 104
Data is at the heart of Product organization's decision making process. Yet, only a few pieces of valuable product usage data are deployed for financials analytics. In this session learn how RecVue processes trillions of product usage data points with Oracle Bigdata cloud and build analytics to monetize product usage, reduce customer churn and retain customer success.
Anto is a technologist, innovator, and product strategist. He leads the engineering and operations teams at RecVue. His teams are responsible for delivering RecVue (SaaS), including the user interface, applications and frameworks. He believes deeply in building technology products... Read More →
Tuesday March 20, 2018 3:35pm - 4:05pm PDT
2-Rm 102
Don't miss the Architecture Live experience! In this interactive session, you'll witness two industry experts digitally illustrating data-driven architectures live, with input and feedback from the audience.
Kafka, Lambda, and Streaming Analytics will be all covered. We’ll tell you what these words mean, and more importantly how they affect the choices we make building an enterprise architecture. With the Oracle Information Management Reference Architecture as the backdrop, we'll clarify and delineate the different components involved in delivering big data, fast data, and all the gray area in between. The Architecture Live experience will be fun and different, and we'll all learn something along the way.
Stewart is the Owner and Co-founder of Red Pill Analytics, and has been designing and implementing BI systems since 1996. He is an Oracle ACE Director, is an internationally recognized writer and speaker in the Oracle Community, and co-host for the audio and video podcast See Data... Read More →
Highly experienced product manager with 15 years of experience in enterprise software and enterprise data. Currently responsible for all product management aspects (technology, GTM, Enablement etc.) of Oracle Big Data Appliance and Big Data Cloud Service at Oracle.Active interest... Read More →
Tuesday March 20, 2018 4:20pm - 5:10pm PDT
1-Auditorium
Hadoop, through schema on read offers a tremendous advantage in analytics: flexibility to work with data before it is completely modeled and loaded, enabling analysis to start much sooner. But, as we learned from Hive in Hadoop, it does come with relatively poor performance. This session will unveil and explain technological innovations from Oracle, addressing these performance issues. We will share how we achieved parquet query performance with the flexibility of schema on read enabling the best of both worlds while fitting comfortably within the open source community. On top of that, this session addresses further research into methods of securing data. Security will be done at a much more granular level than possible today. If you are interested in the latest innovations around Hadoop, Object Stores, performance and security, you should not miss this session.
Highly experienced product manager with 15 years of experience in enterprise software and enterprise data. Currently responsible for all product management aspects (technology, GTM, Enablement etc.) of Oracle Big Data Appliance and Big Data Cloud Service at Oracle.Active interest... Read More →
Wednesday March 21, 2018 9:00am - 9:50am PDT
1-Auditorium
This session providers an overview of Oracle's modern, cloud native big data 2.0 platform as a cloud service. Learn how various technological innovations within the portfolio of Big Data Cloud forms big data foundation services-around data integration, data management and analytics. Also learn about Oracle Event Hub Cloud that is offered as cloud native Kafka as a managed service, that his low latency and high performance streaming platform in the cloud.
Over 17 Years of Product Development experience building Enterprise Technology Solutions - Analytics & Data Platforms, Big Data, Distributed Systems and Cloud Computing (public/private/hybrid). Functions:- Product Management & Strategy, Product Engineering and Business Develop... Read More →
Wednesday March 21, 2018 10:05am - 10:55am PDT
2-Rm 102
Limited Capacityfull Adding this to your schedule will put you on the waitlist.
Organizations are capturing data in a variety of data stores including Oracle Database, NoSQL, and Hadoop. Oracle Big Data SQL allows you to use Oracle SQL and security policies to seamlessly query across these stores. This means that existing Oracle Database query applications—without changes—will now be able to take advantage of this new information. This hands-on lab introduces you to how this is achieved: providing access to big data sources, implementing unified security, and delivering analyses that combine and correlate all data.
I'm part of Oracle's big data product management team - with a key focus on Big Data SQL. I have a long history at Oracle - starting in the sales consulting organization - and then working in the BI, OLAP and Big Data PM groups
Wednesday March 21, 2018 11:10am - 1:50pm PDT
6-Bldg 500
Criminal analysis and prediction involves using many, varied data sources with various data formats. A city level public security department in China looked to Oracle's big data technologies to address their business requirements. We delivered this project to our customer using Oracle Database 11.2, Big Data Spatial and Graph 2.2, BIEE 12c, Data Visualization, Big Data Discovery 1.3.2 products. We integrated various data including emergency call (911), demographic, hotel bookings, criminal records, live traffic, and other data sets, to perform accommodation analysis, criminal prediction analysis, correlation analysis, and operation efficiency dashboard and reporting. The property graph features in Big Data Spatial and Graph were key to performing these analytics, and the solution integrated Oracle's big data technology into one platform.
Lending Club’s Big Data Analytics platform is going through a major transformation to adopt the Lambda Architecture with Real-Time reporting and execution of AI/ML models on Real-Time data. Lending Club's OLTP platform is transitioning to a Microservice-based architecture. At the heart of both solutions is GoldenGate for BigData. We are using a single server configured to run the replicats serving JSON payloads and Avro formats using same set of trail files. We implemented a custom java replicat to stream JSON messages to AWS Kinesis. We are using HDFS handler to stream data in Avro format to HDFS and AWS S3 to be consumed by AWS EMR, Hive on Tez/Presto/Spark. The solutions have been architected in a secure, reliable, and resilient manner.
Venky Guruswamy is a seasoned database professional with over 15 years of progressive experience in various tech domains of banking, financial, retail, ecommerce primarily managing Oracle, Mysql, MongoDB ,Teradata databases. He is currently the Principal DBA at Lending Club leading... Read More →
Independent of the source of data, the integration and analysis of event streams gets more important in the world of sensors, social media streams and Internet of Things. Events have to be accepted quickly and reliably, they have to be distributed and analyzed, often with many consumers or systems interested in all or part of the events. In this session we compare two popular Streaming Analytics solutions: Spark Streaming and Kafka Streams. Spark is fast and general engine for large-scale data processing and has been designed to provide a more efficient alternative to Hadoop MapReduce. Spark Streaming brings Spark's language-integrated API to stream processing, letting you write streaming applications the same way you write batch jobs. It supports both Java and Scala. Kafka Streams is the stream processing solution which is part of Kafka. It is provided as a Java library and by that can be easily integrated with any Java application.
Guido Schmutz works for the Oracle Platinum Partner Trivadis. He has more than 30 years of technology experience. At Trivadis he leads the Trivadis Architecture Board. He has long-time experience as developer, coach, trainer, and architect in the area of building IT systems. In the... Read More →
Wednesday March 21, 2018 2:20pm - 3:10pm PDT
5-Rm 105
There are zillions of Open Source big data projects these days. However, in this session you will learn about the core principles of 4 key technologies that are most often used in projects: Hadoop, Spark, Hive, and Kafka. We first explain the fundamentals of those 4 big data technologies. Then you will see how to make the first easy steps into the big data world yourself with Oracle Big Data Cloud Service Compute Edition and Oracle Event Hub Cloud Service live demos.
Frank Munz is an expert in middleware and distributed computing. He earned a Ph.D. in computer science from the Technische Universität München with summa cum laude for his work on distributed computing and brain research. He published more than 20 peer reviewed scientific papers... Read More →
Wednesday March 21, 2018 3:25pm - 4:15pm PDT
4-Rm 104
You may have heard that Oracle Big Data SQL allows applications to query data across Oracle Database, Hadoop and NoSQL stores. This session will step you thru its architecture. You'll learn how Big Data SQL fits into the Oracle Database and Hadoop infrastructure, how it leverages standards for data access and how it optimizes query execution to achieve good performance. The session will help you understand how you can leverage your Oracle Database investment and effectively augment your analyses with data from across the big data stack.
I'm part of Oracle's big data product management team - with a key focus on Big Data SQL. I have a long history at Oracle - starting in the sales consulting organization - and then working in the BI, OLAP and Big Data PM groups
Thursday March 22, 2018 12:00pm - 12:30pm PDT
1-Auditorium
Traditional Business intelligence (BI) has delivered a great value to organizations, enabling them to better data-based decision making. However, today the speed to make business decisions has outpaced the traditional BI's ability to adapt and answer quickly new questions.
In this talk we introduce a very powerful engine like Apache Solr and demonstrate how easy it is to set it up and start getting answers right away. Since a picture is better than a thousand words we will also introduce Banana (a fork of Kibana) as its graphical front end for data analysis.
Finally, We will also discuss about Solr Analytics 2.0 which, provides the ability to search on live data perfectly complementing the use of a traditional data warehouse database.
Over 16 years of experience in all phases of Architecting, implementing, and supporting BI and Data Warehousing solutions for customers in the USA, Europe and Latin America. Worked with many industries including Banking, Insurance, Telco, and Retail.
Highly experienced product manager with 15 years of experience in enterprise software and enterprise data. Currently responsible for all product management aspects (technology, GTM, Enablement etc.) of Oracle Big Data Appliance and Big Data Cloud Service at Oracle.Active interest... Read More →
Thursday March 22, 2018 2:30pm - 3:20pm PDT
4-Rm 104
While relational databases are still the king of transactional processing systems, they have a hard time keeping up with the increasing demand for real-time analytics. In this session we will build and demonstrate an end-to-end data processing pipeline. We will discuss how to turn changes in database state into events and stream them into Apache Kafka. We will explain the basic concepts of streaming transformations using windows and KSQL before ingesting the transformed stream in a dashboard application. And lastly we will explore the possibilities of adding microservices as subscribers.
Stewart is the Owner and Co-founder of Red Pill Analytics, and has been designing and implementing BI systems since 1996. He is an Oracle ACE Director, is an internationally recognized writer and speaker in the Oracle Community, and co-host for the audio and video podcast See Data... Read More →
Björn Rost is an Oracle ACE Director, and one of Pythian’s top Oracle experts. A popular presenter, Björn travels the world attending technology conferences, sharing insights, and learning with his wide network of peers. Björn also served as president of IOUG’s RAC special... Read More →
Thursday March 22, 2018 3:40pm - 4:30pm PDT
1-Auditorium