Skip to content
This repository has been archived by the owner on Aug 2, 2019. It is now read-only.
/ ProcessKafka Public archive

Spark streaming consume kafka and insert to hive

Notifications You must be signed in to change notification settings

THPT/ProcessKafka

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 

Repository files navigation

ProcessKafka

Spark streaming consume kafka and insert to hadoop hive

Create table event in hive:

CREATE TABLE IF NOT EXISTS events (ip STRING, created_at BIGINT, agent STRING, user_agent_family STRING, user_agent_major STRING, user_agent_minor STRING, os_family STRING, os_major STRING, os_minor STRING, device_family STRING, uuid STRING, referrer STRING, url STRING, metric STRING, product_id STRING, video_id STRING, order_id BIGINT, customer_id BIGINT);

Run:

mvn clean && mvn compile && mvn package && spark-submit --class ProcessKafka target/ProcessKafka-0.0.1-SNAPSHOT-jar-with-dependencies.jar localhost:2181 page_view page_view_logs,click_logs,order_logs 1

About

Spark streaming consume kafka and insert to hive

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages