楼主: longgb246
1447 0

[程序分享] hive自学代码[自学笔记] [推广有奖]

  • 3关注
  • 1粉丝

已卖:139份资源

讲师

68%

还不是VIP/贵宾

-

威望
0
论坛币
2491 个
通用积分
8.4272
学术水平
10 点
热心指数
16 点
信用等级
8 点
经验
2008 点
帖子
346
精华
0
在线时间
613 小时
注册时间
2013-4-26
最后登录
2023-3-18

楼主
longgb246 学生认证  发表于 2016-3-15 14:09:06 |AI写论文

+2 论坛币
k人 参与回答

经管之家送您一份

应届毕业生专属福利!

求职就业群
赵安豆老师微信:zhaoandou666

经管之家联合CDA

送您一个全额奖学金名额~ !

感谢您参与论坛问题回答

经管之家送您两个论坛币!

+2 论坛币
  1. # start
  2. sh $HIVE_HOME/bin/hive --service hwi

  3. hadoop fs -text /user/admin/daiqf/createspu_fp/input/cateinfo | head

  4. # 005 means the tab
  5. CREATE TABLE IF NOT EXISTS table_name ROW FORMAT DELIMITED FIELDS TERMINATED BY '\005' STORED AS textfile

  6. hadoop fs -ls

  7. # partition
  8. # 1.build table
  9. CREATE TABLE c02_clickstat_fatdtl
  10. (
  11.         yyyymmdd string,
  12.         id int,
  13.         ip string,
  14.         country string,
  15.         cookie_id string,
  16.         page_id string,
  17.         clickstat_url_id int,
  18.         query_string string,
  19.         refer string
  20. )PARTITIONED BY(dt STRING)
  21. row format delimited fields terminated by '\005'
  22. # load data
  23. LOAD DATA INPATH '/user/admin/SqlldrDat/CnClickstat/20101101/19/clickstat_gp_fatdt0/0'
  24. OVERWRITE INTO TABLE c02_clickstat_fatdtl PARTITION(dt='20101101');
  25. # visit a partition
  26. SELECT count(*)
  27. FROM c02_clickstat_fatdtl a
  28. WHERE a.dt >= '20101101'  AND a.dt < '20101102';
  29. select * from ods_dpi_mob_data_log a where a.dt >= '2014051814'  and a.dt < '2014051815'
  30. # add a partition
  31. alter table ods_dpi_mob_data_log add partition(dt='2014051814')
  32. show partitions ods_dpi_mob_data_log;

  33. ALTER TABLE table_name RENAME TO new_table_name

  34. hive> desc xi;
  35. OK
  36. id int
  37. cont string
  38. dw_ins_date string
  39. hive> create table xibak like xi;
  40. OK
  41. # add
  42. hive> alter table xibak add columns (ins_date1 string);
  43. OK
  44. hive> desc xibak;
  45. OK
  46. id int
  47. cont string
  48. dw_ins_date string
  49. ins_date1 string
  50. # repalce all
  51. hive> alter table xibak repalce columns (ins_date2 string);
  52. OK
  53. hive> desc xibak;
  54. OK
  55. ins_date2 string

  56. SHOW TABLES
  57. SHOW TABLES 'page.*'
  58. SHOW TABLES '.*view'
  59. SHOW PARTITIONS page_view
  60. DESCRIBE invites
  61. SELECT a.foo FROM invites a WHERE a.ds='2008-08-15'
  62. SELECT a.foo FROM invites a limit 3;
  63. DESCRIBE EXTENDED page_view PARTITION (ds='2008-08-08')

  64. # add to the origin table
  65. LOAD DATA LOCAL INPATH '/tmp/pv_2008-06-08_us.txt' INTO TABLE c02 PARTITION(date='2008-06-08', country='US')
  66. LOAD DATA LOCAL INPATH './examples/files/kvl.txt' INTO TABLE pokes;
  67. # overwirte
  68. LOAD DATA INPATH '/user/admin/SqlldrDat/CnClickstat/20101101/19/clickstat_gp_fatdt0/0'
  69. OVERWRITE INTO TABLE c02_clickstat_fatdtl PARTITION(dt='20101101');

  70. hive> FROM invites a INSERT OVERWRITE TABLE events SELECT a.bar, count(*)
  71. WHERE a.foo > 0 GROUP  BY a.bar;
  72. hiev> INSERT OVERWRITE TABLE events SELECT a.bar, count(*) FROM invites a
  73. WHERE a.foo > 0 GROUP  BY a.bar;

  74. SELECT a.* FROM a JOIN b ON (a.id = b.id)
  75. SELECT a.* FROM a JOIN b ON (a.id = b.id AND a.department = b.department)
  76. SELECT a.val, b.val, c.val FROM a JOIN b ON (a.key = b.key1) JOIN c ON (c.key = b.key2)

  77. SELECT a.val, b.val, c.val
  78. FROM OUTER JOIN b ON (a.key = b.key)
  79. WHERE a.ds='2009-07-07' AND b.ds = '2009-07-07'

  80. SELECT a.val, b.val FROM a LEFT OUTER JOIN b
  81. ON (a.key = b.key AND b.ds = '2009-07-07' AND a.ds = '2009-07-07')

  82. # hive
  83. # -i <filename>   initialization sql from file
  84. # -e 'quoted query string'     sql from command line
  85. # -f <filename>     sql from file



  86. hive -e "use ku; show tables"        #查看ku库中的所有表
  87. hive -e "show databases"        #查看所有的databases
  88.         show tables like '*name*'        #查看名字包含name的数据库
  89.         desc formatted table_name        #查看数据库的结构
  90.         lbs_lifedata_         
  91.         lbs_lifebase_s
  92. desc formatted lbs_lifedata_lbs_batch_plan
  93. hive -S -e "select * FROM mytable LIMIT 3"
  94. less -SN table.desc        #可以查看table按照格式

  95. hadoop fs -lsr /apps/hduser4899
  96. hadoop fs -dus /apps/hduser4899/heater/heater_agraph
  97. hadoop fs -get /apps/hduser4899/heater/heater_idno_name/* ./

  98. table=$1
  99. hive -e "select * from $1 limit 1000" | gzip > $1.1000.gz
复制代码







二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

关键词:Hive HIV Department partition formatted

已有 1 人评分经验 论坛币 收起 理由
三世相思2013 + 20 + 20 鼓励积极发帖讨论

总评分: 经验 + 20  论坛币 + 20   查看全部评分

您需要登录后才可以回帖 登录 | 我要注册

本版微信群
加好友,备注cda
拉您进交流群
GMT+8, 2025-12-30 11:30