spark sql对单表数据在1.6亿左右的表进行复杂查询,速度有多快

sql大致如下
select t.numIid numIid,t.keyCount as keyCount,t.ssrqCount as ssrqCount,

        ((t.keyCount-t2.keyCount)/t.keywordCount) as growthIndex,t.keyCount/#{KeyCont} as keyPercent
        from
        (
            select numIid, 
            sum(keyCount) as keyCount,
            sum(ssrqCount) as ssrqCount
            from
            tablename
            where
            day="201601"
            and type in(1,2,3)
            group by numIid
        ) t
        left join
        (
            select numIid, 
            sum(keywordCount) as keywordCount,
            sum(ssrqCount) as ssrqCount
            from tablename
            where
            day="201602"
            and type in(1,2,3)
            group by numIid
        ) t2
        on t.numIid =t2.numIid
    order by keycount desc
阅读 4.4k
撰写回答
你尚未登录,登录后可以
  • 和开发者交流问题的细节
  • 关注并接收问题和回答的更新提醒
  • 参与内容的编辑和改进,让解决方法与时俱进
推荐问题