adn_request_daily.sh 1.8 KB
Newer Older
wang-jinfeng committed
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51
#!/bin/bash

# # # # # # # # # # # # # # # # # # # # # # 
# @file  : adn_request_daily.sh
# @author: walt
# @date  : 16-12-01
# @desc  : 从ods_adn_trackingnew_request中抽取appid并匹配出package_name,保存到etl_adn_sdk_request_daily
# # # # # # # # # # # # # # # # # # # # # #


source ../../dmp_env.sh

LOG_TIME=$(date -d "$ScheduleTime 1 days ago" "+%Y%m%d")
old_path=$(date -d "$ScheduleTime 2 days ago" "+%Y/%m/%d")
date_path=$(date -d "$ScheduleTime 1 days ago" "+%Y/%m/%d")
task_date=$(date -d "$ScheduleTime 1 days ago" +"%Y-%m-%d")

INPUT_ADN_SDK_PKG_DAILY="$ETL_ADN_ORG_REQ_HOURS/${date_path}"
INPUT_MAPPING_PATH="$APP_ID_MAPPING/$date_path"
TMP_OUTPUT_PATH="${TMP_AND_REQUEST_DAILY_PATH}/$date_path"
OUTPUT="$ETL_ADN_SDK_REQUEST_DAILY/$date_path"


check_await "$INPUT_ADN_SDK_PKG_DAILY/virginia/23/_SUCCESS"
check_await "$INPUT_MAPPING_PATH/_SUCCESS"

hadoop fs -rm -r "$TMP_OUTPUT_PATH"

## REDUCE_NUM=$(calculate_reduce_num "${INPUT_ADN_SDK_PKG_DAILY};${INPUT_MAPPING_PATH}")

export HADOOP_CLIENT_OPTS="-Xmx2496m $HADOOP_CLIENT_OPTS"
hadoop jar ../../${JAR} mobvista.dmp.datasource.adn.mapreduce.AdnSdkRequestPkgDailyMR \
    -Dtask.date="${task_date}" \
    -Dmapreduce.map.memory.mb=4072 \
    -Dmapreduce.map.java.opts=-Xmx3458m \
    -Dmapreduce.reduce.memory.mb=4072 \
    -Dmapreduce.reduce.java.opts=-Xmx3458m \
    -Dmapreduce.job.reduces=300 \
    -Dmapreduce.fileoutputcommitter.algorithm.version=2 \
    "$INPUT_ADN_SDK_PKG_DAILY" "$INPUT_MAPPING_PATH" "$TMP_OUTPUT_PATH" "${DIM_MANUAL_MAPPING}"

if [[ $? -ne 0 ]]; then
    exit 255
fi

# 为bundleId 匹配 packageName
matchBundlePackage "$date_path" "$old_path" "2" "4" "$TMP_OUTPUT_PATH" "$OUTPUT" "adn_request" "../../${JAR}"

mount_partition "etl_adn_sdk_request_daily" "\`date\`='$LOG_TIME'" "$OUTPUT" || exit 1

echo "[Etl Adn Sdk request Daily End!]"