Compare commits
No commits in common. "master" and "v1.8.5.fx1" have entirely different histories.
master
...
v1.8.5.fx1
1
.gitignore
vendored
1
.gitignore
vendored
@ -1 +0,0 @@
|
||||
.history/
|
||||
@ -1,278 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
# --------------------- Settings
|
||||
# SOURCE_HOST=cloud21.cvtt.vpn
|
||||
# SOURCE_ROOT_DIR=/opt/store/cvtt/md_archive/crypto
|
||||
# DB_SOURCES=cloud28,cloud29
|
||||
# OUTPUT_DIR=/tmp
|
||||
# DATE=20250516
|
||||
# RSYNC_TARGETS="cvtt@hs01.cvtt.vpn:/works/cvtt/md_archive/crypto/sim/ cvtt@cloud21.cvtt.vpn:/opt/store/cvtt/md_archive/crypto/sim/"
|
||||
# --------------------- Settings
|
||||
|
||||
if [ -z ${DATE} ] ; then
|
||||
DATE=$(date -d 'yesterday' +'%Y%m%d')
|
||||
fi
|
||||
|
||||
if [ -z ${OUTPUT_DIR} ] ; then
|
||||
OUTPUT_DIR=.
|
||||
fi
|
||||
|
||||
echo "DATE=${DATE} SOURCE_HOST=${SOURCE_HOST}"
|
||||
|
||||
mkdir -p ${OUTPUT_DIR}
|
||||
|
||||
year=$(date -d ${DATE} +"%Y")
|
||||
month=$(date -d ${DATE} +"%m")
|
||||
|
||||
if [ -z "${DB_SOURCES}" ]; then
|
||||
echo "DB_SOURCES is empty"
|
||||
exit
|
||||
fi
|
||||
|
||||
IFS=',' read -r -a db_source_hosts <<< "${DB_SOURCES}"
|
||||
|
||||
SourceFile="${DATE}.mktdata.db.gz"
|
||||
SelectedSourceHost=""
|
||||
SelectedSourceFilePath=""
|
||||
SelectedSourceSize=0
|
||||
|
||||
for db_source_host in "${db_source_hosts[@]}"; do
|
||||
SourceDir="${SOURCE_ROOT_DIR}/${db_source_host}/${year}/${month}"
|
||||
CandidatePath="${SourceDir}/${SourceFile}"
|
||||
remote_stat_cmd="if [ -f '${CandidatePath}' ]; then stat -c %s '${CandidatePath}'; else exit 1; fi"
|
||||
CandidateSize=$(ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null ${SOURCE_HOST} "${remote_stat_cmd}" 2>/dev/null)
|
||||
if [ $? -eq 0 ] && [ -n "${CandidateSize}" ]; then
|
||||
echo "Found ${SOURCE_HOST}:${CandidatePath} (${CandidateSize} bytes)"
|
||||
if [ -z "${SelectedSourceHost}" ] || [ "${CandidateSize}" -gt "${SelectedSourceSize}" ]; then
|
||||
SelectedSourceHost=${db_source_host}
|
||||
SelectedSourceFilePath=${CandidatePath}
|
||||
SelectedSourceSize=${CandidateSize}
|
||||
fi
|
||||
fi
|
||||
done
|
||||
|
||||
|
||||
if [ -z "${SelectedSourceHost}" ]; then
|
||||
echo "File ${SourceFile} NOT FOUND on any DB_SOURCES host"
|
||||
exit
|
||||
fi
|
||||
|
||||
echo "Using source ${SelectedSourceHost} with ${SelectedSourceFilePath} (${SelectedSourceSize} bytes)"
|
||||
|
||||
Cmd="/usr/bin/rsync -ahv"
|
||||
Cmd+=" --mkpath"
|
||||
Cmd+=" -e 'ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null'"
|
||||
Cmd+=" ${SOURCE_HOST}:${SelectedSourceFilePath}"
|
||||
Cmd+=" $OUTPUT_DIR/"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
if [ ! -f ${OUTPUT_DIR}/${SourceFile} ] ; then
|
||||
echo "File ${OUTPUT_DIR}/${SourceFile} NOT FOUND"
|
||||
exit
|
||||
fi
|
||||
|
||||
Cmd="(cd ${OUTPUT_DIR} && gunzip -f *.db.gz)"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
SourceDbFile="${OUTPUT_DIR}/${DATE}.mktdata.db"
|
||||
ResultDbFile="${OUTPUT_DIR}/${DATE}.crypto_sim_md.db"
|
||||
|
||||
echo "SourceDbFile=${SourceDbFile}"
|
||||
echo "Creating Result Database File ${ResultDbFile}"
|
||||
cleanup() {
|
||||
rm ${SourceDbFile}
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
echo "Creating table md_trades ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
CREATE TABLE IF NOT EXISTS md_trades (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
exch text,
|
||||
px real,
|
||||
qty real,
|
||||
trade_id text,
|
||||
condition text,
|
||||
tape text
|
||||
);
|
||||
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_trades_uidx
|
||||
ON md_trades(tstamp_ns, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
echo "Creating table md_quotes ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
CREATE TABLE IF NOT EXISTS md_quotes (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
bid_exch text,
|
||||
bid_px real,
|
||||
bid_qty real,
|
||||
ask_exch text,
|
||||
ask_px real,
|
||||
ask_qty real
|
||||
);
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_quotes_uidx
|
||||
ON md_quotes(tstamp_ns, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
echo "Creating table md_1min_bars ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
CREATE TABLE IF NOT EXISTS md_1min_bars (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
open real,
|
||||
high real,
|
||||
low real,
|
||||
close real,
|
||||
volume real,
|
||||
vwap real,
|
||||
num_trades integer
|
||||
);
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_1min_bars_uidx
|
||||
ON md_1min_bars(tstamp, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
echo "Loading md_trades ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
ATTACH '${SourceDbFile}' AS source_db;
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_trades SELECT
|
||||
datetime(exchange_ts_ns / 1000000000, 'unixepoch') || '.' || printf('%06d', (exchange_ts_ns % 1000000000) / 1000) as tstamp,
|
||||
time as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
"" as exch,
|
||||
price as px,
|
||||
quantity as qty,
|
||||
"" as trade_id,
|
||||
taker_side as condition,
|
||||
"" as tape
|
||||
from source_db.bnbspot_md_trades;
|
||||
COMMIT;
|
||||
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_trades
|
||||
SELECT
|
||||
datetime(exchange_ts_ns / 1000000000, 'unixepoch') || '.' || printf('%06d', (exchange_ts_ns % 1000000000) / 1000) as tstamp,
|
||||
time as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
"" as exch,
|
||||
price as px,
|
||||
quantity as qty,
|
||||
"" as trade_id,
|
||||
taker_side as condition,
|
||||
"" as tape
|
||||
from source_db.coinbase_md_trades;
|
||||
COMMIT;
|
||||
|
||||
DETACH source_db;
|
||||
EOF
|
||||
|
||||
echo "Loading md_quotes ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
ATTACH '${SourceDbFile}' AS source_db;
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_quotes SELECT
|
||||
datetime(exchange_ts_ns / 1000000000, 'unixepoch') || '.' || printf('%06d', (exchange_ts_ns % 1000000000) / 1000) as tstamp,
|
||||
time as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
exchange_id as bid_exch,
|
||||
bid_price as bid_px,
|
||||
bid_quantity as bid_qty,
|
||||
exchange_id as ask_exch,
|
||||
ask_price as ask_px,
|
||||
ask_quantity as ask_qty
|
||||
from bnbspot_md_booktops;
|
||||
COMMIT;
|
||||
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_quotes SELECT
|
||||
datetime(exchange_ts_ns / 1000000000, 'unixepoch') || '.' || printf('%06d', (exchange_ts_ns % 1000000000) / 1000) as tstamp,
|
||||
time as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
exchange_id as bid_exch,
|
||||
bid_price as bid_px,
|
||||
bid_quantity as bid_qty,
|
||||
exchange_id as ask_exch,
|
||||
ask_price as ask_px,
|
||||
ask_quantity as ask_qty
|
||||
from coinbase_md_booktops;
|
||||
COMMIT;
|
||||
|
||||
DETACH source_db;
|
||||
EOF
|
||||
|
||||
### --- REPLACE 0 with num_trades ---
|
||||
|
||||
echo "Loading md_1min_bars ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
ATTACH '${SourceDbFile}' AS source_db;
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_1min_bars SELECT
|
||||
datetime(tstamp / 1000000000, 'unixepoch') || '.' || printf('%06d', (tstamp % 1000000000) / 1000) as tsatmp,
|
||||
tstamp as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
open,
|
||||
high,
|
||||
low,
|
||||
close,
|
||||
volume,
|
||||
vwap,
|
||||
0 as num_trades
|
||||
from bnbspot_ohlcv_1min;
|
||||
COMMIT;
|
||||
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_1min_bars SELECT
|
||||
datetime(tstamp / 1000000000, 'unixepoch') || '.' || printf('%06d', (tstamp % 1000000000) / 1000) as tstamp,
|
||||
tstamp as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
open,
|
||||
high,
|
||||
low,
|
||||
close,
|
||||
volume,
|
||||
vwap,
|
||||
0 as num_trades
|
||||
from coinbase_ohlcv_1min;
|
||||
COMMIT;
|
||||
|
||||
DETACH source_db;
|
||||
EOF
|
||||
|
||||
Cmd="gzip ${ResultDbFile}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
for tgt in ${RSYNC_TARGETS} ; do
|
||||
tgt="${tgt}/${year}/${month}/"
|
||||
Cmd="/usr/bin/rsync -ahv"
|
||||
Cmd+=" --mkpath"
|
||||
Cmd+=" -e 'ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null'"
|
||||
Cmd+=" ${ResultDbFile}.gz ${tgt}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
done
|
||||
|
||||
echo Done $0 ${*}
|
||||
@ -1,224 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
# --------------------- Settings
|
||||
# SOURCE_HOST=cloud21.cvtt.vpn
|
||||
# SOURCE_ROOT_DIR=/opt/store/cvtt/md_archive/equity/alpaca_md
|
||||
# SYMBOL_LIST=CAN #,COIN,GBTC,HOOD,MSTR,PYPL,XYZ
|
||||
# OUTPUT_DIR=/tmp
|
||||
# DATE=20250514
|
||||
# RSYNC_TARGETS="cvtt@hs01.cvtt.vpn:/works/cvtt/md_archive/equity/test/alpaca_md/sim/"
|
||||
# RSYNC_TARGETS="cvtt@hs01.cvtt.vpn:/works/cvtt/md_archive/equity/alpaca_md/sim/
|
||||
#cvtt@cloud21.cvtt.vpn:/opt/store/cvtt/md_archive/equity/alpaca_md/sim/"
|
||||
# --------------------- Settings
|
||||
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
date=$(date -d "${dt}" +"%Y-%m-%d")
|
||||
|
||||
CalendarURL=https://trading-calendar.cvtt.net/api/v1/markets/hours?mic=XNYS
|
||||
URL="${CalendarURL}&start=${date}&end=${date}"
|
||||
echo "URL=${URL}"
|
||||
# curl $URL
|
||||
open_time=$(curl -s "${URL}" | jq '.[] | .open_time')
|
||||
echo "open_time=${open_time}"
|
||||
|
||||
if [ -n "${open_time}" ]; then
|
||||
return 0
|
||||
else
|
||||
return 1
|
||||
fi
|
||||
|
||||
}
|
||||
export -f is_business_day
|
||||
|
||||
if [ -z ${SYMBOL_LIST} ] ; then
|
||||
echo "symbols are not specified"
|
||||
exit
|
||||
fi
|
||||
if [ -z ${DATE} ] ; then
|
||||
DATE=$(date -d 'yesterday' +'%Y%m%d')
|
||||
fi
|
||||
|
||||
mkdir -p ${OUTPUT_DIR}
|
||||
|
||||
if ! is_business_day ${DATE}; then
|
||||
echo "${DATE} is not business day"
|
||||
exit
|
||||
fi
|
||||
|
||||
OLD_IFS=${IFS}
|
||||
IFS=","
|
||||
read -ra Symbols <<< "${SYMBOL_LIST}"
|
||||
IFS=${OLD_IFS}
|
||||
|
||||
echo "DATE=${DATE} Symbols=${Symbols[@]} OUTPUT_DIR=${OUTPUT_DIR}"
|
||||
echo Getting data from ${DataHost} ...
|
||||
|
||||
year=$(date -d ${DATE} +"%Y")
|
||||
month=$(date -d ${DATE} +"%m")
|
||||
|
||||
for sym in ${Symbols[@]}; do
|
||||
inst_id="STOCK-${sym}"
|
||||
capital=${sym:0:1}
|
||||
SourceDir="${SOURCE_ROOT_DIR}/${year}/${capital}/${sym}"
|
||||
SourceHbarFile="${SourceDir}/${DATE}.${sym}.alpaca_1m_bars.db.gz"
|
||||
SourceQatFile="${SourceDir}/${DATE}.${sym}.alpaca_qat.db.gz"
|
||||
|
||||
for src_file in ${SourceHbarFile} ${SourceQatFile}; do
|
||||
Cmd="rsync -ahv"
|
||||
Cmd+=" -e 'ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null'"
|
||||
Cmd+=" ${SOURCE_HOST}:${src_file}"
|
||||
Cmd+=" $OUTPUT_DIR/"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
done
|
||||
done
|
||||
|
||||
Cmd="(cd ${OUTPUT_DIR} && gunzip *.db.gz)"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
ResultDbFile="${OUTPUT_DIR}/${DATE}.alpaca_sim_md.db"
|
||||
echo "Creating Result Database File ${ResultDbFile}"
|
||||
|
||||
echo "Creating table md_trades ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
CREATE TABLE IF NOT EXISTS md_trades (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
exch text,
|
||||
px real,
|
||||
qty real,
|
||||
trade_id text,
|
||||
condition text,
|
||||
tape text
|
||||
);
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_trades_uidx
|
||||
ON md_trades(tstamp_ns, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
echo "Creating table md_quotes ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
CREATE TABLE IF NOT EXISTS md_quotes (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
bid_exch text,
|
||||
bid_px real,
|
||||
bid_qty real,
|
||||
ask_exch text,
|
||||
ask_px real,
|
||||
ask_qty real
|
||||
);
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_quotes_uidx
|
||||
ON md_quotes(tstamp_ns, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
echo "Creating table md_1min_bars ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
CREATE TABLE IF NOT EXISTS md_1min_bars (
|
||||
tstamp text,
|
||||
tstamp_ns integer,
|
||||
exchange_id text,
|
||||
instrument_id text,
|
||||
open real,
|
||||
high real,
|
||||
low real,
|
||||
close real,
|
||||
volume real,
|
||||
vwap real,
|
||||
num_trades integer
|
||||
);
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_1min_bars_uidx
|
||||
ON md_1min_bars(tstamp, exchange_id, instrument_id);
|
||||
EOF
|
||||
|
||||
|
||||
# set -f # not to expand *
|
||||
|
||||
for sym in ${Symbols[@]}; do
|
||||
src_hbar_db=${OUTPUT_DIR}/${DATE}.${sym}.alpaca_1m_bars.db
|
||||
src_qat_db=${OUTPUT_DIR}/${DATE}.${sym}.alpaca_qat.db
|
||||
|
||||
echo "Loading md_trades and md_quotes from ${src_qat_db} ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
ATTACH '${src_qat_db}' AS source_db;
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_trades SELECT
|
||||
tstamp,
|
||||
tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
exch,
|
||||
px,
|
||||
qty,
|
||||
trade_id,
|
||||
condition,
|
||||
tape
|
||||
FROM source_db.md_trades;
|
||||
COMMIT;
|
||||
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_quotes SELECT
|
||||
tstamp,
|
||||
tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
bid_exch,
|
||||
bid_px,
|
||||
bid_qty,
|
||||
ask_exch,
|
||||
ask_px,
|
||||
ask_qty
|
||||
FROM source_db.md_quotes;
|
||||
COMMIT;
|
||||
EOF
|
||||
|
||||
echo "Loading md_1min_bars from ${src_hbar_db} ..."
|
||||
sqlite3 ${ResultDbFile} <<EOF
|
||||
.echo ON
|
||||
ATTACH '${src_hbar_db}' AS source_db;
|
||||
BEGIN;
|
||||
INSERT OR IGNORE INTO md_1min_bars SELECT
|
||||
tstamp,
|
||||
tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
open,
|
||||
high,
|
||||
low,
|
||||
close,
|
||||
volume,
|
||||
vwap,
|
||||
num_trades
|
||||
FROM source_db.md_1min_bars;
|
||||
COMMIT;
|
||||
|
||||
DETACH source_db;
|
||||
EOF
|
||||
|
||||
Cmd="rm ${src_hbar_db} ${src_qat_db}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
done
|
||||
|
||||
Cmd="gzip ${ResultDbFile}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
for tgt in ${RSYNC_TARGETS} ; do
|
||||
Cmd="/usr/bin/rsync -ahv"
|
||||
Cmd+=" --mkpath"
|
||||
Cmd+=" -e 'ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null'"
|
||||
Cmd+=" ${ResultDbFile}.gz ${tgt}/${year}/${month}/"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
done
|
||||
|
||||
echo Done $0 ${*}
|
||||
@ -1,6 +1,6 @@
|
||||
#!/bin/env bash
|
||||
|
||||
Registry=http://hs01.cvtt.vpn:5500
|
||||
Registry=http://homestore.cvtt.vpn:5500
|
||||
|
||||
Catalog=${Registry}/v2/_catalog
|
||||
|
||||
@ -11,12 +11,5 @@ jstr=$(curl -s -X GET ${Catalog})
|
||||
echo "${jstr}" | jq -r '.repositories[]' | while read repo; do
|
||||
TagListURL="${Registry}/v2/${repo}/tags/list"
|
||||
# echo $repo
|
||||
Cmd="curl -s -X GET ${TagListURL}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
# TagList=$(eval ${Cmd})
|
||||
# # echo ${TagList}
|
||||
# echo "${TagList}" | jq -r '.tags[]' | while read tag; do
|
||||
# echo "${Registry}/${repo}:${tag}"
|
||||
# done
|
||||
curl -s -X GET ${TagListURL}
|
||||
done
|
||||
@ -3,20 +3,20 @@
|
||||
# ---------------- Settings
|
||||
declare -A git_repo_arr
|
||||
|
||||
git_repo_arr[cvttpy]=git@cloud21.cvtt.vpn:/opt/store/git/cvtt2/cvttpy.git
|
||||
git_repo_arr[ops]=git@cloud21.cvtt.vpn:/opt/store/git/cvtt2/ops.git
|
||||
git_repo_arr[research]=git@cloud21.cvtt.vpn:/opt/store/git/cvtt2/research.git
|
||||
git_repo_arr[cvttpy]=git@cloud21.cvtt.vpn:/opt/store/git/cvttpy.git
|
||||
git_repo_arr[ops]=git@cloud21.cvtt.vpn:/opt/store/git/ops.git
|
||||
git_repo_arr[research]=git@cloud21.cvtt.vpn:/opt/store/git/research.git
|
||||
git_repo_arr[cvtt-rust]=git@cloud21.cvtt.vpn:/opt/store/git/cvtt2/cvtt-rust.git
|
||||
git_repo_arr[docker_dev]=git@cloud21.cvtt.vpn:/opt/store/git/cvtt2/docker_dev.git
|
||||
git_repo_arr[docker_dev]=git@cloud21.cvtt.vpn:/opt/store/git/docker_dev.git
|
||||
|
||||
dist_root=/home/cvttdist/software/cvtt2
|
||||
dist_user=cvttdist
|
||||
dist_host="cloud21.cvtt.vpn"
|
||||
dist_ssh_port="22"
|
||||
|
||||
dist_locations="cloud21.cvtt.vpn:22 homestore.cvtt.vpn:22"
|
||||
# ---------------- Settings
|
||||
dist_locations="cloud21.cvtt.vpn:22 hs01.cvtt.vpn:22"
|
||||
version_file="VERSION"
|
||||
|
||||
# ---------------- cmdline
|
||||
prj=
|
||||
brnch=master
|
||||
@ -145,12 +145,11 @@ if [ "${?}" != "0" ]; then
|
||||
fi
|
||||
|
||||
|
||||
release_version=$(cat ${version_file} | awk -F',' '{print $1}')
|
||||
whats_new=$(cat ${version_file} | awk -F',' '{print $2}')
|
||||
release_version=$(cat release_version.txt | awk -F',' '{print $1}')
|
||||
whats_new=$(cat release_version.txt | awk -F',' '{print $2}')
|
||||
|
||||
|
||||
echo "--------------------------------"
|
||||
echo "Version file: ${version_file}"
|
||||
echo "Release version: ${release_version}"
|
||||
|
||||
confirm
|
||||
@ -176,7 +175,7 @@ if [ "${project}" == "cvtt-rust" ]; then
|
||||
release_dir=${prj_dir}/dist/${project}
|
||||
|
||||
cmd_arr+=("mkdir -p ${release_dir}")
|
||||
cmd_arr+=("cp ${version_file} ${release_dir}")
|
||||
cmd_arr+=("cp release_version.txt ${release_dir}")
|
||||
|
||||
jq_cmd="jq '.packages[] | select(.targets[].kind[] == \"bin\") | .name'"
|
||||
apps=$(cargo metadata --no-deps --format-version 1 --manifest-path=${prj_dir}/Cargo.toml | eval $jq_cmd | uniq)
|
||||
|
||||
@ -5,11 +5,6 @@ usage() {
|
||||
echo
|
||||
exit 1
|
||||
}
|
||||
|
||||
error_exit() {
|
||||
echo "ERROR: ${1}"
|
||||
exit 1
|
||||
}
|
||||
# --- Settings
|
||||
RootDir=${1}
|
||||
if [ -z "${RootDir}" ] ; then
|
||||
@ -19,7 +14,7 @@ fi
|
||||
RegistryService=cloud21.cvtt.vpn:5500
|
||||
RegistryProtocol=http
|
||||
|
||||
ConfigUrl=http://cloud16.cvtt.vpn:6789/admin/docker_images
|
||||
ConfigUrl=http://cloud23.cvtt.vpn:6789/admin/docker_images
|
||||
|
||||
|
||||
ReleaseHost="cloud21.cvtt.vpn"
|
||||
@ -42,7 +37,6 @@ build_docker_image() {
|
||||
image_name=${2}
|
||||
docker_dev_path=${3}
|
||||
project=${4}
|
||||
test_script=${5}
|
||||
|
||||
echo "Building ${title}..."
|
||||
echo "ImageName=${image_name}"
|
||||
@ -72,9 +66,7 @@ build_docker_image() {
|
||||
convert_symlink_to_dir ${project}
|
||||
|
||||
DockerDir=./docker_dev/${docker_dev_path}
|
||||
echo "================================================================================"
|
||||
echo "Building ${title}... on ${DockerDir}"
|
||||
echo "================================================================================"
|
||||
echo "Building ${title}... on ${DockerDev}"
|
||||
|
||||
Cmd="docker build"
|
||||
Cmd+=" -t ${image_name}"
|
||||
@ -82,93 +74,29 @@ build_docker_image() {
|
||||
Cmd+=" -t ${image_name}:${image_version}"
|
||||
Cmd+=" -f ${DockerDir}/Dockerfile ${RootDir}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || error_exit "Docker build failed"
|
||||
|
||||
if [ ! -z "${test_script}" ] && [ "${test_script}" != "null" ]; then
|
||||
if [ ! -f ${DockerDir}/${test_script} ]; then
|
||||
error_exit "Test script ${test_script} is missing"
|
||||
fi
|
||||
echo "Running test script ${test_script} for ${image_name}"
|
||||
|
||||
Cmd="docker cp ${DockerDir}/${test_script} ${image_name}:/"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || error_exit "Test script copy failed"
|
||||
|
||||
Cmd="docker exec ${image_name} chmod +x /${test_script}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || error_exit "Test script chmod failed"
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd+=" --rm"
|
||||
Cmd+=" -d"
|
||||
Cmd+=" --name test-container"
|
||||
Cmd+=" ${image_name}"
|
||||
Cmd+=" /bin/bash -c /${TestScript}"
|
||||
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || error_exit "Test script run failed"
|
||||
|
||||
echo "Removing test script ${test_script} from ${image_name}"
|
||||
Cmd="docker exec ${image_name} rm /${test_script}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || error_exit "Test script remove failed"
|
||||
fi
|
||||
|
||||
|
||||
: <<'COMMENT'
|
||||
-------------------------
|
||||
Every single docker_dev project must have a test script
|
||||
image_test.sh
|
||||
that is copied to the image's root directory.
|
||||
-------------------------
|
||||
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd+=" --rm"
|
||||
Cmd+=" -d"
|
||||
Cmd+=" --name test-container"
|
||||
Cmd+=" ${image_name}"
|
||||
Cmd+=" /bin/bash -c /${TestScript}"
|
||||
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} 1>&2 || exit 1
|
||||
|
||||
|
||||
# Get the exit status of the test script
|
||||
docker wait test-container
|
||||
|
||||
# Check the exit status of the container
|
||||
EXIT_STATUS=$?
|
||||
|
||||
# Analyze the exit status
|
||||
if [ ${EXIT_STATUS} -ne 0 ]; then
|
||||
echo "Tests Failed"
|
||||
exit 1
|
||||
fi
|
||||
COMMENT
|
||||
|
||||
eval ${Cmd} 1>&2 || exit
|
||||
|
||||
Cmd="docker tag"
|
||||
Cmd+=" ${image_name}:latest"
|
||||
Cmd+=" ${RegistryService}/${image_name}:latest"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} || error_exit "Docker tag failed"
|
||||
eval ${Cmd} || exit
|
||||
|
||||
Cmd="docker tag"
|
||||
Cmd+=" ${image_name}:${image_version}"
|
||||
Cmd+=" ${RegistryService}/${image_name}:${image_version}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} || error_exit "Docker tag 2failed"
|
||||
eval ${Cmd} || exit
|
||||
|
||||
Cmd="docker push"
|
||||
Cmd+=" ${RegistryService}/${image_name}:latest"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} || error_exit "Docker push failed"
|
||||
eval ${Cmd} || exit
|
||||
|
||||
Cmd="docker push"
|
||||
Cmd+=" ${RegistryService}/${image_name}:${image_version}"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd} || error_exit "Docker push 2 failed"
|
||||
eval ${Cmd} || exit
|
||||
|
||||
echo "Removing the copy of ${project} from ${RootDir}"
|
||||
Cmd="rm -rf ${RootDir}/${project}"
|
||||
@ -236,7 +164,6 @@ get_latest_dist_version() {
|
||||
echo "${Version}"
|
||||
}
|
||||
|
||||
echo "------------------------------------- Starting ${0} ${*}"
|
||||
echo RootDir=${RootDir}
|
||||
|
||||
cd ${RootDir}
|
||||
@ -259,15 +186,13 @@ for item in "${Jobs[@]}"; do
|
||||
image_name=$(echo "$item" | jq -r '.value.image_name')
|
||||
docker_dev_path=$(echo "$item" | jq -r '.value.docker_dev_path')
|
||||
project=$(echo "$item" | jq -r '.value.based_on_project')
|
||||
test_script=$(echo "$item" | jq -r '.value.test_script')
|
||||
|
||||
build_docker_image "$title" "$image_name" "$docker_dev_path" "$project" "${test_script}"
|
||||
build_docker_image "$title" "$image_name" "$docker_dev_path" "$project"
|
||||
done
|
||||
|
||||
Cmd="rm -rf ${RootDir}/docker_dev"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
echo "Done ${0} ${*}"
|
||||
#===============================================================================
|
||||
#===============================================================================
|
||||
|
||||
@ -1,33 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
usage() {
|
||||
echo -n "Usage: ${0} <grep_filter>"
|
||||
echo
|
||||
exit 1
|
||||
}
|
||||
# --- Settings
|
||||
RegistryService=cloud21.cvtt.vpn:5500
|
||||
RegistryProtocol=http
|
||||
|
||||
print_all_reg_images() {
|
||||
project=${1}
|
||||
|
||||
repositories=$(curl -s "${RegistryProtocol}://${RegistryService}/v2/_catalog" | jq -r '.repositories[]')
|
||||
|
||||
for repo in $repositories; do
|
||||
# Fetch all tags for the repository
|
||||
tags=$(curl -s "${RegistryProtocol}://${RegistryService}/v2/$repo/tags/list" | jq -r '.tags[]')
|
||||
|
||||
# List each tag
|
||||
echo "REPO: $repo"
|
||||
for tag in $tags; do
|
||||
if [ "${project}" == "" ];then
|
||||
echo "$repo:$tag"
|
||||
else
|
||||
echo "$repo:$tag" | grep ${project}
|
||||
fi
|
||||
done
|
||||
done
|
||||
}
|
||||
|
||||
print_all_reg_images ${1}
|
||||
@ -10,7 +10,7 @@ if [ "" == "${prj}" ] ; then
|
||||
usage
|
||||
fi
|
||||
|
||||
Cmd="pushd /home/oleg/develop/cvtt2-ops"
|
||||
Cmd="pushd /home/oleg/develop/cvtt2"
|
||||
Cmd="${Cmd} && (cd ${prj}"
|
||||
Cmd="${Cmd} && git pushall)"
|
||||
Cmd="${Cmd} && ./build_release.sh -p ${prj}"
|
||||
|
||||
183
build_ops.sh
183
build_ops.sh
@ -1,183 +0,0 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
# ---------------- Settings
|
||||
|
||||
repo=git@cloud21.cvtt.vpn:/works/git/cvtt2/ops.git
|
||||
|
||||
dist_root=/home/cvttdist/software/cvtt2
|
||||
dist_user=cvttdist
|
||||
dist_host="cloud21.cvtt.vpn"
|
||||
dist_ssh_port="22"
|
||||
|
||||
dist_locations="cloud21.cvtt.vpn:22 hs01.cvtt.vpn:22"
|
||||
version_file="VERSION"
|
||||
|
||||
prj=ops
|
||||
brnch=master
|
||||
interactive=N
|
||||
|
||||
# ---------------- Settings
|
||||
|
||||
# ---------------- cmdline
|
||||
|
||||
usage() {
|
||||
echo "Usage: $0 [-b <branch (master)> -i (interactive)"
|
||||
exit 1
|
||||
}
|
||||
|
||||
while getopts "b:i" opt; do
|
||||
case ${opt} in
|
||||
b )
|
||||
brnch=$OPTARG
|
||||
;;
|
||||
i )
|
||||
interactive=Y
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
;;
|
||||
: )
|
||||
echo "Option -$OPTARG requires an argument." >&2
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
# ---------------- cmdline
|
||||
|
||||
confirm() {
|
||||
if [ "${interactive}" == "Y" ]; then
|
||||
echo "--------------------------------"
|
||||
echo -n "Press <Enter> to continue" && read
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
if [ "${interactive}" == "Y" ]; then
|
||||
echo -n "Enter project [${prj}]: "
|
||||
read project
|
||||
if [ "${project}" == "" ]
|
||||
then
|
||||
project=${prj}
|
||||
fi
|
||||
else
|
||||
project=${prj}
|
||||
fi
|
||||
|
||||
# repo=${git_repo_arr[${project}]}
|
||||
if [ -z ${repo} ]; then
|
||||
echo "ERROR: Project repository for ${project} not found"
|
||||
exit -1
|
||||
fi
|
||||
echo "Project repo: ${repo}"
|
||||
|
||||
if [ "${interactive}" == "Y" ]; then
|
||||
echo -n "Enter branch to build release from [${brnch}]: "
|
||||
read branch
|
||||
if [ "${branch}" == "" ]
|
||||
then
|
||||
branch=${brnch}
|
||||
fi
|
||||
else
|
||||
branch=${brnch}
|
||||
fi
|
||||
|
||||
tmp_dir=$(mktemp -d)
|
||||
function cleanup {
|
||||
cd ${HOME}
|
||||
rm -rf ${tmp_dir}
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
|
||||
prj_dir="${tmp_dir}/${prj}"
|
||||
|
||||
cmd_arr=()
|
||||
Cmd="git clone ${repo} ${prj_dir}"
|
||||
cmd_arr+=("${Cmd}")
|
||||
|
||||
Cmd="cd ${prj_dir}"
|
||||
cmd_arr+=("${Cmd}")
|
||||
|
||||
if [ "${interactive}" == "Y" ]; then
|
||||
echo "------------------------------------"
|
||||
echo "The following commands will execute:"
|
||||
echo "------------------------------------"
|
||||
for cmd in "${cmd_arr[@]}"
|
||||
do
|
||||
echo ${cmd}
|
||||
done
|
||||
fi
|
||||
|
||||
confirm
|
||||
|
||||
for cmd in "${cmd_arr[@]}"
|
||||
do
|
||||
echo ${cmd} && eval ${cmd}
|
||||
done
|
||||
|
||||
Cmd="git checkout ${branch}"
|
||||
echo ${Cmd} && eval ${Cmd}
|
||||
if [ "${?}" != "0" ]; then
|
||||
echo "ERROR: Branch ${branch} is not found"
|
||||
cd ${HOME} && rm -rf ${tmp_dir}
|
||||
exit -1
|
||||
fi
|
||||
|
||||
|
||||
release_version=$(cat ${version_file} | awk -F',' '{print $1}')
|
||||
whats_new=$(cat ${version_file} | awk -F',' '{print $2}')
|
||||
|
||||
|
||||
echo "--------------------------------"
|
||||
echo "Version file: ${version_file}"
|
||||
echo "Release version: ${release_version}"
|
||||
|
||||
confirm
|
||||
|
||||
version_tag="v${release_version}"
|
||||
version_comment="'${version_tag} ${project} ${branch} $(date +%Y-%m-%d)\n${whats_new}'"
|
||||
|
||||
cmd_arr=()
|
||||
Cmd="git tag -a ${version_tag} -m ${version_comment}"
|
||||
cmd_arr+=("${Cmd}")
|
||||
|
||||
Cmd="git push origin --tags"
|
||||
cmd_arr+=("${Cmd}")
|
||||
|
||||
Cmd="rm -rf .git"
|
||||
cmd_arr+=("${Cmd}")
|
||||
|
||||
SourceLoc=../${project}
|
||||
|
||||
dist_path="${dist_root}/${project}/${release_version}"
|
||||
|
||||
for dist_loc in ${dist_locations}; do
|
||||
dhp=(${dist_loc//:/ })
|
||||
dist_host=${dhp[0]}
|
||||
dist_port=${dhp[1]}
|
||||
Cmd="rsync -avzh"
|
||||
Cmd="${Cmd} --rsync-path=\"mkdir -p ${dist_path}"
|
||||
Cmd="${Cmd} && rsync\" -e \"ssh -p ${dist_ssh_port}\""
|
||||
Cmd="${Cmd} $SourceLoc ${dist_user}@${dist_host}:${dist_path}/"
|
||||
cmd_arr+=("${Cmd}")
|
||||
done
|
||||
|
||||
if [ "${interactive}" == "Y" ]; then
|
||||
echo "------------------------------------"
|
||||
echo "The following commands will execute:"
|
||||
echo "------------------------------------"
|
||||
for cmd in "${cmd_arr[@]}"
|
||||
do
|
||||
echo ${cmd}
|
||||
done
|
||||
fi
|
||||
|
||||
confirm
|
||||
|
||||
for cmd in "${cmd_arr[@]}"
|
||||
do
|
||||
pwd && echo ${cmd} && eval ${cmd}
|
||||
done
|
||||
|
||||
echo "$0 Done ${project} ${release_version}"
|
||||
@ -9,7 +9,7 @@ get_user_hosts() {
|
||||
local User=${1}
|
||||
local Domain=${2}
|
||||
|
||||
Cmd="curl -s http://cloud16.cvtt.vpn:6789/admin/cvtt_hosts"
|
||||
Cmd="curl -s http://cloud23.cvtt.vpn:6789/admin/cvtt_hosts"
|
||||
Cmd="${Cmd} | ${HOME}/bin/hjson -j"
|
||||
Cmd="${Cmd} | jq -r"
|
||||
Cmd="${Cmd} --arg domain \"${Domain}\""
|
||||
|
||||
@ -17,7 +17,7 @@ is_container_running() {
|
||||
fi
|
||||
}
|
||||
|
||||
export CalendarURL=https://trading-calendar.cvtt.net/api/v1/markets/hours?mic=XNYS
|
||||
export CalendarURL=http://cloud23.cvtt.vpn:8000/api/v1/markets/hours?mic=XNYS
|
||||
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
25
docker/market_data/hist/crypto_md_day.sh
Executable file
25
docker/market_data/hist/crypto_md_day.sh
Executable file
@ -0,0 +1,25 @@
|
||||
#!/bin/bash
|
||||
|
||||
host=${1}
|
||||
date=${2}
|
||||
|
||||
DockerRegistry=cloud21.cvtt.vpn:5500
|
||||
DockerImage=${DockerRegistry}/crypto_md_day
|
||||
|
||||
if [ -z ${date} ] ; then
|
||||
date=$(date -d "yesterday" +%Y%m%d)
|
||||
fi
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd="${Cmd} --pull=always"
|
||||
Cmd="${Cmd} --network=host"
|
||||
Cmd="${Cmd} --name=crypto_md_day.${host}.${date}"
|
||||
Cmd="${Cmd} --rm"
|
||||
Cmd="${Cmd} ${DockerImage}"
|
||||
Cmd="${Cmd} -h ${host}"
|
||||
Cmd="${Cmd} -d ${date}"
|
||||
# Cmd="${Cmd} -s coinbase,bnbspot,bnbfut"
|
||||
Cmd="${Cmd} -s coinbase,bnbspot"
|
||||
|
||||
echo $Cmd
|
||||
eval $Cmd
|
||||
105
docker/market_data/hist/load_alpaca_hbar.sh
Executable file
105
docker/market_data/hist/load_alpaca_hbar.sh
Executable file
@ -0,0 +1,105 @@
|
||||
#!/bin/bash
|
||||
|
||||
usage() {
|
||||
echo "Usage: $0 -N <num_symbols> [-L <LogDir>] [-d <YYYYMMDD Date>]"
|
||||
exit 1
|
||||
}
|
||||
|
||||
is_container_running() {
|
||||
local container_name=$1
|
||||
|
||||
if [ "$(docker ps --filter "name=^/${container_name}$" --filter "status=running" -q)" ]; then
|
||||
return 0 # true
|
||||
else
|
||||
return 1 # false
|
||||
fi
|
||||
}
|
||||
|
||||
export CalendarURL=http://cloud23.cvtt.vpn:8000/api/v1/markets/hours?mic=XNYS
|
||||
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
|
||||
open_time=$(curl -s "${CalendarURL}&start=${dt}&end=${dt}" | jq '.[] | .open_time')
|
||||
if [ -n "${open_time}" ]; then
|
||||
return 0
|
||||
else
|
||||
return 1
|
||||
fi
|
||||
|
||||
}
|
||||
export -f is_business_day
|
||||
|
||||
get_prev_business_day() {
|
||||
Start=${1}
|
||||
while true; do
|
||||
if is_business_day ${Start}; then
|
||||
break
|
||||
fi
|
||||
echo "${Start} is not business day in US" >&2
|
||||
Start=$(date -d "${Start} - 1 day" "+%Y-%m-%d")
|
||||
done
|
||||
echo ${Start}
|
||||
}
|
||||
export -f get_prev_business_day
|
||||
|
||||
# ----- Settings
|
||||
DockerRegistry=cloud21.cvtt.vpn:5500
|
||||
DockerImage=${DockerRegistry}/alpaca_hbar #:latest
|
||||
ContainerName=alpaca_hbar
|
||||
LogDir=/home/cvtt/prod/logs
|
||||
# ----- Settings
|
||||
|
||||
while getopts ":d:N:L:" opt; do
|
||||
case ${opt} in
|
||||
d )
|
||||
date_to_load=$OPTARG
|
||||
;;
|
||||
N )
|
||||
NumSymbols=$OPTARG
|
||||
;;
|
||||
L )
|
||||
LogDir=$OPTARG
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
;;
|
||||
: )
|
||||
echo "Option -$OPTARG requires an argument." >&2
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if is_container_running "$ContainerName"; then
|
||||
echo "Container ${ContainerName} is already running."
|
||||
exit 3
|
||||
fi
|
||||
|
||||
if [ -z "${date_to_load}" ]; then
|
||||
echo "date_to_load is empty"
|
||||
date_to_load=$(get_prev_business_day $(date -d "yesterday" '+%Y-%m-%d'))
|
||||
echo "Historical Data for ${date_to_load}"
|
||||
fi
|
||||
|
||||
echo "date_to_load=${date_to_load}"
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd="${Cmd} --pull=always"
|
||||
Cmd="${Cmd} --network=host"
|
||||
Cmd="${Cmd} --name=${ContainerName}"
|
||||
Cmd="${Cmd} --rm"
|
||||
Cmd="${Cmd} --volume=${LogDir}:/logs"
|
||||
Cmd="${Cmd} ${DockerImage}"
|
||||
Cmd="${Cmd} -d ${date_to_load}"
|
||||
if [ -n "${NumSymbols}" ]; then
|
||||
Cmd="${Cmd} -N ${NumSymbols}"
|
||||
fi
|
||||
|
||||
echo $Cmd
|
||||
eval $Cmd
|
||||
|
||||
if [ "$?" != "0" ] ; then
|
||||
exit 1 # if killed we do not save last day
|
||||
fi
|
||||
108
docker/market_data/hist/load_alpaca_qat.sh
Executable file
108
docker/market_data/hist/load_alpaca_qat.sh
Executable file
@ -0,0 +1,108 @@
|
||||
#!/bin/bash
|
||||
|
||||
usage() {
|
||||
echo "Usage: $0 -S <symbols> [-L <LogDir>] [-d <YYYYMMDD Date>]"
|
||||
exit 1
|
||||
}
|
||||
|
||||
is_container_running() {
|
||||
local container_name=$1
|
||||
|
||||
if [ "$(docker ps --filter "name=^/${container_name}$" --filter "status=running" -q)" ]; then
|
||||
return 0 # true
|
||||
else
|
||||
return 1 # false
|
||||
fi
|
||||
}
|
||||
|
||||
export CalendarURL=http://cloud23.cvtt.vpn:8000/api/v1/markets/hours?mic=XNYS
|
||||
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
|
||||
open_time=$(curl -s "${CalendarURL}&start=${dt}&end=${dt}" | jq '.[] | .open_time')
|
||||
if [ -n "${open_time}" ]; then
|
||||
return 0
|
||||
else
|
||||
return 1
|
||||
fi
|
||||
|
||||
}
|
||||
export -f is_business_day
|
||||
|
||||
get_prev_business_day() {
|
||||
Start=${1}
|
||||
while true; do
|
||||
if is_business_day ${Start}; then
|
||||
break
|
||||
fi
|
||||
echo "${Start} is not business day in US" >&2
|
||||
Start=$(date -d "${Start} - 1 day" "+%Y-%m-%d")
|
||||
done
|
||||
echo ${Start}
|
||||
}
|
||||
export -f get_prev_business_day
|
||||
|
||||
# ----- Settings
|
||||
DockerRegistry=cloud21.cvtt.vpn:5500
|
||||
DockerImage=${DockerRegistry}/alpaca_qat #:latest
|
||||
ContainerName=alpaca_qat
|
||||
LogDir=/home/cvtt/prod/logs
|
||||
# ----- Settings
|
||||
|
||||
while getopts ":d:S:L:" opt; do
|
||||
case ${opt} in
|
||||
d )
|
||||
date_to_load=$OPTARG
|
||||
;;
|
||||
S )
|
||||
Symbols=$OPTARG
|
||||
;;
|
||||
L )
|
||||
LogDir=$OPTARG
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
;;
|
||||
: )
|
||||
echo "Option -$OPTARG requires an argument." >&2
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ -z ${Symbols} ] ; then
|
||||
echo "symbols are not specified"
|
||||
usage
|
||||
fi
|
||||
|
||||
if is_container_running "$ContainerName"; then
|
||||
echo "Container ${ContainerName} is already running."
|
||||
exit 3
|
||||
fi
|
||||
|
||||
if [ -z "${date_to_load}" ]; then
|
||||
echo "date_to_load is empty"
|
||||
date_to_load=$(get_prev_business_day $(date -d "yesterday" '+%Y-%m-%d'))
|
||||
echo "Historical Data for ${date_to_load}"
|
||||
fi
|
||||
|
||||
echo "date_to_load=${date_to_load}"
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd="${Cmd} --pull=always"
|
||||
Cmd="${Cmd} --network=host"
|
||||
Cmd="${Cmd} --name=${ContainerName}"
|
||||
Cmd="${Cmd} --rm"
|
||||
Cmd="${Cmd} --volume=${LogDir}:/logs"
|
||||
Cmd="${Cmd} ${DockerImage}"
|
||||
Cmd="${Cmd} -d ${date_to_load}"
|
||||
Cmd="${Cmd} -S ${Symbols}"
|
||||
|
||||
echo $Cmd
|
||||
eval $Cmd
|
||||
|
||||
if [ "$?" != "0" ] ; then
|
||||
exit 1 # if killed we do not save last day
|
||||
fi
|
||||
@ -3,7 +3,7 @@
|
||||
ValidJobs=('BNBFUT_CLD21' 'BNBSPOT_CLD21' 'COINBASE_CLD21' 'BNBFUT_CVTTDATA' 'BNBSPOT_CVTTDATA' 'COINBASE_CVTTDATA')
|
||||
# runs on host to start container
|
||||
usage() {
|
||||
echo "Usage: $0 <job_name; one of (${ValidJobs[@]})> [image_tag]"
|
||||
echo "Usage: $0 <job_name; one of (${ValidJobs[@]})>"
|
||||
exit 1
|
||||
}
|
||||
|
||||
@ -35,12 +35,6 @@ if ! is_valid "${job}"; then
|
||||
usage
|
||||
fi
|
||||
|
||||
ImageTag=${2}
|
||||
if [ "${ImageTag}" == "" ] ; then
|
||||
ImageTag="1.6.9"
|
||||
fi
|
||||
|
||||
DockerImage=cloud21.cvtt.vpn:5500/md_recorder:${ImageTag}
|
||||
ContainerName="md_recorder.${job}"
|
||||
|
||||
if is_container_running "$ContainerName"; then
|
||||
@ -50,17 +44,15 @@ fi
|
||||
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd+=" -d"
|
||||
Cmd+=" --rm"
|
||||
Cmd+=" --network=host"
|
||||
# Cmd+=" --pull=always"
|
||||
Cmd+=" --name=${ContainerName}"
|
||||
Cmd+=" -v /home/cvtt/.creds:/.creds"
|
||||
Cmd+=" -v /home/cvtt/prod/data:/app/data"
|
||||
Cmd+=" -v /home/cvtt/prod/logs:/logs"
|
||||
Cmd+=" -e CONFIG_SERVICE=cloud16.cvtt.vpn:6789"
|
||||
Cmd+=" ${DockerImage}"
|
||||
Cmd+=" ${job}"
|
||||
Cmd="${Cmd} -d"
|
||||
Cmd="${Cmd} --rm"
|
||||
Cmd="${Cmd} --network=host"
|
||||
Cmd="${Cmd} --pull=always"
|
||||
Cmd="${Cmd} --name=${ContainerName}"
|
||||
Cmd="${Cmd} -v /home/cvtt/prod/data:/app/data"
|
||||
Cmd="${Cmd} -v /home/cvtt/prod/logs:/logs"
|
||||
Cmd="${Cmd} cloud21.cvtt.vpn:5500/md_recorder:latest"
|
||||
Cmd="${Cmd} ${job}"
|
||||
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
@ -27,6 +27,9 @@ is_container_running() {
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
Instance=${1}
|
||||
if ! is_valid "${Instance}"; then
|
||||
usage
|
||||
@ -45,16 +48,14 @@ if is_container_running "$ContainerName"; then
|
||||
fi
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd+=" -d"
|
||||
Cmd+=" --rm"
|
||||
Cmd+=" --network=host"
|
||||
Cmd+=" --pull=always"
|
||||
Cmd+=" --name=${ContainerName}"
|
||||
Cmd+=" -v /home/cvtt/.creds:/.creds"
|
||||
Cmd+=" -v /home/cvtt/prod/logs:/logs"
|
||||
Cmd+=" -e CONFIG_SERVICE=cloud16.cvtt.vpn:6789"
|
||||
Cmd+=" cloud21.cvtt.vpn:5500/md_recorder_monitor:latest"
|
||||
Cmd+=" ${Instance} ${AdminPort}"
|
||||
Cmd="${Cmd} -d"
|
||||
Cmd="${Cmd} --rm"
|
||||
Cmd="${Cmd} --network=host"
|
||||
Cmd="${Cmd} --pull=always"
|
||||
Cmd="${Cmd} --name=${ContainerName}"
|
||||
Cmd="${Cmd} -v /home/cvtt/prod/logs:/logs"
|
||||
Cmd="${Cmd} cloud21.cvtt.vpn:5500/md_recorder_monitor:latest"
|
||||
Cmd="${Cmd} ${Instance} ${AdminPort}"
|
||||
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
@ -1,17 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
Cmd="docker run"
|
||||
Cmd+=" -d"
|
||||
Cmd+=" --rm"
|
||||
Cmd+=" --pull=always"
|
||||
Cmd+=" --network=host"
|
||||
Cmd+=" --name=crypto_exch_stats"
|
||||
Cmd+=" --volume=${HOME}/prod/data:/app/data"
|
||||
Cmd+=" --volume=${HOME}/prod/logs:/logs"
|
||||
Cmd+=" -e CONFIG_SERVICE=cloud16.cvtt.vpn:6789"
|
||||
Cmd+=" cloud21.cvtt.vpn:5500/crypto_exch_stats:latest"
|
||||
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
# Cmd+=" cloud21.cvtt.vpn:5500/relative_liquidity:latest"
|
||||
|
||||
1
release_version.txt
Normal file
1
release_version.txt
Normal file
@ -0,0 +1 @@
|
||||
1.8.5.fx1,docker_image_builder
|
||||
@ -2,7 +2,7 @@
|
||||
|
||||
# by chatGPT
|
||||
# ==========================================
|
||||
# hs01.cvtt.vpn SETTINGS
|
||||
# homestore.cvtt.vpn SETTINGS
|
||||
# ==========================================
|
||||
|
||||
# Define source and destination base directories
|
||||
|
||||
@ -1,275 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
if [ $# -ne 2 ]; then
|
||||
echo "Usage: $0 <source_database_file> <features_database_file>"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
SRC_DB=$1
|
||||
DEST_DB=$2
|
||||
|
||||
if [ ! -f "$SRC_DB" ]; then
|
||||
echo "Error: Source database file $SRC_DB does not exist"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo "Creating feature tables in $DEST_DB using data from $SRC_DB..."
|
||||
|
||||
# Create md_1min_trade_features table
|
||||
echo "Creating md_1min_trade_features table..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
DROP TABLE IF EXISTS md_1min_trade_features;
|
||||
CREATE TABLE IF NOT EXISTS md_1min_trade_features (
|
||||
bin_tstamp TEXT,
|
||||
tstamp_ns INTEGER,
|
||||
exchange_id TEXT,
|
||||
instrument_id TEXT,
|
||||
price_mean REAL,
|
||||
price_median REAL,
|
||||
volume REAL,
|
||||
vwap REAL,
|
||||
signed_volume REAL,
|
||||
order_flow_imbalance REAL,
|
||||
num_trades INTEGER,
|
||||
avg_trade_size REAL,
|
||||
PRIMARY KEY (bin_tstamp, exchange_id, instrument_id)
|
||||
);"
|
||||
|
||||
# Create index for md_1min_trade_features
|
||||
echo "Creating index for md_1min_trade_features..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_1min_trade_features_uidx
|
||||
ON md_1min_trade_features(bin_tstamp, exchange_id, instrument_id);"
|
||||
|
||||
# Populate md_1min_trade_features using source database
|
||||
echo "Populating md_1min_trade_features..."
|
||||
sqlite3 "$SRC_DB" "ATTACH DATABASE '$DEST_DB' AS dest;
|
||||
WITH trade_metrics AS (
|
||||
SELECT
|
||||
tstamp,
|
||||
strftime('%Y-%m-%d %H:%M:00', tstamp) as bin_tstamp,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
px as price,
|
||||
qty,
|
||||
CASE
|
||||
WHEN condition = 'B' THEN qty
|
||||
WHEN condition = 'S' THEN -qty
|
||||
ELSE 0
|
||||
END as signed_qty
|
||||
FROM md_trades
|
||||
),
|
||||
trade_metrics_agg AS (
|
||||
SELECT
|
||||
bin_tstamp,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
COUNT(*) as cnt,
|
||||
MIN(tstamp) as min_tstamp,
|
||||
MAX(tstamp) as max_tstamp
|
||||
FROM trade_metrics
|
||||
GROUP BY bin_tstamp, exchange_id, instrument_id
|
||||
)
|
||||
INSERT INTO dest.md_1min_trade_features
|
||||
SELECT
|
||||
tm.bin_tstamp,
|
||||
CAST(strftime('%s', tm.bin_tstamp) * 1000000000 AS INTEGER) as tstamp_ns,
|
||||
tm.exchange_id,
|
||||
tm.instrument_id,
|
||||
AVG(price) as price_mean,
|
||||
AVG(CASE WHEN rank_num >= FLOOR(cnt/2.0) AND rank_num <= CEIL(cnt/2.0) THEN price ELSE NULL END) as price_median,
|
||||
SUM(qty) as volume,
|
||||
SUM(price * qty) / SUM(qty) as vwap,
|
||||
SUM(signed_qty) as signed_volume,
|
||||
SUM(CASE WHEN signed_qty > 0 THEN signed_qty ELSE 0 END) -
|
||||
SUM(CASE WHEN signed_qty < 0 THEN ABS(signed_qty) ELSE 0 END) as order_flow_imbalance,
|
||||
COUNT(*) as num_trades,
|
||||
AVG(qty) as avg_trade_size
|
||||
FROM (
|
||||
SELECT
|
||||
tm.*,
|
||||
tma.cnt,
|
||||
ROW_NUMBER() OVER (PARTITION BY tm.bin_tstamp, tm.exchange_id, tm.instrument_id ORDER BY price) as rank_num
|
||||
FROM trade_metrics tm
|
||||
JOIN trade_metrics_agg tma
|
||||
ON tm.bin_tstamp = tma.bin_tstamp
|
||||
AND tm.exchange_id = tma.exchange_id
|
||||
AND tm.instrument_id = tma.instrument_id
|
||||
) tm
|
||||
GROUP BY tm.bin_tstamp, tm.exchange_id, tm.instrument_id;"
|
||||
|
||||
# Create md_1min_quote_features table in destination database
|
||||
echo "Creating md_1min_quote_features table..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
DROP TABLE IF EXISTS md_1min_quote_features;
|
||||
CREATE TABLE IF NOT EXISTS md_1min_quote_features (
|
||||
bin_tstamp TEXT,
|
||||
tstamp_ns INTEGER,
|
||||
exchange_id TEXT,
|
||||
instrument_id TEXT,
|
||||
mid_price_open REAL,
|
||||
mid_price_high REAL,
|
||||
mid_price_low REAL,
|
||||
mid_price_close REAL,
|
||||
mid_price_mean REAL,
|
||||
rel_spread_mean REAL,
|
||||
rel_spread_min REAL,
|
||||
rel_spread_max REAL,
|
||||
rel_spread_first REAL,
|
||||
rel_spread_last REAL,
|
||||
l1_imbalance_mean REAL,
|
||||
l1_imbalance_min REAL,
|
||||
l1_imbalance_max REAL,
|
||||
l1_imbalance_first REAL,
|
||||
l1_imbalance_last REAL,
|
||||
micro_price_mean REAL,
|
||||
micro_price_min REAL,
|
||||
micro_price_max REAL,
|
||||
micro_price_first REAL,
|
||||
micro_price_last REAL,
|
||||
weighted_mid_mean REAL,
|
||||
weighted_mid_min REAL,
|
||||
weighted_mid_max REAL,
|
||||
weighted_mid_first REAL,
|
||||
weighted_mid_last REAL,
|
||||
PRIMARY KEY (bin_tstamp, exchange_id, instrument_id)
|
||||
);"
|
||||
|
||||
# Create index for md_1min_quote_features
|
||||
echo "Creating index for md_1min_quote_features..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_1min_quote_features_uidx
|
||||
ON md_1min_quote_features(bin_tstamp, exchange_id, instrument_id);"
|
||||
|
||||
# Populate md_1min_quote_features using source database
|
||||
echo "Populating md_1min_quote_features..."
|
||||
sqlite3 "$SRC_DB" "ATTACH DATABASE '$DEST_DB' AS dest;
|
||||
INSERT INTO dest.md_1min_quote_features
|
||||
SELECT
|
||||
strftime('%Y-%m-%d %H:%M:00', tstamp) as bin_tstamp,
|
||||
CAST(strftime('%s', tstamp) * 1000000000 AS INTEGER) as tstamp_ns,
|
||||
exchange_id,
|
||||
instrument_id,
|
||||
FIRST_VALUE((ask_px + bid_px) / 2.0) OVER w as mid_price_open,
|
||||
MAX((ask_px + bid_px) / 2.0) as mid_price_high,
|
||||
MIN((ask_px + bid_px) / 2.0) as mid_price_low,
|
||||
LAST_VALUE((ask_px + bid_px) / 2.0) OVER w as mid_price_close,
|
||||
AVG((ask_px + bid_px) / 2.0) as mid_price_mean,
|
||||
AVG((ask_px - bid_px) / ((ask_px + bid_px) / 2.0)) as rel_spread_mean,
|
||||
MIN((ask_px - bid_px) / ((ask_px + bid_px) / 2.0)) as rel_spread_min,
|
||||
MAX((ask_px - bid_px) / ((ask_px + bid_px) / 2.0)) as rel_spread_max,
|
||||
FIRST_VALUE((ask_px - bid_px) / ((ask_px + bid_px) / 2.0)) OVER w as rel_spread_first,
|
||||
LAST_VALUE((ask_px - bid_px) / ((ask_px + bid_px) / 2.0)) OVER w as rel_spread_last,
|
||||
AVG((bid_qty - ask_qty) / (bid_qty + ask_qty)) as l1_imbalance_mean,
|
||||
MIN((bid_qty - ask_qty) / (bid_qty + ask_qty)) as l1_imbalance_min,
|
||||
MAX((bid_qty - ask_qty) / (bid_qty + ask_qty)) as l1_imbalance_max,
|
||||
FIRST_VALUE((bid_qty - ask_qty) / (bid_qty + ask_qty)) OVER w as l1_imbalance_first,
|
||||
LAST_VALUE((bid_qty - ask_qty) / (bid_qty + ask_qty)) OVER w as l1_imbalance_last,
|
||||
AVG((ask_px * bid_qty + bid_px * ask_qty) / (bid_qty + ask_qty)) as micro_price_mean,
|
||||
MIN((ask_px * bid_qty + bid_px * ask_qty) / (bid_qty + ask_qty)) as micro_price_min,
|
||||
MAX((ask_px * bid_qty + bid_px * ask_qty) / (bid_qty + ask_qty)) as micro_price_max,
|
||||
FIRST_VALUE((ask_px * bid_qty + bid_px * ask_qty) / (bid_qty + ask_qty)) OVER w as micro_price_first,
|
||||
LAST_VALUE((ask_px * bid_qty + bid_px * ask_qty) / (bid_qty + ask_qty)) OVER w as micro_price_last,
|
||||
AVG((ask_px * ask_qty + bid_px * bid_qty) / (bid_qty + ask_qty)) as weighted_mid_mean,
|
||||
MIN((ask_px * ask_qty + bid_px * bid_qty) / (bid_qty + ask_qty)) as weighted_mid_min,
|
||||
MAX((ask_px * ask_qty + bid_px * bid_qty) / (bid_qty + ask_qty)) as weighted_mid_max,
|
||||
FIRST_VALUE((ask_px * ask_qty + bid_px * bid_qty) / (bid_qty + ask_qty)) OVER w as weighted_mid_first,
|
||||
LAST_VALUE((ask_px * ask_qty + bid_px * bid_qty) / (bid_qty + ask_qty)) OVER w as weighted_mid_last
|
||||
FROM md_quotes
|
||||
GROUP BY strftime('%Y-%m-%d %H:%M:00', tstamp), exchange_id, instrument_id
|
||||
WINDOW w AS (
|
||||
PARTITION BY strftime('%Y-%m-%d %H:%M:00', tstamp), exchange_id, instrument_id
|
||||
ORDER BY tstamp
|
||||
RANGE BETWEEN UNBOUNDED PRECEDING AND UNBOUNDED FOLLOWING
|
||||
);"
|
||||
|
||||
# Copy the md_1min_bars table to destination database
|
||||
echo "Copying md_1min_bars to destination..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
DROP TABLE IF EXISTS md_1min_bars;
|
||||
CREATE TABLE IF NOT EXISTS md_1min_bars (
|
||||
bin_tstamp TEXT,
|
||||
tstamp_ns INTEGER,
|
||||
exchange_id TEXT,
|
||||
instrument_id TEXT,
|
||||
open REAL,
|
||||
high REAL,
|
||||
low REAL,
|
||||
close REAL,
|
||||
volume REAL,
|
||||
vwap REAL,
|
||||
num_trades INTEGER,
|
||||
PRIMARY KEY (bin_tstamp, exchange_id, instrument_id)
|
||||
);"
|
||||
|
||||
echo "Creating index for md_1min_bars..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
CREATE UNIQUE INDEX IF NOT EXISTS md_1min_bars_uidx
|
||||
ON md_1min_bars(bin_tstamp, exchange_id, instrument_id);"
|
||||
|
||||
echo "Populating md_1min_bars..."
|
||||
sqlite3 "$SRC_DB" "ATTACH DATABASE '$DEST_DB' AS dest;
|
||||
INSERT INTO dest.md_1min_bars
|
||||
SELECT * FROM md_1min_bars;"
|
||||
|
||||
# Create the combined features view in destination database
|
||||
echo "Creating combined features view..."
|
||||
sqlite3 "$DEST_DB" "
|
||||
DROP VIEW IF EXISTS md_1min_features_view;
|
||||
CREATE VIEW IF NOT EXISTS md_1min_features_view AS
|
||||
SELECT
|
||||
b.bin_tstamp,
|
||||
b.tstamp_ns,
|
||||
b.exchange_id,
|
||||
b.instrument_id,
|
||||
-- OHLCV data from md_1min_bars
|
||||
b.open,
|
||||
b.high,
|
||||
b.low,
|
||||
b.close,
|
||||
b.volume,
|
||||
b.vwap,
|
||||
b.num_trades,
|
||||
-- Quote features
|
||||
q.mid_price_open,
|
||||
q.mid_price_high,
|
||||
q.mid_price_low,
|
||||
q.mid_price_close,
|
||||
q.mid_price_mean,
|
||||
q.rel_spread_mean,
|
||||
q.rel_spread_min,
|
||||
q.rel_spread_max,
|
||||
q.rel_spread_first,
|
||||
q.rel_spread_last,
|
||||
q.l1_imbalance_mean,
|
||||
q.l1_imbalance_min,
|
||||
q.l1_imbalance_max,
|
||||
q.l1_imbalance_first,
|
||||
q.l1_imbalance_last,
|
||||
q.micro_price_mean,
|
||||
q.micro_price_min,
|
||||
q.micro_price_max,
|
||||
q.micro_price_first,
|
||||
q.micro_price_last,
|
||||
q.weighted_mid_mean,
|
||||
q.weighted_mid_min,
|
||||
q.weighted_mid_max,
|
||||
q.weighted_mid_first,
|
||||
q.weighted_mid_last,
|
||||
-- Trade features
|
||||
t.price_mean as trade_price_mean,
|
||||
t.price_median as trade_price_median,
|
||||
t.signed_volume,
|
||||
t.order_flow_imbalance,
|
||||
t.avg_trade_size
|
||||
FROM md_1min_bars b
|
||||
LEFT JOIN md_1min_quote_features q
|
||||
ON b.bin_tstamp = q.bin_tstamp
|
||||
AND b.exchange_id = q.exchange_id
|
||||
AND b.instrument_id = q.instrument_id
|
||||
LEFT JOIN md_1min_trade_features t
|
||||
ON b.bin_tstamp = t.bin_tstamp
|
||||
AND b.exchange_id = t.exchange_id
|
||||
AND b.instrument_id = t.instrument_id;"
|
||||
|
||||
echo "Feature tables created and populated successfully in $DEST_DB!"
|
||||
@ -1,10 +1,7 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
usage() {
|
||||
echo -n "Usage: $0 [-h <host (hs01*/cloud21)>]"
|
||||
echo -n " [-d <YYYYMMDD> (yesterday*)]"
|
||||
echo -n " [-s <source> (cloud28/cloud21*)>]"
|
||||
echo " [-t <target_dir> (/opt/jupyter_gpu/data/crypto_md)]"
|
||||
echo "Usage: $0 [-h <host (homestore*/cloud21)>] [-d <YYYYMMDD> (yesterday*)] [-s <source> (cvttdata/cloud21*)>]"
|
||||
exit 1
|
||||
}
|
||||
|
||||
@ -22,15 +19,11 @@ is_valid() {
|
||||
return 1
|
||||
}
|
||||
|
||||
# ------- D E F A U L T S
|
||||
date=""
|
||||
host=hs01
|
||||
host=homestore
|
||||
source=cloud21
|
||||
TargetDir="/opt/jupyter_gpu/data/crypto_md"
|
||||
# ------- D E F A U L T S
|
||||
|
||||
|
||||
while getopts ":h:d:s:t:" opt; do
|
||||
while getopts ":h:d:s:" opt; do
|
||||
case ${opt} in
|
||||
d )
|
||||
date=$OPTARG
|
||||
@ -41,9 +34,6 @@ while getopts ":h:d:s:t:" opt; do
|
||||
s )
|
||||
source=$OPTARG
|
||||
;;
|
||||
t )
|
||||
TargetDir=$OPTARG
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
@ -61,13 +51,13 @@ fi
|
||||
|
||||
echo "$date $host $source"
|
||||
|
||||
valid_hosts=('hs01' 'cloud21')
|
||||
valid_hosts=('homestore' 'cloud21')
|
||||
if ! is_valid "${host}" "${valid_hosts[@]}" ; then
|
||||
echo "Host '${host}' is not valid"
|
||||
usage
|
||||
fi
|
||||
|
||||
valid_sources=('cloud21' 'cloud28')
|
||||
valid_sources=('cloud21' 'cvttdata')
|
||||
if ! is_valid ${source} ${valid_sources[@]} ; then
|
||||
echo "Source '${source}' is not valid"
|
||||
usage
|
||||
@ -77,8 +67,8 @@ if [ "${host}" == "cloud21" ] ; then
|
||||
SourceHost=cloud21.cvtt.vpn
|
||||
SourceUser=cvtt
|
||||
SourceRootDir="/opt/store/cvtt/md_archive/crypto/${source}"
|
||||
elif [ "${host}" == "hs01" ]; then
|
||||
SourceHost=hs01.cvtt.vpn
|
||||
elif [ "${host}" == "homestore" ]; then
|
||||
SourceHost=homestore.cvtt.vpn
|
||||
SourceUser=cvtt
|
||||
SourceRootDir=/works/cvtt/md_archive/crypto/${source}
|
||||
else
|
||||
@ -92,6 +82,7 @@ SourceFile=$(date -d ${date} "+%Y%m%d.mktdata.db")
|
||||
SourceFileZip="${SourceFile}.gz"
|
||||
SourceFilePath=$(date -d ${date} "+${SourceRootDir}/%Y/%m/${SourceFileZip}")
|
||||
|
||||
TargetDir="/opt/jupyter_gpu/data/crypto_md"
|
||||
TargetFile=$(date -d ${date} "+%Y%m%d.mktdata.ohlcv.db")
|
||||
TargetFilePath="${TargetDir}/${TargetFile}"
|
||||
|
||||
|
||||
@ -2,7 +2,7 @@
|
||||
|
||||
usage() {
|
||||
echo -n "Usage: $0"
|
||||
echo -n " [-h <host (hs01*/cloud21)>]"
|
||||
echo -n " [-h <host (homestore*/cloud21)>]"
|
||||
echo -n " [-d <YYYYMMDD> (yesterday*)]"
|
||||
echo -n " [-s <stocks comma separated>"
|
||||
echo -n " [-t <target directory>"
|
||||
@ -19,7 +19,7 @@ fi
|
||||
|
||||
# ------------------ Settings
|
||||
md_date=""
|
||||
host=hs01
|
||||
host=homestore
|
||||
stocks=COIN,GBTC,SQ
|
||||
|
||||
TargetDir="/opt/jupyter_gpu/data/eqty_md"
|
||||
@ -81,7 +81,7 @@ fi
|
||||
|
||||
echo "$md_date $host $source"
|
||||
|
||||
valid_hosts=('hs01' 'cloud21')
|
||||
valid_hosts=('homestore' 'cloud21')
|
||||
if ! is_valid "${host}" "${valid_hosts[@]}" ; then
|
||||
echo "Host '${host}' is not valid"
|
||||
usage
|
||||
@ -108,8 +108,8 @@ if [ "${host}" == "cloud21" ] ; then
|
||||
SourceHost=cloud21.cvtt.vpn
|
||||
SourceUser=cvtt
|
||||
SourceRootDir="/opt/store/cvtt/md_archive/equity/alpaca_md"
|
||||
elif [ "${host}" == "hs01" ]; then
|
||||
SourceHost=hs01.cvtt.vpn
|
||||
elif [ "${host}" == "homestore" ]; then
|
||||
SourceHost=homestore.cvtt.vpn
|
||||
SourceUser=cvtt
|
||||
SourceRootDir="/works/cvtt/md_archive/equity/alpaca_md"
|
||||
else
|
||||
|
||||
44
scripts/__DEPRECATED__/archive_yesterday_md.sh
Executable file
44
scripts/__DEPRECATED__/archive_yesterday_md.sh
Executable file
@ -0,0 +1,44 @@
|
||||
#!/bin/bash
|
||||
|
||||
Python=/home/cvtt/.pyenv/python3.12-venv/bin/python3.12
|
||||
RootDir=/home/cvtt/prod
|
||||
export PYTHONPATH=${RootDir}
|
||||
|
||||
host=${1}
|
||||
if [ "${host}" == "cvttdata" ]
|
||||
then
|
||||
ArchiveRootDir=/home/cvtt/prod/archive/md_archive/crypto/cvttdata
|
||||
CredKey=TSDB_MD_CVTTDATA_RO
|
||||
elif [ "${host}" == "cloud21" ]
|
||||
then
|
||||
ArchiveRootDir=/home/cvtt/prod/archive/md_archive/crypto/cloud21
|
||||
CredKey=TSDB_MD_CLD21_RO
|
||||
else
|
||||
echo "Unknown host ${host}. ${0} Aborted."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
mkdir -p ${ArchiveRootDir}
|
||||
|
||||
yesterday=$(date -d "yesterday" +%Y%m%d)
|
||||
Schemas=${2}
|
||||
if [ "${Schemas}" == "" ]
|
||||
then
|
||||
Schemas="coinbase,bnbspot,bnbfut"
|
||||
fi
|
||||
echo "Schemas=${Schemas}"
|
||||
|
||||
Cmd=
|
||||
Cmd="${Python}"
|
||||
Cmd="${Cmd} ${RootDir}/cvttpy/research/utils/archive_ts_md.py"
|
||||
Cmd="${Cmd} --config=http://cloud16.cvtt.vpn:6789/apps/md_recorder"
|
||||
Cmd="${Cmd} --db_credentials_key=${CredKey}"
|
||||
Cmd="${Cmd} --date=${yesterday}"
|
||||
Cmd="${Cmd} --schemas=${Schemas}"
|
||||
Cmd="${Cmd} --root_dir=${ArchiveRootDir}"
|
||||
Cmd="${Cmd} --format=SQLite"
|
||||
Cmd="${Cmd} --compress"
|
||||
echo ${Cmd}
|
||||
eval ${Cmd}
|
||||
|
||||
echo "${0} ${*} Done."
|
||||
@ -3,7 +3,7 @@
|
||||
usage() {
|
||||
echo -n "Usage: ${0}"
|
||||
echo -n " [-c <config (dflt: apps/cvtt_eqt_alpaca)>]"
|
||||
echo -n " [-s <config_serverice (dflt: http://cloud16.cvtt.vpn:6789)>]"
|
||||
echo -n " [-s <config_serverice (dflt: http://cloud23.cvtt.vpn:6789)>]"
|
||||
echo -n " [-N <name (dflt: ALPACA-SNDBX)>]"
|
||||
echo
|
||||
exit 1
|
||||
@ -23,7 +23,7 @@ export PYTHONPATH=${RootDir}
|
||||
StatusChannel=Status-CVTT
|
||||
AlertChannel=Alerts-CVTT
|
||||
Sender=${RootDir}/ops/utils/send_mmost.sh
|
||||
ConfigService=http://cloud16.cvtt.vpn:6789
|
||||
ConfigService=http://cloud23.cvtt.vpn:6789
|
||||
Config=apps/cvtt_eqt_alpaca
|
||||
Name="ALPACA-SNDBX"
|
||||
|
||||
|
||||
@ -5,15 +5,15 @@
|
||||
# "users": ["oleg"],
|
||||
# "type": "cloud",
|
||||
# "ssh_port": 7822,
|
||||
# "to_check": "false"
|
||||
# "timeout_sec": 5
|
||||
# # "to_check": "No"
|
||||
# },
|
||||
|
||||
RootDir="${HOME}/prod"
|
||||
# RootDir=/home/oleg/develop/cvtt2 ###### D E B U G
|
||||
|
||||
AlertChannel=Alerts-CVTT
|
||||
Sender=${RootDir}/ops/utils/send_mmost.sh
|
||||
ConfigUrl=http://cloud16.cvtt.vpn:6789/admin/cvtt_hosts
|
||||
ConfigUrl=http://cloud23.cvtt.vpn:6789/admin/cvtt_hosts
|
||||
|
||||
HOSTS_CONFIG=$(curl -s ${ConfigUrl} | ${HOME}/bin/hjson -j)
|
||||
|
||||
@ -46,8 +46,6 @@ function host_alert() {
|
||||
User=oleg
|
||||
Hosts=()
|
||||
DEFAULT_SSH_PORT=22
|
||||
DEFAULT_TIMEOUT=5
|
||||
DEFAULT_TO_CHECK="true"
|
||||
Domains=("${Domains[@]}" "$(get_domains)")
|
||||
|
||||
for Domain in ${Domains[@]} ; do
|
||||
@ -58,21 +56,16 @@ for Host in ${Hosts[@]} ; do
|
||||
host=$(echo $Host | cut -d'.' -f1)
|
||||
Domain=$(echo $Host | cut -d'.' -f2-)
|
||||
|
||||
|
||||
PortSSH=$(echo "$HOSTS_CONFIG" | jq -r --arg domain "$Domain" --arg host "$host" '.[$domain][$host].ssh_port // '"$DEFAULT_SSH_PORT"'')
|
||||
|
||||
Timeout=$(echo "$HOSTS_CONFIG" | jq -r --arg domain "$Domain" --arg host "$host" '.[$domain][$host].timeout_sec // '"$DEFAULT_TIMEOUT"'')
|
||||
|
||||
ToCheck=$(echo "$HOSTS_CONFIG" | jq -r --arg domain "$Domain" --arg host "$host" '.[$domain][$host].to_check // '"$DEFAULT_TO_CHECK"'')
|
||||
to_check="${ToCheck^^}"
|
||||
if [ "${to_check}" == "TRUE" -o "${to_check}" == "YES" -o "${to_check}" == "Y" -o "${to_check}" == "T" ] ; then
|
||||
echo "Checking host: $Host on port $PortSSH"
|
||||
else
|
||||
ToCheck=$(echo "$HOSTS_CONFIG" | jq -r --arg domain "$Domain" --arg host "$host" '.[$domain][$host].to_check // "Yes"')
|
||||
if [ "${ToCheck^^}" == "NO" ]; then
|
||||
continue
|
||||
fi
|
||||
PortSSH=$(echo "$HOSTS_CONFIG" | jq -r --arg domain "$Domain" --arg host "$host" '.[$domain][$host].ssh_port // '"$DEFAULT_SSH_PORT"'')
|
||||
|
||||
echo "Checking host: $Host on port $PortSSH"
|
||||
|
||||
# Use nc to check if the specified port is open
|
||||
if ! nc -z -w ${Timeout} "$Host" "$PortSSH"; then
|
||||
if ! nc -z -w5 "$Host" "$PortSSH"; then
|
||||
echo "Host $Host is not available on port $PortSSH"
|
||||
host_alert "Host $Host is not available on port $PortSSH"
|
||||
fi
|
||||
|
||||
@ -12,7 +12,7 @@ RootDir="${HOME}/prod"
|
||||
|
||||
AlertChannel=Alerts-CVTT
|
||||
Sender=${RootDir}/ops/utils/send_mmost.sh
|
||||
ConfigUrl=http://cloud16.cvtt.vpn:6789/admin/cvtt_services
|
||||
ConfigUrl=http://cloud23.cvtt.vpn:6789/admin/cvtt_services
|
||||
|
||||
SERVICES_CONFIG=$(curl -s ${ConfigUrl} | ${HOME}/bin/hjson -j)
|
||||
|
||||
|
||||
@ -18,14 +18,14 @@ get_user_hosts() {
|
||||
local User=${1}
|
||||
local Domain=${2}
|
||||
|
||||
Cmd="curl -s http://cloud16.cvtt.vpn:6789/admin/cvtt_hosts"
|
||||
Cmd+=" | ${HOME}/bin/hjson -j"
|
||||
Cmd+=" | jq -r"
|
||||
Cmd+=" --arg domain \"${Domain}\""
|
||||
Cmd+=" --arg usr \"${User}\""
|
||||
Cmd+=" '.[\$domain] | to_entries[] | select(.value.users[] | contains(\$usr)) | .key'"
|
||||
Cmd="curl -s http://cloud23.cvtt.vpn:6789/admin/cvtt_hosts"
|
||||
Cmd="${Cmd} | ${HOME}/bin/hjson -j"
|
||||
Cmd="${Cmd} | jq -r"
|
||||
Cmd="${Cmd} --arg domain \"${Domain}\""
|
||||
Cmd="${Cmd} --arg usr \"${User}\""
|
||||
Cmd="${Cmd} '.[\$domain] | to_entries[] | select(.value.users[] | contains(\$usr)) | .key'"
|
||||
|
||||
Cmd+=" | sed 's/\$/.${Domain}/'"
|
||||
Cmd="${Cmd} | sed 's/\$/.${Domain}/'"
|
||||
eval ${Cmd}
|
||||
}
|
||||
|
||||
@ -66,16 +66,13 @@ function storage_check() {
|
||||
else
|
||||
port=22
|
||||
fi
|
||||
Cmd="ssh -p ${port}"
|
||||
Cmd+=" -o StrictHostKeyChecking=no"
|
||||
Cmd+=" -o UserKnownHostsFile=/dev/null"
|
||||
Cmd+=" $host"
|
||||
Cmd+=" eval \"df -hTl"
|
||||
Cmd+=" -x squashfs"
|
||||
Cmd+=" -x tmpfs"
|
||||
Cmd+=" -x vfat"
|
||||
Cmd+=" -x devtmpfs"
|
||||
Cmd+=" | grep -v Filesystem\""
|
||||
Cmd="ssh -p ${port} $host"
|
||||
Cmd="${Cmd} eval \"df -hTl"
|
||||
Cmd="${Cmd} -x squashfs"
|
||||
Cmd="${Cmd} -x tmpfs"
|
||||
Cmd="${Cmd} -x vfat"
|
||||
Cmd="${Cmd} -x devtmpfs"
|
||||
Cmd="${Cmd} | grep -v Filesystem\""
|
||||
|
||||
IFS=$'\n' ; lines=$(eval ${Cmd})
|
||||
for ln in $lines
|
||||
|
||||
@ -4,7 +4,7 @@ Start=${1}
|
||||
NumJobs=${2}
|
||||
InstListFile=${3}
|
||||
|
||||
export CalendarURL=https://trading-calendar.cvtt.net/api/v1/markets/hours?mic=XNYS
|
||||
export CalendarURL=http://cloud16.cvtt.vpn:8000/api/v1/markets/hours?mic=XNYS
|
||||
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
|
||||
@ -5,7 +5,7 @@ export PYTHONPATH=${HOME}/prod
|
||||
Python=${HOME}/.pyenv/python3.12-venv/bin/python3
|
||||
Script=${HOME}/prod/cvttpy/exchanges/alpaca/hist_md/rl_calc_loader.py
|
||||
DbFile=${HOME}/prod/data/rel_liquidity.db
|
||||
Config=http://cloud16.cvtt.vpn:6789/apps/minimal_md_eqt
|
||||
Config=http://cloud23.cvtt.vpn:6789/apps/minimal_md_eqt
|
||||
|
||||
Cmd="${Python}"
|
||||
Cmd="${Cmd} ${Script}"
|
||||
|
||||
@ -3,7 +3,7 @@
|
||||
is_business_day() {
|
||||
dt=${1}
|
||||
|
||||
CalendarURL=https://trading-calendar.cvtt.net/api/v1/markets/hours?mic=XNYS
|
||||
CalendarURL=http://cloud16.cvtt.vpn:8000/api/v1/markets/hours?mic=XNYS
|
||||
open_time=$(curl -s "${CalendarURL}&start=${dt}&end=${dt}" | jq '.[] | .open_time')
|
||||
if [ -n "${open_time}" ]; then
|
||||
return 0
|
||||
|
||||
@ -25,12 +25,10 @@ run_checklist() {
|
||||
|
||||
declare -A Commands
|
||||
Commands=(
|
||||
["hs01:cloud21"]="ssh cvtt@hs01.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cloud21/${yr}/${mn} | tail -5"
|
||||
["hs01:cloud28"]="ssh cvtt@hs01.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cloud28/${yr}/${mn} | tail -5"
|
||||
["hs01:sim"]="ssh cvtt@hs01.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/sim/ | tail -5"
|
||||
["cloud21:cloud21"]="ssh cvtt@cloud21.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cloud21/${yr}/${mn} | tail -5"
|
||||
["cloud21:cloud28"]="ssh cvtt@cloud21.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cloud28/${yr}/${mn} | tail -5"
|
||||
["cloud21:sim"]="ssh cvtt@cloud21.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/sim | tail -5"
|
||||
["homestore:cloud21"]="ssh cvtt@homestore.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cloud21/${yr}/${mn} | tail -5"
|
||||
["homestore:cvttdata"]="ssh cvtt@homestore.cvtt.vpn ls -l /works/cvtt/md_archive/crypto/cvttdata/${yr}/${mn} | tail -5"
|
||||
["cloud21:cloud21"]="ssh cvtt@cloud21.cvtt.vpn ls -l /opt/store/cvtt/md_archive/crypto/cloud21/${yr}/${mn} | tail -5"
|
||||
["cloud21:cvttdata"]="ssh cvtt@cloud21.cvtt.vpn ls -l /opt/store/cvtt/md_archive/crypto/cvttdata/${yr}/${mn} | tail -5"
|
||||
["gpushnik"]="ssh oleg@gpushnik.cvtt.vpn 'ls -l /opt/jupyter_gpu/data/crypto_md | tail -10'"
|
||||
)
|
||||
|
||||
|
||||
@ -24,19 +24,17 @@ run_checklist() {
|
||||
yr=$(date -d 'yesterday' '+%Y')
|
||||
CheckSymbols="A/AAPL N/NVDA M/META"
|
||||
declare -A Commands
|
||||
Commands["hs01"]=""
|
||||
Commands["homestore"]=""
|
||||
for sym in ${CheckSymbols}; do
|
||||
Commands["hs01"]+="ssh cvtt@hs01.cvtt.vpn ls -l /works/cvtt/md_archive/equity/alpaca_md/${yr}/${sym} | tail -3; "
|
||||
Commands["homestore"]+="ssh cvtt@homestore.cvtt.vpn ls -l /works/cvtt/md_archive/equity/alpaca_md/${yr}/${sym} | tail -3; "
|
||||
done
|
||||
Commands["hs01"]+="echo"
|
||||
Commands["hs01:sim"]="ssh cvtt@hs01.cvtt.vpn ls -l /works/cvtt/md_archive/equity/alpaca_md/sim | tail -5"
|
||||
Commands["homestore"]+="echo"
|
||||
|
||||
Commands["cloud21"]=""
|
||||
for sym in ${CheckSymbols}; do
|
||||
Commands["cloud21"]+="ssh cvtt@cloud21.cvtt.vpn ls -l /works/cvtt/md_archive/equity/alpaca_md/${yr}/${sym} | tail -3; "
|
||||
Commands["cloud21"]+="ssh cvtt@cloud21.cvtt.vpn ls -l /opt/store/cvtt/md_archive/equity/alpaca_md/${yr}/${sym} | tail -3; "
|
||||
done
|
||||
Commands["cloud21"]+="echo"
|
||||
Commands["cloud21:sim"]="ssh cvtt@cloud21.cvtt.vpn ls -l /works/cvtt/md_archive/equity/alpaca_md/sim | tail -5"
|
||||
|
||||
Commands["gpushnik"]="ssh oleg@gpushnik.cvtt.vpn 'ls -l /opt/jupyter_gpu/data/eqty_md | tail -10'"
|
||||
|
||||
@ -48,13 +46,6 @@ run_checklist() {
|
||||
done
|
||||
}
|
||||
|
||||
tmpfile=$(mktemp)
|
||||
function cleanup {
|
||||
cd ${HOME}
|
||||
rm -f ${tmpfile}
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
# run_checklist
|
||||
|
||||
tmpfile=$(mktemp)
|
||||
|
||||
@ -2,7 +2,7 @@
|
||||
|
||||
Source=/home/cvtt/prod/archive/md_archive/
|
||||
Targets=
|
||||
Targets="${Targets} cvtt@hs01.cvtt.vpn:/works/cvtt/md_archive/"
|
||||
Targets="${Targets} cvtt@homestore.cvtt.vpn:/works/cvtt/md_archive/"
|
||||
Targets="${Targets} cvtt@cloud21.cvtt.vpn:/opt/store/cvtt/md_archive/"
|
||||
|
||||
|
||||
|
||||
@ -47,7 +47,7 @@ function rsync_load_version() {
|
||||
if ssh -q -p ${port} ${user}@${host} "test -d ${rel_dir}/${Project}/${Version}"
|
||||
then
|
||||
echo "Directory found..."
|
||||
rsync_cmd="rsync -ahv -e \"ssh -p ${port}\""
|
||||
rsync_cmd="rsync -ahvv -e \"ssh -p ${port}\""
|
||||
rsync_cmd="${rsync_cmd} ${user}@${host}:${rel_dir}/${Project}/${Version}"
|
||||
rsync_cmd="${rsync_cmd} ${LocalSoftwareDir}/${Project}/"
|
||||
echo ${rsync_cmd}
|
||||
|
||||
@ -6,7 +6,7 @@ SUBDIRS += configs
|
||||
|
||||
|
||||
FILES=
|
||||
FILES += VERSION
|
||||
FILES += release_version.txt
|
||||
|
||||
all: install
|
||||
|
||||
|
||||
@ -1,63 +1,36 @@
|
||||
#!/bin/bash
|
||||
|
||||
|
||||
function usage {
|
||||
echo -n "Usage: ${0}"
|
||||
echo -n " -L <log directory>"
|
||||
echo -n " [ -A <archive_logs_dir> (default /works/archive/logs)]"
|
||||
echo -n " [-D <older than time criteria> (default: '2 days ago')]"
|
||||
echo
|
||||
echo "Usage: ${0} <log directory> [days (default 2)]"
|
||||
exit 1
|
||||
}
|
||||
|
||||
echo Starting $0 $*
|
||||
|
||||
# ---- D e f a u l t s
|
||||
LogArchiveDir=/works/archive/logs
|
||||
DateCriteria="2 days ago"
|
||||
# ---- D e f a u l t s
|
||||
|
||||
|
||||
# ---------------- cmdline
|
||||
while getopts "A:L:D:" opt; do
|
||||
case ${opt} in
|
||||
A )
|
||||
LogArchiveDir=$OPTARG
|
||||
;;
|
||||
L )
|
||||
LogDir=$OPTARG
|
||||
;;
|
||||
D )
|
||||
DateCriteria=$OPTARG
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
;;
|
||||
: )
|
||||
echo "Option -$OPTARG requires an argument." >&2
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
# ---------------- cmdline
|
||||
|
||||
|
||||
LogDir=${1}
|
||||
if [ "${LogDir}" == "" ]
|
||||
then
|
||||
usage
|
||||
fi
|
||||
|
||||
Cmd="mkdir -p ${LogArchiveDir}"
|
||||
echo ${Cmd} && eval ${Cmd}
|
||||
Days=${2}
|
||||
if [ "${Days}" == "" ]
|
||||
then
|
||||
Days=2
|
||||
fi
|
||||
|
||||
echo "Looking for log files older than '${DateCriteria}' in ${LogDir}"
|
||||
DateCriteria="${Days} days ago"
|
||||
|
||||
Oldest=$(date -d "${DateCriteria}" '+%Y-%m-%d %H:%M:%S')
|
||||
Oldest=$(date -d "${DateCriteria}" '+%Y-%m-%d')
|
||||
|
||||
Cmd="find ${LogDir}/ '(' -name '*.log' -o -name '*.log.*' ')' -type f -not -newermt \"${Oldest}\""
|
||||
echo $Cmd
|
||||
echo "Looking for log files older than ${DateCriteria} in ${LogDir}"
|
||||
|
||||
files=$(eval ${Cmd})
|
||||
LogArchiveDir=${HOME}/prod/archive/logs
|
||||
mkdir -p ${LogArchiveDir}
|
||||
|
||||
echo "find ${LogDir}/ '(' -name '*.log' -o -name '*.log.*' ')' -type f -not -newermt ${Oldest})"
|
||||
files=$(find ${LogDir}/ '(' -name '*.log' -o -name '*.log.*' ')' -type f -not -newermt ${Oldest})
|
||||
if [ "$files" == "" ]
|
||||
then
|
||||
echo "No files found older than ${Oldest} in ${LogDir}"
|
||||
|
||||
@ -1,57 +1,40 @@
|
||||
#!/bin/bash
|
||||
# FOR hosts with limited disk space - move to storage server
|
||||
|
||||
# FOR cloud hosts with limited disk space - move to storage server
|
||||
|
||||
function usage {
|
||||
echo -n "Usage: ${0}"
|
||||
echo -n " -H <host_label>"
|
||||
echo -n " [ -A <archive_dir> (default /works/archive)]"
|
||||
echo -n " [-D <older than time criteria> (default: '2 days ago')]"
|
||||
echo
|
||||
echo "Usage: ${0} <host> <from_dir> <days>"
|
||||
exit 1
|
||||
}
|
||||
|
||||
echo Starting $0 $*
|
||||
|
||||
# ---- D e f a u l t s
|
||||
ArchiveDir=/works/archive
|
||||
DateCriteria="2 days ago"
|
||||
|
||||
FromHost=$(hostname -s)
|
||||
# ---- D e f a u l t s
|
||||
|
||||
# ---------------- cmdline
|
||||
while getopts "A:H:D:" opt; do
|
||||
case ${opt} in
|
||||
A )
|
||||
ArchiveDir=$OPTARG
|
||||
;;
|
||||
H )
|
||||
FromHost=$OPTARG
|
||||
;;
|
||||
D )
|
||||
DateCriteria=$OPTARG
|
||||
;;
|
||||
\? )
|
||||
echo "Invalid option: -$OPTARG" >&2
|
||||
usage
|
||||
;;
|
||||
: )
|
||||
echo "Option -$OPTARG requires an argument." >&2
|
||||
usage
|
||||
;;
|
||||
esac
|
||||
done
|
||||
# ---------------- cmdline
|
||||
|
||||
|
||||
if [ "${FromHost}" == "" ]
|
||||
then
|
||||
usage
|
||||
fi
|
||||
|
||||
ArchiveDir=${2}
|
||||
if [ "${ArchiveDir}" == "" ]
|
||||
then
|
||||
usage
|
||||
fi
|
||||
|
||||
Days=${3}
|
||||
if [ "${Days}" == "" ]
|
||||
then
|
||||
Days=2
|
||||
fi
|
||||
|
||||
DateCriteria="${Days} days ago"
|
||||
|
||||
TargetHost=cloud21.cvtt.vpn
|
||||
TargetRootDir=/opt/store/cvtt/archive
|
||||
|
||||
Oldest=$(date -d "${DateCriteria}" '+%Y-%m-%d %H:%M:%S')
|
||||
Oldest=$(date -d "${DateCriteria}" '+%Y-%m-%d')
|
||||
Now=$(date '+%Y%m%d_%H%M%S')
|
||||
|
||||
echo "Looking for log files older than ${DateCriteria} in ${ArchiveDir}"
|
||||
Cmd="find ${ArchiveDir}/"
|
||||
@ -62,7 +45,7 @@ Cmd="${Cmd} -o -name '*.logs.*'"
|
||||
Cmd="${Cmd} -o -name '*.tgz'"
|
||||
Cmd="${Cmd} ')'"
|
||||
Cmd="${Cmd} -type f"
|
||||
Cmd="${Cmd} -not -newermt \"${Oldest}\""
|
||||
Cmd="${Cmd} -not -newermt ${Oldest}"
|
||||
|
||||
echo ${Cmd}
|
||||
files=$(eval ${Cmd})
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user