commit
b73adda637
27
.travis.yml
27
.travis.yml
|
@ -1,4 +1,4 @@
|
||||||
dist: trusty
|
dist: xenial
|
||||||
version: ~> 1.0
|
version: ~> 1.0
|
||||||
language: php
|
language: php
|
||||||
php:
|
php:
|
||||||
|
@ -10,8 +10,6 @@ php:
|
||||||
- 7.0
|
- 7.0
|
||||||
# folks who prefer running on 5.x should be using 5.6 in most cases, 5.4 is no in the matrix since noone should use it
|
# folks who prefer running on 5.x should be using 5.6 in most cases, 5.4 is no in the matrix since noone should use it
|
||||||
- 5.6
|
- 5.6
|
||||||
# this is in for centos support, it's still the default on CentOS 7.3 and there were some lang changes after 5.4
|
|
||||||
- 5.4
|
|
||||||
services:
|
services:
|
||||||
- postgresql
|
- postgresql
|
||||||
- rabbitmq
|
- rabbitmq
|
||||||
|
@ -49,6 +47,13 @@ addons:
|
||||||
apt:
|
apt:
|
||||||
packages:
|
packages:
|
||||||
- silan
|
- silan
|
||||||
|
- libgirepository1.0-dev
|
||||||
|
- gir1.2-gstreamer-1.0
|
||||||
|
- gstreamer1.0-plugins-base
|
||||||
|
- gstreamer1.0-plugins-good
|
||||||
|
- gstreamer1.0-plugins-bad
|
||||||
|
- gstreamer1.0-plugins-ugly
|
||||||
|
- libcairo2-dev
|
||||||
- liquidsoap
|
- liquidsoap
|
||||||
- liquidsoap-plugin-mad
|
- liquidsoap-plugin-mad
|
||||||
- liquidsoap-plugin-taglib
|
- liquidsoap-plugin-taglib
|
||||||
|
@ -58,10 +63,10 @@ addons:
|
||||||
- liquidsoap-plugin-faad
|
- liquidsoap-plugin-faad
|
||||||
- liquidsoap-plugin-vorbis
|
- liquidsoap-plugin-vorbis
|
||||||
- liquidsoap-plugin-opus
|
- liquidsoap-plugin-opus
|
||||||
- python-nose
|
- python3
|
||||||
- python-rgain
|
- python3-nose
|
||||||
- python-gst-1.0
|
- python3-gst-1.0
|
||||||
- python-magic
|
- python3-magic
|
||||||
- dos2unix
|
- dos2unix
|
||||||
install:
|
install:
|
||||||
- >
|
- >
|
||||||
|
@ -70,9 +75,11 @@ install:
|
||||||
fi
|
fi
|
||||||
- >
|
- >
|
||||||
if [[ "$PYTHON" == true ]]; then
|
if [[ "$PYTHON" == true ]]; then
|
||||||
pip install --user mkdocs
|
pyenv local 3.7
|
||||||
pushd python_apps/airtime_analyzer
|
pip3 install -U pip wheel
|
||||||
python setup.py install --dry-run --no-init-script
|
pip3 install --user mkdocs rgain3
|
||||||
|
pushd python_apps/airtime_analyzer
|
||||||
|
python3 setup.py install --dry-run --no-init-script
|
||||||
popd
|
popd
|
||||||
fi
|
fi
|
||||||
before_script:
|
before_script:
|
||||||
|
|
|
@ -3,8 +3,6 @@
|
||||||
echo "Updating Apt."
|
echo "Updating Apt."
|
||||||
apt-get update > /dev/null
|
apt-get update > /dev/null
|
||||||
echo "Ensuring Pip is installed."
|
echo "Ensuring Pip is installed."
|
||||||
DEBIAN_FRONTEND=noninteractive apt-get install -y -qq python-pip > /dev/null
|
DEBIAN_FRONTEND=noninteractive apt-get install -y -qq python3-pip > /dev/null
|
||||||
echo "Updating Pip."
|
|
||||||
pip install pip -q -q --upgrade > /dev/null
|
|
||||||
echo "Ensuring Mkdocs is installed."
|
echo "Ensuring Mkdocs is installed."
|
||||||
pip install -q mkdocs > /dev/null
|
pip3 install mkdocs
|
||||||
|
|
30
install
30
install
|
@ -465,7 +465,7 @@ while :; do
|
||||||
;;
|
;;
|
||||||
--no-rabbitmq)
|
--no-rabbitmq)
|
||||||
skip_rabbitmq=1
|
skip_rabbitmq=1
|
||||||
;;
|
;;
|
||||||
--)
|
--)
|
||||||
shift
|
shift
|
||||||
break
|
break
|
||||||
|
@ -923,22 +923,12 @@ loud "\n-----------------------------------------------------"
|
||||||
loud " * Installing Airtime Services * "
|
loud " * Installing Airtime Services * "
|
||||||
loud "-----------------------------------------------------"
|
loud "-----------------------------------------------------"
|
||||||
|
|
||||||
verbose "\n * Installing necessary python services..."
|
python_version=$(python3 --version 2>&1 | awk '{ print $2 }')
|
||||||
loudCmd "pip install setuptools --upgrade"
|
|
||||||
loudCmd "pip install zipp==1.0.0"
|
|
||||||
verbose "...Done"
|
|
||||||
|
|
||||||
# Ubuntu Trusty and Debian Wheezy needs a workaround for python version SSL downloads
|
|
||||||
# This affects all python installs where python < 2.7.9
|
|
||||||
python_version=$(python --version 2>&1 | awk '{ print $2 }')
|
|
||||||
verbose "Detected Python version: $python_version"
|
verbose "Detected Python version: $python_version"
|
||||||
# Convert version so each segment is zero padded for easy comparison
|
|
||||||
python_version_formatted=$(awk 'BEGIN {FS = "."} {printf "%03d.%03d.%03d\n", $1,$2,$3}' <<< $python_version)
|
verbose "\n * Installing necessary python services..."
|
||||||
if [[ "$python_version_formatted" < "002.007.009" ]]; then
|
loudCmd "pip3 install setuptools --upgrade"
|
||||||
verbose "\n * Installing pyOpenSSL and ca db for SNI support..."
|
verbose "...Done"
|
||||||
loudCmd "pip install pyOpenSSL cryptography idna certifi --upgrade"
|
|
||||||
verbose "...Done"
|
|
||||||
fi
|
|
||||||
|
|
||||||
verbose "\n * Creating /run/airtime..."
|
verbose "\n * Creating /run/airtime..."
|
||||||
mkdir -p /run/airtime
|
mkdir -p /run/airtime
|
||||||
|
@ -960,11 +950,11 @@ if [ ! -d /var/log/airtime ]; then
|
||||||
fi
|
fi
|
||||||
|
|
||||||
verbose "\n * Installing API client..."
|
verbose "\n * Installing API client..."
|
||||||
loudCmd "python ${AIRTIMEROOT}/python_apps/api_clients/setup.py install --install-scripts=/usr/bin"
|
loudCmd "python3 ${AIRTIMEROOT}/python_apps/api_clients/setup.py install --install-scripts=/usr/bin"
|
||||||
verbose "...Done"
|
verbose "...Done"
|
||||||
|
|
||||||
verbose "\n * Installing pypo and liquidsoap..."
|
verbose "\n * Installing pypo and liquidsoap..."
|
||||||
loudCmd "python ${AIRTIMEROOT}/python_apps/pypo/setup.py install --install-scripts=/usr/bin --no-init-script"
|
loudCmd "python3 ${AIRTIMEROOT}/python_apps/pypo/setup.py install --install-scripts=/usr/bin --no-init-script"
|
||||||
loudCmd "mkdir -p /var/log/airtime/{pypo,pypo-liquidsoap} /var/tmp/airtime/pypo/{cache,files,tmp} /var/tmp/airtime/show-recorder/"
|
loudCmd "mkdir -p /var/log/airtime/{pypo,pypo-liquidsoap} /var/tmp/airtime/pypo/{cache,files,tmp} /var/tmp/airtime/show-recorder/"
|
||||||
loudCmd "chown -R ${web_user}:${web_user} /var/log/airtime/{pypo,pypo-liquidsoap} /var/tmp/airtime/pypo/{cache,files,tmp} /var/tmp/airtime/show-recorder/"
|
loudCmd "chown -R ${web_user}:${web_user} /var/log/airtime/{pypo,pypo-liquidsoap} /var/tmp/airtime/pypo/{cache,files,tmp} /var/tmp/airtime/show-recorder/"
|
||||||
systemInitInstall airtime-liquidsoap $web_user
|
systemInitInstall airtime-liquidsoap $web_user
|
||||||
|
@ -972,7 +962,7 @@ systemInitInstall airtime-playout $web_user
|
||||||
verbose "...Done"
|
verbose "...Done"
|
||||||
|
|
||||||
verbose "\n * Installing airtime-celery..."
|
verbose "\n * Installing airtime-celery..."
|
||||||
loudCmd "python ${AIRTIMEROOT}/python_apps/airtime-celery/setup.py install --no-init-script"
|
loudCmd "python3 ${AIRTIMEROOT}/python_apps/airtime-celery/setup.py install --no-init-script"
|
||||||
# Create the Celery user
|
# Create the Celery user
|
||||||
if $is_centos_dist; then
|
if $is_centos_dist; then
|
||||||
loudCmd "id celery 2>/dev/null || adduser --no-create-home -c 'LibreTime Celery' -r celery || true"
|
loudCmd "id celery 2>/dev/null || adduser --no-create-home -c 'LibreTime Celery' -r celery || true"
|
||||||
|
@ -988,7 +978,7 @@ systemInitInstall airtime-celery
|
||||||
verbose "...Done"
|
verbose "...Done"
|
||||||
|
|
||||||
verbose "\n * Installing airtime_analyzer..."
|
verbose "\n * Installing airtime_analyzer..."
|
||||||
loudCmd "python ${AIRTIMEROOT}/python_apps/airtime_analyzer/setup.py install --install-scripts=/usr/bin --no-init-script"
|
loudCmd "python3 ${AIRTIMEROOT}/python_apps/airtime_analyzer/setup.py install --install-scripts=/usr/bin --no-init-script"
|
||||||
systemInitInstall airtime_analyzer $web_user
|
systemInitInstall airtime_analyzer $web_user
|
||||||
verbose "...Done"
|
verbose "...Done"
|
||||||
|
|
||||||
|
|
|
@ -1,71 +1,59 @@
|
||||||
apache2
|
apache2
|
||||||
|
coreutils
|
||||||
|
curl
|
||||||
|
ecasound
|
||||||
|
flac
|
||||||
git
|
git
|
||||||
|
gstreamer1.0-plugins-bad
|
||||||
|
gstreamer1.0-plugins-good
|
||||||
|
gstreamer1.0-plugins-ugly
|
||||||
|
icecast2
|
||||||
|
lame
|
||||||
|
libao-ocaml
|
||||||
libapache2-mod-php7.3
|
libapache2-mod-php7.3
|
||||||
php7.3
|
libcairo2-dev
|
||||||
php7.3-dev
|
libcamomile-ocaml-data
|
||||||
php7.3-bcmath
|
libfaad2
|
||||||
php7.3-mbstring
|
libmad-ocaml
|
||||||
php-pear
|
libopus0
|
||||||
php7.3-gd
|
libportaudio2
|
||||||
php-amqplib
|
libpulse0
|
||||||
|
libsamplerate0
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php7.3-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
libsoundtouch-ocaml
|
||||||
libtaglib-ocaml
|
libtaglib-ocaml
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
libvo-aacenc0
|
libvo-aacenc0
|
||||||
|
liquidsoap
|
||||||
python-rgain
|
lsb-release
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
curl
|
|
||||||
php7.3-curl
|
|
||||||
mpg123
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsb-release
|
lsb-release
|
||||||
lsof
|
lsof
|
||||||
vorbisgain
|
mpg123
|
||||||
flac
|
patch
|
||||||
vorbis-tools
|
php7.3
|
||||||
pwgen
|
php7.3-bcmath
|
||||||
libfaad2
|
php7.3-curl
|
||||||
|
php7.3-dev
|
||||||
|
php7.3-gd
|
||||||
|
php7.3-mbstring
|
||||||
|
php7.3-pgsql
|
||||||
|
php-amqplib
|
||||||
php-apcu
|
php-apcu
|
||||||
|
php-pear
|
||||||
lame
|
pkg-config
|
||||||
|
postgresql
|
||||||
|
postgresql-client
|
||||||
|
pwgen
|
||||||
|
python3
|
||||||
|
python3-gst-1.0
|
||||||
|
python3-pika
|
||||||
|
python3-pip
|
||||||
|
python3-virtualenv
|
||||||
|
python3-cairo
|
||||||
|
rabbitmq-server
|
||||||
silan
|
silan
|
||||||
coreutils
|
|
||||||
|
|
||||||
liquidsoap
|
|
||||||
|
|
||||||
libopus0
|
|
||||||
|
|
||||||
systemd-sysv
|
systemd-sysv
|
||||||
|
unzip
|
||||||
|
vorbisgain
|
||||||
|
vorbis-tools
|
||||||
|
vorbis-tools
|
||||||
xmlstarlet
|
xmlstarlet
|
||||||
|
zip
|
||||||
|
|
|
@ -1,67 +0,0 @@
|
||||||
apache2
|
|
||||||
libapache2-mod-php5
|
|
||||||
php5
|
|
||||||
php-pear
|
|
||||||
php5-gd
|
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php5-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
|
||||||
libtaglib-ocaml
|
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
libvo-aacenc0
|
|
||||||
|
|
||||||
python-rgain
|
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
curl
|
|
||||||
php5-curl
|
|
||||||
mpg123
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsb-release
|
|
||||||
lsof
|
|
||||||
vorbisgain
|
|
||||||
flac
|
|
||||||
vorbis-tools
|
|
||||||
pwgen
|
|
||||||
libfaad2
|
|
||||||
php-apc
|
|
||||||
|
|
||||||
lame
|
|
||||||
|
|
||||||
coreutils
|
|
||||||
|
|
||||||
liquidsoap
|
|
||||||
|
|
||||||
libopus0
|
|
||||||
|
|
||||||
sysvinit
|
|
||||||
sysvinit-utils
|
|
||||||
|
|
||||||
xmlstarlet
|
|
|
@ -1,71 +1,56 @@
|
||||||
apache2
|
apache2
|
||||||
|
coreutils
|
||||||
|
curl
|
||||||
|
ecasound
|
||||||
|
flac
|
||||||
git
|
git
|
||||||
|
gstreamer1.0-plugins-bad
|
||||||
|
gstreamer1.0-plugins-good
|
||||||
|
gstreamer1.0-plugins-ugly
|
||||||
|
icecast2
|
||||||
|
lame
|
||||||
|
libao-ocaml
|
||||||
libapache2-mod-php7.0
|
libapache2-mod-php7.0
|
||||||
php7.0
|
libcamomile-ocaml-data
|
||||||
php7.0-dev
|
libfaad2
|
||||||
php7.0-bcmath
|
libmad-ocaml
|
||||||
php7.0-mbstring
|
libopus0
|
||||||
php-pear
|
libportaudio2
|
||||||
php7.0-gd
|
libpulse0
|
||||||
php-amqplib
|
libsamplerate0
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php7.0-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
libsoundtouch-ocaml
|
||||||
libtaglib-ocaml
|
libtaglib-ocaml
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
libvo-aacenc0
|
libvo-aacenc0
|
||||||
|
liquidsoap
|
||||||
python-rgain
|
lsb-release
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
curl
|
|
||||||
php7.0-curl
|
|
||||||
mpg123
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsb-release
|
lsb-release
|
||||||
lsof
|
lsof
|
||||||
vorbisgain
|
mpg123
|
||||||
flac
|
patch
|
||||||
vorbis-tools
|
php7.0
|
||||||
pwgen
|
php7.0-bcmath
|
||||||
libfaad2
|
php7.0-curl
|
||||||
|
php7.0-dev
|
||||||
|
php7.0-gd
|
||||||
|
php7.0-mbstring
|
||||||
|
php7.0-pgsql
|
||||||
|
php-amqplib
|
||||||
php-apcu
|
php-apcu
|
||||||
|
php-pear
|
||||||
lame
|
postgresql
|
||||||
|
postgresql-client
|
||||||
coreutils
|
pwgen
|
||||||
|
python3
|
||||||
liquidsoap
|
python3-gst-1.0
|
||||||
|
python3-pika
|
||||||
libopus0
|
python3-pip
|
||||||
|
python3-virtualenv
|
||||||
|
python3-cairo
|
||||||
|
rabbitmq-server
|
||||||
systemd-sysv
|
systemd-sysv
|
||||||
|
unzip
|
||||||
|
vorbisgain
|
||||||
|
vorbis-tools
|
||||||
|
vorbis-tools
|
||||||
xmlstarlet
|
xmlstarlet
|
||||||
|
zip
|
||||||
|
|
|
@ -1,62 +1,27 @@
|
||||||
apache2
|
apache2
|
||||||
libapache2-mod-php7.2
|
build-essential
|
||||||
php7.2
|
|
||||||
php-pear
|
|
||||||
php7.2-gd
|
|
||||||
php-bcmath
|
|
||||||
php-mbstring
|
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php7.2-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
|
||||||
libtaglib-ocaml
|
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
|
|
||||||
python-rgain
|
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
php7.2-curl
|
|
||||||
mpg123
|
|
||||||
curl
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsof
|
|
||||||
vorbisgain
|
|
||||||
flac
|
|
||||||
vorbis-tools
|
|
||||||
pwgen
|
|
||||||
libfaad2
|
|
||||||
php-apcu
|
|
||||||
|
|
||||||
lame
|
|
||||||
|
|
||||||
coreutils
|
coreutils
|
||||||
|
curl
|
||||||
|
ecasound
|
||||||
|
flac
|
||||||
|
gstreamer1.0-plugins-bad
|
||||||
|
gstreamer1.0-plugins-good
|
||||||
|
gstreamer1.0-plugins-ugly
|
||||||
|
icecast2
|
||||||
|
lame
|
||||||
|
libao-ocaml
|
||||||
|
libapache2-mod-php7.2
|
||||||
|
libcamomile-ocaml-data
|
||||||
|
libfaad2
|
||||||
|
libffi-dev
|
||||||
|
libmad-ocaml
|
||||||
|
libopus0
|
||||||
|
libportaudio2
|
||||||
|
libpulse0
|
||||||
|
libsamplerate0
|
||||||
|
libsoundtouch-ocaml
|
||||||
|
libssl-dev
|
||||||
|
libtaglib-ocaml
|
||||||
liquidsoap
|
liquidsoap
|
||||||
liquidsoap-plugin-alsa
|
liquidsoap-plugin-alsa
|
||||||
liquidsoap-plugin-ao
|
liquidsoap-plugin-ao
|
||||||
|
@ -71,15 +36,34 @@ liquidsoap-plugin-pulseaudio
|
||||||
liquidsoap-plugin-taglib
|
liquidsoap-plugin-taglib
|
||||||
liquidsoap-plugin-voaacenc
|
liquidsoap-plugin-voaacenc
|
||||||
liquidsoap-plugin-vorbis
|
liquidsoap-plugin-vorbis
|
||||||
|
lsb-release
|
||||||
|
lsof
|
||||||
|
mpg123
|
||||||
|
patch
|
||||||
|
php7.2
|
||||||
|
php7.2-curl
|
||||||
|
php7.2-gd
|
||||||
|
php7.2-pgsql
|
||||||
|
php-apcu
|
||||||
|
php-bcmath
|
||||||
|
php-mbstring
|
||||||
|
php-pear
|
||||||
|
postgresql
|
||||||
|
postgresql-client
|
||||||
|
pwgen
|
||||||
|
python3
|
||||||
|
python3-dev
|
||||||
|
python3-gst-1.0
|
||||||
|
python3-pika
|
||||||
|
python3-pip
|
||||||
|
python3-virtualenv
|
||||||
|
python3-cairo
|
||||||
|
rabbitmq-server
|
||||||
silan
|
silan
|
||||||
libopus0
|
|
||||||
|
|
||||||
sysvinit-utils
|
sysvinit-utils
|
||||||
|
unzip
|
||||||
build-essential
|
vorbisgain
|
||||||
libssl-dev
|
vorbis-tools
|
||||||
libffi-dev
|
vorbis-tools
|
||||||
python-dev
|
|
||||||
|
|
||||||
xmlstarlet
|
xmlstarlet
|
||||||
|
zip
|
||||||
|
|
|
@ -1,74 +0,0 @@
|
||||||
apache2
|
|
||||||
libapache2-mod-php5
|
|
||||||
php5
|
|
||||||
php-pear
|
|
||||||
php5-gd
|
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php5-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
|
||||||
libtaglib-ocaml
|
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
|
|
||||||
python-rgain
|
|
||||||
python-gst0.10
|
|
||||||
gstreamer0.10-plugins-ugly
|
|
||||||
gir1.2-gstreamer-0.10
|
|
||||||
patch
|
|
||||||
|
|
||||||
curl
|
|
||||||
php5-curl
|
|
||||||
mpg123
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsb-release
|
|
||||||
lsof
|
|
||||||
vorbisgain
|
|
||||||
flac
|
|
||||||
vorbis-tools
|
|
||||||
pwgen
|
|
||||||
libfaad2
|
|
||||||
php-apc
|
|
||||||
dbus
|
|
||||||
|
|
||||||
lame
|
|
||||||
|
|
||||||
coreutils
|
|
||||||
|
|
||||||
liquidsoap
|
|
||||||
liquidsoap-plugin-alsa
|
|
||||||
liquidsoap-plugin-ao
|
|
||||||
liquidsoap-plugin-faad
|
|
||||||
liquidsoap-plugin-flac
|
|
||||||
liquidsoap-plugin-icecast
|
|
||||||
liquidsoap-plugin-lame
|
|
||||||
liquidsoap-plugin-mad
|
|
||||||
liquidsoap-plugin-ogg
|
|
||||||
liquidsoap-plugin-portaudio
|
|
||||||
liquidsoap-plugin-pulseaudio
|
|
||||||
liquidsoap-plugin-taglib
|
|
||||||
liquidsoap-plugin-voaacenc
|
|
||||||
liquidsoap-plugin-vorbis
|
|
||||||
|
|
||||||
xmlstarlet
|
|
|
@ -1,62 +1,27 @@
|
||||||
apache2
|
apache2
|
||||||
libapache2-mod-php7.0
|
build-essential
|
||||||
php7.0
|
|
||||||
php-pear
|
|
||||||
php7.0-gd
|
|
||||||
php-bcmath
|
|
||||||
php-mbstring
|
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
rabbitmq-server
|
|
||||||
|
|
||||||
postgresql
|
|
||||||
postgresql-client
|
|
||||||
php7.0-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
|
||||||
libtaglib-ocaml
|
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
|
|
||||||
python-rgain
|
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
php7.0-curl
|
|
||||||
mpg123
|
|
||||||
curl
|
|
||||||
|
|
||||||
icecast2
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsof
|
|
||||||
vorbisgain
|
|
||||||
flac
|
|
||||||
vorbis-tools
|
|
||||||
pwgen
|
|
||||||
libfaad2
|
|
||||||
php-apcu
|
|
||||||
|
|
||||||
lame
|
|
||||||
|
|
||||||
coreutils
|
coreutils
|
||||||
|
curl
|
||||||
|
ecasound
|
||||||
|
flac
|
||||||
|
gstreamer1.0-plugins-bad
|
||||||
|
gstreamer1.0-plugins-good
|
||||||
|
gstreamer1.0-plugins-ugly
|
||||||
|
icecast2
|
||||||
|
lame
|
||||||
|
libao-ocaml
|
||||||
|
libapache2-mod-php7.0
|
||||||
|
libcamomile-ocaml-data
|
||||||
|
libfaad2
|
||||||
|
libffi-dev
|
||||||
|
libmad-ocaml
|
||||||
|
libopus0
|
||||||
|
libportaudio2
|
||||||
|
libpulse0
|
||||||
|
libsamplerate0
|
||||||
|
libsoundtouch-ocaml
|
||||||
|
libssl-dev
|
||||||
|
libtaglib-ocaml
|
||||||
liquidsoap
|
liquidsoap
|
||||||
liquidsoap-plugin-alsa
|
liquidsoap-plugin-alsa
|
||||||
liquidsoap-plugin-ao
|
liquidsoap-plugin-ao
|
||||||
|
@ -71,15 +36,34 @@ liquidsoap-plugin-pulseaudio
|
||||||
liquidsoap-plugin-taglib
|
liquidsoap-plugin-taglib
|
||||||
liquidsoap-plugin-voaacenc
|
liquidsoap-plugin-voaacenc
|
||||||
liquidsoap-plugin-vorbis
|
liquidsoap-plugin-vorbis
|
||||||
|
lsb-release
|
||||||
|
lsof
|
||||||
|
mpg123
|
||||||
|
patch
|
||||||
|
php7.0
|
||||||
|
php7.0-curl
|
||||||
|
php7.0-gd
|
||||||
|
php7.0-pgsql
|
||||||
|
php-apcu
|
||||||
|
php-bcmath
|
||||||
|
php-mbstring
|
||||||
|
php-pear
|
||||||
|
postgresql
|
||||||
|
postgresql-client
|
||||||
|
pwgen
|
||||||
|
python3
|
||||||
|
python3-dev
|
||||||
|
python3-gst-1.0
|
||||||
|
python3-pika
|
||||||
|
python3-pip
|
||||||
|
python3-virtualenv
|
||||||
|
python3-cairo
|
||||||
|
rabbitmq-server
|
||||||
silan
|
silan
|
||||||
libopus0
|
|
||||||
|
|
||||||
sysvinit-utils
|
sysvinit-utils
|
||||||
|
unzip
|
||||||
build-essential
|
vorbisgain
|
||||||
libssl-dev
|
vorbis-tools
|
||||||
libffi-dev
|
vorbis-tools
|
||||||
python-dev
|
|
||||||
|
|
||||||
xmlstarlet
|
xmlstarlet
|
||||||
|
zip
|
||||||
|
|
|
@ -1,57 +1,26 @@
|
||||||
apache2
|
apache2
|
||||||
libapache2-mod-php7.0
|
build-essential
|
||||||
php7.0
|
|
||||||
php-pear
|
|
||||||
php7.0-gd
|
|
||||||
php-bcmath
|
|
||||||
php-mbstring
|
|
||||||
|
|
||||||
lsb-release
|
|
||||||
|
|
||||||
zip
|
|
||||||
unzip
|
|
||||||
|
|
||||||
postgresql-client
|
|
||||||
php7.0-pgsql
|
|
||||||
|
|
||||||
python
|
|
||||||
python-virtualenv
|
|
||||||
python-pip
|
|
||||||
|
|
||||||
libsoundtouch-ocaml
|
|
||||||
libtaglib-ocaml
|
|
||||||
libao-ocaml
|
|
||||||
libmad-ocaml
|
|
||||||
ecasound
|
|
||||||
libportaudio2
|
|
||||||
libsamplerate0
|
|
||||||
|
|
||||||
python-rgain
|
|
||||||
python-gst-1.0
|
|
||||||
gstreamer1.0-plugins-ugly
|
|
||||||
python-pika
|
|
||||||
|
|
||||||
patch
|
|
||||||
|
|
||||||
php7.0-curl
|
|
||||||
mpg123
|
|
||||||
curl
|
|
||||||
|
|
||||||
libcamomile-ocaml-data
|
|
||||||
libpulse0
|
|
||||||
vorbis-tools
|
|
||||||
lsof
|
|
||||||
vorbisgain
|
|
||||||
flac
|
|
||||||
vorbis-tools
|
|
||||||
pwgen
|
|
||||||
libfaad2
|
|
||||||
php-apcu
|
|
||||||
|
|
||||||
lame
|
|
||||||
|
|
||||||
coreutils
|
coreutils
|
||||||
|
curl
|
||||||
|
ecasound
|
||||||
|
flac
|
||||||
|
gstreamer1.0-plugins-bad
|
||||||
|
gstreamer1.0-plugins-good
|
||||||
|
gstreamer1.0-plugins-ugly
|
||||||
|
lame
|
||||||
|
libao-ocaml
|
||||||
|
libapache2-mod-php7.0
|
||||||
|
libcamomile-ocaml-data
|
||||||
|
libfaad2
|
||||||
|
libffi-dev
|
||||||
|
libmad-ocaml
|
||||||
|
libopus0
|
||||||
|
libportaudio2
|
||||||
|
libpulse0
|
||||||
|
libsamplerate0
|
||||||
|
libsoundtouch-ocaml
|
||||||
|
libssl-dev
|
||||||
|
libtaglib-ocaml
|
||||||
liquidsoap
|
liquidsoap
|
||||||
liquidsoap-plugin-alsa
|
liquidsoap-plugin-alsa
|
||||||
liquidsoap-plugin-ao
|
liquidsoap-plugin-ao
|
||||||
|
@ -66,15 +35,32 @@ liquidsoap-plugin-pulseaudio
|
||||||
liquidsoap-plugin-taglib
|
liquidsoap-plugin-taglib
|
||||||
liquidsoap-plugin-voaacenc
|
liquidsoap-plugin-voaacenc
|
||||||
liquidsoap-plugin-vorbis
|
liquidsoap-plugin-vorbis
|
||||||
|
lsb-release
|
||||||
|
lsof
|
||||||
|
mpg123
|
||||||
|
patch
|
||||||
|
php7.0
|
||||||
|
php7.0-curl
|
||||||
|
php7.0-gd
|
||||||
|
php7.0-pgsql
|
||||||
|
php-apcu
|
||||||
|
php-bcmath
|
||||||
|
php-mbstring
|
||||||
|
php-pear
|
||||||
|
postgresql-client
|
||||||
|
pwgen
|
||||||
|
python3
|
||||||
|
python3-dev
|
||||||
|
python3-gst-1.0
|
||||||
|
python3-pika
|
||||||
|
python3-pip
|
||||||
|
python3-virtualenv
|
||||||
|
python3-cairo
|
||||||
silan
|
silan
|
||||||
libopus0
|
|
||||||
|
|
||||||
sysvinit-utils
|
sysvinit-utils
|
||||||
|
unzip
|
||||||
build-essential
|
vorbisgain
|
||||||
libssl-dev
|
vorbis-tools
|
||||||
libffi-dev
|
vorbis-tools
|
||||||
python-dev
|
|
||||||
|
|
||||||
xmlstarlet
|
xmlstarlet
|
||||||
|
zip
|
||||||
|
|
|
@ -86,8 +86,12 @@ def soundcloud_download(token, callback_url, api_key, track_id):
|
||||||
auth=requests.auth.HTTPBasicAuth(api_key, ""),
|
auth=requests.auth.HTTPBasicAuth(api_key, ""),
|
||||||
)
|
)
|
||||||
re.raise_for_status()
|
re.raise_for_status()
|
||||||
|
try:
|
||||||
|
response = re.content.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
response = re.content
|
||||||
f = json.loads(
|
f = json.loads(
|
||||||
re.content
|
response
|
||||||
) # Read the response from the media API to get the file id
|
) # Read the response from the media API to get the file id
|
||||||
obj["fileid"] = f["id"]
|
obj["fileid"] = f["id"]
|
||||||
else:
|
else:
|
||||||
|
@ -203,8 +207,12 @@ def podcast_download(
|
||||||
auth=requests.auth.HTTPBasicAuth(api_key, ""),
|
auth=requests.auth.HTTPBasicAuth(api_key, ""),
|
||||||
)
|
)
|
||||||
re.raise_for_status()
|
re.raise_for_status()
|
||||||
|
try:
|
||||||
|
response = re.content.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
response = re.content
|
||||||
f = json.loads(
|
f = json.loads(
|
||||||
re.content
|
response
|
||||||
) # Read the response from the media API to get the file id
|
) # Read the response from the media API to get the file id
|
||||||
obj["fileid"] = f["id"]
|
obj["fileid"] = f["id"]
|
||||||
obj["status"] = 1
|
obj["status"] = 1
|
||||||
|
|
|
@ -45,7 +45,7 @@ setup(
|
||||||
author_email="duncan.sommerville@sourcefabric.org",
|
author_email="duncan.sommerville@sourcefabric.org",
|
||||||
license="MIT",
|
license="MIT",
|
||||||
packages=["airtime-celery"],
|
packages=["airtime-celery"],
|
||||||
install_requires=["soundcloud", "celery < 4", "kombu < 3.1", "configobj"],
|
install_requires=["soundcloud", "celery", "kombu", "configobj"],
|
||||||
zip_safe=False,
|
zip_safe=False,
|
||||||
data_files=data_files,
|
data_files=data_files,
|
||||||
)
|
)
|
||||||
|
|
|
@ -5,12 +5,12 @@ import logging.handlers
|
||||||
import sys
|
import sys
|
||||||
import signal
|
import signal
|
||||||
import traceback
|
import traceback
|
||||||
import config_file
|
from . import config_file
|
||||||
from functools import partial
|
from functools import partial
|
||||||
from metadata_analyzer import MetadataAnalyzer
|
from .metadata_analyzer import MetadataAnalyzer
|
||||||
from replaygain_analyzer import ReplayGainAnalyzer
|
from .replaygain_analyzer import ReplayGainAnalyzer
|
||||||
from status_reporter import StatusReporter
|
from .status_reporter import StatusReporter
|
||||||
from message_listener import MessageListener
|
from .message_listener import MessageListener
|
||||||
|
|
||||||
|
|
||||||
class AirtimeAnalyzerServer:
|
class AirtimeAnalyzerServer:
|
||||||
|
@ -76,7 +76,7 @@ class AirtimeAnalyzerServer:
|
||||||
def dump_stacktrace(stack):
|
def dump_stacktrace(stack):
|
||||||
''' Dump a stacktrace for all threads '''
|
''' Dump a stacktrace for all threads '''
|
||||||
code = []
|
code = []
|
||||||
for threadId, stack in sys._current_frames().items():
|
for threadId, stack in list(sys._current_frames().items()):
|
||||||
code.append("\n# ThreadID: %s" % threadId)
|
code.append("\n# ThreadID: %s" % threadId)
|
||||||
for filename, lineno, name, line in traceback.extract_stack(stack):
|
for filename, lineno, name, line in traceback.extract_stack(stack):
|
||||||
code.append('File: "%s", line %d, in %s' % (filename, lineno, name))
|
code.append('File: "%s", line %d, in %s' % (filename, lineno, name))
|
||||||
|
|
|
@ -1,19 +1,19 @@
|
||||||
""" Analyzes and imports an audio file into the Airtime library.
|
""" Analyzes and imports an audio file into the Airtime library.
|
||||||
"""
|
"""
|
||||||
import logging
|
import logging
|
||||||
import threading
|
import threading
|
||||||
import multiprocessing
|
import multiprocessing
|
||||||
import Queue
|
from queue import Queue
|
||||||
import ConfigParser
|
import configparser
|
||||||
from metadata_analyzer import MetadataAnalyzer
|
from .metadata_analyzer import MetadataAnalyzer
|
||||||
from filemover_analyzer import FileMoverAnalyzer
|
from .filemover_analyzer import FileMoverAnalyzer
|
||||||
from cuepoint_analyzer import CuePointAnalyzer
|
from .cuepoint_analyzer import CuePointAnalyzer
|
||||||
from replaygain_analyzer import ReplayGainAnalyzer
|
from .replaygain_analyzer import ReplayGainAnalyzer
|
||||||
from playability_analyzer import *
|
from .playability_analyzer import *
|
||||||
|
|
||||||
class AnalyzerPipeline:
|
class AnalyzerPipeline:
|
||||||
""" Analyzes and imports an audio file into the Airtime library.
|
""" Analyzes and imports an audio file into the Airtime library.
|
||||||
|
|
||||||
This currently performs metadata extraction (eg. gets the ID3 tags from an MP3),
|
This currently performs metadata extraction (eg. gets the ID3 tags from an MP3),
|
||||||
then moves the file to the Airtime music library (stor/imported), and returns
|
then moves the file to the Airtime music library (stor/imported), and returns
|
||||||
the results back to the parent process. This class is used in an isolated process
|
the results back to the parent process. This class is used in an isolated process
|
||||||
|
@ -26,35 +26,35 @@ class AnalyzerPipeline:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def run_analysis(queue, audio_file_path, import_directory, original_filename, storage_backend, file_prefix):
|
def run_analysis(queue, audio_file_path, import_directory, original_filename, storage_backend, file_prefix):
|
||||||
"""Analyze and import an audio file, and put all extracted metadata into queue.
|
"""Analyze and import an audio file, and put all extracted metadata into queue.
|
||||||
|
|
||||||
Keyword arguments:
|
Keyword arguments:
|
||||||
queue: A multiprocessing.queues.Queue which will be used to pass the
|
queue: A multiprocessing.queues.Queue which will be used to pass the
|
||||||
extracted metadata back to the parent process.
|
extracted metadata back to the parent process.
|
||||||
audio_file_path: Path on disk to the audio file to analyze.
|
audio_file_path: Path on disk to the audio file to analyze.
|
||||||
import_directory: Path to the final Airtime "import" directory where
|
import_directory: Path to the final Airtime "import" directory where
|
||||||
we will move the file.
|
we will move the file.
|
||||||
original_filename: The original filename of the file, which we'll try to
|
original_filename: The original filename of the file, which we'll try to
|
||||||
preserve. The file at audio_file_path typically has a
|
preserve. The file at audio_file_path typically has a
|
||||||
temporary randomly generated name, which is why we want
|
temporary randomly generated name, which is why we want
|
||||||
to know what the original name was.
|
to know what the original name was.
|
||||||
storage_backend: String indicating the storage backend (amazon_s3 or file)
|
storage_backend: String indicating the storage backend (amazon_s3 or file)
|
||||||
file_prefix:
|
file_prefix:
|
||||||
"""
|
"""
|
||||||
# It is super critical to initialize a separate log file here so that we
|
# It is super critical to initialize a separate log file here so that we
|
||||||
# don't inherit logging/locks from the parent process. Supposedly
|
# don't inherit logging/locks from the parent process. Supposedly
|
||||||
# this can lead to Bad Things (deadlocks): http://bugs.python.org/issue6721
|
# this can lead to Bad Things (deadlocks): http://bugs.python.org/issue6721
|
||||||
AnalyzerPipeline.python_logger_deadlock_workaround()
|
AnalyzerPipeline.python_logger_deadlock_workaround()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if not isinstance(queue, Queue.Queue):
|
if not isinstance(queue, Queue):
|
||||||
raise TypeError("queue must be a Queue.Queue()")
|
raise TypeError("queue must be a Queue.Queue()")
|
||||||
if not isinstance(audio_file_path, unicode):
|
if not isinstance(audio_file_path, str):
|
||||||
raise TypeError("audio_file_path must be unicode. Was of type " + type(audio_file_path).__name__ + " instead.")
|
raise TypeError("audio_file_path must be unicode. Was of type " + type(audio_file_path).__name__ + " instead.")
|
||||||
if not isinstance(import_directory, unicode):
|
if not isinstance(import_directory, str):
|
||||||
raise TypeError("import_directory must be unicode. Was of type " + type(import_directory).__name__ + " instead.")
|
raise TypeError("import_directory must be unicode. Was of type " + type(import_directory).__name__ + " instead.")
|
||||||
if not isinstance(original_filename, unicode):
|
if not isinstance(original_filename, str):
|
||||||
raise TypeError("original_filename must be unicode. Was of type " + type(original_filename).__name__ + " instead.")
|
raise TypeError("original_filename must be unicode. Was of type " + type(original_filename).__name__ + " instead.")
|
||||||
if not isinstance(file_prefix, unicode):
|
if not isinstance(file_prefix, str):
|
||||||
raise TypeError("file_prefix must be unicode. Was of type " + type(file_prefix).__name__ + " instead.")
|
raise TypeError("file_prefix must be unicode. Was of type " + type(file_prefix).__name__ + " instead.")
|
||||||
|
|
||||||
|
|
||||||
|
@ -72,7 +72,7 @@ class AnalyzerPipeline:
|
||||||
|
|
||||||
metadata["import_status"] = 0 # Successfully imported
|
metadata["import_status"] = 0 # Successfully imported
|
||||||
|
|
||||||
# Note that the queue we're putting the results into is our interprocess communication
|
# Note that the queue we're putting the results into is our interprocess communication
|
||||||
# back to the main process.
|
# back to the main process.
|
||||||
|
|
||||||
# Pass all the file metadata back to the main analyzer process, which then passes
|
# Pass all the file metadata back to the main analyzer process, which then passes
|
||||||
|
@ -91,7 +91,7 @@ class AnalyzerPipeline:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def python_logger_deadlock_workaround():
|
def python_logger_deadlock_workaround():
|
||||||
# Workaround for: http://bugs.python.org/issue6721#msg140215
|
# Workaround for: http://bugs.python.org/issue6721#msg140215
|
||||||
logger_names = logging.Logger.manager.loggerDict.keys()
|
logger_names = list(logging.Logger.manager.loggerDict.keys())
|
||||||
logger_names.append(None) # Root logger
|
logger_names.append(None) # Root logger
|
||||||
for name in logger_names:
|
for name in logger_names:
|
||||||
for handler in logging.getLogger(name).handlers:
|
for handler in logging.getLogger(name).handlers:
|
||||||
|
|
|
@ -1,15 +1,16 @@
|
||||||
import ConfigParser
|
|
||||||
|
import configparser
|
||||||
|
|
||||||
def read_config_file(config_path):
|
def read_config_file(config_path):
|
||||||
"""Parse the application's config file located at config_path."""
|
"""Parse the application's config file located at config_path."""
|
||||||
config = ConfigParser.SafeConfigParser()
|
config = configparser.SafeConfigParser()
|
||||||
try:
|
try:
|
||||||
config.readfp(open(config_path))
|
config.readfp(open(config_path))
|
||||||
except IOError as e:
|
except IOError as e:
|
||||||
print "Failed to open config file at " + config_path + ": " + e.strerror
|
print("Failed to open config file at {}: {}".format(config_path, e.strerror))
|
||||||
exit(-1)
|
exit(-1)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print e.strerror
|
print(e.strerror)
|
||||||
exit(-1)
|
exit(-1)
|
||||||
|
|
||||||
return config
|
return config
|
||||||
|
|
|
@ -3,7 +3,7 @@ import logging
|
||||||
import traceback
|
import traceback
|
||||||
import json
|
import json
|
||||||
import datetime
|
import datetime
|
||||||
from analyzer import Analyzer
|
from .analyzer import Analyzer
|
||||||
|
|
||||||
|
|
||||||
class CuePointAnalyzer(Analyzer):
|
class CuePointAnalyzer(Analyzer):
|
||||||
|
@ -27,6 +27,10 @@ class CuePointAnalyzer(Analyzer):
|
||||||
command = [CuePointAnalyzer.SILAN_EXECUTABLE, '-b', '-F', '0.99', '-f', 'JSON', '-t', '1.0', filename]
|
command = [CuePointAnalyzer.SILAN_EXECUTABLE, '-b', '-F', '0.99', '-f', 'JSON', '-t', '1.0', filename]
|
||||||
try:
|
try:
|
||||||
results_json = subprocess.check_output(command, stderr=subprocess.STDOUT, close_fds=True)
|
results_json = subprocess.check_output(command, stderr=subprocess.STDOUT, close_fds=True)
|
||||||
|
try:
|
||||||
|
results_json = results_json.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
silan_results = json.loads(results_json)
|
silan_results = json.loads(results_json)
|
||||||
|
|
||||||
# Defensive coding against Silan wildly miscalculating the cue in and out times:
|
# Defensive coding against Silan wildly miscalculating the cue in and out times:
|
||||||
|
@ -64,7 +68,7 @@ class CuePointAnalyzer(Analyzer):
|
||||||
except OSError as e: # silan was not found
|
except OSError as e: # silan was not found
|
||||||
logging.warn("Failed to run: %s - %s. %s" % (command[0], e.strerror, "Do you have silan installed?"))
|
logging.warn("Failed to run: %s - %s. %s" % (command[0], e.strerror, "Do you have silan installed?"))
|
||||||
except subprocess.CalledProcessError as e: # silan returned an error code
|
except subprocess.CalledProcessError as e: # silan returned an error code
|
||||||
logging.warn("%s %s %s", e.cmd, e.message, e.returncode)
|
logging.warn("%s %s %s", e.cmd, e.output, e.returncode)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.warn(e)
|
logging.warn(e)
|
||||||
|
|
||||||
|
|
|
@ -4,24 +4,24 @@ import time
|
||||||
import shutil
|
import shutil
|
||||||
import os, errno
|
import os, errno
|
||||||
import time
|
import time
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
from analyzer import Analyzer
|
from .analyzer import Analyzer
|
||||||
|
|
||||||
class FileMoverAnalyzer(Analyzer):
|
class FileMoverAnalyzer(Analyzer):
|
||||||
"""This analyzer copies a file over from a temporary directory (stor/organize)
|
"""This analyzer copies a file over from a temporary directory (stor/organize)
|
||||||
into the Airtime library (stor/imported).
|
into the Airtime library (stor/imported).
|
||||||
"""
|
"""
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def analyze(audio_file_path, metadata):
|
def analyze(audio_file_path, metadata):
|
||||||
"""Dummy method because we need more info than analyze gets passed to it"""
|
"""Dummy method because we need more info than analyze gets passed to it"""
|
||||||
raise Exception("Use FileMoverAnalyzer.move() instead.")
|
raise Exception("Use FileMoverAnalyzer.move() instead.")
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def move(audio_file_path, import_directory, original_filename, metadata):
|
def move(audio_file_path, import_directory, original_filename, metadata):
|
||||||
"""Move the file at audio_file_path over into the import_directory/import,
|
"""Move the file at audio_file_path over into the import_directory/import,
|
||||||
renaming it to original_filename.
|
renaming it to original_filename.
|
||||||
|
|
||||||
Keyword arguments:
|
Keyword arguments:
|
||||||
audio_file_path: Path to the file to be imported.
|
audio_file_path: Path to the file to be imported.
|
||||||
import_directory: Path to the "import" directory inside the Airtime stor directory.
|
import_directory: Path to the "import" directory inside the Airtime stor directory.
|
||||||
|
@ -29,26 +29,28 @@ class FileMoverAnalyzer(Analyzer):
|
||||||
original_filename: The filename of the file when it was uploaded to Airtime.
|
original_filename: The filename of the file when it was uploaded to Airtime.
|
||||||
metadata: A dictionary where the "full_path" of where the file is moved to will be added.
|
metadata: A dictionary where the "full_path" of where the file is moved to will be added.
|
||||||
"""
|
"""
|
||||||
if not isinstance(audio_file_path, unicode):
|
if not isinstance(audio_file_path, str):
|
||||||
raise TypeError("audio_file_path must be unicode. Was of type " + type(audio_file_path).__name__)
|
raise TypeError("audio_file_path must be string. Was of type " + type(audio_file_path).__name__)
|
||||||
if not isinstance(import_directory, unicode):
|
if not isinstance(import_directory, str):
|
||||||
raise TypeError("import_directory must be unicode. Was of type " + type(import_directory).__name__)
|
raise TypeError("import_directory must be string. Was of type " + type(import_directory).__name__)
|
||||||
if not isinstance(original_filename, unicode):
|
if not isinstance(original_filename, str):
|
||||||
raise TypeError("original_filename must be unicode. Was of type " + type(original_filename).__name__)
|
raise TypeError("original_filename must be string. Was of type " + type(original_filename).__name__)
|
||||||
if not isinstance(metadata, dict):
|
if not isinstance(metadata, dict):
|
||||||
raise TypeError("metadata must be a dict. Was of type " + type(metadata).__name__)
|
raise TypeError("metadata must be a dict. Was of type " + type(metadata).__name__)
|
||||||
|
if not os.path.exists(audio_file_path):
|
||||||
|
raise FileNotFoundError("audio file not found: {}".format(audio_file_path))
|
||||||
|
|
||||||
#Import the file over to it's final location.
|
#Import the file over to it's final location.
|
||||||
# TODO: Also, handle the case where the move fails and write some code
|
# TODO: Also, handle the case where the move fails and write some code
|
||||||
# to possibly move the file to problem_files.
|
# to possibly move the file to problem_files.
|
||||||
|
|
||||||
max_dir_len = 48
|
max_dir_len = 48
|
||||||
max_file_len = 48
|
max_file_len = 48
|
||||||
final_file_path = import_directory
|
final_file_path = import_directory
|
||||||
orig_file_basename, orig_file_extension = os.path.splitext(original_filename)
|
orig_file_basename, orig_file_extension = os.path.splitext(original_filename)
|
||||||
if metadata.has_key("artist_name"):
|
if "artist_name" in metadata:
|
||||||
final_file_path += "/" + metadata["artist_name"][0:max_dir_len] # truncating with array slicing
|
final_file_path += "/" + metadata["artist_name"][0:max_dir_len] # truncating with array slicing
|
||||||
if metadata.has_key("album_title"):
|
if "album_title" in metadata:
|
||||||
final_file_path += "/" + metadata["album_title"][0:max_dir_len]
|
final_file_path += "/" + metadata["album_title"][0:max_dir_len]
|
||||||
# Note that orig_file_extension includes the "." already
|
# Note that orig_file_extension includes the "." already
|
||||||
final_file_path += "/" + orig_file_basename[0:max_file_len] + orig_file_extension
|
final_file_path += "/" + orig_file_basename[0:max_file_len] + orig_file_extension
|
||||||
|
@ -58,11 +60,11 @@ class FileMoverAnalyzer(Analyzer):
|
||||||
|
|
||||||
#If a file with the same name already exists in the "import" directory, then
|
#If a file with the same name already exists in the "import" directory, then
|
||||||
#we add a unique string to the end of this one. We never overwrite a file on import
|
#we add a unique string to the end of this one. We never overwrite a file on import
|
||||||
#because if we did that, it would mean Airtime's database would have
|
#because if we did that, it would mean Airtime's database would have
|
||||||
#the wrong information for the file we just overwrote (eg. the song length would be wrong!)
|
#the wrong information for the file we just overwrote (eg. the song length would be wrong!)
|
||||||
#If the final file path is the same as the file we've been told to import (which
|
#If the final file path is the same as the file we've been told to import (which
|
||||||
#you often do when you're debugging), then don't move the file at all.
|
#you often do when you're debugging), then don't move the file at all.
|
||||||
|
|
||||||
if os.path.exists(final_file_path):
|
if os.path.exists(final_file_path):
|
||||||
if os.path.samefile(audio_file_path, final_file_path):
|
if os.path.samefile(audio_file_path, final_file_path):
|
||||||
metadata["full_path"] = final_file_path
|
metadata["full_path"] = final_file_path
|
||||||
|
@ -77,14 +79,14 @@ class FileMoverAnalyzer(Analyzer):
|
||||||
|
|
||||||
#Ensure the full path to the file exists
|
#Ensure the full path to the file exists
|
||||||
mkdir_p(os.path.dirname(final_file_path))
|
mkdir_p(os.path.dirname(final_file_path))
|
||||||
|
|
||||||
#Move the file into its final destination directory
|
#Move the file into its final destination directory
|
||||||
logging.debug("Moving %s to %s" % (audio_file_path, final_file_path))
|
logging.debug("Moving %s to %s" % (audio_file_path, final_file_path))
|
||||||
shutil.move(audio_file_path, final_file_path)
|
shutil.move(audio_file_path, final_file_path)
|
||||||
|
|
||||||
metadata["full_path"] = final_file_path
|
metadata["full_path"] = final_file_path
|
||||||
return metadata
|
return metadata
|
||||||
|
|
||||||
def mkdir_p(path):
|
def mkdir_p(path):
|
||||||
""" Make all directories in a tree (like mkdir -p)"""
|
""" Make all directories in a tree (like mkdir -p)"""
|
||||||
if path == "":
|
if path == "":
|
||||||
|
|
|
@ -6,9 +6,9 @@ import select
|
||||||
import signal
|
import signal
|
||||||
import logging
|
import logging
|
||||||
import multiprocessing
|
import multiprocessing
|
||||||
import Queue
|
import queue
|
||||||
from analyzer_pipeline import AnalyzerPipeline
|
from .analyzer_pipeline import AnalyzerPipeline
|
||||||
from status_reporter import StatusReporter
|
from .status_reporter import StatusReporter
|
||||||
|
|
||||||
EXCHANGE = "airtime-uploads"
|
EXCHANGE = "airtime-uploads"
|
||||||
EXCHANGE_TYPE = "topic"
|
EXCHANGE_TYPE = "topic"
|
||||||
|
@ -112,8 +112,7 @@ class MessageListener:
|
||||||
self._channel.queue_bind(exchange=EXCHANGE, queue=QUEUE, routing_key=ROUTING_KEY)
|
self._channel.queue_bind(exchange=EXCHANGE, queue=QUEUE, routing_key=ROUTING_KEY)
|
||||||
|
|
||||||
logging.info(" Listening for messages...")
|
logging.info(" Listening for messages...")
|
||||||
self._channel.basic_consume(self.msg_received_callback,
|
self._channel.basic_consume(QUEUE, self.msg_received_callback, auto_ack=False)
|
||||||
queue=QUEUE, no_ack=False)
|
|
||||||
|
|
||||||
def wait_for_messages(self):
|
def wait_for_messages(self):
|
||||||
'''Wait until we've received a RabbitMQ message.'''
|
'''Wait until we've received a RabbitMQ message.'''
|
||||||
|
@ -158,6 +157,10 @@ class MessageListener:
|
||||||
We avoid cascading failure this way.
|
We avoid cascading failure this way.
|
||||||
'''
|
'''
|
||||||
try:
|
try:
|
||||||
|
try:
|
||||||
|
body = body.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
msg_dict = json.loads(body)
|
msg_dict = json.loads(body)
|
||||||
api_key = msg_dict["api_key"]
|
api_key = msg_dict["api_key"]
|
||||||
callback_url = msg_dict["callback_url"]
|
callback_url = msg_dict["callback_url"]
|
||||||
|
@ -198,7 +201,7 @@ class MessageListener:
|
||||||
if callback_url: # If we got an invalid message, there might be no callback_url in the JSON
|
if callback_url: # If we got an invalid message, there might be no callback_url in the JSON
|
||||||
# Report this as a failed upload to the File Upload REST API.
|
# Report this as a failed upload to the File Upload REST API.
|
||||||
StatusReporter.report_failure_to_callback_url(callback_url, api_key, import_status=2,
|
StatusReporter.report_failure_to_callback_url(callback_url, api_key, import_status=2,
|
||||||
reason=u'An error occurred while importing this file')
|
reason='An error occurred while importing this file')
|
||||||
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
|
@ -224,7 +227,7 @@ class MessageListener:
|
||||||
'''
|
'''
|
||||||
metadata = {}
|
metadata = {}
|
||||||
|
|
||||||
q = Queue.Queue()
|
q = queue.Queue()
|
||||||
try:
|
try:
|
||||||
AnalyzerPipeline.run_analysis(q, audio_file_path, import_directory, original_filename, storage_backend, file_prefix)
|
AnalyzerPipeline.run_analysis(q, audio_file_path, import_directory, original_filename, storage_backend, file_prefix)
|
||||||
metadata = q.get()
|
metadata = q.get()
|
||||||
|
|
|
@ -6,22 +6,24 @@ import wave
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import hashlib
|
import hashlib
|
||||||
from analyzer import Analyzer
|
from .analyzer import Analyzer
|
||||||
|
|
||||||
class MetadataAnalyzer(Analyzer):
|
class MetadataAnalyzer(Analyzer):
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def analyze(filename, metadata):
|
def analyze(filename, metadata):
|
||||||
''' Extract audio metadata from tags embedded in the file (eg. ID3 tags)
|
''' Extract audio metadata from tags embedded in the file (eg. ID3 tags)
|
||||||
|
|
||||||
Keyword arguments:
|
Keyword arguments:
|
||||||
filename: The path to the audio file to extract metadata from.
|
filename: The path to the audio file to extract metadata from.
|
||||||
metadata: A dictionary that the extracted metadata will be added to.
|
metadata: A dictionary that the extracted metadata will be added to.
|
||||||
'''
|
'''
|
||||||
if not isinstance(filename, unicode):
|
if not isinstance(filename, str):
|
||||||
raise TypeError("filename must be unicode. Was of type " + type(filename).__name__)
|
raise TypeError("filename must be string. Was of type " + type(filename).__name__)
|
||||||
if not isinstance(metadata, dict):
|
if not isinstance(metadata, dict):
|
||||||
raise TypeError("metadata must be a dict. Was of type " + type(metadata).__name__)
|
raise TypeError("metadata must be a dict. Was of type " + type(metadata).__name__)
|
||||||
|
if not os.path.exists(filename):
|
||||||
|
raise FileNotFoundError("audio file not found: {}".format(filename))
|
||||||
|
|
||||||
#Airtime <= 2.5.x nonsense:
|
#Airtime <= 2.5.x nonsense:
|
||||||
metadata["ftype"] = "audioclip"
|
metadata["ftype"] = "audioclip"
|
||||||
|
@ -40,7 +42,7 @@ class MetadataAnalyzer(Analyzer):
|
||||||
m.update(data)
|
m.update(data)
|
||||||
metadata["md5"] = m.hexdigest()
|
metadata["md5"] = m.hexdigest()
|
||||||
|
|
||||||
# Mutagen doesn't handle WAVE files so we use a different package
|
# Mutagen doesn't handle WAVE files so we use a different package
|
||||||
ms = magic.open(magic.MIME_TYPE)
|
ms = magic.open(magic.MIME_TYPE)
|
||||||
ms.load()
|
ms.load()
|
||||||
with open(filename, 'rb') as fh:
|
with open(filename, 'rb') as fh:
|
||||||
|
@ -57,15 +59,15 @@ class MetadataAnalyzer(Analyzer):
|
||||||
if audio_file == None: # Don't use "if not" here. It is wrong due to mutagen's design.
|
if audio_file == None: # Don't use "if not" here. It is wrong due to mutagen's design.
|
||||||
return metadata
|
return metadata
|
||||||
# Note that audio_file can equal {} if the file is valid but there's no metadata tags.
|
# Note that audio_file can equal {} if the file is valid but there's no metadata tags.
|
||||||
# We can still try to grab the info variables below.
|
# We can still try to grab the info variables below.
|
||||||
|
|
||||||
#Grab other file information that isn't encoded in a tag, but instead usually
|
#Grab other file information that isn't encoded in a tag, but instead usually
|
||||||
#in the file header. Mutagen breaks that out into a separate "info" object:
|
#in the file header. Mutagen breaks that out into a separate "info" object:
|
||||||
info = audio_file.info
|
info = audio_file.info
|
||||||
if hasattr(info, "sample_rate"): # Mutagen is annoying and inconsistent
|
if hasattr(info, "sample_rate"): # Mutagen is annoying and inconsistent
|
||||||
metadata["sample_rate"] = info.sample_rate
|
metadata["sample_rate"] = info.sample_rate
|
||||||
if hasattr(info, "length"):
|
if hasattr(info, "length"):
|
||||||
metadata["length_seconds"] = info.length
|
metadata["length_seconds"] = info.length
|
||||||
#Converting the length in seconds (float) to a formatted time string
|
#Converting the length in seconds (float) to a formatted time string
|
||||||
track_length = datetime.timedelta(seconds=info.length)
|
track_length = datetime.timedelta(seconds=info.length)
|
||||||
metadata["length"] = str(track_length) #time.strftime("%H:%M:%S.%f", track_length)
|
metadata["length"] = str(track_length) #time.strftime("%H:%M:%S.%f", track_length)
|
||||||
|
@ -77,12 +79,12 @@ class MetadataAnalyzer(Analyzer):
|
||||||
|
|
||||||
if hasattr(info, "bitrate"):
|
if hasattr(info, "bitrate"):
|
||||||
metadata["bit_rate"] = info.bitrate
|
metadata["bit_rate"] = info.bitrate
|
||||||
|
|
||||||
# Use the mutagen to get the MIME type, if it has one. This is more reliable and
|
# Use the mutagen to get the MIME type, if it has one. This is more reliable and
|
||||||
# consistent for certain types of MP3s or MPEG files than the MIMEs returned by magic.
|
# consistent for certain types of MP3s or MPEG files than the MIMEs returned by magic.
|
||||||
if audio_file.mime:
|
if audio_file.mime:
|
||||||
metadata["mime"] = audio_file.mime[0]
|
metadata["mime"] = audio_file.mime[0]
|
||||||
|
|
||||||
#Try to get the number of channels if mutagen can...
|
#Try to get the number of channels if mutagen can...
|
||||||
try:
|
try:
|
||||||
#Special handling for getting the # of channels from MP3s. It's in the "mode" field
|
#Special handling for getting the # of channels from MP3s. It's in the "mode" field
|
||||||
|
@ -97,18 +99,18 @@ class MetadataAnalyzer(Analyzer):
|
||||||
except (AttributeError, KeyError):
|
except (AttributeError, KeyError):
|
||||||
#If mutagen can't figure out the number of channels, we'll just leave it out...
|
#If mutagen can't figure out the number of channels, we'll just leave it out...
|
||||||
pass
|
pass
|
||||||
|
|
||||||
#Try to extract the number of tracks on the album if we can (the "track total")
|
#Try to extract the number of tracks on the album if we can (the "track total")
|
||||||
try:
|
try:
|
||||||
track_number = audio_file["tracknumber"]
|
track_number = audio_file["tracknumber"]
|
||||||
if isinstance(track_number, list): # Sometimes tracknumber is a list, ugh
|
if isinstance(track_number, list): # Sometimes tracknumber is a list, ugh
|
||||||
track_number = track_number[0]
|
track_number = track_number[0]
|
||||||
track_number_tokens = track_number
|
track_number_tokens = track_number
|
||||||
if u'/' in track_number:
|
if '/' in track_number:
|
||||||
track_number_tokens = track_number.split(u'/')
|
track_number_tokens = track_number.split('/')
|
||||||
track_number = track_number_tokens[0]
|
track_number = track_number_tokens[0]
|
||||||
elif u'-' in track_number:
|
elif '-' in track_number:
|
||||||
track_number_tokens = track_number.split(u'-')
|
track_number_tokens = track_number.split('-')
|
||||||
track_number = track_number_tokens[0]
|
track_number = track_number_tokens[0]
|
||||||
metadata["track_number"] = track_number
|
metadata["track_number"] = track_number
|
||||||
track_total = track_number_tokens[1]
|
track_total = track_number_tokens[1]
|
||||||
|
@ -118,7 +120,7 @@ class MetadataAnalyzer(Analyzer):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
#We normalize the mutagen tags slightly here, so in case mutagen changes,
|
#We normalize the mutagen tags slightly here, so in case mutagen changes,
|
||||||
#we find the
|
#we find the
|
||||||
mutagen_to_airtime_mapping = {
|
mutagen_to_airtime_mapping = {
|
||||||
'title': 'track_title',
|
'title': 'track_title',
|
||||||
'artist': 'artist_name',
|
'artist': 'artist_name',
|
||||||
|
@ -146,20 +148,20 @@ class MetadataAnalyzer(Analyzer):
|
||||||
#'mime_type': 'mime',
|
#'mime_type': 'mime',
|
||||||
}
|
}
|
||||||
|
|
||||||
for mutagen_tag, airtime_tag in mutagen_to_airtime_mapping.iteritems():
|
for mutagen_tag, airtime_tag in mutagen_to_airtime_mapping.items():
|
||||||
try:
|
try:
|
||||||
metadata[airtime_tag] = audio_file[mutagen_tag]
|
metadata[airtime_tag] = audio_file[mutagen_tag]
|
||||||
|
|
||||||
# Some tags are returned as lists because there could be multiple values.
|
# Some tags are returned as lists because there could be multiple values.
|
||||||
# This is unusual so we're going to always just take the first item in the list.
|
# This is unusual so we're going to always just take the first item in the list.
|
||||||
if isinstance(metadata[airtime_tag], list):
|
if isinstance(metadata[airtime_tag], list):
|
||||||
if metadata[airtime_tag]:
|
if metadata[airtime_tag]:
|
||||||
metadata[airtime_tag] = metadata[airtime_tag][0]
|
metadata[airtime_tag] = metadata[airtime_tag][0]
|
||||||
else: # Handle empty lists
|
else: # Handle empty lists
|
||||||
metadata[airtime_tag] = ""
|
metadata[airtime_tag] = ""
|
||||||
|
|
||||||
except KeyError:
|
except KeyError:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
return metadata
|
return metadata
|
||||||
|
|
||||||
|
@ -174,7 +176,7 @@ class MetadataAnalyzer(Analyzer):
|
||||||
track_length = datetime.timedelta(seconds=length_seconds)
|
track_length = datetime.timedelta(seconds=length_seconds)
|
||||||
metadata["length"] = str(track_length) #time.strftime("%H:%M:%S.%f", track_length)
|
metadata["length"] = str(track_length) #time.strftime("%H:%M:%S.%f", track_length)
|
||||||
metadata["length_seconds"] = length_seconds
|
metadata["length_seconds"] = length_seconds
|
||||||
metadata["cueout"] = metadata["length"]
|
metadata["cueout"] = metadata["length"]
|
||||||
except wave.Error as ex:
|
except wave.Error as ex:
|
||||||
logging.error("Invalid WAVE file: {}".format(str(ex)))
|
logging.error("Invalid WAVE file: {}".format(str(ex)))
|
||||||
raise
|
raise
|
||||||
|
|
|
@ -2,7 +2,7 @@ __author__ = 'asantoni'
|
||||||
|
|
||||||
import subprocess
|
import subprocess
|
||||||
import logging
|
import logging
|
||||||
from analyzer import Analyzer
|
from .analyzer import Analyzer
|
||||||
|
|
||||||
class UnplayableFileError(Exception):
|
class UnplayableFileError(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
|
@ -1,12 +1,13 @@
|
||||||
import subprocess
|
import subprocess
|
||||||
import logging
|
import logging
|
||||||
from analyzer import Analyzer
|
from .analyzer import Analyzer
|
||||||
|
import re
|
||||||
|
|
||||||
|
|
||||||
class ReplayGainAnalyzer(Analyzer):
|
class ReplayGainAnalyzer(Analyzer):
|
||||||
''' This class extracts the ReplayGain using a tool from the python-rgain package. '''
|
''' This class extracts the ReplayGain using a tool from the python-rgain package. '''
|
||||||
|
|
||||||
REPLAYGAIN_EXECUTABLE = 'replaygain' # From the python-rgain package
|
REPLAYGAIN_EXECUTABLE = 'replaygain' # From the rgain3 python package
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def analyze(filename, metadata):
|
def analyze(filename, metadata):
|
||||||
|
@ -19,17 +20,16 @@ class ReplayGainAnalyzer(Analyzer):
|
||||||
'''
|
'''
|
||||||
command = [ReplayGainAnalyzer.REPLAYGAIN_EXECUTABLE, '-d', filename]
|
command = [ReplayGainAnalyzer.REPLAYGAIN_EXECUTABLE, '-d', filename]
|
||||||
try:
|
try:
|
||||||
results = subprocess.check_output(command, stderr=subprocess.STDOUT, close_fds=True)
|
results = subprocess.check_output(command, stderr=subprocess.STDOUT,
|
||||||
filename_token = "%s: " % filename
|
close_fds=True, text=True)
|
||||||
rg_pos = results.find(filename_token, results.find("Calculating Replay Gain information")) + len(filename_token)
|
gain_match = r'Calculating Replay Gain information \.\.\.(?:\n|.)*?:([\d.-]*) dB'
|
||||||
db_pos = results.find(" dB", rg_pos)
|
replaygain = re.search(gain_match, results).group(1)
|
||||||
replaygain = results[rg_pos:db_pos]
|
|
||||||
metadata['replay_gain'] = float(replaygain)
|
metadata['replay_gain'] = float(replaygain)
|
||||||
|
|
||||||
except OSError as e: # replaygain was not found
|
except OSError as e: # replaygain was not found
|
||||||
logging.warn("Failed to run: %s - %s. %s" % (command[0], e.strerror, "Do you have python-rgain installed?"))
|
logging.warn("Failed to run: %s - %s. %s" % (command[0], e.strerror, "Do you have python-rgain installed?"))
|
||||||
except subprocess.CalledProcessError as e: # replaygain returned an error code
|
except subprocess.CalledProcessError as e: # replaygain returned an error code
|
||||||
logging.warn("%s %s %s", e.cmd, e.message, e.returncode)
|
logging.warn("%s %s %s", e.cmd, e.output, e.returncode)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.warn(e)
|
logging.warn(e)
|
||||||
|
|
||||||
|
|
|
@ -2,12 +2,12 @@ import requests
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
import collections
|
import collections
|
||||||
import Queue
|
import queue
|
||||||
import time
|
import time
|
||||||
import traceback
|
import traceback
|
||||||
import pickle
|
import pickle
|
||||||
import threading
|
import threading
|
||||||
from urlparse import urlparse
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
# Disable urllib3 warnings because these can cause a rare deadlock due to Python 2's crappy internal non-reentrant locking
|
# Disable urllib3 warnings because these can cause a rare deadlock due to Python 2's crappy internal non-reentrant locking
|
||||||
# around POSIX stuff. See SAAS-714. The hasattr() is for compatibility with older versions of requests.
|
# around POSIX stuff. See SAAS-714. The hasattr() is for compatibility with older versions of requests.
|
||||||
|
@ -68,7 +68,7 @@ def process_http_requests(ipc_queue, http_retry_queue_path):
|
||||||
break
|
break
|
||||||
if not isinstance(request, PicklableHttpRequest):
|
if not isinstance(request, PicklableHttpRequest):
|
||||||
raise TypeError("request must be a PicklableHttpRequest. Was of type " + type(request).__name__)
|
raise TypeError("request must be a PicklableHttpRequest. Was of type " + type(request).__name__)
|
||||||
except Queue.Empty:
|
except queue.Empty:
|
||||||
request = None
|
request = None
|
||||||
|
|
||||||
# If there's no new HTTP request we need to execute, let's check our "retry
|
# If there's no new HTTP request we need to execute, let's check our "retry
|
||||||
|
@ -159,7 +159,7 @@ class StatusReporter():
|
||||||
''' We use multiprocessing.Process again here because we need a thread for this stuff
|
''' We use multiprocessing.Process again here because we need a thread for this stuff
|
||||||
anyways, and Python gives us process isolation for free (crash safety).
|
anyways, and Python gives us process isolation for free (crash safety).
|
||||||
'''
|
'''
|
||||||
_ipc_queue = Queue.Queue()
|
_ipc_queue = queue.Queue()
|
||||||
#_http_thread = multiprocessing.Process(target=process_http_requests,
|
#_http_thread = multiprocessing.Process(target=process_http_requests,
|
||||||
# args=(_ipc_queue,))
|
# args=(_ipc_queue,))
|
||||||
_http_thread = None
|
_http_thread = None
|
||||||
|
@ -222,7 +222,7 @@ class StatusReporter():
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def report_failure_to_callback_url(self, callback_url, api_key, import_status, reason):
|
def report_failure_to_callback_url(self, callback_url, api_key, import_status, reason):
|
||||||
if not isinstance(import_status, (int, long) ):
|
if not isinstance(import_status, int ):
|
||||||
raise TypeError("import_status must be an integer. Was of type " + type(import_status).__name__)
|
raise TypeError("import_status must be an integer. Was of type " + type(import_status).__name__)
|
||||||
|
|
||||||
logging.debug("Reporting import failure to Airtime REST API...")
|
logging.debug("Reporting import failure to Airtime REST API...")
|
||||||
|
|
|
@ -2,6 +2,7 @@
|
||||||
"""Runs the airtime_analyzer application.
|
"""Runs the airtime_analyzer application.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
import daemon
|
import daemon
|
||||||
import argparse
|
import argparse
|
||||||
import os
|
import os
|
||||||
|
@ -14,7 +15,7 @@ DEFAULT_HTTP_RETRY_PATH = '/tmp/airtime_analyzer_http_retries'
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
'''Entry-point for this application'''
|
'''Entry-point for this application'''
|
||||||
print "Airtime Analyzer " + VERSION
|
print("Airtime Analyzer {}".format(VERSION))
|
||||||
parser = argparse.ArgumentParser()
|
parser = argparse.ArgumentParser()
|
||||||
parser.add_argument("-d", "--daemon", help="run as a daemon", action="store_true")
|
parser.add_argument("-d", "--daemon", help="run as a daemon", action="store_true")
|
||||||
parser.add_argument("--debug", help="log full debugging output", action="store_true")
|
parser.add_argument("--debug", help="log full debugging output", action="store_true")
|
||||||
|
@ -22,8 +23,6 @@ def run():
|
||||||
parser.add_argument("--http-retry-queue-file", help="specify where incompleted HTTP requests will be serialized (default is %s)" % DEFAULT_HTTP_RETRY_PATH)
|
parser.add_argument("--http-retry-queue-file", help="specify where incompleted HTTP requests will be serialized (default is %s)" % DEFAULT_HTTP_RETRY_PATH)
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
check_if_media_monitor_is_running()
|
|
||||||
|
|
||||||
#Default config file path
|
#Default config file path
|
||||||
rmq_config_path = DEFAULT_RMQ_CONFIG_PATH
|
rmq_config_path = DEFAULT_RMQ_CONFIG_PATH
|
||||||
http_retry_queue_path = DEFAULT_HTTP_RETRY_PATH
|
http_retry_queue_path = DEFAULT_HTTP_RETRY_PATH
|
||||||
|
@ -35,32 +34,12 @@ def run():
|
||||||
if args.daemon:
|
if args.daemon:
|
||||||
with daemon.DaemonContext():
|
with daemon.DaemonContext():
|
||||||
aa.AirtimeAnalyzerServer(rmq_config_path=rmq_config_path,
|
aa.AirtimeAnalyzerServer(rmq_config_path=rmq_config_path,
|
||||||
http_retry_queue_path=http_retry_queue_path,
|
http_retry_queue_path=http_retry_queue_path,
|
||||||
debug=args.debug)
|
debug=args.debug)
|
||||||
else:
|
else:
|
||||||
# Run without daemonizing
|
# Run without daemonizing
|
||||||
aa.AirtimeAnalyzerServer(rmq_config_path=rmq_config_path,
|
aa.AirtimeAnalyzerServer(rmq_config_path=rmq_config_path,
|
||||||
http_retry_queue_path=http_retry_queue_path,
|
http_retry_queue_path=http_retry_queue_path,
|
||||||
debug=args.debug)
|
debug=args.debug)
|
||||||
|
|
||||||
|
|
||||||
def check_if_media_monitor_is_running():
|
|
||||||
"""Ensure media_monitor isn't running before we start.
|
|
||||||
|
|
||||||
We do this because media_monitor will move newly uploaded
|
|
||||||
files into the library on us and screw up the operation of airtime_analyzer.
|
|
||||||
media_monitor is deprecated.
|
|
||||||
"""
|
|
||||||
pids = [pid for pid in os.listdir('/proc') if pid.isdigit()]
|
|
||||||
|
|
||||||
for pid in pids:
|
|
||||||
try:
|
|
||||||
process_name = open(os.path.join('/proc', pid, 'cmdline'), 'rb').read()
|
|
||||||
if 'media_monitor.py' in process_name:
|
|
||||||
print "Error: This process conflicts with media_monitor, and media_monitor is running."
|
|
||||||
print " Please terminate the running media_monitor.py process and try again."
|
|
||||||
exit(1)
|
|
||||||
except IOError: # proc has already terminated
|
|
||||||
continue
|
|
||||||
|
|
||||||
run()
|
run()
|
||||||
|
|
|
@ -1,3 +1,4 @@
|
||||||
|
from __future__ import print_function
|
||||||
from setuptools import setup
|
from setuptools import setup
|
||||||
from subprocess import call
|
from subprocess import call
|
||||||
import sys
|
import sys
|
||||||
|
@ -5,7 +6,7 @@ import os
|
||||||
|
|
||||||
# Change directory since setuptools uses relative paths
|
# Change directory since setuptools uses relative paths
|
||||||
script_path = os.path.dirname(os.path.realpath(__file__))
|
script_path = os.path.dirname(os.path.realpath(__file__))
|
||||||
print script_path
|
print(script_path)
|
||||||
os.chdir(script_path)
|
os.chdir(script_path)
|
||||||
|
|
||||||
# Allows us to avoid installing the upstart init script when deploying airtime_analyzer
|
# Allows us to avoid installing the upstart init script when deploying airtime_analyzer
|
||||||
|
@ -16,7 +17,7 @@ if '--no-init-script' in sys.argv:
|
||||||
else:
|
else:
|
||||||
data_files = [('/etc/init', ['install/upstart/airtime_analyzer.conf']),
|
data_files = [('/etc/init', ['install/upstart/airtime_analyzer.conf']),
|
||||||
('/etc/init.d', ['install/sysvinit/airtime_analyzer'])]
|
('/etc/init.d', ['install/sysvinit/airtime_analyzer'])]
|
||||||
print data_files
|
print(data_files)
|
||||||
|
|
||||||
setup(name='airtime_analyzer',
|
setup(name='airtime_analyzer',
|
||||||
version='0.1',
|
version='0.1',
|
||||||
|
@ -28,16 +29,15 @@ setup(name='airtime_analyzer',
|
||||||
packages=['airtime_analyzer'],
|
packages=['airtime_analyzer'],
|
||||||
scripts=['bin/airtime_analyzer'],
|
scripts=['bin/airtime_analyzer'],
|
||||||
install_requires=[
|
install_requires=[
|
||||||
'mutagen~=1.43.0', # got rid of specific version requirement
|
'mutagen~=1.43',
|
||||||
'pika',
|
'pika~=1.1.0',
|
||||||
'daemon',
|
|
||||||
'file-magic',
|
'file-magic',
|
||||||
'nose',
|
'nose',
|
||||||
'coverage',
|
'coverage',
|
||||||
'mock',
|
'mock',
|
||||||
'python-daemon==1.6',
|
'python-daemon',
|
||||||
'requests>=2.7.0',
|
'requests>=2.7.0',
|
||||||
'rgain',
|
'rgain3',
|
||||||
# These next 3 are required for requests to support SSL with SNI. Learned this the hard way...
|
# These next 3 are required for requests to support SSL with SNI. Learned this the hard way...
|
||||||
# What sucks is that GCC is required to pip install these.
|
# What sucks is that GCC is required to pip install these.
|
||||||
#'ndg-httpsclient',
|
#'ndg-httpsclient',
|
||||||
|
@ -49,8 +49,8 @@ setup(name='airtime_analyzer',
|
||||||
|
|
||||||
# Remind users to reload the initctl config so that "service start airtime_analyzer" works
|
# Remind users to reload the initctl config so that "service start airtime_analyzer" works
|
||||||
if data_files:
|
if data_files:
|
||||||
print "Remember to reload the initctl configuration"
|
print("Remember to reload the initctl configuration")
|
||||||
print "Run \"sudo initctl reload-configuration; sudo service airtime_analyzer restart\" now."
|
print("Run \"sudo initctl reload-configuration; sudo service airtime_analyzer restart\" now.")
|
||||||
print "Or on Ubuntu Xenial (16.04)"
|
print("Or on Ubuntu Xenial (16.04)")
|
||||||
print "Remember to reload the systemd configuration"
|
print("Remember to reload the systemd configuration")
|
||||||
print "Run \"sudo systemctl daemon-reload; sudo service airtime_analyzer restart\" now."
|
print("Run \"sudo systemctl daemon-reload; sudo service airtime_analyzer restart\" now.")
|
||||||
|
|
|
@ -1,9 +1,8 @@
|
||||||
from nose.tools import *
|
from nose.tools import *
|
||||||
from ConfigParser import SafeConfigParser
|
|
||||||
import os
|
import os
|
||||||
import shutil
|
import shutil
|
||||||
import multiprocessing
|
import multiprocessing
|
||||||
import Queue
|
from queue import Queue
|
||||||
import datetime
|
import datetime
|
||||||
from airtime_analyzer.analyzer_pipeline import AnalyzerPipeline
|
from airtime_analyzer.analyzer_pipeline import AnalyzerPipeline
|
||||||
from airtime_analyzer import config_file
|
from airtime_analyzer import config_file
|
||||||
|
@ -21,7 +20,7 @@ def teardown():
|
||||||
|
|
||||||
def test_basic():
|
def test_basic():
|
||||||
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
||||||
q = Queue.Queue()
|
q = Queue()
|
||||||
file_prefix = u''
|
file_prefix = u''
|
||||||
storage_backend = "file"
|
storage_backend = "file"
|
||||||
#This actually imports the file into the "./Test Artist" directory.
|
#This actually imports the file into the "./Test Artist" directory.
|
||||||
|
@ -39,17 +38,17 @@ def test_basic():
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_wrong_type_queue_param():
|
def test_wrong_type_queue_param():
|
||||||
AnalyzerPipeline.run_analysis(Queue.Queue(), u'', u'', u'')
|
AnalyzerPipeline.run_analysis(Queue(), u'', u'', u'')
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_wrong_type_string_param2():
|
def test_wrong_type_string_param2():
|
||||||
AnalyzerPipeline.run_analysis(Queue.Queue(), '', u'', u'')
|
AnalyzerPipeline.run_analysis(Queue(), '', u'', u'')
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_wrong_type_string_param3():
|
def test_wrong_type_string_param3():
|
||||||
AnalyzerPipeline.run_analysis(Queue.Queue(), u'', '', u'')
|
AnalyzerPipeline.run_analysis(Queue(), u'', '', u'')
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_wrong_type_string_param4():
|
def test_wrong_type_string_param4():
|
||||||
AnalyzerPipeline.run_analysis(Queue.Queue(), u'', u'', '')
|
AnalyzerPipeline.run_analysis(Queue(), u'', u'', '')
|
||||||
|
|
||||||
|
|
|
@ -2,7 +2,6 @@ from nose.tools import *
|
||||||
import os
|
import os
|
||||||
import shutil
|
import shutil
|
||||||
import multiprocessing
|
import multiprocessing
|
||||||
import Queue
|
|
||||||
import time
|
import time
|
||||||
import mock
|
import mock
|
||||||
from pprint import pprint
|
from pprint import pprint
|
||||||
|
@ -23,30 +22,34 @@ def test_dont_use_analyze():
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_move_wrong_string_param1():
|
def test_move_wrong_string_param1():
|
||||||
FileMoverAnalyzer.move('', u'', u'', dict())
|
FileMoverAnalyzer.move(42, '', '', dict())
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_move_wrong_string_param2():
|
def test_move_wrong_string_param2():
|
||||||
FileMoverAnalyzer.move(u'', '', u'', dict())
|
FileMoverAnalyzer.move(u'', 23, u'', dict())
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_move_wrong_string_param3():
|
def test_move_wrong_string_param3():
|
||||||
FileMoverAnalyzer.move(u'', u'', '', dict())
|
FileMoverAnalyzer.move('', '', 5, dict())
|
||||||
|
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_move_wrong_dict_param():
|
def test_move_wrong_dict_param():
|
||||||
FileMoverAnalyzer.move(u'', u'', u'', 12345)
|
FileMoverAnalyzer.move('', '', '', 12345)
|
||||||
|
|
||||||
|
@raises(FileNotFoundError)
|
||||||
|
def test_move_wrong_string_param3():
|
||||||
|
FileMoverAnalyzer.move('', '', '', dict())
|
||||||
|
|
||||||
def test_basic():
|
def test_basic():
|
||||||
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
||||||
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
||||||
#Move the file back
|
#Move the file back
|
||||||
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
||||||
|
|
||||||
def test_basic_samefile():
|
def test_basic_samefile():
|
||||||
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
||||||
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'tests/test_data', filename, dict())
|
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'tests/test_data', filename, dict())
|
||||||
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
||||||
|
|
||||||
def test_duplicate_file():
|
def test_duplicate_file():
|
||||||
|
@ -55,9 +58,9 @@ def test_duplicate_file():
|
||||||
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
||||||
#Copy it back to the original location
|
#Copy it back to the original location
|
||||||
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
#Import it again. It shouldn't overwrite the old file and instead create a new
|
#Import it again. It shouldn't overwrite the old file and instead create a new
|
||||||
metadata = dict()
|
metadata = dict()
|
||||||
metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, metadata)
|
metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, metadata)
|
||||||
#Cleanup: move the file (eg. 44100Hz-16bit-mono.mp3) back
|
#Cleanup: move the file (eg. 44100Hz-16bit-mono.mp3) back
|
||||||
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
#Remove the renamed duplicate, eg. 44100Hz-16bit-mono_03-26-2014-11-58.mp3
|
#Remove the renamed duplicate, eg. 44100Hz-16bit-mono_03-26-2014-11-58.mp3
|
||||||
|
@ -71,7 +74,7 @@ def test_duplicate_file():
|
||||||
it's imported within 1 second of the second file (ie. if the timestamp is the same).
|
it's imported within 1 second of the second file (ie. if the timestamp is the same).
|
||||||
'''
|
'''
|
||||||
def test_double_duplicate_files():
|
def test_double_duplicate_files():
|
||||||
# Here we use mock to patch out the time.localtime() function so that it
|
# Here we use mock to patch out the time.localtime() function so that it
|
||||||
# always returns the same value. This allows us to consistently simulate this test cases
|
# always returns the same value. This allows us to consistently simulate this test cases
|
||||||
# where the last two of the three files are imported at the same time as the timestamp.
|
# where the last two of the three files are imported at the same time as the timestamp.
|
||||||
with mock.patch('airtime_analyzer.filemover_analyzer.time') as mock_time:
|
with mock.patch('airtime_analyzer.filemover_analyzer.time') as mock_time:
|
||||||
|
@ -83,17 +86,17 @@ def test_double_duplicate_files():
|
||||||
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename, dict())
|
||||||
#Copy it back to the original location
|
#Copy it back to the original location
|
||||||
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
#Import it again. It shouldn't overwrite the old file and instead create a new
|
#Import it again. It shouldn't overwrite the old file and instead create a new
|
||||||
first_dup_metadata = dict()
|
first_dup_metadata = dict()
|
||||||
first_dup_metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename,
|
first_dup_metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename,
|
||||||
first_dup_metadata)
|
first_dup_metadata)
|
||||||
#Copy it back again!
|
#Copy it back again!
|
||||||
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.copy("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
#Reimport for the third time, which should have the same timestamp as the second one
|
#Reimport for the third time, which should have the same timestamp as the second one
|
||||||
#thanks to us mocking out time.localtime()
|
#thanks to us mocking out time.localtime()
|
||||||
second_dup_metadata = dict()
|
second_dup_metadata = dict()
|
||||||
second_dup_metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename,
|
second_dup_metadata = FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, u'.', filename,
|
||||||
second_dup_metadata)
|
second_dup_metadata)
|
||||||
#Cleanup: move the file (eg. 44100Hz-16bit-mono.mp3) back
|
#Cleanup: move the file (eg. 44100Hz-16bit-mono.mp3) back
|
||||||
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
shutil.move("./" + filename, DEFAULT_AUDIO_FILE)
|
||||||
#Remove the renamed duplicate, eg. 44100Hz-16bit-mono_03-26-2014-11-58.mp3
|
#Remove the renamed duplicate, eg. 44100Hz-16bit-mono_03-26-2014-11-58.mp3
|
||||||
|
@ -105,7 +108,7 @@ def test_double_duplicate_files():
|
||||||
def test_bad_permissions_destination_dir():
|
def test_bad_permissions_destination_dir():
|
||||||
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
filename = os.path.basename(DEFAULT_AUDIO_FILE)
|
||||||
dest_dir = u'/sys/foobar' # /sys is using sysfs on Linux, which is unwritable
|
dest_dir = u'/sys/foobar' # /sys is using sysfs on Linux, which is unwritable
|
||||||
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, dest_dir, filename, dict())
|
FileMoverAnalyzer.move(DEFAULT_AUDIO_FILE, dest_dir, filename, dict())
|
||||||
#Move the file back
|
#Move the file back
|
||||||
shutil.move(os.path.join(dest_dir, filename), DEFAULT_AUDIO_FILE)
|
shutil.move(os.path.join(dest_dir, filename), DEFAULT_AUDIO_FILE)
|
||||||
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
assert os.path.exists(DEFAULT_AUDIO_FILE)
|
||||||
|
|
|
@ -1,9 +1,10 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
from __future__ import print_function
|
||||||
import datetime
|
import datetime
|
||||||
import mutagen
|
import mutagen
|
||||||
import mock
|
import mock
|
||||||
from nose.tools import *
|
from nose.tools import *
|
||||||
from airtime_analyzer.metadata_analyzer import MetadataAnalyzer
|
from airtime_analyzer.metadata_analyzer import MetadataAnalyzer
|
||||||
|
|
||||||
def setup():
|
def setup():
|
||||||
pass
|
pass
|
||||||
|
@ -12,115 +13,115 @@ def teardown():
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def check_default_metadata(metadata):
|
def check_default_metadata(metadata):
|
||||||
assert metadata['track_title'] == u'Test Title'
|
assert metadata['track_title'] == 'Test Title'
|
||||||
assert metadata['artist_name'] == u'Test Artist'
|
assert metadata['artist_name'] == 'Test Artist'
|
||||||
assert metadata['album_title'] == u'Test Album'
|
assert metadata['album_title'] == 'Test Album'
|
||||||
assert metadata['year'] == u'1999'
|
assert metadata['year'] == '1999'
|
||||||
assert metadata['genre'] == u'Test Genre'
|
assert metadata['genre'] == 'Test Genre'
|
||||||
assert metadata['track_number'] == u'1'
|
assert metadata['track_number'] == '1'
|
||||||
assert metadata["length"] == str(datetime.timedelta(seconds=metadata["length_seconds"]))
|
assert metadata["length"] == str(datetime.timedelta(seconds=metadata["length_seconds"]))
|
||||||
|
|
||||||
def test_mp3_mono():
|
def test_mp3_mono():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-mono.mp3', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-mono.mp3', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 1
|
assert metadata['channels'] == 1
|
||||||
assert metadata['bit_rate'] == 63998
|
assert metadata['bit_rate'] == 63998
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3' # Not unicode because MIMEs aren't.
|
assert metadata['mime'] == 'audio/mp3' # Not unicode because MIMEs aren't.
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
#Mutagen doesn't extract comments from mp3s it seems
|
#Mutagen doesn't extract comments from mp3s it seems
|
||||||
|
|
||||||
def test_mp3_jointstereo():
|
def test_mp3_jointstereo():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-jointstereo.mp3', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-jointstereo.mp3', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] == 127998
|
assert metadata['bit_rate'] == 127998
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3'
|
assert metadata['mime'] == 'audio/mp3'
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
|
|
||||||
def test_mp3_simplestereo():
|
def test_mp3_simplestereo():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-simplestereo.mp3', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-simplestereo.mp3', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] == 127998
|
assert metadata['bit_rate'] == 127998
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3'
|
assert metadata['mime'] == 'audio/mp3'
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
|
|
||||||
def test_mp3_dualmono():
|
def test_mp3_dualmono():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-dualmono.mp3', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-dualmono.mp3', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] == 127998
|
assert metadata['bit_rate'] == 127998
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3'
|
assert metadata['mime'] == 'audio/mp3'
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
|
|
||||||
|
|
||||||
def test_ogg_mono():
|
def test_ogg_mono():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-mono.ogg', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-mono.ogg', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 1
|
assert metadata['channels'] == 1
|
||||||
assert metadata['bit_rate'] == 80000
|
assert metadata['bit_rate'] == 80000
|
||||||
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
||||||
assert metadata['mime'] == 'audio/vorbis'
|
assert metadata['mime'] == 'audio/vorbis'
|
||||||
assert metadata['comment'] == u'Test Comment'
|
assert metadata['comment'] == 'Test Comment'
|
||||||
|
|
||||||
def test_ogg_stereo():
|
def test_ogg_stereo():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-stereo.ogg', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-stereo.ogg', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] == 112000
|
assert metadata['bit_rate'] == 112000
|
||||||
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
||||||
assert metadata['mime'] == 'audio/vorbis'
|
assert metadata['mime'] == 'audio/vorbis'
|
||||||
assert metadata['comment'] == u'Test Comment'
|
assert metadata['comment'] == 'Test Comment'
|
||||||
|
|
||||||
''' faac and avconv can't seem to create a proper mono AAC file... ugh
|
''' faac and avconv can't seem to create a proper mono AAC file... ugh
|
||||||
def test_aac_mono():
|
def test_aac_mono():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-mono.m4a')
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-mono.m4a')
|
||||||
print "Mono AAC metadata:"
|
print("Mono AAC metadata:")
|
||||||
print metadata
|
print(metadata)
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 1
|
assert metadata['channels'] == 1
|
||||||
assert metadata['bit_rate'] == 80000
|
assert metadata['bit_rate'] == 80000
|
||||||
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp4'
|
assert metadata['mime'] == 'audio/mp4'
|
||||||
assert metadata['comment'] == u'Test Comment'
|
assert metadata['comment'] == 'Test Comment'
|
||||||
'''
|
'''
|
||||||
|
|
||||||
def test_aac_stereo():
|
def test_aac_stereo():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-stereo.m4a', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-stereo.m4a', dict())
|
||||||
check_default_metadata(metadata)
|
check_default_metadata(metadata)
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] == 102619
|
assert metadata['bit_rate'] == 102619
|
||||||
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
assert abs(metadata['length_seconds'] - 3.8) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp4'
|
assert metadata['mime'] == 'audio/mp4'
|
||||||
assert metadata['comment'] == u'Test Comment'
|
assert metadata['comment'] == 'Test Comment'
|
||||||
|
|
||||||
def test_mp3_utf8():
|
def test_mp3_utf8():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-stereo-utf8.mp3', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-stereo-utf8.mp3', dict())
|
||||||
# Using a bunch of different UTF-8 codepages here. Test data is from:
|
# Using a bunch of different UTF-8 codepages here. Test data is from:
|
||||||
# http://winrus.com/utf8-jap.htm
|
# http://winrus.com/utf8-jap.htm
|
||||||
assert metadata['track_title'] == u'アイウエオカキクケコサシスセソタチツテ'
|
assert metadata['track_title'] == 'アイウエオカキクケコサシスセソタチツテ'
|
||||||
assert metadata['artist_name'] == u'てすと'
|
assert metadata['artist_name'] == 'てすと'
|
||||||
assert metadata['album_title'] == u'Ä ä Ü ü ß'
|
assert metadata['album_title'] == 'Ä ä Ü ü ß'
|
||||||
assert metadata['year'] == u'1999'
|
assert metadata['year'] == '1999'
|
||||||
assert metadata['genre'] == u'Я Б Г Д Ж Й'
|
assert metadata['genre'] == 'Я Б Г Д Ж Й'
|
||||||
assert metadata['track_number'] == u'1'
|
assert metadata['track_number'] == '1'
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
assert metadata['bit_rate'] < 130000
|
assert metadata['bit_rate'] < 130000
|
||||||
assert metadata['bit_rate'] > 127000
|
assert metadata['bit_rate'] > 127000
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3'
|
assert metadata['mime'] == 'audio/mp3'
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
|
|
||||||
def test_invalid_wma():
|
def test_invalid_wma():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-stereo-invalid.wma', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-stereo-invalid.wma', dict())
|
||||||
assert metadata['mime'] == 'audio/x-ms-wma'
|
assert metadata['mime'] == 'audio/x-ms-wma'
|
||||||
|
|
||||||
def test_wav_stereo():
|
def test_wav_stereo():
|
||||||
metadata = MetadataAnalyzer.analyze(u'tests/test_data/44100Hz-16bit-stereo.wav', dict())
|
metadata = MetadataAnalyzer.analyze('tests/test_data/44100Hz-16bit-stereo.wav', dict())
|
||||||
assert metadata['mime'] == 'audio/x-wav'
|
assert metadata['mime'] == 'audio/x-wav'
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['channels'] == 2
|
assert metadata['channels'] == 2
|
||||||
|
@ -128,7 +129,7 @@ def test_wav_stereo():
|
||||||
|
|
||||||
|
|
||||||
# Make sure the parameter checking works
|
# Make sure the parameter checking works
|
||||||
@raises(TypeError)
|
@raises(FileNotFoundError)
|
||||||
def test_move_wrong_string_param1():
|
def test_move_wrong_string_param1():
|
||||||
not_unicode = 'asdfasdf'
|
not_unicode = 'asdfasdf'
|
||||||
MetadataAnalyzer.analyze(not_unicode, dict())
|
MetadataAnalyzer.analyze(not_unicode, dict())
|
||||||
|
@ -136,12 +137,12 @@ def test_move_wrong_string_param1():
|
||||||
@raises(TypeError)
|
@raises(TypeError)
|
||||||
def test_move_wrong_metadata_dict():
|
def test_move_wrong_metadata_dict():
|
||||||
not_a_dict = list()
|
not_a_dict = list()
|
||||||
MetadataAnalyzer.analyze(u'asdfasdf', not_a_dict)
|
MetadataAnalyzer.analyze('asdfasdf', not_a_dict)
|
||||||
|
|
||||||
# Test an mp3 file where the number of channels is invalid or missing:
|
# Test an mp3 file where the number of channels is invalid or missing:
|
||||||
def test_mp3_bad_channels():
|
def test_mp3_bad_channels():
|
||||||
filename = u'tests/test_data/44100Hz-16bit-mono.mp3'
|
filename = 'tests/test_data/44100Hz-16bit-mono.mp3'
|
||||||
'''
|
'''
|
||||||
It'd be a pain in the ass to construct a real MP3 with an invalid number
|
It'd be a pain in the ass to construct a real MP3 with an invalid number
|
||||||
of channels by hand because that value is stored in every MP3 frame in the file
|
of channels by hand because that value is stored in every MP3 frame in the file
|
||||||
'''
|
'''
|
||||||
|
@ -157,8 +158,8 @@ def test_mp3_bad_channels():
|
||||||
assert metadata['bit_rate'] == 63998
|
assert metadata['bit_rate'] == 63998
|
||||||
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
assert abs(metadata['length_seconds'] - 3.9) < 0.1
|
||||||
assert metadata['mime'] == 'audio/mp3' # Not unicode because MIMEs aren't.
|
assert metadata['mime'] == 'audio/mp3' # Not unicode because MIMEs aren't.
|
||||||
assert metadata['track_total'] == u'10' # MP3s can have a track_total
|
assert metadata['track_total'] == '10' # MP3s can have a track_total
|
||||||
#Mutagen doesn't extract comments from mp3s it seems
|
#Mutagen doesn't extract comments from mp3s it seems
|
||||||
|
|
||||||
def test_unparsable_file():
|
def test_unparsable_file():
|
||||||
MetadataAnalyzer.analyze(u'README.rst', dict())
|
MetadataAnalyzer.analyze('README.rst', dict())
|
||||||
|
|
|
@ -1,20 +1,7 @@
|
||||||
|
from __future__ import print_function
|
||||||
from nose.tools import *
|
from nose.tools import *
|
||||||
from airtime_analyzer.replaygain_analyzer import ReplayGainAnalyzer
|
from airtime_analyzer.replaygain_analyzer import ReplayGainAnalyzer
|
||||||
|
|
||||||
'''
|
|
||||||
The tests in here were all tagged with the 'rgain' tag so the can be exluded from being run
|
|
||||||
with nosetest -a '!rgain'. This was needed due to the fact that it is not readily possible
|
|
||||||
to install replaygain on a containerized travis instance.
|
|
||||||
|
|
||||||
We can either give running replaygain test on travis another shot after ubuntu getsan updated
|
|
||||||
gi instrospection allowing us to install gi and gobject into the virtualenv, or we can switch
|
|
||||||
to a full machine and stop using 'sudo: false' on travis.
|
|
||||||
|
|
||||||
Deactivating these tests is a bad fix for now and I plan on looking into it again after
|
|
||||||
most everything else is up and running. For those interesed the tests seem to work locally
|
|
||||||
albeit my results not being up to the given tolerance of 0.30 (which I'm assuming is my rig's
|
|
||||||
problem and would work on travis if replaygain was available).
|
|
||||||
'''
|
|
||||||
|
|
||||||
def check_default_metadata(metadata):
|
def check_default_metadata(metadata):
|
||||||
''' Check that the values extract by Silan/CuePointAnalyzer on our test audio files match what we expect.
|
''' Check that the values extract by Silan/CuePointAnalyzer on our test audio files match what we expect.
|
||||||
|
@ -28,7 +15,7 @@ def check_default_metadata(metadata):
|
||||||
'''
|
'''
|
||||||
tolerance = 0.30
|
tolerance = 0.30
|
||||||
expected_replaygain = 5.0
|
expected_replaygain = 5.0
|
||||||
print metadata['replay_gain']
|
print(metadata['replay_gain'])
|
||||||
assert abs(metadata['replay_gain'] - expected_replaygain) < tolerance
|
assert abs(metadata['replay_gain'] - expected_replaygain) < tolerance
|
||||||
|
|
||||||
def test_missing_replaygain():
|
def test_missing_replaygain():
|
||||||
|
|
|
@ -8,8 +8,7 @@
|
||||||
###############################################################################
|
###############################################################################
|
||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
import urllib
|
import urllib.request, urllib.error, urllib.parse
|
||||||
import urllib2
|
|
||||||
import requests
|
import requests
|
||||||
import socket
|
import socket
|
||||||
import logging
|
import logging
|
||||||
|
@ -21,26 +20,6 @@ from configobj import ConfigObj
|
||||||
AIRTIME_API_VERSION = "1.1"
|
AIRTIME_API_VERSION = "1.1"
|
||||||
|
|
||||||
|
|
||||||
# TODO : Place these functions in some common module. Right now, media
|
|
||||||
# monitor uses the same functions and it would be better to reuse them
|
|
||||||
# instead of copy pasting them around
|
|
||||||
|
|
||||||
def to_unicode(obj, encoding='utf-8'):
|
|
||||||
if isinstance(obj, basestring):
|
|
||||||
if not isinstance(obj, unicode):
|
|
||||||
obj = unicode(obj, encoding)
|
|
||||||
return obj
|
|
||||||
|
|
||||||
def encode_to(obj, encoding='utf-8'):
|
|
||||||
if isinstance(obj, unicode):
|
|
||||||
obj = obj.encode(encoding)
|
|
||||||
return obj
|
|
||||||
|
|
||||||
def convert_dict_value_to_utf8(md):
|
|
||||||
#list comprehension to convert all values of md to utf-8
|
|
||||||
return dict([(item[0], encode_to(item[1], "utf-8")) for item in md.items()])
|
|
||||||
|
|
||||||
|
|
||||||
api_config = {}
|
api_config = {}
|
||||||
|
|
||||||
# URL to get the version number of the server API
|
# URL to get the version number of the server API
|
||||||
|
@ -114,7 +93,7 @@ class ApcUrl(object):
|
||||||
|
|
||||||
def params(self, **params):
|
def params(self, **params):
|
||||||
temp_url = self.base_url
|
temp_url = self.base_url
|
||||||
for k, v in params.iteritems():
|
for k, v in params.items():
|
||||||
wrapped_param = "%%" + k + "%%"
|
wrapped_param = "%%" + k + "%%"
|
||||||
if wrapped_param in temp_url:
|
if wrapped_param in temp_url:
|
||||||
temp_url = temp_url.replace(wrapped_param, str(v))
|
temp_url = temp_url.replace(wrapped_param, str(v))
|
||||||
|
@ -138,12 +117,13 @@ class ApiRequest(object):
|
||||||
|
|
||||||
def __call__(self,_post_data=None, **kwargs):
|
def __call__(self,_post_data=None, **kwargs):
|
||||||
final_url = self.url.params(**kwargs).url()
|
final_url = self.url.params(**kwargs).url()
|
||||||
if _post_data is not None: _post_data = urllib.urlencode(_post_data)
|
if _post_data is not None:
|
||||||
|
_post_data = urllib.parse.urlencode(_post_data).encode('utf-8')
|
||||||
self.logger.debug(final_url)
|
self.logger.debug(final_url)
|
||||||
try:
|
try:
|
||||||
req = urllib2.Request(final_url, _post_data)
|
req = urllib.request.Request(final_url, _post_data)
|
||||||
f = urllib2.urlopen(req, timeout=ApiRequest.API_HTTP_REQUEST_TIMEOUT)
|
f = urllib.request.urlopen(req, timeout=ApiRequest.API_HTTP_REQUEST_TIMEOUT)
|
||||||
content_type = f.info().getheader('Content-Type')
|
content_type = f.info().get_content_type()
|
||||||
response = f.read()
|
response = f.read()
|
||||||
#Everything that calls an ApiRequest should be catching URLError explicitly
|
#Everything that calls an ApiRequest should be catching URLError explicitly
|
||||||
#(according to the other comments in this file and a cursory grep through the code)
|
#(according to the other comments in this file and a cursory grep through the code)
|
||||||
|
@ -151,20 +131,22 @@ class ApiRequest(object):
|
||||||
except socket.timeout:
|
except socket.timeout:
|
||||||
self.logger.error('HTTP request to %s timed out', final_url)
|
self.logger.error('HTTP request to %s timed out', final_url)
|
||||||
raise
|
raise
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
#self.logger.error('Exception: %s', e)
|
#self.logger.exception(e)
|
||||||
#self.logger.error("traceback: %s", traceback.format_exc())
|
|
||||||
raise
|
raise
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if content_type == 'application/json':
|
if content_type == 'application/json':
|
||||||
|
try:
|
||||||
|
response = response.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
data = json.loads(response)
|
data = json.loads(response)
|
||||||
return data
|
return data
|
||||||
else:
|
else:
|
||||||
raise InvalidContentType()
|
raise InvalidContentType()
|
||||||
except Exception:
|
except Exception:
|
||||||
#self.logger.error(response)
|
#self.logger.exception(e)
|
||||||
#self.logger.error("traceback: %s", traceback.format_exc())
|
|
||||||
raise
|
raise
|
||||||
|
|
||||||
def req(self, *args, **kwargs):
|
def req(self, *args, **kwargs):
|
||||||
|
@ -193,18 +175,20 @@ class RequestProvider(object):
|
||||||
self.config["general"]["base_dir"], self.config["api_base"],
|
self.config["general"]["base_dir"], self.config["api_base"],
|
||||||
'%%action%%'))
|
'%%action%%'))
|
||||||
# Now we must discover the possible actions
|
# Now we must discover the possible actions
|
||||||
actions = dict( (k,v) for k,v in cfg.iteritems() if '%%api_key%%' in v)
|
actions = dict( (k,v) for k,v in cfg.items() if '%%api_key%%' in v)
|
||||||
for action_name, action_value in actions.iteritems():
|
for action_name, action_value in actions.items():
|
||||||
new_url = self.url.params(action=action_value).params(
|
new_url = self.url.params(action=action_value).params(
|
||||||
api_key=self.config["general"]['api_key'])
|
api_key=self.config["general"]['api_key'])
|
||||||
self.requests[action_name] = ApiRequest(action_name, new_url)
|
self.requests[action_name] = ApiRequest(action_name, new_url)
|
||||||
|
|
||||||
def available_requests(self) : return self.requests.keys()
|
def available_requests(self) : return list(self.requests.keys())
|
||||||
def __contains__(self, request) : return request in self.requests
|
def __contains__(self, request) : return request in self.requests
|
||||||
|
|
||||||
def __getattr__(self, attr):
|
def __getattr__(self, attr):
|
||||||
if attr in self: return self.requests[attr]
|
if attr in self:
|
||||||
else: return super(RequestProvider, self).__getattribute__(attr)
|
return self.requests[attr]
|
||||||
|
else:
|
||||||
|
return super(RequestProvider, self).__getattribute__(attr)
|
||||||
|
|
||||||
|
|
||||||
class AirtimeApiClient(object):
|
class AirtimeApiClient(object):
|
||||||
|
@ -217,17 +201,16 @@ class AirtimeApiClient(object):
|
||||||
self.config = ConfigObj(config_path)
|
self.config = ConfigObj(config_path)
|
||||||
self.config.update(api_config)
|
self.config.update(api_config)
|
||||||
self.services = RequestProvider(self.config)
|
self.services = RequestProvider(self.config)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error('Error loading config file: %s', config_path)
|
self.logger.exception('Error loading config file: %s', config_path)
|
||||||
self.logger.error("traceback: %s", traceback.format_exc())
|
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
def __get_airtime_version(self):
|
def __get_airtime_version(self):
|
||||||
try: return self.services.version_url()[u'airtime_version']
|
try: return self.services.version_url()['airtime_version']
|
||||||
except Exception: return -1
|
except Exception: return -1
|
||||||
|
|
||||||
def __get_api_version(self):
|
def __get_api_version(self):
|
||||||
try: return self.services.version_url()[u'api_version']
|
try: return self.services.version_url()['api_version']
|
||||||
except Exception: return -1
|
except Exception: return -1
|
||||||
|
|
||||||
def is_server_compatible(self, verbose=True):
|
def is_server_compatible(self, verbose=True):
|
||||||
|
@ -259,8 +242,8 @@ class AirtimeApiClient(object):
|
||||||
def notify_liquidsoap_started(self):
|
def notify_liquidsoap_started(self):
|
||||||
try:
|
try:
|
||||||
self.services.notify_liquidsoap_started()
|
self.services.notify_liquidsoap_started()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
|
|
||||||
def notify_media_item_start_playing(self, media_id):
|
def notify_media_item_start_playing(self, media_id):
|
||||||
""" This is a callback from liquidsoap, we use this to notify
|
""" This is a callback from liquidsoap, we use this to notify
|
||||||
|
@ -268,15 +251,15 @@ class AirtimeApiClient(object):
|
||||||
which we handed to liquidsoap in get_liquidsoap_data(). """
|
which we handed to liquidsoap in get_liquidsoap_data(). """
|
||||||
try:
|
try:
|
||||||
return self.services.update_start_playing_url(media_id=media_id)
|
return self.services.update_start_playing_url(media_id=media_id)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def get_shows_to_record(self):
|
def get_shows_to_record(self):
|
||||||
try:
|
try:
|
||||||
return self.services.show_schedule_url()
|
return self.services.show_schedule_url()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def upload_recorded_show(self, files, show_id):
|
def upload_recorded_show(self, files, show_id):
|
||||||
|
@ -321,15 +304,14 @@ class AirtimeApiClient(object):
|
||||||
"""
|
"""
|
||||||
break
|
break
|
||||||
|
|
||||||
except requests.exceptions.HTTPError, e:
|
except requests.exceptions.HTTPError as e:
|
||||||
logger.error("Http error code: %s", e.code)
|
logger.error("Http error code: %s", e.code)
|
||||||
logger.error("traceback: %s", traceback.format_exc())
|
logger.error("traceback: %s", traceback.format_exc())
|
||||||
except requests.exceptions.ConnectionError, e:
|
except requests.exceptions.ConnectionError as e:
|
||||||
logger.error("Server is down: %s", e.args)
|
logger.error("Server is down: %s", e.args)
|
||||||
logger.error("traceback: %s", traceback.format_exc())
|
logger.error("traceback: %s", traceback.format_exc())
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
logger.error("Exception: %s", e)
|
self.logger.exception(e)
|
||||||
logger.error("traceback: %s", traceback.format_exc())
|
|
||||||
|
|
||||||
#wait some time before next retry
|
#wait some time before next retry
|
||||||
time.sleep(retries_wait)
|
time.sleep(retries_wait)
|
||||||
|
@ -340,8 +322,8 @@ class AirtimeApiClient(object):
|
||||||
try:
|
try:
|
||||||
return self.services.check_live_stream_auth(
|
return self.services.check_live_stream_auth(
|
||||||
username=username, password=password, djtype=dj_type)
|
username=username, password=password, djtype=dj_type)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
return {}
|
return {}
|
||||||
|
|
||||||
def construct_url(self,config_action_key):
|
def construct_url(self,config_action_key):
|
||||||
|
@ -407,7 +389,7 @@ class AirtimeApiClient(object):
|
||||||
# Note that we must prefix every key with: mdX where x is a number
|
# Note that we must prefix every key with: mdX where x is a number
|
||||||
# Is there a way to format the next line a little better? The
|
# Is there a way to format the next line a little better? The
|
||||||
# parenthesis make the code almost unreadable
|
# parenthesis make the code almost unreadable
|
||||||
md_list = dict((("md%d" % i), json.dumps(convert_dict_value_to_utf8(md))) \
|
md_list = dict((("md%d" % i), json.dumps(md)) \
|
||||||
for i,md in enumerate(valid_actions))
|
for i,md in enumerate(valid_actions))
|
||||||
# For testing we add the following "dry" parameter to tell the
|
# For testing we add the following "dry" parameter to tell the
|
||||||
# controller not to actually do any changes
|
# controller not to actually do any changes
|
||||||
|
@ -422,10 +404,10 @@ class AirtimeApiClient(object):
|
||||||
def list_all_db_files(self, dir_id, all_files=True):
|
def list_all_db_files(self, dir_id, all_files=True):
|
||||||
logger = self.logger
|
logger = self.logger
|
||||||
try:
|
try:
|
||||||
all_files = u"1" if all_files else u"0"
|
all_files = "1" if all_files else "0"
|
||||||
response = self.services.list_all_db_files(dir_id=dir_id,
|
response = self.services.list_all_db_files(dir_id=dir_id,
|
||||||
all=all_files)
|
all=all_files)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
response = {}
|
response = {}
|
||||||
logger.error("Exception: %s", e)
|
logger.error("Exception: %s", e)
|
||||||
try:
|
try:
|
||||||
|
@ -483,23 +465,20 @@ class AirtimeApiClient(object):
|
||||||
post_data = {"msg_post": msg}
|
post_data = {"msg_post": msg}
|
||||||
|
|
||||||
#encoded_msg is no longer used server_side!!
|
#encoded_msg is no longer used server_side!!
|
||||||
encoded_msg = urllib.quote('dummy')
|
encoded_msg = urllib.parse.quote('dummy')
|
||||||
self.services.update_liquidsoap_status.req(post_data,
|
self.services.update_liquidsoap_status.req(post_data,
|
||||||
msg=encoded_msg,
|
msg=encoded_msg,
|
||||||
stream_id=stream_id,
|
stream_id=stream_id,
|
||||||
boot_time=time).retry(5)
|
boot_time=time).retry(5)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
#TODO
|
self.logger.exception(e)
|
||||||
logger.error("Exception: %s", e)
|
|
||||||
|
|
||||||
def notify_source_status(self, sourcename, status):
|
def notify_source_status(self, sourcename, status):
|
||||||
try:
|
try:
|
||||||
logger = self.logger
|
|
||||||
return self.services.update_source_status.req(sourcename=sourcename,
|
return self.services.update_source_status.req(sourcename=sourcename,
|
||||||
status=status).retry(5)
|
status=status).retry(5)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
#TODO
|
self.logger.exception(e)
|
||||||
logger.error("Exception: %s", e)
|
|
||||||
|
|
||||||
def get_bootstrap_info(self):
|
def get_bootstrap_info(self):
|
||||||
""" Retrieve infomations needed on bootstrap time """
|
""" Retrieve infomations needed on bootstrap time """
|
||||||
|
@ -514,8 +493,8 @@ class AirtimeApiClient(object):
|
||||||
#http://localhost/api/get-files-without-replay-gain/dir_id/1
|
#http://localhost/api/get-files-without-replay-gain/dir_id/1
|
||||||
try:
|
try:
|
||||||
return self.services.get_files_without_replay_gain(dir_id=dir_id)
|
return self.services.get_files_without_replay_gain(dir_id=dir_id)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
return []
|
return []
|
||||||
|
|
||||||
def get_files_without_silan_value(self):
|
def get_files_without_silan_value(self):
|
||||||
|
@ -526,8 +505,8 @@ class AirtimeApiClient(object):
|
||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
return self.services.get_files_without_silan_value()
|
return self.services.get_files_without_silan_value()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
return []
|
return []
|
||||||
|
|
||||||
def update_replay_gain_values(self, pairs):
|
def update_replay_gain_values(self, pairs):
|
||||||
|
@ -569,9 +548,8 @@ class AirtimeApiClient(object):
|
||||||
try:
|
try:
|
||||||
response = self.services.update_stream_setting_table(_post_data={'data': json.dumps(data)})
|
response = self.services.update_stream_setting_table(_post_data={'data': json.dumps(data)})
|
||||||
return response
|
return response
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
#TODO
|
self.logger.exception(e)
|
||||||
self.logger.error(str(e))
|
|
||||||
|
|
||||||
def update_metadata_on_tunein(self):
|
def update_metadata_on_tunein(self):
|
||||||
self.services.update_metadata_on_tunein()
|
self.services.update_metadata_on_tunein()
|
||||||
|
|
|
@ -1,10 +1,11 @@
|
||||||
|
from __future__ import print_function
|
||||||
from setuptools import setup
|
from setuptools import setup
|
||||||
from subprocess import call
|
from subprocess import call
|
||||||
import sys
|
import sys
|
||||||
import os
|
import os
|
||||||
|
|
||||||
script_path = os.path.dirname(os.path.realpath(__file__))
|
script_path = os.path.dirname(os.path.realpath(__file__))
|
||||||
print script_path
|
print(script_path)
|
||||||
os.chdir(script_path)
|
os.chdir(script_path)
|
||||||
|
|
||||||
setup(name='api_clients',
|
setup(name='api_clients',
|
||||||
|
@ -16,18 +17,7 @@ setup(name='api_clients',
|
||||||
packages=['api_clients'],
|
packages=['api_clients'],
|
||||||
scripts=[],
|
scripts=[],
|
||||||
install_requires=[
|
install_requires=[
|
||||||
# 'amqplib',
|
|
||||||
# 'anyjson',
|
|
||||||
# 'argparse',
|
|
||||||
'configobj'
|
'configobj'
|
||||||
# 'docopt',
|
|
||||||
# 'kombu',
|
|
||||||
# 'mutagen',
|
|
||||||
# 'poster',
|
|
||||||
# 'PyDispatcher',
|
|
||||||
# 'pyinotify',
|
|
||||||
# 'pytz',
|
|
||||||
# 'wsgiref'
|
|
||||||
],
|
],
|
||||||
zip_safe=False,
|
zip_safe=False,
|
||||||
data_files=[])
|
data_files=[])
|
||||||
|
|
|
@ -1,20 +1,20 @@
|
||||||
import unittest
|
import unittest
|
||||||
from .. api_client import ApcUrl, UrlBadParam, IncompleteUrl
|
from api_clients.api_client import ApcUrl, UrlBadParam, IncompleteUrl
|
||||||
|
|
||||||
class TestApcUrl(unittest.TestCase):
|
class TestApcUrl(unittest.TestCase):
|
||||||
def test_init(self):
|
def test_init(self):
|
||||||
url = "/testing"
|
url = "/testing"
|
||||||
u = ApcUrl(url)
|
u = ApcUrl(url)
|
||||||
self.assertEquals( u.base_url, url)
|
self.assertEqual( u.base_url, url)
|
||||||
|
|
||||||
def test_params_1(self):
|
def test_params_1(self):
|
||||||
u = ApcUrl("/testing/%%key%%")
|
u = ApcUrl("/testing/%%key%%")
|
||||||
self.assertEquals(u.params(key='val').url(), '/testing/val')
|
self.assertEqual(u.params(key='val').url(), '/testing/val')
|
||||||
|
|
||||||
def test_params_2(self):
|
def test_params_2(self):
|
||||||
u = ApcUrl('/testing/%%key%%/%%api%%/more_testing')
|
u = ApcUrl('/testing/%%key%%/%%api%%/more_testing')
|
||||||
full_url = u.params(key="AAA",api="BBB").url()
|
full_url = u.params(key="AAA",api="BBB").url()
|
||||||
self.assertEquals(full_url, '/testing/AAA/BBB/more_testing')
|
self.assertEqual(full_url, '/testing/AAA/BBB/more_testing')
|
||||||
|
|
||||||
def test_params_ex(self):
|
def test_params_ex(self):
|
||||||
u = ApcUrl("/testing/%%key%%")
|
u = ApcUrl("/testing/%%key%%")
|
||||||
|
@ -23,7 +23,7 @@ class TestApcUrl(unittest.TestCase):
|
||||||
|
|
||||||
def test_url(self):
|
def test_url(self):
|
||||||
u = "one/two/three"
|
u = "one/two/three"
|
||||||
self.assertEquals( ApcUrl(u).url(), u )
|
self.assertEqual( ApcUrl(u).url(), u )
|
||||||
|
|
||||||
def test_url_ex(self):
|
def test_url_ex(self):
|
||||||
u = ApcUrl('/%%one%%/%%two%%/three').params(two='testing')
|
u = ApcUrl('/%%one%%/%%two%%/three').params(two='testing')
|
||||||
|
|
|
@ -1,21 +1,26 @@
|
||||||
import unittest
|
import unittest
|
||||||
import json
|
import json
|
||||||
from mock import MagicMock, patch
|
from mock import MagicMock, patch
|
||||||
from .. api_client import ApcUrl, ApiRequest
|
from api_clients.api_client import ApcUrl, ApiRequest
|
||||||
|
|
||||||
|
class ResponseInfo:
|
||||||
|
def get_content_type(self):
|
||||||
|
return 'application/json'
|
||||||
|
|
||||||
class TestApiRequest(unittest.TestCase):
|
class TestApiRequest(unittest.TestCase):
|
||||||
def test_init(self):
|
def test_init(self):
|
||||||
u = ApiRequest('request_name', ApcUrl('/test/ing'))
|
u = ApiRequest('request_name', ApcUrl('/test/ing'))
|
||||||
self.assertEquals(u.name, "request_name")
|
self.assertEqual(u.name, "request_name")
|
||||||
|
|
||||||
def test_call(self):
|
def test_call(self):
|
||||||
ret = json.dumps( {u'ok':u'ok'} )
|
ret = json.dumps( {'ok':'ok'} )
|
||||||
read = MagicMock()
|
read = MagicMock()
|
||||||
read.read = MagicMock(return_value=ret)
|
read.read = MagicMock(return_value=ret)
|
||||||
u = '/testing'
|
read.info = MagicMock(return_value=ResponseInfo())
|
||||||
with patch('urllib2.urlopen') as mock_method:
|
u = 'http://localhost/testing'
|
||||||
|
with patch('urllib.request.urlopen') as mock_method:
|
||||||
mock_method.return_value = read
|
mock_method.return_value = read
|
||||||
request = ApiRequest('mm', ApcUrl(u))()
|
request = ApiRequest('mm', ApcUrl(u))()
|
||||||
self.assertEquals(request, json.loads(ret))
|
self.assertEqual(request, json.loads(ret))
|
||||||
|
|
||||||
if __name__ == '__main__': unittest.main()
|
if __name__ == '__main__': unittest.main()
|
||||||
|
|
|
@ -2,13 +2,19 @@ import unittest
|
||||||
import json
|
import json
|
||||||
from mock import patch, MagicMock
|
from mock import patch, MagicMock
|
||||||
from configobj import ConfigObj
|
from configobj import ConfigObj
|
||||||
from .. api_client import RequestProvider
|
from api_clients.api_client import RequestProvider, api_config
|
||||||
|
|
||||||
class TestRequestProvider(unittest.TestCase):
|
class TestRequestProvider(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.cfg = ConfigObj('api_client.cfg')
|
self.cfg = api_config
|
||||||
|
self.cfg['general'] = {}
|
||||||
|
self.cfg['general']['base_dir'] = '/test'
|
||||||
|
self.cfg['general']['base_port'] = 80
|
||||||
|
self.cfg['general']['base_url'] = 'localhost'
|
||||||
|
self.cfg['general']['api_key'] = 'TEST_KEY'
|
||||||
|
self.cfg['api_base'] = 'api'
|
||||||
def test_test(self):
|
def test_test(self):
|
||||||
self.assertTrue('api_key' in self.cfg)
|
self.assertTrue('general' in self.cfg)
|
||||||
def test_init(self):
|
def test_init(self):
|
||||||
rp = RequestProvider(self.cfg)
|
rp = RequestProvider(self.cfg)
|
||||||
self.assertTrue( len( rp.available_requests() ) > 0 )
|
self.assertTrue( len( rp.available_requests() ) > 0 )
|
||||||
|
@ -16,17 +22,6 @@ class TestRequestProvider(unittest.TestCase):
|
||||||
rp = RequestProvider(self.cfg)
|
rp = RequestProvider(self.cfg)
|
||||||
methods = ['upload_recorded', 'update_media_url', 'list_all_db_files']
|
methods = ['upload_recorded', 'update_media_url', 'list_all_db_files']
|
||||||
for meth in methods:
|
for meth in methods:
|
||||||
self.assertTrue( meth in rp )
|
self.assertTrue( meth in rp.requests )
|
||||||
|
|
||||||
def test_notify_webstream_data(self):
|
|
||||||
ret = json.dumps( {u'testing' : u'123' } )
|
|
||||||
rp = RequestProvider(self.cfg)
|
|
||||||
read = MagicMock()
|
|
||||||
read.read = MagicMock(return_value=ret)
|
|
||||||
with patch('urllib2.urlopen') as mock_method:
|
|
||||||
mock_method.return_value = read
|
|
||||||
response = rp.notify_webstream_data(media_id=123)
|
|
||||||
mock_method.called_once_with(media_id=123)
|
|
||||||
self.assertEquals(json.loads(ret), response)
|
|
||||||
|
|
||||||
if __name__ == '__main__': unittest.main()
|
if __name__ == '__main__': unittest.main()
|
||||||
|
|
|
@ -1,11 +1,12 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
|
||||||
import shutil
|
import shutil
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
if os.geteuid() != 0:
|
if os.geteuid() != 0:
|
||||||
print "Please run this as root."
|
print("Please run this as root.")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
def get_current_script_dir():
|
def get_current_script_dir():
|
||||||
|
@ -17,6 +18,6 @@ try:
|
||||||
current_script_dir = get_current_script_dir()
|
current_script_dir = get_current_script_dir()
|
||||||
shutil.copy(current_script_dir+"/../airtime-icecast-status.xsl", "/usr/share/icecast2/web")
|
shutil.copy(current_script_dir+"/../airtime-icecast-status.xsl", "/usr/share/icecast2/web")
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
print "exception: %s" % e
|
print("exception: {}".format(e))
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
#!/usr/bin/env python
|
#!/usr/bin/env python
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
import traceback
|
import traceback
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
@ -75,7 +76,7 @@ logger = rootLogger
|
||||||
try:
|
try:
|
||||||
config = ConfigObj('/etc/airtime/airtime.conf')
|
config = ConfigObj('/etc/airtime/airtime.conf')
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
logger.error('Error loading config file: %s', e)
|
logger.error('Error loading config file: %s', e)
|
||||||
sys.exit()
|
sys.exit()
|
||||||
|
|
||||||
|
@ -133,21 +134,20 @@ class Notify:
|
||||||
elif options.liquidsoap_started:
|
elif options.liquidsoap_started:
|
||||||
self.notify_liquidsoap_started()
|
self.notify_liquidsoap_started()
|
||||||
else:
|
else:
|
||||||
logger.debug("Unrecognized option in options(%s). Doing nothing" \
|
logger.debug("Unrecognized option in options({}). Doing nothing".format(options))
|
||||||
% str(options))
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
print
|
print()
|
||||||
print '#########################################'
|
print('#########################################')
|
||||||
print '# *** pypo *** #'
|
print('# *** pypo *** #')
|
||||||
print '# pypo notification gateway #'
|
print('# pypo notification gateway #')
|
||||||
print '#########################################'
|
print('#########################################')
|
||||||
|
|
||||||
# initialize
|
# initialize
|
||||||
try:
|
try:
|
||||||
n = Notify()
|
n = Notify()
|
||||||
n.run_with_options(options)
|
n.run_with_options(options)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print( traceback.format_exc() )
|
print(traceback.format_exc())
|
||||||
|
|
||||||
|
|
|
@ -1,29 +1,29 @@
|
||||||
""" Runs Airtime liquidsoap
|
""" Runs Airtime liquidsoap
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
import os
|
import os
|
||||||
import generate_liquidsoap_cfg
|
from . import generate_liquidsoap_cfg
|
||||||
import logging
|
import logging
|
||||||
import subprocess
|
import subprocess
|
||||||
|
from pypo import pure
|
||||||
|
|
||||||
PYPO_HOME = '/var/tmp/airtime/pypo/'
|
PYPO_HOME = '/var/tmp/airtime/pypo/'
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
'''Entry-point for this application'''
|
'''Entry-point for this application'''
|
||||||
print "Airtime Liquidsoap"
|
print("Airtime Liquidsoap")
|
||||||
parser = argparse.ArgumentParser()
|
parser = argparse.ArgumentParser()
|
||||||
parser.add_argument("-d", "--debug", help="run in debug mode", action="store_true")
|
parser.add_argument("-d", "--debug", help="run in debug mode", action="store_true")
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
os.environ["HOME"] = PYPO_HOME
|
os.environ["HOME"] = PYPO_HOME
|
||||||
|
|
||||||
if args.debug:
|
if args.debug:
|
||||||
logging.basicConfig(level=getattr(logging, 'DEBUG', None))
|
logging.basicConfig(level=getattr(logging, 'DEBUG', None))
|
||||||
|
|
||||||
generate_liquidsoap_cfg.run()
|
generate_liquidsoap_cfg.run()
|
||||||
''' check liquidsoap version if less than 1.3 use legacy liquidsoap script '''
|
''' check liquidsoap version if less than 1.3 use legacy liquidsoap script '''
|
||||||
liquidsoap_version=subprocess.check_output("liquidsoap --version", shell=True)
|
liquidsoap_version = subprocess.check_output("liquidsoap --version", shell=True, universal_newlines=True)
|
||||||
if "1.1.1" not in liquidsoap_version:
|
if "1.1.1" not in liquidsoap_version:
|
||||||
script_path = os.path.join(os.path.dirname(__file__), 'ls_script.liq')
|
script_path = os.path.join(os.path.dirname(__file__), 'ls_script.liq')
|
||||||
else:
|
else:
|
||||||
|
|
|
@ -1,3 +1,4 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
@ -13,7 +14,7 @@ def generate_liquidsoap_config(ss):
|
||||||
fh.write("################################################\n")
|
fh.write("################################################\n")
|
||||||
fh.write("# The ignore() lines are to squash unused variable warnings\n")
|
fh.write("# The ignore() lines are to squash unused variable warnings\n")
|
||||||
|
|
||||||
for key, value in data.iteritems():
|
for key, value in data.items():
|
||||||
try:
|
try:
|
||||||
if not "port" in key and not "bitrate" in key: # Stupid hack
|
if not "port" in key and not "bitrate" in key: # Stupid hack
|
||||||
raise ValueError()
|
raise ValueError()
|
||||||
|
@ -27,29 +28,29 @@ def generate_liquidsoap_config(ss):
|
||||||
except: #Everything else is a string
|
except: #Everything else is a string
|
||||||
str_buffer = "%s = \"%s\"\n" % (key, value)
|
str_buffer = "%s = \"%s\"\n" % (key, value)
|
||||||
|
|
||||||
fh.write(str_buffer.encode('utf-8'))
|
fh.write(str_buffer)
|
||||||
# ignore squashes unused variable errors from Liquidsoap
|
# ignore squashes unused variable errors from Liquidsoap
|
||||||
fh.write(("ignore(%s)\n" % key).encode('utf-8'))
|
fh.write("ignore(%s)\n" % key)
|
||||||
|
|
||||||
auth_path = os.path.dirname(os.path.realpath(__file__))
|
auth_path = os.path.dirname(os.path.realpath(__file__))
|
||||||
fh.write('log_file = "/var/log/airtime/pypo-liquidsoap/<script>.log"\n')
|
fh.write('log_file = "/var/log/airtime/pypo-liquidsoap/<script>.log"\n')
|
||||||
fh.write('auth_path = "%s/liquidsoap_auth.py"\n' % auth_path)
|
fh.write('auth_path = "%s/liquidsoap_auth.py"\n' % auth_path)
|
||||||
fh.close()
|
fh.close()
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
logging.basicConfig(format='%(message)s')
|
logging.basicConfig(format='%(message)s')
|
||||||
attempts = 0
|
attempts = 0
|
||||||
max_attempts = 10
|
max_attempts = 10
|
||||||
successful = False
|
successful = False
|
||||||
|
|
||||||
while not successful:
|
while not successful:
|
||||||
try:
|
try:
|
||||||
ac = AirtimeApiClient(logging.getLogger())
|
ac = AirtimeApiClient(logging.getLogger())
|
||||||
ss = ac.get_stream_setting()
|
ss = ac.get_stream_setting()
|
||||||
generate_liquidsoap_config(ss)
|
generate_liquidsoap_config(ss)
|
||||||
successful = True
|
successful = True
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
print "Unable to connect to the Airtime server."
|
print("Unable to connect to the Airtime server.")
|
||||||
logging.error(str(e))
|
logging.error(str(e))
|
||||||
logging.error("traceback: %s", traceback.format_exc())
|
logging.error("traceback: %s", traceback.format_exc())
|
||||||
if attempts == max_attempts:
|
if attempts == max_attempts:
|
||||||
|
|
|
@ -1,3 +1,4 @@
|
||||||
|
|
||||||
from api_clients import *
|
from api_clients import *
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
@ -16,8 +17,8 @@ elif dj_type == '--dj':
|
||||||
response = api_clients.check_live_stream_auth(username, password, source_type)
|
response = api_clients.check_live_stream_auth(username, password, source_type)
|
||||||
|
|
||||||
if 'msg' in response and response['msg'] == True:
|
if 'msg' in response and response['msg'] == True:
|
||||||
print response['msg']
|
print(response['msg'])
|
||||||
sys.exit(0)
|
sys.exit(0)
|
||||||
else:
|
else:
|
||||||
print False
|
print(False)
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
|
@ -13,7 +13,7 @@ try:
|
||||||
tn.write('exit\n')
|
tn.write('exit\n')
|
||||||
tn.read_all()
|
tn.read_all()
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
print('Error loading config file: %s', e)
|
print("Error loading config file: {}".format(e))
|
||||||
sys.exit()
|
sys.exit()
|
||||||
|
|
||||||
|
|
|
@ -244,7 +244,7 @@ def check_auth(user="", password="", ~type="master") =
|
||||||
log("#{type} user #{user} connected",label="#{type}_source")
|
log("#{type} user #{user} connected",label="#{type}_source")
|
||||||
|
|
||||||
# Check auth based on return value from auth script
|
# Check auth based on return value from auth script
|
||||||
ret = snd(snd(run_process("python #{auth_path} --#{type} #{user} #{password}"))) == "0"
|
ret = snd(snd(run_process("python3 #{auth_path} --#{type} #{user} #{password}"))) == "0"
|
||||||
|
|
||||||
if ret then
|
if ret then
|
||||||
log("#{type} user #{user} authenticated",label="#{type}_source")
|
log("#{type} user #{user} authenticated",label="#{type}_source")
|
||||||
|
|
|
@ -239,7 +239,7 @@ end
|
||||||
def check_master_dj_client(user,password) =
|
def check_master_dj_client(user,password) =
|
||||||
log("master connected")
|
log("master connected")
|
||||||
#get the output of the php script
|
#get the output of the php script
|
||||||
ret = get_process_lines("python #{auth_path} --master #{user} #{password}")
|
ret = get_process_lines("python3 #{auth_path} --master #{user} #{password}")
|
||||||
#ret has now the value of the live client (dj1,dj2, or djx), or "ERROR"/"unknown" ...
|
#ret has now the value of the live client (dj1,dj2, or djx), or "ERROR"/"unknown" ...
|
||||||
ret = list.hd(ret)
|
ret = list.hd(ret)
|
||||||
|
|
||||||
|
@ -250,7 +250,7 @@ end
|
||||||
def check_dj_client(user,password) =
|
def check_dj_client(user,password) =
|
||||||
log("live dj connected")
|
log("live dj connected")
|
||||||
#get the output of the php script
|
#get the output of the php script
|
||||||
ret = get_process_lines("python #{auth_path} --dj #{user} #{password}")
|
ret = get_process_lines("python3 #{auth_path} --dj #{user} #{password}")
|
||||||
#ret has now the value of the live client (dj1,dj2, or djx), or "ERROR"/"unknown" ...
|
#ret has now the value of the live client (dj1,dj2, or djx), or "ERROR"/"unknown" ...
|
||||||
hd = list.hd(ret)
|
hd = list.hd(ret)
|
||||||
log("Live DJ authenticated: #{hd}")
|
log("Live DJ authenticated: #{hd}")
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
"""
|
"""
|
||||||
Python part of radio playout (pypo)
|
Python part of radio playout (pypo)
|
||||||
"""
|
"""
|
||||||
from __future__ import absolute_import
|
|
||||||
|
|
||||||
import locale
|
import locale
|
||||||
import logging
|
import logging
|
||||||
|
@ -16,10 +16,11 @@ from api_clients import api_client
|
||||||
from configobj import ConfigObj
|
from configobj import ConfigObj
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from optparse import OptionParser
|
from optparse import OptionParser
|
||||||
|
import importlib
|
||||||
try:
|
try:
|
||||||
from queue import Queue
|
from queue import Queue
|
||||||
except ImportError: # Python 2.7.5 (CentOS 7)
|
except ImportError: # Python 2.7.5 (CentOS 7)
|
||||||
from Queue import Queue
|
from queue import Queue
|
||||||
from threading import Lock
|
from threading import Lock
|
||||||
|
|
||||||
from .listenerstat import ListenerStat
|
from .listenerstat import ListenerStat
|
||||||
|
@ -119,62 +120,9 @@ try:
|
||||||
consoleHandler.setFormatter(logFormatter)
|
consoleHandler.setFormatter(logFormatter)
|
||||||
rootLogger.addHandler(consoleHandler)
|
rootLogger.addHandler(consoleHandler)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print("Couldn't configure logging", e)
|
print("Couldn't configure logging: {}".format(e))
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
def configure_locale():
|
|
||||||
"""
|
|
||||||
Silly hacks to force Python 2.x to run in UTF-8 mode. Not portable at all,
|
|
||||||
however serves our purpose at the moment.
|
|
||||||
|
|
||||||
More information available here:
|
|
||||||
http://stackoverflow.com/questions/3828723/why-we-need-sys-setdefaultencodingutf-8-in-a-py-script
|
|
||||||
"""
|
|
||||||
logger.debug("Before %s", locale.nl_langinfo(locale.CODESET))
|
|
||||||
current_locale = locale.getlocale()
|
|
||||||
|
|
||||||
if current_locale[1] is None:
|
|
||||||
logger.debug("No locale currently set. Attempting to get default locale.")
|
|
||||||
default_locale = locale.getdefaultlocale()
|
|
||||||
|
|
||||||
if default_locale[1] is None:
|
|
||||||
logger.debug(
|
|
||||||
"No default locale exists. Let's try loading from \
|
|
||||||
/etc/default/locale"
|
|
||||||
)
|
|
||||||
if os.path.exists("/etc/default/locale"):
|
|
||||||
locale_config = ConfigObj("/etc/default/locale")
|
|
||||||
lang = locale_config.get("LANG")
|
|
||||||
new_locale = lang
|
|
||||||
else:
|
|
||||||
logger.error(
|
|
||||||
"/etc/default/locale could not be found! Please \
|
|
||||||
run 'sudo update-locale' from command-line."
|
|
||||||
)
|
|
||||||
sys.exit(1)
|
|
||||||
else:
|
|
||||||
new_locale = default_locale
|
|
||||||
|
|
||||||
logger.info(
|
|
||||||
"New locale set to: %s", locale.setlocale(locale.LC_ALL, new_locale)
|
|
||||||
)
|
|
||||||
|
|
||||||
reload(sys)
|
|
||||||
sys.setdefaultencoding("UTF-8")
|
|
||||||
current_locale_encoding = locale.getlocale()[1].lower()
|
|
||||||
logger.debug("sys default encoding %s", sys.getdefaultencoding())
|
|
||||||
logger.debug("After %s", locale.nl_langinfo(locale.CODESET))
|
|
||||||
|
|
||||||
if current_locale_encoding not in ["utf-8", "utf8"]:
|
|
||||||
logger.error(
|
|
||||||
"Need a UTF-8 locale. Currently '%s'. Exiting..." % current_locale_encoding
|
|
||||||
)
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
|
|
||||||
configure_locale()
|
|
||||||
|
|
||||||
# loading config file
|
# loading config file
|
||||||
try:
|
try:
|
||||||
config = ConfigObj("/etc/airtime/airtime.conf")
|
config = ConfigObj("/etc/airtime/airtime.conf")
|
||||||
|
@ -207,11 +155,11 @@ def liquidsoap_get_info(telnet_lock, host, port, logger):
|
||||||
telnet_lock.acquire()
|
telnet_lock.acquire()
|
||||||
tn = telnetlib.Telnet(host, port)
|
tn = telnetlib.Telnet(host, port)
|
||||||
msg = "version\n"
|
msg = "version\n"
|
||||||
tn.write(msg)
|
tn.write(msg.encode("utf-8"))
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode("utf-8"))
|
||||||
response = tn.read_all()
|
response = tn.read_all().decode("utf-8")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(str(e))
|
logger.error(e)
|
||||||
return None
|
return None
|
||||||
finally:
|
finally:
|
||||||
telnet_lock.release()
|
telnet_lock.release()
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
import urllib2
|
import urllib.request, urllib.error, urllib.parse
|
||||||
import defusedxml.minidom
|
import defusedxml.minidom
|
||||||
import base64
|
import base64
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
@ -44,13 +44,13 @@ class ListenerStat(Thread):
|
||||||
user_agent = "Mozilla/5.0 (Linux; rv:22.0) Gecko/20130405 Firefox/22.0"
|
user_agent = "Mozilla/5.0 (Linux; rv:22.0) Gecko/20130405 Firefox/22.0"
|
||||||
header["User-Agent"] = user_agent
|
header["User-Agent"] = user_agent
|
||||||
|
|
||||||
req = urllib2.Request(
|
req = urllib.request.Request(
|
||||||
#assuming that the icecast stats path is /admin/stats.xml
|
#assuming that the icecast stats path is /admin/stats.xml
|
||||||
#need to fix this
|
#need to fix this
|
||||||
url=url,
|
url=url,
|
||||||
headers=header)
|
headers=header)
|
||||||
|
|
||||||
f = urllib2.urlopen(req, timeout=ListenerStat.HTTP_REQUEST_TIMEOUT)
|
f = urllib.request.urlopen(req, timeout=ListenerStat.HTTP_REQUEST_TIMEOUT)
|
||||||
document = f.read()
|
document = f.read()
|
||||||
|
|
||||||
return document
|
return document
|
||||||
|
@ -146,7 +146,7 @@ class ListenerStat(Thread):
|
||||||
|
|
||||||
if stats:
|
if stats:
|
||||||
self.push_stream_stats(stats)
|
self.push_stream_stats(stats)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error('Exception: %s', e)
|
self.logger.error('Exception: %s', e)
|
||||||
|
|
||||||
time.sleep(120)
|
time.sleep(120)
|
||||||
|
|
|
@ -1,10 +1,14 @@
|
||||||
import re
|
import re
|
||||||
|
from packaging.version import Version, parse
|
||||||
|
|
||||||
def version_cmp(version1, version2):
|
def version_cmp(version1, version2):
|
||||||
def normalize(v):
|
version1 = parse(version1)
|
||||||
return [int(x) for x in re.sub(r'(\.0+)*$','', v).split(".")]
|
version2 = parse(version2)
|
||||||
return cmp(normalize(version1), normalize(version2))
|
if version1 > version2:
|
||||||
|
return 1
|
||||||
|
if version1 == version2:
|
||||||
|
return 0
|
||||||
|
return -1
|
||||||
|
|
||||||
def date_interval_to_seconds(interval):
|
def date_interval_to_seconds(interval):
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -10,15 +10,14 @@ import copy
|
||||||
import subprocess
|
import subprocess
|
||||||
import signal
|
import signal
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import traceback
|
|
||||||
import pure
|
|
||||||
import mimetypes
|
import mimetypes
|
||||||
from Queue import Empty
|
from . import pure
|
||||||
|
from queue import Empty
|
||||||
from threading import Thread, Timer
|
from threading import Thread, Timer
|
||||||
from subprocess import Popen, PIPE
|
from subprocess import Popen, PIPE
|
||||||
|
|
||||||
from api_clients import api_client
|
from api_clients import api_client
|
||||||
from timeout import ls_timeout
|
from .timeout import ls_timeout
|
||||||
|
|
||||||
|
|
||||||
def keyboardInterruptHandler(signum, frame):
|
def keyboardInterruptHandler(signum, frame):
|
||||||
|
@ -65,7 +64,7 @@ class PypoFetch(Thread):
|
||||||
"""
|
"""
|
||||||
self.logger.debug("Cache dir does not exist. Creating...")
|
self.logger.debug("Cache dir does not exist. Creating...")
|
||||||
os.makedirs(dir)
|
os.makedirs(dir)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
self.schedule_data = []
|
self.schedule_data = []
|
||||||
|
@ -79,6 +78,10 @@ class PypoFetch(Thread):
|
||||||
try:
|
try:
|
||||||
self.logger.info("Received event from Pypo Message Handler: %s" % message)
|
self.logger.info("Received event from Pypo Message Handler: %s" % message)
|
||||||
|
|
||||||
|
try:
|
||||||
|
message = message.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
m = json.loads(message)
|
m = json.loads(message)
|
||||||
command = m['event_type']
|
command = m['event_type']
|
||||||
self.logger.info("Handling command: " + command)
|
self.logger.info("Handling command: " + command)
|
||||||
|
@ -120,11 +123,8 @@ class PypoFetch(Thread):
|
||||||
if self.listener_timeout < 0:
|
if self.listener_timeout < 0:
|
||||||
self.listener_timeout = 0
|
self.listener_timeout = 0
|
||||||
self.logger.info("New timeout: %s" % self.listener_timeout)
|
self.logger.info("New timeout: %s" % self.listener_timeout)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
self.logger.exception("Exception in handling Message Handler message")
|
||||||
self.logger.error('Exception: %s', e)
|
|
||||||
self.logger.error("traceback: %s", top)
|
|
||||||
self.logger.error("Exception in handling Message Handler message: %s", e)
|
|
||||||
|
|
||||||
|
|
||||||
def switch_source_temp(self, sourcename, status):
|
def switch_source_temp(self, sourcename, status):
|
||||||
|
@ -151,13 +151,12 @@ class PypoFetch(Thread):
|
||||||
self.logger.debug('Getting information needed on bootstrap from Airtime')
|
self.logger.debug('Getting information needed on bootstrap from Airtime')
|
||||||
try:
|
try:
|
||||||
info = self.api_client.get_bootstrap_info()
|
info = self.api_client.get_bootstrap_info()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error('Unable to get bootstrap info.. Exiting pypo...')
|
self.logger.exception('Unable to get bootstrap info.. Exiting pypo...')
|
||||||
self.logger.error(str(e))
|
|
||||||
|
|
||||||
self.logger.debug('info:%s', info)
|
self.logger.debug('info:%s', info)
|
||||||
commands = []
|
commands = []
|
||||||
for k, v in info['switch_status'].iteritems():
|
for k, v in info['switch_status'].items():
|
||||||
commands.append(self.switch_source_temp(k, v))
|
commands.append(self.switch_source_temp(k, v))
|
||||||
|
|
||||||
stream_format = info['stream_label']
|
stream_format = info['stream_label']
|
||||||
|
@ -190,16 +189,16 @@ class PypoFetch(Thread):
|
||||||
while True:
|
while True:
|
||||||
try:
|
try:
|
||||||
tn = telnetlib.Telnet(self.config['ls_host'], self.config['ls_port'])
|
tn = telnetlib.Telnet(self.config['ls_host'], self.config['ls_port'])
|
||||||
tn.write("exit\n")
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all()
|
||||||
self.logger.info("Liquidsoap is up and running")
|
self.logger.info("Liquidsoap is up and running")
|
||||||
break
|
break
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
#sleep 0.5 seconds and try again
|
#sleep 0.5 seconds and try again
|
||||||
time.sleep(0.5)
|
time.sleep(0.5)
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(e)
|
self.logger.exception(e)
|
||||||
finally:
|
finally:
|
||||||
if self.telnet_lock.locked():
|
if self.telnet_lock.locked():
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
@ -226,19 +225,19 @@ class PypoFetch(Thread):
|
||||||
# we are manually adjusting the bootup time variable so the status msg will get
|
# we are manually adjusting the bootup time variable so the status msg will get
|
||||||
# updated.
|
# updated.
|
||||||
current_time = time.time()
|
current_time = time.time()
|
||||||
boot_up_time_command = "vars.bootup_time " + str(current_time) + "\n"
|
boot_up_time_command = ("vars.bootup_time " + str(current_time) + "\n").encode('utf-8')
|
||||||
self.logger.info(boot_up_time_command)
|
self.logger.info(boot_up_time_command)
|
||||||
tn.write(boot_up_time_command)
|
tn.write(boot_up_time_command)
|
||||||
|
|
||||||
connection_status = "streams.connection_status\n"
|
connection_status = ("streams.connection_status\n").encode('utf-8')
|
||||||
self.logger.info(connection_status)
|
self.logger.info(connection_status)
|
||||||
tn.write(connection_status)
|
tn.write(connection_status)
|
||||||
|
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
|
|
||||||
output = tn.read_all()
|
output = tn.read_all()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -269,10 +268,10 @@ class PypoFetch(Thread):
|
||||||
command = ('vars.stream_metadata_type %s\n' % stream_format).encode('utf-8')
|
command = ('vars.stream_metadata_type %s\n' % stream_format).encode('utf-8')
|
||||||
self.logger.info(command)
|
self.logger.info(command)
|
||||||
tn.write(command)
|
tn.write(command)
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Exception %s", e)
|
self.logger.exception(e)
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -286,10 +285,10 @@ class PypoFetch(Thread):
|
||||||
command = ('vars.default_dj_fade %s\n' % fade).encode('utf-8')
|
command = ('vars.default_dj_fade %s\n' % fade).encode('utf-8')
|
||||||
self.logger.info(command)
|
self.logger.info(command)
|
||||||
tn.write(command)
|
tn.write(command)
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Exception %s", e)
|
self.logger.exception(e)
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -304,14 +303,14 @@ class PypoFetch(Thread):
|
||||||
command = ('vars.station_name %s\n' % station_name).encode('utf-8')
|
command = ('vars.station_name %s\n' % station_name).encode('utf-8')
|
||||||
self.logger.info(command)
|
self.logger.info(command)
|
||||||
tn.write(command)
|
tn.write(command)
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.exception(e)
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Exception %s", e)
|
self.logger.exception(e)
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Process the schedule
|
Process the schedule
|
||||||
|
@ -336,7 +335,7 @@ class PypoFetch(Thread):
|
||||||
download_dir = self.cache_dir
|
download_dir = self.cache_dir
|
||||||
try:
|
try:
|
||||||
os.makedirs(download_dir)
|
os.makedirs(download_dir)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
media_copy = {}
|
media_copy = {}
|
||||||
|
@ -344,20 +343,21 @@ class PypoFetch(Thread):
|
||||||
media_item = media[key]
|
media_item = media[key]
|
||||||
if (media_item['type'] == 'file'):
|
if (media_item['type'] == 'file'):
|
||||||
fileExt = self.sanity_check_media_item(media_item)
|
fileExt = self.sanity_check_media_item(media_item)
|
||||||
dst = os.path.join(download_dir, unicode(media_item['id']) + unicode(fileExt))
|
dst = os.path.join(download_dir, "{}{}".format(media_item['id'], fileExt))
|
||||||
media_item['dst'] = dst
|
media_item['dst'] = dst
|
||||||
media_item['file_ready'] = False
|
media_item['file_ready'] = False
|
||||||
media_filtered[key] = media_item
|
media_filtered[key] = media_item
|
||||||
|
|
||||||
media_item['start'] = datetime.strptime(media_item['start'],
|
media_item['start'] = datetime.strptime(media_item['start'],
|
||||||
"%Y-%m-%d-%H-%M-%S")
|
"%Y-%m-%d-%H-%M-%S")
|
||||||
media_item['end'] = datetime.strptime(media_item['end'],
|
media_item['end'] = datetime.strptime(media_item['end'],
|
||||||
"%Y-%m-%d-%H-%M-%S")
|
"%Y-%m-%d-%H-%M-%S")
|
||||||
media_copy[key] = media_item
|
media_copy[key] = media_item
|
||||||
|
|
||||||
|
|
||||||
self.media_prepare_queue.put(copy.copy(media_filtered))
|
self.media_prepare_queue.put(copy.copy(media_filtered))
|
||||||
except Exception, e: self.logger.error("%s", e)
|
except Exception as e:
|
||||||
|
self.logger.exception(e)
|
||||||
|
|
||||||
# Send the data to pypo-push
|
# Send the data to pypo-push
|
||||||
self.logger.debug("Pushing to pypo-push")
|
self.logger.debug("Pushing to pypo-push")
|
||||||
|
@ -365,8 +365,10 @@ class PypoFetch(Thread):
|
||||||
|
|
||||||
|
|
||||||
# cleanup
|
# cleanup
|
||||||
try: self.cache_cleanup(media)
|
try:
|
||||||
except Exception, e: self.logger.error("%s", e)
|
self.cache_cleanup(media)
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.exception(e)
|
||||||
|
|
||||||
#do basic validation of file parameters. Useful for debugging
|
#do basic validation of file parameters. Useful for debugging
|
||||||
#purposes
|
#purposes
|
||||||
|
@ -408,7 +410,9 @@ class PypoFetch(Thread):
|
||||||
for mkey in media:
|
for mkey in media:
|
||||||
media_item = media[mkey]
|
media_item = media[mkey]
|
||||||
if media_item['type'] == 'file':
|
if media_item['type'] == 'file':
|
||||||
scheduled_file_set.add(unicode(media_item["id"]) + unicode(media_item["file_ext"]))
|
if "file_ext" not in media_item.keys():
|
||||||
|
media_item["file_ext"] = mimetypes.guess_extension(media_item['metadata']['mime'], strict=False)
|
||||||
|
scheduled_file_set.add("{}{}".format(media_item["id"], media_item["file_ext"]))
|
||||||
|
|
||||||
expired_files = cached_file_set - scheduled_file_set
|
expired_files = cached_file_set - scheduled_file_set
|
||||||
|
|
||||||
|
@ -426,9 +430,8 @@ class PypoFetch(Thread):
|
||||||
self.logger.info("File '%s' removed" % path)
|
self.logger.info("File '%s' removed" % path)
|
||||||
else:
|
else:
|
||||||
self.logger.info("File '%s' not removed. Still busy!" % path)
|
self.logger.info("File '%s' not removed. Still busy!" % path)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Problem removing file '%s'" % f)
|
self.logger.exception("Problem removing file '%s'" % f)
|
||||||
self.logger.error(traceback.format_exc())
|
|
||||||
|
|
||||||
def manual_schedule_fetch(self):
|
def manual_schedule_fetch(self):
|
||||||
success, self.schedule_data = self.api_client.get_schedule()
|
success, self.schedule_data = self.api_client.get_schedule()
|
||||||
|
@ -498,18 +501,13 @@ class PypoFetch(Thread):
|
||||||
self.logger.info("Queue timeout. Fetching schedule manually")
|
self.logger.info("Queue timeout. Fetching schedule manually")
|
||||||
manual_fetch_needed = True
|
manual_fetch_needed = True
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
self.logger.exception(e)
|
||||||
self.logger.error('Exception: %s', e)
|
|
||||||
self.logger.error("traceback: %s", top)
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if manual_fetch_needed:
|
if manual_fetch_needed:
|
||||||
self.persistent_manual_schedule_fetch(max_attempts=5)
|
self.persistent_manual_schedule_fetch(max_attempts=5)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
self.logger.exception('Failed to manually fetch the schedule.')
|
||||||
self.logger.error('Failed to manually fetch the schedule.')
|
|
||||||
self.logger.error('Exception: %s', e)
|
|
||||||
self.logger.error("traceback: %s", top)
|
|
||||||
|
|
||||||
loops += 1
|
loops += 1
|
||||||
|
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
from Queue import Empty
|
from queue import Empty
|
||||||
from ConfigParser import NoOptionError
|
from configparser import NoOptionError
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import shutil
|
import shutil
|
||||||
|
@ -12,7 +12,7 @@ import os
|
||||||
import sys
|
import sys
|
||||||
import stat
|
import stat
|
||||||
import requests
|
import requests
|
||||||
import ConfigParser
|
import configparser
|
||||||
import json
|
import json
|
||||||
import hashlib
|
import hashlib
|
||||||
from requests.exceptions import ConnectionError, HTTPError, Timeout
|
from requests.exceptions import ConnectionError, HTTPError, Timeout
|
||||||
|
@ -44,7 +44,7 @@ class PypoFile(Thread):
|
||||||
dst_exists = True
|
dst_exists = True
|
||||||
try:
|
try:
|
||||||
dst_size = os.path.getsize(dst)
|
dst_size = os.path.getsize(dst)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
dst_exists = False
|
dst_exists = False
|
||||||
|
|
||||||
do_copy = False
|
do_copy = False
|
||||||
|
@ -69,11 +69,11 @@ class PypoFile(Thread):
|
||||||
baseurl = self._config.get(CONFIG_SECTION, 'base_url')
|
baseurl = self._config.get(CONFIG_SECTION, 'base_url')
|
||||||
try:
|
try:
|
||||||
port = self._config.get(CONFIG_SECTION, 'base_port')
|
port = self._config.get(CONFIG_SECTION, 'base_port')
|
||||||
except NoOptionError, e:
|
except NoOptionError as e:
|
||||||
port = 80
|
port = 80
|
||||||
try:
|
try:
|
||||||
protocol = self._config.get(CONFIG_SECTION, 'protocol')
|
protocol = self._config.get(CONFIG_SECTION, 'protocol')
|
||||||
except NoOptionError, e:
|
except NoOptionError as e:
|
||||||
protocol = str(("http", "https")[int(port) == 443])
|
protocol = str(("http", "https")[int(port) == 443])
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
@ -103,7 +103,7 @@ class PypoFile(Thread):
|
||||||
media_item["filesize"] = file_size
|
media_item["filesize"] = file_size
|
||||||
|
|
||||||
media_item['file_ready'] = True
|
media_item['file_ready'] = True
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Could not copy from %s to %s" % (src, dst))
|
self.logger.error("Could not copy from %s to %s" % (src, dst))
|
||||||
self.logger.error(e)
|
self.logger.error(e)
|
||||||
|
|
||||||
|
@ -172,7 +172,7 @@ class PypoFile(Thread):
|
||||||
|
|
||||||
def read_config_file(self, config_path):
|
def read_config_file(self, config_path):
|
||||||
"""Parse the application's config file located at config_path."""
|
"""Parse the application's config file located at config_path."""
|
||||||
config = ConfigParser.SafeConfigParser(allow_no_value=True)
|
config = configparser.SafeConfigParser(allow_no_value=True)
|
||||||
try:
|
try:
|
||||||
config.readfp(open(config_path))
|
config.readfp(open(config_path))
|
||||||
except IOError as e:
|
except IOError as e:
|
||||||
|
@ -202,14 +202,14 @@ class PypoFile(Thread):
|
||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
self.media = self.media_queue.get_nowait()
|
self.media = self.media_queue.get_nowait()
|
||||||
except Empty, e:
|
except Empty as e:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
media_item = self.get_highest_priority_media_item(self.media)
|
media_item = self.get_highest_priority_media_item(self.media)
|
||||||
if media_item is not None:
|
if media_item is not None:
|
||||||
self.copy_file(media_item)
|
self.copy_file(media_item)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
import traceback
|
import traceback
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
@ -221,7 +221,7 @@ class PypoFile(Thread):
|
||||||
Entry point of the thread
|
Entry point of the thread
|
||||||
"""
|
"""
|
||||||
try: self.main()
|
try: self.main()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error('PypoFile Exception: %s', top)
|
self.logger.error('PypoFile Exception: %s', top)
|
||||||
time.sleep(5)
|
time.sleep(5)
|
||||||
|
|
|
@ -7,7 +7,7 @@ import sys
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
|
||||||
from Queue import Empty
|
from queue import Empty
|
||||||
|
|
||||||
import signal
|
import signal
|
||||||
def keyboardInterruptHandler(signum, frame):
|
def keyboardInterruptHandler(signum, frame):
|
||||||
|
@ -38,7 +38,7 @@ class PypoLiqQueue(Thread):
|
||||||
time_until_next_play)
|
time_until_next_play)
|
||||||
media_schedule = self.queue.get(block=True, \
|
media_schedule = self.queue.get(block=True, \
|
||||||
timeout=time_until_next_play)
|
timeout=time_until_next_play)
|
||||||
except Empty, e:
|
except Empty as e:
|
||||||
#Time to push a scheduled item.
|
#Time to push a scheduled item.
|
||||||
media_item = schedule_deque.popleft()
|
media_item = schedule_deque.popleft()
|
||||||
self.pypo_liquidsoap.play(media_item)
|
self.pypo_liquidsoap.play(media_item)
|
||||||
|
@ -82,7 +82,7 @@ class PypoLiqQueue(Thread):
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
try: self.main()
|
try: self.main()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error('PypoLiqQueue Exception: %s', traceback.format_exc())
|
self.logger.error('PypoLiqQueue Exception: %s', traceback.format_exc())
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,10 @@
|
||||||
from pypofetch import PypoFetch
|
from .pypofetch import PypoFetch
|
||||||
from telnetliquidsoap import TelnetLiquidsoap
|
from .telnetliquidsoap import TelnetLiquidsoap
|
||||||
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from datetime import timedelta
|
from datetime import timedelta
|
||||||
|
|
||||||
import eventtypes
|
from . import eventtypes
|
||||||
import time
|
import time
|
||||||
|
|
||||||
class PypoLiquidsoap():
|
class PypoLiquidsoap():
|
||||||
|
@ -22,7 +22,7 @@ class PypoLiquidsoap():
|
||||||
logger,\
|
logger,\
|
||||||
host,\
|
host,\
|
||||||
port,\
|
port,\
|
||||||
self.liq_queue_tracker.keys())
|
list(self.liq_queue_tracker.keys()))
|
||||||
|
|
||||||
def get_telnet_dispatcher(self):
|
def get_telnet_dispatcher(self):
|
||||||
return self.telnet_liquidsoap
|
return self.telnet_liquidsoap
|
||||||
|
@ -97,36 +97,37 @@ class PypoLiquidsoap():
|
||||||
|
|
||||||
|
|
||||||
def verify_correct_present_media(self, scheduled_now):
|
def verify_correct_present_media(self, scheduled_now):
|
||||||
#verify whether Liquidsoap is currently playing the correct files.
|
"""
|
||||||
#if we find an item that Liquidsoap is not playing, then push it
|
verify whether Liquidsoap is currently playing the correct files.
|
||||||
#into one of Liquidsoap's queues. If Liquidsoap is already playing
|
if we find an item that Liquidsoap is not playing, then push it
|
||||||
#it do nothing. If Liquidsoap is playing a track that isn't in
|
into one of Liquidsoap's queues. If Liquidsoap is already playing
|
||||||
#currently_playing then stop it.
|
it do nothing. If Liquidsoap is playing a track that isn't in
|
||||||
|
currently_playing then stop it.
|
||||||
|
|
||||||
#Check for Liquidsoap media we should source.skip
|
Check for Liquidsoap media we should source.skip
|
||||||
#get liquidsoap items for each queue. Since each queue can only have one
|
get liquidsoap items for each queue. Since each queue can only have one
|
||||||
#item, we should have a max of 8 items.
|
item, we should have a max of 8 items.
|
||||||
|
|
||||||
#2013-03-21-22-56-00_0: {
|
2013-03-21-22-56-00_0: {
|
||||||
#id: 1,
|
id: 1,
|
||||||
#type: "stream_output_start",
|
type: "stream_output_start",
|
||||||
#row_id: 41,
|
row_id: 41,
|
||||||
#uri: "http://stream2.radioblackout.org:80/blackout.ogg",
|
uri: "http://stream2.radioblackout.org:80/blackout.ogg",
|
||||||
#start: "2013-03-21-22-56-00",
|
start: "2013-03-21-22-56-00",
|
||||||
#end: "2013-03-21-23-26-00",
|
end: "2013-03-21-23-26-00",
|
||||||
#show_name: "Untitled Show",
|
show_name: "Untitled Show",
|
||||||
#independent_event: true
|
independent_event: true
|
||||||
#},
|
},
|
||||||
|
"""
|
||||||
|
|
||||||
try:
|
try:
|
||||||
scheduled_now_files = \
|
scheduled_now_files = \
|
||||||
filter(lambda x: x["type"] == eventtypes.FILE, scheduled_now)
|
[x for x in scheduled_now if x["type"] == eventtypes.FILE]
|
||||||
|
|
||||||
scheduled_now_webstream = \
|
scheduled_now_webstream = \
|
||||||
filter(lambda x: x["type"] == eventtypes.STREAM_OUTPUT_START, \
|
[x for x in scheduled_now if x["type"] == eventtypes.STREAM_OUTPUT_START]
|
||||||
scheduled_now)
|
|
||||||
|
|
||||||
schedule_ids = set(map(lambda x: x["row_id"], scheduled_now_files))
|
schedule_ids = set([x["row_id"] for x in scheduled_now_files])
|
||||||
|
|
||||||
row_id_map = {}
|
row_id_map = {}
|
||||||
liq_queue_ids = set()
|
liq_queue_ids = set()
|
||||||
|
@ -156,7 +157,6 @@ class PypoLiquidsoap():
|
||||||
to_be_removed.add(i["row_id"])
|
to_be_removed.add(i["row_id"])
|
||||||
to_be_added.add(i["row_id"])
|
to_be_added.add(i["row_id"])
|
||||||
|
|
||||||
|
|
||||||
to_be_removed.update(liq_queue_ids - schedule_ids)
|
to_be_removed.update(liq_queue_ids - schedule_ids)
|
||||||
to_be_added.update(schedule_ids - liq_queue_ids)
|
to_be_added.update(schedule_ids - liq_queue_ids)
|
||||||
|
|
||||||
|
|
|
@ -53,7 +53,7 @@ class PypoMessageHandler(Thread):
|
||||||
heartbeat = 5) as connection:
|
heartbeat = 5) as connection:
|
||||||
rabbit = RabbitConsumer(connection, [schedule_queue], self)
|
rabbit = RabbitConsumer(connection, [schedule_queue], self)
|
||||||
rabbit.run()
|
rabbit.run()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(e)
|
self.logger.error(e)
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
@ -64,6 +64,10 @@ class PypoMessageHandler(Thread):
|
||||||
try:
|
try:
|
||||||
self.logger.info("Received event from RabbitMQ: %s" % message)
|
self.logger.info("Received event from RabbitMQ: %s" % message)
|
||||||
|
|
||||||
|
try:
|
||||||
|
message = message.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
m = json.loads(message)
|
m = json.loads(message)
|
||||||
command = m['event_type']
|
command = m['event_type']
|
||||||
self.logger.info("Handling command: " + command)
|
self.logger.info("Handling command: " + command)
|
||||||
|
@ -98,13 +102,13 @@ class PypoMessageHandler(Thread):
|
||||||
self.recorder_queue.put(message)
|
self.recorder_queue.put(message)
|
||||||
else:
|
else:
|
||||||
self.logger.info("Unknown command: %s" % command)
|
self.logger.info("Unknown command: %s" % command)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error("Exception in handling RabbitMQ message: %s", e)
|
self.logger.error("Exception in handling RabbitMQ message: %s", e)
|
||||||
|
|
||||||
def main(self):
|
def main(self):
|
||||||
try:
|
try:
|
||||||
self.init_rabbit_mq()
|
self.init_rabbit_mq()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error('Exception: %s', e)
|
self.logger.error('Exception: %s', e)
|
||||||
self.logger.error("traceback: %s", traceback.format_exc())
|
self.logger.error("traceback: %s", traceback.format_exc())
|
||||||
self.logger.error("Error connecting to RabbitMQ Server. Trying again in few seconds")
|
self.logger.error("Error connecting to RabbitMQ Server. Trying again in few seconds")
|
||||||
|
|
|
@ -13,15 +13,15 @@ import math
|
||||||
import traceback
|
import traceback
|
||||||
import os
|
import os
|
||||||
|
|
||||||
from pypofetch import PypoFetch
|
from .pypofetch import PypoFetch
|
||||||
from pypoliqqueue import PypoLiqQueue
|
from .pypoliqqueue import PypoLiqQueue
|
||||||
|
|
||||||
from Queue import Empty, Queue
|
from queue import Empty, Queue
|
||||||
|
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
|
|
||||||
from api_clients import api_client
|
from api_clients import api_client
|
||||||
from timeout import ls_timeout
|
from .timeout import ls_timeout
|
||||||
|
|
||||||
logging.captureWarnings(True)
|
logging.captureWarnings(True)
|
||||||
|
|
||||||
|
@ -67,7 +67,7 @@ class PypoPush(Thread):
|
||||||
while True:
|
while True:
|
||||||
try:
|
try:
|
||||||
media_schedule = self.queue.get(block=True)
|
media_schedule = self.queue.get(block=True)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
raise
|
raise
|
||||||
else:
|
else:
|
||||||
|
@ -138,7 +138,7 @@ class PypoPush(Thread):
|
||||||
tn.write("exit\n")
|
tn.write("exit\n")
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all())
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
@ -146,7 +146,7 @@ class PypoPush(Thread):
|
||||||
def run(self):
|
def run(self):
|
||||||
while True:
|
while True:
|
||||||
try: self.main()
|
try: self.main()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error('Pypo Push Exception: %s', top)
|
self.logger.error('Pypo Push Exception: %s', top)
|
||||||
time.sleep(5)
|
time.sleep(5)
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import json
|
import json
|
||||||
import time
|
import time
|
||||||
|
@ -14,9 +15,6 @@ import re
|
||||||
|
|
||||||
from configobj import ConfigObj
|
from configobj import ConfigObj
|
||||||
|
|
||||||
from poster.encode import multipart_encode
|
|
||||||
from poster.streaminghttp import register_openers
|
|
||||||
|
|
||||||
from subprocess import Popen
|
from subprocess import Popen
|
||||||
from subprocess import PIPE
|
from subprocess import PIPE
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
|
@ -35,8 +33,8 @@ def api_client(logger):
|
||||||
# loading config file
|
# loading config file
|
||||||
try:
|
try:
|
||||||
config = ConfigObj('/etc/airtime/airtime.conf')
|
config = ConfigObj('/etc/airtime/airtime.conf')
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
print ('Error loading config file: %s', e)
|
print("Error loading config file: {}".format(e))
|
||||||
sys.exit()
|
sys.exit()
|
||||||
|
|
||||||
# TODO : add docstrings everywhere in this module
|
# TODO : add docstrings everywhere in this module
|
||||||
|
@ -94,7 +92,7 @@ class ShowRecorder(Thread):
|
||||||
|
|
||||||
self.logger.info("starting record")
|
self.logger.info("starting record")
|
||||||
self.logger.info("command " + command)
|
self.logger.info("command " + command)
|
||||||
|
|
||||||
self.p = Popen(args,stdout=PIPE,stderr=PIPE)
|
self.p = Popen(args,stdout=PIPE,stderr=PIPE)
|
||||||
|
|
||||||
#blocks at the following line until the child process
|
#blocks at the following line until the child process
|
||||||
|
@ -126,11 +124,8 @@ class ShowRecorder(Thread):
|
||||||
|
|
||||||
filename = os.path.split(filepath)[1]
|
filename = os.path.split(filepath)[1]
|
||||||
|
|
||||||
# Register the streaming http handlers with urllib2
|
|
||||||
register_openers()
|
|
||||||
|
|
||||||
# files is what requests actually expects
|
# files is what requests actually expects
|
||||||
files = {'file': open(filepath, "rb"), 'name': filename, 'show_instance': str(self.show_instance)}
|
files = {'file': open(filepath, "rb"), 'name': filename, 'show_instance': self.show_instance}
|
||||||
|
|
||||||
self.api_client.upload_recorded_show(files, self.show_instance)
|
self.api_client.upload_recorded_show(files, self.show_instance)
|
||||||
|
|
||||||
|
@ -152,10 +147,10 @@ class ShowRecorder(Thread):
|
||||||
recorded_file['title'] = "%s-%s-%s" % (self.show_name,
|
recorded_file['title'] = "%s-%s-%s" % (self.show_name,
|
||||||
full_date, full_time)
|
full_date, full_time)
|
||||||
#You cannot pass ints into the metadata of a file. Even tracknumber needs to be a string
|
#You cannot pass ints into the metadata of a file. Even tracknumber needs to be a string
|
||||||
recorded_file['tracknumber'] = unicode(self.show_instance)
|
recorded_file['tracknumber'] = self.show_instance
|
||||||
recorded_file.save()
|
recorded_file.save()
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error('Exception: %s', e)
|
self.logger.error('Exception: %s', e)
|
||||||
self.logger.error("traceback: %s", top)
|
self.logger.error("traceback: %s", top)
|
||||||
|
@ -172,7 +167,7 @@ class ShowRecorder(Thread):
|
||||||
|
|
||||||
self.upload_file(filepath)
|
self.upload_file(filepath)
|
||||||
os.remove(filepath)
|
os.remove(filepath)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(e)
|
self.logger.error(e)
|
||||||
else:
|
else:
|
||||||
self.logger.info("problem recording show")
|
self.logger.info("problem recording show")
|
||||||
|
@ -195,14 +190,18 @@ class Recorder(Thread):
|
||||||
try:
|
try:
|
||||||
self.api_client.register_component('show-recorder')
|
self.api_client.register_component('show-recorder')
|
||||||
success = True
|
success = True
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
time.sleep(10)
|
time.sleep(10)
|
||||||
|
|
||||||
def handle_message(self):
|
def handle_message(self):
|
||||||
if not self.queue.empty():
|
if not self.queue.empty():
|
||||||
message = self.queue.get()
|
message = self.queue.get()
|
||||||
msg = json.loads(message)
|
try:
|
||||||
|
message = message.decode()
|
||||||
|
except (UnicodeDecodeError, AttributeError):
|
||||||
|
pass
|
||||||
|
msg = json.loads(message)
|
||||||
command = msg["event_type"]
|
command = msg["event_type"]
|
||||||
self.logger.info("Received msg from Pypo Message Handler: %s", msg)
|
self.logger.info("Received msg from Pypo Message Handler: %s", msg)
|
||||||
if command == 'cancel_recording':
|
if command == 'cancel_recording':
|
||||||
|
@ -220,12 +219,12 @@ class Recorder(Thread):
|
||||||
temp_shows_to_record = {}
|
temp_shows_to_record = {}
|
||||||
shows = m['shows']
|
shows = m['shows']
|
||||||
for show in shows:
|
for show in shows:
|
||||||
show_starts = getDateTimeObj(show[u'starts'])
|
show_starts = getDateTimeObj(show['starts'])
|
||||||
show_end = getDateTimeObj(show[u'ends'])
|
show_end = getDateTimeObj(show['ends'])
|
||||||
time_delta = show_end - show_starts
|
time_delta = show_end - show_starts
|
||||||
|
|
||||||
temp_shows_to_record[show[u'starts']] = [time_delta,
|
temp_shows_to_record[show['starts']] = [time_delta,
|
||||||
show[u'instance_id'], show[u'name'], m['server_timezone']]
|
show['instance_id'], show['name'], m['server_timezone']]
|
||||||
self.shows_to_record = temp_shows_to_record
|
self.shows_to_record = temp_shows_to_record
|
||||||
|
|
||||||
def get_time_till_next_show(self):
|
def get_time_till_next_show(self):
|
||||||
|
@ -245,11 +244,11 @@ class Recorder(Thread):
|
||||||
self.logger.debug("Next show %s", next_show)
|
self.logger.debug("Next show %s", next_show)
|
||||||
self.logger.debug("Now %s", tnow)
|
self.logger.debug("Now %s", tnow)
|
||||||
return out
|
return out
|
||||||
|
|
||||||
def cancel_recording(self):
|
def cancel_recording(self):
|
||||||
self.sr.cancel_recording()
|
self.sr.cancel_recording()
|
||||||
self.sr = None
|
self.sr = None
|
||||||
|
|
||||||
def currently_recording(self):
|
def currently_recording(self):
|
||||||
if self.sr is not None and self.sr.is_recording():
|
if self.sr is not None and self.sr.is_recording():
|
||||||
return True
|
return True
|
||||||
|
@ -277,27 +276,27 @@ class Recorder(Thread):
|
||||||
start_time_formatted = '%(year)d-%(month)02d-%(day)02d %(hour)02d:%(min)02d:%(sec)02d' % \
|
start_time_formatted = '%(year)d-%(month)02d-%(day)02d %(hour)02d:%(min)02d:%(sec)02d' % \
|
||||||
{'year': start_time_on_server.year, 'month': start_time_on_server.month, 'day': start_time_on_server.day, \
|
{'year': start_time_on_server.year, 'month': start_time_on_server.month, 'day': start_time_on_server.day, \
|
||||||
'hour': start_time_on_server.hour, 'min': start_time_on_server.minute, 'sec': start_time_on_server.second}
|
'hour': start_time_on_server.hour, 'min': start_time_on_server.minute, 'sec': start_time_on_server.second}
|
||||||
|
|
||||||
|
|
||||||
seconds_waiting = 0
|
seconds_waiting = 0
|
||||||
|
|
||||||
#avoiding CC-5299
|
#avoiding CC-5299
|
||||||
while(True):
|
while(True):
|
||||||
if self.currently_recording():
|
if self.currently_recording():
|
||||||
self.logger.info("Previous record not finished, sleeping 100ms")
|
self.logger.info("Previous record not finished, sleeping 100ms")
|
||||||
seconds_waiting = seconds_waiting + 0.1
|
seconds_waiting = seconds_waiting + 0.1
|
||||||
time.sleep(0.1)
|
time.sleep(0.1)
|
||||||
else:
|
else:
|
||||||
show_length_seconds = show_length.seconds - seconds_waiting
|
show_length_seconds = show_length.seconds - seconds_waiting
|
||||||
|
|
||||||
self.sr = ShowRecorder(show_instance, show_name, show_length_seconds, start_time_formatted)
|
self.sr = ShowRecorder(show_instance, show_name, show_length_seconds, start_time_formatted)
|
||||||
self.sr.start()
|
self.sr.start()
|
||||||
break
|
break
|
||||||
|
|
||||||
#remove show from shows to record.
|
#remove show from shows to record.
|
||||||
del self.shows_to_record[start_time]
|
del self.shows_to_record[start_time]
|
||||||
#self.time_till_next_show = self.get_time_till_next_show()
|
#self.time_till_next_show = self.get_time_till_next_show()
|
||||||
except Exception, e :
|
except Exception as e :
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error('Exception: %s', e)
|
self.logger.error('Exception: %s', e)
|
||||||
self.logger.error("traceback: %s", top)
|
self.logger.error("traceback: %s", top)
|
||||||
|
@ -317,7 +316,7 @@ class Recorder(Thread):
|
||||||
if temp is not None:
|
if temp is not None:
|
||||||
self.process_recorder_schedule(temp)
|
self.process_recorder_schedule(temp)
|
||||||
self.logger.info("Bootstrap recorder schedule received: %s", temp)
|
self.logger.info("Bootstrap recorder schedule received: %s", temp)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error( traceback.format_exc() )
|
self.logger.error( traceback.format_exc() )
|
||||||
self.logger.error(e)
|
self.logger.error(e)
|
||||||
|
|
||||||
|
@ -337,16 +336,16 @@ class Recorder(Thread):
|
||||||
if temp is not None:
|
if temp is not None:
|
||||||
self.process_recorder_schedule(temp)
|
self.process_recorder_schedule(temp)
|
||||||
self.logger.info("updated recorder schedule received: %s", temp)
|
self.logger.info("updated recorder schedule received: %s", temp)
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error( traceback.format_exc() )
|
self.logger.error( traceback.format_exc() )
|
||||||
self.logger.error(e)
|
self.logger.error(e)
|
||||||
try: self.handle_message()
|
try: self.handle_message()
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error( traceback.format_exc() )
|
self.logger.error( traceback.format_exc() )
|
||||||
self.logger.error('Pypo Recorder Exception: %s', e)
|
self.logger.error('Pypo Recorder Exception: %s', e)
|
||||||
time.sleep(PUSH_INTERVAL)
|
time.sleep(PUSH_INTERVAL)
|
||||||
self.loops += 1
|
self.loops += 1
|
||||||
except Exception, e :
|
except Exception as e :
|
||||||
top = traceback.format_exc()
|
top = traceback.format_exc()
|
||||||
self.logger.error('Exception: %s', e)
|
self.logger.error('Exception: %s', e)
|
||||||
self.logger.error("traceback: %s", top)
|
self.logger.error("traceback: %s", top)
|
||||||
|
|
|
@ -1,5 +1,7 @@
|
||||||
|
|
||||||
import telnetlib
|
import telnetlib
|
||||||
from timeout import ls_timeout
|
from .timeout import ls_timeout
|
||||||
|
import traceback
|
||||||
|
|
||||||
def create_liquidsoap_annotation(media):
|
def create_liquidsoap_annotation(media):
|
||||||
# We need liq_start_next value in the annotate. That is the value that controls overlap duration of crossfade.
|
# We need liq_start_next value in the annotate. That is the value that controls overlap duration of crossfade.
|
||||||
|
@ -51,8 +53,8 @@ class TelnetLiquidsoap:
|
||||||
return True
|
return True
|
||||||
tn = self.__connect()
|
tn = self.__connect()
|
||||||
msg = '%s.queue\nexit\n' % queue_id
|
msg = '%s.queue\nexit\n' % queue_id
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
output = tn.read_all().splitlines()
|
output = tn.read_all().decode('utf-8').splitlines()
|
||||||
if len(output) == 3:
|
if len(output) == 3:
|
||||||
return len(output[0]) == 0
|
return len(output[0]) == 0
|
||||||
else:
|
else:
|
||||||
|
@ -67,10 +69,10 @@ class TelnetLiquidsoap:
|
||||||
for i in self.queues:
|
for i in self.queues:
|
||||||
msg = 'queues.%s_skip\n' % i
|
msg = 'queues.%s_skip\n' % i
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
except Exception:
|
except Exception:
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
|
@ -84,10 +86,10 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
msg = 'queues.%s_skip\n' % queue_id
|
msg = 'queues.%s_skip\n' % queue_id
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
except Exception:
|
except Exception:
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
|
@ -104,17 +106,17 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
tn = self.__connect()
|
tn = self.__connect()
|
||||||
annotation = create_liquidsoap_annotation(media_item)
|
annotation = create_liquidsoap_annotation(media_item)
|
||||||
msg = '%s.push %s\n' % (queue_id, annotation.encode('utf-8'))
|
msg = '%s.push %s\n' % (queue_id, annotation)
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
show_name = media_item['show_name']
|
show_name = media_item['show_name']
|
||||||
msg = 'vars.show_name %s\n' % show_name.encode('utf-8')
|
msg = 'vars.show_name %s\n' % show_name
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
except Exception:
|
except Exception:
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
|
@ -130,17 +132,18 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
msg = 'http.stop\n'
|
msg = 'http.stop\n'
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
msg = 'dynamic_source.id -1\n'
|
msg = 'dynamic_source.id -1\n'
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -153,13 +156,14 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
msg = 'dynamic_source.output_stop\n'
|
msg = 'dynamic_source.output_stop\n'
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
|
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -171,21 +175,22 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
#TODO: DO we need this?
|
#TODO: DO we need this?
|
||||||
msg = 'streams.scheduled_play_start\n'
|
msg = 'streams.scheduled_play_start\n'
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
msg = 'dynamic_source.output_start\n'
|
msg = 'dynamic_source.output_start\n'
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
|
|
||||||
self.current_prebuffering_stream_id = None
|
self.current_prebuffering_stream_id = None
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
@ls_timeout
|
@ls_timeout
|
||||||
def start_web_stream_buffer(self, media_item):
|
def start_web_stream_buffer(self, media_item):
|
||||||
try:
|
try:
|
||||||
|
@ -194,18 +199,19 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
msg = 'dynamic_source.id %s\n' % media_item['row_id']
|
msg = 'dynamic_source.id %s\n' % media_item['row_id']
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
msg = 'http.restart %s\n' % media_item['uri'].encode('latin-1')
|
msg = 'http.restart %s\n' % media_item['uri']
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
self.logger.debug(tn.read_all())
|
self.logger.debug(tn.read_all().decode('utf-8'))
|
||||||
|
|
||||||
self.current_prebuffering_stream_id = media_item['row_id']
|
self.current_prebuffering_stream_id = media_item['row_id']
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -217,15 +223,16 @@ class TelnetLiquidsoap:
|
||||||
|
|
||||||
msg = 'dynamic_source.get_id\n'
|
msg = 'dynamic_source.get_id\n'
|
||||||
self.logger.debug(msg)
|
self.logger.debug(msg)
|
||||||
tn.write(msg)
|
tn.write(msg.encode('utf-8'))
|
||||||
|
|
||||||
tn.write("exit\n")
|
tn.write("exit\n".encode('utf-8'))
|
||||||
stream_id = tn.read_all().splitlines()[0]
|
stream_id = tn.read_all().decode('utf-8').splitlines()[0]
|
||||||
self.logger.debug("stream_id: %s" % stream_id)
|
self.logger.debug("stream_id: %s" % stream_id)
|
||||||
|
|
||||||
return stream_id
|
return stream_id
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -242,10 +249,10 @@ class TelnetLiquidsoap:
|
||||||
self.telnet_lock.acquire()
|
self.telnet_lock.acquire()
|
||||||
tn = telnetlib.Telnet(self.ls_host, self.ls_port)
|
tn = telnetlib.Telnet(self.ls_host, self.ls_port)
|
||||||
self.logger.info(command)
|
self.logger.info(command)
|
||||||
tn.write(command)
|
tn.write(command.encode('utf-8'))
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all().decode('utf-8')
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(traceback.format_exc())
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
@ -258,12 +265,15 @@ class TelnetLiquidsoap:
|
||||||
tn = telnetlib.Telnet(self.ls_host, self.ls_port)
|
tn = telnetlib.Telnet(self.ls_host, self.ls_port)
|
||||||
for i in commands:
|
for i in commands:
|
||||||
self.logger.info(i)
|
self.logger.info(i)
|
||||||
|
if type(i) is str:
|
||||||
|
i = i.encode('utf-8')
|
||||||
tn.write(i)
|
tn.write(i)
|
||||||
|
|
||||||
tn.write('exit\n')
|
tn.write('exit\n'.encode('utf-8'))
|
||||||
tn.read_all()
|
tn.read_all().decode('utf-8')
|
||||||
except Exception, e:
|
except Exception as e:
|
||||||
self.logger.error(str(e))
|
self.logger.error(str(e))
|
||||||
|
self.logger.error(traceback.format_exc())
|
||||||
finally:
|
finally:
|
||||||
self.telnet_lock.release()
|
self.telnet_lock.release()
|
||||||
|
|
||||||
|
@ -302,7 +312,7 @@ class DummyTelnetLiquidsoap:
|
||||||
|
|
||||||
self.logger.info("Pushing %s to queue %s" % (media_item, queue_id))
|
self.logger.info("Pushing %s to queue %s" % (media_item, queue_id))
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
print "Time now: %s" % datetime.utcnow()
|
print("Time now: {:s}".format(datetime.utcnow()))
|
||||||
|
|
||||||
annotation = create_liquidsoap_annotation(media_item)
|
annotation = create_liquidsoap_annotation(media_item)
|
||||||
self.liquidsoap_mock_queues[queue_id].append(annotation)
|
self.liquidsoap_mock_queues[queue_id].append(annotation)
|
||||||
|
@ -318,7 +328,7 @@ class DummyTelnetLiquidsoap:
|
||||||
|
|
||||||
self.logger.info("Purging queue %s" % queue_id)
|
self.logger.info("Purging queue %s" % queue_id)
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
print "Time now: %s" % datetime.utcnow()
|
print("Time now: {:s}".format(datetime.utcnow()))
|
||||||
|
|
||||||
except Exception:
|
except Exception:
|
||||||
raise
|
raise
|
||||||
|
|
|
@ -1,8 +1,9 @@
|
||||||
from pypoliqqueue import PypoLiqQueue
|
|
||||||
from telnetliquidsoap import DummyTelnetLiquidsoap, TelnetLiquidsoap
|
from .pypoliqqueue import PypoLiqQueue
|
||||||
|
from .telnetliquidsoap import DummyTelnetLiquidsoap, TelnetLiquidsoap
|
||||||
|
|
||||||
|
|
||||||
from Queue import Queue
|
from queue import Queue
|
||||||
from threading import Lock
|
from threading import Lock
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
|
@ -45,7 +46,7 @@ plq.daemon = True
|
||||||
plq.start()
|
plq.start()
|
||||||
|
|
||||||
|
|
||||||
print "Time now: %s" % datetime.utcnow()
|
print("Time now: {:s}".format(datetime.utcnow()))
|
||||||
|
|
||||||
media_schedule = {}
|
media_schedule = {}
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
import threading
|
import threading
|
||||||
import pypofetch
|
from . import pypofetch
|
||||||
|
|
||||||
def __timeout(func, timeout_duration, default, args, kwargs):
|
def __timeout(func, timeout_duration, default, args, kwargs):
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,11 @@
|
||||||
|
from __future__ import print_function
|
||||||
from setuptools import setup
|
from setuptools import setup
|
||||||
from subprocess import call
|
from subprocess import call
|
||||||
import sys
|
import sys
|
||||||
import os
|
import os
|
||||||
|
|
||||||
script_path = os.path.dirname(os.path.realpath(__file__))
|
script_path = os.path.dirname(os.path.realpath(__file__))
|
||||||
print script_path
|
print(script_path)
|
||||||
os.chdir(script_path)
|
os.chdir(script_path)
|
||||||
|
|
||||||
# Allows us to avoid installing the upstart init script when deploying on Airtime Pro:
|
# Allows us to avoid installing the upstart init script when deploying on Airtime Pro:
|
||||||
|
@ -16,7 +17,7 @@ else:
|
||||||
for root, dirnames, filenames in os.walk('pypo'):
|
for root, dirnames, filenames in os.walk('pypo'):
|
||||||
for filename in filenames:
|
for filename in filenames:
|
||||||
pypo_files.append(os.path.join(root, filename))
|
pypo_files.append(os.path.join(root, filename))
|
||||||
|
|
||||||
data_files = [
|
data_files = [
|
||||||
('/etc/init', ['install/upstart/airtime-playout.conf.template']),
|
('/etc/init', ['install/upstart/airtime-playout.conf.template']),
|
||||||
('/etc/init', ['install/upstart/airtime-liquidsoap.conf.template']),
|
('/etc/init', ['install/upstart/airtime-liquidsoap.conf.template']),
|
||||||
|
@ -29,7 +30,7 @@ else:
|
||||||
('/var/tmp/airtime/pypo/files', []),
|
('/var/tmp/airtime/pypo/files', []),
|
||||||
('/var/tmp/airtime/pypo/tmp', []),
|
('/var/tmp/airtime/pypo/tmp', []),
|
||||||
]
|
]
|
||||||
print data_files
|
print(data_files)
|
||||||
|
|
||||||
setup(name='airtime-playout',
|
setup(name='airtime-playout',
|
||||||
version='1.0',
|
version='1.0',
|
||||||
|
@ -54,19 +55,18 @@ setup(name='airtime-playout',
|
||||||
'future',
|
'future',
|
||||||
'kombu',
|
'kombu',
|
||||||
'mutagen',
|
'mutagen',
|
||||||
'poster',
|
|
||||||
'PyDispatcher',
|
'PyDispatcher',
|
||||||
'pyinotify',
|
'pyinotify',
|
||||||
'pytz',
|
'pytz',
|
||||||
'requests',
|
'requests',
|
||||||
'wsgiref',
|
'defusedxml',
|
||||||
'defusedxml'
|
'packaging',
|
||||||
],
|
],
|
||||||
zip_safe=False,
|
zip_safe=False,
|
||||||
data_files=data_files)
|
data_files=data_files)
|
||||||
|
|
||||||
# Reload the initctl config so that playout services works
|
# Reload the initctl config so that playout services works
|
||||||
if data_files:
|
if data_files:
|
||||||
print "Reloading initctl configuration"
|
print("Reloading initctl configuration")
|
||||||
#call(['initctl', 'reload-configuration'])
|
#call(['initctl', 'reload-configuration'])
|
||||||
print "Run \"sudo service airtime-playout start\" and \"sudo service airtime-liquidsoap start\""
|
print("Run \"sudo service airtime-playout start\" and \"sudo service airtime-liquidsoap start\"")
|
||||||
|
|
|
@ -1,18 +0,0 @@
|
||||||
#!/bin/bash
|
|
||||||
|
|
||||||
which py.test
|
|
||||||
pytest_exist=$?
|
|
||||||
|
|
||||||
if [ "$pytest_exist" != "0" ]; then
|
|
||||||
echo "Need to have py.test installed. Exiting..."
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
SCRIPT=`readlink -f $0`
|
|
||||||
# Absolute directory this script is in
|
|
||||||
SCRIPTPATH=`dirname $SCRIPT`
|
|
||||||
|
|
||||||
export PYTHONPATH=$PYTHONPATH:$SCRIPTPATH/..:$SCRIPTPATH/../..
|
|
||||||
|
|
||||||
py.test
|
|
||||||
|
|
|
@ -1,26 +0,0 @@
|
||||||
from pypopush import PypoPush
|
|
||||||
from threading import Lock
|
|
||||||
from Queue import Queue
|
|
||||||
|
|
||||||
import datetime
|
|
||||||
|
|
||||||
pypoPush_q = Queue()
|
|
||||||
telnet_lock = Lock()
|
|
||||||
|
|
||||||
pp = PypoPush(pypoPush_q, telnet_lock)
|
|
||||||
|
|
||||||
def test_modify_cue_in():
|
|
||||||
link = pp.modify_first_link_cue_point([])
|
|
||||||
assert len(link) == 0
|
|
||||||
|
|
||||||
min_ago = datetime.datetime.utcnow() - datetime.timedelta(minutes = 1)
|
|
||||||
link = [{"start":min_ago.strftime("%Y-%m-%d-%H-%M-%S"),
|
|
||||||
"cue_in":"0", "cue_out":"30"}]
|
|
||||||
link = pp.modify_first_link_cue_point(link)
|
|
||||||
assert len(link) == 0
|
|
||||||
|
|
||||||
link = [{"start":min_ago.strftime("%Y-%m-%d-%H-%M-%S"),
|
|
||||||
"cue_in":"0", "cue_out":"70"}]
|
|
||||||
link = pp.modify_first_link_cue_point(link)
|
|
||||||
assert len(link) == 1
|
|
||||||
|
|
|
@ -4,11 +4,15 @@ set -xe
|
||||||
|
|
||||||
[[ "$PYTHON" == false ]] && exit 0
|
[[ "$PYTHON" == false ]] && exit 0
|
||||||
|
|
||||||
|
pyenv local 3.7
|
||||||
pushd python_apps/airtime_analyzer
|
pushd python_apps/airtime_analyzer
|
||||||
pyenv local 2.7
|
pip3 install -e .
|
||||||
pip install -e .
|
nosetests
|
||||||
nosetests -a '!rgain'
|
popd
|
||||||
echo "replaygain tests where skipped due to not having a reliable replaygain install on travis."
|
|
||||||
|
pushd python_apps/api_clients
|
||||||
|
pip3 install -e .
|
||||||
|
nosetests
|
||||||
popd
|
popd
|
||||||
|
|
||||||
echo "Building docs..."
|
echo "Building docs..."
|
||||||
|
|
|
@ -65,7 +65,7 @@ echo -e "The following files, directories, and services will be removed:\n"
|
||||||
for i in ${FILES[*]}; do
|
for i in ${FILES[*]}; do
|
||||||
echo $i
|
echo $i
|
||||||
done
|
done
|
||||||
echo "pip airtime-playout"
|
echo "pip3 airtime-playout"
|
||||||
|
|
||||||
echo -e "\nIf your web root is not listed, you will need to manually remove it."
|
echo -e "\nIf your web root is not listed, you will need to manually remove it."
|
||||||
|
|
||||||
|
@ -103,6 +103,6 @@ if [[ "$IN" = "y" || "$IN" = "Y" ]]; then
|
||||||
dropAirtimeDatabase
|
dropAirtimeDatabase
|
||||||
fi
|
fi
|
||||||
|
|
||||||
pip uninstall -y airtime-playout airtime-media-monitor airtime-analyzer
|
pip3 uninstall -y airtime-playout airtime-media-monitor airtime-analyzer
|
||||||
service apache2 restart
|
service apache2 restart
|
||||||
echo "...Done"
|
echo "...Done"
|
||||||
|
|
Loading…
Reference in New Issue