import os import sys import time import logging import logging.config import shutil import pickle import random import string import json import telnetlib import math from threading import Thread # For RabbitMQ from kombu.connection import BrokerConnection from kombu.messaging import Exchange, Queue, Consumer, Producer from api_clients import api_client from util import CueFile from configobj import ConfigObj # configure logging logging.config.fileConfig("logging.cfg") # loading config file try: config = ConfigObj('config.cfg') LS_HOST = config['ls_host'] LS_PORT = config['ls_port'] POLL_INTERVAL = int(config['poll_interval']) except Exception, e: print 'Error loading config file: ', e sys.exit() # Yuk - using a global, i know! SCHEDULE_PUSH_MSG = [] """ Handle a message from RabbitMQ, put it into our yucky global var. Hopefully there is a better way to do this. """ def handle_message(body, message): logger = logging.getLogger('fetch') global SCHEDULE_PUSH_MSG logger.info("Received schedule from RabbitMQ: " + message.body) SCHEDULE_PUSH_MSG = json.loads(message.body) # ACK the message to take it off the queue message.ack() class PypoFetch(Thread): def __init__(self, q): Thread.__init__(self) logger = logging.getLogger('fetch') self.api_client = api_client.api_client_factory(config) self.cue_file = CueFile() self.set_export_source('scheduler') self.queue = q logger.info("Initializing RabbitMQ stuff") schedule_exchange = Exchange("airtime-schedule", "direct", durable=True, auto_delete=True) schedule_queue = Queue("pypo-fetch", exchange=schedule_exchange, key="foo") self.connection = BrokerConnection(config["rabbitmq_host"], config["rabbitmq_user"], config["rabbitmq_password"], "/") channel = self.connection.channel() consumer = Consumer(channel, schedule_queue) consumer.register_callback(handle_message) consumer.consume() logger.info("PypoFetch: init complete"); def set_export_source(self, export_source): self.export_source = export_source self.cache_dir = config["cache_dir"] + self.export_source + '/' """ Process the schedule - Reads the scheduled entries of a given range (actual time +/- "prepare_ahead" / "cache_for") - Saves a serialized file of the schedule - playlists are prepared. (brought to liquidsoap format) and, if not mounted via nsf, files are copied to the cache dir (Folder-structure: cache/YYYY-MM-DD-hh-mm-ss) - runs the cleanup routine, to get rid of unused cashed files """ def process_schedule(self, schedule_data, export_source): logger = logging.getLogger('fetch') self.schedule = schedule_data["playlists"] # Push stream metadata to liquidsoap # TODO: THIS LIQUIDSOAP STUFF NEEDS TO BE MOVED TO PYPO-PUSH!!! stream_metadata = schedule_data['stream_metadata'] try: tn = telnetlib.Telnet(LS_HOST, LS_PORT) #encode in latin-1 due to telnet protocol not supporting utf-8 tn.write(('vars.stream_metadata_type %s\n' % stream_metadata['format']).encode('latin-1')) tn.write(('vars.station_name %s\n' % stream_metadata['station_name']).encode('latin-1')) tn.write('exit\n') tn.read_all() except Exception, e: logger.error("Exception %s", e) status = 0 # Download all the media and put playlists in liquidsoap format try: playlists = self.prepare_playlists() except Exception, e: logger.error("%s", e) # Send the data to pypo-push scheduled_data = dict() scheduled_data['playlists'] = playlists scheduled_data['schedule'] = self.schedule scheduled_data['stream_metadata'] = schedule_data["stream_metadata"] self.queue.put(scheduled_data) # cleanup try: self.cleanup(self.export_source) except Exception, e: logger.error("%s", e) """ In this function every audio file is cut as necessary (cue_in/cue_out != 0) and stored in a playlist folder. file is e.g. 2010-06-23-15-00-00/17_cue_10.132-123.321.mp3 """ def prepare_playlists(self): logger = logging.getLogger('fetch') schedule = self.schedule playlists = dict() # Dont do anything if schedule is empty if not schedule: logger.debug("Schedule is empty.") return playlists scheduleKeys = sorted(schedule.iterkeys()) try: for pkey in scheduleKeys: logger.info("Playlist starting at %s", pkey) playlist = schedule[pkey] # create playlist directory try: os.mkdir(self.cache_dir + str(pkey)) except Exception, e: pass #logger.debug('*****************************************') #logger.debug('pkey: ' + str(pkey)) #logger.debug('cached at : ' + self.cache_dir + str(pkey)) #logger.debug('subtype: ' + str(playlist['subtype'])) #logger.debug('played: ' + str(playlist['played'])) #logger.debug('schedule id: ' + str(playlist['schedule_id'])) #logger.debug('duration: ' + str(playlist['duration'])) #logger.debug('source id: ' + str(playlist['x_ident'])) #logger.debug('*****************************************') if int(playlist['played']) == 1: logger.info("playlist %s already played / sent to liquidsoap, so will ignore it", pkey) elif int(playlist['subtype']) > 0 and int(playlist['subtype']) < 5: ls_playlist = self.handle_media_file(playlist, pkey) playlists[pkey] = ls_playlist except Exception, e: logger.info("%s", e) return playlists """ Download and cache the media files. This handles both remote and local files. Returns an updated ls_playlist string. """ def handle_media_file(self, playlist, pkey): ls_playlist = [] logger = logging.getLogger('fetch') for media in playlist['medias']: logger.debug("Processing track %s", media['uri']) fileExt = os.path.splitext(media['uri'])[1] try: if str(media['cue_in']) == '0' and str(media['cue_out']) == '0': #logger.debug('No cue in/out detected for this file') dst = "%s%s/%s%s" % (self.cache_dir, str(pkey), str(media['id']), str(fileExt)) do_cue = False else: #logger.debug('Cue in/out detected') dst = "%s%s/%s_cue_%s-%s%s" % \ (self.cache_dir, str(pkey), str(media['id']), str(float(media['cue_in']) / 1000), str(float(media['cue_out']) / 1000), str(fileExt)) do_cue = True # check if it is a remote file, if yes download if media['uri'][0:4] == 'http': self.handle_remote_file(media, dst, do_cue) else: logger.debug("invalid media uri: %s", media['uri']) if True == os.access(dst, os.R_OK): # check filesize (avoid zero-byte files) try: fsize = os.path.getsize(dst) except Exception, e: logger.error("%s", e) fsize = 0 if fsize > 0: pl_entry = \ 'annotate:export_source="%s",media_id="%s",liq_start_next="%s",liq_fade_in="%s",liq_fade_out="%s",schedule_table_id="%s":%s'\ % (str(media['export_source']), media['id'], 0, str(float(media['fade_in']) / 1000), \ str(float(media['fade_out']) / 1000), media['row_id'],dst) #logger.debug(pl_entry) """ Tracks are only added to the playlist if they are accessible on the file system and larger than 0 bytes. So this can lead to playlists shorter than expectet. (there is a hardware silence detector for this cases...) """ entry = dict() entry['type'] = 'file' entry['annotate'] = pl_entry entry['show_name'] = playlist['show_name'] ls_playlist.append(entry) #logger.debug("everything ok, adding %s to playlist", pl_entry) else: print 'zero-file: ' + dst + ' from ' + media['uri'] logger.warning("zero-size file - skipping %s. will not add it to playlist", dst) else: logger.warning("something went wrong. file %s not available. will not add it to playlist", dst) except Exception, e: logger.info("%s", e) return ls_playlist """ Download a file from a remote server and store it in the cache. """ def handle_remote_file(self, media, dst, do_cue): logger = logging.getLogger('fetch') if do_cue == False: if os.path.isfile(dst): pass #logger.debug("file already in cache: %s", dst) else: logger.debug("try to download %s", media['uri']) self.api_client.get_media(media['uri'], dst) else: if os.path.isfile(dst): logger.debug("file already in cache: %s", dst) else: logger.debug("try to download and cue %s", media['uri']) fileExt = os.path.splitext(media['uri'])[1] dst_tmp = config["tmp_dir"] + "".join([random.choice(string.letters) for i in xrange(10)]) + fileExt self.api_client.get_media(media['uri'], dst_tmp) # cue logger.debug("STARTING CUE") debugDst = self.cue_file.cue(dst_tmp, dst, float(media['cue_in']) / 1000, float(media['cue_out']) / 1000) logger.debug(debugDst) logger.debug("END CUE") if True == os.access(dst, os.R_OK): try: fsize = os.path.getsize(dst) except Exception, e: logger.error("%s", e) fsize = 0 if fsize > 0: logger.debug('try to remove temporary file: %s' + dst_tmp) try: os.remove(dst_tmp) except Exception, e: logger.error("%s", e) else: logger.warning('something went wrong cueing: %s - using uncued file' + dst) try: os.rename(dst_tmp, dst) except Exception, e: logger.error("%s", e) """ Cleans up folders in cache_dir. Look for modification date older than "now - CACHE_FOR" and deletes them. """ def cleanup(self, export_source): logger = logging.getLogger('fetch') offset = 3600 * int(config["cache_for"]) now = time.time() for r, d, f in os.walk(self.cache_dir): for dir in d: try: timestamp = time.mktime(time.strptime(dir, "%Y-%m-%d-%H-%M-%S")) if (now - timestamp) > offset: try: logger.debug('trying to remove %s - timestamp: %s', os.path.join(r, dir), timestamp) shutil.rmtree(os.path.join(r, dir)) except Exception, e: logger.error("%s", e) pass else: logger.info('sucessfully removed %s', os.path.join(r, dir)) except Exception, e: print e logger.error("%s", e) """ Main loop of the thread: Wait for schedule updates from RabbitMQ, but in case there arent any, poll the server to get the upcoming schedule. """ def run(self): logger = logging.getLogger('fetch') try: os.mkdir(self.cache_dir) except Exception, e: pass # Bootstrap: since we are just starting up, we need to grab the # most recent schedule. After that we can just wait for updates. status, schedule_data = self.api_client.get_schedule() if status == 1: self.process_schedule(schedule_data, "scheduler") logger.info("Bootstrap complete: got initial copy of the schedule") loops = 1 while True: logger.info("Loop #"+str(loops)) try: # Wait for messages from RabbitMQ. Timeout if we # dont get any after POLL_INTERVAL. self.connection.drain_events(timeout=POLL_INTERVAL) # Hooray for globals! schedule_data = SCHEDULE_PUSH_MSG status = 1 except: # We didnt get a message for a while, so poll the server # to get an updated schedule. status, schedule_data = self.api_client.get_schedule() if status == 1: self.process_schedule(schedule_data, "scheduler") loops += 1