diff mbox

[Branch,~linaro-image-tools/linaro-image-tools/trunk] Rev 358: Merging in index_server.py and supporting files.

Message ID 20110617183734.27855.82357.launchpad@loganberry.canonical.com
State Accepted
Headers show

Commit Message

James Tunnicliffe June 17, 2011, 6:37 p.m. UTC
Merge authors:
  James Tunnicliffe (dooferlad)
Related merge proposals:
  https://code.launchpad.net/~dooferlad/linaro-image-tools/fetch_image_server_indexer/+merge/64709
  proposed by: James Tunnicliffe (dooferlad)
  review: Approve - James Westby (james-w)
------------------------------------------------------------
revno: 358 [merge]
committer: James Tunnicliffe <james.tunnicliffe@linaro.org>
branch nick: linaro-image-tools
timestamp: Fri 2011-06-17 19:07:51 +0100
message:
  Merging in index_server.py and supporting files.
added:
  linaro_image_tools/FetchImage.py
  linaro_image_tools/fetch_image_settings.yaml
  linaro_image_tools/index_server.py


--
lp:linaro-image-tools
https://code.launchpad.net/~linaro-image-tools/linaro-image-tools/trunk

You are subscribed to branch lp:linaro-image-tools.
To unsubscribe from this branch go to https://code.launchpad.net/~linaro-image-tools/linaro-image-tools/trunk/+edit-subscription
diff mbox

Patch

=== added file 'linaro_image_tools/FetchImage.py'
--- linaro_image_tools/FetchImage.py	1970-01-01 00:00:00 +0000
+++ linaro_image_tools/FetchImage.py	2011-06-17 18:04:47 +0000
@@ -0,0 +1,1102 @@ 
+#!/usr/bin/env python
+# Copyright (C) 2010, 2011 Linaro
+#
+# Author: James Tunnicliffe <james.tunnicliffe@linaro.org>
+#
+# This file is part of Linaro Image Tools.
+#
+# Linaro Image Tools is free software; you can redistribute it and/or
+# modify it under the terms of the GNU General Public License
+# as published by the Free Software Foundation; either version 2
+# of the License, or (at your option) any later version.
+#
+# Linaro Image Tools is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License
+# along with Linaro Image Tools; if not, write to the Free Software
+# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
+# USA.
+
+import os
+import sys
+import re
+import urllib2
+import argparse
+import sqlite3
+import yaml
+import urlparse
+import logging
+import bz2
+import time
+import shutil
+import datetime
+import threading
+import subprocess
+
+
+class FileHandler():
+    """Downloads files and creates images from them by calling
+    linaro-media-create"""
+    def __init__(self):
+        import xdg.BaseDirectory as xdgBaseDir
+        self.homedir = os.path.join(xdgBaseDir.xdg_config_home,
+                                     "linaro",
+                                     "image-tools",
+                                     "fetch_image")
+
+        self.cachedir = os.path.join(xdgBaseDir.xdg_cache_home,
+                                     "linaro",
+                                     "image-tools",
+                                     "fetch_image")
+
+    class DummyEventHandler():
+        """Just a sink for events if no event handler is provided to
+        create_media"""
+        def event_start(self, event):
+            pass
+
+        def event_end(self, event):
+            pass
+
+    def has_key_and_evaluates_True(self, dictionary, key):
+        return bool(key in dictionary and dictionary[key])
+
+    def append_setting_to(self, list, dictionary, key, setting_name=None):
+        if not setting_name:
+            setting_name = "--" + key
+
+        if self.has_key_and_evaluates_True(dictionary, key):
+            list.append(setting_name)
+            list.append(dictionary[key])
+
+    def create_media(self, image_url, hwpack_url, settings, tools_dir,
+                     run_in_gui=False, event_handler=None):
+        """Create a command line for linaro-media-create based on the settings
+        provided then run linaro-media-create, either in a separate thread
+        (GUI mode) or in the current one (CLI mode)."""
+
+        import linaro_image_tools.utils
+
+        if event_handler == None:
+            event_handler = self.DummyEventHandler()
+
+        args = []
+        args.append("pkexec")
+
+        # Prefer a local linaro-media-create (from a bzr checkout for instance)
+        # to an installed one
+        lmc_command = linaro_image_tools.utils.find_command(
+                                                    'linaro-media-create',
+                                                    tools_dir)
+
+        if lmc_command:
+            args.append(os.path.abspath(lmc_command))
+        else:
+            args.append("linaro-media-create")
+
+        if run_in_gui:
+            args.append("--nocheck-mmc")
+
+        event_handler.event_start("download OS")
+        try:
+            binary_file = self.download(image_url,
+                                        settings["force_download"],
+                                        show_wx_progress=run_in_gui,
+                                        wx_progress_title=
+                                         "Downloading file 1 of 2")
+        except Exception:
+            # Download error. Hardly matters what, we can't continue.
+            print "Unexpected error:", sys.exc_info()[0]
+            logging.error("Unable to download " + image_url + " - aborting.")
+        event_handler.event_end("download OS")
+
+        if binary_file == None:  # User hit cancel when downloading
+            sys.exit(0)
+
+        event_handler.event_start("download hwpack")
+        try:
+            hwpack_file = self.download(hwpack_url,
+                                        settings["force_download"],
+                                        show_wx_progress=run_in_gui,
+                                        wx_progress_title=
+                                         "Downloading file 2 of 2")
+        except Exception:
+            # Download error. Hardly matters what, we can't continue.
+            print "Unexpected error:", sys.exc_info()[0]
+            logging.error("Unable to download " + hwpack_url + " - aborting.")
+        event_handler.event_end("download hwpack")
+
+        if hwpack_file == None:  # User hit cancel when downloading
+            sys.exit(0)
+
+        logging.info("Have downloaded OS binary to", binary_file,
+                     "and hardware pack to", hwpack_file)
+
+        if 'rootfs' in settings and settings['rootfs']:
+            args.append("--rootfs")
+            args.append(settings['rootfs'])
+
+        assert(self.has_key_and_evaluates_True(settings, 'image_file') ^
+               self.has_key_and_evaluates_True(settings, 'mmc')), ("Please "
+               "specify either an image file, or an mmc target, not both.")
+
+        self.append_setting_to(args, settings, 'mmc')
+        self.append_setting_to(args, settings, 'image_file')
+        self.append_setting_to(args, settings, 'swap_size')
+        self.append_setting_to(args, settings, 'swap_file')
+        self.append_setting_to(args, settings, 'yes_to_mmc_selection',
+                                               '--nocheck_mmc')
+
+        args.append("--dev")
+        args.append(settings['hardware'])
+        args.append("--binary")
+        args.append(binary_file)
+        args.append("--hwpack")
+        args.append(hwpack_file)
+
+        logging.info(args)
+
+        if run_in_gui:
+            self.lmcargs        = args
+            self.event_handler  = event_handler
+            self.started_lmc    = False
+            return
+
+        else:
+            self.create_process = subprocess.Popen(args)
+            self.create_process.wait()
+
+    class LinaroMediaCreate(threading.Thread):
+        """Thread class for running linaro-media-create"""
+        def __init__(self, event_handler, lmcargs, event_queue):
+            threading.Thread.__init__(self)
+            self.event_handler = event_handler
+            self.lmcargs = lmcargs
+            self.event_queue = event_queue
+
+        def run(self):
+            """Start linaro-media-create and look for lines in the output that:
+            1. Tell us that an event has happened that we can use to update the
+               UI progress.
+            2. Tell us that linaro-media-create is asking a question that needs
+               to be re-directed to the GUI"""
+
+            self.create_process = subprocess.Popen(self.lmcargs,
+                                                   stdin=subprocess.PIPE,
+                                                   stdout=subprocess.PIPE,
+                                                   stderr=subprocess.STDOUT)
+
+            self.line                       = ""
+            self.saved_lines                = ""
+            self.save_lines                 = False
+            self.state                      = None
+            self.waiting_for_event_response = False
+            self.event_queue.put(("start", "unpack"))
+
+            while(1):
+                if self.create_process.poll() != None:   # linaro-media-create
+                                                         # has finished.
+                    self.event_queue.put(("terminate"))  # Tell the GUI
+                    return                               # Terminate the thread
+
+                self.input = self.create_process.stdout.read(1)
+
+                # We build up lines by extracting data from linaro-media-create
+                # a single character at a time. This is because if we fetch
+                # whole lines, using Popen.communicate a character is
+                # automatically sent back to the running process, which if the
+                # process is waiting for user input, can trigger a default
+                # action. By using stdout.read(1) we pick off a character at a
+                # time and avoid this problem.
+                if self.input == '\n':
+                    if self.save_lines:
+                        self.saved_lines += self.line
+
+                    self.line = ""
+                else:
+                    self.line += self.input
+
+                if self.line == "Updating apt package lists ...":
+                    self.event_queue.put(("end", "unpack"))
+                    self.event_queue.put(("start", "installing packages"))
+                    self.state = "apt"
+
+                elif(self.line ==
+                     "WARNING: The following packages cannot be authenticated!"):
+                    self.saved_lines = ""
+                    self.save_lines = True
+
+                elif(self.line ==
+                     "Install these packages without verification [y/N]? "):
+                    self.saved_lines = re.sub("WARNING: The following packages"
+                                              " cannot be authenticated!",
+                                              "", self.saved_lines)
+                    self.event_queue.put(("start",
+                                          "unverified_packages:"
+                                          + self.saved_lines))
+                    self.line = ""
+                    self.waiting_for_event_response = True  # Wait for restart
+
+                elif self.line == "Do you want to continue [Y/n]? ":
+                    self.send_to_create_process("y")
+                    self.line = ""
+
+                elif self.line == "Done" and self.state == "apt":
+                    self.state = "create file system"
+                    self.event_queue.put(("end", "installing packages"))
+                    self.event_queue.put(("start", "create file system"))
+
+                elif(    self.line == "Created:"
+                     and self.state == "create file system"):
+                    self.event_queue.put(("end", "create file system"))
+                    self.event_queue.put(("start", "populate file system"))
+                    self.state = "populate file system"
+
+                while self.waiting_for_event_response:
+                    time.sleep(0.2)
+
+        def send_to_create_process(self, text):
+            print >> self.create_process.stdin, text
+            self.waiting_for_event_response = False
+
+    def start_lmc_gui_thread(self, event_queue):
+        self.lmc_thread = self.LinaroMediaCreate(self.event_handler,
+                                                 self.lmcargs, event_queue)
+        self.lmc_thread.start()
+
+    def kill_create_media(self):
+        pass  # TODO: Something!
+              # Need to make sure all child processes are terminated.
+
+    def send_to_create_process(self, text):
+        self.lmc_thread.send_to_create_process(text)
+
+    def name_and_path_from_url(self, url):
+        """Return the file name and the path at which the file will be stored
+        on the local system based on the URL we are downloading from"""
+        # Use urlparse to get rid of everything but the host and path
+        scheme, host, path, params, query, fragment = urlparse.urlparse(url)
+
+        url_search = re.search(r"^(.*)/(.*?)$", host + path)
+        assert url_search, "URL in unexpectd format" + host + path
+
+        # Everything in url_search.group(1) should be directories and the
+        # server name and url_search.group(2) the file name
+        file_path = self.cachedir + os.sep + url_search.group(1)
+        file_name = url_search.group(2)
+
+        return file_name, file_path
+
+    def create_wx_progress(self, title, message):
+        """Create a standard WX progrss dialog"""
+        import wx
+        self.dlg = wx.ProgressDialog(title,
+                                     message,
+                                     maximum=1000,
+                                     parent=None,
+                                     style=wx.PD_CAN_ABORT
+                                      | wx.PD_APP_MODAL
+                                      | wx.PD_ELAPSED_TIME
+                                      | wx.PD_AUTO_HIDE
+                                      | wx.PD_REMAINING_TIME)
+
+    def timer_ping(self):
+        self.update_wx_process(self.download_count)
+
+    def update_wx_progress(self, count):
+        self.download_count = count
+        (self.do_download, skip) = self.dlg.Update(count)
+
+    def download(self, url, force_download=False,
+                 show_wx_progress=False, wx_progress_title=None):
+        """Downloads the file requested buy URL to the local cache and returns
+        the full path to the downloaded file"""
+
+        file_name, file_path = self.name_and_path_from_url(url)
+
+        just_file_name = file_name
+        file_name = file_path + os.sep + file_name
+
+        if not os.path.isdir(file_path):
+            os.makedirs(file_path)
+
+        if force_download != True and os.path.exists(file_name):
+            logging.info(file_name + " already cached. Not downloading (use "
+                                     "--force-download to override).")
+            return file_name
+
+        logging.info("Fetching", url)
+
+        maxtries = 10
+        for trycount in range(0, maxtries):
+            try:
+                response = urllib2.urlopen(url)
+            except:
+                if trycount < maxtries - 1:
+                    print "Unable to download", url, "retrying in 5 seconds..."
+                    time.sleep(5)
+                    continue
+                else:
+                    print "Download failed for some reason:", url
+                    raise
+                    return
+            else:
+                break
+
+        self.do_download = True
+        file_out = open(file_name, 'w')
+        download_size_in_bytes = int(response.info()
+                                     .getheader('Content-Length').strip())
+        chunks_downloaded = 0
+
+        show_progress = download_size_in_bytes > 1024 * 200
+
+        if show_progress:
+            chunk_size = download_size_in_bytes / 1000
+            if show_wx_progress:
+                if wx_progress_title == None:
+                    wx_progress_title = "Downloading File"
+                self.create_wx_progress(wx_progress_title,
+                                        "Downloading " + just_file_name)
+            else:
+                print "Fetching", url
+        else:
+            chunk_size = download_size_in_bytes
+
+        if show_progress and show_wx_progress:
+            # Just update the download box before we get the first %
+            self.update_wx_progress(0)
+
+        printed_progress = False
+        while self.do_download:
+            chunk = response.read(chunk_size)
+            if len(chunk):
+                # Print a % download complete so we don't get too bored
+                if show_progress:
+                    if show_wx_progress:
+                        self.update_wx_progress(chunks_downloaded)
+                    else:
+                        # Have 1000 chunks so div by 10 to get %...
+                        sys.stdout.write("\r%d%%" % (chunks_downloaded / 10))
+                        printed_progress = True
+                sys.stdout.flush()
+
+                file_out.write(chunk)
+                chunks_downloaded += 1
+
+            else:
+                if printed_progress:
+                    print ""
+                break
+
+        file_out.close()
+
+        if self.do_download == False:
+            os.remove(file_name)
+            return None
+
+        return file_name
+
+    def download_if_old(self, url, force_download, show_wx_progress=False):
+        file_name, file_path = self.name_and_path_from_url(url)
+
+        file_path_and_name = file_path + os.sep + file_name
+
+        if(not os.path.isdir(file_path)):
+            os.makedirs(file_path)
+        try:
+            force_download = (force_download == True
+                              or (    time.mktime(time.localtime())
+                                    - os.path.getmtime(file_path_and_name)
+                                  > 60 * 60 * 24))
+        except OSError:
+            force_download = True  # File not found...
+
+        return self.download(url, force_download, show_wx_progress)
+
+    def update_files_from_server(self, force_download=False,
+                                 show_wx_progress=False):
+
+        settings_url     = "http://z.nanosheep.org/fetch_image_settings.yaml"
+        server_index_url = "http://z.nanosheep.org/server_index.bz2"
+
+        self.settings_file = self.download_if_old(settings_url,
+                                                  force_download,
+                                                  show_wx_progress)
+
+        self.index_file = self.download_if_old(server_index_url,
+                                               force_download,
+                                               show_wx_progress)
+
+        zip_search = re.search(r"^(.*)\.bz2$", self.index_file)
+
+        if zip_search:
+            # index file is compressed, unzip it
+            zipped = bz2.BZ2File(self.index_file)
+            unzipped = open(zip_search.group(1), "w")
+
+            unzipped.write(zipped.read())
+
+            zipped.close()
+            unzipped.close()
+
+            self.index_file = zip_search.group(1)
+
+    def clean_cache(self):
+        shutil.rmtree(self.cachedir)
+
+
+class FetchImageConfig():
+    """Reads settings from the settings YAML file as well as those from the
+       command line providing a central settings repository."""
+
+    def __init__(self):
+        self.settings = {}
+
+    def read_config(self, file_name):
+        try:
+            f = open(file_name, "r")
+
+            if(f):
+                self.settings = dict(self.settings.items() +
+                                     yaml.load(f.read()).items())
+
+        except IOError:
+            print "Unable to read settings file %s", file_name
+            logging.error("Unable to read settings file %s", file_name)
+            sys.exit(0)
+
+        self.settings['write_to_file_or_device'] = "file"
+
+        # At this point, assume settings have been loaded.
+        # We need some reverse mappings set up
+        self.settings['UI']['reverse-descriptions'] = {}
+        for (key, value) in self.settings['UI']['descriptions'].items():
+            if isinstance(value, basestring):
+                value = re.sub('LEB:\s*', '', value)
+                self.settings['UI']['reverse-descriptions'][value] = key
+
+        self.settings['UI']['reverse-translate'] = {}
+        for (key, value) in self.settings['UI']['translate'].items():
+            self.settings['UI']['reverse-translate'][value] = key
+
+    def parse_args(self, args):
+        parser = argparse.ArgumentParser(description=
+                                         "Create a board image, first "
+                                         "downloading any required files.")
+
+        for (key, value) in self.settings['choice'].items():
+            parser.add_argument(
+                       "-" + self.settings['UI']['cmdline short names'][key],
+                       "--" + key,
+                       help=self.settings['UI']['descriptions']['choice'][key],
+                       required=True)
+
+        parser.add_argument("-x", "--clean-cache",
+                            help="Delete all cached downloads",
+                            action='store_true')
+        parser.add_argument("-d", "--force-download",
+                            help="Force re-downloading of cached files",
+                            action='store_true')
+        parser.add_argument(
+                    "-t", "--image-file",
+                    help="Where to write image file to (use this XOR mmc)")
+        parser.add_argument(
+                  "-m", "--mmc",
+                  help="What disk to write image to (use this XOR image-file)")
+        parser.add_argument("--swap-file",
+                            help="Swap file size for created image")
+        parser.add_argument("--image-size",
+                            help="Image file size for created image")
+        parser.add_argument("--rootfs",
+                            help="Root file system type for created image")
+
+        self.args = vars(parser.parse_args(args))
+
+
+class DB():
+    """Interacts with the database storing URLs of files that are available
+       for download to create images with. Provides functions for both creation
+       and reading."""
+
+    def __init__(self, index_name):
+        # http://www.sqlite.org/lang_transaction.html - defer acquiring a locK
+        # until it is required.
+        self.db_file_name = index_name
+        self.database = sqlite3.connect(index_name, isolation_level="DEFERRED")
+        self.c = self.database.cursor()
+        self.touched_urls = {}
+
+    def close(self):
+        self.database.close()
+        self.database = None
+
+    def set_url_parse_info(self, url_parse):
+        self.url_parse = url_parse
+
+    def record_url(self, url, table):
+        """Check to see if the record exists in the index, if not, add it"""
+
+        assert self.url_parse[table]["base_url"] != None, ("Can not match the "
+               "URL received (%s) to an entry provided by add_url_parse_list",
+               url)
+        assert re.search('^' + self.url_parse[table]["base_url"], url)
+
+        if(not re.search(self.url_parse[table]["url_validator"], url)):
+            #Make sure that the URL matches the validator
+            return
+
+        logging.info("Recording URL", url)
+
+        assert url not in self.touched_urls, ("URLs expected to only be added "
+                                              "to 1 place\n" + url)
+
+        self.touched_urls[url] = True
+
+        # Do not add the record if it already exists
+        self.c.execute("select url from " + table + " where url == ?", (url,))
+        if(self.c.fetchone()):
+            return
+
+        url_match = re.search(self.url_parse[table]["base_url"] + r"(.*)$",
+                              url)
+        url_chunks = url_match.group(1).lstrip('/').encode('ascii').split('/')
+        # url_chunks now contains all parts of the url, split on /,
+        # not including the base URL
+
+        # We now construct an SQL command to insert the index data into the
+        # database using the information we have.
+
+        # Work out how many values we will insert into the database
+        length = 0
+        for name in self.url_parse[table]["url_chunks"]:
+            if(name != ""):
+                length += 1
+
+        sqlcmd = "insert into " + table + " values ("
+
+        # Add the appropriate number of ?s (+1 is so we have room for url)
+        sqlcmd += "".join(["?, " for x in range(length + 1)])
+        sqlcmd = sqlcmd.rstrip(" ")  # get rid of unwanted space
+        sqlcmd = sqlcmd.rstrip(",")  # get rid of unwanted comma
+        sqlcmd += ")"
+
+        # Get the parameters from the URL to record in the SQL database
+        sqlparams = []
+        chunk_index = 0
+        for name in self.url_parse[table]["url_chunks"]:
+            # If this part of the URL isn't a parameter, don't insert it
+            if(name != ""):
+                # If the entry is a tuple, it indicates it is of the form
+                # name, regexp
+                if(isinstance(name, tuple)):
+                    # use stored regexp to extract data for the database
+                    match = re.search(name[1], url_chunks[chunk_index])
+                    assert match, ("Unable to match regexp to string ",
+                                  + url_chunks[chunk_index] + " " + name[1])
+                    sqlparams.append(match.group(1))
+
+                else:
+                    sqlparams.append(url_chunks[chunk_index])
+
+            chunk_index += 1
+
+        sqlparams.append(url)
+
+        self.c.execute(sqlcmd, tuple(sqlparams))
+
+    def commit(self):
+        self.database.commit()
+
+    def __del__(self):
+        if(self.database):
+            self.commit()
+            self.database.close()
+
+    def create_table_with_url_text_items(self, table, items):
+        cmd = "create table if not exists "
+        cmd += table + " ("
+
+        for item in items:
+            cmd += item + " TEXT, "
+
+        cmd += "url TEXT)"
+
+        self.execute(cmd)
+
+    def execute(self, cmd, params=None):
+        self.c = self.database.cursor()
+        logging.info(cmd)
+        if(params):
+            logging.info(params)
+            self.c.execute(cmd, params)
+        else:
+            self.c.execute(cmd)
+
+    def execute_return_list(self, cmd, params=None):
+        self.execute(cmd, params)
+        list = []
+        item = self.c.fetchone()
+        while item:
+            list.append(item)
+            item = self.c.fetchone()
+
+        return list
+
+    def delete_by_url(self, table, url):
+        self.execute("delete from " + table + " where url == ?", (url,))
+
+    def clean_removed_urls_from_db(self):
+        self.c = self.database.cursor()
+
+        for table, info in self.url_parse.items():
+            self.c.execute("select url from " + table)
+            to_delete = []
+
+            while(1):
+                url = self.c.fetchone()
+                if(url == None):
+                    break
+
+                if(url[0] not in self.touched_urls):
+                    to_delete.append(url[0])
+
+            # We can't delete an item from the database while iterating on the
+            # results of a previous query, so we store the URLs of entries to
+            # delete, and batch them up
+            for url in to_delete:
+                self.delete_by_url(table, url)
+
+        self.commit()
+
+    def get_url(self, table, key_value_pairs, sql_extras=None):
+        """Return the first matching URL from the specified table building up a
+        SQL query based on the values inkey_value_pairs creating key == value
+        sections of the query. sql_extras is appended to the created SQL query,
+         so we can get the first entry in an ordered list for instance"""
+        self.c = self.database.cursor()
+
+        cmd = "select url from " + table + " where "
+        params = []
+
+        first = True
+        for key, value in key_value_pairs:
+            if(first == False):
+                cmd += " AND "
+            else:
+                first = False
+
+            cmd += key + " == ? "
+            params.append(value)
+
+        if(sql_extras):
+            cmd += " " + sql_extras
+
+        self.c.execute(cmd, tuple(params))
+        url = self.c.fetchone()
+
+        return url
+
+    def get_platforms(self, table):
+        self.execute("select distinct platform from ?", table)
+
+        platforms = []
+        platform = self.c.fetchone()
+
+        while(platform):
+            platforms.append(platform)
+            platform = self.c.fetchone()
+
+        return platforms
+
+    def get_builds(self, platform, image=None):
+        """return a complete list of builds available from the releases
+        repository"""
+
+        build_tuples = self.execute_return_list(
+                                'select distinct build from release_binaries '
+                                'where platform == "' + platform + '"')
+        hwpack_build_tuples = self.execute_return_list(
+                                   'select distinct build from release_hwpacks'
+                                   ' where platform == "' + platform + '"')
+
+        # Need to also check hwpacks and image -> some builds are only
+        # Available for a particular image (OS). Guess this makes sense as OS
+        # customisations may come and go.
+
+        image_builds = [build[0] for build in build_tuples]
+        hwpack_builds = [build[0] for build in hwpack_build_tuples]
+
+        builds = []
+        for build in image_builds:
+            if build in hwpack_builds:
+                # Only use a build if it exists for the hardware pack as well
+                # as the image and platform chosen
+                builds.append(build)
+
+        builds.sort()
+
+        # Just so final is always the last item, if it exists...
+        if "final" in builds:
+            builds.remove("final")
+            builds.append("final")
+
+        return builds
+
+    def get_hwpacks(self, table, platform=None):
+        """Return a list of all the hardware packs available in the specified
+        table"""
+
+        query = 'select distinct hardware from ' + table
+        if platform:
+            query += ' where platform == "' + platform + '"'
+
+        results = self.execute_return_list(query)
+
+        hwpacks = [item[0] for item in results]
+        return hwpacks
+
+    def get_dates_and_builds_of_snapshots(self):
+        """Return all the dates that match an an OS and hardware pack build"""
+        # First get all dates from hwpack table
+        self.execute("select distinct date from snapshot_hwpacks "
+                     "order by date")
+
+        date_with_hwpack = {}
+        date = self.c.fetchone()
+
+        while date:
+            date_with_hwpack[date] = 1
+            date = self.c.fetchone()
+
+        # Now make sure there is a file system image for that date as well
+        self.execute("select distinct date from snapshot_binaries "
+                     "order by date")
+
+        date = self.c.fetchone()
+        while date:
+            if date in date_with_hwpack:
+                date_with_hwpack[date] = 2
+            date = self.c.fetchone()
+
+        # Now anything with both a hwpack and a file system has a value of 2
+        # recorded in date_with_hwpack[date]
+        date_with_build = {}
+
+        for key, date in date_with_hwpack.items():
+            if(key == 2):
+                date_with_build[date] = True
+
+        return date_with_build
+
+    def get_hardware_from_db(self, table):
+        """Get a list of hardware available from the given table"""
+        self.execute("select distinct hardware from " + table +
+                     " order by hardware")
+
+        hardware = []
+        hw = self.c.fetchone()
+        while hw:
+            hardware.append(hw[0])
+            hw = self.c.fetchone()
+
+        return hardware
+
+    def image_hardware_combo_available(self,
+                                       release_or_snapshot,
+                                       image,
+                                       hwpacks):
+        """Try and find a matching plaform, build pair for both the provided
+        image and one of the hwpacks in the list."""
+        binary_table = release_or_snapshot + "_binaries"
+        hwpack_table = release_or_snapshot + "_hwpacks"
+
+        binary_list = self.execute_return_list(
+                            'select distinct platform, build from '
+                             + binary_table + ' where image == ?', (image,))
+
+        for hwpack in hwpacks:
+            hwpack_list = self.execute_return_list(
+                                'select distinct platform, build from '
+                                + hwpack_table +
+                                ' where hardware == ?', (hwpack,))
+
+            for item in binary_list:
+                if item in hwpack_list:
+                    return True
+
+        return False
+
+    def hardware_is_available_in_table(self, table, hardware):
+        return len(self.execute_return_list(
+                    'select url from ' + table +
+                    ' where hardware == ?', (hardware,))) > 0
+
+    def hardware_is_available_for_platform(self,
+                                           hardware_list,
+                                           platform,
+                                           table="release_hwpacks"):
+
+        for hardware in self.execute_return_list(
+                              'select distinct hardware from ' + table +
+                              ' where platform == ?', (platform,)):
+            if hardware[0] in hardware_list:
+                return True
+
+        return False
+
+    def get_available_hwpacks_for_hardware_build_plaform(self,
+                                                         hardware_list,
+                                                         platform,
+                                                         build):
+        hwpacks = []
+        for hardware in self.execute_return_list(
+                             'select distinct hardware from release_hwpacks '
+                             'where platform == ? and build == ?',
+                             (platform, build)):
+
+            if hardware[0] in hardware_list:
+                hwpacks.append(hardware[0])
+
+        return hwpacks
+
+    def image_is_available_for_platform(self,
+                                        table,
+                                        platform,
+                                        image):
+
+        return len(self.execute_return_list(
+                        'select * from ' + table +
+                        ' where platform == ? and image == ?',
+                        (platform, image))) > 0
+
+    def hardware_is_available_for_platform_build(self,
+                                                 hardware_list,
+                                                 platform,
+                                                 build):
+
+        for hardware in self.execute_return_list(
+                             'select distinct hardware from release_hwpacks '
+                             'where platform == ? and build == ?',
+                              (platform, build)):
+
+            if hardware[0] in hardware_list:
+                return True
+
+        return False
+
+    def build_is_available_for_platform_image(self,
+                                              table,
+                                              platform,
+                                              image,
+                                              build):
+
+        return len(self.execute_return_list(
+                        'select * from ' + table +
+                        ' where platform == ? and image == ? and build == ?',
+                        (platform, image, build))) > 0
+
+    def get_available_hwpacks_for_hardware_snapshot_build(self,
+                                                          hardware_list,
+                                                          platform,
+                                                          date,
+                                                          build):
+        hwpacks = []
+        for hardware in self.execute_return_list(
+                             'select distinct hardware from snapshot_hwpacks '
+                             'where platform == ? '
+                             'and   date == ? '
+                             'and   build == ?',
+                              (platform, date, build)):
+            if hardware[0] in hardware_list:
+                hwpacks.append(hardware[0])
+
+        return hwpacks
+
+    def get_binary_builds_on_day_from_db(self, image, date, hwpacks):
+        """Return a list of build numbers that are available on the given date
+        for the given image ensuring that the reqiested hardware is also
+        available."""
+
+        # Remove the dashes from the date we get in (YYYY-MM-DD --> YYYYMMDD)
+        date = re.sub('-', '', date)
+
+        binaries = self.execute_return_list(
+                        "select build from snapshot_binaries "
+                        "where date == ? and image == ?",
+                        (date, image))
+
+        if len(binaries) == 0:
+            # Try adding "linaro-" to the beginning of the image name
+            binaries = self.execute_return_list(
+                            "select build from snapshot_binaries "
+                            "where date == ? and image == ?",
+                            (date, "linaro-" + image))
+
+        for hwpack in hwpacks:
+            builds = self.execute_return_list(
+                           "select build from snapshot_hwpacks "
+                           "where date == ? and hardware == ?",
+                           (date, hwpack))
+
+            if len(builds):
+                # A hardware pack exists for that date, return what we found
+                # for binaries
+                return binaries
+
+        # No hardware pack exists for the date requested, return empty table
+        return []
+
+    def get_next_prev_day_with_builds(self, image, date, hwpacks):
+        """Searches forwards and backwards in the database for a date with a
+        build. Will look 1 year in each direction. Returns a tuple of dates
+        in YYYYMMDD format."""
+
+        # Split ISO date into year, month, day
+        date_chunks = date.split('-')
+
+        current_date = datetime.date(int(date_chunks[0]),
+                                     int(date_chunks[1]),
+                                     int(date_chunks[2]))
+
+        one_day = datetime.timedelta(days=1)
+
+        # In case of time zone issues we add 1 day to max_search_date
+        max_search_date = datetime.date.today() + one_day
+
+        day_count = 0
+        # Look in the future & past from the given date until we find a day
+        # with a build on it
+
+        test_date = {'future': current_date,
+                      'past':   current_date}
+
+        for in_the in ["future", "past"]:
+            test_date[in_the] = None
+
+            if in_the == "future":
+                loop_date_increment = one_day
+            else:
+                loop_date_increment = -one_day
+
+            test_date[in_the] = current_date
+            
+            while test_date[in_the] <= max_search_date:
+                test_date[in_the] += loop_date_increment
+
+                builds = []
+                for hwpack in hwpacks:
+                    builds = self.get_binary_builds_on_day_from_db(
+                                    image,
+                                    test_date[in_the].isoformat(),
+                                    [hwpack])
+                    if len(builds):
+                        break
+
+                if len(builds):
+                    break
+
+                day_count += 1
+                if day_count > 365:
+                    test_date[in_the] = None
+                    break
+
+            if test_date[in_the] > max_search_date:
+                test_date[in_the] = None
+
+            if test_date[in_the]:
+                test_date[in_the] = test_date[in_the].isoformat()
+
+        return(test_date['future'], test_date['past'])
+
+    def get_os_list_from(self, table):
+        return [item[0] for item in self.execute_return_list(
+                                        "select distinct image from "
+                                         + table)]
+
+    def get_image_and_hwpack_urls(self, args):
+        """ We need a platform image and a hardware pack. These are specified
+            either as
+            Release:
+                Image: platform, image, build
+                HW Pack: platform, build, hardware
+            Snapshot:
+                Image: platform, image, date, build
+                HW Pack: platform, hardware, date, build"""
+        image_url = None
+        hwpack_url = None
+
+        if(args['release_or_snapshot'] == "snapshot"):
+            count = 0
+            while(1):  # Just so we can try several times...
+                if(args['build'] == "latest"):
+                    # First result of SQL query is latest build (ORDER BY)
+                    image_url = self.get_url("snapshot_binaries",
+                                    [
+                                     ("image", args['image'])],
+                                    "ORDER BY date DESC, build DESC")
+
+                    hwpack_url = self.get_url("snapshot_hwpacks",
+                                    [
+                                     ("hardware", args['hwpack'])],
+                                    "ORDER BY date DESC, build DESC")
+
+                else:
+                    build_bits = args['build'].split(":")
+                    assert re.search("^\d+$", build_bits[0]), (
+                            "Unexpected date format in build parameter "
+                            + build_bits[0])
+                    assert re.search("^\d+$", build_bits[1]), (
+                            "Build number in shapshot build parameter should "
+                            "be an integer " + build_bits[1])
+
+                    image_url = self.get_url("snapshot_binaries",
+                                    [
+                                     ("image", args['image']),
+                                     ("build", build_bits[1]),
+                                     ("date", build_bits[0])])
+
+                    hwpack_url = self.get_url("snapshot_hwpacks",
+                                    [
+                                     ("hardware", args['hwpack']),
+                                     ("build", build_bits[1]),
+                                     ("date", build_bits[0])])
+
+                if(count == 1):
+                    break
+                count += 1
+
+                if(image_url == None):
+                    # If we didn't match anything, try prepending "linaro-"
+                    # to the image name
+                    args['image'] = "linaro-" + args['image']
+                else:
+                    break  # Got a URL, go.
+
+        else:
+            image_url = self.get_url("release_binaries",
+                            [
+                             ("image", args['image']),
+                             ("build", args['build']),
+                             ("platform", args['platform'])])
+
+            hwpack_url = self.get_url("release_hwpacks",
+                            [
+                             ("hardware", args['hwpack']),
+                             ("build", args['build']),
+                             ("platform", args['platform'])])
+
+        if(not image_url):
+            # If didn't get an image URL set up something so the return line
+            # doesn't crash
+            image_url = [None]
+
+        if(not hwpack_url):
+            # If didn't get a hardware pack URL set up something so the return
+            # line doesn't crash
+            hwpack_url = [None]
+
+        return(image_url[0], hwpack_url[0])

=== added file 'linaro_image_tools/fetch_image_settings.yaml'
--- linaro_image_tools/fetch_image_settings.yaml	1970-01-01 00:00:00 +0000
+++ linaro_image_tools/fetch_image_settings.yaml	2011-06-17 18:04:47 +0000
@@ -0,0 +1,109 @@ 
+UI:
+  descriptions:
+    alip: ARM Internet Platform
+    choice:
+      build: The build (alpha-1 / beta / final) or date of snapshot (YYYYMMDD:build_number or 'latest')
+        to fetch.
+      hardware: The board you are building the image to run on.
+      hwpack: The hardware pack to use when building the image.
+      image: A distribuion image, e.g. headless.
+      platform: Which Linaro platform to build an image from. Specify 'snapshot' to
+        use a snapshot rather than a release.
+    developer: Developer Tools
+    graphics: Graphics
+    multimedia: Multimedia
+    nano: Nano
+    ubuntu-desktop: 'LEB: Linaro Ubuntu Desktop'
+    ubuntu-desktop::long: Linux for humans on low power machines
+    ubuntu-desktop::release_note: Shiny!
+  help:
+    latest snapshot: Latest Snapshot help text
+    release: Release help text
+    snapshot: Snapshot help text
+  translate:
+    linaro-m: '10.11'
+    linaro-n: '11.05'
+  cmdline short names:
+    build: b
+    hardware: w
+    platform: a
+    image: i
+    hwpack: p
+choice:
+  build: {}
+  hardware:
+    beagle: Beagle Board
+    igep: IGEPv2
+    panda: Panda Board
+    vexpress: Versatile Express
+    ux500: UX500
+    efikamx: EFIKA MX Smarttop 
+    efikasb: EFIKA MX Smartbook
+    mx51evk: i.MX51
+    mx53loco: i.MX53 Loco
+    # I don't have a --dev setting to linaro-media-create for a u8500 yet.
+    # u8500: U8500
+    overo: Overo
+    smdkv310: S5PV310
+  
+  hwpack:
+    beagle:
+    - omap3
+    - omap3-x11-base
+    
+    igep:
+    - igep
+    
+    panda:
+    - panda
+    - panda-x11-base
+    - bsp-omap4
+    
+    vexpress:
+    - vexpress
+    
+    ux500:
+    - bsp-ux500
+    
+    efikamx:
+    - efikamx
+    
+    efikasb:
+    - efikamx
+    
+    mx51evk:
+    - imx51
+    
+    mx53loco:
+    - lt-mx53loco
+    - s5pv310
+     
+    u8500:
+     - lt-u8500
+     
+    overo:
+    - overo
+     
+    smdkv310:
+     - lt-s5pv310
+     - s5pv310
+
+  image:
+  - alip
+  - developer
+  - headless
+  - ubuntu-desktop
+  platform:
+    linaro-m:
+    - final
+    linaro-n:
+    - alpha-1
+    - alpha-2
+    - alpha-3
+    - beta
+    - rc
+    - final
+
+platform: linaro-n
+repository: release
+

=== added file 'linaro_image_tools/index_server.py'
--- linaro_image_tools/index_server.py	1970-01-01 00:00:00 +0000
+++ linaro_image_tools/index_server.py	2011-06-17 18:04:47 +0000
@@ -0,0 +1,157 @@ 
+#!/usr/bin/env python
+# Copyright (C) 2010, 2011 Linaro
+#
+# Author: James Tunnicliffe <james.tunnicliffe@linaro.org>
+#
+# This file is part of Linaro Image Tools.
+#
+# Linaro Image Tools is free software; you can redistribute it and/or
+# modify it under the terms of the GNU General Public License
+# as published by the Free Software Foundation; either version 2
+# of the License, or (at your option) any later version.
+# 
+# Linaro Image Tools is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+# GNU General Public License for more details.
+# 
+# You should have received a copy of the GNU General Public License
+# along with Linaro Image Tools; if not, write to the Free Software
+# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
+# USA.
+
+import os
+import re
+import FetchImage
+import urlparse
+import logging
+import bz2
+
+RELEASES_WWW_DOCUMENT_ROOT  = "/srv/releases.linaro.org/www/platform/"
+RELEASE_URL                 = "http://releases.linaro.org/platform/"
+SNAPSHOTS_WWW_DOCUMENT_ROOT = "/srv/snapshots.linaro.org/www/"
+SNAPSHOTS_URL               = "http://snapshots.linaro.org/"
+
+class ServerIndexer():
+    """Create a database of files on the linaro image servers for use by image
+       creation tools."""
+    def reset(self):
+        self.url_parse = {}
+
+    def __init__(self):
+        self.reset()
+        self.db_file_name = "server_index"
+        self.db = FetchImage.DB(self.db_file_name)
+
+    def crawl(self):
+        self.db.set_url_parse_info(self.url_parse)
+        logging.info(self.url_parse.items())
+        
+        for table, info in self.url_parse.items():
+            logging.info(info["base_dir"], ":", info["base_url"], table,
+                         info["url_validator"], info["url_chunks"])
+            self.go(info["base_dir"], info["base_url"], table)
+            logging.info("")
+
+    def go(self, root_dir_, root_url_, table_):
+        for root, subFolders, files in os.walk( root_dir_ ):
+            for file in files:
+                if(re.search('\.gz$', file)):
+                    # Construct a URL to the file and save in the database
+                    relative_location = re.sub(root_dir_, "", 
+                                               os.path.join(root, file))
+                    url = urlparse.urljoin(root_url_, relative_location)
+                    url = urlparse.urljoin(url, file)
+                   
+                    if not re.search('/leb-panda/', url):
+                        logging.info(url)
+                        self.db.record_url(url, table_)
+                    
+        self.dump() 
+
+    def dump(self):
+        self.db.commit()
+        
+    def close_and_bzip2(self):
+        # After finishing creating the database, create a compressed version
+        # for more efficient downloads
+        self.db.close()
+        bz2_db_file = bz2.BZ2File(self.db_file_name + ".bz2", "w")
+        db_file = open(self.db_file_name)
+        bz2_db_file.write(db_file.read())
+        bz2_db_file.close()
+
+    def add_directory_parse_list(self,
+                                 base_dir_,
+                                 base_url_,
+                                 url_validator_,
+                                 id_,
+                                 url_chunks_):
+        
+        if(not id_ in self.url_parse):
+            
+            print base_dir_
+            self.url_parse[id_] = {"base_dir":      base_dir_,
+                                   "base_url":      base_url_,
+                                   "url_validator": url_validator_,
+                                   "url_chunks":    url_chunks_}
+            logging.info(self.url_parse[id_]["base_dir"])
+
+            # Construct data needed to create the table
+            items = []
+            for item in url_chunks_:
+                if(item != ""):
+                    # If the entry is a tuple, it indicates it is of the
+                    # form name, regexp
+                    if(isinstance(item, tuple)):
+                        items.append(item[0])
+                    else:
+                        items.append(item)
+
+            self.db.create_table_with_url_text_items(id_, items)
+
+    def clean_removed_urls_from_db(self):
+        self.db.clean_removed_urls_from_db()
+
+if __name__ == '__main__':
+    crawler = ServerIndexer()
+
+    # The use of a zero width assertion here to look for links that don't 
+    # contain /hwpacks/ is a bit scary and could be replaced by a tuple of
+    # (False, r"hwpacks"), where the first parameter could indicate that we
+    # want the regexp to fail if we are to use the URL. May be a bit nicer.
+    
+    #http://releases.linaro.org/platform/linaro-m/plasma/final/
+    crawler.add_directory_parse_list(RELEASES_WWW_DOCUMENT_ROOT,
+                                     RELEASE_URL,
+                                     r"^((?!hwpack).)*$",
+                                     "release_binaries",
+                                     ["platform", "image", "build"])
+
+    #http://releases.linaro.org/platform/linaro-m/hwpacks/final/hwpack_linaro-bsp-omap4_20101109-1_armel_unsupported.tar.gz
+    crawler.add_directory_parse_list(RELEASES_WWW_DOCUMENT_ROOT,
+                                     RELEASE_URL,
+                                     r"/hwpacks/",
+                                     "release_hwpacks",
+                                     ["platform", "", "build",
+                                      ("hardware", r"hwpack_linaro-(.*?)_")])
+    
+    #http://snapshots.linaro.org/11.05-daily/linaro-alip/20110420/0/images/tar/
+    crawler.add_directory_parse_list(SNAPSHOTS_WWW_DOCUMENT_ROOT,
+                                     SNAPSHOTS_URL,
+                                     r"^((?!hwpack).)*$",
+                                     "snapshot_binaries",
+                                     ["platform", "image", "date", "build"])
+
+    #http://snapshots.linaro.org/11.05-daily/linaro-hwpacks/omap3/20110420/0/images/hwpack/
+    crawler.add_directory_parse_list(SNAPSHOTS_WWW_DOCUMENT_ROOT,
+                                     SNAPSHOTS_URL,
+                                     r"/hwpack/",
+                                     "snapshot_hwpacks",
+                                     ["platform", "", "hardware", "date",
+                                      "build"])
+
+    crawler.crawl()
+    crawler.clean_removed_urls_from_db()
+    crawler.dump()
+    crawler.close_and_bzip2()