LGTM
Also change "Quue" to "Queue".

On Wed, Jul 1, 2009 at 11:32 PM, Michael Hanselmann<[email protected]> wrote:
>
> Signed-off-by: Michael Hanselmann <[email protected]>
> ---
>  doc/hooks.rst        |    2 +-
>  doc/iallocator.rst   |    2 +-
>  lib/backend.py       |   18 +++++++++---------
>  lib/bdev.py          |   14 +++++++-------
>  lib/bootstrap.py     |    2 +-
>  lib/cli.py           |    8 ++++----
>  lib/cmdlib.py        |   38 +++++++++++++++++++-------------------
>  lib/config.py        |    8 ++++----
>  lib/http/__init__.py |    4 ++--
>  lib/http/auth.py     |    4 ++--
>  lib/http/server.py   |    4 ++--
>  lib/jqueue.py        |   10 +++++-----
>  lib/locking.py       |   10 +++++-----
>  lib/luxi.py          |    2 +-
>  lib/mcpu.py          |    4 ++--
>  lib/rpc.py           |   10 +++++-----
>  lib/ssh.py           |    2 +-
>  lib/utils.py         |   20 ++++++++++----------
>  man/gnt-node.sgml    |    2 +-
>  tools/lvmstrap       |    6 +++---
>  20 files changed, 85 insertions(+), 85 deletions(-)
>
> diff --git a/doc/hooks.rst b/doc/hooks.rst
> index 7dbe7d5..b2f05ce 100644
> --- a/doc/hooks.rst
> +++ b/doc/hooks.rst
> @@ -104,7 +104,7 @@ The scripts will be run as follows:
>   be left
>
>
> -All informations about the cluster is passed using environment
> +All information about the cluster is passed using environment
>  variables. Different operations will have sligthly different
>  environments, but most of the variables are common.
>
> diff --git a/doc/iallocator.rst b/doc/iallocator.rst
> index f4b8bfc..9dfc847 100644
> --- a/doc/iallocator.rst
> +++ b/doc/iallocator.rst
> @@ -233,7 +233,7 @@ The response message is much more simple than the input 
> one. It is
>  also a dict having three keys:
>
>  success
> -  a boolean value denoting if the allocation was successfull or not
> +  a boolean value denoting if the allocation was successful or not
>
>  info
>   a string with information from the scripts; if the allocation fails,
> diff --git a/lib/backend.py b/lib/backend.py
> index bca3512..3f965a6 100644
> --- a/lib/backend.py
> +++ b/lib/backend.py
> @@ -308,7 +308,7 @@ def LeaveCluster():
>
>
>  def GetNodeInfo(vgname, hypervisor_type):
> -  """Gives back a hash with different informations about the node.
> +  """Gives back a hash with different information about the node.
>
>   @type vgname: C{string}
>   @param vgname: the name of the volume group to ask for disk space 
> information
> @@ -581,7 +581,7 @@ def GetInstanceList(hypervisor_list):
>
>
>  def GetInstanceInfo(instance, hname):
> -  """Gives back the informations about an instance as a dictionary.
> +  """Gives back the information about an instance as a dictionary.
>
>   @type instance: string
>   @param instance: the instance name
> @@ -746,7 +746,7 @@ def RunRenameInstance(instance, old_name):
>
>
>  def _GetVGInfo(vg_name):
> -  """Get informations about the volume group.
> +  """Get information about the volume group.
>
>   @type vg_name: str
>   @param vg_name: the volume group which we query
> @@ -1032,7 +1032,7 @@ def AcceptInstance(instance, info, target):
>     msg = "Failed to accept instance"
>     logging.exception(msg)
>     return (False, '%s: %s' % (msg, err))
> -  return (True, "Accept successfull")
> +  return (True, "Accept successful")
>
>
>  def FinalizeMigration(instance, info, success):
> @@ -1080,7 +1080,7 @@ def MigrateInstance(instance, target, live):
>     msg = "Failed to migrate instance"
>     logging.exception(msg)
>     return (False, "%s: %s" % (msg, err))
> -  return (True, "Migration successfull")
> +  return (True, "Migration successful")
>
>
>  def BlockdevCreate(disk, size, owner, on_primary, info):
> @@ -1273,7 +1273,7 @@ def BlockdevAssemble(disk, owner, as_primary):
>  def BlockdevShutdown(disk):
>   """Shut down a block device.
>
> -  First, if the device is assembled (Attach() is successfull), then
> +  First, if the device is assembled (Attach() is successful), then
>   the device is shutdown. Then the children of the device are
>   shutdown.
>
> @@ -1391,7 +1391,7 @@ def BlockdevGetmirrorstatus(disks):
>  def _RecursiveFindBD(disk):
>   """Check if a device is activated.
>
> -  If so, return informations about the real device.
> +  If so, return information about the real device.
>
>   @type disk: L{objects.Disk}
>   @param disk: the disk object we need to find
> @@ -1411,7 +1411,7 @@ def _RecursiveFindBD(disk):
>  def BlockdevFind(disk):
>   """Check if a device is activated.
>
> -  If it is, return informations about the real device.
> +  If it is, return information about the real device.
>
>   @type disk: L{objects.Disk}
>   @param disk: the disk to find
> @@ -2094,7 +2094,7 @@ def RemoveFileStorageDir(file_storage_dir):
>   @param file_storage_dir: the directory we should cleanup
>   @rtype: tuple (success,)
>   @return: tuple of one element, C{success}, denoting
> -      whether the operation was successfull
> +      whether the operation was successful
>
>   """
>   file_storage_dir = _TransformFileStorageDir(file_storage_dir)
> diff --git a/lib/bdev.py b/lib/bdev.py
> index 006e5e3..4971b53 100644
> --- a/lib/bdev.py
> +++ b/lib/bdev.py
> @@ -161,7 +161,7 @@ class BlockDev(object):
>     """Remove this device.
>
>     This makes sense only for some of the device types: LV and file
> -    storeage. Also note that if the device can't attach, the removal
> +    storage. Also note that if the device can't attach, the removal
>     can't be completed.
>
>     """
> @@ -444,7 +444,7 @@ class LogicalVolume(BlockDev):
>   def Assemble(self):
>     """Assemble the device.
>
> -    We alway run `lvchange -ay` on the LV to ensure it's active before
> +    We always run `lvchange -ay` on the LV to ensure it's active before
>     use, as there were cases when xenvg was not active after boot
>     (also possibly after disk issues).
>
> @@ -1258,14 +1258,14 @@ class DRBD8(BaseDRBD):
>
>
>     If sync_percent is None, it means all is ok
> -    If estimated_time is None, it means we can't esimate
> +    If estimated_time is None, it means we can't estimate
>     the time needed, otherwise it's the time left in seconds.
>
>
>     We set the is_degraded parameter to True on two conditions:
>     network not connected or local disk missing.
>
> -    We compute the ldisk parameter based on wheter we have a local
> +    We compute the ldisk parameter based on whether we have a local
>     disk or not.
>
>     @rtype: tuple
> @@ -1335,14 +1335,14 @@ class DRBD8(BaseDRBD):
>
>     ever_disconnected = _IgnoreError(self._ShutdownNet, self.minor)
>     timeout_limit = time.time() + self._NET_RECONFIG_TIMEOUT
> -    sleep_time = 0.100 # we start the retry time at 100 miliseconds
> +    sleep_time = 0.100 # we start the retry time at 100 milliseconds
>     while time.time() < timeout_limit:
>       status = self.GetProcStatus()
>       if status.is_standalone:
>         break
>       # retry the disconnect, it seems possible that due to a
>       # well-time disconnect on the peer, my disconnect command might
> -      # be ingored and forgotten
> +      # be ignored and forgotten
>       ever_disconnected = _IgnoreError(self._ShutdownNet, self.minor) or \
>                           ever_disconnected
>       time.sleep(sleep_time)
> @@ -1647,7 +1647,7 @@ class FileStorage(BlockDev):
>   def Shutdown(self):
>     """Shutdown the device.
>
> -    This is a no-op for the file type, as we don't deacivate
> +    This is a no-op for the file type, as we don't deactivate
>     the file on shutdown.
>
>     """
> diff --git a/lib/bootstrap.py b/lib/bootstrap.py
> index 0308484..171d830 100644
> --- a/lib/bootstrap.py
> +++ b/lib/bootstrap.py
> @@ -477,7 +477,7 @@ def GatherMasterVotes(node_list):
>
>   @type node_list: list
>   @param node_list: the list of nodes to query for master info; the current
> -      node wil be removed if it is in the list
> +      node will be removed if it is in the list
>   @rtype: list
>   @return: list of (node, votes)
>
> diff --git a/lib/cli.py b/lib/cli.py
> index d351f2f..03c2ab1 100644
> --- a/lib/cli.py
> +++ b/lib/cli.py
> @@ -320,7 +320,7 @@ keyval_option = KeyValOption
>  def _ParseArgs(argv, commands, aliases):
>   """Parser for the command line arguments.
>
> -  This function parses the arguements and returns the function which
> +  This function parses the arguments and returns the function which
>   must be executed together with its (modified) arguments.
>
>   @param argv: the command line
> @@ -438,10 +438,10 @@ def AskUser(text, choices=None):
>     choices = [('y', True, 'Perform the operation'),
>                ('n', False, 'Do not perform the operation')]
>   if not choices or not isinstance(choices, list):
> -    raise errors.ProgrammerError("Invalid choiches argument to AskUser")
> +    raise errors.ProgrammerError("Invalid choices argument to AskUser")
>   for entry in choices:
>     if not isinstance(entry, tuple) or len(entry) < 3 or entry[0] == '?':
> -      raise errors.ProgrammerError("Invalid choiches element to AskUser")
> +      raise errors.ProgrammerError("Invalid choices element to AskUser")
>
>   answer = choices[-1][1]
>   new_text = []
> @@ -747,7 +747,7 @@ def GenericMain(commands, override=None, aliases=None):
>   except (errors.GenericError, luxi.ProtocolError,
>           JobSubmittedException), err:
>     result, err_msg = FormatError(err)
> -    logging.exception("Error durring command processing")
> +    logging.exception("Error during command processing")
>     ToStderr(err_msg)
>
>   return result
> diff --git a/lib/cmdlib.py b/lib/cmdlib.py
> index cb3bb31..0809489 100644
> --- a/lib/cmdlib.py
> +++ b/lib/cmdlib.py
> @@ -68,7 +68,7 @@ class LogicalUnit(object):
>   def __init__(self, processor, op, context, rpc):
>     """Constructor for LogicalUnit.
>
> -    This needs to be overriden in derived classes in order to check op
> +    This needs to be overridden in derived classes in order to check op
>     validity.
>
>     """
> @@ -116,7 +116,7 @@ class LogicalUnit(object):
>     CheckPrereq, doing these separate is better because:
>
>       - ExpandNames is left as as purely a lock-related function
> -      - CheckPrereq is run after we have aquired locks (and possible
> +      - CheckPrereq is run after we have acquired locks (and possible
>         waited for them)
>
>     The function is allowed to change the self.op attribute so that
> @@ -477,7 +477,7 @@ def _BuildInstanceHookEnv(name, primary_node, 
> secondary_nodes, os_type, status,
>   @param nics: list of tuples (ip, bridge, mac) representing
>       the NICs the instance  has
>   @type disk_template: string
> - �...@param disk_template: the distk template of the instance
> + �...@param disk_template: the disk template of the instance
>   @type disks: list
>   @param disks: the list of (size, mode) pairs
>   @type bep: dict
> @@ -592,10 +592,10 @@ def _AdjustCandidatePool(lu):
>
>
>  def _CheckInstanceBridgesExist(lu, instance):
> -  """Check that the brigdes needed by an instance exist.
> +  """Check that the bridges needed by an instance exist.
>
>   """
> -  # check bridges existance
> +  # check bridges existence
>   brlist = [nic.bridge for nic in instance.nics]
>   result = lu.rpc.call_bridges_exist(instance.primary_node, brlist)
>   result.Raise()
> @@ -616,7 +616,7 @@ class LUDestroyCluster(NoHooksLU):
>
>     This checks whether the cluster is empty.
>
> -    Any errors are signalled by raising errors.OpPrereqError.
> +    Any errors are signaled by raising errors.OpPrereqError.
>
>     """
>     master = self.cfg.GetMasterNode()
> @@ -669,7 +669,7 @@ class LUVerifyCluster(LogicalUnit):
>     Test list:
>
>       - compares ganeti version
> -      - checks vg existance and size > 20G
> +      - checks vg existence and size > 20G
>       - checks config file checksum
>       - checks ssh to other nodes
>
> @@ -908,7 +908,7 @@ class LUVerifyCluster(LogicalUnit):
>           if bep[constants.BE_AUTO_BALANCE]:
>             needed_mem += bep[constants.BE_MEMORY]
>         if nodeinfo['mfree'] < needed_mem:
> -          feedback_fn("  - ERROR: not enough memory on node %s to accomodate"
> +          feedback_fn("  - ERROR: not enough memory on node %s to 
> accommodate"
>                       " failovers should node %s fail" % (node, prinode))
>           bad = True
>     return bad
> @@ -927,7 +927,7 @@ class LUVerifyCluster(LogicalUnit):
>   def BuildHooksEnv(self):
>     """Build hooks env.
>
> -    Cluster-Verify hooks just rone in the post phase and their failure makes
> +    Cluster-Verify hooks just ran in the post phase and their failure makes
>     the output be logged in the verify output and the verification to fail.
>
>     """
> @@ -1194,7 +1194,7 @@ class LUVerifyCluster(LogicalUnit):
>     return not bad
>
>   def HooksCallBack(self, phase, hooks_results, feedback_fn, lu_result):
> -    """Analize the post-hooks' result
> +    """Analyze the post-hooks' result
>
>     This method analyses the hook result, handles it, and sends some
>     nicely-formatted feedback back to the user.
> @@ -1414,7 +1414,7 @@ def _RecursiveCheckIfLVMBased(disk):
>
>   @type disk: L{objects.Disk}
>   @param disk: the disk to check
> - �...@rtype: booleean
> + �...@rtype: boolean
>   @return: boolean indicating whether a LD_LV dev_type was found or not
>
>   """
> @@ -1815,7 +1815,7 @@ class LURemoveNode(LogicalUnit):
>      - it does not have primary or secondary instances
>      - it's not the master
>
> -    Any errors are signalled by raising errors.OpPrereqError.
> +    Any errors are signaled by raising errors.OpPrereqError.
>
>     """
>     node = self.cfg.GetNodeInfo(self.cfg.ExpandNodeName(self.op.node_name))
> @@ -2136,7 +2136,7 @@ class LUAddNode(LogicalUnit):
>      - it is resolvable
>      - its parameters (single/dual homed) matches the cluster
>
> -    Any errors are signalled by raising errors.OpPrereqError.
> +    Any errors are signaled by raising errors.OpPrereqError.
>
>     """
>     node_name = self.op.node_name
> @@ -2190,7 +2190,7 @@ class LUAddNode(LogicalUnit):
>         raise errors.OpPrereqError("The master has a private ip but the"
>                                    " new node doesn't have one")
>
> -    # checks reachablity
> +    # checks reachability
>     if not utils.TcpPing(primary_ip, constants.DEFAULT_NODED_PORT):
>       raise errors.OpPrereqError("Node not reachable by ping")
>
> @@ -2866,7 +2866,7 @@ class LUStartupInstance(LogicalUnit):
>     _CheckNodeOnline(self, instance.primary_node)
>
>     bep = self.cfg.GetClusterInfo().FillBE(instance)
> -    # check bridges existance
> +    # check bridges existence
>     _CheckInstanceBridgesExist(self, instance)
>
>     remote_info = self.rpc.call_instance_info(instance.primary_node,
> @@ -2944,7 +2944,7 @@ class LURebootInstance(LogicalUnit):
>
>     _CheckNodeOnline(self, instance.primary_node)
>
> -    # check bridges existance
> +    # check bridges existence
>     _CheckInstanceBridgesExist(self, instance)
>
>   def Exec(self, feedback_fn):
> @@ -3615,7 +3615,7 @@ class LUFailoverInstance(LogicalUnit):
>       self.LogInfo("Not checking memory on the secondary node as"
>                    " instance will not be started")
>
> -    # check bridge existance
> +    # check bridge existence
>     brlist = [nic.bridge for nic in instance.nics]
>     result = self.rpc.call_bridges_exist(target_node, brlist)
>     result.Raise()
> @@ -3753,7 +3753,7 @@ class LUMigrateInstance(LogicalUnit):
>                          instance.name, i_be[constants.BE_MEMORY],
>                          instance.hypervisor)
>
> -    # check bridge existance
> +    # check bridge existence
>     brlist = [nic.bridge for nic in instance.nics]
>     result = self.rpc.call_bridges_exist(target_node, brlist)
>     if result.failed or not result.data:
> @@ -6289,7 +6289,7 @@ class LUExportInstance(LogicalUnit):
>     # remove it from its current node. In the future we could fix this by:
>     #  - making a tasklet to search (share-lock all), then create the new one,
>     #    then one to remove, after
> -    #  - removing the removal operation altoghether
> +    #  - removing the removal operation altogether
>     self.needed_locks[locking.LEVEL_NODE] = locking.ALL_SET
>
>   def DeclareLocks(self, level):
> diff --git a/lib/config.py b/lib/config.py
> index 4075156..ceea4e7 100644
> --- a/lib/config.py
> +++ b/lib/config.py
> @@ -796,7 +796,7 @@ class ConfigWriter:
>                                     self._config_data.instances.keys())
>
>   def _UnlockedGetInstanceInfo(self, instance_name):
> -    """Returns informations about an instance.
> +    """Returns information about an instance.
>
>     This function is for internal use, when the config lock is already held.
>
> @@ -808,9 +808,9 @@ class ConfigWriter:
>
>   @locking.ssynchronized(_config_lock, shared=1)
>   def GetInstanceInfo(self, instance_name):
> -    """Returns informations about an instance.
> +    """Returns information about an instance.
>
> -    It takes the information from the configuration file. Other informations 
> of
> +    It takes the information from the configuration file. Other information 
> of
>     an instance are taken from the live systems.
>
>     @param instance_name: name of the instance, e.g.
> @@ -1208,7 +1208,7 @@ class ConfigWriter:
>
>   @locking.ssynchronized(_config_lock, shared=1)
>   def GetClusterInfo(self):
> -    """Returns informations about the cluster
> +    """Returns information about the cluster
>
>     @rtype: L{objects.Cluster}
>     @return: the cluster object
> diff --git a/lib/http/__init__.py b/lib/http/__init__.py
> index 008cf9c..b7455dc 100644
> --- a/lib/http/__init__.py
> +++ b/lib/http/__init__.py
> @@ -744,7 +744,7 @@ class HttpMessageWriter(object):
>   def HasMessageBody(self):
>     """Checks whether the HTTP message contains a body.
>
> -    Can be overriden by subclasses.
> +    Can be overridden by subclasses.
>
>     """
>     return bool(self._msg.body)
> @@ -937,7 +937,7 @@ class HttpMessageReader(object):
>   def ParseStartLine(self, start_line):
>     """Parses the start line of a message.
>
> -    Must be overriden by subclass.
> +    Must be overridden by subclass.
>
>     @type start_line: string
>     @param start_line: Start line string
> diff --git a/lib/http/auth.py b/lib/http/auth.py
> index 8a8d720..ff880ac 100644
> --- a/lib/http/auth.py
> +++ b/lib/http/auth.py
> @@ -80,7 +80,7 @@ class HttpServerRequestAuthentication(object):
>   def GetAuthRealm(self, req):
>     """Returns the authentication realm for a request.
>
> -    MAY be overriden by a subclass, which then can return different realms 
> for
> +    MAY be overridden by a subclass, which then can return different realms 
> for
>     different paths. Returning "None" means no authentication is needed for a
>     request.
>
> @@ -195,7 +195,7 @@ class HttpServerRequestAuthentication(object):
>   def Authenticate(self, req, user, password):
>     """Checks the password for a user.
>
> -    This function MUST be overriden by a subclass.
> +    This function MUST be overridden by a subclass.
>
>     """
>     raise NotImplementedError()
> diff --git a/lib/http/server.py b/lib/http/server.py
> index b74eb36..95a3756 100644
> --- a/lib/http/server.py
> +++ b/lib/http/server.py
> @@ -536,14 +536,14 @@ class HttpServer(http.HttpBase):
>   def PreHandleRequest(self, req):
>     """Called before handling a request.
>
> -    Can be overriden by a subclass.
> +    Can be overridden by a subclass.
>
>     """
>
>   def HandleRequest(self, req):
>     """Handles a request.
>
> -    Must be overriden by subclass.
> +    Must be overridden by subclass.
>
>     """
>     raise NotImplementedError()
> diff --git a/lib/jqueue.py b/lib/jqueue.py
> index 3364a93..e3ce73b 100644
> --- a/lib/jqueue.py
> +++ b/lib/jqueue.py
> @@ -69,7 +69,7 @@ def TimeStampNow():
>
>
>  class _QueuedOpCode(object):
> -  """Encasulates an opcode object.
> +  """Encapsulates an opcode object.
>
>   @ivar log: holds the execution log and consists of tuples
>   of the form C{(log_serial, timestamp, level, message)}
> @@ -286,7 +286,7 @@ class _QueuedJob(object):
>     """Selectively returns the log entries.
>
>     @type newer_than: None or int
> -   �...@param newer_than: if this is None, return all log enties,
> +   �...@param newer_than: if this is None, return all log entries,
>         otherwise return only the log entries with serial higher
>         than this value
>     @rtype: list
> @@ -469,7 +469,7 @@ class _JobQueueWorkerPool(workerpool.WorkerPool):
>
>
>  class JobQueue(object):
> -  """Quue used to manaage the jobs.
> +  """Quue used to manage the jobs.
>
>   @cvar _RE_JOB_FILE: regex matching the valid job file names
>
> @@ -651,7 +651,7 @@ class JobQueue(object):
>
>     Since we aim to keep consistency should this node (the current
>     master) fail, we will log errors if our rpc fail, and especially
> -    log the case when more than half of the nodes failes.
> +    log the case when more than half of the nodes fails.
>
>     @param result: the data as returned from the rpc call
>     @type nodes: list
> @@ -934,7 +934,7 @@ class JobQueue(object):
>     and in the future we might merge them.
>
>     @type drain_flag: boolean
> -   �...@param drain_flag: wheter to set or unset the drain flag
> +   �...@param drain_flag: Whether to set or unset the drain flag
>
>     """
>     if drain_flag:
> diff --git a/lib/locking.py b/lib/locking.py
> index 647e14f..16f302e 100644
> --- a/lib/locking.py
> +++ b/lib/locking.py
> @@ -297,7 +297,7 @@ class SharedLock:
>
>
>  # Whenever we want to acquire a full LockSet we pass None as the value
> -# to acquire.  Hide this behing this nicely named constant.
> +# to acquire.  Hide this behind this nicely named constant.
>  ALL_SET = None
>
>
> @@ -689,7 +689,7 @@ BGL = 'BGL'
>  class GanetiLockManager:
>   """The Ganeti Locking Library
>
> -  The purpouse of this small library is to manage locking for ganeti clusters
> +  The purpose of this small library is to manage locking for ganeti clusters
>   in a central place, while at the same time doing dynamic checks against
>   possible deadlocks. It will also make it easier to transition to a different
>   lock type should we migrate away from python threads.
> @@ -774,7 +774,7 @@ class GanetiLockManager:
>     """Acquire a set of resource locks, at the same level.
>
>     @param level: the level at which the locks shall be acquired;
> -        it must be a memmber of LEVELS.
> +        it must be a member of LEVELS.
>     @param names: the names of the locks which shall be acquired
>         (special lock names, or instance/node names)
>     @param shared: whether to acquire in shared mode; by default
> @@ -809,7 +809,7 @@ class GanetiLockManager:
>     mode, before releasing them.
>
>     @param level: the level at which the locks shall be released;
> -        it must be a memmber of LEVELS
> +        it must be a member of LEVELS
>     @param names: the names of the locks which shall be released
>         (defaults to all the locks acquired at that level)
>
> @@ -827,7 +827,7 @@ class GanetiLockManager:
>     """Add locks at the specified level.
>
>     @param level: the level at which the locks shall be added;
> -        it must be a memmber of LEVELS_MOD.
> +        it must be a member of LEVELS_MOD.
>     @param names: names of the locks to acquire
>     @param acquired: whether to acquire the newly added locks
>     @param shared: whether the acquisition will be shared
> diff --git a/lib/luxi.py b/lib/luxi.py
> index 308de9f..85c3aa5 100644
> --- a/lib/luxi.py
> +++ b/lib/luxi.py
> @@ -191,7 +191,7 @@ class Transport:
>       raise TimeoutError("Sending timeout: %s" % str(err))
>
>   def Recv(self):
> -    """Try to receive a messae from the socket.
> +    """Try to receive a message from the socket.
>
>     In case we already have messages queued, we just return from the
>     queue. Otherwise, we try to read data with a _rwtimeout network
> diff --git a/lib/mcpu.py b/lib/mcpu.py
> index 2512762..a6014a1 100644
> --- a/lib/mcpu.py
> +++ b/lib/mcpu.py
> @@ -158,7 +158,7 @@ class Processor(object):
>           self.context.glm.add(level, add_locks, acquired=1, shared=share)
>         except errors.LockError:
>           raise errors.OpPrereqError(
> -            "Coudn't add locks (%s), probably because of a race condition"
> +            "Couldn't add locks (%s), probably because of a race condition"
>             " with another job, who added them first" % add_locks)
>       try:
>         try:
> @@ -187,7 +187,7 @@ class Processor(object):
>     @type run_notifier: callable (no arguments) or None
>     @param run_notifier:  this function (if callable) will be called when
>                           we are about to call the lu's Exec() method, that
> -                          is, after we have aquired all locks
> +                          is, after we have acquired all locks
>
>     """
>     if not isinstance(op, opcodes.OpCode):
> diff --git a/lib/rpc.py b/lib/rpc.py
> index b50c3d1..fc99fc2 100644
> --- a/lib/rpc.py
> +++ b/lib/rpc.py
> @@ -83,7 +83,7 @@ class RpcResult(object):
>   calls we can't raise an exception just because one one out of many
>   failed, and therefore we use this class to encapsulate the result.
>
> - �...@ivar data: the data payload, for successfull results, or None
> + �...@ivar data: the data payload, for successful results, or None
>   @type failed: boolean
>   @ivar failed: whether the operation failed at RPC level (not
>       application level on the remote node)
> @@ -161,7 +161,7 @@ class Client:
>   list of nodes, will contact (in parallel) all nodes, and return a
>   dict of results (key: node name, value: result).
>
> -  One current bug is that generic failure is still signalled by
> +  One current bug is that generic failure is still signaled by
>   'False' result, which is not good. This overloading of values can
>   cause bugs.
>
> @@ -220,7 +220,7 @@ class Client:
>     @return: List of RPC results
>
>     """
> -    assert _http_manager, "RPC module not intialized"
> +    assert _http_manager, "RPC module not initialized"
>
>     _http_manager.ExecRequests(self.nc.values())
>
> @@ -269,9 +269,9 @@ class RpcRunner(object):
>     @type instance: L{objects.Instance}
>     @param instance: an Instance object
>     @type hvp: dict or None
> -   �...@param hvp: a dictionary with overriden hypervisor parameters
> +   �...@param hvp: a dictionary with overridden hypervisor parameters
>     @type bep: dict or None
> -   �...@param bep: a dictionary with overriden backend parameters
> +   �...@param bep: a dictionary with overridden backend parameters
>     @rtype: dict
>     @return: the instance dict, with the hvparams filled with the
>         cluster defaults
> diff --git a/lib/ssh.py b/lib/ssh.py
> index 40df999..f0362b4 100644
> --- a/lib/ssh.py
> +++ b/lib/ssh.py
> @@ -201,7 +201,7 @@ class SshRunner:
>     connected to).
>
>     This is used to detect problems in ssh known_hosts files
> -    (conflicting known hosts) and incosistencies between dns/hosts
> +    (conflicting known hosts) and inconsistencies between dns/hosts
>     entries and local machine names
>
>     @param node: nodename of a host to check; can be short or
> diff --git a/lib/utils.py b/lib/utils.py
> index ac781fb..3d33694 100644
> --- a/lib/utils.py
> +++ b/lib/utils.py
> @@ -136,7 +136,7 @@ def RunCmd(cmd, env=None, output=None, cwd='/'):
>       directory for the command; the default will be /
>   @rtype: L{RunResult}
>   @return: RunResult instance
> - �...@raise erors.ProgrammerError: if we call this when forks are disabled
> + �...@raise errors.ProgrammerError: if we call this when forks are disabled
>
>   """
>   if no_fork:
> @@ -701,7 +701,7 @@ def IsValidIP(ip):
>   @type ip: str
>   @param ip: the address to be checked
>   @rtype: a regular expression match object
> - �...@return: a regular epression match object, or None if the
> + �...@return: a regular expression match object, or None if the
>       address is not valid
>
>   """
> @@ -734,7 +734,7 @@ def BuildShellCmd(template, *args):
>
>   This function will check all arguments in the args list so that they
>   are valid shell parameters (i.e. they don't contain shell
> -  metacharaters). If everything is ok, it will return the result of
> +  metacharacters). If everything is ok, it will return the result of
>   template % args.
>
>   @type template: str
> @@ -1063,7 +1063,7 @@ def ShellQuoteArgs(args):
>   @type args: list
>   @param args: list of arguments to be quoted
>   @rtype: str
> - �...@return: the quoted arguments concatenaned with spaces
> + �...@return: the quoted arguments concatenated with spaces
>
>   """
>   return ' '.join([ShellQuote(i) for i in args])
> @@ -1080,7 +1080,7 @@ def TcpPing(target, port, timeout=10, 
> live_port_needed=False, source=None):
>   @type port: int
>   @param port: the port to connect to
>   @type timeout: int
> - �...@param timeout: the timeout on the connection attemp
> + �...@param timeout: the timeout on the connection attempt
>   @type live_port_needed: boolean
>   @param live_port_needed: whether a closed port will cause the
>       function to return failure, as if there was a timeout
> @@ -1122,7 +1122,7 @@ def OwnIpAddress(address):
>   address.
>
>   @type address: string
> - �...@param address: the addres to check
> + �...@param address: the address to check
>   @rtype: bool
>   @return: True if we own the address
>
> @@ -1218,7 +1218,7 @@ def ReadFile(file_name, size=None):
>   @type size: None or int
>   @param size: Read at most size bytes
>   @rtype: str
> - �...@return: the (possibly partial) conent of the file
> + �...@return: the (possibly partial) content of the file
>
>   """
>   f = open(file_name, "r")
> @@ -1378,7 +1378,7 @@ def UniqueSequence(seq):
>   Element order is preserved.
>
>   @type seq: sequence
> - �...@param seq: the sequence with the source elementes
> + �...@param seq: the sequence with the source elements
>   @rtype: list
>   @return: list of unique elements from seq
>
> @@ -1390,7 +1390,7 @@ def UniqueSequence(seq):
>  def IsValidMac(mac):
>   """Predicate to check if a MAC address is valid.
>
> -  Checks wether the supplied MAC address is formally correct, only
> +  Checks whether the supplied MAC address is formally correct, only
>   accepts colon separated format.
>
>   @type mac: str
> @@ -1831,7 +1831,7 @@ def SafeEncode(text):
>
>   """
>   if isinstance(text, unicode):
> -    # onli if unicode; if str already, we handle it below
> +    # only if unicode; if str already, we handle it below
>     text = text.encode('ascii', 'backslashreplace')
>   resu = ""
>   for char in text:
> diff --git a/man/gnt-node.sgml b/man/gnt-node.sgml
> index 07b6f10..5071e5a 100644
> --- a/man/gnt-node.sgml
> +++ b/man/gnt-node.sgml
> @@ -91,7 +91,7 @@
>         discussion in <citerefentry>
>         <refentrytitle>gnt-cluster</refentrytitle>
>         <manvolnum>8</manvolnum> </citerefentry> for more
> -        informations.
> +        information.
>       </para>
>
>       <para>
> diff --git a/tools/lvmstrap b/tools/lvmstrap
> index 8af2f61..fdbf1fb 100755
> --- a/tools/lvmstrap
> +++ b/tools/lvmstrap
> @@ -267,7 +267,7 @@ def CheckSysDev(name, devnum):
>    devnum: the device number, e.g. 0x803 (2051 in decimal) for sda3
>
>   Returns:
> -    None; failure of the check is signalled by raising a
> +    None; failure of the check is signaled by raising a
>       SysconfigError exception
>   """
>
> @@ -449,7 +449,7 @@ def GetMountInfo():
>
>
>  def DevInfo(name, dev, mountinfo):
> -  """Computes miscellaneous informations about a block device.
> +  """Computes miscellaneous information about a block device.
>
>   Args:
>     name: the device name, e.g. sda
> @@ -478,7 +478,7 @@ def DevInfo(name, dev, mountinfo):
>  def ShowDiskInfo(opts):
>   """Shows a nicely formatted block device list for this system.
>
> -  This function shows the user a table with the informations gathered
> +  This function shows the user a table with the information gathered
>   by the other functions defined, in order to help the user make a
>   choice about which disks should be allocated to our volume group.
>
> --
> 1.5.4.3
>
>



-- 
olive

Reply via email to