Various non-functional changes to a number of .bbclass files: * Spelling * Grammar * Ridiculously long lines
Signed-off-by: Robert P. J. Day <rpj...@crashcourse.ca> --- while perusing numerous .bbclass files to document a few things for myself, i figured i might as well do some proofreading. all of this is strictly non-functional changes to comment lines in .bbclass files. diff --git a/meta/classes/allarch.bbclass b/meta/classes/allarch.bbclass index c953e7c..0a64588 100644 --- a/meta/classes/allarch.bbclass +++ b/meta/classes/allarch.bbclass @@ -1,5 +1,6 @@ # -# This class is used for architecture independent recipes/data files (usally scripts) +# This class is used for architecture-independent recipes/data files, such as +# configuration files, media files or scripts. # # Expand STAGING_DIR_HOST since for cross-canadian/native/nativesdk, this will @@ -15,8 +16,8 @@ python () { # No need for virtual/libc or a cross compiler d.setVar("INHIBIT_DEFAULT_DEPS","1") - # Set these to a common set of values, we shouldn't be using them other that for WORKDIR directory - # naming anyway + # Set these to a common set of values, we shouldn't be using them + # other than for WORKDIR directory naming, anyway. d.setVar("TARGET_ARCH", "allarch") d.setVar("TARGET_OS", "linux") d.setVar("TARGET_CC_ARCH", "none") @@ -33,7 +34,7 @@ python () { # packages. d.setVar("LDFLAGS", "") - # No need to do shared library processing or debug symbol handling + # No need to do shared library processing or debug symbol handling. d.setVar("EXCLUDE_FROM_SHLIBS", "1") d.setVar("INHIBIT_PACKAGE_DEBUG_SPLIT", "1") d.setVar("INHIBIT_PACKAGE_STRIP", "1") diff --git a/meta/classes/archiver.bbclass b/meta/classes/archiver.bbclass index efd413b..7f04387 100644 --- a/meta/classes/archiver.bbclass +++ b/meta/classes/archiver.bbclass @@ -1,7 +1,7 @@ # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # -# This bbclass is used for creating archive for: +# This bbclass is used for creating archives for: # 1) original (or unpacked) source: ARCHIVER_MODE[src] = "original" # 2) patched source: ARCHIVER_MODE[src] = "patched" (default) # 3) configured source: ARCHIVER_MODE[src] = "configured" @@ -48,7 +48,6 @@ do_ar_original[dirs] = "${ARCHIVER_OUTDIR} ${ARCHIVER_WORKDIR}" # This is a convenience for the shell script to use it - python () { pn = d.getVar('PN', True) @@ -71,7 +70,7 @@ python () { d.appendVarFlag('do_deploy_archives', 'depends', ' %s:do_ar_patched' % pn) elif ar_src == "configured": # We can't use "addtask do_ar_configured after do_configure" since it - # will cause the deptask of do_populate_sysroot to run not matter what + # will cause the deptask of do_populate_sysroot to run no matter what # archives we need, so we add the depends here. d.appendVarFlag('do_ar_configured', 'depends', ' %s:do_configure' % pn) d.appendVarFlag('do_deploy_archives', 'depends', ' %s:do_ar_configured' % pn) @@ -122,7 +121,7 @@ python () { d.setVarFlag('do_unpack_and_patch', 'stamp-base-clean', flag_clean) } -# Take all the sources for a recipe and puts them in WORKDIR/archiver-work/. +# Take all the sources for a recipe and put them in WORKDIR/archiver-work/. # Files in SRC_URI are copied directly, anything that's a directory # (e.g. git repositories) is "unpacked" and then put into a tarball. python do_ar_original() { @@ -190,8 +189,7 @@ python do_ar_configured() { bb.note('Archiving the configured source...') # The libtool-native's do_configure will remove the # ${STAGING_DATADIR}/aclocal/libtool.m4, so we can't re-run the - # do_configure, we archive the already configured ${S} to - # instead of. + # do_configure, we archive the already configured ${S} instead. if d.getVar('PN', True) != 'libtool-native': # Change the WORKDIR to make do_configure run in another dir. d.setVar('WORKDIR', d.getVar('ARCHIVER_WORKDIR', True)) @@ -276,9 +274,9 @@ python do_unpack_and_patch() { # Change the WORKDIR to make do_unpack do_patch run in another dir. d.setVar('WORKDIR', d.getVar('ARCHIVER_WORKDIR', True)) - # The changed 'WORKDIR' also casued 'B' changed, create dir 'B' for the - # possibly requiring of the following tasks (such as some recipes's - # do_patch required 'B' existed). + # The changed 'WORKDIR' also caused 'B' to change, create dir 'B' for + # possibly requiring the following tasks (such as some recipe's + # do_patch requiring that 'B' exists). bb.utils.mkdirhier(d.getVar('B', True)) # The kernel source is ready after do_validate_branches diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass index 8114cf6..3c24727 100644 --- a/meta/classes/base.bbclass +++ b/meta/classes/base.bbclass @@ -68,7 +68,7 @@ def base_dep_prepend(d): # deps = "" - # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not + # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not # we need that built is the responsibility of the patch function / class, not # the application. if not d.getVar('INHIBIT_DEFAULT_DEPS'): @@ -81,8 +81,8 @@ BASEDEPENDS = "${@base_dep_prepend(d)}" DEPENDS_prepend="${BASEDEPENDS} " FILESPATH = "${@base_set_filespath(["${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files"], d)}" -# THISDIR only works properly with imediate expansion as it has to run -# in the context of the location its used (:=) +# THISDIR only works properly with immediate expansion as it has to run +# in the context of the location where it's used (:=) THISDIR = "${@os.path.dirname(d.getVar('FILE', True))}" def extra_path_elements(d): @@ -311,8 +311,8 @@ python base_eventhandler() { bb.plain('\n%s\n%s\n' % (statusheader, '\n'.join(statuslines))) # This code is to silence warnings where the SDK variables overwrite the - # target ones and we'd see dulpicate key names overwriting each other - # for various PREFERRED_PROVIDERS + # target ones and we'd see duplicate key names overwriting each other + # for various PREFERRED_PROVIDERs. if isinstance(e, bb.event.RecipePreFinalise): if e.data.getVar("TARGET_PREFIX", True) == e.data.getVar("SDK_PREFIX", True): e.data.delVar("PREFERRED_PROVIDER_virtual/${TARGET_PREFIX}binutils") @@ -465,7 +465,7 @@ python () { else: appendVar('EXTRA_OECONF', extraconf) - # If PRINC is set, try and increase the PR value by the amount specified + # If PRINC is set, try and increase the PR value by the amount specified. # The PR server is now the preferred way to handle PR changes based on # the checksum of the recipe (including bbappend). The PRINC is now # obsolete. Return a warning to the user. @@ -495,7 +495,7 @@ python () { " whitelisted in LICENSE_FLAGS_WHITELIST") # If we're building a target package we need to use fakeroot (pseudo) - # in order to capture permissions, owners, groups and special files + # in order to capture permissions, owners, groups and special files. if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d): d.setVarFlag('do_unpack', 'umask', '022') d.setVarFlag('do_configure', 'umask', '022') @@ -587,12 +587,12 @@ python () { elif "osc://" in srcuri: d.appendVarFlag('do_fetch', 'depends', ' osc-native:do_populate_sysroot') - # *.lz4 should depends on lz4-native for unpacking + # *.lz4 should depend on lz4-native for unpacking # Not endswith because of "*.patch.lz4;patch=1". Need bb.fetch.decodeurl in future if '.lz4' in srcuri: d.appendVarFlag('do_unpack', 'depends', ' lz4-native:do_populate_sysroot') - # *.xz should depends on xz-native for unpacking + # *.xz should depend on xz-native for unpacking # Not endswith because of "*.patch.xz;patch=1". Need bb.fetch.decodeurl in future if '.xz' in srcuri: d.appendVarFlag('do_unpack', 'depends', ' xz-native:do_populate_sysroot') @@ -616,7 +616,7 @@ python () { return # - # We always try to scan SRC_URI for urls with machine overrides + # We always try to scan SRC_URI for URLs with machine overrides # unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0 # override = d.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', True) diff --git a/meta/classes/blacklist.bbclass b/meta/classes/blacklist.bbclass index a0141a8..b988808 100644 --- a/meta/classes/blacklist.bbclass +++ b/meta/classes/blacklist.bbclass @@ -1,15 +1,11 @@ -# anonymous support class from originally from angstrom +# anonymous support class originally from angstrom # -# To use the blacklist, a distribution should include this -# class in the INHERIT_DISTRO -# -# No longer use ANGSTROM_BLACKLIST, instead use a table of -# recipes in PNBLACKLIST +# This class is already included by default from defaultsetup.conf. # # Features: # # * To add a package to the blacklist, set: -# PNBLACKLIST[pn] = "message" +# PNBLACKLIST[pn] = "message explaining rejection" # # Cope with PNBLACKLIST flags for multilib case diff --git a/meta/classes/boot-directdisk.bbclass b/meta/classes/boot-directdisk.bbclass index 995d3e7..8c8039b 100644 --- a/meta/classes/boot-directdisk.bbclass +++ b/meta/classes/boot-directdisk.bbclass @@ -1,19 +1,19 @@ # boot-directdisk.bbclass -# (loosly based off bootimg.bbclass Copyright (C) 2004, Advanced Micro Devices, Inc.) +# (loosely based off bootimg.bbclass Copyright (C) 2004, Advanced Micro Devices, Inc.) # # Create an image which can be placed directly onto a harddisk using dd and then # booted. # -# This uses syslinux. extlinux would have been nice but required the ext2/3 +# This uses SYSLINUX. extlinux would have been nice but required the ext2/3 # partition to be mounted. grub requires to run itself as part of the install # process. # -# The end result is a 512 boot sector populated with an MBR and partition table -# followed by an msdos fat16 partition containing syslinux and a linux kernel -# completed by the ext2/3 rootfs. +# The end result is a 512-byte boot sector populated with an MBR and +# partition table followed by an MSDOS fat16 partition containing SYSLINUX +# and a linux kernel completed by the ext2/3 rootfs. # -# We have to push the msdos parition table size > 16MB so fat 16 is used as parted -# won't touch fat12 partitions. +# We have to push the MSDOS partition table size > 16MB so fat16 is used +# as parted won't touch fat12 partitions. # External variables needed diff --git a/meta/classes/bugzilla.bbclass b/meta/classes/bugzilla.bbclass index 3fc8956..234d964 100644 --- a/meta/classes/bugzilla.bbclass +++ b/meta/classes/bugzilla.bbclass @@ -1,7 +1,7 @@ # # Small event handler to automatically open URLs and file -# bug reports at a bugzilla of your choiche -# it uses XML-RPC interface, so you must have it enabled +# bug reports at a bugzilla of your choice; +# it uses XML-RPC interface, so you must have it enabled. # # Before using you must define BUGZILLA_USER, BUGZILLA_PASS credentials, # BUGZILLA_XMLRPC - uri of xmlrpc.cgi, diff --git a/meta/classes/buildhistory.bbclass b/meta/classes/buildhistory.bbclass index 20382ce..7c6d384 100644 --- a/meta/classes/buildhistory.bbclass +++ b/meta/classes/buildhistory.bbclass @@ -24,7 +24,7 @@ sstate_install[vardepsexclude] += "buildhistory_emit_pkghistory" SSTATEPOSTINSTFUNCS[vardepvalueexclude] .= "| buildhistory_emit_pkghistory" # -# Write out metadata about this package for comparision when writing future packages +# Write out metadata about this package for comparison when writing future packages # python buildhistory_emit_pkghistory() { if not d.getVar('BB_CURRENTTASK', True) in ['packagedata', 'packagedata_setscene']: @@ -431,7 +431,8 @@ buildhistory_get_sdk_installed_target() { buildhistory_list_files() { # List the files in the specified directory, but exclude date/time etc. - # This awk script is somewhat messy, but handles where the size is not printed for device files under pseudo + # This awk script is somewhat messy, but handles where the size is + # not printed for device files under pseudo ( cd $1 && find . -printf "%M %-10u %-10g %10s %p -> %l\n" | sort -k5 | sed 's/ * -> $//' > $2 ) } @@ -490,7 +491,8 @@ ROOTFS_POSTPROCESS_COMMAND =+ " buildhistory_list_installed_image ;\ IMAGE_POSTPROCESS_COMMAND += " buildhistory_get_imageinfo ; " -# We want these to be the last run so that we get called after complementary package installation +# We want these to be the last run so that we get called after +# complementary package installation. POPULATE_SDK_POST_TARGET_COMMAND_append = " buildhistory_list_installed_sdk_target ;\ buildhistory_get_sdk_installed_target ; " POPULATE_SDK_POST_HOST_COMMAND_append = " buildhistory_list_installed_sdk_host ;\ @@ -505,7 +507,8 @@ def buildhistory_get_layers(d): return layertext def buildhistory_get_metadata_revs(d): - # We want an easily machine-readable format here, so get_layers_branch_rev isn't quite what we want + # We want an easily machine-readable format here, so get_layers_branch_rev + # isn't quite what we want layers = (d.getVar("BBLAYERS", True) or "").split() medadata_revs = ["%-17s = %s:%s" % (os.path.basename(i), \ base_get_metadata_git_branch(i, None).strip(), \ @@ -555,7 +558,8 @@ def buildhistory_get_cmdline(d): buildhistory_commit() { if [ ! -d ${BUILDHISTORY_DIR} ] ; then - # Code above that creates this dir never executed, so there can't be anything to commit + # Code above that creates this dir never executed, so there + # can't be anything to commit return fi diff --git a/meta/classes/core-image.bbclass b/meta/classes/core-image.bbclass index 1b36cba..b897c5b 100644 --- a/meta/classes/core-image.bbclass +++ b/meta/classes/core-image.bbclass @@ -7,8 +7,8 @@ LIC_FILES_CHKSUM = "file://${COREBASE}/LICENSE;md5=4d92cd373abda3937c2bc47fbc49d # IMAGE_FEATURES control content of the core reference images # -# By default we install packagegroup-core-boot and packagegroup-base packages - this gives us -# working (console only) rootfs. +# By default we install packagegroup-core-boot and packagegroup-base packages; +# this gives us working (console only) rootfs. # # Available IMAGE_FEATURES: # diff --git a/meta/classes/cross-canadian.bbclass b/meta/classes/cross-canadian.bbclass index 6da43fe..54322b4 100644 --- a/meta/classes/cross-canadian.bbclass +++ b/meta/classes/cross-canadian.bbclass @@ -1,5 +1,5 @@ # -# NOTE - When using this class the user is repsonsible for ensuring that +# NOTE - When using this class the user is responsible for ensuring that # TRANSLATED_TARGET_ARCH is added into PN. This ensures that if the TARGET_ARCH # is changed, another nativesdk xxx-canadian-cross can be installed # @@ -114,7 +114,7 @@ do_populate_sysroot[stamp-extra-info] = "" USE_NLS = "${SDKUSE_NLS}" -# We have to us TARGET_ARCH but we care about the absolute value +# We have to use TARGET_ARCH but we care about the absolute value # and not any particular tune that is enabled. TARGET_ARCH[vardepsexclude] = "TUNE_ARCH" diff --git a/meta/classes/crosssdk.bbclass b/meta/classes/crosssdk.bbclass index 261a374..27a3e2b 100644 --- a/meta/classes/crosssdk.bbclass +++ b/meta/classes/crosssdk.bbclass @@ -29,7 +29,7 @@ baselib = "lib" do_populate_sysroot[stamp-extra-info] = "" do_packagedata[stamp-extra-info] = "" -# Need to force this to ensure consitency accross architectures +# Need to force this to ensure consistency accross architectures EXTRA_OECONF_FPU = "" USE_NLS = "no" diff --git a/meta/classes/debian.bbclass b/meta/classes/debian.bbclass index 1ddb56f..2eca2db 100644 --- a/meta/classes/debian.bbclass +++ b/meta/classes/debian.bbclass @@ -1,7 +1,7 @@ -# Debian package renaming only occurs when a package is built +# Debian package renaming only occurs when a package is built. # We therefore have to make sure we build all runtime packages -# before building the current package to make the packages runtime -# depends are correct +# before building the current package to make sure the packages +# runtime depends are correct. # # Custom library package names can be defined setting # DEBIANNAME_ + pkgname to the desired name. diff --git a/meta/classes/devshell.bbclass b/meta/classes/devshell.bbclass index 41164a3..be994dd 100644 --- a/meta/classes/devshell.bbclass +++ b/meta/classes/devshell.bbclass @@ -22,8 +22,8 @@ do_devshell[nostamp] = "1" # devshell and fakeroot/pseudo need careful handling since only the final # command should run under fakeroot emulation, any X connection should -# be done as the normal user. We therfore carefully construct the envionment -# manually +# be done as the normal user. We therefore carefully construct the environment +# manually. python () { if d.getVarFlag("do_devshell", "fakeroot"): # We need to signal our code that we want fakeroot however we diff --git a/meta/classes/grub-efi.bbclass b/meta/classes/grub-efi.bbclass index 47bd35e..189b102 100644 --- a/meta/classes/grub-efi.bbclass +++ b/meta/classes/grub-efi.bbclass @@ -13,7 +13,7 @@ # ${LABELS} - a list of targets for the automatic config # ${APPEND} - an override list of append strings for each label # ${GRUB_OPTS} - additional options to add to the config, ';' delimited # (optional) -# ${GRUB_TIMEOUT} - timeout before executing the deault label (optional) +# ${GRUB_TIMEOUT} - timeout before executing the default label (optional) do_bootimg[depends] += "${MLPREFIX}grub-efi:do_deploy" do_bootdirectdisk[depends] += "${MLPREFIX}grub-efi:do_deploy" diff --git a/meta/classes/icecc.bbclass b/meta/classes/icecc.bbclass index 3ec8c06..db44703 100644 --- a/meta/classes/icecc.bbclass +++ b/meta/classes/icecc.bbclass @@ -3,26 +3,30 @@ # Stages directories with symlinks from gcc/g++ to icecc, for both # native and cross compilers. Depending on each configure or compile, # the directories are added at the head of the PATH list and ICECC_CXX -# and ICEC_CC are set. +# and ICECC_CC are set. # # For the cross compiler, creates a tar.gz of our toolchain and sets # ICECC_VERSION accordingly. # -# The class now handles all 3 different compile 'stages' (i.e native ,cross-kernel and target) creating the -# necessary environment tar.gz file to be used by the remote machines. -# It also supports meta-toolchain generation +# The class now handles all 3 different compile 'stages' (i.e native, +# cross-kernel and target) creating the necessary environment tar.gz file +# to be used by the remote machines. It also supports meta-toolchain generation. # -# If ICECC_PATH is not set in local.conf then the class will try to locate it using 'bb.utils.which' -# but nothing is sure ;) +# If ICECC_PATH is not set in local.conf then the class will try to locate it +# using 'bb.utils.which' but nothing is sure ;) # -# If ICECC_ENV_EXEC is set in local.conf, then it should point to the icecc-create-env script provided by the user -# or the default one provided by icecc-create-env.bb will be used -# (NOTE that this is a modified version of the script need it and *not the one that comes with icecc* +# If ICECC_ENV_EXEC is set in local.conf, then it should point to the +# icecc-create-env script provided by the user or the default one +# provided by icecc-create-env.bb will be used. +# (NOTE that this is a modified version of the script and *not the one +# that comes with icecc*. # -# User can specify if specific packages or packages belonging to class should not use icecc to distribute -# compile jobs to remote machines, but handled locally, by defining ICECC_USER_CLASS_BL and ICECC_USER_PACKAGE_BL -# with the appropriate values in local.conf. In addition the user can force to enable icecc for packages -# which set an empty PARALLEL_MAKE variable by defining ICECC_USER_PACKAGE_WL. +# User can specify if specific packages or packages belonging to class +# should not use icecc to distribute compile jobs to remote machines, +# but handled locally, by defining ICECC_USER_CLASS_BL and ICECC_USER_PACKAGE_BL +# with the appropriate values in local.conf. In addition, the user can +# force to enable icecc for packages # which set an empty PARALLEL_MAKE +# variable by defining ICECC_USER_PACKAGE_WL. # ######################################################################################### #Error checking is kept to minimum so double check any parameters you pass to the class @@ -33,7 +37,7 @@ BB_HASHBASE_WHITELIST += "ICECC_PARALLEL_MAKE ICECC_DISABLED ICECC_USER_PACKAGE_ ICECC_ENV_EXEC ?= "${STAGING_BINDIR_NATIVE}/icecc-create-env" def icecc_dep_prepend(d): - # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not + # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not # we need that built is the responsibility of the patch function / class, not # the application. if not d.getVar('INHIBIT_DEFAULT_DEPS'): @@ -66,7 +70,7 @@ def create_path(compilers, bb, d): if icc_is_kernel(bb, d): staging += "-kernel" - #check if the icecc path is set by the user + # Check if the icecc path is set by the user icecc = get_icecc(d) # Create the dir if necessary diff --git a/meta/classes/insane.bbclass b/meta/classes/insane.bbclass index 55bfaf2..51b1993 100644 --- a/meta/classes/insane.bbclass +++ b/meta/classes/insane.bbclass @@ -8,7 +8,7 @@ # -Check the RUNTIME path for the $TMPDIR # -Check if .la files wrongly point to workdir # -Check if .pc files wrongly point to workdir -# -Check if packages contains .debug directories or .so files +# -Check if packages contain .debug directories or .so files # where they should be in -dev or -dbg # -Check if config.log contains traces to broken autoconf tests # -Ensure that binaries in base_[bindir|sbindir|libdir] do not link @@ -24,7 +24,7 @@ QADEPENDS_class-native = "" QADEPENDS_class-nativesdk = "" QA_SANE = "True" -# Elect whether a given type of error is a warning or error, they may +# Select whether a given type of error is a warning or error, they may # have been set by other files. WARN_QA ?= "ldflags useless-rpaths rpaths staticdev libdir xorg-driver-abi \ textrel already-stripped incompatible-license files-invalid \ @@ -236,8 +236,8 @@ def package_qa_check_useless_rpaths(file, name, d, elf, messages): if m: rpath = m.group(1) if rpath_eq(rpath, libdir) or rpath_eq(rpath, base_libdir): - # The dynamic linker searches both these places anyway. There is no point in - # looking there again. + # The dynamic linker searches both these places anyway. + # There is no point in looking there again. messages["useless-rpaths"] = "%s: %s contains probably-redundant RPATH %s" % (name, package_qa_clean_path(file, d), rpath) QAPATHTEST[dev-so] = "package_qa_check_dev" diff --git a/meta/classes/kernel-grub.bbclass b/meta/classes/kernel-grub.bbclass index a63f482..b9079e9 100644 --- a/meta/classes/kernel-grub.bbclass +++ b/meta/classes/kernel-grub.bbclass @@ -4,14 +4,14 @@ # you to fall back to the original kernel as well. # # - In kernel-image's preinstall scriptlet, it backs up original kernel to avoid -# probable confliction with the new one. +# probable conflict with the new one. # # - In kernel-image's postinstall scriptlet, it modifies grub's config file to # updates the new kernel as the boot priority. # pkg_preinst_kernel-image_append () { - # Parsing confliction + # Parsing conflict [ -f "$D/boot/grub/menu.list" ] && grubcfg="$D/boot/grub/menu.list" [ -f "$D/boot/grub/grub.cfg" ] && grubcfg="$D/boot/grub/grub.cfg" if [ -n "$grubcfg" ]; then diff --git a/meta/classes/kernel-yocto.bbclass b/meta/classes/kernel-yocto.bbclass index 6010dc9..8fcc7fe 100644 --- a/meta/classes/kernel-yocto.bbclass +++ b/meta/classes/kernel-yocto.bbclass @@ -44,7 +44,7 @@ def find_kernel_feature_dirs(d): return feature_dirs -# find the master/machine source branch. In the same way that the fetcher proceses +# find the master/machine source branch. In the same way that the fetcher processes # git repositories in the SRC_URI we take the first repo found, first branch. def get_machine_branch(d, default): fetch = bb.fetch2.Fetch([], d) diff --git a/meta/classes/libc-package.bbclass b/meta/classes/libc-package.bbclass index c1bc399..2d399b8 100644 --- a/meta/classes/libc-package.bbclass +++ b/meta/classes/libc-package.bbclass @@ -1,6 +1,6 @@ # -# This class knows how to package up [e]glibc. Its shared since prebuild binary toolchains -# may need packaging and its pointless to duplicate this code. +# This class knows how to package up [e]glibc. It's shared since prebuild binary +# toolchains may need packaging and it's pointless to duplicate this code. # # Caller should set GLIBC_INTERNAL_USE_BINARY_LOCALE to one of: # "compile" - Use QEMU to generate the binary locale files diff --git a/meta/classes/license.bbclass b/meta/classes/license.bbclass index 601f561..1042405 100644 --- a/meta/classes/license.bbclass +++ b/meta/classes/license.bbclass @@ -110,7 +110,7 @@ python do_populate_lic() { copy_license_files(lic_files_paths, destdir) } -# it would be better to copy them in do_install_append, but find_license_filesa is python +# it would be better to copy them in do_install_append, but find_license_files is python python perform_packagecopy_prepend () { enabled = oe.data.typed_value('LICENSE_CREATE_PACKAGE', d) if d.getVar('CLASSOVERRIDE', True) == 'class-target' and enabled: diff --git a/meta/classes/module.bbclass b/meta/classes/module.bbclass index ad6f7af..d8450ff 100644 --- a/meta/classes/module.bbclass +++ b/meta/classes/module.bbclass @@ -26,7 +26,7 @@ module_do_install() { EXPORT_FUNCTIONS do_compile do_install -# add all splitted modules to PN RDEPENDS, PN can be empty now +# add all split modules to PN RDEPENDS, PN can be empty now KERNEL_MODULES_META_PACKAGE = "${PN}" FILES_${PN} = "" ALLOW_EMPTY_${PN} = "1" diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass index 6a552d9..77186f8 100644 --- a/meta/classes/package.bbclass +++ b/meta/classes/package.bbclass @@ -26,7 +26,7 @@ # a list of affected files in FILER{PROVIDES,DEPENDS}FLIST_pkg # # h) package_do_shlibs - Look at the shared libraries generated and autotmatically add any -# depenedencies found. Also stores the package name so anyone else using this library +# dependencies found. Also stores the package name so anyone else using this library # knows which package to depend on. # # i) package_do_pkgconfig - Keep track of which packages need and provide which .pc files @@ -288,7 +288,7 @@ def splitdebuginfo(file, debugfile, debugsrcdir, sourcefile, d): return 0 def copydebugsources(debugsrcdir, d): - # The debug src information written out to sourcefile is further procecessed + # The debug src information written out to sourcefile is further processed # and copied to the destination here. import stat @@ -814,7 +814,7 @@ python split_and_strip_files () { continue if not s: continue - # Check its an excutable + # Check it's an excutable if (s[stat.ST_MODE] & stat.S_IXUSR) or (s[stat.ST_MODE] & stat.S_IXGRP) or (s[stat.ST_MODE] & stat.S_IXOTH) \ or ((file.startswith(libdir) or file.startswith(baselibdir)) and ".so" in f): # If it's a symlink, and points to an ELF file, we capture the readlink target @@ -844,7 +844,7 @@ python split_and_strip_files () { elffiles[file] = elf_file # - # First lets process debug splitting + # First, let's process debug splitting # if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'): hardlinkmap = {} @@ -1463,7 +1463,7 @@ python package_do_shlibs() { rpath = [] p = sub.Popen([d.expand("${HOST_PREFIX}otool"), '-l', file],stdout=sub.PIPE,stderr=sub.PIPE) err, out = p.communicate() - # If returned succesfully, process stderr for results + # If returned successfully, process stderr for results if p.returncode == 0: for l in err.split("\n"): l = l.strip() @@ -1472,7 +1472,7 @@ python package_do_shlibs() { p = sub.Popen([d.expand("${HOST_PREFIX}otool"), '-L', file],stdout=sub.PIPE,stderr=sub.PIPE) err, out = p.communicate() - # If returned succesfully, process stderr for results + # If returned successfully, process stderr for results if p.returncode == 0: for l in err.split("\n"): l = l.strip() @@ -1938,9 +1938,9 @@ python do_package () { # Optimisations ########################################################################### - # Contunually rexpanding complex expressions is inefficient, particularly when - # we write to the datastore and invalidate the expansion cache. This code - # pre-expands some frequently used variables + # Continually re-expanding complex expressions is inefficient, particularly + # when we write to the datastore and invalidate the expansion cache. This + # code pre-expands some frequently used variables. def expandVar(x, d): d.setVar(x, d.getVar(x, True)) diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass index 0a32b3e..b24b91a 100644 --- a/meta/classes/package_rpm.bbclass +++ b/meta/classes/package_rpm.bbclass @@ -7,7 +7,7 @@ RPMBUILD="rpmbuild" PKGWRITEDIRRPM = "${WORKDIR}/deploy-rpms" -# Maintaining the perfile dependencies has singificant overhead when writing the +# Maintaining the per-file dependencies has significant overhead when writing the # packages. When set, this value merges them for efficiency. MERGEPERFILEDEPS = "1" @@ -171,7 +171,7 @@ python write_specfile () { depends = bb.utils.join_deps(newdeps_dict) d.setVar(varname, depends.strip()) - # We need to change the style the dependency from BB to RPM + # We need to change the style of the dependency from BB to RPM. # This needs to happen AFTER the mapping_rename_hook def print_deps(variable, tag, array, d): depends = variable @@ -635,7 +635,7 @@ python do_package_rpm () { return # Construct the spec file... - # If the spec file already exist, and has not been stored into + # If the spec file already exists, and has not been stored into # pseudo's files.db, it maybe cause rpmbuild src.rpm fail, # so remove it before doing rpmbuild src.rpm. srcname = strip_multilib(d.getVar('PN', True), d) diff --git a/meta/classes/rm_work.bbclass b/meta/classes/rm_work.bbclass index f0f6d18..7d656de 100644 --- a/meta/classes/rm_work.bbclass +++ b/meta/classes/rm_work.bbclass @@ -1,7 +1,7 @@ # # Removes source after build # -# To use it add that line to conf/local.conf: +# To use it add this line to conf/local.conf: # # INHERIT += "rm_work" # @@ -31,7 +31,7 @@ do_rm_work () { for dir in * do # Retain only logs and other files in temp, safely ignore - # failures of removing pseudo folers on NFS2/3 server. + # failures of removing pseudo folders on NFS2/3 server. if [ $dir = 'pseudo' ]; then rm -rf $dir 2> /dev/null || true elif [ $dir != 'temp' ]; then @@ -39,7 +39,7 @@ do_rm_work () { fi done - # Need to add pseudo back or subsqeuent work in this workdir + # Need to add pseudo back or subsequent work in this workdir # might fail since setscene may not rerun to recreate it mkdir -p ${WORKDIR}/pseudo/ diff --git a/meta/classes/spdx.bbclass b/meta/classes/spdx.bbclass index 55ce3af..d6302eb 100644 --- a/meta/classes/spdx.bbclass +++ b/meta/classes/spdx.bbclass @@ -1,5 +1,5 @@ # This class integrates real-time license scanning, generation of SPDX standard -# output and verifiying license info during the building process. +# output and verifying license info during the building process. # It is a combination of efforts from the OE-Core, SPDX and Fossology projects. # # For more information on FOSSology: -- ======================================================================== Robert P. J. Day Ottawa, Ontario, CANADA http://crashcourse.ca Twitter: http://twitter.com/rpjday LinkedIn: http://ca.linkedin.com/in/rpjday ======================================================================== -- _______________________________________________ Openembedded-core mailing list Openembedded-core@lists.openembedded.org http://lists.openembedded.org/mailman/listinfo/openembedded-core