v81.0-2
parent
d1dc5bfcff
commit
42d221db49
33
PKGBUILD
33
PKGBUILD
|
@ -6,8 +6,8 @@
|
|||
pkgname=librewolf
|
||||
_pkgname=LibreWolf
|
||||
# how to get ci vars instead?
|
||||
pkgver=80.0.1
|
||||
pkgrel=1
|
||||
pkgver=81.0
|
||||
pkgrel=2
|
||||
pkgdesc="Community-maintained fork of Firefox, focused on privacy, security and freedom."
|
||||
arch=(x86_64 aarch64)
|
||||
license=(MPL GPL LGPL)
|
||||
|
@ -15,21 +15,26 @@ url="https://librewolf-community.gitlab.io/"
|
|||
depends=(gtk3 libxt mime-types dbus-glib ffmpeg nss ttf-font libpulse)
|
||||
makedepends=(unzip zip diffutils yasm mesa imake inetutils xorg-server-xvfb
|
||||
autoconf2.13 rust clang llvm jack gtk2 nodejs cbindgen nasm
|
||||
python-setuptools python-psutil git binutils lld)
|
||||
python-setuptools python-psutil python-zstandard git binutils lld)
|
||||
optdepends=('networkmanager: Location detection via available WiFi networks'
|
||||
'libnotify: Notification integration'
|
||||
'pulseaudio: Audio support'
|
||||
'speech-dispatcher: Text-to-Speech'
|
||||
'hunspell-en_US: Spell checking, American English')
|
||||
options=(!emptydirs !makeflags !strip)
|
||||
_arch_svn=https://git.archlinux.org/svntogit/packages.git/plain/trunk
|
||||
source_x86_64=(https://archive.mozilla.org/pub/firefox/releases/$pkgver/source/firefox-$pkgver.source.tar.xz
|
||||
$pkgname.desktop
|
||||
"0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch::${_arch_svn}/0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch?h=packages/firefox"
|
||||
"0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch::${_arch_svn}/0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch?h=packages/firefox"
|
||||
"git+https://gitlab.com/${pkgname}-community/browser/common.git"
|
||||
"git+https://gitlab.com/${pkgname}-community/settings.git"
|
||||
"megabar.patch"
|
||||
"remove_addons.patch")
|
||||
source_aarch64=(https://archive.mozilla.org/pub/firefox/releases/$pkgver/source/firefox-$pkgver.source.tar.xz
|
||||
$pkgname.desktop
|
||||
"0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch::${_arch_svn}/0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch?h=packages/firefox"
|
||||
"0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch::${_arch_svn}/0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch?h=packages/firefox"
|
||||
"git+https://gitlab.com/${pkgname}-community/browser/common.git"
|
||||
"git+https://gitlab.com/${pkgname}-community/settings.git"
|
||||
"megabar.patch"
|
||||
|
@ -37,18 +42,22 @@ source_aarch64=(https://archive.mozilla.org/pub/firefox/releases/$pkgver/source/
|
|||
arm.patch
|
||||
https://raw.githubusercontent.com/archlinuxarm/PKGBUILDs/master/extra/firefox/build-arm-libopus.patch)
|
||||
|
||||
sha256sums_x86_64=('596b085e32a2d683ba960e161ea65c6271f90f576d4bf956e0d48e83af992c21'
|
||||
sha256sums_x86_64=('9328745012178aee5a4f47c833539f7872cc6e0f20a853568a313e60cabd1ec8'
|
||||
'0b28ba4cc2538b7756cb38945230af52e8c4659b2006262da6f3352345a8bed2'
|
||||
'c2489a4ad3bfb65c064e07180a1de9a2fbc3b1b72d6bc4cd3985484d1b6b7b29'
|
||||
'52cc26cda4117f79fae1a0ad59e1404b299191a1c53d38027ceb178dab91f3dc'
|
||||
'SKIP'
|
||||
'SKIP'
|
||||
'2bef819c55935f6c72a7aa28273ecddfce0888429a32465feb6c34a16ff1ed9c'
|
||||
'd191e65a0ce3eeba0a3171c143fc93e3ded6c29eb751b90d58a7d3bf1983aca6')
|
||||
sha256sums_aarch64=('596b085e32a2d683ba960e161ea65c6271f90f576d4bf956e0d48e83af992c21'
|
||||
'682bf4bf5d79db0080aa132235a95b25745c8ef944d2a2e1fed985489d894df5'
|
||||
'41719289b309912c4b6bc86b41594f671427979481a90c32a9d3d0bf1cdd6d44')
|
||||
sha256sums_aarch64=('9328745012178aee5a4f47c833539f7872cc6e0f20a853568a313e60cabd1ec8'
|
||||
'0b28ba4cc2538b7756cb38945230af52e8c4659b2006262da6f3352345a8bed2'
|
||||
'c2489a4ad3bfb65c064e07180a1de9a2fbc3b1b72d6bc4cd3985484d1b6b7b29'
|
||||
'52cc26cda4117f79fae1a0ad59e1404b299191a1c53d38027ceb178dab91f3dc'
|
||||
'SKIP'
|
||||
'SKIP'
|
||||
'2bef819c55935f6c72a7aa28273ecddfce0888429a32465feb6c34a16ff1ed9c'
|
||||
'd191e65a0ce3eeba0a3171c143fc93e3ded6c29eb751b90d58a7d3bf1983aca6'
|
||||
'682bf4bf5d79db0080aa132235a95b25745c8ef944d2a2e1fed985489d894df5'
|
||||
'41719289b309912c4b6bc86b41594f671427979481a90c32a9d3d0bf1cdd6d44'
|
||||
'6ca87d2ac7dc48e6f595ca49ac8151936afced30d268a831c6a064b52037f6b7'
|
||||
'2d4d91f7e35d0860225084e37ec320ca6cae669f6c9c8fe7735cdbd542e3a7c9')
|
||||
|
||||
|
@ -56,6 +65,11 @@ prepare() {
|
|||
mkdir mozbuild
|
||||
cd firefox-$pkgver
|
||||
|
||||
# https://bugs.archlinux.org/task/67978
|
||||
# https://bugzilla.mozilla.org/show_bug.cgi?id=1660901
|
||||
patch -Np1 -i ../0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch
|
||||
patch -Np1 -i ../0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch
|
||||
|
||||
cat >../mozconfig <<END
|
||||
ac_add_options --enable-application=browser
|
||||
mk_add_options MOZ_OBJDIR=${PWD@Q}/obj
|
||||
|
@ -160,6 +174,7 @@ build() {
|
|||
|
||||
export MOZ_NOSPAM=1
|
||||
export MOZBUILD_STATE_PATH="$srcdir/mozbuild"
|
||||
export MACH_USE_SYSTEM_PYTHON=1
|
||||
|
||||
# LTO needs more open files
|
||||
ulimit -n 4096
|
||||
|
|
|
@ -53,10 +53,10 @@ apt-get -y -qq install $_DEPENDENCIES;
|
|||
if [[ $CARCH == 'x86_64' ]];then
|
||||
# Installs (non-ancient) clang
|
||||
apt install -y software-properties-common apt-transport-https ca-certificates
|
||||
apt-add-repository "deb http://apt.llvm.org/xenial/ llvm-toolchain-xenial-9 main"
|
||||
apt-add-repository "deb http://apt.llvm.org/xenial/ llvm-toolchain-xenial-11 main"
|
||||
wget -O - https://apt.llvm.org/llvm-snapshot.gpg.key | apt-key add
|
||||
apt-get update
|
||||
apt-get -y install clang-9 libclang-9-dev
|
||||
apt-get -y install clang-11 libclang-11-dev
|
||||
else
|
||||
apt-get -y install clang-8 libclang-8-dev
|
||||
fi
|
||||
|
|
|
@ -99,11 +99,11 @@ else
|
|||
# ubuntu seems to recommend this
|
||||
ac_add_options --disable-elf-hack
|
||||
|
||||
export CC='clang-9'
|
||||
export CXX='clang++-9'
|
||||
export AR=llvm-ar-9
|
||||
export NM=llvm-nm-9
|
||||
export RANLIB=llvm-ranlib-9
|
||||
export CC='clang-11'
|
||||
export CXX='clang++-11'
|
||||
export AR=llvm-ar-11
|
||||
export NM=llvm-nm-11
|
||||
export RANLIB=llvm-ranlib-11
|
||||
|
||||
# probably not needed, enabled by default?
|
||||
ac_add_options --enable-optimize
|
||||
|
@ -111,6 +111,16 @@ END
|
|||
|
||||
fi
|
||||
|
||||
# some patches used by Arch upstream
|
||||
# https://bugs.archlinux.org/task/67978
|
||||
# https://bugzilla.mozilla.org/show_bug.cgi?id=1660901
|
||||
_arch_svn=https://git.archlinux.org/svntogit/packages.git/plain/trunk
|
||||
wget "${_arch_svn}/0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch?h=packages/firefox" -O 0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch
|
||||
wget "${_arch_svn}/0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch?h=packages/firefox" -O 0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch
|
||||
|
||||
patch -Np1 -i ./0002-Bug-1660901-Support-the-fstat-like-subset-of-fstatat.patch
|
||||
patch -Np1 -i ./0003-Bug-1660901-ignore-AT_NO_AUTOMOUNT-in-fstatat-system.patch
|
||||
|
||||
# hopefully the magic sauce that makes things build on 16.04 and later on work "everywhere":
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/lower-python3-requirement.patch"
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/armhf-reduce-linker-memory-use.patch"
|
||||
|
@ -118,6 +128,8 @@ patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/build-with-libstdc++-7.patch"
|
|||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/fix-armhf-webrtc-build.patch"
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/webrtc-fix-compiler-flags-for-armhf.patch"
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/python3-remove-variable-annotations.patch"
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/python3-remove-fstrings.patch"
|
||||
patch -p1 -i "${CI_PROJECT_DIR}/deb_patches/python3-remove-pep487.patch"
|
||||
|
||||
# Remove some pre-installed addons that might be questionable
|
||||
patch -p1 -i ${CI_PROJECT_DIR}/remove_addons.patch
|
||||
|
|
|
@ -33,6 +33,7 @@ fi
|
|||
|
||||
export MOZ_NOSPAM=1
|
||||
export MOZBUILD_STATE_PATH="${_MOZBUILD}"
|
||||
export MACH_USE_SYSTEM_PYTHON=1
|
||||
|
||||
if [[ $CARCH == 'aarch64' ]]; then
|
||||
export MOZ_DEBUG_FLAGS=" "
|
||||
|
|
|
@ -37,9 +37,9 @@ diff -r 53fd96ca5aa4 media/webrtc/trunk/webrtc/system_wrappers/cpu_features_linu
|
|||
- "/media/webrtc/trunk/webrtc/system_wrappers/source/cpu_features_linux.c"
|
||||
-]
|
||||
-
|
||||
if CONFIG["MOZ_DEBUG"] == "1":
|
||||
if not CONFIG["MOZ_DEBUG"]:
|
||||
|
||||
DEFINES["DYNAMIC_ANNOTATIONS_ENABLED"] = "1"
|
||||
DEFINES["DYNAMIC_ANNOTATIONS_ENABLED"] = "0"
|
||||
@@ -53,20 +49,36 @@
|
||||
DEFINES["WEBRTC_ARCH_ARM64"] = True
|
||||
DEFINES["WEBRTC_HAS_NEON"] = True
|
||||
|
|
|
@ -0,0 +1,392 @@
|
|||
Description: remove f-strings that require Python 3.6 (https://docs.python.org/3/whatsnew/3.6.html#whatsnew36-pep498)
|
||||
Author: Olivier Tilloy <olivier.tilloy@canonical.com>
|
||||
|
||||
--- a/third_party/python/glean_parser/glean_parser/lint.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/lint.py
|
||||
@@ -42,10 +42,10 @@ def _english_list(items: List[str]) -> s
|
||||
if len(items) == 0:
|
||||
return ""
|
||||
elif len(items) == 1:
|
||||
- return f"'{items[0]}'"
|
||||
+ return "'{}'".format(items[0])
|
||||
else:
|
||||
return "{}, or '{}'".format(
|
||||
- ", ".join([f"'{x}'" for x in items[:-1]]), items[-1]
|
||||
+ ", ".join(["'{}'".format(x) for x in items[:-1]]), items[-1]
|
||||
)
|
||||
|
||||
|
||||
@@ -88,10 +88,11 @@ def check_common_prefix(
|
||||
if i > 0:
|
||||
common_prefix = "_".join(first[:i])
|
||||
yield (
|
||||
- f"Within category '{category_name}', all metrics begin with "
|
||||
- f"prefix '{common_prefix}'."
|
||||
+ "Within category '{}', all metrics begin with "
|
||||
+ "prefix '{}'."
|
||||
"Remove the prefixes on the metric names and (possibly) "
|
||||
"rename the category."
|
||||
+ .format(category_name, common_prefix)
|
||||
)
|
||||
|
||||
|
||||
@@ -131,17 +132,19 @@ def check_unit_in_name(
|
||||
or unit_in_name == time_unit.name
|
||||
):
|
||||
yield (
|
||||
- f"Suffix '{unit_in_name}' is redundant with time_unit "
|
||||
- f"'{time_unit.name}'. Only include time_unit."
|
||||
+ "Suffix '{}' is redundant with time_unit "
|
||||
+ "'{}'. Only include time_unit."
|
||||
+ .format(unit_in_name, time_unit.name)
|
||||
)
|
||||
elif (
|
||||
unit_in_name in TIME_UNIT_ABBREV.keys()
|
||||
or unit_in_name in TIME_UNIT_ABBREV.values()
|
||||
):
|
||||
yield (
|
||||
- f"Suffix '{unit_in_name}' doesn't match time_unit "
|
||||
- f"'{time_unit.name}'. "
|
||||
+ "Suffix '{}' doesn't match time_unit "
|
||||
+ "'{}'. "
|
||||
"Confirm the unit is correct and only include time_unit."
|
||||
+ .format(unit_in_name, time_unit.name)
|
||||
)
|
||||
|
||||
elif memory_unit is not None:
|
||||
@@ -150,26 +153,29 @@ def check_unit_in_name(
|
||||
or unit_in_name == memory_unit.name
|
||||
):
|
||||
yield (
|
||||
- f"Suffix '{unit_in_name}' is redundant with memory_unit "
|
||||
- f"'{memory_unit.name}'. "
|
||||
+ "Suffix '{}' is redundant with memory_unit "
|
||||
+ "'{}'. "
|
||||
"Only include memory_unit."
|
||||
+ .format(unit_in_name, memory_unit.name)
|
||||
)
|
||||
elif (
|
||||
unit_in_name in MEMORY_UNIT_ABBREV.keys()
|
||||
or unit_in_name in MEMORY_UNIT_ABBREV.values()
|
||||
):
|
||||
yield (
|
||||
- f"Suffix '{unit_in_name}' doesn't match memory_unit "
|
||||
- f"{memory_unit.name}'. "
|
||||
+ "Suffix '{}' doesn't match memory_unit "
|
||||
+ "{}'. "
|
||||
"Confirm the unit is correct and only include memory_unit."
|
||||
+ .format(unit_in_name, memory_unit.name)
|
||||
)
|
||||
|
||||
elif unit is not None:
|
||||
if unit_in_name == unit:
|
||||
yield (
|
||||
- f"Suffix '{unit_in_name}' is redundant with unit param "
|
||||
- f"'{unit}'. "
|
||||
+ "Suffix '{}' is redundant with unit param "
|
||||
+ "'{}'. "
|
||||
"Only include unit."
|
||||
+ .format(unit_in_name, unit)
|
||||
)
|
||||
|
||||
|
||||
@@ -183,8 +189,9 @@ def check_category_generic(
|
||||
|
||||
if category_name in GENERIC_CATEGORIES:
|
||||
yield (
|
||||
- f"Category '{category_name}' is too generic. "
|
||||
- f"Don't use {_english_list(GENERIC_CATEGORIES)} for category names"
|
||||
+ "Category '{}' is too generic. "
|
||||
+ "Don't use {} for category names"
|
||||
+ .format(category_name, _english_list(GENERIC_CATEGORIES))
|
||||
)
|
||||
|
||||
|
||||
@@ -195,9 +202,10 @@ def check_bug_number(
|
||||
|
||||
if len(number_bugs):
|
||||
yield (
|
||||
- f"For bugs {', '.join(number_bugs)}: "
|
||||
+ "For bugs {}: "
|
||||
"Bug numbers are deprecated and should be changed to full URLs. "
|
||||
"For example, use 'http://bugzilla.mozilla.org/12345' instead of '12345'."
|
||||
+ .format(', '.join(number_bugs))
|
||||
)
|
||||
|
||||
|
||||
@@ -220,7 +228,7 @@ def check_misspelled_pings(
|
||||
for builtin in pings.RESERVED_PING_NAMES:
|
||||
distance = _hamming_distance(ping, builtin)
|
||||
if distance == 1:
|
||||
- yield f"Ping '{ping}' seems misspelled. Did you mean '{builtin}'?"
|
||||
+ yield "Ping '{}' seems misspelled. Did you mean '{}'?".format(ping, builtin)
|
||||
|
||||
|
||||
def check_user_lifetime_expiration(
|
||||
@@ -263,8 +271,9 @@ class GlinterNit:
|
||||
|
||||
def format(self):
|
||||
return (
|
||||
- f"{self.check_type.name.upper()}: {self.check_name}: "
|
||||
- f"{self.name}: {self.msg}"
|
||||
+ "{}: {}: "
|
||||
+ "{}: {}"
|
||||
+ .format(self.check_type.name.upper(), self.check_name, self.name, self.msg)
|
||||
)
|
||||
|
||||
|
||||
@@ -324,8 +333,9 @@ def lint_metrics(
|
||||
"SUPERFLUOUS_NO_LINT",
|
||||
".".join([metric.category, metric.name]),
|
||||
(
|
||||
- f"Superfluous no_lint entry '{check_name}'. "
|
||||
+ "Superfluous no_lint entry '{}'. "
|
||||
"Please remove it."
|
||||
+ .format(check_name)
|
||||
),
|
||||
CheckType.warning,
|
||||
)
|
||||
@@ -372,7 +382,7 @@ def lint_yaml_files(input_filepaths: Ite
|
||||
if len(nits):
|
||||
print("Sorry, Glean found some glinter nits:", file=file)
|
||||
for (path, p) in nits:
|
||||
- print(f"{path} ({p.line}:{p.column}) - {p.message}")
|
||||
+ print("{} ({}:{}) - {}".format(path, p.line, p.column, p.message))
|
||||
print("", file=file)
|
||||
print("Please fix the above nits to continue.", file=file)
|
||||
|
||||
--- a/third_party/python/glean_parser/glean_parser/markdown.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/markdown.py
|
||||
@@ -70,7 +70,7 @@ def metrics_docs(obj_name: str) -> str:
|
||||
if obj_name.startswith("labeled_"):
|
||||
fixedup_name += "s"
|
||||
|
||||
- return f"https://mozilla.github.io/glean/book/user/metrics/{fixedup_name}.html"
|
||||
+ return "https://mozilla.github.io/glean/book/user/metrics/{}.html".format(fixedup_name)
|
||||
|
||||
|
||||
def ping_docs(ping_name: str) -> str:
|
||||
@@ -81,7 +81,7 @@ def ping_docs(ping_name: str) -> str:
|
||||
if ping_name not in pings.RESERVED_PING_NAMES:
|
||||
return ""
|
||||
|
||||
- return f"https://mozilla.github.io/glean/book/user/pings/{ping_name}.html"
|
||||
+ return "https://mozilla.github.io/glean/book/user/pings/{}.html".format(ping_name)
|
||||
|
||||
|
||||
def if_empty(ping_name: str, custom_pings_cache: Dict[str, pings.Ping] = {}) -> bool:
|
||||
--- a/third_party/python/glean_parser/glean_parser/parser.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/parser.py
|
||||
@@ -46,7 +46,8 @@ def _update_validator(validator):
|
||||
if len(missing_properties):
|
||||
missing_properties = sorted(list(missing_properties))
|
||||
yield ValidationError(
|
||||
- f"Missing required properties: {', '.join(missing_properties)}"
|
||||
+ "Missing required properties: {}"
|
||||
+ .format(', '.join(missing_properties))
|
||||
)
|
||||
|
||||
validator.VALIDATORS["required"] = required
|
||||
@@ -65,7 +66,7 @@ def _load_file(
|
||||
return {}, None
|
||||
|
||||
if content is None:
|
||||
- yield util.format_error(filepath, "", f"'{filepath}' file can not be empty.")
|
||||
+ yield util.format_error(filepath, "", "'{}' file can not be empty.".format(filepath))
|
||||
return {}, None
|
||||
|
||||
if not isinstance(content, dict):
|
||||
@@ -76,7 +77,7 @@ def _load_file(
|
||||
|
||||
schema_key = content.get("$schema")
|
||||
if not isinstance(schema_key, str):
|
||||
- raise TypeError(f"Invalid schema key {schema_key}")
|
||||
+ raise TypeError("Invalid schema key {}".format(schema_key))
|
||||
|
||||
filetype = FILE_TYPES.get(schema_key)
|
||||
|
||||
@@ -115,7 +116,7 @@ def _get_schema(
|
||||
if schema_id not in schemas:
|
||||
raise ValueError(
|
||||
util.format_error(
|
||||
- filepath, "", f"$schema key must be one of {', '.join(schemas.keys())}",
|
||||
+ filepath, "", "$schema key must be one of {}".format(', '.join(schemas.keys())),
|
||||
)
|
||||
)
|
||||
return schemas[schema_id]
|
||||
@@ -187,7 +188,7 @@ def _instantiate_metrics(
|
||||
if not config.get("allow_reserved") and category_key.split(".")[0] == "glean":
|
||||
yield util.format_error(
|
||||
filepath,
|
||||
- f"For category '{category_key}'",
|
||||
+ "For category '{}'".format(category_key),
|
||||
"Categories beginning with 'glean' are reserved for "
|
||||
"Glean internal use.",
|
||||
)
|
||||
@@ -195,7 +196,7 @@ def _instantiate_metrics(
|
||||
all_objects.setdefault(category_key, OrderedDict())
|
||||
|
||||
if not isinstance(category_val, dict):
|
||||
- raise TypeError(f"Invalid content for {category_key}")
|
||||
+ raise TypeError("Invalid content for {}".format(category_key))
|
||||
|
||||
for metric_key, metric_val in category_val.items():
|
||||
try:
|
||||
@@ -204,7 +205,7 @@ def _instantiate_metrics(
|
||||
)
|
||||
except Exception as e:
|
||||
yield util.format_error(
|
||||
- filepath, f"On instance {category_key}.{metric_key}", str(e),
|
||||
+ filepath, "On instance {}.{}".format(category_key, metric_key), str(e),
|
||||
)
|
||||
metric_obj = None
|
||||
else:
|
||||
@@ -214,7 +215,7 @@ def _instantiate_metrics(
|
||||
):
|
||||
yield util.format_error(
|
||||
filepath,
|
||||
- f"On instance {category_key}.{metric_key}",
|
||||
+ "On instance {}.{}".format(category_key, metric_key),
|
||||
'Only internal metrics may specify "all-pings" '
|
||||
'in "send_in_pings"',
|
||||
)
|
||||
@@ -230,8 +231,9 @@ def _instantiate_metrics(
|
||||
filepath,
|
||||
"",
|
||||
(
|
||||
- f"Duplicate metric name '{category_key}.{metric_key}' "
|
||||
- f"already defined in '{already_seen}'"
|
||||
+ "Duplicate metric name '{}.{}' "
|
||||
+ "already defined in '{}'"
|
||||
+ .format(category_key, metric_key, already_seen)
|
||||
),
|
||||
)
|
||||
else:
|
||||
@@ -257,17 +259,17 @@ def _instantiate_pings(
|
||||
if ping_key in RESERVED_PING_NAMES:
|
||||
yield util.format_error(
|
||||
filepath,
|
||||
- f"For ping '{ping_key}'",
|
||||
- f"Ping uses a reserved name ({RESERVED_PING_NAMES})",
|
||||
+ "For ping '{}'".format(ping_key),
|
||||
+ "Ping uses a reserved name ({})".format(RESERVED_PING_NAMES),
|
||||
)
|
||||
continue
|
||||
if not isinstance(ping_val, dict):
|
||||
- raise TypeError(f"Invalid content for ping {ping_key}")
|
||||
+ raise TypeError("Invalid content for ping {}".format(ping_key))
|
||||
ping_val["name"] = ping_key
|
||||
try:
|
||||
ping_obj = Ping(**ping_val)
|
||||
except Exception as e:
|
||||
- yield util.format_error(filepath, f"On instance '{ping_key}'", str(e))
|
||||
+ yield util.format_error(filepath, "On instance '{}'".format(ping_key), str(e))
|
||||
continue
|
||||
|
||||
already_seen = sources.get(ping_key)
|
||||
@@ -276,8 +278,9 @@ def _instantiate_pings(
|
||||
yield util.format_error(
|
||||
filepath,
|
||||
"",
|
||||
- f"Duplicate ping name '{ping_key}' "
|
||||
- f"already defined in '{already_seen}'",
|
||||
+ "Duplicate ping name '{}' "
|
||||
+ "already defined in '{}'"
|
||||
+ .format(ping_key, already_seen),
|
||||
)
|
||||
else:
|
||||
all_objects.setdefault("pings", {})[ping_key] = ping_obj
|
||||
--- a/third_party/python/glean_parser/glean_parser/util.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/util.py
|
||||
@@ -132,7 +132,7 @@ def load_yaml_or_json(path: Path, ordere
|
||||
else:
|
||||
return yaml.load(fd, Loader=_NoDatesSafeLoader)
|
||||
else:
|
||||
- raise ValueError(f"Unknown file extension {path.suffix}")
|
||||
+ raise ValueError("Unknown file extension {}".format(path.suffix))
|
||||
|
||||
|
||||
def ensure_list(value: Any) -> Sequence[Any]:
|
||||
@@ -331,9 +331,9 @@ def format_error(filepath: Union[str, Pa
|
||||
else:
|
||||
filepath = "<string>"
|
||||
if header:
|
||||
- return f"{filepath}: {header}\n{_utils.indent(content)}"
|
||||
+ return "{}: {}\n{}".format(filepath, header, _utils.indent(content))
|
||||
else:
|
||||
- return f"{filepath}:\n{_utils.indent(content)}"
|
||||
+ return "{}:\n{}".format(filepath, _utils.indent(content))
|
||||
|
||||
|
||||
def is_expired(expires: str) -> bool:
|
||||
@@ -353,8 +353,8 @@ def is_expired(expires: str) -> bool:
|
||||
date = datetime.date.fromisoformat(expires)
|
||||
except ValueError:
|
||||
raise ValueError(
|
||||
- f"Invalid expiration date '{expires}'. "
|
||||
- "Must be of the form yyyy-mm-dd in UTC."
|
||||
+ "Invalid expiration date '{}'. "
|
||||
+ "Must be of the form yyyy-mm-dd in UTC.".format(expires)
|
||||
)
|
||||
return date <= datetime.datetime.utcnow().date()
|
||||
|
||||
--- a/third_party/python/glean_parser/tools/extract_data_categories.py
|
||||
+++ b/third_party/python/glean_parser/tools/extract_data_categories.py
|
||||
@@ -62,7 +62,7 @@ def fetch_url(url: str) -> str:
|
||||
if content is not None:
|
||||
return content
|
||||
|
||||
- print(f"Fetching {url}")
|
||||
+ print("Fetching {}".format(url))
|
||||
content = urlopen(url).read()
|
||||
cache[url] = content
|
||||
time.sleep(0.5)
|
||||
@@ -97,7 +97,7 @@ def categories_as_strings(categories: Se
|
||||
"""
|
||||
if len(categories):
|
||||
return [
|
||||
- CATEGORY_MAP.get(x, f"!!!UNKNOWN CATEGORY {x}")
|
||||
+ CATEGORY_MAP.get(x, "!!!UNKNOWN CATEGORY {}".format(x))
|
||||
for x in sorted(list(categories))
|
||||
]
|
||||
else:
|
||||
@@ -119,24 +119,24 @@ def update_lines(
|
||||
|
||||
for line in lines_iter:
|
||||
output.append(line)
|
||||
- if line.startswith(f"{category_name}:"):
|
||||
+ if line.startswith("{}:".format(category_name)):
|
||||
break
|
||||
|
||||
for line in lines_iter:
|
||||
output.append(line)
|
||||
- if line.startswith(f" {metric_name}:"):
|
||||
+ if line.startswith(" {}:".format(metric_name)):
|
||||
break
|
||||
|
||||
for line in lines_iter:
|
||||
output.append(line)
|
||||
- if line.startswith(f" data_reviews:"):
|
||||
+ if line.startswith(" data_reviews:"):
|
||||
break
|
||||
|
||||
for line in lines_iter:
|
||||
if not line.strip().startswith("- "):
|
||||
output.append(" data_sensitivity:\n")
|
||||
for data_sensitivity in data_sensitivity_values:
|
||||
- output.append(f" - {data_sensitivity}\n")
|
||||
+ output.append(" - {}\n".format(data_sensitivity))
|
||||
output.append(line)
|
||||
break
|
||||
else:
|
||||
--- a/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
@@ -133,7 +133,7 @@ class Metric:
|
||||
"""
|
||||
metric_type = metric_info["type"]
|
||||
if not isinstance(metric_type, str):
|
||||
- raise TypeError(f"Unknown metric type {metric_type}")
|
||||
+ raise TypeError("Unknown metric type {}".format(metric_type))
|
||||
return cls.metric_types[metric_type](
|
||||
category=category,
|
||||
name=name,
|
|
@ -0,0 +1,23 @@
|
|||
Description: revert a change that requires Python 3.6 (https://docs.python.org/3/whatsnew/3.6.html#whatsnew36-pep487)
|
||||
Author: Olivier Tilloy <olivier.tilloy@canonical.com>
|
||||
|
||||
--- a/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
@@ -15,6 +15,8 @@ from typing import Any, Dict, List, Opti
|
||||
from . import pings
|
||||
from . import util
|
||||
|
||||
+import pep487
|
||||
+
|
||||
|
||||
# Important: if the values are ever changing here, make sure
|
||||
# to also fix mozilla/glean. Otherwise language bindings may
|
||||
@@ -32,7 +34,7 @@ class DataSensitivity(enum.Enum):
|
||||
highly_sensitive = 4
|
||||
|
||||
|
||||
-class Metric:
|
||||
+class Metric(pep487.PEP487Object):
|
||||
typename = "ERROR"
|
||||
glean_internal_metric_cat = "glean.internal.metrics"
|
||||
metric_types = {}
|
|
@ -52,3 +52,97 @@ Author: Olivier Tilloy <olivier.tilloy@canonical.com>
|
|||
def __init__(self):
|
||||
self.root_node = Node(None, is_root_node=True)
|
||||
self.end_node = Node(None, is_end_node=True)
|
||||
--- a/third_party/python/glean_parser/glean_parser/lint.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/lint.py
|
||||
@@ -237,9 +237,7 @@ def check_user_lifetime_expiration(
|
||||
|
||||
# The checks that operate on an entire category of metrics:
|
||||
# {NAME: (function, is_error)}
|
||||
-CATEGORY_CHECKS: Dict[
|
||||
- str, Tuple[Callable[[str, Iterable[metrics.Metric]], LintGenerator], CheckType]
|
||||
-] = {
|
||||
+CATEGORY_CHECKS = {
|
||||
"COMMON_PREFIX": (check_common_prefix, CheckType.error),
|
||||
"CATEGORY_GENERIC": (check_category_generic, CheckType.error),
|
||||
}
|
||||
@@ -247,9 +245,7 @@ CATEGORY_CHECKS: Dict[
|
||||
|
||||
# The checks that operate on individual metrics:
|
||||
# {NAME: (function, is_error)}
|
||||
-INDIVIDUAL_CHECKS: Dict[
|
||||
- str, Tuple[Callable[[metrics.Metric, dict], LintGenerator], CheckType]
|
||||
-] = {
|
||||
+INDIVIDUAL_CHECKS = {
|
||||
"UNIT_IN_NAME": (check_unit_in_name, CheckType.error),
|
||||
"BUG_NUMBER": (check_bug_number, CheckType.error),
|
||||
"BASELINE_PING": (check_valid_in_baseline, CheckType.error),
|
||||
@@ -282,7 +278,7 @@ def lint_metrics(
|
||||
:param file: The stream to write errors to.
|
||||
:returns: List of nits.
|
||||
"""
|
||||
- nits: List[GlinterNit] = []
|
||||
+ nits = []
|
||||
for (category_name, category) in sorted(list(objs.items())):
|
||||
if category_name == "pings":
|
||||
continue
|
||||
@@ -363,7 +359,7 @@ def lint_yaml_files(input_filepaths: Ite
|
||||
|
||||
# Generic type since the actual type comes from yamllint, which we don't
|
||||
# control.
|
||||
- nits: List = []
|
||||
+ nits = []
|
||||
for path in input_filepaths:
|
||||
# yamllint needs both the file content and the path.
|
||||
file_content = None
|
||||
--- a/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/metrics.py
|
||||
@@ -33,10 +33,10 @@ class DataSensitivity(enum.Enum):
|
||||
|
||||
|
||||
class Metric:
|
||||
- typename: str = "ERROR"
|
||||
- glean_internal_metric_cat: str = "glean.internal.metrics"
|
||||
- metric_types: Dict[str, Any] = {}
|
||||
- default_store_names: List[str] = ["metrics"]
|
||||
+ typename = "ERROR"
|
||||
+ glean_internal_metric_cat = "glean.internal.metrics"
|
||||
+ metric_types = {}
|
||||
+ default_store_names = ["metrics"]
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
--- a/third_party/python/glean_parser/glean_parser/pings.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/pings.py
|
||||
@@ -49,7 +49,7 @@ class Ping:
|
||||
# _validated indicates whether this metric has already been jsonschema
|
||||
# validated (but not any of the Python-level validation).
|
||||
if not _validated:
|
||||
- data: Dict[str, util.JSONType] = {
|
||||
+ data = {
|
||||
"$schema": parser.PINGS_ID,
|
||||
self.name: self.serialize(),
|
||||
}
|
||||
--- a/third_party/python/glean_parser/glean_parser/util.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/util.py
|
||||
@@ -269,7 +269,7 @@ def fetch_remote_url(url: str, cache: bo
|
||||
if key in dc:
|
||||
return dc[key]
|
||||
|
||||
- contents: str = urllib.request.urlopen(url).read()
|
||||
+ contents = urllib.request.urlopen(url).read()
|
||||
|
||||
if cache:
|
||||
with diskcache.Cache(cache_dir) as dc:
|
||||
--- a/third_party/python/glean_parser/glean_parser/parser.py
|
||||
+++ b/third_party/python/glean_parser/glean_parser/parser.py
|
||||
@@ -339,8 +339,8 @@ def parse_objects(
|
||||
value from the `metrics.yaml`, rather than having it overridden when
|
||||
the metric expires.
|
||||
"""
|
||||
- all_objects: ObjectTree = OrderedDict()
|
||||
- sources: Dict[Any, Path] = {}
|
||||
+ all_objects = OrderedDict()
|
||||
+ sources = {}
|
||||
filepaths = util.ensure_list(filepaths)
|
||||
for filepath in filepaths:
|
||||
content, filetype = yield from _load_file(filepath)
|
||||
|
|
|
@ -83,7 +83,7 @@ index 6b8cac0..54648c6 100644
|
|||
@@ -48,8 +48,8 @@
|
||||
display: block;
|
||||
text-shadow: none;
|
||||
overflow: -moz-hidden-unscrollable;
|
||||
overflow: clip;
|
||||
- margin-inline: @urlbarViewMarginInline@;
|
||||
- width: calc(100% - 2 * @urlbarViewMarginInline@);
|
||||
+ margin-inline: 0;
|
||||
|
|
|
@ -61,14 +61,10 @@ diff --git a/browser/locales/l10n.toml b/browser/locales/l10n.toml
|
|||
index 0b53c52..f8b605d 100644
|
||||
--- a/browser/locales/l10n.toml
|
||||
+++ b/browser/locales/l10n.toml
|
||||
@@ -129,14 +129,6 @@ locales = [
|
||||
@@ -133,10 +133,6 @@ locales = [
|
||||
reference = "browser/extensions/fxmonitor/locales/en-US/**"
|
||||
l10n = "{l}browser/extensions/fxmonitor/**"
|
||||
|
||||
-[[paths]]
|
||||
- reference = "browser/extensions/webcompat-reporter/locales/en-US/**"
|
||||
- l10n = "{l}browser/extensions/webcompat-reporter/**"
|
||||
-
|
||||
-[[paths]]
|
||||
- reference = "browser/extensions/report-site-issue/locales/en-US/**"
|
||||
- l10n = "{l}browser/extensions/report-site-issue/**"
|
||||
|
|
Loading…
Reference in New Issue