electron: Support updating electron-bin with update.py
This commits extends the functionality of electron's update script by - making it possible to update both binary and from source built releases. - adding the `--source-only` and `--bin-only` options for limiting the update mechanism to source or binary releases. - adding a `--commit` option to automatically commit the changes for you.
This commit is contained in:
parent
251ba97441
commit
59b9b178c9
5 changed files with 696 additions and 231 deletions
|
@ -1,44 +1,11 @@
|
|||
{ callPackage }:
|
||||
{ lib, callPackage }:
|
||||
|
||||
let
|
||||
mkElectron = callPackage ./generic.nix { };
|
||||
infoJson = builtins.fromJSON (builtins.readFile ./info.json);
|
||||
in
|
||||
rec {
|
||||
electron-bin = electron_29-bin;
|
||||
|
||||
electron_24-bin = mkElectron "24.8.6" {
|
||||
armv7l-linux = "8f46901667a904a62df7043991f20dc1c2a00370a42159976855458562cda8fc";
|
||||
aarch64-linux = "599e78a3a8127828ea3fa444927e7e51035dba9811ce0d81d59ad9b0bd02b4f6";
|
||||
x86_64-linux = "61e87bbd361da101c6a8363cc9c1f8b8b51db61b076cf495d3f4424303265a96";
|
||||
x86_64-darwin = "067ce05d628b44e1393369c506268915081ac9d96c0973d367262c71dcd91078";
|
||||
aarch64-darwin = "d9093e6928b2247336b3f0811e4f66c4ae50a719ec9399c393ac9556c8e56cee";
|
||||
headers = "009p1ffh2cyn98fcmprrjzq79jysp7h565v4f54wvjxjsq2nkr97";
|
||||
};
|
||||
|
||||
electron_27-bin = mkElectron "27.3.11" {
|
||||
armv7l-linux = "012127a3edf79e0e4623a08e853286e1cba512438a0414b1ab19b75d929c1cf2";
|
||||
aarch64-linux = "ddbfcd5e04450178ca4e3113f776893454822af6757761adc792692f7978e0df";
|
||||
x86_64-linux = "e3a6f55e54e7a623bba1a15016541248408eef5a19ab82a59d19c807aab14563";
|
||||
x86_64-darwin = "357e70a1c8848d4ac7655346bec98dd18a7c0cee82452a7edf76142017779049";
|
||||
aarch64-darwin = "a687b199fcb9890f43af90ac8a4d19dc7b15522394de89e42abd5f5c6b735804";
|
||||
headers = "0vrjdvqllfyz09sw2y078mds1di219hnmska8bw8ni7j35wxr2br";
|
||||
};
|
||||
|
||||
electron_28-bin = mkElectron "28.3.1" {
|
||||
armv7l-linux = "2e22fbab2376a9bbeb8cbdd7d9bb3ca69fda6adeafa2b22ffb67157fcfcdb6ff";
|
||||
aarch64-linux = "3e46c3076041386213f7b9ebc12335889fbad5822ffc306cf7514abb88de8512";
|
||||
x86_64-linux = "e3be93e1a15d61f72e074aee021e12f20465b81f51b8c1170bd9072d7d695c3a";
|
||||
x86_64-darwin = "bd8a220fd906625ad4a8edf92e80e8eff89d51f40c22168e05090daa7c12bd66";
|
||||
aarch64-darwin = "53fc040cd09e955e013254f784cf51712029ded4a574559cf5fa19c9a911d75d";
|
||||
headers = "07iv5fh0yxv17c1akb2j4ab5xhv29d9zsgi6dm2r0n4pnf72wxwr";
|
||||
};
|
||||
|
||||
electron_29-bin = mkElectron "29.3.0" {
|
||||
armv7l-linux = "51a8b2d67ae58b01919d6eb9e8eef255cd4bb3475b3acaf58ed1b8dc2448f206";
|
||||
aarch64-linux = "bd74743eb03a77f40b65739b9ca751af264c6f428e16728d7e0332a4c94789a9";
|
||||
x86_64-linux = "7274fe2bbb2e3b71f8fc084921e22d10e529220d380a354827b274f9567261da";
|
||||
x86_64-darwin = "88873a315ddd2a70b82e83f2cb7495c0d9d7c7fb5c9ad14fcfee16af4ab89d5e";
|
||||
aarch64-darwin = "b3145bbd45007918c2365b1df59a35b4d0636222cd43eea4803580de36b9a17d";
|
||||
headers = "1smvjlgdp3ailmh0fvxj96p7cnvls19w7kdwn62v1s3xpl84b915";
|
||||
};
|
||||
}
|
||||
lib.mapAttrs' (majorVersion: info:
|
||||
lib.nameValuePair
|
||||
"electron_${majorVersion}-bin"
|
||||
(mkElectron info.version info.hashes)
|
||||
) infoJson
|
||||
|
|
46
pkgs/development/tools/electron/binary/info.json
Normal file
46
pkgs/development/tools/electron/binary/info.json
Normal file
|
@ -0,0 +1,46 @@
|
|||
{
|
||||
"24": {
|
||||
"hashes": {
|
||||
"aarch64-darwin": "d9093e6928b2247336b3f0811e4f66c4ae50a719ec9399c393ac9556c8e56cee",
|
||||
"aarch64-linux": "599e78a3a8127828ea3fa444927e7e51035dba9811ce0d81d59ad9b0bd02b4f6",
|
||||
"armv7l-linux": "8f46901667a904a62df7043991f20dc1c2a00370a42159976855458562cda8fc",
|
||||
"headers": "009p1ffh2cyn98fcmprrjzq79jysp7h565v4f54wvjxjsq2nkr97",
|
||||
"x86_64-darwin": "067ce05d628b44e1393369c506268915081ac9d96c0973d367262c71dcd91078",
|
||||
"x86_64-linux": "61e87bbd361da101c6a8363cc9c1f8b8b51db61b076cf495d3f4424303265a96"
|
||||
},
|
||||
"version": "24.8.6"
|
||||
},
|
||||
"27": {
|
||||
"hashes": {
|
||||
"aarch64-darwin": "a687b199fcb9890f43af90ac8a4d19dc7b15522394de89e42abd5f5c6b735804",
|
||||
"aarch64-linux": "ddbfcd5e04450178ca4e3113f776893454822af6757761adc792692f7978e0df",
|
||||
"armv7l-linux": "012127a3edf79e0e4623a08e853286e1cba512438a0414b1ab19b75d929c1cf2",
|
||||
"headers": "0vrjdvqllfyz09sw2y078mds1di219hnmska8bw8ni7j35wxr2br",
|
||||
"x86_64-darwin": "357e70a1c8848d4ac7655346bec98dd18a7c0cee82452a7edf76142017779049",
|
||||
"x86_64-linux": "e3a6f55e54e7a623bba1a15016541248408eef5a19ab82a59d19c807aab14563"
|
||||
},
|
||||
"version": "27.3.11"
|
||||
},
|
||||
"28": {
|
||||
"hashes": {
|
||||
"aarch64-darwin": "53fc040cd09e955e013254f784cf51712029ded4a574559cf5fa19c9a911d75d",
|
||||
"aarch64-linux": "3e46c3076041386213f7b9ebc12335889fbad5822ffc306cf7514abb88de8512",
|
||||
"armv7l-linux": "2e22fbab2376a9bbeb8cbdd7d9bb3ca69fda6adeafa2b22ffb67157fcfcdb6ff",
|
||||
"headers": "07iv5fh0yxv17c1akb2j4ab5xhv29d9zsgi6dm2r0n4pnf72wxwr",
|
||||
"x86_64-darwin": "bd8a220fd906625ad4a8edf92e80e8eff89d51f40c22168e05090daa7c12bd66",
|
||||
"x86_64-linux": "e3be93e1a15d61f72e074aee021e12f20465b81f51b8c1170bd9072d7d695c3a"
|
||||
},
|
||||
"version": "28.3.1"
|
||||
},
|
||||
"29": {
|
||||
"hashes": {
|
||||
"aarch64-darwin": "b3145bbd45007918c2365b1df59a35b4d0636222cd43eea4803580de36b9a17d",
|
||||
"aarch64-linux": "bd74743eb03a77f40b65739b9ca751af264c6f428e16728d7e0332a4c94789a9",
|
||||
"armv7l-linux": "51a8b2d67ae58b01919d6eb9e8eef255cd4bb3475b3acaf58ed1b8dc2448f206",
|
||||
"headers": "1smvjlgdp3ailmh0fvxj96p7cnvls19w7kdwn62v1s3xpl84b915",
|
||||
"x86_64-darwin": "88873a315ddd2a70b82e83f2cb7495c0d9d7c7fb5c9ad14fcfee16af4ab89d5e",
|
||||
"x86_64-linux": "7274fe2bbb2e3b71f8fc084921e22d10e529220d380a354827b274f9567261da"
|
||||
},
|
||||
"version": "29.3.0"
|
||||
}
|
||||
}
|
|
@ -1,38 +0,0 @@
|
|||
#!/usr/bin/env bash
|
||||
|
||||
set -eu -o pipefail
|
||||
|
||||
if [[ $# -lt 1 ]]; then
|
||||
echo "$0: version" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
VERSION="$1"
|
||||
|
||||
declare -A SYSTEMS HASHES
|
||||
SYSTEMS=(
|
||||
[i686-linux]=linux-ia32
|
||||
[x86_64-linux]=linux-x64
|
||||
[armv7l-linux]=linux-armv7l
|
||||
[aarch64-linux]=linux-arm64
|
||||
[x86_64-darwin]=darwin-x64
|
||||
[aarch64-darwin]=darwin-arm64
|
||||
)
|
||||
|
||||
hashfile="$(nix-prefetch-url --print-path "https://github.com/electron/electron/releases/download/v${VERSION}/SHASUMS256.txt" | tail -n1)"
|
||||
headers="$(nix-prefetch-url "https://artifacts.electronjs.org/headers/dist/v${VERSION}/node-v${VERSION}-headers.tar.gz")"
|
||||
|
||||
# Entry similar to the following goes in default.nix:
|
||||
|
||||
echo " electron_${VERSION%%.*}-bin = mkElectron \"${VERSION}\" {"
|
||||
|
||||
for S in "${!SYSTEMS[@]}"; do
|
||||
hash="$(grep " *electron-v${VERSION}-${SYSTEMS[$S]}.zip$" "$hashfile"|cut -f1 -d' ' || :)"
|
||||
if [[ -n $hash ]]; then
|
||||
echo " $S = \"$hash\";"
|
||||
fi
|
||||
done
|
||||
|
||||
echo " headers = \"$headers\";"
|
||||
|
||||
echo " };"
|
|
@ -1,81 +1,105 @@
|
|||
#! /usr/bin/env nix-shell
|
||||
#! nix-shell -i python -p python3.pkgs.joblib python3.pkgs.click python3.pkgs.click-log nix nix-prefetch-git nix-universal-prefetch prefetch-yarn-deps prefetch-npm-deps
|
||||
"""
|
||||
electron updater
|
||||
|
||||
import logging
|
||||
import click_log
|
||||
import click
|
||||
import random
|
||||
import traceback
|
||||
import csv
|
||||
A script for updating both binary and source hashes.
|
||||
|
||||
It supports the following modes:
|
||||
|
||||
| Mode | Description |
|
||||
|------------- | ----------------------------------------------- |
|
||||
| `update` | for updating a specific Electron release |
|
||||
| `update-all` | for updating all electron releases at once |
|
||||
| `eval` | just print the necessary sources to fetch |
|
||||
|
||||
The `eval` and `update` commands accept an optional `--version` flag
|
||||
to restrict the mechanism only to a given major release.
|
||||
|
||||
The `update` and `update-all` commands accept an optional `--commit`
|
||||
flag to automatically commit the changes for you.
|
||||
|
||||
The `update` and `update-all` commands accept optional `--bin-only`
|
||||
and `--source-only` flags to restict the update to binary or source
|
||||
releases.
|
||||
"""
|
||||
import base64
|
||||
import os
|
||||
import re
|
||||
import tempfile
|
||||
import subprocess
|
||||
import csv
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import random
|
||||
import re
|
||||
import subprocess
|
||||
import sys
|
||||
from joblib import Parallel, delayed, Memory
|
||||
import tempfile
|
||||
import traceback
|
||||
import urllib.request
|
||||
|
||||
from abc import ABC
|
||||
from codecs import iterdecode
|
||||
from datetime import datetime
|
||||
from typing import Iterable, Optional, Tuple
|
||||
from urllib.request import urlopen
|
||||
|
||||
os.chdir(os.path.dirname(__file__))
|
||||
import click
|
||||
import click_log
|
||||
|
||||
from joblib import Parallel, delayed, Memory
|
||||
|
||||
depot_tools_checkout = tempfile.TemporaryDirectory()
|
||||
subprocess.check_call([
|
||||
"nix-prefetch-git",
|
||||
"--builder", "--quiet",
|
||||
"--url", "https://chromium.googlesource.com/chromium/tools/depot_tools",
|
||||
"--out", depot_tools_checkout.name,
|
||||
"--rev", "7a69b031d58081d51c9e8e89557b343bba8518b1"])
|
||||
subprocess.check_call(
|
||||
[
|
||||
"nix-prefetch-git",
|
||||
"--builder",
|
||||
"--quiet",
|
||||
"--url",
|
||||
"https://chromium.googlesource.com/chromium/tools/depot_tools",
|
||||
"--out",
|
||||
depot_tools_checkout.name,
|
||||
"--rev",
|
||||
"7a69b031d58081d51c9e8e89557b343bba8518b1",
|
||||
]
|
||||
)
|
||||
sys.path.append(depot_tools_checkout.name)
|
||||
|
||||
import gclient_eval
|
||||
import gclient_utils
|
||||
|
||||
memory = Memory("cache", verbose=0)
|
||||
|
||||
@memory.cache
|
||||
def get_repo_hash(fetcher, args):
|
||||
cmd = ['nix-universal-prefetch', fetcher]
|
||||
for arg_name, arg in args.items():
|
||||
cmd.append(f'--{arg_name}')
|
||||
cmd.append(arg)
|
||||
# Relative path to the electron-source info.json
|
||||
SOURCE_INFO_JSON = "info.json"
|
||||
|
||||
print(" ".join(cmd), file=sys.stderr)
|
||||
out = subprocess.check_output(cmd)
|
||||
return out.decode('utf-8').strip()
|
||||
# Relatice path to the electron-bin info.json
|
||||
BINARY_INFO_JSON = "binary/info.json"
|
||||
|
||||
@memory.cache
|
||||
def _get_yarn_hash(file):
|
||||
print(f'prefetch-yarn-deps', file=sys.stderr)
|
||||
with tempfile.TemporaryDirectory() as tmp_dir:
|
||||
with open(tmp_dir + '/yarn.lock', 'w') as f:
|
||||
f.write(file)
|
||||
return subprocess.check_output(['prefetch-yarn-deps', tmp_dir + '/yarn.lock']).decode('utf-8').strip()
|
||||
def get_yarn_hash(repo, yarn_lock_path = 'yarn.lock'):
|
||||
return _get_yarn_hash(repo.get_file(yarn_lock_path))
|
||||
# Number of spaces used for each indentation level
|
||||
JSON_INDENT = 4
|
||||
|
||||
os.chdir(os.path.dirname(__file__))
|
||||
|
||||
memory: Memory = Memory("cache", verbose=0)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
click_log.basic_config(logger)
|
||||
|
||||
@memory.cache
|
||||
def _get_npm_hash(file):
|
||||
print(f'prefetch-npm-deps', file=sys.stderr)
|
||||
with tempfile.TemporaryDirectory() as tmp_dir:
|
||||
with open(tmp_dir + '/package-lock.json', 'w') as f:
|
||||
f.write(file)
|
||||
return subprocess.check_output(['prefetch-npm-deps', tmp_dir + '/package-lock.json']).decode('utf-8').strip()
|
||||
def get_npm_hash(repo, package_lock_path = 'package-lock.json'):
|
||||
return _get_npm_hash(repo.get_file(package_lock_path))
|
||||
|
||||
class Repo:
|
||||
def __init__(self):
|
||||
self.deps = {}
|
||||
fetcher: str
|
||||
args: dict
|
||||
|
||||
def __init__(self) -> None:
|
||||
self.deps: dict = {}
|
||||
self.hash = "sha256-AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA="
|
||||
|
||||
def get_deps(self, repo_vars, path):
|
||||
print("evaluating " + json.dumps(self, default = vars), file=sys.stderr)
|
||||
def get_deps(self, repo_vars: dict, path: str) -> None:
|
||||
print(
|
||||
"evaluating " + json.dumps(self, default=vars, sort_keys=True),
|
||||
file=sys.stderr,
|
||||
)
|
||||
|
||||
deps_file = self.get_file("DEPS")
|
||||
evaluated = gclient_eval.Parse(deps_file, filename='DEPS')
|
||||
evaluated = gclient_eval.Parse(deps_file, filename="DEPS")
|
||||
|
||||
repo_vars = dict(evaluated["vars"]) | repo_vars
|
||||
|
||||
|
@ -84,7 +108,12 @@ class Repo:
|
|||
self.deps = {
|
||||
prefix + dep_name: repo_from_dep(dep)
|
||||
for dep_name, dep in evaluated["deps"].items()
|
||||
if (gclient_eval.EvaluateCondition(dep["condition"], repo_vars) if "condition" in dep else True) and repo_from_dep(dep) != None
|
||||
if (
|
||||
gclient_eval.EvaluateCondition(dep["condition"], repo_vars)
|
||||
if "condition" in dep
|
||||
else True
|
||||
)
|
||||
and repo_from_dep(dep) != None
|
||||
}
|
||||
|
||||
for key in evaluated.get("recursedeps", []):
|
||||
|
@ -92,80 +121,244 @@ class Repo:
|
|||
if dep_path in self.deps and dep_path != "src/third_party/squirrel.mac":
|
||||
self.deps[dep_path].get_deps(repo_vars, dep_path)
|
||||
|
||||
def prefetch(self):
|
||||
def prefetch(self) -> None:
|
||||
self.hash = get_repo_hash(self.fetcher, self.args)
|
||||
|
||||
def prefetch_all(self):
|
||||
return sum([dep.prefetch_all() for [_, dep] in self.deps.items()], [delayed(self.prefetch)()])
|
||||
def prefetch_all(self) -> int:
|
||||
return sum(
|
||||
[dep.prefetch_all() for [_, dep] in self.deps.items()],
|
||||
[delayed(self.prefetch)()],
|
||||
)
|
||||
|
||||
def flatten_repr(self):
|
||||
return {
|
||||
"fetcher": self.fetcher,
|
||||
"hash": self.hash,
|
||||
**self.args
|
||||
}
|
||||
def flatten_repr(self) -> dict:
|
||||
return {"fetcher": self.fetcher, "hash": self.hash, **self.args}
|
||||
|
||||
def flatten(self, path):
|
||||
out = {
|
||||
path: self.flatten_repr()
|
||||
}
|
||||
def flatten(self, path: str) -> dict:
|
||||
out = {path: self.flatten_repr()}
|
||||
for dep_path, dep in self.deps.items():
|
||||
out |= dep.flatten(dep_path)
|
||||
return out
|
||||
|
||||
def get_file(self, filepath: str) -> str:
|
||||
raise NotImplementedError
|
||||
|
||||
|
||||
class GitRepo(Repo):
|
||||
def __init__(self, url, rev):
|
||||
def __init__(self, url: str, rev: str) -> None:
|
||||
super().__init__()
|
||||
self.fetcher = 'fetchgit'
|
||||
self.fetcher = "fetchgit"
|
||||
self.args = {
|
||||
"url": url,
|
||||
"rev": rev,
|
||||
}
|
||||
|
||||
|
||||
class GitHubRepo(Repo):
|
||||
def __init__(self, owner, repo, rev):
|
||||
def __init__(self, owner: str, repo: str, rev: str) -> None:
|
||||
super().__init__()
|
||||
self.fetcher = 'fetchFromGitHub'
|
||||
self.fetcher = "fetchFromGitHub"
|
||||
self.args = {
|
||||
"owner": owner,
|
||||
"repo": repo,
|
||||
"rev": rev,
|
||||
}
|
||||
|
||||
def get_file(self, filepath):
|
||||
return urlopen(f"https://raw.githubusercontent.com/{self.args['owner']}/{self.args['repo']}/{self.args['rev']}/{filepath}").read().decode('utf-8')
|
||||
def get_file(self, filepath: str) -> str:
|
||||
return (
|
||||
urlopen(
|
||||
f"https://raw.githubusercontent.com/{self.args['owner']}/{self.args['repo']}/{self.args['rev']}/{filepath}"
|
||||
)
|
||||
.read()
|
||||
.decode("utf-8")
|
||||
)
|
||||
|
||||
|
||||
class GitilesRepo(Repo):
|
||||
def __init__(self, url, rev):
|
||||
def __init__(self, url: str, rev: str) -> None:
|
||||
super().__init__()
|
||||
self.fetcher = 'fetchFromGitiles'
|
||||
#self.fetcher = 'fetchgit'
|
||||
self.fetcher = "fetchFromGitiles"
|
||||
# self.fetcher = 'fetchgit'
|
||||
self.args = {
|
||||
"url": url,
|
||||
"rev": rev,
|
||||
#"fetchSubmodules": "false",
|
||||
# "fetchSubmodules": "false",
|
||||
}
|
||||
|
||||
if url == "https://chromium.googlesource.com/chromium/src.git":
|
||||
self.args['postFetch'] = "rm -r $out/third_party/blink/web_tests; "
|
||||
self.args['postFetch'] += "rm -r $out/third_party/hunspell/tests; "
|
||||
self.args['postFetch'] += "rm -r $out/content/test/data; "
|
||||
self.args['postFetch'] += "rm -r $out/courgette/testdata; "
|
||||
self.args['postFetch'] += "rm -r $out/extensions/test/data; "
|
||||
self.args['postFetch'] += "rm -r $out/media/test/data; "
|
||||
self.args["postFetch"] = "rm -r $out/third_party/blink/web_tests; "
|
||||
self.args["postFetch"] += "rm -r $out/third_party/hunspell/tests; "
|
||||
self.args["postFetch"] += "rm -r $out/content/test/data; "
|
||||
self.args["postFetch"] += "rm -r $out/courgette/testdata; "
|
||||
self.args["postFetch"] += "rm -r $out/extensions/test/data; "
|
||||
self.args["postFetch"] += "rm -r $out/media/test/data; "
|
||||
|
||||
def get_file(self, filepath):
|
||||
return base64.b64decode(urlopen(f"{self.args['url']}/+/{self.args['rev']}/{filepath}?format=TEXT").read()).decode('utf-8')
|
||||
def get_file(self, filepath: str) -> str:
|
||||
return base64.b64decode(
|
||||
urlopen(
|
||||
f"{self.args['url']}/+/{self.args['rev']}/{filepath}?format=TEXT"
|
||||
).read()
|
||||
).decode("utf-8")
|
||||
|
||||
def repo_from_dep(dep):
|
||||
|
||||
class ElectronBinRepo(GitHubRepo):
|
||||
def __init__(self, owner: str, repo: str, rev: str) -> None:
|
||||
super().__init__(owner, repo, rev)
|
||||
self.systems = {
|
||||
"i686-linux": "linux-ia32",
|
||||
"x86_64-linux": "linux-x64",
|
||||
"armv7l-linux": "linux-armv7l",
|
||||
"aarch64-linux": "linux-arm64",
|
||||
"x86_64-darwin": "darwin-x64",
|
||||
"aarch64-darwin": "darwin-arm64",
|
||||
}
|
||||
|
||||
def get_shasums256(self, version: str) -> list:
|
||||
"""Returns the contents of SHASUMS256.txt"""
|
||||
try:
|
||||
called_process: subprocess.CompletedProcess = subprocess.run(
|
||||
[
|
||||
"nix-prefetch-url",
|
||||
"--print-path",
|
||||
f"https://github.com/electron/electron/releases/download/v{version}/SHASUMS256.txt",
|
||||
],
|
||||
capture_output=True,
|
||||
check=True,
|
||||
text=True,
|
||||
)
|
||||
|
||||
hash_file_path = called_process.stdout.split("\n")[1]
|
||||
|
||||
with open(hash_file_path, "r") as f:
|
||||
return f.read().split("\n")
|
||||
|
||||
except subprocess.CalledProcessError as err:
|
||||
print(err.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
def get_headers(self, version: str) -> str:
|
||||
"""Returns the hash of the release headers tarball"""
|
||||
try:
|
||||
called_process: subprocess.CompletedProcess = subprocess.run(
|
||||
[
|
||||
"nix-prefetch-url",
|
||||
f"https://artifacts.electronjs.org/headers/dist/v{version}/node-v{version}-headers.tar.gz",
|
||||
],
|
||||
capture_output=True,
|
||||
check=True,
|
||||
text=True,
|
||||
)
|
||||
return called_process.stdout.split("\n")[0]
|
||||
except subprocess.CalledProcessError as err:
|
||||
print(err.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
def get_hashes(self, major_version: str) -> dict:
|
||||
"""Returns a dictionary of hashes for a given major version"""
|
||||
m, _ = get_latest_version(major_version)
|
||||
version: str = m["version"]
|
||||
|
||||
out = {}
|
||||
out[major_version] = {
|
||||
"hashes": {},
|
||||
"version": version,
|
||||
}
|
||||
|
||||
hashes: list = self.get_shasums256(version)
|
||||
|
||||
for nix_system, electron_system in self.systems.items():
|
||||
filename = f"*electron-v{version}-{electron_system}.zip"
|
||||
if any([x.endswith(filename) for x in hashes]):
|
||||
out[major_version]["hashes"][nix_system] = [
|
||||
x.split(" ")[0] for x in hashes if x.endswith(filename)
|
||||
][0]
|
||||
out[major_version]["hashes"]["headers"] = self.get_headers(version)
|
||||
|
||||
return out
|
||||
|
||||
|
||||
# Releases that have reached end-of-life no longer receive any updates
|
||||
# and it is rather pointless trying to update those.
|
||||
#
|
||||
# https://endoflife.date/electron
|
||||
@memory.cache
|
||||
def supported_version_range() -> range:
|
||||
"""Returns a range of electron releases that have not reached end-of-life yet"""
|
||||
releases_json = json.loads(
|
||||
urlopen("https://endoflife.date/api/electron.json").read()
|
||||
)
|
||||
supported_releases = [
|
||||
int(x["cycle"])
|
||||
for x in releases_json
|
||||
if x["eol"] == False
|
||||
or datetime.strptime(x["eol"], "%Y-%m-%d") > datetime.today()
|
||||
]
|
||||
|
||||
return range(
|
||||
min(supported_releases), # incl.
|
||||
# We have also packaged the beta release in nixpkgs,
|
||||
# but it is not tracked by endoflife.date
|
||||
max(supported_releases) + 2, # excl.
|
||||
1,
|
||||
)
|
||||
|
||||
|
||||
@memory.cache
|
||||
def get_repo_hash(fetcher: str, args: dict) -> str:
|
||||
cmd = ["nix-universal-prefetch", fetcher]
|
||||
for arg_name, arg in args.items():
|
||||
cmd.append(f"--{arg_name}")
|
||||
cmd.append(arg)
|
||||
|
||||
print(" ".join(cmd), file=sys.stderr)
|
||||
out = subprocess.check_output(cmd)
|
||||
return out.decode("utf-8").strip()
|
||||
|
||||
|
||||
@memory.cache
|
||||
def _get_yarn_hash(path: str) -> str:
|
||||
print(f"prefetch-yarn-deps", file=sys.stderr)
|
||||
with tempfile.TemporaryDirectory() as tmp_dir:
|
||||
with open(tmp_dir + "/yarn.lock", "w") as f:
|
||||
f.write(path)
|
||||
return (
|
||||
subprocess.check_output(["prefetch-yarn-deps", tmp_dir + "/yarn.lock"])
|
||||
.decode("utf-8")
|
||||
.strip()
|
||||
)
|
||||
|
||||
|
||||
def get_yarn_hash(repo: Repo, yarn_lock_path: str = "yarn.lock") -> str:
|
||||
return _get_yarn_hash(repo.get_file(yarn_lock_path))
|
||||
|
||||
|
||||
@memory.cache
|
||||
def _get_npm_hash(filename: str) -> str:
|
||||
print(f"prefetch-npm-deps", file=sys.stderr)
|
||||
with tempfile.TemporaryDirectory() as tmp_dir:
|
||||
with open(tmp_dir + "/package-lock.json", "w") as f:
|
||||
f.write(filename)
|
||||
return (
|
||||
subprocess.check_output(
|
||||
["prefetch-npm-deps", tmp_dir + "/package-lock.json"]
|
||||
)
|
||||
.decode("utf-8")
|
||||
.strip()
|
||||
)
|
||||
|
||||
|
||||
def get_npm_hash(repo: Repo, package_lock_path: str = "package-lock.json") -> str:
|
||||
return _get_npm_hash(repo.get_file(package_lock_path))
|
||||
|
||||
|
||||
def repo_from_dep(dep: dict) -> Optional[Repo]:
|
||||
if "url" in dep:
|
||||
url, rev = gclient_utils.SplitUrlRevision(dep["url"])
|
||||
|
||||
search_object = re.search(r'https://github.com/(.+)/(.+?)(\.git)?$', url)
|
||||
search_object = re.search(r"https://github.com/(.+)/(.+?)(\.git)?$", url)
|
||||
if search_object:
|
||||
return GitHubRepo(search_object.group(1), search_object.group(2), rev)
|
||||
|
||||
if re.match(r'https://.+.googlesource.com', url):
|
||||
if re.match(r"https://.+.googlesource.com", url):
|
||||
return GitilesRepo(url, rev)
|
||||
|
||||
return GitRepo(url, rev)
|
||||
|
@ -173,107 +366,404 @@ def repo_from_dep(dep):
|
|||
# Not a git dependency; skip
|
||||
return None
|
||||
|
||||
def get_gn_source(repo):
|
||||
|
||||
def get_gn_source(repo: Repo) -> dict:
|
||||
gn_pattern = r"'gn_version': 'git_revision:([0-9a-f]{40})'"
|
||||
gn_commit = re.search(gn_pattern, repo.get_file("DEPS")).group(1)
|
||||
gn = subprocess.check_output([
|
||||
"nix-prefetch-git",
|
||||
"--quiet",
|
||||
"https://gn.googlesource.com/gn",
|
||||
"--rev", gn_commit
|
||||
])
|
||||
gn = json.loads(gn)
|
||||
gn_prefetch: bytes = subprocess.check_output(
|
||||
[
|
||||
"nix-prefetch-git",
|
||||
"--quiet",
|
||||
"https://gn.googlesource.com/gn",
|
||||
"--rev",
|
||||
gn_commit,
|
||||
]
|
||||
)
|
||||
gn: dict = json.loads(gn_prefetch)
|
||||
return {
|
||||
"gn": {
|
||||
"version": datetime.fromisoformat(gn["date"]).date().isoformat(),
|
||||
"url": gn["url"],
|
||||
"rev": gn["rev"],
|
||||
"hash": gn["hash"]
|
||||
"hash": gn["hash"],
|
||||
}
|
||||
}
|
||||
|
||||
def get_electron_info(major_version):
|
||||
electron_releases = json.loads(urlopen("https://releases.electronjs.org/releases.json").read())
|
||||
major_version_releases = filter(lambda item: item["version"].startswith(f"{major_version}."), electron_releases)
|
||||
|
||||
def get_latest_version(major_version: str) -> Tuple[str, str]:
|
||||
"""Returns the latest version for a given major version"""
|
||||
electron_releases: dict = json.loads(
|
||||
urlopen("https://releases.electronjs.org/releases.json").read()
|
||||
)
|
||||
major_version_releases = filter(
|
||||
lambda item: item["version"].startswith(f"{major_version}."), electron_releases
|
||||
)
|
||||
m = max(major_version_releases, key=lambda item: item["date"])
|
||||
|
||||
rev=f"v{m['version']}"
|
||||
rev = f"v{m['version']}"
|
||||
return (m, rev)
|
||||
|
||||
electron_repo = GitHubRepo("electron", "electron", rev)
|
||||
electron_repo.recurse = True
|
||||
|
||||
electron_repo.get_deps({
|
||||
f"checkout_{platform}": platform == "linux"
|
||||
for platform in ["ios", "chromeos", "android", "mac", "win", "linux"]
|
||||
}, "src/electron")
|
||||
def get_electron_bin_info(major_version: str) -> Tuple[str, str, ElectronBinRepo]:
|
||||
m, rev = get_latest_version(major_version)
|
||||
|
||||
electron_repo: ElectronBinRepo = ElectronBinRepo("electron", "electron", rev)
|
||||
return (major_version, m, electron_repo)
|
||||
|
||||
|
||||
def get_electron_info(major_version: str) -> Tuple[str, str, GitHubRepo]:
|
||||
m, rev = get_latest_version(major_version)
|
||||
|
||||
electron_repo: GitHubRepo = GitHubRepo("electron", "electron", rev)
|
||||
electron_repo.get_deps(
|
||||
{
|
||||
f"checkout_{platform}": platform == "linux"
|
||||
for platform in ["ios", "chromeos", "android", "mac", "win", "linux"]
|
||||
},
|
||||
"src/electron",
|
||||
)
|
||||
|
||||
return (major_version, m, electron_repo)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
click_log.basic_config(logger)
|
||||
|
||||
@click.group()
|
||||
def cli():
|
||||
pass
|
||||
|
||||
@cli.command("eval")
|
||||
@click.option("--version", help="The major version, e.g. '23'")
|
||||
def eval(version):
|
||||
(_, _, repo) = electron_repo = get_electron_info(version)
|
||||
tree = electron_repo.flatten("src/electron")
|
||||
print(json.dumps(tree, indent=4, default = vars))
|
||||
|
||||
def get_update(repo):
|
||||
def get_update(repo: Tuple[str, str, Repo]) -> Tuple[str, dict]:
|
||||
(major_version, m, electron_repo) = repo
|
||||
|
||||
tasks = electron_repo.prefetch_all()
|
||||
a = lambda: (
|
||||
("electron_yarn_hash", get_yarn_hash(electron_repo))
|
||||
)
|
||||
a = lambda: (("electron_yarn_hash", get_yarn_hash(electron_repo)))
|
||||
tasks.append(delayed(a)())
|
||||
a = lambda: (
|
||||
("chromium_npm_hash", get_npm_hash(electron_repo.deps["src"], "third_party/node/package-lock.json"))
|
||||
(
|
||||
"chromium_npm_hash",
|
||||
get_npm_hash(
|
||||
electron_repo.deps["src"], "third_party/node/package-lock.json"
|
||||
),
|
||||
)
|
||||
)
|
||||
tasks.append(delayed(a)())
|
||||
random.shuffle(tasks)
|
||||
|
||||
task_results = {n[0]: n[1] for n in Parallel(n_jobs=3, require='sharedmem', return_as="generator")(tasks) if n != None}
|
||||
task_results = {
|
||||
n[0]: n[1]
|
||||
for n in Parallel(n_jobs=3, require="sharedmem", return_as="generator")(tasks)
|
||||
if n != None
|
||||
}
|
||||
|
||||
tree = electron_repo.flatten("src/electron")
|
||||
|
||||
return (f"{major_version}", {
|
||||
"deps": tree,
|
||||
**{key: m[key] for key in ["version", "modules", "chrome", "node"]},
|
||||
"chromium": {
|
||||
"version": m['chrome'],
|
||||
"deps": get_gn_source(electron_repo.deps["src"])
|
||||
},
|
||||
**task_results
|
||||
})
|
||||
return (
|
||||
f"{major_version}",
|
||||
{
|
||||
"deps": tree,
|
||||
**{key: m[key] for key in ["version", "modules", "chrome", "node"]},
|
||||
"chromium": {
|
||||
"version": m["chrome"],
|
||||
"deps": get_gn_source(electron_repo.deps["src"]),
|
||||
},
|
||||
**task_results,
|
||||
},
|
||||
)
|
||||
|
||||
@cli.command("update")
|
||||
@click.option("--version", help="The major version, e.g. '23'")
|
||||
def update(version):
|
||||
|
||||
def load_info_json(path: str) -> dict:
|
||||
"""Load the contents of a JSON file
|
||||
|
||||
Args:
|
||||
path: The path to the JSON file
|
||||
|
||||
Returns: An empty dict if the path does not exist, otherwise the contents of the JSON file.
|
||||
"""
|
||||
try:
|
||||
with open('info.json', 'r') as f:
|
||||
old_info = json.loads(f.read())
|
||||
with open(path, "r") as f:
|
||||
return json.loads(f.read())
|
||||
except:
|
||||
old_info = {}
|
||||
repo = get_electron_info(version)
|
||||
update = get_update(repo)
|
||||
out = old_info | { update[0]: update[1] }
|
||||
with open('info.json', 'w') as f:
|
||||
f.write(json.dumps(out, indent=4, default = vars))
|
||||
f.write('\n')
|
||||
return {}
|
||||
|
||||
@cli.command("update-all")
|
||||
def update_all():
|
||||
repos = Parallel(n_jobs=2, require='sharedmem')(delayed(get_electron_info)(major_version) for major_version in range(28, 24, -1))
|
||||
out = {n[0]: n[1] for n in Parallel(n_jobs=2, require='sharedmem')(delayed(get_update)(repo) for repo in repos)}
|
||||
|
||||
with open('info.json', 'w') as f:
|
||||
f.write(json.dumps(out, indent=4, default = vars))
|
||||
f.write('\n')
|
||||
def save_info_json(path: str, content: dict) -> None:
|
||||
"""Saves the given info to a JSON file
|
||||
|
||||
Args:
|
||||
path: The path where the info should be saved
|
||||
content: The content to be saved as JSON.
|
||||
"""
|
||||
with open(path, "w") as f:
|
||||
f.write(json.dumps(content, indent=JSON_INDENT, default=vars, sort_keys=True))
|
||||
f.write("\n")
|
||||
|
||||
|
||||
def update_bin(major_version: str, commit: bool) -> None:
|
||||
"""Update a given electron-bin release
|
||||
|
||||
Args:
|
||||
major_version: The major version number, e.g. '27'
|
||||
commit: Whether the updater should commit the result
|
||||
"""
|
||||
package_name = f"electron_{major_version}-bin"
|
||||
print(f"Updating {package_name}")
|
||||
|
||||
electron_bin_info = get_electron_bin_info(major_version)
|
||||
(_major_version, _version, repo) = electron_bin_info
|
||||
|
||||
old_info = load_info_json(BINARY_INFO_JSON)
|
||||
new_info = repo.get_hashes(major_version)
|
||||
|
||||
out = old_info | new_info
|
||||
|
||||
save_info_json(BINARY_INFO_JSON, out)
|
||||
|
||||
old_version = (
|
||||
old_info[major_version]["version"] if major_version in old_info else None
|
||||
)
|
||||
new_version = new_info[major_version]["version"]
|
||||
if old_version == new_version:
|
||||
print(f"{package_name} is up-to-date")
|
||||
elif commit:
|
||||
commit_result(package_name, old_version, new_version, BINARY_INFO_JSON)
|
||||
|
||||
|
||||
def update_source(major_version: str, commit: bool) -> None:
|
||||
"""Update a given electron-source release
|
||||
|
||||
Args:
|
||||
major_version: The major version number, e.g. '27'
|
||||
commit: Whether the updater should commit the result
|
||||
"""
|
||||
package_name = f"electron-source.electron_{major_version}"
|
||||
print(f"Updating electron-source.electron_{major_version}")
|
||||
|
||||
old_info = load_info_json(SOURCE_INFO_JSON)
|
||||
old_version = (
|
||||
old_info[str(major_version)]["version"]
|
||||
if str(major_version) in old_info
|
||||
else None
|
||||
)
|
||||
|
||||
electron_source_info = get_electron_info(major_version)
|
||||
new_info = get_update(electron_source_info)
|
||||
out = old_info | {new_info[0]: new_info[1]}
|
||||
|
||||
save_info_json(SOURCE_INFO_JSON, out)
|
||||
|
||||
new_version = new_info[1]["version"]
|
||||
if old_version == new_version:
|
||||
print(f"{package_name} is up-to-date")
|
||||
elif commit:
|
||||
commit_result(package_name, old_version, new_version, SOURCE_INFO_JSON)
|
||||
|
||||
|
||||
def non_eol_releases(releases: Iterable[int]) -> Iterable[int]:
|
||||
"""Returns a list of releases that have not reached end-of-life yet."""
|
||||
return tuple(filter(lambda x: x in supported_version_range(), releases))
|
||||
|
||||
|
||||
def update_all_source(commit: bool) -> None:
|
||||
"""Update all eletron-source releases at once
|
||||
|
||||
Args:
|
||||
commit: Whether to commit the result
|
||||
"""
|
||||
old_info = load_info_json(SOURCE_INFO_JSON)
|
||||
|
||||
filtered_releases = non_eol_releases(tuple(map(lambda x: int(x), old_info.keys())))
|
||||
|
||||
# This might take some time
|
||||
repos = Parallel(n_jobs=2, require="sharedmem")(
|
||||
delayed(get_electron_info)(major_version) for major_version in filtered_releases
|
||||
)
|
||||
new_info = {
|
||||
n[0]: n[1]
|
||||
for n in Parallel(n_jobs=2, require="sharedmem")(
|
||||
delayed(get_update)(repo) for repo in repos
|
||||
)
|
||||
}
|
||||
|
||||
if commit:
|
||||
for major_version in filtered_releases:
|
||||
# Since the sources have been fetched at this point already,
|
||||
# fetching them again will be much faster.
|
||||
update_source(str(major_version), commit)
|
||||
else:
|
||||
out = old_info | {new_info[0]: new_info[1]}
|
||||
save_info_json(SOURCE_INFO_JSON, out)
|
||||
|
||||
|
||||
def parse_cve_numbers(tag_name: str) -> Iterable[str]:
|
||||
"""Returns mentioned CVE numbers from a given release tag"""
|
||||
cve_pattern = r"CVE-\d{4}-\d+"
|
||||
url = f"https://api.github.com/repos/electron/electron/releases/tags/{tag_name}"
|
||||
headers = {
|
||||
"Accept": "application/vnd.github+json",
|
||||
"X-GitHub-Api-Version": "2022-11-28",
|
||||
}
|
||||
request = urllib.request.Request(url=url, headers=headers)
|
||||
release_note = ""
|
||||
try:
|
||||
with urlopen(request) as response:
|
||||
release_note = json.loads(response.read().decode("utf-8"))["body"]
|
||||
except:
|
||||
print(
|
||||
f"WARN: Fetching release note for {tag_name} from GitHub failed!",
|
||||
file=sys.stderr,
|
||||
)
|
||||
|
||||
return sorted(re.findall(cve_pattern, release_note))
|
||||
|
||||
|
||||
def commit_result(
|
||||
package_name: str, old_version: Optional[str], new_version: str, path: str
|
||||
) -> None:
|
||||
"""Creates a git commit with a short description of the change
|
||||
|
||||
Args:
|
||||
package_name: The package name, e.g. `electron-source.electron-{major_version}`
|
||||
or `electron_{major_version}-bin`
|
||||
|
||||
old_version: Version number before the update.
|
||||
Can be left empty when initializing a new release.
|
||||
|
||||
new_version: Version number after the update.
|
||||
|
||||
path: Path to the lockfile to be committed
|
||||
"""
|
||||
assert (
|
||||
isinstance(package_name, str) and len(package_name) > 0
|
||||
), "Argument `package_name` cannot be empty"
|
||||
assert (
|
||||
isinstance(new_version, str) and len(new_version) > 0
|
||||
), "Argument `new_version` cannot be empty"
|
||||
|
||||
if old_version != new_version:
|
||||
major_version = new_version.split(".")[0]
|
||||
cve_fixes_text = "\n".join(
|
||||
list(
|
||||
map(lambda cve: f"- Fixes {cve}", parse_cve_numbers(f"v{new_version}"))
|
||||
)
|
||||
)
|
||||
init_msg = f"init at {new_version}"
|
||||
update_msg = f"{old_version} -> {new_version}"
|
||||
diff = f"- Diff: https://github.com/electron/electron/compare/refs/tags/v{old_version}...v{new_version}\n" if old_version != None else ""
|
||||
commit_message = f"""{package_name}: {update_msg if old_version != None else init_msg}
|
||||
|
||||
- Changelog: https://github.com/electron/electron/releases/tag/v{new_version}
|
||||
{diff}{cve_fixes_text}
|
||||
"""
|
||||
subprocess.run(
|
||||
[
|
||||
"git",
|
||||
"add",
|
||||
path,
|
||||
]
|
||||
)
|
||||
subprocess.run(
|
||||
[
|
||||
"git",
|
||||
"commit",
|
||||
"-m",
|
||||
commit_message,
|
||||
]
|
||||
)
|
||||
|
||||
|
||||
@click.group()
|
||||
def cli() -> None:
|
||||
"""A script for updating electron-bin and electron-source hashes"""
|
||||
pass
|
||||
|
||||
|
||||
@cli.command(
|
||||
"eval", help="Print the necessary sources to fetch for a given major release"
|
||||
)
|
||||
@click.option("--version", help="The major version, e.g. '23'")
|
||||
def eval(version):
|
||||
(_, _, repo) = electron_repo = get_electron_info(version)
|
||||
tree = repo.flatten("src/electron")
|
||||
print(json.dumps(tree, indent=JSON_INDENT, default=vars, sort_keys=True))
|
||||
|
||||
|
||||
@cli.command("update", help="Update a single major release")
|
||||
@click.option("-v", "--version", help="The major version, e.g. '23'")
|
||||
@click.option(
|
||||
"-b",
|
||||
"--bin-only",
|
||||
is_flag=True,
|
||||
default=False,
|
||||
help="Only update electron-bin packages",
|
||||
)
|
||||
@click.option(
|
||||
"-s",
|
||||
"--source-only",
|
||||
is_flag=True,
|
||||
default=False,
|
||||
help="Only update electron-source packages",
|
||||
)
|
||||
@click.option("-c", "--commit", is_flag=True, default=False, help="Commit the result")
|
||||
def update(version: str, bin_only: bool, source_only: bool, commit: bool) -> None:
|
||||
assert isinstance(version, str) and len(version) > 0, "version must be non-empty"
|
||||
|
||||
if bin_only and source_only:
|
||||
print(
|
||||
"Error: Omit --bin-only and --source-only if you want to update both source and binary packages.",
|
||||
file=sys.stderr,
|
||||
)
|
||||
sys.exit(1)
|
||||
|
||||
elif bin_only:
|
||||
update_bin(version, commit)
|
||||
|
||||
elif source_only:
|
||||
update_source(version, commit)
|
||||
|
||||
else:
|
||||
update_bin(version, commit)
|
||||
update_source(version, commit)
|
||||
|
||||
|
||||
@cli.command("update-all", help="Update all releases at once")
|
||||
@click.option(
|
||||
"-b",
|
||||
"--bin-only",
|
||||
is_flag=True,
|
||||
default=False,
|
||||
help="Only update electron-bin packages",
|
||||
)
|
||||
@click.option(
|
||||
"-s",
|
||||
"--source-only",
|
||||
is_flag=True,
|
||||
default=False,
|
||||
help="Only update electron-source packages",
|
||||
)
|
||||
@click.option("-c", "--commit", is_flag=True, default=False, help="Commit the result")
|
||||
def update_all(bin_only: bool, source_only: bool, commit: bool) -> None:
|
||||
# Filter out releases that have reached end-of-life
|
||||
filtered_bin_info = dict(
|
||||
filter(
|
||||
lambda entry: int(entry[0]) in supported_version_range(),
|
||||
load_info_json(BINARY_INFO_JSON).items(),
|
||||
)
|
||||
)
|
||||
|
||||
if bin_only and source_only:
|
||||
print(
|
||||
"Error: omit --bin-only and --source-only if you want to update both source and binary packages.",
|
||||
file=sys.stderr,
|
||||
)
|
||||
sys.exit(1)
|
||||
|
||||
elif bin_only:
|
||||
for major_version, _ in filtered_bin_info.items():
|
||||
update_bin(major_version, commit)
|
||||
|
||||
elif source_only:
|
||||
update_all_source(commit)
|
||||
|
||||
else:
|
||||
for major_version, _ in filtered_bin_info.items():
|
||||
update_bin(major_version, commit)
|
||||
|
||||
update_all_source(commit)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
cli()
|
||||
|
|
|
@ -18095,7 +18095,6 @@ with pkgs;
|
|||
electron-source = callPackage ../development/tools/electron { };
|
||||
|
||||
inherit (callPackages ../development/tools/electron/binary { })
|
||||
electron-bin
|
||||
electron_24-bin
|
||||
electron_27-bin
|
||||
electron_28-bin
|
||||
|
@ -18106,6 +18105,7 @@ with pkgs;
|
|||
electron_28 = if lib.meta.availableOn stdenv.hostPlatform electron-source.electron_28 then electron-source.electron_28 else electron_28-bin;
|
||||
electron_29 = if lib.meta.availableOn stdenv.hostPlatform electron-source.electron_29 then electron-source.electron_29 else electron_29-bin;
|
||||
electron = electron_29;
|
||||
electron-bin = electron_29-bin;
|
||||
|
||||
autobuild = callPackage ../development/tools/misc/autobuild { };
|
||||
|
||||
|
|
Loading…
Reference in a new issue