Merge remote-tracking branch 'origin/master' into staging-next

main
Martin Weinelt 2 years ago
commit fa1bb283d8
  1. 11
      nixos/modules/config/nsswitch.nix
  2. 1
      nixos/modules/module-list.nix
  3. 4
      nixos/modules/system/boot/stage-1.nix
  4. 32
      nixos/modules/system/boot/systemd/initrd-mdraid.nix
  5. 3
      nixos/modules/system/boot/systemd/initrd.nix
  6. 44
      nixos/modules/tasks/swraid.nix
  7. 2
      nixos/tests/all-tests.nix
  8. 8
      nixos/tests/systemd-initrd-swraid.nix
  9. 4
      pkgs/applications/audio/cmus/default.nix
  10. 12
      pkgs/applications/editors/neovim/neovide/default.nix
  11. 4
      pkgs/applications/editors/setzer/default.nix
  12. 12
      pkgs/applications/editors/vscode/generic.nix
  13. 11
      pkgs/applications/misc/megacmd/default.nix
  14. 6
      pkgs/applications/networking/browsers/chromium/upstream-info.json
  15. 20
      pkgs/applications/networking/nextcloud-client/default.nix
  16. 18
      pkgs/applications/office/banking/default.nix
  17. 64
      pkgs/applications/science/physics/xfitter/default.nix
  18. 454
      pkgs/applications/science/physics/xfitter/undefined_behavior.patch
  19. 16
      pkgs/build-support/kernel/make-initrd-ng.nix
  20. 2
      pkgs/development/libraries/CoinMP/default.nix
  21. 2
      pkgs/development/python-modules/dm-haiku/default.nix
  22. 8
      pkgs/development/python-modules/flax/default.nix
  23. 13
      pkgs/development/python-modules/google-cloud-error-reporting/default.nix
  24. 70
      pkgs/development/python-modules/meteofrance-api/default.nix
  25. 40
      pkgs/development/python-modules/oasatelematics/default.nix
  26. 5
      pkgs/development/python-modules/pylama/default.nix
  27. 43
      pkgs/development/python-modules/python-family-hub-local/default.nix
  28. 40
      pkgs/development/python-modules/rova/default.nix
  29. 41
      pkgs/development/python-modules/securetar/default.nix
  30. 34
      pkgs/development/python-modules/shiv/default.nix
  31. 4
      pkgs/development/python-modules/types-setuptools/default.nix
  32. 5
      pkgs/development/python-modules/tzlocal/default.nix
  33. 6
      pkgs/development/python-modules/zimports/default.nix
  34. 4
      pkgs/os-specific/linux/firmware/sof-firmware/default.nix
  35. 27
      pkgs/servers/dgraph/default.nix
  36. 19
      pkgs/servers/home-assistant/component-packages.nix
  37. 4
      pkgs/servers/home-assistant/default.nix
  38. 4
      pkgs/tools/admin/syft/default.nix
  39. 6
      pkgs/tools/graphics/sic-image-cli/default.nix
  40. 4
      pkgs/tools/misc/diffoscope/default.nix
  41. 4
      pkgs/top-level/all-packages.nix
  42. 12
      pkgs/top-level/python-packages.nix

@ -95,11 +95,14 @@ with lib;
config = {
assertions = [
{
# Prevent users from disabling nscd, with nssModules being set.
# If disabling nscd is really necessary, it's still possible to opt out
# by forcing config.system.nssModules to [].
assertion = config.system.nssModules.path != "" -> config.services.nscd.enable;
message = "Loading NSS modules from system.nssModules (${config.system.nssModules.path}), requires services.nscd.enable being set to true.";
message = ''
Loading NSS modules from system.nssModules (${config.system.nssModules.path}),
requires services.nscd.enable being set to true.
If disabling nscd is really necessary, it is possible to disable loading NSS modules
by setting `system.nssModules = lib.mkForce [];` in your configuration.nix.
'';
}
];

@ -1187,7 +1187,6 @@
./system/boot/systemd/tmpfiles.nix
./system/boot/systemd/user.nix
./system/boot/systemd/initrd.nix
./system/boot/systemd/initrd-mdraid.nix
./system/boot/timesyncd.nix
./system/boot/tmp.nix
./system/etc/etc-activation.nix

@ -355,7 +355,7 @@ let
[ { object = bootStage1;
symlink = "/init";
}
{ object = pkgs.writeText "mdadm.conf" config.boot.initrd.services.mdraid.mdadmConf;
{ object = pkgs.writeText "mdadm.conf" config.boot.initrd.services.swraid.mdadmConf;
symlink = "/etc/mdadm.conf";
}
{ object = pkgs.runCommand "initrd-kmod-blacklist-ubuntu" {
@ -731,6 +731,6 @@ in
};
imports = [
(mkRenamedOptionModule [ "boot" "initrd" "mdadmConf" ] [ "boot" "initrd" "services" "mdraid" "mdadmConf" ])
(mkRenamedOptionModule [ "boot" "initrd" "mdadmConf" ] [ "boot" "initrd" "services" "swraid" "mdadmConf" ])
];
}

@ -1,32 +0,0 @@
{ config, pkgs, lib, ... }: let
cfg = config.boot.initrd.services.mdraid;
in {
options.boot.initrd.services.mdraid = {
enable = (lib.mkEnableOption "mdraid support in initrd") // {
visible = false;
};
mdadmConf = lib.mkOption {
description = "Contents of <filename>/etc/mdadm.conf</filename> in initrd.";
type = lib.types.lines;
default = "";
};
};
config = lib.mkIf (config.boot.initrd.systemd.enable && cfg.enable) {
boot.initrd.systemd = {
contents."/etc/mdadm.conf" = lib.mkIf (cfg.mdadmConf != "") {
text = cfg.mdadmConf;
};
initrdBin = [ pkgs.mdadm ];
};
boot.initrd.services.udev.packages = [ pkgs.mdadm ];
boot.initrd.systemd.packages = [ pkgs.mdadm ];
boot.kernelModules = [ "dm-raid" ];
};
}

@ -125,6 +125,9 @@ let
};
initialRamdisk = pkgs.makeInitrdNG {
name = "initrd-${kernel-name}";
inherit (config.boot.initrd) compressor compressorArgs prepend;
contents = map (path: { object = path; symlink = ""; }) (subtractLists cfg.suppressedStorePaths cfg.storePaths)
++ mapAttrsToList (_: v: { object = v.source; symlink = v.target; }) (filterAttrs (_: v: v.enable) cfg.contents);
};

@ -1,17 +1,43 @@
{ pkgs, ... }:
{ config, pkgs, lib, ... }: let
{
cfg = config.boot.initrd.services.swraid;
environment.systemPackages = [ pkgs.mdadm ];
in {
services.udev.packages = [ pkgs.mdadm ];
options.boot.initrd.services.swraid = {
enable = (lib.mkEnableOption "swraid support using mdadm") // {
visible = false; # only has effect when the new stage 1 is in place
};
systemd.packages = [ pkgs.mdadm ];
mdadmConf = lib.mkOption {
description = "Contents of <filename>/etc/mdadm.conf</filename> in initrd.";
type = lib.types.lines;
default = "";
};
};
boot.initrd.availableKernelModules = [ "md_mod" "raid0" "raid1" "raid10" "raid456" ];
config = {
environment.systemPackages = [ pkgs.mdadm ];
boot.initrd.extraUdevRulesCommands = ''
cp -v ${pkgs.mdadm}/lib/udev/rules.d/*.rules $out/
'';
services.udev.packages = [ pkgs.mdadm ];
systemd.packages = [ pkgs.mdadm ];
boot.initrd.availableKernelModules = lib.mkIf (config.boot.initrd.systemd.enable -> cfg.enable) [ "md_mod" "raid0" "raid1" "raid10" "raid456" ];
boot.initrd.extraUdevRulesCommands = lib.mkIf (!config.boot.initrd.systemd.enable) ''
cp -v ${pkgs.mdadm}/lib/udev/rules.d/*.rules $out/
'';
boot.initrd.systemd = lib.mkIf cfg.enable {
contents."/etc/mdadm.conf" = lib.mkIf (cfg.mdadmConf != "") {
text = cfg.mdadmConf;
};
packages = [ pkgs.mdadm ];
initrdBin = [ pkgs.mdadm ];
};
boot.initrd.services.udev.packages = lib.mkIf cfg.enable [ pkgs.mdadm ];
};
}

@ -521,8 +521,8 @@ in
systemd-confinement = handleTest ./systemd-confinement.nix {};
systemd-cryptenroll = handleTest ./systemd-cryptenroll.nix {};
systemd-escaping = handleTest ./systemd-escaping.nix {};
systemd-initrd-mdraid = handleTest ./systemd-initrd-mdraid.nix {};
systemd-initrd-simple = handleTest ./systemd-initrd-simple.nix {};
systemd-initrd-swraid = handleTest ./systemd-initrd-swraid.nix {};
systemd-journal = handleTest ./systemd-journal.nix {};
systemd-machinectl = handleTest ./systemd-machinectl.nix {};
systemd-networkd = handleTest ./systemd-networkd.nix {};

@ -1,5 +1,5 @@
import ./make-test-python.nix ({ lib, pkgs, ... }: {
name = "systemd-initrd-mdraid";
name = "systemd-initrd-swraid";
nodes.machine = { pkgs, ... }: {
# Use systemd-boot
@ -17,7 +17,7 @@ import ./make-test-python.nix ({ lib, pkgs, ... }: {
enable = true;
emergencyAccess = true;
};
services.mdraid = {
services.swraid = {
enable = true;
mdadmConf = ''
ARRAY /dev/md0 devices=/dev/vdc,/dev/vdd
@ -26,7 +26,7 @@ import ./make-test-python.nix ({ lib, pkgs, ... }: {
kernelModules = [ "raid0" ];
};
specialisation.boot-mdraid.configuration.virtualisation.bootDevice = "/dev/disk/by-label/testraid";
specialisation.boot-swraid.configuration.virtualisation.bootDevice = "/dev/disk/by-label/testraid";
};
testScript = ''
@ -36,7 +36,7 @@ import ./make-test-python.nix ({ lib, pkgs, ... }: {
machine.succeed("mkdir -p /mnt && mount /dev/md0 /mnt && echo hello > /mnt/test && umount /mnt")
# Boot from the RAID
machine.succeed("bootctl set-default nixos-generation-1-specialisation-boot-mdraid.conf")
machine.succeed("bootctl set-default nixos-generation-1-specialisation-boot-swraid.conf")
machine.succeed("sync")
machine.crash()
machine.wait_for_unit("multi-user.target")

@ -1,5 +1,5 @@
{ config, lib, stdenv, fetchFromGitHub, runCommand, ncurses, pkg-config
, libiconv, CoreAudio, AudioUnit
, libiconv, CoreAudio, AudioUnit, VideoToolbox
, alsaSupport ? stdenv.isLinux, alsa-lib ? null
# simple fallback for everyone else
@ -121,7 +121,7 @@ stdenv.mkDerivation rec {
nativeBuildInputs = [ pkg-config ];
buildInputs = [ ncurses ]
++ lib.optional stdenv.cc.isClang clangGCC
++ lib.optionals stdenv.isDarwin [ libiconv CoreAudio AudioUnit ]
++ lib.optionals stdenv.isDarwin [ libiconv CoreAudio AudioUnit VideoToolbox ]
++ flatten (concatMap (a: a.deps) opts);
makeFlags = [ "LD=$(CC)" ];

@ -25,16 +25,16 @@
}:
rustPlatform.buildRustPackage rec {
pname = "neovide";
version = "unstable-2022-02-04";
version = "0.8.0";
src = fetchFromGitHub {
owner = "Kethku";
repo = "neovide";
rev = "92bc1725f1733547eb0ae25b740425f03f358c2a";
sha256 = "sha256-bKTteaj6gddp0NuV5Y0pfHotezU9Hmb136xOC9zkJ/M=";
rev = version;
sha256 = "sha256-pbniOWjEw1Z+PoXqbbFOUkW5Ii1UDOMoZpAvVF1uNEg=";
};
cargoSha256 = "sha256-TaZN49ou6bf1vW0mEsmaItp1c73d0M826MMrSGXpnGE=";
cargoSha256 = "sha256-7o7uJXH68pvfuiG1eSNmbPx8OO8QJjCe+oEFl38bFm4=";
SKIA_SOURCE_DIR =
let
@ -46,7 +46,7 @@ rustPlatform.buildRustPackage rec {
sha256 = "sha256-F1DWLm7bdKnuCu5tMMekxSyaGq8gPRNtZwcRVXJxjZQ=";
};
# The externals for skia are taken from skia/DEPS
externals = lib.mapAttrs (n: v: fetchgit v) (lib.importJSON ./skia-externals.json);
externals = lib.mapAttrs (n: fetchgit) (lib.importJSON ./skia-externals.json);
in
runCommand "source" {} (
''
@ -128,7 +128,7 @@ rustPlatform.buildRustPackage rec {
homepage = "https://github.com/Kethku/neovide";
license = with licenses; [ mit ];
maintainers = with maintainers; [ ck3d ];
platforms = platforms.unix;
platforms = platforms.linux;
mainProgram = "neovide";
};
}

@ -18,13 +18,13 @@
python3.pkgs.buildPythonApplication rec {
pname = "setzer";
version = "0.4.4";
version = "0.4.7";
src = fetchFromGitHub {
owner = "cvfosammmm";
repo = "Setzer";
rev = "v${version}";
hash = "sha256-y39u1eSX7vzcIJzoykU7Y7FTmeZnBW/IlvJLzKHfz8Y=";
hash = "sha256-IP56jOiiIK9EW4D5yEdLc49rUzcvegAX3Yyk2ERK/pE=";
};
format = "other";

@ -66,9 +66,9 @@ let
buildInputs = [ libsecret libXScrnSaver libxshmfence ]
++ lib.optionals (!stdenv.isDarwin) ([ at-spi2-atk ] ++ atomEnv.packages);
runtimeDependencies = lib.optional (stdenv.isLinux) [ (lib.getLib systemd) fontconfig.lib libdbusmenu ];
runtimeDependencies = lib.optional stdenv.isLinux [ (lib.getLib systemd) fontconfig.lib libdbusmenu ];
nativeBuildInputs = [unzip] ++ lib.optionals (!stdenv.isDarwin) [ autoPatchelfHook wrapGAppsHook ];
nativeBuildInputs = [ unzip ] ++ lib.optionals stdenv.isLinux [ autoPatchelfHook nodePackages.asar wrapGAppsHook ];
dontBuild = true;
dontConfigure = true;
@ -114,12 +114,18 @@ let
# this is a fix for "save as root" functionality
packed="resources/app/node_modules.asar"
unpacked="resources/app/node_modules"
${nodePackages.asar}/bin/asar extract "$packed" "$unpacked"
asar extract "$packed" "$unpacked"
substituteInPlace $unpacked/@vscode/sudo-prompt/index.js \
--replace "/usr/bin/pkexec" "/run/wrappers/bin/pkexec" \
--replace "/bin/bash" "${bash}/bin/bash"
rm -rf "$packed"
# without this symlink loading JsChardet, the library that is used for auto encoding detection when files.autoGuessEncoding is true,
# fails to load with: electron/js2c/renderer_init: Error: Cannot find module 'jschardet'
# and the window immediately closes which renders VSCode unusable
# see https://github.com/NixOS/nixpkgs/issues/152939 for full log
ln -rs "$unpacked" "$packed"
# this fixes bundled ripgrep
chmod +x resources/app/node_modules/@vscode/ripgrep/bin/rg
'';

@ -1,7 +1,6 @@
{ lib
, stdenv
, autoconf
, automake
, autoreconfHook
, c-ares
, cryptopp
, curl
@ -14,7 +13,6 @@
, libmediainfo
, libraw
, libsodium
, libtool
, libuv
, libzen
, pcre-cpp
@ -35,7 +33,8 @@ stdenv.mkDerivation rec {
fetchSubmodules = true;
};
nativeBuildInputs = [ autoconf automake libtool pkg-config ];
enableParallelBuilding = true;
nativeBuildInputs = [ autoreconfHook pkg-config ];
buildInputs = [
c-ares
@ -54,10 +53,6 @@ stdenv.mkDerivation rec {
sqlite
];
preConfigure = ''
./autogen.sh
'';
configureFlags = [
"--disable-curl-checks"
"--disable-examples"

@ -1,8 +1,8 @@
{
"stable": {
"version": "100.0.4896.88",
"sha256": "0l628x41krsjgzff9996k5wkbcvcjqf4128z32hpj1pkg23719f5",
"sha256bin64": "1wqzs3f70ayi9vy3ncm5mild22xvhwn4d2lcfra31wwnzxi1nqxm",
"version": "100.0.4896.127",
"sha256": "0kgq38dy9mjyc44556i9gxhlsgd7dfvv1xi1ibk92b4p7i2y6427",
"sha256bin64": "0mm6lix14bf4ca440dyccnq54z0qvn5c886ghfyzy2q0bqzbq4nh",
"deps": {
"gn": {
"version": "2022-01-21",

@ -2,7 +2,9 @@
, mkDerivation
, fetchFromGitHub
, cmake
, extra-cmake-modules
, inotify-tools
, installShellFiles
, libcloudproviders
, libsecret
, openssl
@ -15,6 +17,8 @@
, qtwebsockets
, qtquickcontrols2
, qtgraphicaleffects
, plasma5Packages
, sphinx
, sqlite
, inkscape
, xdg-utils
@ -24,6 +28,8 @@ mkDerivation rec {
pname = "nextcloud-client";
version = "3.4.4";
outputs = [ "out" "dev" ];
src = fetchFromGitHub {
owner = "nextcloud";
repo = "desktop";
@ -37,10 +43,19 @@ mkDerivation rec {
./0001-When-creating-the-autostart-entry-do-not-use-an-abso.patch
];
postPatch = ''
for file in src/libsync/vfs/*/CMakeLists.txt; do
substituteInPlace $file \
--replace "PLUGINDIR" "KDE_INSTALL_PLUGINDIR"
done
'';
nativeBuildInputs = [
pkg-config
cmake
extra-cmake-modules
inkscape
sphinx
];
buildInputs = [
@ -49,6 +64,7 @@ mkDerivation rec {
libsecret
openssl
pcre
plasma5Packages.kio
qtbase
qtkeychain
qttools
@ -71,6 +87,10 @@ mkDerivation rec {
"-DNO_SHIBBOLETH=1" # allows to compile without qtwebkit
];
postBuild = ''
make doc-man
'';
meta = with lib; {
description = "Nextcloud themed desktop client";
homepage = "https://nextcloud.com";

@ -5,27 +5,26 @@
, appstream-glib
, desktop-file-utils
, glib
, gtk3
, libxml2
, meson
, ninja
, pkg-config
, wrapGAppsHook
, gobject-introspection
, libhandy
, libadwaita
, librsvg
}:
python3.pkgs.buildPythonApplication rec {
pname = "banking";
version = "0.3.0";
version = "0.4.0";
format = "other";
src = fetchFromGitLab {
owner = "tabos";
repo = "banking";
rev = version;
sha256 = "1w5x9iczw5hb9bfdm1df37n8xhdrida1yfrd82k9l8hb1k4q3h9d";
sha256 = "sha256-VGNCSirQslRfLIFeo375BNlHujoNXm+s55Ty+hB+ZRI=";
};
patches = [
@ -35,17 +34,21 @@ python3.pkgs.buildPythonApplication rec {
url = "https://gitlab.com/tabos/banking/-/commit/c3cc9afc380fe666ae6e331aa8a97659c60397a4.patch";
sha256 = "r9n9l47XU4Tg4U5sfiFdGkbG8QB7O4ol9CB1ya06yOc=";
})
# fix build with libadwaita 1.0.0
(fetchpatch {
url = "https://gitlab.com/tabos/banking/-/commit/27ac4a89ba6047005d43de71a469ef30d1fda8b5.patch";
hash = "sha256-dpDjdYf3gDsyFMTfGes+x27yUxKEnKjLulJxX2encG0=";
})
];
postPatch = ''
patchShebangs meson_post_install.py
patchShebangs meson_post_conf.py meson_post_install.py
'';
nativeBuildInputs = [
appstream-glib # for appstream-util
desktop-file-utils # for desktop-file-validate
glib # for glib-compile-resources
gtk3 # for gtk-update-icon-cache
libxml2 # for xmllint
meson
ninja
@ -55,8 +58,7 @@ python3.pkgs.buildPythonApplication rec {
buildInputs = [
gobject-introspection
gtk3
libhandy
libadwaita
librsvg
];

@ -1,53 +1,57 @@
{ lib, stdenv, fetchurl, apfel, apfelgrid, applgrid, blas, gfortran, lhapdf, lapack, libyaml, lynx
, mela, root5, qcdnum, which, libtirpc
{ lib
, stdenv
, fetchurl
, apfel
, apfelgrid
, applgrid
, blas
, ceres-solver
, cmake
, gfortran
, gsl
, lapack
, lhapdf
, libtirpc
, libyaml
, libyamlcpp
, pkg-config
, qcdnum
, root
, zlib
, memorymappingHook, memstreamHook
}:
stdenv.mkDerivation rec {
pname = "xfitter";
version = "2.0.1";
version = "2.2.0";
src = fetchurl {
name = "${pname}-${version}.tgz";
url = "https://www.xfitter.org/xFitter/xFitter/DownloadPage?action=AttachFile&do=get&target=${pname}-${version}.tgz";
sha256 = "0kmgc67nw5flp92yw5x6l2vsnhwsfi5z2a20404anisdgdjs8zc6";
sha256 = "sha256-ZHIQ5hOY+k0/wmpE0o4Po+RZ4MkVMk+bK1Rc6eqwwH0=";
};
patches = [
./undefined_behavior.patch
];
preConfigure = ''
substituteInPlace CMakeLists.txt \
--replace "-fallow-argument-mismatch" ""
'';
preConfigure =
# Fix F77LD to workaround for a following build error:
#
# gfortran: error: unrecognized command line option '-stdlib=libc++'
#
lib.optionalString stdenv.isDarwin ''
substituteInPlace src/Makefile.in \
--replace "F77LD = \$(F77)" "F77LD = \$(CXXLD)" \
'';
configureFlags = [
"--enable-apfel"
"--enable-apfelgrid"
"--enable-applgrid"
"--enable-mela"
"--enable-lhapdf"
];
nativeBuildInputs = [ gfortran which ];
nativeBuildInputs = [ cmake gfortran pkg-config ];
buildInputs =
[ apfel apfelgrid applgrid blas lhapdf libyaml lapack mela root5 qcdnum ]
[ apfel blas ceres-solver lhapdf lapack libyaml root qcdnum gsl libyamlcpp zlib ]
++ lib.optionals ("5" == lib.versions.major root.version) [ apfelgrid applgrid ]
++ lib.optionals (stdenv.system == "x86_64-darwin") [ memorymappingHook memstreamHook ]
++ lib.optional (stdenv.hostPlatform.libc == "glibc") libtirpc
;
propagatedBuildInputs = [ lynx ];
enableParallelBuilding = true;
NIX_CFLAGS_COMPILE = lib.optional (stdenv.hostPlatform.libc == "glibc") "-I${libtirpc.dev}/include/tirpc";
NIX_LDFLAGS = lib.optional (stdenv.hostPlatform.libc == "glibc") "-ltirpc";
# workaround wrong library IDs
postInstall = lib.optionalString stdenv.isDarwin ''
ln -sv "$out/lib/xfitter/"* "$out/lib/"
'';
meta = with lib; {
description = "The xFitter project is an open source QCD fit framework ready to extract PDFs and assess the impact of new data";
license = licenses.gpl3;

@ -1,454 +0,0 @@
diff --git a/DY/src/finterface.cc b/DY/src/finterface.cc
--- a/DY/src/finterface.cc
+++ b/DY/src/finterface.cc
@@ -14,17 +14,17 @@
using namespace std;
extern "C" {
- int dy_create_calc_(const int *ds_id, const int *chg_prod,
+ void dy_create_calc_(const int *ds_id, const int *chg_prod,
const double *beam_en, const char *boz,
const double *ranges, const char *var_name,
const int *n_bins, const double *bin_edges);
- int dy_do_calc_();
+ void dy_do_calc_();
- int dy_get_res_(const int *ds_id, double *calc_res);
+ void dy_get_res_(const int *ds_id, double *calc_res);
- int dy_release_();
- int dy_set_ewpars_();
+ void dy_release_();
+ void dy_set_ewpars_();
}
typedef map <int, DYcalc* > DCmap;
@@ -34,7 +34,7 @@ vector<BinMatrix*> gBinMatrices;
// initializes Drell-Yan LO calculations with info on
// beam, process, kinematic cuts, and bins.
-int dy_create_calc_(const int *ds_id, const int *chg_prod,
+void dy_create_calc_(const int *ds_id, const int *chg_prod,
const double *beam_en, const char *boz,
const double *ranges, const char *var_name,
const int *n_bins, const double *bin_edges)
@@ -99,13 +99,11 @@ int dy_create_calc_(const int *ds_id, const int *chg_prod,
// create calculator and put to map
DYcalc * dc = new DYcalc(bm, pc, int_steps);
gCalcs.insert( pair<int,DYcalc*>( *ds_id,dc ) );
-
- return 1;
}
// calculate Drell-Yan LO cross sections for all data sets
-int dy_do_calc_()
+void dy_do_calc_()
{
// evolve convolutions
vector<PDFconv*>::iterator ipc = gPDFconvs.begin();
@@ -118,28 +116,24 @@ int dy_do_calc_()
if ( true != idc->second->Integrate() ) {
cout << "Something is wrong with DY integration for "
<< idc->first << " data set." << endl;
- return 0;
+ return;
}
}
-
- return 1;
}
// return DY calculations for data set ds_name
-int dy_get_res_(const int *ds_id, double *calc_res)
+void dy_get_res_(const int *ds_id, double *calc_res)
{
DYcalc * dc = gCalcs.find(*ds_id)->second;
dc->getCalcRes(calc_res);
-
- return 1;
}
-int dy_set_ewpars_(){
+void dy_set_ewpars_(){
PhysPar::setPhysPar();
}
-int dy_release_()
+void dy_release_()
{
vector<PDFconv*>::iterator ipc = gPDFconvs.begin();
for (; ipc!=gPDFconvs.end(); ipc++){
@@ -155,6 +149,4 @@ int dy_release_()
for (; idc != gCalcs.end() ; idc++){
delete (idc->second);
}
-
- return 1;
}
diff --git a/DiffDIS/include/DataTable.h b/DiffDIS/include/DataTable.h
--- a/DiffDIS/include/DataTable.h
+++ b/DiffDIS/include/DataTable.h
@@ -307,6 +307,7 @@ class DataTable_t {
for(ic=0; ic < GetNcols(); ic++) {
for(ir=0; ir < npt; ir++) Data[ic][ir] = A.Data[ic][ir];
}
+ return *this;
}
//@}
diff --git a/FastNLO/src/FastNLOInterface.cc b/FastNLO/src/FastNLOInterface.cc
--- a/FastNLO/src/FastNLOInterface.cc
+++ b/FastNLO/src/FastNLOInterface.cc
@@ -39,14 +39,14 @@ void gauleg(double x1,double x2,double *x,double *w, int n);
extern "C" {
- int fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_FIT_ORDER, bool *PublicationUnits , double* murdef, double* murscale, double *mufdef, double* mufscale);
- int fastnlocalc_(const int *idataset, double *xsec);
- int fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *tot, int *Npt);
- int fastnlopointskip_(const int *idataset, int *point, int *npoints);
- int hf_errlog_(const int* ID, const char* TEXT, long length);
- int hf_stop_();
+ void fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_FIT_ORDER, bool *PublicationUnits , double* murdef, double* murscale, double *mufdef, double* mufscale);
+ void fastnlocalc_(const int *idataset, double *xsec);
+ void fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *tot, int *Npt);
+ void fastnlopointskip_(const int *idataset, int *point, int *npoints);
+ void hf_errlog_(const int* ID, const char* TEXT, long length);
+ void hf_stop_();
double interp_(double *A, double *xx1, double *x, int *NGrid1, double *res);
- int setfastnlotoppar_(const int *idataset);
+ void setfastnlotoppar_(const int *idataset);
}
@@ -58,7 +58,7 @@ map<int, FastNLOxFitter*> gFastNLO_array;
map<int, BoolArray*> gUsedPoints_array;
int CreateUsedPointsArray(int idataset, int npoints);
-int fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_FIT_ORDER, bool *PublicationUnits , double* murdef, double* murscale, double *mufdef, double* mufscale) {
+void fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_FIT_ORDER, bool *PublicationUnits , double* murdef, double* murscale, double *mufdef, double* mufscale) {
map<int, FastNLOxFitter*>::const_iterator FastNLOIterator = gFastNLO_array.find(*idataset);
@@ -67,7 +67,7 @@ int fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_
const char* text = "I: Double initialization of the same fastnlo data set!";
hf_errlog_(&id, text, (long)strlen(text));
//hf_stop_();
- return 1;
+ return;
}
FastNLOxFitter* fnloreader = new FastNLOxFitter( thfile );
@@ -112,10 +112,9 @@ int fastnloinit_(const char *s, const int *idataset, const char *thfile, int *I_
}
gFastNLO_array.insert(pair<int, FastNLOxFitter*>(*idataset, fnloreader) );
- return 0;
}
-int setfastnlotoppar_(const int *idataset) {
+void setfastnlotoppar_(const int *idataset) {
//!< Dedicated settings for difftop
map<int, FastNLOxFitter*>::const_iterator FastNLOIterator = gFastNLO_array.find(*idataset);
map<int, BoolArray*>::const_iterator UsedPointsIterator = gUsedPoints_array.find(*idataset);
@@ -130,11 +129,9 @@ int setfastnlotoppar_(const int *idataset) {
fnloreader->SetExternalFuncForMuF( &Function_Mu );
fnloreader->SetExternalFuncForMuR( &Function_Mu);
//fnloreader->SetScaleFactorsMuRMuF(1.0,1.0); //Be reminded that muR and muF scales are hard coded (that's not true!)
-
- return 0;
}
-int fastnlocalc_(const int *idataset, double *xsec) {
+void fastnlocalc_(const int *idataset, double *xsec) {
map<int, FastNLOxFitter*>::const_iterator FastNLOIterator = gFastNLO_array.find(*idataset);
map<int, BoolArray*>::const_iterator UsedPointsIterator = gUsedPoints_array.find(*idataset);
@@ -176,13 +173,10 @@ int fastnlocalc_(const int *idataset, double *xsec) {
outputidx++;
}
}
-
-
- return 0;
}
//MK14 New function for Difftop calculation: it is called in trunk/src/difftop_fastnlo.f
-int fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *tot, int *Npt){
+void fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *tot, int *Npt){
map<int, FastNLOxFitter*>::const_iterator FastNLOIterator = gFastNLO_array.find(*idataset);
map<int, BoolArray*>::const_iterator UsedPointsIterator = gUsedPoints_array.find(*idataset);
@@ -262,10 +256,6 @@ int fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *to
Total += interpC(xsec,xg[k],thbin,Nthpoints)*wg[k];
*tot = Total;
-
-
-
- return 0;
}
@@ -277,7 +267,7 @@ int fastnlocalctop_(const int *idataset, double *xsec, double *thbin, double *to
-int fastnlopointskip_(const int *idataset, int *point, int *npoints) {
+void fastnlopointskip_(const int *idataset, int *point, int *npoints) {
map<int, BoolArray*>::const_iterator UsedPointsIterator = gUsedPoints_array.find(*idataset);
if(UsedPointsIterator == gUsedPoints_array.end( ))
CreateUsedPointsArray(*idataset, *npoints);
@@ -292,8 +282,6 @@ int fastnlopointskip_(const int *idataset, int *point, int *npoints) {
BoolArray* usedpoints = UsedPointsIterator->second;
usedpoints->at(*point-1) = false;
-
- return 0;
}
int CreateUsedPointsArray(int idataset, int npoints) {
diff --git a/Hathor/src/HathorInterface.cc b/Hathor/src/HathorInterface.cc
--- a/Hathor/src/HathorInterface.cc
+++ b/Hathor/src/HathorInterface.cc
@@ -6,9 +6,9 @@
#include "../interface/xFitterPdf.h"
extern "C" {
- int hathorinit_(const int* idataset, const double& sqrtS, const bool& ppbar, const double& mt,
+ void hathorinit_(const int* idataset, const double& sqrtS, const bool& ppbar, const double& mt,
const unsigned int& pertubOrder, const unsigned int& precisionLevel);
- int hathorcalc_(const int *idataset, double *xsec);
+ void hathorcalc_(const int *idataset, double *xsec);
}
extern "C" {
@@ -19,7 +19,7 @@ extern "C" {
}
extern "C" {
- int hf_errlog_(const int* ID, const char* TEXT, long length);
+ void hf_errlog_(const int* ID, const char* TEXT, long length);
}
// FIXME: delete pointers at the end! (in some hathordestroy_ or so)
@@ -28,7 +28,7 @@ xFitterPdf* pdf;
int* rndStore;
double mtop;
-int hathorinit_(const int* idataset, const double& sqrtS, const bool& ppbar, const double& mt,
+void hathorinit_(const int* idataset, const double& sqrtS, const bool& ppbar, const double& mt,
const unsigned int& pertubOrder, const unsigned int& precisionLevel) {
if(hathor_array.size()==0) {
@@ -69,7 +69,7 @@ int hathorinit_(const int* idataset, const double& sqrtS, const bool& ppbar, con
return 0;
}
-int hathorcalc_(const int *idataset, double *xsec) {
+void hathorcalc_(const int *idataset, double *xsec) {
rlxd_reset(rndStore);
std::map<int, Hathor*>::const_iterator hathorIter = hathor_array.find(*idataset);
diff --git a/src/TheorEval.cc b/src/TheorEval.cc
--- a/src/TheorEval.cc
+++ b/src/TheorEval.cc
@@ -62,6 +62,7 @@ TheorEval::initTheory()
list<tToken> sl;
this->assignTokens(sl);
this->convertToRPN(sl);
+ return 0;
}
int
@@ -167,6 +168,7 @@ TheorEval::assignTokens(list<tToken> &sl)
sl.push_back(t);
}
}
+ return 0;
}
int
@@ -217,6 +219,7 @@ TheorEval::convertToRPN(list<tToken> &sl)
cout << endl;
*/
+ return 0;
}
int
@@ -236,6 +239,7 @@ TheorEval::initTerm(int iterm, valarray<double> *val)
hf_errlog_(id, text, textlen);
return -1;
}
+ return 0;
}
int
@@ -348,6 +352,7 @@ TheorEval::initGridTerm(int iterm, valarray<double> *val)
// associate grid and valarray pointers in token
_mapGridToken[g] = val;
+ return 0;
}
int
@@ -430,6 +435,7 @@ TheorEval::initKfTerm(int iterm, valarray<double> *val)
// write k-factor array to the token valarray
*val = valarray<double>(vkf.data(), vkf.size());
+ return 0;
}
int
@@ -465,6 +471,7 @@ TheorEval::setCKM(const vector<double> &v_ckm)
int textlen = strlen(text);
hf_errlog_(id, text, textlen);
#endif
+ return 0;
}
int
@@ -531,6 +538,7 @@ TheorEval::Evaluate(valarray<double> &vte )
}
//vte /= _units;
}
+ return 0;
}
int
@@ -555,6 +563,7 @@ TheorEval::getGridValues()
}
+ return 0;
}
int
diff --git a/src/ftheor_eval.cc b/src/ftheor_eval.cc
--- a/src/ftheor_eval.cc
+++ b/src/ftheor_eval.cc
@@ -19,15 +19,15 @@
using namespace std;
extern "C" {
- int set_theor_eval_(int *dsId);//, int *nTerms, char **TermName, char **TermType,
+ void set_theor_eval_(int *dsId);//, int *nTerms, char **TermName, char **TermType,
// char **TermSource, char *TermExpr);
- int set_theor_bins_(int *dsId, int *nBinDimension, int *nPoints, int *binFlags,
+ void set_theor_bins_(int *dsId, int *nBinDimension, int *nPoints, int *binFlags,
double *allBins);
// int set_theor_units_(int *dsId, double *units);
- int init_theor_eval_(int *dsId);
- int update_theor_ckm_();
- int get_theor_eval_(int *dsId, int* np, int* idx);
- int close_theor_eval_();
+ void init_theor_eval_(int *dsId);
+ void update_theor_ckm_();
+ void get_theor_eval_(int *dsId, int* np, int* idx);
+ void close_theor_eval_();
}
/// global dataset to theory evaluation pointer map
@@ -59,7 +59,7 @@ extern struct ord_scales {
dataset ID.
write details on argumets
*/
-int set_theor_eval_(int *dsId)//, int *nTerms, char **TermName, char **TermType,
+void set_theor_eval_(int *dsId)//, int *nTerms, char **TermName, char **TermType,
// char **TermSource, char *TermExpr)
{
// convert fortran strings to c++
@@ -90,15 +90,13 @@ int set_theor_eval_(int *dsId)//, int *nTerms, char **TermName, char **TermType,
<< " already exists." << endl;
exit(1); // make proper exit later
}
-
- return 1;
}
/*!
Sets datasets bins in theory evaluations.
write details on argumets
*/
-int set_theor_bins_(int *dsId, int *nBinDimension, int *nPoints, int *binFlags,
+void set_theor_bins_(int *dsId, int *nBinDimension, int *nPoints, int *binFlags,
double *allBins)
{
tTEmap::iterator it = gTEmap.find(*dsId);
@@ -110,7 +108,6 @@ int set_theor_bins_(int *dsId, int *nBinDimension, int *nPoints, int *binFlags,
TheorEval *te = gTEmap.at(*dsId);
te->setBins(*nBinDimension, *nPoints, binFlags, allBins);
- return 1;
}
/*
@@ -132,7 +129,7 @@ int set_theor_units_(int *dsId, double *units)
/*!
Initializes theory for requested dataset.
*/
-int init_theor_eval_(int *dsId)
+void init_theor_eval_(int *dsId)
{
tTEmap::iterator it = gTEmap.find(*dsId);
if (it == gTEmap.end() ) {
@@ -148,7 +145,7 @@ int init_theor_eval_(int *dsId)
/*!
Updates the CKM matrix to all the initialized appl grids
*/
-int update_theor_ckm_()
+void update_theor_ckm_()
{
double a_ckm[] = { ckm_matrix_.Vud, ckm_matrix_.Vus, ckm_matrix_.Vub,
ckm_matrix_.Vcd, ckm_matrix_.Vcs, ckm_matrix_.Vcb,
@@ -164,7 +161,7 @@ int update_theor_ckm_()
/*!
Evaluates theory for requested dataset and writes it to the global THEO array.
*/
-int get_theor_eval_(int *dsId, int *np, int*idx)
+void get_theor_eval_(int *dsId, int *np, int*idx)
{
tTEmap::iterator it = gTEmap.find(*dsId);
@@ -194,11 +191,11 @@ int get_theor_eval_(int *dsId, int *np, int*idx)
// write the predictions to THEO array
if( ip != *np ){
cout << "ERROR in get_theor_eval_: number of points mismatch" << endl;
- return -1;
+ return;
}
}
-int close_theor_eval_()
+void close_theor_eval_()
{
tTEmap::iterator it = gTEmap.begin();
for (; it!= gTEmap.end(); it++){
diff --git a/src/lhapdf6_output.c b/src/lhapdf6_output.c
--- a/src/lhapdf6_output.c
+++ b/src/lhapdf6_output.c
@@ -64,7 +64,7 @@ extern double bvalij_(int *,int *,int *,int *,int *);
extern double bvalxq_(int *,int *,double *,double *,int *);
extern double hf_get_alphas_(double *);
extern int getord_(int *);
-extern int grpars_(int *, double *, double *, int *, double *, double *, int *);
+extern void grpars_(int *, double *, double *, int *, double *, double *, int *);
extern int getcbt_(int *, double *, double *, double *);
extern void getpdfunctype_heraf_(int *mc, int *asymh, int *symh, char *name, size_t size);
extern void hf_errlog_(int *, char *, size_t);
diff --git a/tools/draw/include/FileOpener.h b/tools/draw/include/FileOpener.h
--- a/tools/draw/include/FileOpener.h
+++ b/tools/draw/include/FileOpener.h
@@ -61,7 +61,7 @@ class InFileOpener_t {
string GetPath() const {return ind < 0 ? "" : Flist[ind];}
// ==================================
- int Add(const string& fname) {
+ void Add(const string& fname) {
Flist.push_back(fname);
}

@ -66,14 +66,28 @@ in
compressorArgs = _compressorArgsReal;
};
inherit extension makeUInitrd uInitrdArch prepend;
${if makeUInitrd then "uInitrdCompression" else null} = uInitrdCompression;
passAsFile = ["contents"];
contents = lib.concatMapStringsSep "\n" ({ object, symlink, ... }: "${object}\n${if symlink == null then "" else symlink}") contents + "\n";
nativeBuildInputs = [makeInitrdNGTool patchelf glibc cpio];
nativeBuildInputs = [makeInitrdNGTool patchelf glibc cpio] ++ lib.optional makeUInitrd ubootTools;
} ''
mkdir ./root
make-initrd-ng "$contentsPath" ./root
mkdir "$out"
(cd root && find * .[^.*] -exec touch -h -d '@1' '{}' +)
for PREP in $prepend; do
cat $PREP >> $out/initrd
done
(cd root && find * .[^.*] -print0 | sort -z | cpio -o -H newc -R +0:+0 --reproducible --null | eval -- $compress >> "$out/initrd")
if [ -n "$makeUInitrd" ]; then
mkimage -A "$uInitrdArch" -O linux -T ramdisk -C "$uInitrdCompression" -d "$out/initrd" $out/initrd.img
# Compatibility symlink
ln -sf "initrd.img" "$out/initrd"
else
ln -s "initrd" "$out/initrd$extension"
fi
''

@ -9,6 +9,8 @@ stdenv.mkDerivation rec {
sha256 = "13d3j1sdcjzpijp4qks3n0zibk649ac3hhv88hkk8ffxrc6gnn9l";
};
enableParallelBuilding = true;
hardeningDisable = [ "format" ];
meta = with lib; {

@ -53,6 +53,8 @@ buildPythonPackage rec {
# likely that's the reason the upstream uses TF-nightly for tests?
# `nixpkgs` doesn't have the corresponding TF version packaged.
"haiku/_src/integration/jax2tf_test.py"
# `TypeError: lax.conv_general_dilated requires arguments to have the same dtypes, got float32, float16`.
"haiku/_src/integration/numpy_inputs_test.py"
];
disabledTests = [

@ -14,13 +14,13 @@
buildPythonPackage rec {
pname = "flax";
version = "0.4.0";
version = "0.4.1";
src = fetchFromGitHub {
owner = "google";
repo = pname;
rev = "v${version}";
sha256 = "0rvdaxyf68qmm5d77gbizpcibyz2ic2pb2x7rgf7p8qwijyc39ws";
sha256 = "0j5ngdndm9nm49gcda7m36qzwk5lcbi4jnij9fi96vld54ip6f6v";
};
buildInputs = [ jaxlib ];
@ -43,6 +43,10 @@ buildPythonPackage rec {
tensorflow
];
pytestFlagsArray = [
"-W ignore::FutureWarning"
];
disabledTestPaths = [
# Docs test, needs extra deps + we're not interested in it.
"docs/_ext/codediff_test.py"

@ -8,22 +8,21 @@
, mock
, proto-plus
, pytest-asyncio
, pythonOlder
}:
buildPythonPackage rec {
pname = "google-cloud-error-reporting";
version = "1.5.1";
version = "1.5.2";
format = "setuptools";
disabled = pythonOlder "3.7";
src = fetchPypi {
inherit pname version;
sha256 = "sha256-7gFpvFxtYneABxW2mOwW1V2E7kobmj0yzV0pxaRhs8c=";
hash = "sha256-wjRUPBZwyGP+2528vZ/x4EqiZwqH+9ZvK5rx4ISklHE=";
};
postPatch = ''
substituteInPlace setup.py \
--replace 'google-cloud-logging>=1.14.0, <2.4' 'google-cloud-logging>=1.14.0'
'';
propagatedBuildInputs = [
google-cloud-logging
libcst

@ -0,0 +1,70 @@
{ lib
, buildPythonPackage
, fetchFromGitHub
, poetry
, pytestCheckHook
, pythonOlder
, pytz
, requests
, requests-mock
, typing-extensions
, urllib3
}:
buildPythonPackage rec {
pname = "meteofrance-api";
version = "1.0.2";
format = "pyproject";
disabled = pythonOlder "3.6";
src = fetchFromGitHub {
owner = "hacf-fr";
repo = pname;
rev = "v${version}";
hash = "sha256-X8f0z9ZPXH7Wc3GqHmPptxpNxbHeezdOzw4gZCprumU=";
};
nativeBuildInputs = [
# Doesn't work with poetry-core at the moment
poetry
];
propagatedBuildInputs = [
pytz
requests
urllib3
] ++ lib.optionals (pythonOlder "3.7") [
typing-extensions
];
checkInputs = [
pytestCheckHook
requests-mock
];
pythonImportsCheck = [
"meteofrance_api"
];
disabledTests = [
# Tests require network access
"test_currentphenomenons"
"test_forecast"
"test_full_with_coastal_bulletint"
"test_fulls"
"test_no_rain_expected"
"test_picture_of_the_day"
"test_places"
"test_rain"
"test_session"
"test_workflow"
];
meta = with lib; {
description = "Module to access information from the Meteo-France API";
homepage = "https://github.com/hacf-fr/meteofrance-api";
license = licenses.mit;
maintainers = with maintainers; [ fab ];
};
}

@ -0,0 +1,40 @@
{ lib
, buildPythonPackage
, fetchFromGitHub
, pytestCheckHook
, pythonOlder
, requests
}:
buildPythonPackage rec {
pname = "oasatelematics";
version = "0.3";
format = "setuptools";
disabled = pythonOlder "3.7";
src = fetchFromGitHub {
owner = "panosmz";
repo = pname;
rev = "v${version}";
hash = "sha256-3O7XbNVj1S3ZwheklEhm0ivw16Tj7drML/xYC9383Kg=";
};
propagatedBuildInputs = [
requests
];
# Module has no tests
doCheck = false;
pythonImportsCheck = [
"oasatelematics"
];
meta = with lib; {
description = "Python wrapper for the OASA Telematics API";
homepage = "https://github.com/panosmz/oasatelematics";
license = licenses.gpl3Only;
maintainers = with maintainers; [ fab ];
};
}

@ -17,7 +17,7 @@
buildPythonPackage rec {
pname = "pylama";
version = "8.3.7";
version = "8.3.8";
format = "setuptools";
@ -26,7 +26,7 @@ buildPythonPackage rec {
owner = "klen";
repo = "pylama";
rev = version;
hash = "sha256-//mrvZb4bT4aATURqa4g1DUagYe9SoP3o3OrwmiEJnI=";
hash = "sha256-g6Lq5NaieUI/alxqoVFfL5VaCHwB/jLcp02/N1W69yE=";
};
patches = [
@ -72,6 +72,7 @@ buildPythonPackage rec {
meta = with lib; {
description = "Code audit tool for python";
homepage = "https://github.com/klen/pylama";
changelog = "https://github.com/klen/pylama/blob/${version}/Changelog";
license = licenses.mit;
maintainers = with maintainers; [ dotlambda ];
};

@ -0,0 +1,43 @@
{ lib
, buildPythonPackage
, aiohttp
, async-timeout
, pillow
, fetchPypi
, pytestCheckHook
, pythonOlder
}:
buildPythonPackage rec {
pname = "python-family-hub-local";
version = "0.0.2";
format = "setuptools";
disabled = pythonOlder "3.9";
src = fetchPypi {
inherit pname version;
extension = "zip";
hash = "sha256-bbOBlUJ4g+HOcJihEBAz3lsHR9Gn07z8st14FRFeJbc=";
};
propagatedBuildInputs = [
aiohttp
async-timeout
pillow
];
# Module has no tests
doCheck = false;
pythonImportsCheck = [
"pyfamilyhublocal"
];
meta = with lib; {
description = "Module to accesse information from Samsung FamilyHub fridges locally";
homepage = "https://github.com/Klathmon/python-family-hub-local";
license = licenses.asl20;
maintainers = with maintainers; [ fab ];
};
}

@ -0,0 +1,40 @@
{ lib
, buildPythonPackage
, fetchFromGitHub
, pytestCheckHook
, pythonOlder
, requests
}:
buildPythonPackage rec {
pname = "rova";
version = "0.3.0";
format = "setuptools";
disabled = pythonOlder "3.7";
src = fetchFromGitHub {
owner = "GidoHakvoort";
repo = pname;
rev = "v${version}";
hash = "sha256-TLL9Ta+7Xd55hGTTXjc6CBMj+tW1LpFrprpsnGqZvkQ=";
};
propagatedBuildInputs = [
requests
];
# Module has no tests
doCheck = false;
pythonImportsCheck = [
"rova"
];
meta = with lib; {
description = "Module to access for ROVA calendars";
homepage = "https://github.com/GidoHakvoort/rova";
license = licenses.mit;
maintainers = with maintainers; [ fab ];
};
}

@ -0,0 +1,41 @@
{ lib
, buildPythonPackage
, cryptography
, fetchFromGitHub
, pytestCheckHook
, pythonOlder
}:
buildPythonPackage rec {
pname = "securetar";
version = "2022.02.0";
format = "setuptools";
disabled = pythonOlder "3.9";
src = fetchFromGitHub {
owner = "pvizeli";
repo = pname;
rev = version;
hash = "sha256-FwQp08jwcGh07zpHqRNoUUmeLZJh78wI8wCXySi3Tdc=";
};
propagatedBuildInputs = [
cryptography
];
checkInputs = [
pytestCheckHook
];
pythonImportsCheck = [
"securetar"
];
meta = with lib; {
description = "Module to handle tarfile backups";
homepage = "https://github.com/pvizeli/securetar";
license = licenses.asl20;
maintainers = with maintainers; [ fab ];
};
}

@ -0,0 +1,34 @@
{ stdenv
, lib
, buildPythonPackage
, fetchPypi
, click
, pip
, setuptools
, wheel
, pytestCheckHook
}:
buildPythonPackage rec {
pname = "shiv";
version = "1.0.1";
format = "pyproject";
src = fetchPypi {
inherit pname version;
sha256 = "ec16095a0565906536af7f5e57771e9ae7a061b646ed63ad66ebbc70c30f4d2a";
};
propagatedBuildInputs = [ click pip setuptools wheel ];
pythonImportsCheck = [ "shiv" ];
checkInputs = [ pytestCheckHook ];
meta = with lib; {
description = "Command line utility for building fully self contained Python zipapps";
homepage = "https://github.com/linkedin/shiv";
license = licenses.bsd2;
maintainers = with maintainers; [ prusnak ];
};
}

@ -5,12 +5,12 @@
buildPythonPackage rec {
pname = "types-setuptools";
version = "57.4.11";
version = "57.4.14";
format = "setuptools";
src = fetchPypi {
inherit pname version;
sha256 = "sha256-Ji90BuDH1wWta7RSa1t2H6UAv5nqt03oWsNZIYfWKTU=";
sha256 = "sha256-3wL+HdJE9Yz05nz8PQqXkwotYact2J8h2BxxAXzYP5o=";
};
# Module doesn't have tests

@ -10,13 +10,13 @@
buildPythonPackage rec {
pname = "tzlocal";
version = "4.1"; # version needs to be compatible with APScheduler
version = "4.2"; # version needs to be compatible with APScheduler
disabled = pythonOlder "3.6";
src = fetchPypi {
inherit pname version;
sha256 = "sha256-DygBWsaKXAZyEEAKkZf8XTa6m8P46vHaPL1ZrN/tngk=";
sha256 = "ee5842fa3a795f023514ac2d801c4a81d1743bbe642e3940143326b3a00addd7";
};
propagatedBuildInputs = [
@ -39,6 +39,7 @@ buildPythonPackage rec {
meta = with lib; {
description = "Tzinfo object for the local timezone";
homepage = "https://github.com/regebro/tzlocal";
changelog = "https://github.com/regebro/tzlocal/blob/${version}/CHANGES.txt";
license = licenses.cddl;
maintainers = with maintainers; [ dotlambda ];
};

@ -4,8 +4,9 @@
, buildPythonPackage
, flake8-import-order
, pyflakes
, mock
, tomli
, setuptools
, mock
}:
buildPythonPackage rec {
@ -22,9 +23,10 @@ buildPythonPackage rec {
disabled = !isPy3k;
propagatedBuildInputs = [
pyflakes
flake8-import-order
pyflakes
setuptools
tomli
];
checkInputs = [

@ -5,13 +5,13 @@
stdenvNoCC.mkDerivation rec {
pname = "sof-firmware";
version = "2.0";
version = "2.1.1";
src = fetchFromGitHub {
owner = "thesofproject";
repo = "sof-bin";
rev = "v${version}";
sha256 = "sha256-pDxNcDe/l1foFYuHB0w3YZidKIeH6h0IuwRmMzeMteE=";
sha256 = "sha256-/OYYfIJWMT+rBBhSCtHaSWvwRMlReEQ5y4FuMfk5zUg=";
};
dontFixup = true; # binaries must not be stripped or patchelfed

@ -1,30 +1,37 @@
{ lib, buildGoModule, fetchFromGitHub, installShellFiles }:
{ lib, buildGoModule, fetchFromGitHub, installShellFiles, jemalloc, nodejs }:
buildGoModule rec {
pname = "dgraph";
version = "20.07.3";
version = "21.12.0";
src = fetchFromGitHub {
owner = "dgraph-io";
repo = "dgraph";
rev = "v${version}";
sha256 = "sha256-yuKXcxewt64T0ToDPid37WUEhwLu+yt4tjhDQobj/Ls=";
sha256 = "sha256-OYDWr+wJEIP7raIHsXSjvuFr2ENJOllufO5ff6lxoR4=";
};
vendorSha256 = "sha256-2Ub0qdEaVSHHE5K0bNSXJFukGeSSXNpIBoUldF8jGpI=";
vendorSha256 = "sha256-YtU3Yeq/lNeq7cOB+KvHbvlH9g40WuJk1ovHxCQMG60=";
doCheck = false;
ldflags = [
"-X github.com/dgraph-io/dgraph/x.dgraphVersion=${version}-oss"
];
tags = [
"oss"
];
nativeBuildInputs = [ installShellFiles ];
# see licensing
buildPhase = ''
make oss BUILD_VERSION=${version}
'';
# todo those dependencies are required in the makefile, but verify how they are used
# actually
buildInputs = [ jemalloc nodejs ];
installPhase = ''
install dgraph/dgraph -Dt $out/bin
subPackages = [ "dgraph" ];
postInstall = ''
for shell in bash zsh; do
$out/bin/dgraph completion $shell > dgraph.$shell
installShellCompletion dgraph.$shell

@ -2,7 +2,7 @@
# Do not edit!
{
version = "2022.4.4";
version = "2022.4.5";
components = {
"abode" = ps: with ps; [
abodepy
@ -237,7 +237,8 @@
];
"backup" = ps: with ps; [
aiohttp-cors
]; # missing inputs: securetar
securetar
];
"baidu" = ps: with ps; [
]; # missing inputs: baidu-aip
"balboa" = ps: with ps; [
@ -773,7 +774,8 @@
"fail2ban" = ps: with ps; [
];
"familyhub" = ps: with ps; [
]; # missing inputs: python-family-hub-local
python-family-hub-local
];
"fan" = ps: with ps; [
];
"fastdotcom" = ps: with ps; [
@ -1534,7 +1536,8 @@
pymeteireann
];
"meteo_france" = ps: with ps; [
]; # missing inputs: meteofrance-api
meteofrance-api
];
"meteoalarm" = ps: with ps; [
meteoalertapi
];
@ -1806,7 +1809,8 @@
"nzbget" = ps: with ps; [
]; # missing inputs: pynzbgetapi
"oasa_telematics" = ps: with ps; [
]; # missing inputs: oasatelematics
oasatelematics
];
"obihai" = ps: with ps; [
pyobihai
];
@ -2189,7 +2193,8 @@
boto3
];
"rova" = ps: with ps; [
]; # missing inputs: rova
rova
];
"rpi_camera" = ps: with ps; [
];
"rpi_gpio" = ps: with ps; [
@ -3161,6 +3166,7 @@
"axis"
"azure_devops"
"azure_event_hub"
"backup"
"balboa"
"bayesian"
"binary_sensor"
@ -3416,6 +3422,7 @@
"meraki"
"met"
"met_eireann"
"meteo_france"
"meteoclimatic"
"microsoft_face"
"microsoft_face_detect"

@ -168,7 +168,7 @@ let
extraPackagesFile = writeText "home-assistant-packages" (lib.concatMapStringsSep "\n" (pkg: pkg.pname) extraBuildInputs);
# Don't forget to run parse-requirements.py after updating
hassVersion = "2022.4.4";
hassVersion = "2022.4.5";
in python.pkgs.buildPythonApplication rec {
pname = "homeassistant";
@ -186,7 +186,7 @@ in python.pkgs.buildPythonApplication rec {
owner = "home-assistant";
repo = "core";
rev = version;
hash = "sha256-NcFQjLNBGT0tNrlgkVLB3uFr9Yh4rc4HnLzALYvcrYA=";
hash = "sha256-rt1hNkjt000Wd/Wg4ThK/GJxQ/q7wysmSlzrvc4joCE=";
};
# leave this in, so users don't have to constantly update their downstream patch handling

@ -2,13 +2,13 @@
buildGoModule rec {
pname = "syft";
version = "0.44.0";
version = "0.44.1";
src = fetchFromGitHub {
owner = "anchore";
repo = pname;
rev = "v${version}";
sha256 = "sha256-w5/lTGkwH6cLzbj6/ZUJyFtwx9EyA5Q7hs/CwtsdsJA=";
sha256 = "sha256-kDTTOc2sPCKWGeVuXlGhOo2dHjBZ1QEE7jAEVs4a70U=";
# populate values that require us to use git. By doing this in postFetch we
# can delete .git afterwards and maintain better reproducibility of the src.
leaveDotGit = true;

@ -2,16 +2,16 @@
rustPlatform.buildRustPackage rec {
pname = "sic-image-cli";
version = "0.19.1";
version = "0.20.0";
src = fetchFromGitHub {
owner = "foresterre";
repo = "sic";
rev = "v${version}";
sha256 = "sha256-lFb1roQd6CK/9fKWO4SC0CbOLZglB+pS9ZqTdGRHk0Y=";
sha256 = "sha256-KoDX/d457dTHsmz8VTPhfF2MiB5vZzLFKG46/L351SQ=";
};
cargoSha256 = "sha256-9m8gdbrVv+6tGMg1GTUPy4eRqRQa9zy9BZwfZqNqSS0=";
cargoSha256 = "sha256-sKEZhJivLbos0KLzPCEnGgTCgbyWSIOvHMhoC1IaJRo=";
nativeBuildInputs = [ installShellFiles nasm ];

@ -11,11 +11,11 @@
# Note: when upgrading this package, please run the list-missing-tools.sh script as described below!
python3Packages.buildPythonApplication rec {
pname = "diffoscope";
version = "209";
version = "210";
src = fetchurl {
url = "https://diffoscope.org/archive/diffoscope-${version}.tar.bz2";
sha256 = "sha256-ZyHec9EbLrHmzCGd/Cwg/y9XCUQPtZhRH0xCouHKMp0=";
sha256 = "sha256-KaECH995scVdBbxQ4YhZUFmm04ptgi4yx14oDFX+zSw=";
};
outputs = [ "out" "man" ];

@ -9864,6 +9864,8 @@ with pkgs;
shen-sources = callPackage ../development/interpreters/shen-sources { };
shiv = with python3Packages; toPythonApplication shiv;
shocco = callPackage ../tools/text/shocco { };
shopify-cli = callPackage ../development/web/shopify-cli { };
@ -25163,7 +25165,7 @@ with pkgs;
cmctl = callPackage ../applications/networking/cluster/cmctl { };
cmus = callPackage ../applications/audio/cmus {
inherit (darwin.apple_sdk.frameworks) AudioUnit CoreAudio;
inherit (darwin.apple_sdk.frameworks) AudioUnit CoreAudio VideoToolbox;
libjack = libjack2;
};

@ -5152,6 +5152,8 @@ in {
meteoalertapi = callPackage ../development/python-modules/meteoalertapi { };
meteofrance-api = callPackage ../development/python-modules/meteofrance-api { };
mezzanine = callPackage ../development/python-modules/mezzanine { };
micawber = callPackage ../development/python-modules/micawber { };
@ -5725,6 +5727,8 @@ in {
nwdiag = callPackage ../development/python-modules/nwdiag { };
oasatelematics = callPackage ../development/python-modules/oasatelematics { };
oath = callPackage ../development/python-modules/oath { };
oauth2 = callPackage ../development/python-modules/oauth2 { };
@ -8148,6 +8152,8 @@ in {
python-editor = callPackage ../development/python-modules/python-editor { };
python-family-hub-local = callPackage ../development/python-modules/python-family-hub-local { };
python-fsutil = callPackage ../development/python-modules/python-fsutil { };
pythonefl = callPackage ../development/python-modules/python-efl { };
@ -8958,6 +8964,8 @@ in {
routes = callPackage ../development/python-modules/routes { };
rova = callPackage ../development/python-modules/rova { };
rpcq = callPackage ../development/python-modules/rpcq { };
rpdb = callPackage ../development/python-modules/rpdb { };
@ -9166,6 +9174,8 @@ in {
secure = callPackage ../development/python-modules/secure { };
securetar = callPackage ../development/python-modules/securetar { };
seekpath = callPackage ../development/python-modules/seekpath { };
segments = callPackage ../development/python-modules/segments { };
@ -9261,6 +9271,8 @@ in {
shippai = callPackage ../development/python-modules/shippai { };
shiv = callPackage ../development/python-modules/shiv { };
shodan = callPackage ../development/python-modules/shodan { };
shortuuid = callPackage ../development/python-modules/shortuuid { };

Loading…
Cancel
Save