Differences between revisions 42 and 43
Revision 42 as of 2022-11-07 20:33:27
Size: 11454
Editor: TunayDurmaz
Comment:
Revision 43 as of 2022-11-07 20:38:51
Size: 11128
Editor: TunayDurmaz
Comment:
Deletions are marked like this. Additions are marked like this.
Line 182: Line 182:
plugins(?)


Line 190: Line 186:
 1. Triggers
  1. GitHub webhooks
  1. Cron
  1. Binary build trigger
Line 196: Line 188:
 1. Jenkins Docker image, docker-coompose or docker stack deploy
Line 199: Line 190:
 1. plugins
 1. config, jcasc, config.xml, users.xml, jobs/*.xml?, gpg encrypt
Line 203: Line 192:
  1. Master only
  1. Master and agent per machine
  1. Single master and OS agents

Jenkins

Login info Jobs Settings, Plugins Binary builds require conda-build, constructor

Feedstocks

eman-dev eman-deps pydusa

General instructions Existing feedstocks Files to edit: recipe/, conda-build.yaml, conda-forge.yaml conda create -n smithy conda-smithy -c conda-forge conda-smithy rerender More info in conda-smithy/README.md, conda smithy -h, conda-forge.org/docs New feedstocks conda-smithy/README.md, conda smithy -h

Docker ?

Build System Notes

CMake

  1. Dependency binaries are pulled from Anaconda. CMake uses conda environment location to find packages.
  2. Make targets can be listed with make help. Some convenience targets are:

    $ make help
    The following are some of the valid targets for this Makefile:
    ... .....
    ... .....
    ... PythonFiles
    ... test-rt
    ... test-py-compile
    ... test-verbose-broken
    ... test-progs
    ... test-verbose
    ... .....
    ... .....
  3. libpython can be linked statically or dynamically when python is built. It is important for python extensions to be aware of the type of linking in order to avoid segfaults. This can be accomplished by querying Py_ENABLE_SHARED.

       1 python -c "import sysconfig; print(sysconfig.get_config_var('Py_ENABLE_SHARED'))"
    

    In EMAN, it is done in cmake/FindPython.cmake

  4. OpenGL detection when Anaconda's compilers are used is done using a cmake toolchain file.

  5. glext.h file needed for OpenGL related module compilation is already present on Linux and Mac. On Windows, it is manually copied once into C:\Program Files\Microsoft SDKs\Windows\v6.0A\Include\gl. On Appveyor it is downloaded as part of env setup every time a test is run.
  6. Compiler warnings are turned off by default and can be turned on by setting ENABLE_WARNINGS=ON
       1 cmake <source-dir> -DENABLE_WARNINGS=ON
    
  7. Setting compiler and linker options by include_directories, add_definitions that have global affects are avoided and target-focused design employing modern cmake concepts like interface libraries are used as much as possible.

Anaconda

Dependencies not available on anaconda or conda-forge are available cryoem. The binaries are built and uploaded using conda-forge's conda-smithy. conda-smithy takes care of generating feedstocks, registering them on GitHub and online CI services and building conda recipes.

Feedstocks

Initial Setup

Maintenance

Conda-smithy Workflow

Conda smithy uses tokens to authenticate with GitHub.

  •    1 # conda-forge.yml
       2 
       3 channels:
       4   sources: [cryoem, defaults, conda-forge]
       5   targets:
       6   - [cryoem, main]
       7 
       8 github:
       9   user_or_org: cryoem
      10   repo_name: <package>-feedstock
      11 
      12 provider:
      13   linux: circle
      14   osx: travis
      15   win: appveyor
      16 
      17 azure:
      18   build_id: blank
    
       1 # recipe/conda_build_config.yaml
       2 
       3 channel_sources:
       4 - cryoem, defaults,conda-forge
       5 channel_targets:
       6 - cryoem dev
    

Conda-smithy commands:

  •    1 conda create -n smithy conda-smithy
       2 conda activate smithy
       3 conda smithy init <recipe_directory>
       4 conda smithy register-github <feedstock_directory> --organization cryoem
       5 conda smithy register-ci --organization cryoem --without-azure --without-drone
       6 conda smithy rerender --no-check-uptodate
    

TODO

  • ABI compat, gcc dual ABI interface

Continuous Integration

  1. GitHub webhooks are setup to send notifications to blake. Blake forwards those to three build machines, although only Linux is sufficient. Linux runs the server that drives the the Jenkins jobs.

  2. Binary builds on local build machines.
    1. Manually triggered by including "[ci build]" anywhere in the last commit message. Manually triggered builds on master branch are uploaded as continuous builds and builds triggered from any other branch are uploaded to testing area.

    2. Triggered by cron builds daily.

    3. Any branch in the form of "release-" triggers continuous builds without having to include "[ci build]" in the commit message. Once the release branch is ready, release binaries are manually copied from cont. builds folder into the release folder on the server.
  3. CI configurations files:
    1. CircleCI: .circleci/config.yml

    2. TravisCI: .travis.yml

    3. Appveyor: appveyor.yml

    4. JenkinsCI: Jenkinsfile

      1. Secrets like ssh keys are stored locally in Jenkins
      2. Some env vars need to be set by agents:
        1. HOME_DIR, DEPLOY_PATH, PATH+EXTRA (to add miniconda to PATH).
        2. PATH+EXTRA is not set on win. (?)
          1. Now, it is set on win, too.
      3.  Launch method: via SSH
           Advanced:
             Prefix Start Agent Command: "D: && "
      4. On windows for sh calls in jenkins to work "Git for Windows" might need to be installed.

Jenkins Setup

Jenkins run command(?)

Server on Linux, agents on Linux, Mac and Windows

Jobs:

  • multi: Triggers job cryoem-eman2 on agents
  • cryoem-eman2: Test(?) and binary builds
  • eman-dev(?): Triggers new build of eman-dev

Jenkins Setup on Linux

Credentials

PATH

Under Construction

Jenkins Setup

  1. Jenkins master needs PATH prepended with $CONDA_PREFIX/bin
  2. docker-compose.yml at home dir in build machines
  3. TZ: https://stackoverflow.com/a/46384925

  4. Agent nodes setup, agent nodes auto-start
  5. Server and agent per machine vs single server and os agents

Linux

docker run -d -u root -p 8080:8080 -p 50000:50000 --restart unless-stopped -v /home/eman2/jenkins_home:/var/jenkins_home -v /var/run/docker.sock:/var/run/docker.sock jenkins/jenkins:lts &

docker run -d -u root --name jenkins-master -p 8080:8080 -p 50000:50000 --restart unless-stopped -v /home/eman2/jenkins_home:/var/jenkins_home -v /var/run/docker.sock:/var/run/docker.sock -e PLUGINS_FORCE_UPGRADE=true -e TRY_UPGRADE_IF_NO_MARKER=true --restart unless-stopped cryoem/jenkins:dev

Mac

slave clock sync https://blog.shameerc.com/2017/03/quick-tip-fixing-time-drift-issue-on-docker-for-mac docker run --rm --privileged alpine hwclock -s

Windows

OPENGL: https://github.com/conda/conda-recipes/blob/master/qt5/notes.md

Distribution

Binaries on cryoem.bcm.edu

EMAN2 on anaconda.org

Under Construction

EMAN2 is built with conda-build using binaries from https://anaconda.org, packaged into an installer with constructor as of v2.2.

  1. conda is the package manager.

  2. https://anaconda.org is the online repository of binaries.

  3. conda-build is the tool to build from source.

  4. constructor is the tool to package eman2 and dependency binaries into a single installer file.

EMAN2 is distributed as a single installer which includes all its dependencies. However, EMAN2 is not available as a conda-package on https://anaconda.org. In other words it is not possible to install EMAN2 by typing conda install eman2.

Conda

Packages that are available on https://anaconda.org can be installed into any conda environment by issuing the command conda install <package>. Conda installs the package along with its dependencies. In order for packages to benefit from this automation, they need to be packaged in a specific way. That can be done with conda-build. conda-build builds packages according to instructions provided in a recipe. A recipe consists of a file with package metadata, meta.yaml, and any other necessary resources like build scripts, (build.sh, bld.bat), patches and so on.

Recipes, Feedstocks and anaconda.org channel: cryoem

Most of EMAN2 dependencies can be found on anaconda's channels, defaults and conda-forge. A few that do not exist or need to be customized have been built and uploaded to channel cryoem. The recipes are hosted in separate repositories on GitHub. Every recipe repository follows the feedstock approach of conda-forge. See here for a complete list.

Binary Distribution

Constructor

Packaging is done with constructor, a tool for making installers from conda packages. In order to slightly customize the installers the project was forked. The customized project is at https://github.com/cryoem/constructor. The input files for constructor are maintained at https://github.com/cryoem/docker-images https://github.com/cryoem/build-scripts.

The installer has additional tools like conda, conda-build, pip bundled. The installer is setup so that the packages are kept in the installed EMAN2 conda environment cache for convenience.

Docker

Docker images and helper scripts are at https://github.com/cryoem/docker-images https://github.com/cryoem/build-scripts.

Command to run docker with GUI support, CentOS7:

xhost + local:root

docker run -it -v /tmp/.X11-unix:/tmp/.X11-unix -e DISPLAY=unix$DISPLAY cryoem/eman-nvidia-cuda8-centos7

# When done with eman
xhost - local:root

:FIXME: Runs as root on Linux. chown doesn't work, the resulting installer has root ownership.

EMAN2/BuildSystem (last edited 2022-12-12 03:42:44 by TunayDurmaz)