├── OWNER ├── include ├── applets │ └── __init__.py └── installclass │ └── record_partitions.py ├── DESCRIPTION ├── src ├── 411 │ ├── four11getrc │ └── version.mk ├── sec-channel │ ├── linux.mk │ ├── client │ │ └── version.mk │ └── server │ │ ├── sec-channel.service │ │ └── version.mk ├── updates.img │ ├── version.mk │ ├── pixmaps │ │ ├── first.png │ │ ├── splash.png │ │ ├── rnotes │ │ │ ├── rome.png │ │ │ └── rome.graffle │ │ │ │ └── image7.jpg │ │ ├── first-lowres.png │ │ ├── progress_first.png │ │ ├── anaconda_header.png │ │ ├── progress_first-375.png │ │ ├── first.graffle │ │ │ └── image1.png │ │ ├── splash.graffle │ │ │ └── image1.png │ │ ├── first-lowres.graffle │ │ │ └── image1.pdf │ │ ├── progress_first.graffle │ │ │ └── image2.pdf │ │ └── anaconda_header.graffle │ │ │ └── image1.pdf │ └── Makefile ├── piece-pipe │ ├── version.mk │ └── Makefile ├── rocks-java │ └── version.mk ├── screens │ ├── logo.png │ ├── bg-blue.png │ ├── version.mk │ ├── bg-purple.png │ ├── bg-yellow.png │ ├── logo-nourl.png │ ├── watermark.png │ ├── hidden.html │ ├── header.html │ ├── make-timezones.py │ ├── getrolls.html │ ├── status.html │ ├── help.html │ ├── common.css │ ├── selected.html │ ├── palette-blue.css │ ├── palette-purple.css │ └── palette-yellow.css ├── environment-modules │ ├── version.mk │ ├── modules.sh │ └── modules.csh ├── foundation-python-xml-26 │ └── version.mk ├── pxeflash │ ├── LICENSE.FreeDOS │ ├── addon │ │ └── config.sys │ ├── version.mk │ └── README ├── serf │ ├── serf.conf │ ├── scriptlets │ ├── logrotate_serf │ ├── serf.json │ ├── rocks-serf.service │ └── version.mk ├── usersguide │ ├── images │ │ ├── ipmi.png │ │ ├── rocks.png │ │ ├── cluster.png │ │ ├── central-1.png │ │ ├── central-2.png │ │ ├── central-3.png │ │ ├── install │ │ │ ├── 2.png │ │ │ ├── 20.png │ │ │ ├── 21.png │ │ │ ├── 22.png │ │ │ ├── 23.png │ │ │ ├── 24.png │ │ │ ├── 25.png │ │ │ ├── 26.png │ │ │ ├── 27.png │ │ │ ├── 271.png │ │ │ ├── 28.png │ │ │ ├── 29.png │ │ │ ├── 3.png │ │ │ ├── 30.png │ │ │ ├── 31.png │ │ │ ├── 32.png │ │ │ ├── 33.png │ │ │ ├── 4.png │ │ │ ├── 41.png │ │ │ ├── 5.png │ │ │ ├── 7.png │ │ │ └── splash.png │ │ ├── scanbarcode.png │ │ ├── LogicalVlans.png │ │ ├── PhysicalVlans.png │ │ ├── ipmi-sun-node.png │ │ ├── PhysicalNetworks.png │ │ ├── insert-ethers-1.png │ │ ├── insert-ethers-2.png │ │ ├── insert-ethers-3.png │ │ ├── insert-ethers-4.png │ │ ├── insert-ethers-5.png │ │ ├── add-new-appliance.png │ │ ├── GoogleAuthenticator1.png │ │ ├── GoogleAuthenticator2.png │ │ ├── GooglePlayAuthenticator.png │ │ ├── LogicalVlansAssociated.png │ │ ├── install7 │ │ │ ├── list-rolls-640.png │ │ │ ├── network-up-640.png │ │ │ ├── no-network-640.png │ │ │ ├── begin-ready-640.png │ │ │ ├── mid-install-640.png │ │ │ ├── need-rootpw-640.png │ │ │ ├── rolls-added-640.png │ │ │ ├── choose-language-640.png │ │ │ ├── ipv6-link-local-640.png │ │ │ ├── network-public-640.png │ │ │ ├── rolls-selected-640.png │ │ │ ├── configure-cluster-640.png │ │ │ ├── configure-contact-640.png │ │ │ ├── downloading-rolls-640.png │ │ │ ├── network-configured-640.png │ │ │ ├── partition-confirm-640.png │ │ │ ├── rocks-boot-screen-640.png │ │ │ ├── root-password-set-640.png │ │ │ ├── timezone-selection-640.png │ │ │ ├── configure-cluster-1-640.png │ │ │ ├── no-network-scrolled-640.png │ │ │ ├── select-partitioning-640.png │ │ │ ├── installation-complete-640.png │ │ │ ├── manual-network-frontend-640.png │ │ │ ├── private-network-setup-640.png │ │ │ ├── private-network-complete-640.png │ │ │ ├── configure-partition-manual-640.png │ │ │ └── rolls-selected-main-screen-640.png │ │ └── AppleiTunesAuthenticator.png │ ├── appendix │ │ ├── licenses.sgml │ │ ├── index.sgml │ │ ├── entities.sgml │ │ └── faq │ │ │ ├── mysql.sgml │ │ │ ├── no-kickstart.sgml │ │ │ ├── startx.sgml │ │ │ ├── removenode.sgml │ │ │ ├── smalldisk.sgml │ │ │ ├── mirroring.sgml │ │ │ ├── entities.sgml │ │ │ ├── apache-compute.sgml │ │ │ ├── index.sgml │ │ │ ├── installation-troubleshoot.sgml │ │ │ ├── mpirun-linpack.sgml │ │ │ ├── floppy.sgml │ │ │ ├── security-updates.sgml │ │ │ └── 411.sgml │ ├── community │ │ ├── wiki.sgml │ │ ├── training.sgml │ │ ├── rockson.sqml │ │ ├── foundation.sgml │ │ ├── index.sgml │ │ ├── releases.sgml │ │ ├── entities.sgml │ │ ├── office-hours.sgml │ │ └── mailing-list.sgml │ ├── networking │ │ ├── entities.sgml │ │ └── index.sgml │ ├── advanced │ │ ├── index.sgml │ │ ├── entities.sgml │ │ └── central.sgml │ ├── version.mk │ ├── clusterinstall │ │ ├── entities.sgml │ │ ├── index.sgml │ │ ├── supported.sgml │ │ ├── minimum.sgml │ │ └── physical.sgml │ ├── customizing │ │ ├── index.sgml │ │ ├── entities.sgml │ │ ├── adding-applications.sgml │ │ ├── boot-floppy.sgml │ │ ├── rsh.sgml │ │ ├── postconfig.sgml │ │ └── password.sgml │ └── preface.sgml ├── 411-alert │ ├── 411alert.json │ ├── version.mk │ └── 411event.py ├── config-server-profile │ └── version.mk ├── developersguiderst │ ├── README │ ├── refs │ │ ├── modules.rst │ │ ├── rocks.commands.rst │ │ ├── rocks.db.mappings.rst │ │ └── rocks.db.rst │ └── index.rst ├── sec_attr_plugins │ ├── version.mk │ └── plugins │ │ ├── ssh_host_rsa_key.py │ │ ├── ssh_host_dsa_key.py │ │ ├── ssh_host_ecdsa_key.py │ │ ├── ssh_host_ed25519_key.py │ │ └── root_pw.py ├── phpMyAdmin │ └── version.mk ├── anaconda-yum-plugins │ ├── version.mk │ └── anaconda-yum-plugins.spec.in ├── developersguide │ ├── images │ │ ├── base-graph.png │ │ ├── developer.png │ │ ├── base-valgrind-zoom.png │ │ ├── base-valgrind-graph.png │ │ └── base-valgrind-simple-graph.png │ ├── version.mk │ ├── overview.sgml │ └── index.sgml ├── rocks-cluster-homepage │ ├── rocks-logo.png │ └── version.mk ├── rocks-anaconda-updates │ ├── version.mk │ ├── Makefile │ ├── .rpmmacros │ ├── rocks-anaconda-updates.spec.mk │ └── rocks-anaconda-updates.spec ├── ssl │ └── version.mk ├── channel │ ├── sunos.mk │ ├── channeld.service │ └── version.mk ├── foundation-ant │ └── version.mk ├── snmp-status │ ├── version.mk │ └── index.html.in ├── backup │ └── version.mk ├── yum-extras │ ├── i386.mk │ ├── x86_64.mk │ └── version.mk ├── java │ ├── rpmname.awk │ ├── i386.mk │ └── x86_64.mk ├── kickstart │ └── version.mk ├── mosh │ └── version.mk ├── anaconda │ ├── build │ └── version.mk ├── foundation-tidy │ └── version.mk ├── foundation-gawk │ └── version.mk ├── protobuf │ ├── README.md │ └── version.mk ├── ekv │ ├── x86_64.mk │ ├── version.mk │ └── check_port.py ├── 411-master │ └── version.mk ├── qrencode │ └── version.mk ├── tentakel │ └── version.mk ├── google-otp │ ├── google-otp │ └── version.mk ├── firerox │ ├── version.mk │ ├── user.js │ ├── mozconfig │ └── userChrome.css ├── pdsh │ └── version.mk └── extras │ └── version.mk ├── README.hacking ├── version.mk ├── nodes ├── compilers.xml ├── scripting.xml ├── mosh.xml ├── client-rootpw.xml ├── ipmi-server.xml ├── md-rebuild.xml ├── server-rootpw.xml ├── ipmi-client.xml ├── nas-partition.xml ├── profile-server.xml ├── null.xml ├── homepage.xml ├── x11-server.xml ├── partition.xml ├── bootstrap-dummy.xml ├── compute.xml ├── custom-partition.xml ├── grub2-client.xml ├── devel-server.xml ├── yum-extras.xml ├── client-firewall.xml ├── serf-common.xml ├── google-otp-server.xml ├── nas.xml ├── wan.xml ├── rlo.xml ├── foundation-client.xml ├── devel.xml ├── ipmi-base.xml ├── auto-kickstart.xml ├── selinux.xml ├── google-otp.xml ├── foundation-server.xml ├── root.xml ├── fail2ban.xml ├── routes-client.xml ├── ethers-server.xml ├── fortran-development.xml ├── nas-exports.xml ├── rpc.xml ├── serf-server.xml ├── postshell.xml ├── networking-server.xml ├── tcl-development.xml ├── docbook.xml ├── partitions-save.xml ├── ip-diag.xml ├── 411-base.xml ├── autofs-client.xml ├── debug-post.xml ├── iscsi.xml ├── debug-pre.xml ├── emacs.xml └── snmp-client.xml ├── README ├── tests ├── t2007 │ ├── interfaces │ ├── rolls.xml │ └── test_includes.py ├── t6000 │ └── plugin_device.py ├── t2009 │ ├── test2.xml │ ├── test1.xml │ ├── test3.xml │ └── test.xml ├── aggregate-results.sh ├── t2012-test-config-host-interface.sh ├── t2010-rocks-appliances.sh ├── t2011-creation-db-schema.sh ├── t2002-rocks-dhcpd.sh └── t2009-ks-file-tag.sh ├── include-version ├── 6 │ ├── graphs │ │ └── default │ │ │ └── rocks6.xml │ └── nodes │ │ └── upstart.xml └── 7 │ ├── graphs │ └── default │ │ └── rocks6.xml │ └── nodes │ └── upstart.xml ├── graphs └── default │ ├── rocks7.xml │ ├── serf.xml │ └── root.xml ├── OSROLL ├── roll-distro-update.xml.in └── version.mk └── manifest /OWNER: -------------------------------------------------------------------------------- 1 | Rocks 2 | -------------------------------------------------------------------------------- /include/applets/__init__.py: -------------------------------------------------------------------------------- 1 | -------------------------------------------------------------------------------- /DESCRIPTION: -------------------------------------------------------------------------------- 1 | Core of Rocks 2 | -------------------------------------------------------------------------------- /src/sec-channel/linux.mk: -------------------------------------------------------------------------------- 1 | SRCDIRS += server 2 | -------------------------------------------------------------------------------- /README.hacking: -------------------------------------------------------------------------------- 1 | Checkout src/developersguiderst/hacking.rst 2 | -------------------------------------------------------------------------------- /src/updates.img/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-anaconda-updates 2 | RELEASE = 2 3 | 4 | -------------------------------------------------------------------------------- /version.mk: -------------------------------------------------------------------------------- 1 | RELEASE = 3 2 | COLOR = yellow 3 | 4 | REDHAT.ROOT = $(CURDIR) 5 | -------------------------------------------------------------------------------- /src/piece-pipe/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-piece-pipe 2 | RELEASE = 0 3 | RPM.FILES = /usr/sbin/* 4 | -------------------------------------------------------------------------------- /src/rocks-java/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-java 2 | RELEASE = 1 3 | RPM.FILES = "/etc/profile.d/*" 4 | -------------------------------------------------------------------------------- /src/screens/logo.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/logo.png -------------------------------------------------------------------------------- /nodes/compilers.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | -------------------------------------------------------------------------------- /nodes/scripting.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | -------------------------------------------------------------------------------- /src/screens/bg-blue.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/bg-blue.png -------------------------------------------------------------------------------- /src/screens/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-screens 2 | RELEASE = 0 3 | RPM.FILES = /opt/rocks/screens 4 | -------------------------------------------------------------------------------- /src/environment-modules/version.mk: -------------------------------------------------------------------------------- 1 | NAME = environment-modules 2 | VERSION = 3.2.9c 3 | RELEASE = 1 4 | -------------------------------------------------------------------------------- /src/screens/bg-purple.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/bg-purple.png -------------------------------------------------------------------------------- /src/screens/bg-yellow.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/bg-yellow.png -------------------------------------------------------------------------------- /src/screens/logo-nourl.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/logo-nourl.png -------------------------------------------------------------------------------- /src/screens/watermark.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/screens/watermark.png -------------------------------------------------------------------------------- /README: -------------------------------------------------------------------------------- 1 | 2 | 3 | This is the base roll of the Rocks clusters toolkit. 4 | http://www.rocksclusters.org 5 | 6 | -------------------------------------------------------------------------------- /src/foundation-python-xml-26/version.mk: -------------------------------------------------------------------------------- 1 | NAME = foundation-python-xml-26 2 | VERSION = 0.8.4 3 | RELEASE = 0 4 | -------------------------------------------------------------------------------- /src/pxeflash/LICENSE.FreeDOS: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/pxeflash/LICENSE.FreeDOS -------------------------------------------------------------------------------- /src/serf/serf.conf: -------------------------------------------------------------------------------- 1 | if $programname == 'serf' then /var/log/serf.log 2 | if $programname == 'serf' then ~ 3 | -------------------------------------------------------------------------------- /src/sec-channel/client/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-sec-channel-client 2 | RELEASE = 1 3 | RPM.FILES = /opt/rocks/bin/* 4 | -------------------------------------------------------------------------------- /src/usersguide/images/ipmi.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/ipmi.png -------------------------------------------------------------------------------- /src/usersguide/images/rocks.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/rocks.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/first.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/first.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/splash.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/splash.png -------------------------------------------------------------------------------- /src/usersguide/images/cluster.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/cluster.png -------------------------------------------------------------------------------- /src/411-alert/411alert.json: -------------------------------------------------------------------------------- 1 | { 2 | "event_handlers": [ 3 | "user:411Alert=/opt/rocks/bin/411event" 4 | ] 5 | } 6 | 7 | -------------------------------------------------------------------------------- /src/config-server-profile/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-config-server-profile 2 | RELEASE = 1 3 | RPM.FILES = /etc/profile.d/* 4 | -------------------------------------------------------------------------------- /src/developersguiderst/README: -------------------------------------------------------------------------------- 1 | 2 | Run 3 | make html 4 | to compile the developersguide. Output will be in _build/html/ 5 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-secattr-plugins 2 | RELEASE = 1 3 | RPM.FILES = /opt/rocks/var/plugins/sec_attr 4 | -------------------------------------------------------------------------------- /src/usersguide/appendix/licenses.sgml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/appendix/licenses.sgml -------------------------------------------------------------------------------- /src/usersguide/images/central-1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/central-1.png -------------------------------------------------------------------------------- /src/usersguide/images/central-2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/central-2.png -------------------------------------------------------------------------------- /src/usersguide/images/central-3.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/central-3.png -------------------------------------------------------------------------------- /src/usersguide/images/install/2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/2.png -------------------------------------------------------------------------------- /src/usersguide/images/install/20.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/20.png -------------------------------------------------------------------------------- /src/usersguide/images/install/21.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/21.png -------------------------------------------------------------------------------- /src/usersguide/images/install/22.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/22.png -------------------------------------------------------------------------------- /src/usersguide/images/install/23.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/23.png -------------------------------------------------------------------------------- /src/usersguide/images/install/24.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/24.png -------------------------------------------------------------------------------- /src/usersguide/images/install/25.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/25.png -------------------------------------------------------------------------------- /src/usersguide/images/install/26.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/26.png -------------------------------------------------------------------------------- /src/usersguide/images/install/27.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/27.png -------------------------------------------------------------------------------- /src/usersguide/images/install/271.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/271.png -------------------------------------------------------------------------------- /src/usersguide/images/install/28.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/28.png -------------------------------------------------------------------------------- /src/usersguide/images/install/29.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/29.png -------------------------------------------------------------------------------- /src/usersguide/images/install/3.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/3.png -------------------------------------------------------------------------------- /src/usersguide/images/install/30.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/30.png -------------------------------------------------------------------------------- /src/usersguide/images/install/31.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/31.png -------------------------------------------------------------------------------- /src/usersguide/images/install/32.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/32.png -------------------------------------------------------------------------------- /src/usersguide/images/install/33.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/33.png -------------------------------------------------------------------------------- /src/usersguide/images/install/4.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/4.png -------------------------------------------------------------------------------- /src/usersguide/images/install/41.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/41.png -------------------------------------------------------------------------------- /src/usersguide/images/install/5.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/5.png -------------------------------------------------------------------------------- /src/usersguide/images/install/7.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/7.png -------------------------------------------------------------------------------- /src/usersguide/images/scanbarcode.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/scanbarcode.png -------------------------------------------------------------------------------- /src/phpMyAdmin/version.mk: -------------------------------------------------------------------------------- 1 | NAME = phpMyAdmin 2 | VERSION = 2.6.4 3 | RELEASE = pl1 4 | RPM.FILES = /var/www/html/admin/phpMyAdmin 5 | -------------------------------------------------------------------------------- /src/updates.img/pixmaps/rnotes/rome.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/rnotes/rome.png -------------------------------------------------------------------------------- /src/usersguide/images/LogicalVlans.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/LogicalVlans.png -------------------------------------------------------------------------------- /src/usersguide/images/PhysicalVlans.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/PhysicalVlans.png -------------------------------------------------------------------------------- /src/usersguide/images/ipmi-sun-node.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/ipmi-sun-node.png -------------------------------------------------------------------------------- /src/anaconda-yum-plugins/version.mk: -------------------------------------------------------------------------------- 1 | NAME = anaconda-yum-plugins 2 | RELEASE = 5 3 | MAKE.iscontrib = 1 4 | 5 | VERSION = 1.0 6 | 7 | -------------------------------------------------------------------------------- /src/developersguide/images/base-graph.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/developersguide/images/base-graph.png -------------------------------------------------------------------------------- /src/developersguide/images/developer.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/developersguide/images/developer.png -------------------------------------------------------------------------------- /src/developersguiderst/refs/modules.rst: -------------------------------------------------------------------------------- 1 | API Reference 2 | ============= 3 | 4 | .. toctree:: 5 | :maxdepth: 4 6 | 7 | rocks 8 | -------------------------------------------------------------------------------- /src/rocks-cluster-homepage/rocks-logo.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/rocks-cluster-homepage/rocks-logo.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/first-lowres.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/first-lowres.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/progress_first.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/progress_first.png -------------------------------------------------------------------------------- /src/usersguide/images/PhysicalNetworks.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/PhysicalNetworks.png -------------------------------------------------------------------------------- /src/usersguide/images/insert-ethers-1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/insert-ethers-1.png -------------------------------------------------------------------------------- /src/usersguide/images/insert-ethers-2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/insert-ethers-2.png -------------------------------------------------------------------------------- /src/usersguide/images/insert-ethers-3.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/insert-ethers-3.png -------------------------------------------------------------------------------- /src/usersguide/images/insert-ethers-4.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/insert-ethers-4.png -------------------------------------------------------------------------------- /src/usersguide/images/insert-ethers-5.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/insert-ethers-5.png -------------------------------------------------------------------------------- /src/usersguide/images/install/splash.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install/splash.png -------------------------------------------------------------------------------- /src/pxeflash/addon/config.sys: -------------------------------------------------------------------------------- 1 | FILES=20 2 | BUFFERS=20 3 | !rem SHELL=A:\COMMAND.COM /E:512 /MSG /P 4 | SHELLHIGH=\COMMAND.COM /E:256 /P 5 | -------------------------------------------------------------------------------- /src/updates.img/pixmaps/anaconda_header.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/anaconda_header.png -------------------------------------------------------------------------------- /src/usersguide/images/add-new-appliance.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/add-new-appliance.png -------------------------------------------------------------------------------- /tests/t2007/interfaces: -------------------------------------------------------------------------------- 1 | X-RHN-Provisioning-MAC-0: eth0 00:14:4F:58:0F:E0 none 2 | X-RHN-Provisioning-MAC-1: eth1 00:14:4F:58:0F:E1 none ks 3 | -------------------------------------------------------------------------------- /src/updates.img/pixmaps/progress_first-375.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/progress_first-375.png -------------------------------------------------------------------------------- /src/usersguide/images/GoogleAuthenticator1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/GoogleAuthenticator1.png -------------------------------------------------------------------------------- /src/usersguide/images/GoogleAuthenticator2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/GoogleAuthenticator2.png -------------------------------------------------------------------------------- /src/developersguide/images/base-valgrind-zoom.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/developersguide/images/base-valgrind-zoom.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/first.graffle/image1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/first.graffle/image1.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/splash.graffle/image1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/splash.graffle/image1.png -------------------------------------------------------------------------------- /src/usersguide/images/GooglePlayAuthenticator.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/GooglePlayAuthenticator.png -------------------------------------------------------------------------------- /src/usersguide/images/LogicalVlansAssociated.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/LogicalVlansAssociated.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/list-rolls-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/list-rolls-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/network-up-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/network-up-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/no-network-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/no-network-640.png -------------------------------------------------------------------------------- /src/developersguide/images/base-valgrind-graph.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/developersguide/images/base-valgrind-graph.png -------------------------------------------------------------------------------- /src/rocks-anaconda-updates/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-anaconda-updates 2 | RELEASE = 2 3 | PKGROOT = /RedHat/base 4 | RPM.FILES = $(PKGROOT)/* 5 | 6 | 7 | -------------------------------------------------------------------------------- /src/serf/scriptlets: -------------------------------------------------------------------------------- 1 | %post 2 | # enable the services 3 | /usr/bin/systemctl enable rocks-serf 4 | %postun 5 | /usr/bin/systemctl disable rocks-serf 6 | 7 | -------------------------------------------------------------------------------- /src/ssl/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-ssl 2 | RELEASE = 0 3 | RPM.FILES = \ 4 | /opt/rocks/doc/* \\n\ 5 | /opt/rocks/man/man1/* \\n\ 6 | /opt/rocks/sbin/* 7 | 8 | -------------------------------------------------------------------------------- /src/usersguide/images/AppleiTunesAuthenticator.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/AppleiTunesAuthenticator.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/begin-ready-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/begin-ready-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/mid-install-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/mid-install-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/need-rootpw-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/need-rootpw-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/rolls-added-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/rolls-added-640.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/rnotes/rome.graffle/image7.jpg: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/rnotes/rome.graffle/image7.jpg -------------------------------------------------------------------------------- /src/usersguide/images/install7/choose-language-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/choose-language-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/ipv6-link-local-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/ipv6-link-local-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/network-public-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/network-public-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/rolls-selected-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/rolls-selected-640.png -------------------------------------------------------------------------------- /src/channel/sunos.mk: -------------------------------------------------------------------------------- 1 | install:: 2 | mkdir -p $(ROOT)/$(PKGROOT)/share/channeld/ 3 | $(INSTALL) -m0644 channeld.xml $(ROOT)/$(PKGROOT)/share/channeld/channeld.xml 4 | -------------------------------------------------------------------------------- /src/developersguide/images/base-valgrind-simple-graph.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/developersguide/images/base-valgrind-simple-graph.png -------------------------------------------------------------------------------- /src/foundation-ant/version.mk: -------------------------------------------------------------------------------- 1 | NAME = foundation-ant 2 | VERSION = 1.7.1 3 | RELEASE = 1 4 | RPM.FILES = "/etc/profile.d/*\\n/opt/rocks/bin/*\\n/opt/rocks/lib/*" 5 | -------------------------------------------------------------------------------- /src/snmp-status/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-snmp-status 2 | RELEASE = 0 3 | RPM.FILES = \ 4 | /opt/rocks/bin/* \\n\ 5 | /opt/rocks/doc/* \\n\ 6 | /opt/rocks/man/man1/* 7 | -------------------------------------------------------------------------------- /src/updates.img/pixmaps/first-lowres.graffle/image1.pdf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/first-lowres.graffle/image1.pdf -------------------------------------------------------------------------------- /src/updates.img/pixmaps/progress_first.graffle/image2.pdf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/progress_first.graffle/image2.pdf -------------------------------------------------------------------------------- /src/usersguide/images/install7/configure-cluster-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/configure-cluster-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/configure-contact-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/configure-contact-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/downloading-rolls-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/downloading-rolls-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/network-configured-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/network-configured-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/partition-confirm-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/partition-confirm-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/rocks-boot-screen-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/rocks-boot-screen-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/root-password-set-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/root-password-set-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/timezone-selection-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/timezone-selection-640.png -------------------------------------------------------------------------------- /src/updates.img/pixmaps/anaconda_header.graffle/image1.pdf: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/updates.img/pixmaps/anaconda_header.graffle/image1.pdf -------------------------------------------------------------------------------- /src/usersguide/images/install7/configure-cluster-1-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/configure-cluster-1-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/no-network-scrolled-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/no-network-scrolled-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/select-partitioning-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/select-partitioning-640.png -------------------------------------------------------------------------------- /src/developersguide/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-developersguide 2 | RELEASE = 3 3 | RPM.ARCH = noarch 4 | RPM.FILES = \ 5 | /var/www/html/roll-documentation/developers-guide 6 | -------------------------------------------------------------------------------- /src/usersguide/images/install7/installation-complete-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/installation-complete-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/manual-network-frontend-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/manual-network-frontend-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/private-network-setup-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/private-network-setup-640.png -------------------------------------------------------------------------------- /src/backup/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-backup 2 | RELEASE = 1 3 | RPM.FILES = \ 4 | /etc/cron.daily/* \\n\ 5 | /opt/rocks/bin/* \\n\ 6 | /opt/rocks/tmp \\n\ 7 | /var/www/cgi-bin/* 8 | -------------------------------------------------------------------------------- /src/usersguide/images/install7/private-network-complete-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/private-network-complete-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/configure-partition-manual-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/configure-partition-manual-640.png -------------------------------------------------------------------------------- /src/usersguide/images/install7/rolls-selected-main-screen-640.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/rocksclusters/base/HEAD/src/usersguide/images/install7/rolls-selected-main-screen-640.png -------------------------------------------------------------------------------- /src/yum-extras/i386.mk: -------------------------------------------------------------------------------- 1 | ifeq ($(VERSION.MAJOR),5) 2 | REPORPMS += rpmforge-release-0.5.2-2.el5.rf.i386.rpm 3 | else 4 | REPORPMS += rpmforge-release-0.5.2-2.el6.rf.i686.rpm 5 | endif 6 | -------------------------------------------------------------------------------- /src/java/rpmname.awk: -------------------------------------------------------------------------------- 1 | /^Name/ {name=$NF} 2 | /^Version/ {version=$NF} 3 | /^Release/ {release=$NF} 4 | /^Architecture/ {arch=$NF} 5 | END{printf("%s-%s-%s.%s.rpm",name,version,release,arch)} 6 | -------------------------------------------------------------------------------- /src/kickstart/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-kickstart 2 | RELEASE = 0 3 | RPM.FILES = \ 4 | /etc/rc.d/init.d/* \\n\ 5 | /export/rocks/install/sbin/* \\n\ 6 | /export/rocks/install/site-profiles 7 | -------------------------------------------------------------------------------- /src/mosh/version.mk: -------------------------------------------------------------------------------- 1 | PKGROOT = /usr 2 | NAME = mosh 3 | VERSION = 1.2.4 4 | RELEASE = 1 5 | TARBALL_POSTFIX = tar.gz 6 | RPM.FILES = \ 7 | /usr/bin/* \\n \ 8 | /usr/share/man/man1/* 9 | -------------------------------------------------------------------------------- /src/yum-extras/x86_64.mk: -------------------------------------------------------------------------------- 1 | ifeq ($(VERSION.MAJOR),5) 2 | REPORPMS += rpmforge-release-0.5.2-2.el5.rf.x86_64.rpm 3 | else 4 | REPORPMS += rpmforge-release-0.5.2-2.el6.rf.x86_64.rpm 5 | endif 6 | -------------------------------------------------------------------------------- /src/anaconda/build: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | 3 | echo rpm -Uvh $ROCKS_REBUILD_SRPM 4 | rpm -Uvh $ROCKS_REBUILD_SRPM 5 | 6 | export SPECFILE=anaconda.spec 7 | 8 | rpmbuild -bb /tmp/$SPECFILE 9 | 10 | -------------------------------------------------------------------------------- /src/usersguide/community/wiki.sgml: -------------------------------------------------------------------------------- 1 |
2 | Rocks Wiki 3 | 4 | 5 | Contribute to the wiki. 6 | 7 | 8 |
9 | 10 | 11 | -------------------------------------------------------------------------------- /tests/t2007/rolls.xml: -------------------------------------------------------------------------------- 1 | 2 | 9 | 10 | -------------------------------------------------------------------------------- /src/java/i386.mk: -------------------------------------------------------------------------------- 1 | rpm:: build 2 | ( for j in *i586.rpm; do \ 3 | destname=`/usr/bin/rpm -qip $$j | awk -f rpmname.awk`; \ 4 | /bin/cp $$j $(REDHAT.RPMS)/$(ARCH)/$$destname;\ 5 | done; \ 6 | ) 7 | -------------------------------------------------------------------------------- /src/foundation-tidy/version.mk: -------------------------------------------------------------------------------- 1 | NAME = foundation-tidy 2 | TIDYCVS = :pserver:anonymous@tidy.cvs.sourceforge.net:/cvsroot/tidy 3 | VERSION = $(shell date +%m_%d_%y) 4 | RELEASE = 0 5 | RPM.FILES = "/opt/rocks/bin/*" 6 | -------------------------------------------------------------------------------- /src/java/x86_64.mk: -------------------------------------------------------------------------------- 1 | rpm:: build 2 | ( for j in *i586.rpm *x64.rpm ; do \ 3 | destname=`/usr/bin/rpm -qip $$j | awk -f rpmname.awk`; \ 4 | /bin/cp $$j $(REDHAT.RPMS)/$(ARCH)/$$destname;\ 5 | done; \ 6 | ) 7 | -------------------------------------------------------------------------------- /src/serf/logrotate_serf: -------------------------------------------------------------------------------- 1 | /var/log/serf.log { 2 | daily 3 | missingok 4 | rotate 7 5 | compress 6 | delaycompress 7 | copytruncate 8 | minsize 100k 9 | } 10 | -------------------------------------------------------------------------------- /src/usersguide/community/training.sgml: -------------------------------------------------------------------------------- 1 |
2 | Training Material 3 | 4 | 5 | Rocks slides. 6 | 7 | 8 |
9 | 10 | 11 | -------------------------------------------------------------------------------- /src/411/four11getrc: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 7 | -------------------------------------------------------------------------------- /src/environment-modules/modules.sh: -------------------------------------------------------------------------------- 1 | shell=`/bin/basename \`/bin/ps -p $$ -ocomm=\`` 2 | if [ -f /usr/share/Modules/init/$shell ] 3 | then 4 | . /usr/share/Modules/init/$shell 5 | else 6 | . /usr/share/Modules/init/sh 7 | fi 8 | -------------------------------------------------------------------------------- /src/usersguide/community/rockson.sqml: -------------------------------------------------------------------------------- 1 |
2 | RocksOn Podcast 3 | 4 | 5 | Listen to us blab about Rocks. 6 | 7 | 8 |
9 | 10 | 11 | -------------------------------------------------------------------------------- /include-version/6/graphs/default/rocks6.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | -------------------------------------------------------------------------------- /include-version/7/graphs/default/rocks6.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | -------------------------------------------------------------------------------- /src/foundation-gawk/version.mk: -------------------------------------------------------------------------------- 1 | NAME = foundation-gawk 2 | VERSION = 3.1.5 3 | RELEASE = 1 4 | RPM.FILES = "/opt/rocks/bin/*\\n/opt/rocks/info/*\\n/opt/rocks/libexec/awk\\n/opt/rocks/man/man1/*\\n/opt/rocks/share/awk\\n/opt/rocks/share/locale/*/LC_MESSAGES/*" 5 | -------------------------------------------------------------------------------- /src/protobuf/README.md: -------------------------------------------------------------------------------- 1 | This is google's protocol buffers library for C++, Java, and Python. 2 | See: https://github.com/google/protobuf/ 3 | For full code. 4 | The googletest tar ball was retrieved from git 5 | https://github.com/google/googletest/releases 6 | -------------------------------------------------------------------------------- /src/usersguide/community/foundation.sgml: -------------------------------------------------------------------------------- 1 |
2 | Rocks Foundation 3 | 4 | 5 | Rocks Foundation. 6 | Non-profit. 7 | Cafepress. 8 | 9 | 10 |
11 | 12 | 13 | -------------------------------------------------------------------------------- /nodes/mosh.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | MOSH - Mobile Shell from MIT 7 | 8 | 9 | protobuf 10 | mosh 11 | 12 | 13 | 14 | -------------------------------------------------------------------------------- /src/ekv/x86_64.mk: -------------------------------------------------------------------------------- 1 | # 2 | # $Id: x86_64.mk,v 1.1 2005/03/01 02:02:47 mjk Exp $ 3 | # 4 | # The Architecture dependant Makefile variables for 5 | # AMD Opteron machines. 6 | # 7 | 8 | # Most things regularly in /lib are now in /lib64. 9 | 10 | LIB=lib64 11 | -------------------------------------------------------------------------------- /src/developersguiderst/refs/rocks.commands.rst: -------------------------------------------------------------------------------- 1 | commands Package 2 | ================ 3 | 4 | :mod:`commands` Package 5 | ----------------------- 6 | 7 | .. automodule:: rocks.commands 8 | :members: 9 | :undoc-members: 10 | :show-inheritance: 11 | 12 | -------------------------------------------------------------------------------- /src/developersguiderst/refs/rocks.db.mappings.rst: -------------------------------------------------------------------------------- 1 | mappings Package 2 | ================ 3 | 4 | :mod:`base` Module 5 | ------------------ 6 | 7 | .. automodule:: rocks.db.mappings.base 8 | :members: 9 | :undoc-members: 10 | :show-inheritance: 11 | 12 | -------------------------------------------------------------------------------- /src/pxeflash/version.mk: -------------------------------------------------------------------------------- 1 | NAME = pxeflash 2 | VERSION = 1.0 3 | RELEASE = 0 4 | DOSBOOT = FDSTD.288 5 | PXEIMAGE = $(NAME).img 6 | PKGROOT = /opt/pxeflash 7 | MEMDISK = /usr/share/syslinux/memdisk 8 | RPM.FILES = \ 9 | /opt/pxeflash \\n\ 10 | /tftpboot/pxelinux/* 11 | -------------------------------------------------------------------------------- /src/usersguide/appendix/index.sgml: -------------------------------------------------------------------------------- 1 | &rocks-faq; 2 | &changelog; 3 | &service-411; 4 | &passwd_sec; 5 | 6 | Kickstart Nodes Reference 7 | &base-nodes; 8 | 9 | 10 | &licenses; 11 | 12 | -------------------------------------------------------------------------------- /src/anaconda/version.mk: -------------------------------------------------------------------------------- 1 | NAME = anaconda 2 | RELEASE = 1 3 | MAKE.iscontrib = 1 4 | 5 | ifeq ($(strip $(VERSION.MAJOR)), 5) 6 | VERSION = 11.1.2.259 7 | SRPMRELEASE = $(RELEASE) 8 | else 9 | VERSION = 13.21.239 10 | SRPMRELEASE = $(RELEASE).el6 11 | endif 12 | 13 | -------------------------------------------------------------------------------- /src/protobuf/version.mk: -------------------------------------------------------------------------------- 1 | PKGROOT = /usr 2 | NAME = protobuf 3 | VERSION = 2.6.1 4 | RELEASE = 2 5 | GTEST = googletest-release 6 | GTEST_RELEASE = 1.5.0 7 | TARBALL_POSTFIX = tar.gz 8 | RPM.FILES = "/usr/bin/*\\n/usr/include/google\\n/usr/lib64/lib*\\n/usr/lib64/pkgconfig/*" 9 | -------------------------------------------------------------------------------- /src/screens/hidden.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 |
10 |
11 | 12 |
13 | 14 | 15 | 16 | 17 | 18 | -------------------------------------------------------------------------------- /src/serf/serf.json: -------------------------------------------------------------------------------- 1 | 2 | { 3 | "event_handlers": [ 4 | "user:411_deploy=/opt/rocks/bin/411get --all" 5 | ], 6 | "rejoin_after_leave": true, 7 | "interface": "eth0", 8 | "retry_join": ["comet-dev"], 9 | "enable_syslog": true, 10 | "log_level": "info" 11 | } 12 | 13 | -------------------------------------------------------------------------------- /src/411-master/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-411-master 2 | RELEASE = 2 3 | RPM.ARCH = noarch 4 | RPM.FILE.EXTRAS = "%config /var/411/Files.mk" 5 | RPM.FILES = \ 6 | /etc/411.d \\n\ 7 | /etc/rc.d/init.d/* \\n\ 8 | /opt/rocks/sbin/* \\n\ 9 | /opt/rocks/var/plugins/411 \\n\ 10 | /var/411 11 | -------------------------------------------------------------------------------- /src/usersguide/networking/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | -------------------------------------------------------------------------------- /src/usersguide/networking/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Defining and Modifying Networks and Network Interfaces 3 | 4 | &networking-overview; 5 | &networking-vlan; 6 | &networking-ipmi; 7 | &networking-examples; 8 | 9 | 10 | 11 | -------------------------------------------------------------------------------- /src/rocks-anaconda-updates/Makefile: -------------------------------------------------------------------------------- 1 | REDHAT.ROOT = $(CURDIR)/../../ 2 | 3 | -include $(ROCKSROOT)/etc/Rules.mk 4 | include Rules.mk 5 | 6 | build: 7 | echo "no build required" 8 | 9 | install:: 10 | mkdir -p $(ROOT)/RedHat/base 11 | install -ma+r comps.xml $(ROOT)/$(PKGROOT) 12 | 13 | -------------------------------------------------------------------------------- /src/usersguide/appendix/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | -------------------------------------------------------------------------------- /graphs/default/rocks7.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | -------------------------------------------------------------------------------- /nodes/client-rootpw.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 |
6 | --iscrypted /opt/rocks/sbin/gen_root_pw 7 |
8 |
9 | /opt/rocks/sbin/gen_root_pw 10 |
11 |
12 | -------------------------------------------------------------------------------- /src/qrencode/version.mk: -------------------------------------------------------------------------------- 1 | # Source location: http://fukuchi.org/works/qrencode/qrencode-3.4.0.tar.bz2 2 | PKGROOT = /usr 3 | NAME = qrencode 4 | VERSION = 3.4.0 5 | RELEASE = 2 6 | TARBALL_POSTFIX = tar.bz2 7 | RPM.FILES = "/usr/bin/*\\n/usr/include/*\\n/usr/lib/lib*\\n/usr/lib/pkgconfig/*\\n/usr/share/man/man1/*" 8 | -------------------------------------------------------------------------------- /src/rocks-cluster-homepage/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-cluster-homepage 2 | RELEASE = 1 3 | RPM.ARCH = noarch 4 | PKGROOT = /opt/rocks 5 | RPM.FILES = \ 6 | /opt/rocks/bin/*\\n\ 7 | /var/www/html/images/* 8 | 9 | RPM.DESCRIPTION = \ 10 | A script and images for generating a very simple home page for a rocks \ 11 | cluster. 12 | 13 | -------------------------------------------------------------------------------- /src/tentakel/version.mk: -------------------------------------------------------------------------------- 1 | NAME = tentakel 2 | VERSION = 2.2 3 | RELEASE = 0 4 | COPYRIGHT.EXTRA = Sebastian Stark 5 | TPG-VERSION = 3.1.4 6 | RPM.FILES = \ 7 | /opt/rocks/bin/* \\n\ 8 | /opt/rocks/lib/python2*/site-packages/* \\n\ 9 | /opt/rocks/man/man1/* \\n\ 10 | /opt/rocks/share/doc/tentakel \\n\ 11 | /opt/rocks/var/plugins/insertethers/* 12 | -------------------------------------------------------------------------------- /src/usersguide/advanced/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Advanced Tasks 3 | 4 | &firewall; 5 | &otp; 6 | &pxeflash; 7 | &login; 8 | 9 | &bonding; 10 | ¢ral; 11 | ✗ 12 | &bootflags; 13 | &bootloader; 14 | &updates; 15 | &reconfigure; 16 | 17 | 18 | 19 | -------------------------------------------------------------------------------- /src/google-otp/google-otp: -------------------------------------------------------------------------------- 1 | #%PAM-1.0 2 | # Test ask for OTP if the user is in the google-otp group, if not 3 | # skip the pam_google_authenticator step 4 | auth [success=1 default=ignore] pam_succeed_if.so quiet user notingroup google-otp 5 | auth required pam_google_authenticator.so nullok user=root secret=/export/google-authenticator/${USER} 6 | -------------------------------------------------------------------------------- /src/screens/header.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 |

Welcome to Rocks

9 | 10 | 13 | 14 | 15 | 16 | 17 | 18 | -------------------------------------------------------------------------------- /src/google-otp/version.mk: -------------------------------------------------------------------------------- 1 | NAME = google-authenticator 2 | VERSION = 7SEP2012 3 | RELEASE = 0 4 | RPM.FILES = \ 5 | /etc/pam.d/* \\n\ 6 | /export/google-authenticator \\n\ 7 | /lib64/security/pam_google_authenticator.so \\n\ 8 | /opt/rocks/lib/python2*/site-packages/rocks/commands/sync/users/* \\n\ 9 | /opt/rocks/var/plugins/411/* \\n\ 10 | /usr/bin/* 11 | -------------------------------------------------------------------------------- /src/ekv/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-ekv 2 | RELEASE = 0 3 | FOUNDATION = /opt/rocks 4 | 5 | ifeq ($(strip $(VERSION.MAJOR)), 5) 6 | LIBWRAP = /usr/$(LIB)/libwrap.so* 7 | else 8 | LIBWRAP = /$(LIB)/libwrap.so.* 9 | endif 10 | 11 | RPM.FILES = \ 12 | /rocks/bin/* \\n\ 13 | /usr/bin/* \\n\ 14 | /usr/lib64/* \\n\ 15 | /usr/sbin/* \\n\ 16 | /usr/share/terminfo/v/vt100 17 | -------------------------------------------------------------------------------- /src/sec-channel/server/sec-channel.service: -------------------------------------------------------------------------------- 1 | [Unit] 2 | Description=Rocks sec-channel service 3 | Requires=rpcbind.service 4 | After=rpcbind.service 5 | 6 | [Service] 7 | Type=forking 8 | PIDFile=/var/run/sec-channel.pid 9 | ExecStart=/opt/rocks/sbin/sec_channel_server -p /var/run/sec-channel.pid 10 | Restart=on-abort 11 | 12 | [Install] 13 | WantedBy=multi-user.target 14 | -------------------------------------------------------------------------------- /src/usersguide/community/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Community Resources 3 | 4 | &source-code; 5 | &releases; 6 | &mailing-list; 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | -------------------------------------------------------------------------------- /src/usersguide/community/releases.sgml: -------------------------------------------------------------------------------- 1 |
2 | All Past Rocks Releases 3 | 4 | 5 | All past Rocks releases can be found at 6 | 7 | ftp.rocksclusters.org:/pub/rocks. 8 | 9 | 10 | 11 |
12 | 13 | 14 | -------------------------------------------------------------------------------- /tests/t6000/plugin_device.py: -------------------------------------------------------------------------------- 1 | #!/opt/rocks/bin/python 2 | # 3 | # 4 | 5 | import rocks.commands 6 | 7 | class Plugin(rocks.commands.Plugin): 8 | 9 | def provides(self): 10 | return '%PLUGINNAME%' 11 | 12 | def run(self, node, xml): 13 | """report host vm config xml filter""" 14 | xml.append('') 15 | return xml 16 | 17 | 18 | -------------------------------------------------------------------------------- /src/rocks-anaconda-updates/.rpmmacros: -------------------------------------------------------------------------------- 1 | %_topdir /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/../../ 2 | %_buildrootdir /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot 3 | %buildroot /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot 4 | %_var /var 5 | %debug_package %{nil} 6 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/mysql.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | How do I examine the Rocks MySQL database and/or Ganglia information with 5 | a browser? 6 | 7 | 8 | 9 | 10 | 11 | Point your browser to http://localhost for access to the ganglia info 12 | and MySQL database. 13 | 14 | 15 | 16 | 17 | 18 | -------------------------------------------------------------------------------- /src/usersguide/community/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | -------------------------------------------------------------------------------- /src/channel/channeld.service: -------------------------------------------------------------------------------- 1 | [Unit] 2 | Description=Rocks channeld alert service 3 | Requires=rpcbind.service 4 | After=rpcbind.service 5 | Wants=rocks-rc-post.service 6 | After=rocks-rc-post.service 7 | 8 | [Service] 9 | Type=forking 10 | PIDFile=/var/run/channeld.pid 11 | ExecStart=/opt/rocks/sbin/channeld -p /var/run/channeld.pid 12 | Restart=on-abort 13 | 14 | [Install] 15 | WantedBy=multi-user.target 16 | -------------------------------------------------------------------------------- /src/firerox/version.mk: -------------------------------------------------------------------------------- 1 | NAME = firerox 2 | RELEASE = 0 3 | 4 | ifeq ($(VERSION.MAJOR), 5) 5 | FIREFOX_VERSION = 3.5.3 6 | MOZILLA_VERSION = 1.9.1 7 | endif 8 | 9 | ifeq ($(VERSION.MAJOR), 6) 10 | FIREFOX_VERSION = 3.5.19 11 | MOZILLA_VERSION = 1.9.1 12 | endif 13 | 14 | ifeq ($(VERSION.MAJOR), 7) 15 | FIREFOX_VERSION = 31.0 16 | MOZILLA_VERSION = release 17 | endif 18 | 19 | RPM.FILES = /opt/rocks/firerox 20 | -------------------------------------------------------------------------------- /src/411-alert/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-411-alert 2 | RELEASE = 1 3 | RPM.ARCH = noarch 4 | RPM.FILES = \ 5 | /opt/rocks/bin/*\\n\ 6 | /etc/serf/* 7 | 8 | RPM.REQUIRES = rocks-411 9 | 10 | RPM.DESCRIPTION = \ 11 | This is for handling 411 alerts via serf. This RPM should only be \\n\ 12 | installed on 411 client nodes, not a frontend. This utilizes Hashicorps \\n\ 13 | serf cluster membership for efficiency of operation.\\n 14 | 15 | -------------------------------------------------------------------------------- /src/firerox/user.js: -------------------------------------------------------------------------------- 1 | # Mozilla User Preferences 2 | 3 | /* Do not edit this file. 4 | * 5 | * If you make changes to this file while the browser is running, 6 | * the changes will be overwritten when the browser exits. 7 | * 8 | * To make a manual change to preferences, you can visit the URL about:config 9 | * For more information, see http://www.mozilla.org/unix/customizing.html#prefs 10 | */ 11 | pref("browser.tabs.autoHide", true); 12 | -------------------------------------------------------------------------------- /src/usersguide/version.mk: -------------------------------------------------------------------------------- 1 | NAME = roll-$(ROLL)-usersguide 2 | RELEASE = 2 3 | RPM.ARCH = noarch 4 | 5 | SUMMARY_COMPATIBLE = $(VERSION) 6 | SUMMARY_MAINTAINER = Rocks Group 7 | SUMMARY_ARCHITECTURE = x86_64 8 | 9 | ifeq ($(VERSION.MAJOR),6) 10 | ROLL_REQUIRES = kernel base os 11 | else 12 | ROLL_REQUIRES = kernel core base CentOS 13 | endif 14 | ROLL_CONFLICTS = 15 | RPM.FILES = /var/www/html/roll-documentation/base/* 16 | -------------------------------------------------------------------------------- /src/pdsh/version.mk: -------------------------------------------------------------------------------- 1 | PKGROOT = /opt 2 | NAME = pdsh 3 | VERSION = 2.26 4 | RELEASE = 1 5 | TARBALL_POSTFIX = tar.bz2 6 | RPM.FILES = \ 7 | /etc/ld.so.conf.d/* \\n \ 8 | /etc/profile.d/* \\n \ 9 | /opt/pdsh 10 | RPM.DESCRIPTION = \ 11 | pdsh is a multithreaded remote shell client which executes commands on \\n\ 12 | multiple remote hosts in parallel. Pdsh can use several different \\n\ 13 | remote shell services, including standard "rsh", Kerberos IV, and ssh. 14 | 15 | -------------------------------------------------------------------------------- /include-version/6/nodes/upstart.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Defines Upstart-compatible Services 7 | 8 | 9 | 10 | 11 | $Log: upstart.xml,v $ 12 | Revision 1.1 2012/02/09 21:00:51 phil 13 | Put rocks-pre and rocks-post scripts into CentOS/RHEL upstart framework 14 | 15 | 16 | 17 | rocks-upstart 18 | 19 | 20 | -------------------------------------------------------------------------------- /include-version/7/nodes/upstart.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Defines Upstart-compatible Services 7 | 8 | 9 | 10 | 11 | $Log: upstart.xml,v $ 12 | Revision 1.1 2012/02/09 21:00:51 phil 13 | Put rocks-pre and rocks-post scripts into CentOS/RHEL upstart framework 14 | 15 | 16 | 17 | rocks-upstart 18 | 19 | 20 | -------------------------------------------------------------------------------- /nodes/ipmi-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Server Nodes 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Id: ipmi-server.xml,v 1.4 2012/05/06 05:48:10 phil Exp $ 19 | 20 | 21 | 22 | 23 | -------------------------------------------------------------------------------- /nodes/md-rebuild.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Moderate MD Raid Rebuild Speed 6 | 7 | 8 | 9 | Copyright (c) 2000 - 2017 The Regents of the University of California. 10 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 11 | 12 | 13 |
14 | echo 1000 > /proc/sys/dev/raid/speed_limit_max 
15 | 
16 | 17 |
18 | -------------------------------------------------------------------------------- /nodes/server-rootpw.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | $Id: server-rootpw.xml,v 1.2 2011/06/21 23:26:04 anoop Exp $ 6 | Root Password for the frontend 7 | 8 | 9 | $Log: server-rootpw.xml,v $ 10 | Revision 1.2 2011/06/21 23:26:04 anoop 11 | Add description 12 | 13 | 14 |
15 | --iscrypted &Kickstart_PrivateRootPassword; 16 |
17 |
18 | -------------------------------------------------------------------------------- /src/usersguide/clusterinstall/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | -------------------------------------------------------------------------------- /nodes/ipmi-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Client Nodes 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Id: ipmi-client.xml,v 1.5 2012/05/06 05:48:10 phil Exp $ 19 | 20 | 21 | 22 | 23 | 24 | -------------------------------------------------------------------------------- /src/extras/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-extras 2 | RELEASE = 0 3 | 4 | MAKE.iscontrib = 1 5 | 6 | #DISTPATH = $(shell rocks-dist --path=dist.getRPMSPath paths) 7 | DISTPATH = rocks-dist/$(ARCH)/RedHat/RPMS 8 | OS_VERSION = $(shell mysql --batch --execute='select version from rolls where name="os" and enabled="yes"' -u apache cluster | grep -v version) 9 | 10 | 11 | ifeq ($(strip $(VERSION.MAJOR)), 5) 12 | CACHENAMED = caching-nameserver 13 | else 14 | CACHENAMED = bind-chroot 15 | endif 16 | -------------------------------------------------------------------------------- /OSROLL/roll-distro-update.xml.in: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | -------------------------------------------------------------------------------- /nodes/nas-partition.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Define a partitioning layout for NAS nodes. 7 | 8 | 9 | 12 |
13 | echo "rocks manual" > /tmp/user_partition_info
14 | 
15 | 16 |
17 | 18 | -------------------------------------------------------------------------------- /nodes/profile-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | profile.d entries for servers in Rocks 7 6 | 7 | 8 | 9 | Copyright (c) 2000 - 2017 The Regents of the University of California. 10 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 11 | 12 | 13 | 14 | rocks-config-server-profile 15 | 16 | 17 | -------------------------------------------------------------------------------- /src/developersguiderst/index.rst: -------------------------------------------------------------------------------- 1 | .. Rocks developers documentation master file, created by 2 | sphinx-quickstart on Mon Feb 9 17:07:59 2015. 3 | You can adapt this file completely to your liking, but it should at least 4 | contain the root `toctree` directive. 5 | 6 | Rocks internal coding userguide 7 | =============================== 8 | 9 | Rocks developers guide. 10 | 11 | 12 | Contents: 13 | 14 | .. toctree:: 15 | :maxdepth: 2 16 | 17 | hacking 18 | refs/modules 19 | 20 | 21 | -------------------------------------------------------------------------------- /src/developersguiderst/refs/rocks.db.rst: -------------------------------------------------------------------------------- 1 | db Package 2 | ========== 3 | 4 | :mod:`database` Module 5 | ---------------------- 6 | 7 | .. automodule:: rocks.db.database 8 | :members: 9 | :undoc-members: 10 | :show-inheritance: 11 | 12 | :mod:`helper` Module 13 | -------------------- 14 | 15 | .. automodule:: rocks.db.helper 16 | :members: 17 | :undoc-members: 18 | :show-inheritance: 19 | 20 | Subpackages 21 | ----------- 22 | 23 | .. toctree:: 24 | 25 | rocks.db.mappings 26 | 27 | -------------------------------------------------------------------------------- /nodes/null.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | An empty XML file. 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2008 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v5.0 www.rocksclusters.org 12 | 13 | 14 | 15 | $Log: null.xml,v $ 16 | Revision 1.1 2008/07/29 16:48:05 bruno 17 | save the frontend's partitions in the database. 18 | 19 | 20 | 21 | 22 | 23 | -------------------------------------------------------------------------------- /src/serf/rocks-serf.service: -------------------------------------------------------------------------------- 1 | [Unit] 2 | Description=Serf Dynamic Group Membership (hashicorp) 3 | Wants=rocks-rc-post.service 4 | After=rocks-rc-post.service 5 | 6 | [Service] 7 | Type=simple 8 | StandardOutput=syslog 9 | StandardError=syslog 10 | SyslogIdentifier=serf 11 | Environment=GOMAXPROCS=2 12 | ExecStart=/usr/bin/nohup /opt/rocks/bin/serf agent -config-dir /etc/serf/ < /dev/null 13 | ExecStop=/opt/rocks/bin/serf leave -rpc-auth=$(cat /etc/serf/rpcauth) 14 | 15 | [Install] 16 | WantedBy=multi-user.target 17 | -------------------------------------------------------------------------------- /tests/t2007/test_includes.py: -------------------------------------------------------------------------------- 1 | #!/opt/rocks/bin/python 2 | # 3 | # not used at the moment 4 | 5 | import os 6 | import string 7 | import sys 8 | import syslog 9 | 10 | 11 | currentdir = os.path.dirname(os.path.realpath(__file__)) 12 | newdir = os.path.realpath(currentdir + '/../../include/applets/') 13 | 14 | sys.path.append(newdir) 15 | 16 | import socket 17 | hostname = socket.gethostname() 18 | 19 | import ConfigPartitions 20 | app = ConfigPartitions.App() 21 | app.setHostname(hostname) 22 | app.run() 23 | 24 | -------------------------------------------------------------------------------- /src/usersguide/advanced/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | -------------------------------------------------------------------------------- /src/rocks-anaconda-updates/rocks-anaconda-updates.spec.mk: -------------------------------------------------------------------------------- 1 | # This file is called from the generated spec file. 2 | # It can also be used to debug rpm building. 3 | # make -f rocks-anaconda-updates.spec.mk build|install 4 | 5 | ifndef __RULES_MK 6 | build: 7 | make ROOT=/export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot build 8 | 9 | install: 10 | make ROOT=/export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot install 11 | endif 12 | -------------------------------------------------------------------------------- /src/usersguide/customizing/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Customizing your Rocks Installation 3 | 4 | &customize-adding-packages; 5 | &customize-postconfig; 6 | &customize-adding-applications; 7 | &customize-extra-nic; 8 | &customize-partition; 9 | &customize-kernel; 10 | &customize-rsh; 11 | &customize-ssh; 12 | &customization-adding-appliance; 13 | &customization-driver; 14 | &customization-dns; 15 | &customization-passwd; 16 | &customization-insert-ethers; 17 | 18 | 19 | 20 | -------------------------------------------------------------------------------- /src/firerox/mozconfig: -------------------------------------------------------------------------------- 1 | ac_add_options --enable-application=browser 2 | ac_add_options --prefix=/opt/rocks/firerox 3 | ac_add_options --disable-tests --disable-dbus --disable-crashreporter --disable-oji --with-system-nspr 4 | ac_add_options --disable-postscript --disable-gnomevfs --disable-gnomeui 5 | ac_add_options --disable-printing --disable-plugins 6 | ac_add_options --disable-necko-wifi 7 | ac_add_options --disable-xul 8 | ac_add_options --disable-xpcom-obsolete 9 | ac_add_options --disable-webm 10 | mk_add_options MOZ_CO_PROJECT=browser 11 | -------------------------------------------------------------------------------- /src/screens/make-timezones.py: -------------------------------------------------------------------------------- 1 | #!/opt/rocks/bin/python 2 | 3 | import string 4 | 5 | places = [] 6 | file = open('/usr/share/zoneinfo/zone.tab', 'r') 7 | 8 | localTimezone="" 9 | for line in file.readlines(): 10 | l = string.split(line) 11 | 12 | if len(l) < 2 or l[0][0] == '#': 13 | continue 14 | 15 | if l[2].find("Los_Angeles") >= 0 : 16 | localTimezone=l[2] 17 | else: 18 | places.append(l[2]) 19 | 20 | file.close() 21 | 22 | places.sort() 23 | places.insert(0, localTimezone) 24 | for p in places: 25 | print '' 26 | -------------------------------------------------------------------------------- /src/channel/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-channel 2 | RELEASE = 2 3 | SYSTEMDFOLDER = /etc/systemd/system 4 | SYSTEMDINITSCRIPT = channeld.service 5 | SYSVFOLDER = /etc/rc.d/init.d 6 | SYSVINITSCRIPT = channeld 7 | 8 | ifeq ($(VERSION.MAJOR), 6) 9 | SCRIPTSRC=channel.init.linux 10 | SCRIPT=$(SYSVINITSCRIPT) 11 | SCRIPTDEST=$(SYSVFOLDER) 12 | endif 13 | ifeq ($(VERSION.MAJOR), 7) 14 | SCRIPTSRC=$(SYSTEMDINITSCRIPT) 15 | SCRIPT=$(SYSTEMDINITSCRIPT) 16 | SCRIPTMODE=644 17 | SCRIPTDEST=$(SYSTEMDFOLDER) 18 | endif 19 | 20 | RPM.FILES = "$(SCRIPTDEST)/*\\n/opt/rocks/sbin/*" 21 | -------------------------------------------------------------------------------- /src/usersguide/clusterinstall/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Installing a Rocks Cluster 3 | 4 |
5 | Getting Started 6 | 7 | 8 | This chapter describes the steps to build your cluster and install its software. 9 | 10 | 11 | &supported; 12 | &minimum; 13 | &physical; 14 | 15 |
16 | 17 | &frontend7; 18 | &frontend; 19 | &compute; 20 | &frontend-upgrade; 21 | &frontend-wan; 22 | &enable-www; 23 | 24 |
25 | 26 | -------------------------------------------------------------------------------- /src/sec-channel/server/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-sec-channel-server 2 | RELEASE = 3 3 | 4 | SYSTEMDFOLDER = /etc/systemd/system 5 | SYSTEMDINITSCRIPT = sec-channel.service 6 | SYSVFOLDER = /etc/rc.d/init.d 7 | SYSVINITSCRIPT = sec-channel 8 | 9 | ifeq ($(VERSION.MAJOR), 6) 10 | SCRIPTSRC=sec-channel.init 11 | SCRIPT=$(SYSVINITSCRIPT) 12 | SCRIPTDEST=$(SYSVFOLDER) 13 | endif 14 | ifeq ($(VERSION.MAJOR), 7) 15 | SCRIPTSRC=$(SYSTEMDINITSCRIPT) 16 | SCRIPT=$(SYSTEMDINITSCRIPT) 17 | SCRIPTDEST=$(SYSTEMDFOLDER) 18 | endif 19 | 20 | RPM.FILES = "$(SCRIPTDEST)/*\\n/opt/rocks/sbin/*" 21 | -------------------------------------------------------------------------------- /nodes/homepage.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Create a simple cluster home page 6 | 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2017 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 12 | 13 | 14 | 15 | rocks-cluster-homepage 16 | 17 | 18 | 19 | 20 | 21 | 22 | 23 | 24 | -------------------------------------------------------------------------------- /nodes/x11-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | X11 on Server Nodes for CentOS 7 7 | 8 | 9 | 10 | 11 | Copyright (c) 2000 - 2017 The Regents of the University of California. 12 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 13 | 14 | 15 | 16 | ^graphical-server-environment 17 | 18 | 19 | /usr/bin/systemctl set-default graphical.target 20 | 21 | 22 | 23 | -------------------------------------------------------------------------------- /tests/t2009/test2.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | test2 6 | 7 | 8 | 15 | 16 | test21 17 | test22 18 | 19 | 20 | 21 | grep local 22 | 23 | 24 | cp /etc/hosts /tmp/test2pipesamefile 25 | 26 | grep local 27 | 28 | 29 | 30 | 31 | -------------------------------------------------------------------------------- /src/yum-extras/version.mk: -------------------------------------------------------------------------------- 1 | ifeq ($(VERSION.MAJOR),5) 2 | REPORPMS.NOARCH += elrepo-release-5-3.el5.elrepo.noarch.rpm 3 | REPORPMS.NOARCH += epel-release-5-4.noarch.rpm 4 | endif 5 | ifeq ($(VERSION.MAJOR),6) 6 | REPORPMS.NOARCH += elrepo-release-6-4.el6.elrepo.noarch.rpm 7 | REPORPMS.NOARCH += epel-release-6-5.noarch.rpm 8 | FAIL2BAN.PKGS = fail2ban python-inotify 9 | endif 10 | ifeq ($(VERSION.MAJOR),7) 11 | # REPORPMS.NOARCH += elrepo-release-6-4.el6.elrepo.noarch.rpm 12 | REPORPMS.NOARCH += epel-release-7-11.noarch.rpm 13 | FAIL2BAN.PKGS = fail2ban fail2ban-firewalld fail2ban-sendmail fail2ban-server python-inotify 14 | endif 15 | -------------------------------------------------------------------------------- /nodes/partition.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | $Log: partition.xml,v $ 7 | Revision 1.3 2008/04/22 17:47:25 bruno 8 | tweak partitioning 9 | 10 | Revision 1.2 2008/04/01 19:41:29 bruno 11 | make the default partitioning to only deal with the root disk 12 | 13 | Revision 1.1 2008/02/19 23:20:24 bruno 14 | katz made me do it. 15 | 16 | 17 | 18 | 19 | 20 |
21 | if [ ! -f /tmp/user_partition_info ]
22 | then
23 | 	echo "rocks force-default-root-disk-only" > /tmp/user_partition_info
24 | fi
25 | 
26 | 27 | 28 |
29 | 30 | -------------------------------------------------------------------------------- /graphs/default/serf.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | The serf Roll 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log$ 19 | 20 | 21 | 22 | 23 | serf-client 24 | serf-common 25 | 26 | 27 | 28 | serf-server 29 | serf-common 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /tests/t2009/test1.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | test2 6 | 7 | 8 | 13 | 14 | test11 15 | test12 16 | 17 | 18 | 19 | echo ciao > /tmp/test1file 20 | 21 | 22 | echo ciao > /tmp/test2file 23 | 24 | 25 | echo "this is test1 postsection" 26 | 27 | 28 | 29 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/no-kickstart.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Why can't I re-kickstart compute nodes? Shoot-node fails, and 6 | power cycling the machine doesn't force a re-install. 7 | 8 | 9 | 10 | 11 | 12 | 13 | Older BIOS versions required the boot image to reside in the 14 | first GB of the hard disk. If the boot image, in this case 15 | the kickstart kernel image, resides after the first GB of the 16 | disk the image will not be loaded by GRUB. Update your BIOS 17 | and this problem should get fixed. 18 | 19 | 20 | 21 | 22 | 23 | 24 | -------------------------------------------------------------------------------- /tests/t2009/test3.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | test3 7 | 8 | 9 | 10 | 17 | 18 | test31 19 | test32 20 | 21 | 22 | 23 | 24 | this is a simple file 25 | 26 | 27 | 28 | 29 | cat > /tmp/test3-release << 'SPECIAL' 30 | 31 | cat /etc/rocks-release 32 | 33 | SPECIAL 34 | 35 | 36 | 37 | -------------------------------------------------------------------------------- /src/screens/getrolls.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 11 | 12 | 13 | 16 | 17 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26 | 27 | 28 | 29 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/startx.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Why doesn't startx work on the frontend machine? 6 | 7 | 8 | 9 | 10 | 11 | 12 | Before you can run startx you need to configure XFree86 for 13 | your video card. This is done just like on standard Red Hat machines 14 | using the system-config-display 15 | program. If you do not know anything about 16 | your video card just select "4MB" of video RAM and 16 bit 17 | color 800x600. This video mode should work on any modern VGA 18 | card. 19 | 20 | 21 | 22 | 23 | 24 | 25 | -------------------------------------------------------------------------------- /OSROLL/version.mk: -------------------------------------------------------------------------------- 1 | ## Create 2 | DISTRO=CentOS 3 | TSTARCH=$(shell /bin/arch) 4 | ifeq ($(TSTARCH),i686) 5 | ARCH=i386 6 | else 7 | ARCH=$(TSTARCH) 8 | endif 9 | 10 | ifeq ($(VERSION.MAJOR), 5) 11 | VERSION=5.9 12 | PKGS=RPMS 13 | BASEPATH=centos/$(VERSION)/os/$(ARCH)/CentOS/ 14 | endif 15 | ifeq ($(VERSION.MAJOR), 6) 16 | VERSION=6.8 17 | PKGS=Packages 18 | BASEPATH=$(VERSION)/os/$(ARCH)/$(PKGS)/ 19 | endif 20 | ifeq ($(VERSION.MAJOR), 7) 21 | VERSION=7.5.1804 22 | PKGS=Packages 23 | BASEPATH=$(VERSION)/os/$(ARCH)/$(PKGS)/ 24 | endif 25 | 26 | MIRRORURL=http://linux.mirrors.es.net/centos/ 27 | 28 | UPDATESPATH=$(VERSION)/updates/$(ARCH)/$(PKGS)/ 29 | 30 | ROLLNAME=CentOS-$(VERSION)-Updated 31 | -------------------------------------------------------------------------------- /src/screens/status.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 24 | 25 | 26 | 27 | 28 | 29 |

30 |

31 |

32 | 33 | 34 | 35 | 36 | 37 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/removenode.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | How do I remove a compute node from the cluster? 5 | 6 | 7 | 8 | 9 | 10 | On your frontend end, execute: 11 | 12 | 13 | 14 | # rocks remove host "[your compute node name]" 15 | 16 | 17 | 18 | For example, if the compute node's name is 19 | compute-0-1, you'd execute: 20 | 21 | 22 | 23 | # rocks remove host compute-0-1 24 | # rocks sync config 25 | 26 | 27 | 28 | The compute node has been removed from the cluster. 29 | 30 | 31 | 32 | 33 | 34 | -------------------------------------------------------------------------------- /graphs/default/root.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | The Root of Rocks 8 | 9 | 10 | 11 | 12 | $Log: root.xml,v $ 13 | Revision 1.1 2005/03/04 01:47:16 mjk 14 | *** empty log message *** 15 | 16 | Revision 1.2 2005/02/16 00:40:03 mjk 17 | graph cleanup 18 | 19 | Revision 1.1 2003/07/07 20:47:51 bruno 20 | initial release 21 | 22 | 23 | 24 | 25 | 26 | 27 | server 28 | 29 | 30 | 31 | 32 | 33 | 34 | server 35 | 36 | 37 | 38 | 39 | 40 | -------------------------------------------------------------------------------- /tests/t2009/test.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | test graph 6 | 7 | 8 | 9 | Copyright (c) 2000 - 2017 The Regents of the University of California. 10 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 11 | 12 | 13 | 14 | 17 | 18 | 19 | 20 | test2 21 | 22 | 23 | 24 | test3 25 | 26 | 27 | 28 | 29 | test2 30 | 31 | 32 | 33 | test3 34 | 35 | 36 | 37 | 38 | -------------------------------------------------------------------------------- /nodes/bootstrap-dummy.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Dummy terminal node for bootstrap-packages subgraph 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2017 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 12 | 13 | 14 | 15 | 16 | $Log: bootstrap-dummy.xml,v $ 17 | Revision 1.2 2012/05/06 05:48:10 phil 18 | Copyright Storm for Mamba 19 | 20 | Revision 1.1 2011/11/03 16:32:02 phil 21 | Getting closer to functional bootstrap on a non-Rocks build host 22 | 23 | 24 | 25 | 26 | 27 | -------------------------------------------------------------------------------- /src/411/version.mk: -------------------------------------------------------------------------------- 1 | NAME = rocks-411 2 | RELEASE = 4 3 | RPM.ARCH = noarch 4 | RPM.FILES = \ 5 | /opt/rocks/bin/*\\n\ 6 | /opt/rocks/etc/*\\n\ 7 | /opt/rocks/lib/python2*/site-packages/rocks/* 8 | 9 | RPM.DESCRIPTION = \ 10 | The 411 Secure Information Service. This facility is intended to replace \\n\ 11 | NIS which has security and scaling issues. 411 uses public-key \\n\ 12 | cryptography to securely distribute sensitive files such as group and \\n\ 13 | shadow password files. 411 uses either HTTP or HTTPS as its transport \\n\ 14 | protocol, however 411 is designed to maintain full security even with \\n\ 15 | regular HTTP. The 411 service utilizes Hashicorps serf \\n\ 16 | cluster membership for efficiency of operation.\\n 17 | 18 | -------------------------------------------------------------------------------- /nodes/compute.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Compute Appliance Node. You should create your own extend-compute.xml 8 | file to add functionality to this appliance type. 9 | 10 | 11 | 12 | 13 | $Log: compute.xml,v $ 14 | Revision 1.1 2007/12/10 21:28:33 bruno 15 | the base roll now contains several elements from the HPC roll, thus 16 | making the HPC roll optional. 17 | 18 | this also includes changes to help build and configure VMs for V. 19 | 20 | Revision 1.9 2005/03/04 01:40:08 mjk 21 | massive graph cleanup 22 | 23 | Revision 1.8 2005/02/16 00:40:04 mjk 24 | graph cleanup 25 | 26 | 27 | 28 | 29 | 30 | -------------------------------------------------------------------------------- /nodes/custom-partition.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 |
 5 | # This node should be REPLACED by the user. 
 6 | # REPLACED means:
 7 | # 1. copy this file to:
 8 | # /export/rocks/install/site-profiles/[version]/nodes/replace-custom-partition.xml
 9 | # 2. edit the replace files to create your custom partitioning
10 | # 3. cd /export/rocks/install; rocks create distro
11 | # 4. install the node 
12 | 
13 | # Within the --pre-- section of the node
14 | # the file /tmp/user_partition_info should be created with 
15 | # kickstart partitioning and formatting directions.
16 | # on reinstalls, only /, /var, /boot are formatted
17 | echo "No Custom Partitioning"  > /dev/null
18 | 
19 |
20 | 21 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/smalldisk.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | When I try to install a compute node, the error message on the compute node 6 | says, "Can't mount /tmp. Please press OK to restart". 7 | What should I do? 8 | 9 | 10 | 11 | 12 | 13 | Most likely, this situation arises due to the size of the disk drive on the compute 14 | node. 15 | The installation procedure for Rocks formats the disk on the compute node if Rocks 16 | has never been installed on the compute node before. 17 | 18 | 19 | 20 | The fix requires changing the way Rocks partitions disk drives. 21 | See for details. 22 | 23 | 24 | 25 | 26 | 27 | -------------------------------------------------------------------------------- /nodes/grub2-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Boot loader configuration for cluster appliances. Grub2 edition 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | 19 | 20 | BOOTFLAGS=`/opt/rocks/bin/rocks report host bootflags &hostname;` 21 | echo "addargs=\"$BOOTFLAGS\"" 22 | 23 | /opt/rocks/sbin/modify-grub2 --file=/etc/sysconfig/grub --keyword=GRUB_CMDLINE_LINUX --mode=append $addargs 24 | /usr/sbin/grub2-mkconfig -o /boot/grub2/grub.cfg 25 | 26 | 27 | 28 | 29 | -------------------------------------------------------------------------------- /nodes/devel-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Add Development Appliance Definition to the Frontend 7 | 8 | 9 | 10 | 11 | $Log: devel-server.xml,v $ 12 | Revision 1.2 2011/04/06 15:38:18 phil 13 | Default the yum repository to local, but override this with Attribute 14 | RepositoryLocal 15 | 16 | Revision 1.1 2011/03/27 17:27:30 phil 17 | Add the development appliance (Imported from 5.4 EC2 Roll) 18 | 19 | 20 | 21 | 22 | 23 | /opt/rocks/bin/rocks add appliance devel-server membership="Development Appliance" node=devel-appliance graph=default compute=no public=yes 24 | /opt/rocks/bin/rocks add appliance attr devel-server RepositoryLocal true 25 | 26 | 27 | 28 | 29 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/mirroring.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | When I mirror, I see output that says 6 | No such file `'.. 7 | Is that OK? 8 | 9 | 10 | 11 | 12 | 13 | 14 | Yes, it is ok to see output like: 15 | 16 | 17 | 18 | FINISHED --00:54:08-- 19 | Downloaded: 9 bytes in 1 files 20 | 00:54:08 URL: ftp://www.rocksclusters.org/%2Fpub/rocks/rocks-dist/updates/7.2/en/os/i386/ [9] -> "www.rocksclusters.org/pub/rocks/rocks-dist/updates/7.2/en/os/i386/.listing" [1] 21 | No such file `'. 22 | 23 | 24 | FINISHED --00:54:08-- 25 | Downloaded: 9 bytes in 1 files 26 | No such directory `/pub/rocks/rocks-dist/updates/7.2/en/os/noarch'. 27 | 28 | 29 | 30 | 31 | 32 | -------------------------------------------------------------------------------- /src/serf/version.mk: -------------------------------------------------------------------------------- 1 | PKGROOT = /opt/rocks/bin 2 | NAME = serf 3 | RELEASE = 0 4 | VERSION = 0.8.1 5 | ZIPFILE = $(NAME)_$(VERSION)_linux_amd64.zip 6 | 7 | INITD_DIR = /etc/init.d 8 | INITD_SCRIPTS = rocks-serf 9 | SYSTEMD_DIR = /etc/systemd/system 10 | SYSTEMD_SCRIPTS = rocks-serf.service 11 | 12 | ifeq ($(VERSION.MAJOR),6) 13 | SCRIPTMODE = 755 14 | SCRIPTDIR = $(INITD_DIR) 15 | SCRIPTS = $(INITD_SCRIPTS) 16 | endif 17 | ifeq ($(VERSION.MAJOR),7) 18 | SCRIPTMODE = 644 19 | SCRIPTDIR = $(SYSTEMD_DIR) 20 | SCRIPTS = $(SYSTEMD_SCRIPTS) 21 | RSYSLOGDIR = /etc/rsyslog.d 22 | RSYSLOG_SCRIPT = serf.conf 23 | endif 24 | 25 | RPM.FILES = \ 26 | $(PKGROOT)/*\n\ 27 | /var/opt/rocks/$(NAME)\n\ 28 | /etc/$(NAME)\n\ 29 | /etc/logrotate.d/*\n\ 30 | $(SCRIPTDIR)/*\n\ 31 | 32 | ifeq ($(VERSION.MAJOR),7) 33 | RPM.FILES += $(RSYSLOGDIR)/* 34 | endif 35 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | -------------------------------------------------------------------------------- /nodes/yum-extras.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Extra Yum Repositories. Yum.xml disables by default 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: yum-extras.xml,v $ 19 | Revision 1.2 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.1 2012/03/26 17:52:38 phil 23 | Add extra repositories. Disabled by yum.xml 24 | 25 | 26 | 27 | 28 | elrepo-release 29 | epel-release 30 | rpmforge-release 31 | 32 | 33 | 34 | -------------------------------------------------------------------------------- /src/usersguide/clusterinstall/supported.sgml: -------------------------------------------------------------------------------- 1 |
2 | Supported Hardware 3 | 4 | 5 | Processors 6 | 7 | 8 | 9 | x86 (ia32, AMD Athlon, etc.) 10 | 11 | 12 | 13 | x86_64 (AMD Opteron and EM64T) 14 | 15 | 16 | 17 | 18 | Networks 19 | 20 | 21 | 22 | Ethernet 23 | 24 | 25 | 26 | 27 | 28 | Specialized networks and components (e.g., Myrinet, Infiniband, nVidia GPU) 29 | are also supported. 30 | Hardware requirements and software (Rocks Rolls) can be found on the 31 | respective vendor web sites. 32 | 33 | 34 | 35 |
36 | 37 | -------------------------------------------------------------------------------- /src/411-alert/411event.py: -------------------------------------------------------------------------------- 1 | #!/opt/rocks/bin/python 2 | # This handles a 411Alert event sent through serf 3 | # if no specific url is given, get everything 4 | # 5 | import os 6 | import sys 7 | import subprocess 8 | from rocks.four11handler import Handle411 9 | 10 | # check if this is a 411Alert event 11 | try: 12 | ename = os.environ["SERF_USER_EVENT"] 13 | if ename != "411Alert": 14 | # not for us 15 | os.exit(0) 16 | # fork off a child process because the 411 event handling might take 17 | # a while 18 | pid = os.fork() 19 | if pid != 0: 20 | sys.exit(0) 21 | 22 | # Child 23 | url = sys.stdin.read() 24 | if len(url) > 0: 25 | H = Handle411() 26 | H.apply(url.strip(),None) 27 | else: 28 | subprocess.call(["/opt/rocks/bin/411get","--all"]) 29 | os._exit(0) 30 | except: 31 | # This serf event is not for us 32 | pass 33 | 34 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/apache-compute.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Why is apache running on my compute nodes? 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | The default configuration for compute nodes is to start the 14 | Apache service. This is enabled to allow us to serve (over 15 | HTTP) the Linux /proc filesystem to a future monitoring tool. 16 | UCB's Ganglia will remain the prefered monitoring tool, but 17 | for highly detailed node information only the complete /proc 18 | filesystem will suffice. To disable this feature remove the 19 | following line from your distribution's configuration graph. 20 | 21 | 22 | 23 | 24 | <edge from="slave-node" to="apache"/> 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /src/developersguide/overview.sgml: -------------------------------------------------------------------------------- 1 | 2 | Overview 3 | 4 | 5 | The Roll Developer's Guide is for people who wish to build their own rolls. 6 | In this document, we'll describe the internals of rolls, how to create a roll 7 | and how to test a roll. 8 | 9 | 10 | 11 | Throughout the discussion, it's easy to lose sight of the basic simplicity of a roll. A roll consists of 12 | 13 | 14 | 15 | Software packages (RPMS) 16 | 17 | 18 | 19 | How to configure packages after installation (XML Node Files) 20 | 21 | 22 | 23 | On which types of appliances to install/configure packages (XML Graph) 24 | 25 | 26 | 27 | 28 | 29 | -------------------------------------------------------------------------------- /src/screens/help.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 14 | 15 | 31 | 32 | 33 | 34 | 35 | 36 |

37 |

Help

38 |

39 | 40 |

41 |

42 |
43 |
44 |

45 | 46 | 47 | 48 | 49 | 50 | -------------------------------------------------------------------------------- /nodes/client-firewall.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Client node firewall rules 8 | 9 | 10 | 11 | $Log: client-firewall.xml,v $ 12 | Revision 1.3 2010/05/17 20:58:22 bruno 13 | generate firewall rules from the database 14 | 15 | Revision 1.2 2009/03/30 19:29:50 anoop 16 | Merging xml files from Solaris Roll to the Linux roll. 17 | More files to follow 18 | 19 | Revision 1.1 2007/12/10 21:28:33 bruno 20 | the base roll now contains several elements from the HPC roll, thus 21 | making the HPC roll optional. 22 | 23 | this also includes changes to help build and configure VMs for V. 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | /opt/rocks/bin/rocks report host firewall &hostname; 32 | 33 | 34 | 35 | 36 | 37 | 38 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/plugins/ssh_host_rsa_key.py: -------------------------------------------------------------------------------- 1 | # $Id: ssh_host_rsa_key.py,v 1.1 2012/07/31 23:20:10 phil Exp $ 2 | # set the host rsa key 3 | 4 | import rocks.commands 5 | 6 | import os, sys 7 | import tempfile 8 | 9 | class plugin(rocks.commands.sec_attr_plugin): 10 | def get_sec_attr(self): 11 | return 'ssh_host_rsa_key' 12 | 13 | def filter(self, value): 14 | import grp 15 | keyfile = '/etc/ssh/ssh_host_rsa_key' 16 | pubkeyfile = keyfile + '.pub' 17 | # write the ssh_host_key_rsa file 18 | if os.path.exists(keyfile): 19 | os.chmod(keyfile, 0600) 20 | f = open(keyfile, 'w') 21 | f.write(value) 22 | f.close() 23 | os.chmod(keyfile, 0400) 24 | 25 | # regenerate the public key 26 | if os.path.exists(pubkeyfile): 27 | os.unlink(pubkeyfile) 28 | os.system('ssh-keygen -f %s -y > %s' % (keyfile, pubkeyfile)) 29 | os.chmod(pubkeyfile, 0644) 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/plugins/ssh_host_dsa_key.py: -------------------------------------------------------------------------------- 1 | # $Id: ssh_host_dsa_key.py,v 1.1 2012/07/31 23:20:10 phil Exp $ 2 | # Sync the host dsa key 3 | 4 | import rocks.commands 5 | 6 | import os, sys 7 | import tempfile 8 | 9 | class plugin(rocks.commands.sec_attr_plugin): 10 | def get_sec_attr(self): 11 | return 'ssh_host_dsa_key' 12 | 13 | def filter(self, value): 14 | import grp 15 | keyfile = '/etc/ssh/ssh_host_dsa_key' 16 | pubkeyfile = keyfile + '.pub' 17 | # write the ssh_host_key_rsa file 18 | if os.path.exists(keyfile): 19 | os.chmod(keyfile, 0600) 20 | f = open(keyfile, 'w') 21 | f.write(value) 22 | f.close() 23 | os.chmod(keyfile, 0400) 24 | 25 | # regenerate the public key 26 | if os.path.exists(pubkeyfile): 27 | os.unlink(pubkeyfile) 28 | os.system('ssh-keygen -f %s -y > %s' % (keyfile, pubkeyfile)) 29 | os.chmod(pubkeyfile, 0644) 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/plugins/ssh_host_ecdsa_key.py: -------------------------------------------------------------------------------- 1 | # $Id: ssh_host_dsa_key.py,v 1.1 2012/07/31 23:20:10 phil Exp $ 2 | # Sync the host dsa key 3 | 4 | import rocks.commands 5 | 6 | import os, sys 7 | import tempfile 8 | 9 | class plugin(rocks.commands.sec_attr_plugin): 10 | def get_sec_attr(self): 11 | return 'ssh_host_ecdsa_key' 12 | 13 | def filter(self, value): 14 | import grp 15 | keyfile = '/etc/ssh/ssh_host_ecdsa_key' 16 | pubkeyfile = keyfile + '.pub' 17 | # write the ssh_host_key_rsa file 18 | if os.path.exists(keyfile): 19 | os.chmod(keyfile, 0600) 20 | f = open(keyfile, 'w') 21 | f.write(value) 22 | f.close() 23 | os.chmod(keyfile, 0400) 24 | 25 | # regenerate the public key 26 | if os.path.exists(pubkeyfile): 27 | os.unlink(pubkeyfile) 28 | os.system('ssh-keygen -f %s -y > %s' % (keyfile, pubkeyfile)) 29 | os.chmod(pubkeyfile, 0644) 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /src/screens/common.css: -------------------------------------------------------------------------------- 1 | @import url(palette-blue.css); 2 | 3 | .centeredImage { 4 | text-align: center; 5 | margin-top: 0px; 6 | margin-bottom: 0px; 7 | padding: 0px; 8 | } 9 | 10 | /* place image at the left with text floating around */ 11 | .FloatLeft { 12 | float: left; 13 | margin: 4px; 14 | } 15 | 16 | /* place image at the right with text floating around */ 17 | .FloatRight { 18 | float: right; 19 | margin: 4px; 20 | } 21 | 22 | .TitleLogo { 23 | position: absolute; 24 | top: 0px; 25 | left: 80%; 26 | margin: 7px; 27 | } 28 | 29 | .Padded { 30 | padding-left: 5pt; 31 | padding-right: 5pt; 32 | } 33 | 34 | .Shifted { 35 | padding-left: 10pt; 36 | padding-right: 5pt; 37 | } 38 | 39 | .Watermark { 40 | background-image: url(watermark.png); 41 | background-repeat: no-repeat; 42 | background-attachment: fixed; 43 | background-position: center center; 44 | } 45 | 46 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/plugins/ssh_host_ed25519_key.py: -------------------------------------------------------------------------------- 1 | # $Id: ssh_host_dsa_key.py,v 1.1 2012/07/31 23:20:10 phil Exp $ 2 | # Sync the host dsa key 3 | 4 | import rocks.commands 5 | 6 | import os, sys 7 | import tempfile 8 | 9 | class plugin(rocks.commands.sec_attr_plugin): 10 | def get_sec_attr(self): 11 | return 'ssh_host_ed25519_key' 12 | 13 | def filter(self, value): 14 | import grp 15 | keyfile = '/etc/ssh/ssh_host_ed25519_key' 16 | pubkeyfile = keyfile + '.pub' 17 | # write the ssh_host_key_rsa file 18 | if os.path.exists(keyfile): 19 | os.chmod(keyfile, 0600) 20 | f = open(keyfile, 'w') 21 | f.write(value) 22 | f.close() 23 | os.chmod(keyfile, 0400) 24 | 25 | # regenerate the public key 26 | if os.path.exists(pubkeyfile): 27 | os.unlink(pubkeyfile) 28 | os.system('ssh-keygen -f %s -y > %s' % (keyfile, pubkeyfile)) 29 | os.chmod(pubkeyfile, 0644) 30 | 31 | 32 | 33 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/index.sgml: -------------------------------------------------------------------------------- 1 | 2 | Frequently Asked Questions 3 | 4 |
5 | Installation 6 | 7 | &installation-troubleshoot; 8 | &kickstartfile-compute; 9 | &grub-stage2; 10 | &smalldisk; 11 | &floppy; 12 | 13 |
14 | 15 | 16 |
17 | Configuration 18 | 19 | &removenode; 20 | &startx; 21 | &dell-powerconnect; 22 | &export; 23 | &disable-reinstall; 24 | 25 |
26 | 27 |
28 | System Administration 29 | 30 | &faq-411; 31 | 32 |
33 | 34 |
35 | -------------------------------------------------------------------------------- /src/sec_attr_plugins/plugins/root_pw.py: -------------------------------------------------------------------------------- 1 | # $Id: root_pw.py,v 1.2 2011/08/31 00:53:22 anoop Exp $ 2 | 3 | import rocks.commands 4 | 5 | import os, sys 6 | import tempfile 7 | 8 | class plugin(rocks.commands.sec_attr_plugin): 9 | def get_sec_attr(self): 10 | return 'root_pw' 11 | 12 | def filter(self, value): 13 | # Open the shadow file 14 | f = open('/etc/shadow', 'r') 15 | # Write changes to temporary file 16 | tf, tfname= tempfile.mkstemp() 17 | for line in f.readlines(): 18 | line = line.strip() 19 | split_line = line.split(':') 20 | # Look for root in shadow file 21 | if split_line[0].strip() == 'root': 22 | # if we find it, change it's password 23 | split_line[1] = value 24 | # Write back new file 25 | os.write(tf, ':'.join(split_line) + '\n') 26 | f.close() 27 | os.close(tf) 28 | # Move temporary file back to original file 29 | import shutil 30 | shutil.move(tfname, '/etc/shadow') 31 | -------------------------------------------------------------------------------- /src/piece-pipe/Makefile: -------------------------------------------------------------------------------- 1 | # 2 | # $Id: Makefile,v 1.4 2010/06/22 21:26:24 mjk Exp $ 3 | # 4 | # $Log: Makefile,v $ 5 | # Revision 1.4 2010/06/22 21:26:24 mjk 6 | # Build env is now from src/devel package, nothing outside of base roll. 7 | # 8 | # Revision 1.3 2007/08/15 18:33:14 bruno 9 | # nuke another spec file 10 | # 11 | # 12 | 13 | PKGROOT = /usr/sbin 14 | REDHAT.ROOT = $(CURDIR)/../../ 15 | -include $(ROCKSROOT)/etc/Rules.mk 16 | include Rules.mk 17 | 18 | CPPFLAGS = -DVERSION="\"$(VERSION)\"" 19 | EXECS = csp draino detour detach 20 | 21 | 22 | build: $(EXECS) 23 | 24 | csp: csp.o 25 | $(CC) $(CFLAGS) -o$@ $^ 26 | 27 | draino: draino.o 28 | $(CC) $(CFLAGS) -o$@ $^ 29 | 30 | detour: detour.o 31 | $(CC) $(CFLAGS) -o$@ $^ 32 | 33 | detach: detach.o 34 | $(CC) $(CFLAGS) -o$@ $^ 35 | 36 | 37 | install:: 38 | mkdir -p $(ROOT)/$(PKGROOT) 39 | $(INSTALL) -m755 $(EXECS) $(ROOT)/$(PKGROOT) 40 | 41 | 42 | clean:: 43 | -rm -f *.spec 44 | rm -f *.o $(EXECS) 45 | 46 | -------------------------------------------------------------------------------- /nodes/serf-common.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | serf 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2017 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 12 | 13 | 14 | 15 | serf 16 | 17 | 18 | 19 | # password for local rpc autentication, so non root 20 | # user will not be able to issue command to the agent 21 | localpasswd=$RANDOM$RANDOM 22 | # save it in a local file for easy access later on 23 | # non .json file are not loaded by serf 24 | 25 | $localpasswd 26 | 27 | 28 | # only .json files are loaded by serf as configuration 29 | 30 | { 31 | "rpc_auth" : "$localpasswd" 32 | } 33 | 34 | 35 | 36 | 37 | 38 | -------------------------------------------------------------------------------- /src/usersguide/customizing/entities.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | -------------------------------------------------------------------------------- /src/usersguide/customizing/adding-applications.sgml: -------------------------------------------------------------------------------- 1 |
3 | 4 | Adding Applications to Compute Nodes 5 | 6 | 7 | If you have code you'd like to share among the compute nodes, but your 8 | code isn't in an RPM (or in a roll), then this procedure describes how 9 | you can share it with NFS. 10 | 11 | 12 | 13 | On the frontend, go to the directory 14 | /share/apps. 15 | 16 | 17 | 18 | # cd /share/apps 19 | 20 | 21 | 22 | Then add the files you'd like to share within this directory. 23 | 24 | 25 | 26 | All files will also be available on the compute nodes under: 27 | /share/apps. 28 | For example: 29 | 30 | 31 | 32 | # cd /share/apps 33 | # touch myapp 34 | # ssh compute-0-0 35 | # cd /share/apps 36 | # ls 37 | myapp 38 | 39 | 40 |
41 | 42 | -------------------------------------------------------------------------------- /tests/aggregate-results.sh: -------------------------------------------------------------------------------- 1 | #!/bin/sh 2 | 3 | failed_tests= 4 | fixed=0 5 | success=0 6 | failed=0 7 | broken=0 8 | total=0 9 | 10 | while read file 11 | do 12 | while read type value 13 | do 14 | case $type in 15 | '') 16 | continue ;; 17 | fixed) 18 | fixed=$(($fixed + $value)) ;; 19 | success) 20 | success=$(($success + $value)) ;; 21 | failed) 22 | failed=$(($failed + $value)) 23 | if test $value != 0 24 | then 25 | testnum=$(expr "$file" : 'test-results/\(t[0-9]*\)-') 26 | failed_tests="$failed_tests $testnum" 27 | fi 28 | ;; 29 | broken) 30 | broken=$(($broken + $value)) ;; 31 | total) 32 | total=$(($total + $value)) ;; 33 | esac 34 | done <"$file" 35 | done 36 | 37 | if test -n "$failed_tests" 38 | then 39 | printf "\nfailed test(s):$failed_tests\n\n" 40 | fi 41 | 42 | printf "%-8s%d\n" fixed $fixed 43 | printf "%-8s%d\n" success $success 44 | printf "%-8s%d\n" failed $failed 45 | printf "%-8s%d\n" broken $broken 46 | printf "%-8s%d\n" total $total 47 | -------------------------------------------------------------------------------- /tests/t2012-test-config-host-interface.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | # 3 | # Test dhcpd.conf file creation 4 | # 5 | 6 | test_description='Test rocks config host interface 7 | 8 | Test the command rocks config host interface' 9 | 10 | . ./test-lib.sh 11 | 12 | 13 | 14 | ip=`rocks report nextip private` 15 | test_expect_success 'config host interface - set up tests' ' 16 | rocks add host utest-0-0 membership=compute os=linux cpus=1 rack=0 rank=0 && 17 | rocks add host interface utest-0-0 eth0 ip=$ip mac=F1:F1:F1:F1:F1:F1 subnet=private 18 | ' 19 | 20 | 21 | test_expect_success 'config host interface - verify' ' 22 | #adding virtual host with same name 23 | rocks config host interface utest-0-0 iface=eth0,eth1,eth2 \ 24 | mac=F1:F1:F1:F1:F1:F0,F1:F1:F1:F1:F1:F1,F1:F1:F1:F1:F1:F2 \ 25 | module=none,none,none flag=,ks, && 26 | rocks list host interface utest-0-0 | grep eth1 | grep $ip 27 | ' 28 | 29 | 30 | test_expect_success 'config host interface - tear down' ' 31 | rocks remove host utest-0-0 32 | ' 33 | 34 | test_done 35 | -------------------------------------------------------------------------------- /nodes/google-otp-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Google OTP Authentication Support Server Side 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: google-otp-server.xml,v $ 19 | Revision 1.1 2012/10/16 21:19:43 phil 20 | move most of the logic of google otp into it's own xml files. 21 | 22 | 23 | 24 | 25 | getent group google-otp > /dev/null || groupadd -r google-otp 26 | 27 | 28 | 29 | /usr/sbin/groupadd -g 426 google-otp 30 | 31 | 32 | 33 | /opt/rocks/bin/rocks add attr Info_GoogleOTPUsers yes 34 | /opt/rocks/bin/rocks add attr Info_GoogleOTPRoot yes 35 | 36 | 37 | 38 | 39 | -------------------------------------------------------------------------------- /nodes/nas.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | NAS Appliance Node. You should create your own extend-nas.xml 8 | file to add functionality to this appliance type. 9 | 10 | 11 | 12 | 13 | $Log: nas.xml,v $ 14 | Revision 1.2 2009/01/08 23:36:01 mjk 15 | - rsh edge is conditional (no more uncomment crap) 16 | - add global_attribute commands (list, set, remove, dump) 17 | - attributes are XML entities for kpp pass (both pass1 and pass2) 18 | - attributes are XML entities for kgen pass (not used right now - may go away) 19 | - some node are now interface=public 20 | 21 | Revision 1.1 2007/12/10 21:28:33 bruno 22 | the base roll now contains several elements from the HPC roll, thus 23 | making the HPC roll optional. 24 | 25 | this also includes changes to help build and configure VMs for V. 26 | 27 | Revision 1.8 2005/02/16 00:40:04 mjk 28 | graph cleanup 29 | 30 | 31 | 32 | 33 | 34 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/installation-troubleshoot.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Insert-ethers never sees new compute nodes. I also don't see any DHCP messages 6 | from compute nodes on the frontend. What is wrong? 7 | 8 | 9 | 10 | 11 | 12 | Try bypassing the network switch connecting your nodes to the frontend. The 13 | swich may be configured to squash broadcast messages from unknown IP addresses, 14 | which drops DHCP messages from nodes. To verify your switch is indeed the 15 | problem: 16 | 17 | 18 | 19 | 20 | 21 | Connect an ethernet cable 22 | between a single compute node and the frontend's "eth0" interface. 23 | 24 | 25 | 26 | 27 | 28 | Install the compute node normally (). 29 | You should see the DHCP messages from the node at the frontend. 30 | 31 | 32 | 33 | 34 | 35 | 36 | 37 | 38 | -------------------------------------------------------------------------------- /nodes/wan.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | A new server installed over the wide area network, from a central. 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2008 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v5.0 www.rocksclusters.org 12 | 13 | 14 | 15 | $Log: wan.xml,v $ 16 | Revision 1.4 2009/02/11 19:24:20 bruno 17 | fix 18 | 19 | Revision 1.3 2009/01/23 23:19:31 mjk 20 | tweaks on the entities 21 | 22 | Revision 1.2 2009/01/23 21:50:23 bruno 23 | var tags converted to entities 24 | 25 | Revision 1.1 2008/07/29 16:48:05 bruno 26 | save the frontend's partitions in the database. 27 | 28 | 29 | 30 |
31 | 32 | &Kickstart_Lang; 33 | &Kickstart_Keyboard; 34 | 35 | 36 | 37 | --url http://&Kickstart_PublicHostname;/&Kickstart_PrivateKickstartBasedir;/&distribution;/&arch; 38 | 39 |
40 | 41 |
42 | 43 | -------------------------------------------------------------------------------- /src/usersguide/community/office-hours.sgml: -------------------------------------------------------------------------------- 1 |
2 | Office Hours 3 | 4 | 5 | Another mode in which to get support for Rocks is through "Office Hours". 6 | The idea is to have a set time in which members of the Rocks community can 7 | have a one-on-one live conversation with a Rocks developer. 8 | 9 | 10 | 11 | Office Hours is held every Tuesday from 11:30 AM to 12:30 PM Pacific Time. 12 | If that time slot is inconvienient for you, send a note to 13 | officehours@rocksclusters.org and we'll try to set up an appointment. 14 | 15 | 16 | 17 | To contact a Rocks developer during Office Hours, you can: 18 | 19 | 20 | 21 | 22 | 23 | 24 | Make an audio or video conference call with Skype to the username 25 | "rocksclusters". 26 | 27 | 28 | 29 | 30 | 31 | Call the number 858.866-9422 (country code 1). 32 | 33 | 34 | 35 | 36 | 37 | 38 | 39 |
40 | 41 | -------------------------------------------------------------------------------- /src/snmp-status/index.html.in: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | @NAME@ 6 | 7 | 8 | 9 |

@NAME@

10 | 11 |

Contents

12 | 17 | 18 |

Overview

19 | 20 |

Snmp-status is command line utility to report process and installed 21 | RPM data for compute nodes. Information is gathered over standard SNMP 22 | and reported in ASCII text. It doesn't get much simpler than 23 | that. 24 | 25 |

Who to contact

26 | 27 |

@MAILTO@, San Diego Supercomputer Center. 29 | 30 |


31 | 32 | Last modified: Tue Feb 20 22:52:25 MST 2001 33 | 34 | 35 | -------------------------------------------------------------------------------- /src/rocks-anaconda-updates/rocks-anaconda-updates.spec: -------------------------------------------------------------------------------- 1 | Summary: rocks-anaconda-updates 2 | Name: rocks-anaconda-updates 3 | Version: 7.0 4 | Release: 2 5 | License: University of California 6 | Vendor: Rocks Clusters 7 | Group: System Environment/Base 8 | Source: rocks-anaconda-updates-7.0.tar.gz 9 | Buildroot: /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot 10 | 11 | 12 | 13 | 14 | %description 15 | rocks-anaconda-updates 16 | %prep 17 | %setup 18 | %build 19 | printf "\n\n\n### build ###\n\n\n" 20 | BUILDROOT=/export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot make -f /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.spec.mk build 21 | %install 22 | printf "\n\n\n### install ###\n\n\n" 23 | BUILDROOT=/export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.buildroot make -f /export/home/repositories/rocks/src/roll/base/src/rocks-anaconda-updates/rocks-anaconda-updates.spec.mk install 24 | %files 25 | /RedHat/base/* 26 | 27 | -------------------------------------------------------------------------------- /nodes/rlo.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Remote lights Out 8 | 9 | 10 | 11 | Copyright (c) 2000 - 2017 The Regents of the University of California. 12 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 | ## Check if ksdevice specified. If so, try to find the ethernet device that 23 | ## was defined (either by mac or by device) and make certain that IP 24 | ## connectivity is restored during the install 25 | 26 | ksMac=`/bin/cat /proc/cmdline | /bin/awk 'BEGIN {RS="[[:space:]]"; FS="="} /ksdevice/{print $2}'| /usr/bin/tr [:upper:] [:lower:]` 27 | 28 | if [ "x$ksMac" != "x" ]; then 29 | ethDev=`/sbin/ifconfig -a | grep -i $ksMac | /bin/awk '{print $1}'` 30 | if [ "x$ethDev" != "x" ]; then 31 | /sbin/ifconfig $ethDev netmask &Kickstart_PublicNetmask; &Kickstart_PublicAddress; 32 | 33 | /sbin/route add default gw &Kickstart_PublicGateway; 34 | fi 35 | fi 36 | 37 | 38 | 39 | 40 | -------------------------------------------------------------------------------- /manifest: -------------------------------------------------------------------------------- 1 | anaconda 2 | anaconda-runtime 3 | caching-nameserver 4 | firerox 5 | foundation-ant 6 | foundation-coreutils 7 | foundation-gawk 8 | foundation-graphviz 9 | foundation-libxml2 10 | foundation-mercurial 11 | foundation-mysql 12 | foundation-python 13 | foundation-python-extras 14 | foundation-python-xml 15 | foundation-rcs 16 | foundation-redhat 17 | foundation-tidy 18 | foundation-wget 19 | google-authenticator 20 | jdk-6u16 21 | librocks 22 | pdsh 23 | phpMyAdmin 24 | pxeflash 25 | qrencode 26 | rocks-411 27 | rocks-411-master 28 | rocks-admin 29 | rocks-anaconda-updates 30 | rocks-backup 31 | rocks-channel 32 | rocks-config 33 | rocks-config-server 34 | rocks-devel 35 | rocks-restore-roll 36 | rocks-developersguide 37 | rocks-ekv 38 | rocks-kickstart 39 | rocks-piece-pipe 40 | rocks-pylib 41 | rocks-secattr-plugins 42 | rocks-sec-channel-client 43 | rocks-sec-channel-server 44 | rocks-screens 45 | rocks-snmp-status 46 | rocks-sql 47 | rocks-ssl 48 | roll-base-kickstart 49 | roll-base-usersguide 50 | sun-javadb-client 51 | sun-javadb-common 52 | sun-javadb-core 53 | sun-javadb-demo 54 | sun-javadb-docs 55 | sun-javadb-javadoc 56 | tentakel 57 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/mpirun-linpack.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | When I run the ethernet version of Linpack 5 | (/usr/share/mpi/bin/mpirun -nolocal -np 2 6 | -machinefile machines /opt/hpl-eth/bin/xhpl), 7 | I get the error: "Program /opt/hpl-eth/bin/xhpl either does not exist, is not 8 | executable, or is an erroneous argument to mpirun." 9 | What's up? 10 | 11 | 12 | 13 | 14 | 15 | mpirun requires that the executable be 16 | present on the node that is lauching the job, even if it is launched 17 | with the flag -nolocal. 18 | 19 | 20 | 21 | The workaround is to add the ethernet version hpl to your frontend, by 22 | logging into your frontend and executing: 23 | 24 | 25 | 26 | # rpm -Uvh /home/install/rocks-dist/7.2/en/os/i386/RedHat/RPMS/hpl-eth-1.0-0.i386.rpm 27 | 28 | 29 | 30 | 31 | 32 | This bug will be fixed in the next release of Rocks. 33 | 34 | 35 | 36 | 37 | 38 | 39 | 40 | -------------------------------------------------------------------------------- /src/usersguide/preface.sgml: -------------------------------------------------------------------------------- 1 | 21 | 22 | 23 | Preface 24 | 25 | 26 | 27 | Since May 2000, the Rocks group has been addressing the 28 | difficulties of deploying manageable clusters. We have been driven by one 29 | goal: make clusters easy. By 30 | easy we mean easy to deploy, manage, upgrade and 31 | scale. We are driven by this goal to help deliver the computational power 32 | of clusters to a wide range of scientific users. It is clear that making 33 | stable and manageable parallel computing platforms available to a wide 34 | range of scientists will aid immensely in improving the state of the art 35 | in parallel tools. 36 | 37 | 38 | 39 | 40 | -------------------------------------------------------------------------------- /nodes/foundation-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Rocks Foundation Packages 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: foundation-client.xml,v $ 19 | Revision 1.9 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.8 2011/07/23 02:30:14 phil 23 | Viper Copyright 24 | 25 | Revision 1.7 2010/09/07 23:52:46 bruno 26 | star power for gb 27 | 28 | Revision 1.6 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.5 2008/10/18 00:55:45 mjk 32 | copyright 5.1 33 | 34 | Revision 1.4 2008/03/06 23:41:30 mjk 35 | copyright storm on 36 | 37 | Revision 1.3 2007/06/23 04:03:18 mjk 38 | mars hill copyright 39 | 40 | Revision 1.2 2006/01/18 06:12:42 mjk 41 | more foundation work 42 | 43 | Revision 1.1 2005/07/11 22:00:50 mjk 44 | added foundation to graph 45 | 46 | 47 | 48 | 49 | 50 | 51 | -------------------------------------------------------------------------------- /src/pxeflash/README: -------------------------------------------------------------------------------- 1 | How to Customize your floppy bootimage 2 | 3 | The file FDSTD.288 is a 2.8MB FreeDOS bootable floppy image and hence 4 | this procedure supports DOS-based, floppy flashing programs only. 5 | 6 | The FDSTD.288 file was downloaded from 7 | http://www.fdos.org/bootdisks/autogen/FDSTD.288.gz 8 | 9 | 10 | 1. Place your specific flash files (dos-based flash util, .bin file 11 | and perhaps autoexec.bat) in the addon/ subdirectory 12 | 13 | 2. # make build 14 | this will create a floppy image called pxeflash-FDSTD.288. This image includes 15 | all files in the addon subdirectory. 16 | 17 | 3. # make install 18 | this will copy your floppy image to /tftpboot/pxelinux/pxeflash 19 | 20 | 4. Set a node to pxeboot from your flash image 21 | # rocks set host boot action=pxeflash 22 | 23 | You will need to manually reset the node to install or boot mode using 24 | # rocks set host boot action=os 25 | 26 | 27 | Important Acknowledgement 28 | 29 | We are grateful for the contribution of Jeremy Lawrence for describing the 30 | basic approach in his Rocks discussion list posting 31 | https://lists.sdsc.edu/pipermail/npaci-rocks-discussion/2005-June/012092.html 32 | 33 | -------------------------------------------------------------------------------- /src/ekv/check_port.py: -------------------------------------------------------------------------------- 1 | #!@PYTHON@ 2 | 3 | import os,sys 4 | import socket 5 | import rocks.app 6 | 7 | class App(rocks.app.Application): 8 | 9 | def __init__(self, argv): 10 | rocks.app.Application.__init__(self,argv) 11 | 12 | self.usage_name = 'Check Port' 13 | self.usage_version = '4.2' 14 | 15 | self.node_name = 'localhost' 16 | self.port_num = 5901 17 | 18 | nodehelp = '(nodename default=%s)' % self.node_name 19 | porthelp = '(port number default=%d)' % (self.port_num) 20 | self.getopt.l.extend([ ('node=',nodehelp),('port=',porthelp) ]) 21 | self.getopt.s.extend([('n:',nodehelp),('p:',porthelp)]) 22 | return 23 | 24 | def parseArg(self, c): 25 | rocks.app.Application.parseArg(self,c) 26 | 27 | key, val = c 28 | if key in ('--port'): 29 | self.port_num = int(val) 30 | elif key in ('--node'): 31 | self.node_name = val 32 | 33 | return 34 | 35 | def run(self): 36 | s = socket.socket(socket.AF_INET,socket.SOCK_STREAM) 37 | try: 38 | s.connect((self.node_name,self.port_num)) 39 | except: 40 | print 'closed' 41 | sys.exit(1) 42 | 43 | print 'open' 44 | s.close() 45 | sys.exit(0) 46 | 47 | app = App(sys.argv) 48 | app.parseArgs() 49 | app.run() 50 | 51 | -------------------------------------------------------------------------------- /nodes/devel.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Copyright (c) 2000 - 2017 The Regents of the University of California. 7 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 8 | 9 | 10 | 11 | rocks-devel 12 | cvs 13 | libtool 14 | autoconf 15 | automake 16 | device-mapper-devel 17 | foundation-tidy 18 | isomd5sum-devel 19 | iscsi-initiator-utils-devel 20 | tcp_wrappers-devel 21 | redhat-lsb 22 | 23 | rocks-restore-roll 24 | 25 | NetworkManager-devel 26 | NetworkManager-glib-devel 27 | ncurses 28 | ncurses-devel 29 | 30 | libarchive-devel 31 | libblkid-devel 32 | libIDL 33 | libIDL-devel 34 | 35 | createrepo 36 | squashfs-tools 37 | syslinux-perl 38 | 39 | swig 40 | 41 | glibc-static 42 | 43 | 44 | -------------------------------------------------------------------------------- /src/environment-modules/modules.csh: -------------------------------------------------------------------------------- 1 | if ($?tcsh) then 2 | set modules_shell="tcsh" 3 | else 4 | set modules_shell="csh" 5 | endif 6 | set exec_prefix='/usr/share/Modules/bin' 7 | 8 | set prefix="" 9 | set postfix="" 10 | 11 | if ( $?histchars ) then 12 | set histchar = `echo $histchars | cut -c1` 13 | set _histchars = $histchars 14 | 15 | set prefix = 'unset histchars;' 16 | set postfix = 'set histchars = $_histchars;' 17 | else 18 | set histchar = \! 19 | endif 20 | 21 | if ($?prompt) then 22 | set prefix = "$prefix"'set _prompt="$prompt";set prompt="";' 23 | set postfix = "$postfix"'set prompt="$_prompt";unset _prompt;' 24 | endif 25 | 26 | if ($?noglob) then 27 | set prefix = "$prefix""set noglob;" 28 | set postfix = "$postfix""unset noglob;" 29 | endif 30 | set postfix = "set _exit="'$status'"; $postfix; test 0 = "'$_exit;' 31 | 32 | alias module $prefix'eval `'$exec_prefix'/modulecmd '$modules_shell' '$histchar'*`; '$postfix 33 | unset exec_prefix 34 | unset prefix 35 | unset postfix 36 | 37 | setenv MODULESHOME /usr/share/Modules 38 | 39 | if (! $?MODULEPATH ) then 40 | setenv MODULEPATH `sed -n 's/[ #].*$//; /./H; $ { x; s/^\n//; s/\n/:/g; p; }' ${MODULESHOME}/init/.modulespath` 41 | endif 42 | 43 | if (! $?LOADEDMODULES ) then 44 | setenv LOADEDMODULES "" 45 | endif 46 | -------------------------------------------------------------------------------- /src/usersguide/advanced/central.sgml: -------------------------------------------------------------------------------- 1 |
2 | Frontend Central Server 3 | 4 | 5 | A Central Server is a Rocks Frontend node that can kickstart other frontends and provide rolls over the network, in a process called WAN kickstart. All Rocks frontends have the ability to act as central servers. 6 | 7 | 8 | 9 | The standard Rocks distribution (located under /export/rocks/install) contains 10 | a distribution suitable for WAN kickstart. 11 | The only steps you must take is to open "www" and "https" access on your 12 | frontend for the purpose of RPM package transfer. 13 | See . 14 | 15 | 16 | 17 |
18 | Adding Rolls to Serve from a Central 19 | 20 | 21 | You may wish to serve rolls from your central server that you have not 22 | installed on your central server when the central server was initially 23 | installed. 24 | All frontends will serve the rolls they were built with to client frontends, 25 | but often it is advantageous to serve other rolls as well. 26 | 27 | 28 | 29 | First, you must download the Roll ISO image to your central. 30 | Then, execute: 31 | 32 | 33 | 34 | # rocks add roll <rollname>*.iso 35 | 36 | 37 |
38 | 39 |
40 | -------------------------------------------------------------------------------- /src/usersguide/community/mailing-list.sgml: -------------------------------------------------------------------------------- 1 |
2 | Email Discussion List 3 | 4 | 5 | The primary mode for support for Rocks is through our email discussion list. 6 | 7 | 8 | 9 | Click 10 | here 11 | to subscribe. 12 | 13 | 14 | 15 | The 16 | archives 17 | hold messages from the Rocks discussion list dating back to June 2002. 18 | 19 | 20 | If you need to search old messages you can use the following links: 21 | 22 | 23 | 24 | 25 | 26 | 27 | Mailing list ARChives 28 | 29 | 30 | 31 | 32 | 33 | 35 | Google with the extra directive 36 | site://lists.sdsc.edu/pipermail/npaci-rocks-discussion in the 37 | search bar (to limit search results to Rocks mailing list only). 38 | 39 | 40 | 41 | 42 | 43 | 44 |
45 | 46 | -------------------------------------------------------------------------------- /nodes/ipmi-base.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Common IPMI Support 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Id: ipmi-base.xml,v 1.6 2012/05/06 05:48:10 phil Exp $ 19 | 20 | 21 | 22 | OpenIPMI-python 23 | OpenIPMI-devel 24 | OpenIPMI 25 | OpenIPMI-perl 26 | OpenIPMI-libs 27 | OpenIPMI-tools 28 | 29 | SUNWipmi 30 | 31 | 32 | 33 | if /usr/sbin/dmidecode | fgrep IPMI 34 | then 35 | /sbin/chkconfig ipmi on 36 | else 37 | /sbin/chkconfig ipmi off 38 | fi 39 | 40 | 41 | 42 | 43 | # Load IPMI default settings only once, but keep a log of what we did. 44 | 45 | if [ -f /etc/sysconfig/ipmi-settings ]; then 46 | . /etc/sysconfig/ipmi-settings 47 | mv /etc/sysconfig/ipmi-settings /etc/sysconfig/ipmi-settings.bak 48 | fi 49 | 50 | 51 | 52 | 53 | 54 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/floppy.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | My compute nodes don't have a CD drive and my network cards don't PXE 5 | boot, but my compute nodes do have a floppy drive. 6 | How can I install the compute nodes? 7 | 8 | 9 | 10 | 11 | 12 | You will create a boot floppy that emulates the PXE protocol. 13 | This is accomplished by going to the web site: 14 | 15 | 16 | 17 | ROM-o-matic.net 18 | 19 | 20 | 21 | Then click on the version number under the 22 | Latest Production Release (as of this writing, 23 | this is version 5.4.3). 24 | 25 | 26 | 27 | Select your device driver in item 1. 28 | Keep the default setting in item 2 (Floppy bootable ROM Image). 29 | Then click "Get ROM" in item 4. 30 | 31 | 32 | 33 | We suggest using dd to copy the 34 | downloaded floppy image to the floppy media. 35 | For example: 36 | 37 | 38 | 39 | # dd if=eb-5.4.0-pcnet32.zdsk of=/dev/fd0 40 | 41 | 42 | 43 | Then run insert-ethers on your frontend 44 | and boot your compute node with the floppy. 45 | 46 | 47 | 48 | 49 | 50 | -------------------------------------------------------------------------------- /nodes/auto-kickstart.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Force a kickstart on power fail or system crash. 8 | 9 | 10 | 11 | 12 | 13 | $Log: auto-kickstart.xml,v $ 14 | Revision 1.3 2009/03/30 19:29:50 anoop 15 | Merging xml files from Solaris Roll to the Linux roll. 16 | More files to follow 17 | 18 | Revision 1.2 2008/09/08 19:14:47 bruno 19 | need xen install kernels too 20 | 21 | Revision 1.1 2007/12/10 21:28:33 bruno 22 | the base roll now contains several elements from the HPC roll, thus 23 | making the HPC roll optional. 24 | 25 | this also includes changes to help build and configure VMs for V. 26 | 27 | Revision 1.5 2005/02/16 00:40:04 mjk 28 | graph cleanup 29 | 30 | Revision 1.4 2005/02/01 23:37:10 mjk 31 | remove roll attrs 32 | 33 | Revision 1.3 2004/02/03 00:57:25 mjk 34 | decolorize / derollize 35 | 36 | Revision 1.2 2004/01/30 00:22:54 mjk 37 | color hpc green 38 | 39 | Revision 1.1 2003/07/07 20:47:51 bruno 40 | initial release 41 | 42 | Revision 1.1 2002/04/22 20:32:50 mjk 43 | Added EOL DB2 Profile 44 | 45 | 46 | 47 | rocks-boot-auto 48 | rocks-boot-xen 49 | 50 | 51 | -------------------------------------------------------------------------------- /nodes/selinux.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Configure SELinux 8 | 9 | 10 | 11 | Copyright (c) 2000 - 2017 The Regents of the University of California. 12 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 13 | 14 | 15 | 16 | 17 | $Log: selinux.xml,v $ 18 | Revision 1.10 2012/05/06 05:48:10 phil 19 | Copyright Storm for Mamba 20 | 21 | Revision 1.9 2012/04/02 05:16:55 phil 22 | Not sure why the second config is sometimes needed... but it is. 23 | 24 | Revision 1.8 2011/07/23 02:30:15 phil 25 | Viper Copyright 26 | 27 | Revision 1.7 2010/09/07 23:52:47 bruno 28 | star power for gb 29 | 30 | Revision 1.6 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.5 2009/03/31 00:23:29 anoop 34 | Merged the remainder of the solaris XML files to the base roll 35 | 36 | Revision 1.4 2008/10/18 00:55:46 mjk 37 | copyright 5.1 38 | 39 | Revision 1.3 2008/03/06 23:41:30 mjk 40 | copyright storm on 41 | 42 | Revision 1.2 2007/06/23 04:03:18 mjk 43 | mars hill copyright 44 | 45 | Revision 1.1 2005/03/12 00:05:47 bruno 46 | new files 47 | 48 | 49 | 50 | 51 | 52 | 53 | SELINUX=disabled 54 | 55 | 56 | 57 | 58 | -------------------------------------------------------------------------------- /nodes/google-otp.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Google OTP Authentication Support 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: google-otp.xml,v $ 19 | Revision 1.2 2012/10/18 05:24:48 phil 20 | add qrencode package 21 | 22 | Revision 1.1 2012/10/17 23:18:42 phil 23 | Forgot to check in google-otp.xml 24 | Additional packages for Bootstrap on CentOS 6.3 25 | 26 | 27 | 28 | 29 | google-authenticator 30 | qrencode 31 | 32 | 33 | 34 | sed -e '/ChallengeResponseAuthentication/s/no/yes/g' /etc/ssh/sshd_config > /tmp/sshd_config 35 | 36 | 37 | 38 | /bin/grep -q 'google-otp' /etc/pam.d/sshd 39 | if [ $? -ne 0 ] && [ -f /etc/pam.d/sshd ]; then 40 | sed -e '/#%PAM-1.0/s@#%PAM-1.0@#%PAM-1.0\nauth include google-otp@' /etc/pam.d/sshd > /tmp/pamd_sshd 41 | 42 | /bin/rm /tmp/pamd_sshd 43 | fi 44 | 45 | 46 | 47 | 48 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/security-updates.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | What security patches should I be running with Rocks? 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | Every Rocks release is a snapshot of the most up to date 14 | software packages from Red Hat for whatever release Rocks is 15 | based on. For example, Rocks 2.2 included all the Red Hat 7.2 16 | updates on the of our release. If you use rocks-dist to 17 | mirror your distribution from us you will get updates to our 18 | software and some Red Hat updates as we see fit to push them 19 | out. There is no automated method of updating the software 20 | packages on your frontend machine. This means you will have 21 | to treat it like any other machine on the Internet and track 22 | security updates yourself and updates vulnerable services. 23 | Only you can decide what level of security is appropriate for 24 | your site. Some sites prefer to secure only network services 25 | while others secure the network and local host services. 26 | 27 | 28 | 29 | 30 | 31 | The default Rocks configuration sets up and a firewall on the 32 | frontend and permits only SSH traffic into the cluster. The 33 | minimum security requirement is to track SSH updates and apply 34 | them to your frontend when needed. 35 | 36 | 37 | 38 | 39 | 40 | 41 | 42 | -------------------------------------------------------------------------------- /nodes/foundation-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Rocks Foundation Packages 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | 19 | $Log: foundation-server.xml,v $ 20 | Revision 1.12 2012/05/06 05:48:10 phil 21 | Copyright Storm for Mamba 22 | 23 | Revision 1.11 2011/07/23 02:30:14 phil 24 | Viper Copyright 25 | 26 | Revision 1.10 2010/09/07 23:52:46 bruno 27 | star power for gb 28 | 29 | Revision 1.9 2009/05/20 12:12:00 mjk 30 | nodes have RO access to DB 31 | 32 | Revision 1.8 2009/05/01 19:06:48 mjk 33 | chimi con queso 34 | 35 | Revision 1.7 2009/04/27 19:43:42 mjk 36 | no lib64 in foundation 37 | 38 | Revision 1.6 2008/10/18 00:55:45 mjk 39 | copyright 5.1 40 | 41 | Revision 1.5 2008/03/06 23:41:30 mjk 42 | copyright storm on 43 | 44 | Revision 1.4 2007/06/23 04:03:18 mjk 45 | mars hill copyright 46 | 47 | Revision 1.3 2006/01/18 06:12:42 mjk 48 | more foundation work 49 | 50 | Revision 1.2 2005/09/02 00:05:47 bruno 51 | pushing toward 4.1 beta 52 | 53 | Revision 1.1 2005/07/11 22:00:50 mjk 54 | added foundation to graph 55 | 56 | 57 | 58 | 59 | 60 | 61 | -------------------------------------------------------------------------------- /src/updates.img/Makefile: -------------------------------------------------------------------------------- 1 | REDHAT.ROOT = $(CURDIR)/../../ 2 | ROCKSROOT = ../devel/devel 3 | 4 | -include $(ROCKSROOT)/etc/Rules.mk 5 | include Rules.mk 6 | 7 | SEDSPEC = -e 's%@VERSION.MAJOR@%$(VERSION.MAJOR)%g' 8 | 9 | 10 | IMAGESDIR = ../web-server/src/www.rockclusters.org/images 11 | IMAGES = nsfwantsyou.png \ 12 | pyramids.png rocks-avalanche-logo.png 13 | 14 | 15 | 16 | install-images: 17 | for i in $(IMAGES) ; do \ 18 | cp $(REDHAT.ROOT)/$(IMAGESDIR)/$$i pixmaps/rnotes/ ; \ 19 | done 20 | 21 | build-updates-img: build-updates-img.py 22 | $(SED) $(SEDSPEC) $^ > $@ 23 | chmod +x $@ 24 | 25 | build: build-updates-img install-images 26 | ./build-updates-img 27 | dd if=/dev/zero of=updates.img bs=1024 \ 28 | count=`du -s -k extra | awk '{ print int($$1*1.1) }'` 29 | echo y | /sbin/mkfs -t ext2 updates.img 30 | mkdir mnt 31 | mount -o loop updates.img mnt 32 | mkdir -p mnt/pixmaps 33 | cp pixmaps/*png mnt/pixmaps 34 | mkdir -p mnt/pixmaps/rnotes 35 | cp pixmaps/rnotes/*png mnt/pixmaps/rnotes 36 | (cd extra && cp --recursive . ../mnt) 37 | umount mnt 38 | 39 | install:: 40 | mkdir -p $(ROOT)/images 41 | install -ma+r updates.img $(ROOT)/images 42 | mkdir -p $(ROOT)/RedHat/base 43 | install -ma+r comps.xml $(ROOT)/RedHat/base 44 | 45 | clean:: 46 | rm -f updates.img 47 | rm -f build-updates-img 48 | rm -f $(SPECFILE).in 49 | rm -rf mnt 50 | rm -rf extra 51 | rm -rf rocks-dist 52 | ( cd pixmaps/rnotes ; rm -f $(IMAGES) ) 53 | -------------------------------------------------------------------------------- /nodes/root.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | The root node that is used by rolls in order to tell 'kpp' which node 8 | should be the root of the traversed graph 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: root.xml,v $ 19 | Revision 1.8 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.7 2011/07/23 02:30:15 phil 23 | Viper Copyright 24 | 25 | Revision 1.6 2010/09/07 23:52:47 bruno 26 | star power for gb 27 | 28 | Revision 1.5 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.4 2008/10/18 00:55:45 mjk 32 | copyright 5.1 33 | 34 | Revision 1.3 2008/03/06 23:41:30 mjk 35 | copyright storm on 36 | 37 | Revision 1.2 2007/06/23 04:03:18 mjk 38 | mars hill copyright 39 | 40 | Revision 1.1 2005/03/01 00:22:26 mjk 41 | moved to base roll 42 | 43 | Revision 1.2 2005/02/15 22:58:52 mjk 44 | dtd is gone 45 | 46 | Revision 1.1 2004/03/08 21:01:23 mjk 47 | *** empty log message *** 48 | 49 | Revision 1.3 2004/02/03 20:00:36 mjk 50 | decolorize 51 | 52 | Revision 1.2 2004/01/30 04:53:42 mjk 53 | colorize 54 | 55 | Revision 1.1 2003/07/07 20:04:55 bruno 56 | new for rolls 57 | 58 | 59 | 60 | 61 | 62 | -------------------------------------------------------------------------------- /src/usersguide/customizing/boot-floppy.sgml: -------------------------------------------------------------------------------- 1 |
3 | 4 | Customizing a Boot Floppy 5 | 6 | 7 | For compute nodes without a CD or a PXE-enabled network card, there are 8 | instances where the base RedHat boot floppy image does not support the 9 | node's network card. 10 | The following procedure describes how to customize the drivers that are 11 | included in the boot floppy image. 12 | 13 | 14 | 15 | First, check out the Rocks source code. 16 | See for details. 17 | 18 | 19 | 20 | Change into the directory where you will build the boot floppy: 21 | 22 | 23 | 24 | # cd src/rocks/boot/&redhat-version-x86;/images/i386 25 | 26 | 27 | 28 | Edit the file boot-floppy-drivers and 29 | replace one (or more) of the existing driver names with your 30 | desired drivers. 31 | 32 | 33 | 34 | Build the boot floppy image. 35 | 36 | 37 | 38 | # make 39 | 40 | 41 | 42 | The new boot floppy image is named 43 | bootnet.rocks. 44 | Copy this image to a floppy disk: 45 | 46 | 47 | 48 | # dd if=bootnet.rocks of=/dev/fd0 49 | 50 | 51 | 52 | Put the floppy into your first compute node and follow the procedure 53 | described at . 54 | 55 | 56 |
57 | 58 | -------------------------------------------------------------------------------- /nodes/fail2ban.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | Install fail2ban on node with public connection 6 | 7 | 8 | 11 | python-inotify 12 | fail2ban 13 | 14 | 15 | sed -i 's/\(^ignoreip = 127.0.0.1\/8.*$\)/\1 &Kickstart_PrivateNetwork;\/&Kickstart_PrivateNetmaskCIDR;/' /etc/fail2ban/jail.conf 16 | 17 | 18 | 19 | 20 | # disable sending email everytime an event is triggered in fail2ban 21 | sed -i 's/\([ ]*sendmail-whois.*\)/#\1/' /etc/fail2ban/jail.conf 22 | 23 | 24 | 25 | 26 | [DEFAULT] 27 | bantime = 900 28 | ignoreip = 127.0.0.1/8 10.1.124.0/24 29 | 30 | [sshd] 31 | enabled = true 32 | 33 | 34 | 35 | 36 | # This file is part of the fail2ban-firewalld package to configure the use of 37 | # the firewalld actions as the default actions. You can remove this package 38 | # (along with the empty fail2ban meta-package) if you do not use firewalld 39 | # [DEFAULT] 40 | # banaction = firewallcmd-ipset 41 | 42 | 43 | /usr/bin/systemctl enable fail2ban 44 | 45 | 46 | -------------------------------------------------------------------------------- /src/screens/selected.html: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 |

Selected Rolls

10 | 11 |

12 | No rolls have been selected. 13 |

14 | 15 |

16 | If you have CD/DVD-based rolls (that is, ISO images that have been burned 17 | onto CDs or a DVD), then click the CD/DVD-based Roll button. 18 | The media tray will eject. 19 | Then, place your first roll disk in the tray and click Continue. 20 | Repeat this process for each roll disk. 21 |

22 | 23 |

24 | If you are performing a network-based installation (also known as 25 | a central installation), then input the name of your roll server 26 | into the Hostname of Roll Server field and then click the 27 | Download button. 28 | This will query the roll server and all the rolls that the roll server 29 | has available will be displayed. 30 | Click the selected checkbox for each roll you will to install 31 | from the roll server. 32 |

33 | 34 |

35 | When you have completed your roll selections, click the Next button 36 | to proceed to cluster input screens (e.g., IP address selection, root 37 | password setup, etc.). 38 |

39 | 40 |
42 |
43 | 44 | 45 | 46 | 47 | 48 | -------------------------------------------------------------------------------- /src/usersguide/clusterinstall/minimum.sgml: -------------------------------------------------------------------------------- 1 |
2 | Minimum Hardware Requirements 3 | 4 | 5 | Frontend Node 6 | 7 | 8 | 9 | 10 | 11 | Disk Capacity: 12 | 36 GB 13 | 14 | 15 | 16 | 17 | 18 | Memory Capacity: 19 | 2 GB 20 | 21 | 22 | 23 | 24 | 25 | Ethernet: 26 | 2 physical ports (e.g., "eth0" and "eth1") 27 | 28 | 29 | 30 | 31 | 32 | BIOS Boot Order: 33 | CD, Hard Disk 34 | 35 | 36 | 37 | 38 | 39 | Compute Node 40 | 41 | 42 | 43 | 44 | Disk Capacity: 45 | 36 GB 46 | 47 | 48 | 49 | 50 | 51 | Memory Capacity: 52 | 2 GB 53 | 54 | 55 | 56 | 57 | 58 | Ethernet: 59 | 1 physical port (e.g., "eth0") 60 | 61 | 62 | 63 | 64 | 65 | BIOS Boot Order: 66 | PXE (Network Boot), Hard Disk 67 | 68 | 69 | 70 | 71 |
72 | 73 | -------------------------------------------------------------------------------- /nodes/routes-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Set up static routes 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2017 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 12 | 13 | 14 | 15 | 16 | $Log: routes-client.xml,v $ 17 | Revision 1.13 2012/05/06 05:48:10 phil 18 | Copyright Storm for Mamba 19 | 20 | Revision 1.12 2011/07/23 02:30:15 phil 21 | Viper Copyright 22 | 23 | Revision 1.11 2011/03/24 19:37:00 phil 24 | Wrap routes report inside of XML tag to make it like interfaces,networks. 25 | Add ability to report host addr to output a python dictionary 26 | mod routes-*.xml and sync host network to use new output format 27 | 28 | Revision 1.10 2010/09/07 23:52:47 bruno 29 | star power for gb 30 | 31 | Revision 1.9 2009/05/01 19:06:48 mjk 32 | chimi con queso 33 | 34 | Revision 1.8 2009/03/31 00:23:29 anoop 35 | Merged the remainder of the solaris XML files to the base roll 36 | 37 | Revision 1.7 2009/03/23 23:03:57 bruno 38 | can build frontends and computes 39 | 40 | Revision 1.1 2009/03/12 22:21:04 mjk 41 | - use the new route commands 42 | - kill routes-client (move code to routes-base) 43 | 44 | 45 | 46 | 47 | 48 | 49 | /opt/rocks/bin/rocks report host route &hostname; 50 | 51 | 52 | 53 | 54 | 55 | -------------------------------------------------------------------------------- /src/firerox/userChrome.css: -------------------------------------------------------------------------------- 1 | /* 2 | * Edit this file and copy it as userChrome.css into your 3 | * profile-directory/chrome/ 4 | */ 5 | 6 | /* 7 | * This file can be used to customize the look of Mozilla's user interface 8 | * You should consider using !important on rules which you want to 9 | * override default settings. 10 | */ 11 | 12 | /* 13 | * Do not remove the @namespace line -- it's required for correct functioning 14 | */ 15 | @namespace url("http://www.mozilla.org/keymaster/gatekeeper/there.is.only.xul"); /* set default namespace to XUL */ 16 | 17 | 18 | /* 19 | * Some possible accessibility enhancements: 20 | */ 21 | /* 22 | * Make all the default font sizes 20 pt: 23 | * 24 | * * { 25 | * font-size: 20pt !important 26 | * } 27 | */ 28 | /* 29 | * Make menu items in particular 15 pt instead of the default size: 30 | * 31 | * menupopup > * { 32 | * font-size: 15pt !important 33 | * } 34 | */ 35 | /* 36 | * Give the Location (URL) Bar a fixed-width font 37 | * 38 | * #urlbar { 39 | * font-family: monospace !important; 40 | * } 41 | */ 42 | 43 | /* 44 | * Eliminate the throbber and its annoying movement: 45 | * 46 | * #throbber-box { 47 | * display: none !important; 48 | * } 49 | */ 50 | 51 | /* 52 | * For more examples see http://www.mozilla.org/unix/customizing.html 53 | */ 54 | 55 | #toolbar-menubar { 56 | display: none; 57 | } 58 | 59 | #nav-bar { 60 | display: none; 61 | } 62 | 63 | #PersonalToolbar { 64 | display: none; 65 | } 66 | 67 | .tabs-newtab-button {display: none;} 68 | 69 | -------------------------------------------------------------------------------- /src/usersguide/customizing/rsh.sgml: -------------------------------------------------------------------------------- 1 |
3 | 4 | Enabling RSH on Compute Nodes 5 | 6 | 7 | The default Rocks configuration does not enable rsh commands or login 8 | to compute nodes. Instead, Rocks uses ssh as a drop in replacement 9 | for rsh. There may be some circumstances where ssh does not have 10 | exactly the same semantics of rsh. Further, there may be some users 11 | that cannot modify their application to switch from rsh to ssh. If 12 | you are one of these users you may wish to enable rsh on your cluster. 13 | 14 | 15 | 16 | Enabling rsh on your cluster has serious security implicatations. 17 | While it is true rsh is limited to the private-side network this does 18 | not mean it is as secure as ssh. 19 | 20 | 21 | 22 | Enabling rsh is done by setting an attribute. 23 | To enable rsh on all compute nodes, execute: 24 | 25 | 26 | 27 | # rocks set appliance attr compute rsh true 28 | 29 | 30 | 31 | To apply this configuration change to the compute nodes, reinstall all 32 | your compute nodes. 33 | 34 | 35 | 36 | If you only want to enable rsh on a specific node (e.g., compute-0-0), execute: 37 | 38 | 39 | 40 | # rocks set host attr compute-0-0 rsh true 41 | 42 | 43 | 44 | To apply this configuration change to compute-0-0, reinstall compute-0-0. 45 | 46 | 47 |
48 | 49 | -------------------------------------------------------------------------------- /nodes/ethers-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Probes the ethernet hardware on the frontend and inserts information into 8 | the database. 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: ethers-server.xml,v $ 19 | Revision 1.10 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.9 2011/07/23 02:30:14 phil 23 | Viper Copyright 24 | 25 | Revision 1.8 2010/09/07 23:52:46 bruno 26 | star power for gb 27 | 28 | Revision 1.7 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.6 2009/03/04 23:31:09 bruno 32 | nuked 'dbreport ifcfg' 33 | 34 | Revision 1.5 2008/10/18 00:55:45 mjk 35 | copyright 5.1 36 | 37 | Revision 1.4 2008/03/06 23:41:30 mjk 38 | copyright storm on 39 | 40 | Revision 1.3 2007/06/23 04:03:18 mjk 41 | mars hill copyright 42 | 43 | Revision 1.2 2006/06/21 03:09:53 bruno 44 | updates to put the frontend networking info in the database just like 45 | a compute node 46 | 47 | Revision 1.1 2005/10/11 06:01:15 phil 48 | Call probe-ethers to insert the ethernet HW info into the DB on the 49 | frontend. 50 | 51 | 52 | 53 | 54 | 55 | 57 | 58 | 59 | 60 | 61 | 62 | 63 | -------------------------------------------------------------------------------- /nodes/fortran-development.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Fortran 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: fortran-development.xml,v $ 19 | Revision 1.9 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.8 2011/07/23 02:30:14 phil 23 | Viper Copyright 24 | 25 | Revision 1.7 2010/09/07 23:52:46 bruno 26 | star power for gb 27 | 28 | Revision 1.6 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.5 2008/10/18 00:55:45 mjk 32 | copyright 5.1 33 | 34 | Revision 1.4 2008/03/06 23:41:30 mjk 35 | copyright storm on 36 | 37 | Revision 1.3 2008/01/23 23:14:16 bruno 38 | gcc-g77 is now gcc-gfortran 39 | 40 | Revision 1.2 2007/06/23 04:03:18 mjk 41 | mars hill copyright 42 | 43 | Revision 1.1 2005/03/01 00:22:26 mjk 44 | moved to base roll 45 | 46 | Revision 1.2 2005/02/15 22:58:52 mjk 47 | dtd is gone 48 | 49 | Revision 1.1 2004/03/08 21:01:23 mjk 50 | *** empty log message *** 51 | 52 | Revision 1.3 2004/02/03 20:00:35 mjk 53 | decolorize 54 | 55 | Revision 1.2 2004/01/30 04:53:42 mjk 56 | colorize 57 | 58 | Revision 1.1 2001/09/18 17:39:52 mjk 59 | *** empty log message *** 60 | 61 | 62 | 63 | gcc-gfortran 64 | 65 | 66 | 67 | 68 | -------------------------------------------------------------------------------- /nodes/nas-exports.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Rocks Cluster NAS Node export file system definitions 7 | 8 | 9 | 10 | 11 | $Log: nas-exports.xml,v $ 12 | Revision 1.2 2009/01/23 21:50:23 bruno 13 | var tags converted to entities 14 | 15 | Revision 1.1 2007/12/10 21:28:33 bruno 16 | the base roll now contains several elements from the HPC roll, thus 17 | making the HPC roll optional. 18 | 19 | this also includes changes to help build and configure VMs for V. 20 | 21 | Revision 1.2 2005/10/06 23:57:03 bruno 22 | added 'async' flag to each /etc/exports entry. 23 | 24 | thanks to Bill Hoover for debugging the problem. 25 | 26 | Revision 1.1 2005/09/19 22:08:37 bruno 27 | partitioning tuning. 28 | 29 | - the frontend and NAS appliance will no longer nuke all partitions before 30 | the partitioning screen 31 | - a NAS appliance always displays the partitioning screen 32 | - added a nas-exports node to make it easier for people to customize 33 | their export directories on NAS appliances 34 | - hardened the software raid partitioning code 35 | 36 | 37 | 38 | 39 | 40 | 41 | 45 | 46 | if [ ! -d /export/data1 ]; then mkdir -p /export/data1; fi 47 | 48 | /export/data1 &Kickstart_PrivateNetwork;/&Kickstart_PrivateNetmask;(rw,no_root_squash,async) 49 | 50 | 51 | 52 | 53 | 54 | 55 | 56 | -------------------------------------------------------------------------------- /nodes/rpc.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | RPC support 10 | 11 | 12 | 13 | 14 | Copyright (c) 2000 - 2017 The Regents of the University of California. 15 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 16 | 17 | 18 | 19 | 20 | $Log: rpc.xml,v $ 21 | Revision 1.9 2012/05/06 05:48:10 phil 22 | Copyright Storm for Mamba 23 | 24 | Revision 1.8 2011/07/23 02:30:15 phil 25 | Viper Copyright 26 | 27 | Revision 1.7 2010/09/07 23:52:47 bruno 28 | star power for gb 29 | 30 | Revision 1.6 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.5 2009/03/31 00:23:29 anoop 34 | Merged the remainder of the solaris XML files to the base roll 35 | 36 | Revision 1.4 2008/10/18 00:55:45 mjk 37 | copyright 5.1 38 | 39 | Revision 1.3 2008/03/06 23:41:30 mjk 40 | copyright storm on 41 | 42 | Revision 1.2 2007/06/23 04:03:18 mjk 43 | mars hill copyright 44 | 45 | Revision 1.1 2005/03/01 00:22:26 mjk 46 | moved to base roll 47 | 48 | Revision 1.2 2005/02/15 22:58:52 mjk 49 | dtd is gone 50 | 51 | Revision 1.1 2004/03/08 21:01:23 mjk 52 | *** empty log message *** 53 | 54 | Revision 1.3 2004/02/03 20:00:36 mjk 55 | decolorize 56 | 57 | Revision 1.2 2004/01/30 04:53:42 mjk 58 | colorize 59 | 60 | Revision 1.1 2001/09/05 00:27:18 mjk 61 | main and packages section is correct for compute nodes 62 | 63 | 64 | 65 | portmap 66 | 67 | 68 | -------------------------------------------------------------------------------- /nodes/serf-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | serf 7 | 8 | 9 | 10 | Copyright (c) 2000 - 2017 The Regents of the University of California. 11 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 12 | 13 | 14 | 15 | 16 | 17 | # setup encryption key to talk to the other agent in the cluster 18 | encryption=`cat /etc/411-security/shared.key |grep -v BEGIN |grep -v END | base64 -d |dd bs=1 count=16 |base64` 19 | 20 | 21 | 22 | { 23 | "encrypt_key": "$encryption" 24 | } 25 | 26 | 27 | 28 | # only frontend needs to be stateful (event lost while the daemon 29 | # was down will be replayed when it is back up) 30 | # all other nodes if they lose it is ok so no need to snapshot 31 | mkdir -p /var/opt/rocks/serf/ 32 | 33 | private_ip=`/opt/rocks/bin/rocks list host interface &hostname; output-col=subnet,iface,ip | \ 34 | /bin/awk '{if ($1 == "private" && $2 !~ /vlan[0-9]*/ && $4 ~ /-*/) {print $3}}'` 35 | 36 | { 37 | "snapshot_path": "/var/opt/rocks/serf/serf.snap", 38 | "rejoin_after_leave": true, 39 | "replay_on_join": true, 40 | "bind": "$private_ip", 41 | "retry_join": ["&Kickstart_PrivateHostname;.&Kickstart_PrivateDNSDomain;"] 42 | } 43 | 44 | 45 | if [ -x /etc/init.d/rocks-serf ]; then 46 | /sbin/chkconfig rocks-serf on 47 | else 48 | /usr/bin/systemctl enable rocks-serf 49 | fi 50 | 51 | 52 | 53 | -------------------------------------------------------------------------------- /tests/t2010-rocks-appliances.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | # 3 | # Test kickstart.cgi script 4 | # 5 | 6 | test_description='Test rocks appliance commands 7 | 8 | test various aspects of rocks appliance related commands 9 | ' 10 | 11 | pushd `dirname $0` > /dev/null 12 | export TEST_DIRECTORY=`pwd` 13 | popd > /dev/null 14 | . $TEST_DIRECTORY/test-lib.sh 15 | 16 | hostname=`hostname` 17 | 18 | rocks_build=/export/rocks/install/rocks-dist/x86_64/build 19 | 20 | test_expect_success 'test rocks appliance - add routes' ' 21 | rocks add appliance route nas 10.1.1.0 10.1.1.1 netmask=255.255.255.0 && 22 | ! rocks add appliance route compute nas 10.1.1.0 10.1.1.1 netmask=255.255.255.0 && 23 | test `rocks list appliance route | grep 10.1.1.1 | wc -l` == "1" && 24 | rocks add appliance route compute 10.1.1.0 10.1.1.1 netmask=255.255.255.0 && 25 | test `rocks list appliance route | grep 10.1.1.1 | wc -l` == "2" 26 | ' 27 | 28 | test_expect_success 'test rocks appliance - add routes' ' 29 | /opt/rocks/bin/rocks add appliance test-appliance membership='test-appliance' && 30 | /opt/rocks/bin/rocks set appliance attr test-attr managed false && 31 | rocks list appliance test-appliance 32 | ' 33 | 34 | 35 | test_expect_success 'test rocks appliance - tear down' ' 36 | rocks remove appliance test-appliance && 37 | rocks remove appliance route nas 10.1.1.0 && 38 | rocks remove appliance route compute 10.1.1.0 && 39 | test `rocks list appliance route | grep 10.1.1.1 | wc -l` == "0" 40 | ' 41 | 42 | 43 | test_done 44 | 45 | -------------------------------------------------------------------------------- /nodes/postshell.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Allow interaction after all post sections have run 8 | If postshell is on the command line, wait until /mnt/sysimage/tmp/postshell 9 | is removed interactively. 10 | 11 | 12 | 13 | Copyright (c) 2000 - 2017 The Regents of the University of California. 14 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 15 | 16 | 17 | 18 | 19 | 20 | $Log: postshell.xml,v $ 21 | Revision 1.9 2012/05/06 05:48:10 phil 22 | Copyright Storm for Mamba 23 | 24 | Revision 1.8 2011/07/23 02:30:15 phil 25 | Viper Copyright 26 | 27 | Revision 1.7 2010/09/07 23:52:46 bruno 28 | star power for gb 29 | 30 | Revision 1.6 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.5 2009/03/31 00:23:29 anoop 34 | Merged the remainder of the solaris XML files to the base roll 35 | 36 | Revision 1.4 2008/10/18 00:55:45 mjk 37 | copyright 5.1 38 | 39 | Revision 1.3 2008/03/06 23:41:30 mjk 40 | copyright storm on 41 | 42 | Revision 1.2 2007/06/23 04:03:18 mjk 43 | mars hill copyright 44 | 45 | Revision 1.1 2005/03/31 04:15:10 phil 46 | debugging flag -- if postshell on the command line 47 | e.g. frontend central=rocks postshell 48 | then install will not complete until /mnt/sysimage/tmp/postshell is removed 49 | interactively by the user 50 | 51 | 52 | 53 | 54 | 55 | grep -q "postshell" /proc/cmdline 56 | if [ $? -eq 0 ]; then 57 | touch /tmp/postshell 58 | while [ -e /tmp/postshell ]; do /bin/sleep 5; done 59 | fi 60 | 61 | 62 | 63 | 64 | -------------------------------------------------------------------------------- /src/usersguide/appendix/faq/411.sgml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | How do I use user accounts from an external NIS server on my cluster? 6 | 7 | 8 | 9 | 10 | 11 | While there is no certain method to do this correctly, if necessary we recommend you use "ypcat" to periodically gather external NIS user accounts on the frontend, and let the default 411 system distribute the information inside the cluster. 12 | 13 | 14 | 15 | The following cron script will collect NIS information from your external network onto the frontend. The login files created here will be automatically distributed to cluster nodes via 411. This code courtesy of Chris Dwan at the University of Minnesota. 16 | 17 | 18 | 19 | 20 | (in /etc/cron.hourly/get-NIS on frontend) 21 | 22 | #!/bin/sh 23 | ypcat -k auto.master > /etc/auto.master 24 | ypcat -k auto.home > /etc/auto.home 25 | ypcat -k auto.net > /etc/auto.net 26 | ypcat -k auto.web > /etc/auto.web 27 | 28 | ypcat passwd > /etc/passwd.nis 29 | cat /etc/passwd.local /etc/passwd.nis > /etc/passwd.combined 30 | cp /etc/passwd.combined /etc/passwd 31 | 32 | ypcat group > /etc/group.nis 33 | cat /etc/group.local /etc/group.nis > /etc/group.combined 34 | cp /etc/group.combined /etc/group 35 | 36 | 37 | 38 | 39 | 40 | 41 | There is no way to insure that UIDs GIDs from NIS will not conflict with those already present in the cluster. You must always be careful that such collisions do not occur, as unpredicatble and undefined behavior will result. 42 | 43 | 44 | 45 | 46 | 47 | 48 | -------------------------------------------------------------------------------- /src/usersguide/customizing/postconfig.sgml: -------------------------------------------------------------------------------- 1 |
3 | 4 | Customizing Configuration of Compute Nodes 5 | 6 | 7 | Create a new XML configuration file that will extend 8 | the current compute.xml configuration 9 | file: 10 | 11 | 12 | 13 | # cd /export/rocks/install/site-profiles/&document-version;/nodes/ 14 | # cp skeleton.xml extend-compute.xml 15 | 16 | 17 | 18 | 19 | If you want to include your customizations in all cluster nodes 20 | (and not just the compute nodes) you should use 21 | extend-base.xml as a file name 22 | instead of extend-compute.xml. 23 | 24 | 25 | 26 | 27 | Inside extend-compute.xml, 28 | add your configuration 29 | scripts that will be run in the post configuration 30 | step of the Red Hat installer. 31 | 32 | 33 | 34 | Put your bash scripts in between the tags 35 | <post> 36 | and 37 | </post>: 38 | 39 | 40 | 41 | <post> 42 | < !-- insert your scripts here --> 43 | </post> 44 | 45 | 46 | 47 | To apply your customized configuration scripts to compute nodes, 48 | rebuild the distribution: 49 | 50 | 51 | 52 | # cd /export/rocks/install 53 | # rocks create distro 54 | 55 | 56 | 57 | Then, reinstall your compute nodes. 58 | 59 | 60 |
61 | 62 | -------------------------------------------------------------------------------- /tests/t2011-creation-db-schema.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | # 3 | # Test dhcpd.conf file creation 4 | # 5 | 6 | test_description='Test creation of the database schema 7 | 8 | Test the generation of a rocks database with the schema contained in 9 | node/database-schema.sh' 10 | 11 | pushd `dirname $0` > /dev/null 12 | export TEST_DIRECTORY=`pwd` 13 | popd > /dev/null 14 | . $TEST_DIRECTORY/test-lib.sh 15 | 16 | NODES_DIR=$TEST_DIRECTORY/../nodes 17 | 18 | 19 | attrs="{'os':'linux'}" 20 | 21 | # need to hack the database-schema to create the tables 22 | # in our tempdb instead of the standard cluster db 23 | test_expect_success 'test create db - setup' ' 24 | /opt/rocks/mysql/bin/mysqladmin --defaults-extra-file=/root/.rocks.my.cnf --user=root create tempdb && 25 | test -f $NODES_DIR/database-schema.xml && 26 | cat $NODES_DIR/database-schema.xml | \ 27 | rocks report post attrs="$attrs" | \ 28 | sed "s|/opt/rocks/mysql/bin/mysql|echo |g" | \ 29 | bash 30 | ' 31 | 32 | test_expect_success 'test create db - test schema' ' 33 | echo entering /tmp/tables.sql && 34 | /opt/rocks/mysql/bin/mysql --defaults-extra-file=/root/.rocks.my.cnf \ 35 | --user=root tempdb < /tmp/tables.sql && 36 | echo entering /tmp/categories.sql && 37 | /opt/rocks/mysql/bin/mysql --defaults-extra-file=/root/.rocks.my.cnf \ 38 | --user=root tempdb < /tmp/categories.sql 39 | 40 | ' 41 | 42 | 43 | test_expect_success 'test create db - tear down' ' 44 | /opt/rocks/mysql/bin/mysqladmin --defaults-extra-file=/root/.rocks.my.cnf --user=root -f drop tempdb 45 | ' 46 | 47 | test_done 48 | 49 | -------------------------------------------------------------------------------- /nodes/networking-server.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | Configure a node's network 7 | 8 | 9 | 10 | 11 | Copyright (c) 2000 - 2007 The Regents of the University of California. 12 | All rights reserved. Rocks(tm) v4.3 www.rocksclusters.org 13 | 14 | 15 | 16 | 17 | $Log: networking-server.xml,v $ 18 | Revision 1.6 2009/03/21 22:22:55 bruno 19 | - lights-out install of VM frontends with new node_rolls table 20 | - nuked 'site' columns and tables from database 21 | - worked through some bugs regarding entities 22 | 23 | Revision 1.5 2009/01/23 21:50:23 bruno 24 | var tags converted to entities 25 | 26 | Revision 1.4 2008/09/22 20:20:42 bruno 27 | change 'rocks config host interface|network' to 28 | change 'rocks report host interface|network' 29 | 30 | Revision 1.3 2008/07/24 21:10:32 bruno 31 | fix 32 | 33 | Revision 1.2 2008/07/23 00:01:06 bruno 34 | tweaks 35 | 36 | Revision 1.1 2007/12/10 21:28:33 bruno 37 | the base roll now contains several elements from the HPC roll, thus 38 | making the HPC roll optional. 39 | 40 | this also includes changes to help build and configure VMs for V. 41 | 42 | 43 | 44 | 45 | 46 | 47 | /opt/rocks/bin/rocks report host network &Kickstart_PrivateHostname; | \ 48 | /opt/rocks/bin/rocks report script > /tmp/do-network.sh 49 | 50 | chmod a+x /tmp/do-network.sh 51 | /tmp/do-network.sh 52 | 53 | /opt/rocks/bin/rocks report host interface &Kickstart_PrivateHostname; | \ 54 | /opt/rocks/bin/rocks report script > /tmp/do-interface.sh 55 | 56 | chmod a+x /tmp/do-interface.sh 57 | /tmp/do-interface.sh 58 | 59 | 60 | 61 | 62 | 63 | 64 | -------------------------------------------------------------------------------- /src/usersguide/customizing/password.sgml: -------------------------------------------------------------------------------- 1 |
2 | Changing the Root password 3 | 4 | 5 | Starting with Rocks v5.4.3, we have attempted to improve the security of the 6 | Rocks system by making changes to the way some of the important passwords are 7 | generated and stored. 8 | 9 | To change the root password on the frontend, you can now run the 10 | passwd command. 11 | To change the root password on the client nodes, you can use the rocks secure attribute infrastructure. 13 | Change the root password of a single compute host 14 | 15 | # rocks set host sec_attr compute-0-0 attr=root_pw 16 | Enter root_pw: 17 | Confirm root_pw: 18 | # rocks list host sec_attr compute 19 | HOST ATTR VALUE ENC 20 | compute-0-0: root_pw $1$G55uZRrO$5MI7Nv14U5yWfKAlauqPE0 crypt 21 | # rocks sync host sec_attr compute-0-0 22 | 23 | For more information about changes to the rocks security infrastructure, 24 | refer to the Rocks Password Infrastructure 25 | section of this usersguide 26 | 27 | The default rocks password policy is to set the root password of 28 | the frontend to be the password supplied during the installation. For client 29 | nodes, the root password will be generated at random during installation. Unless 30 | there is compelling reason to do so, (such as giving a user root access to a 31 | compute node) the Rocks group advices against changing the root password on the 32 | client nodes. 33 |
34 | -------------------------------------------------------------------------------- /tests/t2002-rocks-dhcpd.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | # 3 | # Test dhcpd.conf file creation 4 | # 5 | 6 | test_description='Test creation of dhcpd.conf 7 | 8 | Test the generation of dhcpd.conf with some basic network configuration ' 9 | 10 | . ./test-lib.sh 11 | 12 | function testIt(){ 13 | rocks report host dhcpd | rocks report script | bash && 14 | /etc/init.d/dhcpd restart 15 | if [ "$?" != "0" ]; then 16 | echo ---- failing dhcpd.conf creation ----- 17 | rocks remove host test-0-0 18 | rocks remove host test-0-0-0 19 | return -1 20 | fi 21 | return 0 22 | } 23 | 24 | 25 | 26 | test_expect_success setup ' 27 | rocks add host test-0-0 membership=compute os=linux cpus=1 rack=0 rank=0 && 28 | rocks add host test-0-0-0 membership=compute os=linux cpus=1 rack=0 rank=0 29 | ' 30 | 31 | test_expect_success 'dhcp.conf add two interfaces' ' 32 | rocks add host interface test-0-0 eth0 ip=10.1.1.200 mac=F1:F1:F1:F1:F1:F1 subnet=private && 33 | rocks add host interface test-0-0 eth1 mac=F1:F1:F1:F1:F1:F2 && 34 | testIt 35 | ' 36 | 37 | test_expect_success 'dhcp.conf add virtual host with similar name' ' 38 | #adding virtual host with same name 39 | rocks add host interface test-0-0-0 eth0 ip=10.1.1.203 mac=F1:F1:F1:F1:F1:F5 subnet=private && 40 | testIt 41 | ' 42 | 43 | test_expect_success 'dhcp.conf set the ips address' ' 44 | rocks set host interface ip test-0-0 eth1 10.1.1.201 && 45 | rocks set host interface subnet test-0-0 eth1 private && 46 | testIt 47 | ' 48 | 49 | test_expect_success 'dhcp.conf tear down' ' 50 | rocks remove host test-0-0 && 51 | rocks remove host test-0-0-0 52 | ' 53 | 54 | test_done 55 | -------------------------------------------------------------------------------- /nodes/tcl-development.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Tcl support 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: tcl-development.xml,v $ 19 | Revision 1.9 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.8 2011/07/23 02:30:15 phil 23 | Viper Copyright 24 | 25 | Revision 1.7 2010/09/07 23:52:47 bruno 26 | star power for gb 27 | 28 | Revision 1.6 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.5 2009/03/31 00:23:29 anoop 32 | Merged the remainder of the solaris XML files to the base roll 33 | 34 | Revision 1.4 2008/10/18 00:55:46 mjk 35 | copyright 5.1 36 | 37 | Revision 1.3 2008/03/06 23:41:30 mjk 38 | copyright storm on 39 | 40 | Revision 1.2 2007/06/23 04:03:18 mjk 41 | mars hill copyright 42 | 43 | Revision 1.1 2005/03/01 00:22:26 mjk 44 | moved to base roll 45 | 46 | Revision 1.2 2005/02/15 22:58:52 mjk 47 | dtd is gone 48 | 49 | Revision 1.1 2004/03/08 21:01:23 mjk 50 | *** empty log message *** 51 | 52 | Revision 1.4 2004/02/03 20:00:36 mjk 53 | decolorize 54 | 55 | Revision 1.3 2004/01/30 04:53:42 mjk 56 | colorize 57 | 58 | Revision 1.2 2001/09/05 00:27:18 mjk 59 | main and packages section is correct for compute nodes 60 | 61 | 62 | 63 | tcl 64 | expect 65 | 66 | SUNWTcl 67 | SFWexpct 68 | 69 | 70 | -------------------------------------------------------------------------------- /nodes/docbook.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | DOC Book support (needed to build rolls) 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: docbook.xml,v $ 19 | Revision 1.8 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.7 2011/07/23 02:30:14 phil 23 | Viper Copyright 24 | 25 | Revision 1.6 2010/09/07 23:52:46 bruno 26 | star power for gb 27 | 28 | Revision 1.5 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.4 2008/10/18 00:55:45 mjk 32 | copyright 5.1 33 | 34 | Revision 1.3 2008/03/06 23:41:30 mjk 35 | copyright storm on 36 | 37 | Revision 1.2 2007/06/23 04:03:18 mjk 38 | mars hill copyright 39 | 40 | Revision 1.1 2005/03/01 00:22:26 mjk 41 | moved to base roll 42 | 43 | Revision 1.2 2005/02/15 22:58:52 mjk 44 | dtd is gone 45 | 46 | Revision 1.1 2004/03/08 21:01:23 mjk 47 | *** empty log message *** 48 | 49 | Revision 1.3 2004/02/03 20:00:35 mjk 50 | decolorize 51 | 52 | Revision 1.2 2004/01/30 04:53:42 mjk 53 | colorize 54 | 55 | Revision 1.1 2003/11/05 06:26:13 mjk 56 | initial ci 57 | 58 | 59 | 60 | elinks 61 | perl-SGMLSpm 62 | docbook-style-dsssl 63 | docbook-utils 64 | docbook-style-xsl 65 | docbook-utils-pdf 66 | docbook-dtds 67 | 68 | 69 | 70 | 71 | 72 | 73 | 74 | 75 | 76 | 77 | -------------------------------------------------------------------------------- /nodes/partitions-save.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Save the partition info from this node by sending a message back to the 8 | frontend. The frontend will store the partition info in the database 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | 19 | $Log: partitions-save.xml,v $ 20 | Revision 1.9 2012/05/06 05:48:10 phil 21 | Copyright Storm for Mamba 22 | 23 | Revision 1.8 2011/07/23 02:30:15 phil 24 | Viper Copyright 25 | 26 | Revision 1.7 2010/09/07 23:52:46 bruno 27 | star power for gb 28 | 29 | Revision 1.6 2009/05/01 19:06:48 mjk 30 | chimi con queso 31 | 32 | Revision 1.5 2009/03/31 00:23:29 anoop 33 | Merged the remainder of the solaris XML files to the base roll 34 | 35 | Revision 1.4 2008/10/18 00:55:45 mjk 36 | copyright 5.1 37 | 38 | Revision 1.3 2008/03/06 23:41:30 mjk 39 | copyright storm on 40 | 41 | Revision 1.2 2007/12/17 22:23:00 bruno 42 | polish 43 | 44 | Revision 1.1 2007/12/10 21:28:33 bruno 45 | the base roll now contains several elements from the HPC roll, thus 46 | making the HPC roll optional. 47 | 48 | this also includes changes to help build and configure VMs for V. 49 | 50 | 51 | 52 | 53 | chmod a+x /tmp/product/record_partitions.py 54 | /tmp/product/record_partitions.py 55 | 56 | 57 | PYTHONPATH=/tmp /opt/rocks/bin/python /opt/rocks/lib/python2.7/site-packages/rocks/record_partitions.py 58 | 59 | 60 | 61 | -------------------------------------------------------------------------------- /src/developersguide/index.sgml: -------------------------------------------------------------------------------- 1 | 26 | 27 | 30 | 31 | %entities; 32 | 33 | ]> 34 | 35 | 36 | 37 | 38 | Roll Developer's Guide 39 | 40 | 41 | 42 | 43 | 45 | 46 | 47 | 48 | 49 | &document-version; 50 | &document-pubdate; 51 | 52 | &document-year; 53 | &document-copyright; 54 | 55 | 56 | 57 | (PDF version) 58 | 59 | 60 | 61 | 62 | This document is subject to the &rocks; License 63 | (see ). 64 | 65 | 66 | 67 | 68 | 69 | &source-overview; 70 | &source-internals; 71 | &source-building; 72 | &source-testing; 73 | &source-appendix; 74 | &source-roll-rocks-copyright; 75 | 76 | 77 | -------------------------------------------------------------------------------- /include/installclass/record_partitions.py: -------------------------------------------------------------------------------- 1 | #!/opt/rocks/bin/python 2 | 3 | import sys 4 | sys.path.append('/tmp/product') 5 | import rocks_partition 6 | import httplib 7 | import random 8 | import time 9 | import string 10 | sys.path.append('/tmp') 11 | import db_partition_info 12 | 13 | def sendit(server, req, nodepartinfo): 14 | status = 0 15 | 16 | h = httplib.HTTPSConnection(server, key_file = None, cert_file = None) 17 | h.putrequest('GET', '/install/sbin/public/setDbPartitions.cgi') 18 | 19 | h.putheader('X-Rocks-PartitionInfo', repr(nodepartinfo)) 20 | 21 | try: 22 | h.endheaders() 23 | response = h.getresponse() 24 | status = response.status 25 | except: 26 | # 27 | # assume the error occurred due to an 28 | # authorization problem 29 | # 30 | status = 403 31 | pass 32 | 33 | h.close() 34 | return status 35 | 36 | # 37 | # main 38 | # 39 | p = rocks_partition.RocksPartition() 40 | 41 | # 42 | # get the list of hard disks and software raid devices 43 | # 44 | disks = p.getDisks() + p.getRaids() 45 | nodepartinfo = p.getNodePartInfo(disks) 46 | 47 | # 48 | # only try to report the partitions back to the frontend, if this isn't 49 | # a frontend 50 | # 51 | file = open('/proc/cmdline', 'r') 52 | args = string.split(file.readline()) 53 | file.close() 54 | 55 | if 'build' not in args: 56 | random.seed(int(time.time())) 57 | 58 | for i in range(0, 5): 59 | status = sendit(db_partition_info.KickstartHost, 60 | '/install/sbin/public/setDbPartitions.cgi', 61 | nodepartinfo) 62 | 63 | if status == 200: 64 | break 65 | else: 66 | time.sleep(random.randint(0, 10)) 67 | 68 | # 69 | # mark each disk as a 'rocks' disk -- this let's us know that we have 70 | # 'seen' and configured this disk 71 | # 72 | for disk in nodepartinfo.keys(): 73 | p.isRocksDisk(nodepartinfo[disk], touchit = 1) 74 | 75 | -------------------------------------------------------------------------------- /nodes/ip-diag.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | TCP/IP Network diagnostic tools. 10 | 11 | 12 | 13 | 14 | Copyright (c) 2000 - 2017 The Regents of the University of California. 15 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 16 | 17 | 18 | 19 | 20 | $Log: ip-diag.xml,v $ 21 | Revision 1.9 2012/05/06 05:48:10 phil 22 | Copyright Storm for Mamba 23 | 24 | Revision 1.8 2011/07/23 02:30:14 phil 25 | Viper Copyright 26 | 27 | Revision 1.7 2010/09/07 23:52:46 bruno 28 | star power for gb 29 | 30 | Revision 1.6 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.5 2009/03/31 00:23:29 anoop 34 | Merged the remainder of the solaris XML files to the base roll 35 | 36 | Revision 1.4 2008/10/18 00:55:45 mjk 37 | copyright 5.1 38 | 39 | Revision 1.3 2008/03/06 23:41:30 mjk 40 | copyright storm on 41 | 42 | Revision 1.2 2007/06/23 04:03:18 mjk 43 | mars hill copyright 44 | 45 | Revision 1.1 2005/03/01 00:22:26 mjk 46 | moved to base roll 47 | 48 | Revision 1.2 2005/02/15 22:58:52 mjk 49 | dtd is gone 50 | 51 | Revision 1.1 2004/03/08 21:01:23 mjk 52 | *** empty log message *** 53 | 54 | Revision 1.3 2004/02/03 20:00:35 mjk 55 | decolorize 56 | 57 | Revision 1.2 2004/01/30 04:53:42 mjk 58 | colorize 59 | 60 | Revision 1.1 2001/09/05 00:27:17 mjk 61 | main and packages section is correct for compute nodes 62 | 63 | 64 | 65 | tcpdump 66 | iputils 67 | 68 | SFWtdmp 69 | SUNWbip 70 | 71 | 72 | 73 | 74 | 75 | 76 | 77 | 78 | 79 | 80 | -------------------------------------------------------------------------------- /nodes/411-base.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | The packages and other common elements of the 8 | 411 Secure Information Service. 9 | 10 | 11 | 12 | 13 | $Log: 411-base.xml,v $ 14 | Revision 1.2 2009/03/30 19:29:50 anoop 15 | Merging xml files from Solaris Roll to the Linux roll. 16 | More files to follow 17 | 18 | Revision 1.1 2007/12/10 21:28:33 bruno 19 | the base roll now contains several elements from the HPC roll, thus 20 | making the HPC roll optional. 21 | 22 | this also includes changes to help build and configure VMs for V. 23 | 24 | Revision 1.2 2006/02/15 22:08:26 mjk 25 | moved POW to foundation 26 | 27 | Revision 1.1 2005/03/01 17:34:02 mjk 28 | move 411 to hpc roll 29 | 30 | Revision 1.1 2005/03/01 00:22:26 mjk 31 | moved to base roll 32 | 33 | Revision 1.2 2005/02/15 22:58:52 mjk 34 | dtd is gone 35 | 36 | Revision 1.1 2004/03/08 21:01:23 mjk 37 | *** empty log message *** 38 | 39 | Revision 1.5 2004/02/03 20:00:35 mjk 40 | decolorize 41 | 42 | Revision 1.4 2004/01/30 04:53:42 mjk 43 | colorize 44 | 45 | Revision 1.3 2003/08/14 15:32:23 mjk 46 | - Built Python2 for Itanic 47 | - Root part is now 4GB 48 | - Include openssl python tools for 411 on Itanic 49 | 50 | Revision 1.2 2003/07/30 16:23:47 bruno 51 | changes for ia64 52 | 53 | Revision 1.1 2003/07/24 19:14:10 fds 54 | First design 55 | 56 | 57 | 58 | rocks-411 59 | openssl 60 | 61 | ROCKSrocks-411 62 | SUNWopenssl-commands 63 | SUNWopenssl-libraries 64 | SUNWopensslr 65 | SUNWopenssl-man 66 | 67 | 68 | -------------------------------------------------------------------------------- /nodes/autofs-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | AutoFS Client 10 | 11 | 12 | 13 | 14 | Copyright (c) 2000 - 2017 The Regents of the University of California. 15 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 16 | 17 | 18 | 19 | 20 | $Log: autofs-client.xml,v $ 21 | Revision 1.8 2012/05/06 05:48:10 phil 22 | Copyright Storm for Mamba 23 | 24 | Revision 1.7 2011/07/23 02:30:14 phil 25 | Viper Copyright 26 | 27 | Revision 1.6 2010/09/07 23:52:46 bruno 28 | star power for gb 29 | 30 | Revision 1.5 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.4 2008/10/18 00:55:45 mjk 34 | copyright 5.1 35 | 36 | Revision 1.3 2008/03/06 23:41:30 mjk 37 | copyright storm on 38 | 39 | Revision 1.2 2007/06/23 04:03:18 mjk 40 | mars hill copyright 41 | 42 | Revision 1.1 2005/03/01 00:22:26 mjk 43 | moved to base roll 44 | 45 | Revision 1.2 2005/02/15 22:58:52 mjk 46 | dtd is gone 47 | 48 | Revision 1.1 2004/03/08 21:01:23 mjk 49 | *** empty log message *** 50 | 51 | Revision 1.4 2004/02/03 20:00:35 mjk 52 | decolorize 53 | 54 | Revision 1.3 2004/01/30 04:53:42 mjk 55 | colorize 56 | 57 | Revision 1.2 2001/10/02 03:16:39 mjk 58 | - Update to bruno's new auto-part stuff 59 | - Doesn't work but checkin anyway 60 | 61 | Revision 1.1 2001/09/05 00:27:16 mjk 62 | main and packages section is correct for compute nodes 63 | 64 | 65 | 66 | 67 | # we need to fix the Domain to get NFSv4 running 68 | # this regexp will match both initial idmapd.conf as well as following reconfigure 69 | sed -i 's/.*Domain.*=.*/Domain = &Kickstart_PublicHostname;/g' /etc/idmapd.conf 70 | 71 | 72 | 73 | 74 | 75 | -------------------------------------------------------------------------------- /nodes/debug-post.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | Stupid debugging tricks. To use you need to add an edge in your XML 9 | graph file to this node. By setting the weight of the edge you can 10 | control where it the final output this code gets placed. 11 | 12 | This trick is specific to the kickstart post section 13 | 14 | 15 | 16 | 17 | Copyright (c) 2000 - 2017 The Regents of the University of California. 18 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 19 | 20 | 21 | 22 | 23 | $Log: debug-post.xml,v $ 24 | Revision 1.8 2012/05/06 05:48:10 phil 25 | Copyright Storm for Mamba 26 | 27 | Revision 1.7 2011/07/23 02:30:14 phil 28 | Viper Copyright 29 | 30 | Revision 1.6 2010/09/07 23:52:46 bruno 31 | star power for gb 32 | 33 | Revision 1.5 2009/05/01 19:06:48 mjk 34 | chimi con queso 35 | 36 | Revision 1.4 2008/10/18 00:55:45 mjk 37 | copyright 5.1 38 | 39 | Revision 1.3 2008/03/06 23:41:30 mjk 40 | copyright storm on 41 | 42 | Revision 1.2 2007/06/23 04:03:18 mjk 43 | mars hill copyright 44 | 45 | Revision 1.1 2005/03/01 00:22:26 mjk 46 | moved to base roll 47 | 48 | Revision 1.2 2005/02/15 22:58:52 mjk 49 | dtd is gone 50 | 51 | Revision 1.1 2004/03/08 21:01:23 mjk 52 | *** empty log message *** 53 | 54 | Revision 1.5 2004/02/03 20:00:35 mjk 55 | decolorize 56 | 57 | Revision 1.4 2004/01/30 04:53:42 mjk 58 | colorize 59 | 60 | Revision 1.3 2001/11/05 21:21:39 bruno 61 | more fixes for 2.1.1 62 | 63 | Revision 1.2 2001/09/05 00:27:16 mjk 64 | main and packages section is correct for compute nodes 65 | 66 | 67 | 68 | 69 | touch /tmp/yo 70 | while [ -f /tmp/yo ]; do 71 | sleep 5 72 | done 73 | 74 | 75 | 76 | 77 | -------------------------------------------------------------------------------- /src/usersguide/clusterinstall/physical.sgml: -------------------------------------------------------------------------------- 1 |
2 | Physical Assembly 3 | 4 | 5 | The first thing to manage is the physical deployment of a cluster. Much 6 | research exists on the topic of how to physically construct a cluster. 7 | A majority of the O'Reilly 8 | Book Building Linux Clusters is devoted to 9 | the physical setup of a cluster, how to choose a motherboard, etc. 10 | Finally, the book How to Build a Beowulf also has 11 | some good tips on physical construction. 12 | 13 | 14 | 15 | We favor rack-mounted equipment because of its relative reliability and 16 | density. 17 | There are Rocks clusters, however, that are built from mini-towers. 18 | Choose what makes sense for you. 19 | 20 | 21 | 22 | The following diagram shows how the frontend and compute nodes must be 23 | connected: 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | On the compute nodes, the Ethernet interface that Linux maps to 34 | eth0 should be connected to the cluster's Ethernet 35 | switch. 36 | This network is considered private, that is, all 37 | traffic on this network is physically separated from the external public 38 | network (e.g., the internet). 39 | 40 | 41 | 42 | On the frontend, at least two ethernet interfaces are required. 43 | The interface that Linux maps to eth0 should be connected 44 | to the same ethernet network as the compute nodes. 45 | The interface that Linux maps to eth1 should be 46 | connected to the external network (e.g., the internet or your organization's 47 | intranet). 48 | 49 | 50 |
51 | 52 | -------------------------------------------------------------------------------- /nodes/iscsi.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | iSCSI 10 | 11 | 12 | 13 | 14 | Copyright (c) 2000 - 2017 The Regents of the University of California. 15 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 16 | 17 | 18 | 19 | 20 | $Log: iscsi.xml,v $ 21 | Revision 1.8 2012/05/06 05:48:10 phil 22 | Copyright Storm for Mamba 23 | 24 | Revision 1.7 2011/07/23 02:30:14 phil 25 | Viper Copyright 26 | 27 | Revision 1.6 2010/09/07 23:52:46 bruno 28 | star power for gb 29 | 30 | Revision 1.5 2009/05/01 19:06:48 mjk 31 | chimi con queso 32 | 33 | Revision 1.4 2008/10/18 00:55:45 mjk 34 | copyright 5.1 35 | 36 | Revision 1.3 2008/03/06 23:41:30 mjk 37 | copyright storm on 38 | 39 | Revision 1.2 2007/06/23 04:03:18 mjk 40 | mars hill copyright 41 | 42 | Revision 1.1 2005/03/01 00:22:26 mjk 43 | moved to base roll 44 | 45 | Revision 1.2 2005/02/15 22:58:52 mjk 46 | dtd is gone 47 | 48 | Revision 1.1 2004/03/08 21:01:23 mjk 49 | *** empty log message *** 50 | 51 | Revision 1.6 2004/02/03 20:00:35 mjk 52 | decolorize 53 | 54 | Revision 1.5 2004/01/30 04:53:42 mjk 55 | colorize 56 | 57 | Revision 1.4 2003/07/16 20:10:41 bruno 58 | changed to new 'file' tag 59 | 60 | Revision 1.3 2002/06/17 19:50:02 bruno 61 | 7.3-isms 62 | 63 | Revision 1.2 2002/06/11 20:20:28 mjk 64 | Added support for release tag 65 | 66 | Revision 1.1 2002/04/24 00:33:10 mjk 67 | iscsi 68 | 69 | 70 | 71 | 72 | 73 | 74 | 75 | 76 | # iSCSI configuration file 77 | 78 | DiscoveryAddress=10.10.81.81 79 | 80 | 81 | 82 | 83 | 84 | 85 | 86 | 87 | 88 | 89 | 90 | 91 | 92 | 93 | 94 | -------------------------------------------------------------------------------- /nodes/debug-pre.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | Stupid debugging tricks. To use you need to add an edge in your XML 8 | graph file to this node. By setting the weight of the edge you can 9 | control where it the final output this code gets placed. 10 | 11 | This trick is specific to the kickstart pre section 12 | 13 | 14 | 15 | 16 | Copyright (c) 2000 - 2017 The Regents of the University of California. 17 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 18 | 19 | 20 | 21 | 22 | $Log: debug-pre.xml,v $ 23 | Revision 1.8 2012/05/06 05:48:10 phil 24 | Copyright Storm for Mamba 25 | 26 | Revision 1.7 2011/07/23 02:30:14 phil 27 | Viper Copyright 28 | 29 | Revision 1.6 2010/09/07 23:52:46 bruno 30 | star power for gb 31 | 32 | Revision 1.5 2009/05/01 19:06:48 mjk 33 | chimi con queso 34 | 35 | Revision 1.4 2008/10/18 00:55:45 mjk 36 | copyright 5.1 37 | 38 | Revision 1.3 2008/03/06 23:41:30 mjk 39 | copyright storm on 40 | 41 | Revision 1.2 2007/06/23 04:03:18 mjk 42 | mars hill copyright 43 | 44 | Revision 1.1 2005/03/01 00:22:26 mjk 45 | moved to base roll 46 | 47 | Revision 1.2 2005/02/15 22:58:52 mjk 48 | dtd is gone 49 | 50 | Revision 1.1 2004/03/08 21:01:23 mjk 51 | *** empty log message *** 52 | 53 | Revision 1.5 2004/02/03 20:00:35 mjk 54 | decolorize 55 | 56 | Revision 1.4 2004/01/30 04:53:42 mjk 57 | colorize 58 | 59 | Revision 1.3 2001/10/01 19:10:37 bruno 60 | gb tweaks 61 | 62 | Revision 1.2 2001/09/05 00:27:16 mjk 63 | main and packages section is correct for compute nodes 64 | 65 | 66 | 67 |
68 | echo "yo" > /tmp/yo
69 | while [ -f /tmp/yo ]; do
70 | 	sleep 5
71 | done
72 | 
73 | 74 |
75 | 76 | 77 | 78 | 79 | 80 | 81 | 82 | 83 | -------------------------------------------------------------------------------- /nodes/emacs.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | Emacs OS 9 | 10 | 11 | 12 | 13 | Copyright (c) 2000 - 2017 The Regents of the University of California. 14 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 15 | 16 | 17 | 18 | 19 | $Log: emacs.xml,v $ 20 | Revision 1.9 2012/05/06 05:48:10 phil 21 | Copyright Storm for Mamba 22 | 23 | Revision 1.8 2012/04/24 18:43:51 phil 24 | add libotf package explicitly for emacs. 25 | 26 | Revision 1.7 2011/07/23 02:30:14 phil 27 | Viper Copyright 28 | 29 | Revision 1.6 2010/09/07 23:52:46 bruno 30 | star power for gb 31 | 32 | Revision 1.5 2009/05/01 19:06:48 mjk 33 | chimi con queso 34 | 35 | Revision 1.4 2008/10/18 00:55:45 mjk 36 | copyright 5.1 37 | 38 | Revision 1.3 2008/03/06 23:41:30 mjk 39 | copyright storm on 40 | 41 | Revision 1.2 2007/06/23 04:03:18 mjk 42 | mars hill copyright 43 | 44 | Revision 1.1 2005/03/01 00:22:26 mjk 45 | moved to base roll 46 | 47 | Revision 1.2 2005/02/15 22:58:52 mjk 48 | dtd is gone 49 | 50 | Revision 1.1 2004/03/08 21:01:23 mjk 51 | *** empty log message *** 52 | 53 | Revision 1.4 2004/02/03 20:00:35 mjk 54 | decolorize 55 | 56 | Revision 1.3 2004/01/30 04:53:42 mjk 57 | colorize 58 | 59 | Revision 1.2 2002/02/15 23:44:23 mjk 60 | - Added netscape to frontend 61 | - Move package trimming 62 | 63 | Revision 1.1 2001/11/06 18:12:40 mjk 64 | - Made emacs its own node (reuse in frontend and standalone) 65 | - removed Victor's NFS stuff from accounting 66 | - public/private root passwords match on the form 67 | - Fixed private kickstart host on the form 68 | 69 | 70 | 71 | libotf 72 | Emacs 73 | 74 | 75 | 76 | -------------------------------------------------------------------------------- /src/screens/palette-blue.css: -------------------------------------------------------------------------------- 1 | 2 | /* background color for the frames */ 3 | .ProBackground { 4 | background-color: #334455; 5 | } 6 | 7 | /* frame title font */ 8 | .ProTitle { 9 | color: #F4C600; 10 | text-align: center; 11 | } 12 | 13 | /* label in "workarea" and "help" */ 14 | .ProLabel { 15 | color: #CECECE; 16 | font-size: 11pt; 17 | font-weight: bold; 18 | } 19 | 20 | /* form input in "workarea" */ 21 | .ProInput { 22 | color: #112233; 23 | background-color: #b5c9e2; 24 | } 25 | 26 | /* form input in "workarea" when the wrong value is entered */ 27 | .ProInputError { 28 | color: #112233; 29 | background-color: red; 30 | } 31 | 32 | 33 | /* help text in "help" */ 34 | .ProHelpText { 35 | color: #BBCCDD; 36 | font-size: 10pt; 37 | } 38 | 39 | /* status message text in "status" */ 40 | .ProStatusText { 41 | color: #BBCCDD; 42 | font-size: 10pt; 43 | } 44 | 45 | /* button style in "workarea" */ 46 | .ProButton { 47 | color: #112233; 48 | background-color: #b5c9e2; 49 | font-size: 10pt; 50 | font-weight: bold; 51 | text-align: center; 52 | margin-top: 10px; 53 | margin-bottom: 10px; 54 | margin-left: 50px; 55 | margin-right: 20px; 56 | border-top: 2px #cce3ff solid; 57 | border-left: 2px #cce3ff solid; 58 | border-bottom: 2px #223344 solid; 59 | border-right: 2px #223344 solid; 60 | } 61 | 62 | /* button with mouse over in "workarea" */ 63 | .ProButtonHover { 64 | color: #112233; 65 | background-color: #E2Ba03; 66 | border-top: 2px #223344 solid; 67 | border-left: 2px #223344 solid; 68 | border-bottom: 2px #cce3ff solid; 69 | border-right: 2px #cce3ff solid; 70 | } 71 | 72 | /* tiled background */ 73 | .TiledBanner { 74 | text-align: center; 75 | line-height: 90px; 76 | background-image: url(bg-blue.png); 77 | background-repeat: repeat-x; 78 | background-position: 0px 0px; 79 | font-size: 16pt; 80 | font-weight: bold; 81 | } 82 | -------------------------------------------------------------------------------- /src/screens/palette-purple.css: -------------------------------------------------------------------------------- 1 | 2 | /* background color for the frames */ 3 | .ProBackground { 4 | background-color: #9F959E; 5 | } 6 | 7 | /* frame title font */ 8 | .ProTitle { 9 | color: #47003E; 10 | text-align: center; 11 | } 12 | 13 | /* label in "workarea" and "help" */ 14 | .ProLabel { 15 | color: #F3F4DC; 16 | font-size: 11pt; 17 | font-weight: bold; 18 | } 19 | 20 | /* form input in "workarea" */ 21 | .ProInput { 22 | color: #380031; 23 | background-color: #F3F4DC; 24 | } 25 | 26 | /* form input in "workarea" when the wrong value is entered */ 27 | .ProInputError { 28 | color: red; 29 | background-color: #b5c9e2; 30 | } 31 | 32 | /* help text in "help" */ 33 | .ProHelpText { 34 | color: #380031; 35 | font-size: 10pt; 36 | } 37 | 38 | /* status message text in "status" */ 39 | .ProStatusText { 40 | color: #380031; 41 | font-size: 10pt; 42 | } 43 | 44 | /* button style in "workarea" */ 45 | .ProButton { 46 | color: #F3F4DC; 47 | background-color: #47003E; 48 | font-size: 10pt; 49 | font-weight: bold; 50 | text-align: center; 51 | margin-top: 10px; 52 | margin-bottom: 10px; 53 | margin-left: 50px; 54 | margin-right: 20px; 55 | border-top: 2px #DDCAC9 solid; 56 | border-left: 2px #DDCAC9 solid; 57 | border-bottom: 2px #130013 solid; 58 | border-right: 2px #130013 solid; 59 | } 60 | 61 | /* button with mouse over in "workarea" */ 62 | .ProButtonHover { 63 | color: #130013; 64 | background-color: #F4C600; 65 | border-top: 2px #130013 solid; 66 | border-left: 2px #130013 solid; 67 | border-right: 2px #DDCAC9 solid; 68 | border-bottom: 2px #DDCAC9 solid; 69 | } 70 | 71 | /* tiled background */ 72 | .TiledBanner { 73 | text-align: center; 74 | line-height: 90px; 75 | background-image: url(bg-purple.png); 76 | background-repeat: repeat-x; 77 | background-position: 0px 0px; 78 | font-size: 16pt; 79 | font-weight: bold; 80 | } 81 | -------------------------------------------------------------------------------- /src/screens/palette-yellow.css: -------------------------------------------------------------------------------- 1 | 2 | /* background color for the frames */ 3 | .ProBackground { 4 | background-color: #FBF7B0; 5 | } 6 | 7 | /* frame title font */ 8 | .ProTitle { 9 | color: #A13032; 10 | text-align: center; 11 | } 12 | 13 | /* label in "workarea" and "help" */ 14 | .ProLabel { 15 | color: #5D3333; 16 | font-size: 11pt; 17 | font-weight: bold; 18 | } 19 | 20 | /* form input in "workarea" */ 21 | .ProInput { 22 | color: #000000; 23 | background-color: #FCE39D; 24 | } 25 | 26 | /* form input in "workarea" when the wrong value is entered */ 27 | .ProInputError { 28 | color: red; 29 | background-color: #b5c9e2; 30 | } 31 | 32 | /* help text in "help" */ 33 | .ProHelpText { 34 | color: #333333; 35 | font-size: 10pt; 36 | } 37 | 38 | /* status message text in "status" */ 39 | .ProStatusText { 40 | color: #333333; 41 | font-size: 10pt; 42 | } 43 | 44 | /* button style in "workarea" */ 45 | .ProButton { 46 | color: #FCE39D; 47 | background-color: #A13032; 48 | font-size: 10pt; 49 | font-weight: bold; 50 | text-align: center; 51 | margin-top: 10px; 52 | margin-bottom: 10px; 53 | margin-left: 50px; 54 | margin-right: 20px; 55 | border-top: 2px #FCE39D solid; 56 | border-left: 2px #FCE39D solid; 57 | border-bottom: 2px #5D3333 solid; 58 | border-right: 2px #5D3333 solid; 59 | } 60 | 61 | /* button with mouse over in "workarea" */ 62 | .ProButtonHover { 63 | color: #5D3333; 64 | background-color: #EBB144; 65 | border-top: 2px #5D3333 solid; 66 | border-left: 2px #5D3333 solid; 67 | border-right: 2px #FCE390 solid; 68 | border-bottom: 2px #FCE390 solid; 69 | } 70 | 71 | /* tiled background */ 72 | .TiledBanner { 73 | text-align: center; 74 | line-height: 90px; 75 | background-image: url(bg-yellow.png); 76 | background-repeat: repeat-x; 77 | background-position: 0px 0px; 78 | font-size: 16pt; 79 | font-weight: bold; 80 | } 81 | -------------------------------------------------------------------------------- /nodes/snmp-client.xml: -------------------------------------------------------------------------------- 1 | 2 | 3 | 4 | 5 | 6 | 7 | SNMP Client machine. 8 | 9 | 10 | 11 | 12 | Copyright (c) 2000 - 2017 The Regents of the University of California. 13 | All rights reserved. Rocks(r) v6.2/v7.0 www.rocksclusters.org 14 | 15 | 16 | 17 | 18 | $Log: snmp-client.xml,v $ 19 | Revision 1.8 2012/05/06 05:48:10 phil 20 | Copyright Storm for Mamba 21 | 22 | Revision 1.7 2011/07/23 02:30:15 phil 23 | Viper Copyright 24 | 25 | Revision 1.6 2010/09/07 23:52:47 bruno 26 | star power for gb 27 | 28 | Revision 1.5 2009/05/01 19:06:48 mjk 29 | chimi con queso 30 | 31 | Revision 1.4 2008/10/18 00:55:46 mjk 32 | copyright 5.1 33 | 34 | Revision 1.3 2008/03/06 23:41:30 mjk 35 | copyright storm on 36 | 37 | Revision 1.2 2007/06/23 04:03:18 mjk 38 | mars hill copyright 39 | 40 | Revision 1.1 2005/03/04 01:40:04 mjk 41 | massive graph cleanup 42 | 43 | Revision 1.6 2005/02/16 00:40:04 mjk 44 | graph cleanup 45 | 46 | Revision 1.5 2005/02/01 23:37:10 mjk 47 | remove roll attrs 48 | 49 | Revision 1.4 2004/02/03 00:57:25 mjk 50 | decolorize / derollize 51 | 52 | Revision 1.3 2004/01/30 00:22:54 mjk 53 | color hpc green 54 | 55 | Revision 1.2 2003/09/26 00:07:24 bruno 56 | closer to 9 57 | 58 | Revision 1.1 2003/07/07 20:47:52 bruno 59 | initial release 60 | 61 | Revision 1.4 2002/10/09 23:16:27 bruno 62 | fixup of package names to match our new source tree 63 | 64 | Revision 1.3 2001/10/02 03:16:40 mjk 65 | - Update to bruno's new auto-part stuff 66 | - Doesn't work but checkin anyway 67 | 68 | Revision 1.2 2001/09/05 00:27:18 mjk 69 | main and packages section is correct for compute nodes 70 | 71 | 72 | 73 | net-snmp-utils 74 | rocks-snmp-status 75 | 76 | 77 | -------------------------------------------------------------------------------- /src/anaconda-yum-plugins/anaconda-yum-plugins.spec.in: -------------------------------------------------------------------------------- 1 | Summary: Installation-related yum plugins 2 | Name: anaconda-yum-plugins 3 | Epoch: 1 4 | Version: 1.0 5 | Release: 5.1%{?dist} 6 | License: GPLv2+ 7 | Group: Applications/System 8 | URL: http://fedoraproject.org/wiki/Anaconda 9 | 10 | Source0: %{name}-%{version}.tar.bz2 11 | 12 | BuildRoot: %{_tmppath}/%{name}-%{version}-%{release}-root-%(%{__id_u} -n) 13 | BuildArch: noarch 14 | Requires: python, yum 15 | 16 | %description 17 | The anaconda yum-plugins package contains yum plugins that are useful for 18 | anaconda and other system installation-related programs. 19 | 20 | %prep 21 | %setup -q 22 | 23 | %build 24 | # noop 25 | 26 | %install 27 | %{__rm} -rf %{buildroot} 28 | # RPM will take care of the python-compiling stuff 29 | %{__make} install DESTDIR=%{buildroot} NO_PY_COMPILE=1 30 | 31 | %clean 32 | %{__rm} -rf %{buildroot} 33 | 34 | %files 35 | %defattr(-,root,root,-) 36 | %doc COPYING 37 | %config(noreplace) %{_sysconfdir}/yum/pluginconf.d/* 38 | %{_prefix}/lib/yum-plugins/* 39 | 40 | %changelog 41 | * Mon Nov 30 2009 Dennis Gregorovic - 1:1.0-5.1 42 | - Rebuilt for RHEL 6 43 | 44 | * Fri Jul 24 2009 Fedora Release Engineering - 1:1.0-5 45 | - Rebuilt for https://fedoraproject.org/wiki/Fedora_12_Mass_Rebuild 46 | 47 | * Mon Feb 23 2009 Fedora Release Engineering - 1:1.0-4 48 | - Rebuilt for https://fedoraproject.org/wiki/Fedora_11_Mass_Rebuild 49 | 50 | * Fri Sep 19 2008 David Cantrell 1:1.0-3 51 | - Need epoch increased since version is back to 1.0, but it was previously 52 | the same as the anaconda version number. jkeating and wwoods said to do 53 | this too, so blame them. 54 | 55 | * Thu Sep 18 2008 Chris Lumens 1.0-2 56 | - Include the distro tag in the release number. 57 | 58 | * Mon Sep 15 2008 Will Woods - 1.0-1 59 | - Initial packaging (moved out of anaconda package) 60 | -------------------------------------------------------------------------------- /tests/t2009-ks-file-tag.sh: -------------------------------------------------------------------------------- 1 | #!/bin/bash 2 | # 3 | # Test kickstart.cgi script 4 | # 5 | 6 | test_description='Test kickstart syntax and generation 7 | 8 | ' 9 | 10 | pushd `dirname $0` > /dev/null 11 | export TEST_DIRECTORY=`pwd` 12 | popd > /dev/null 13 | . $TEST_DIRECTORY/test-lib.sh 14 | 15 | hostname=`hostname` 16 | 17 | rocks_build=/export/rocks/install/rocks-dist/x86_64/build 18 | 19 | test_expect_success 'test rocks KS syntax - set up' ' 20 | cp $TEST_DIRECTORY/t2009/test.xml $rocks_build/graphs/default/test.xml && 21 | cp $TEST_DIRECTORY/t2009/test1.xml $rocks_build/nodes/test1.xml && 22 | cp $TEST_DIRECTORY/t2009/test2.xml $rocks_build/nodes/test2.xml && 23 | cp $TEST_DIRECTORY/t2009/test3.xml $rocks_build/nodes/test3.xml && 24 | rocks list node xml test1 25 | ' 26 | 27 | test_expect_success 'test rocks KS syntax - file tag' ' 28 | rocks list node xml test1 |rocks report post | bash && 29 | echo test1.xml && 30 | grep ciao /tmp/test1file && 31 | grep localhost /tmp/test1file && 32 | touch /tmp/RCS/test1file,v && 33 | ! grep ciao /tmp/test2file && 34 | grep localhost /tmp/test1file && 35 | echo test2.xml && 36 | grep local /tmp/test2pipe && 37 | grep local /tmp/test2pipesamefile && 38 | echo test3.xml && 39 | touch /tmp/RCS/test2pipe,v && 40 | test ! -f /tmp/RCS/test3simplefile,v && 41 | grep simple /tmp/test3simplefile && 42 | grep Rocks /tmp/test3-release 43 | ' 44 | 45 | 46 | test_expect_success 'test rocks KS syntax - tear down' ' 47 | rm $rocks_build/graphs/default/test.xml && 48 | rm $rocks_build/nodes/test1.xml && 49 | rm $rocks_build/nodes/test2.xml && 50 | rm $rocks_build/nodes/test3.xml && 51 | rm -rf /tmp/RCS/ && 52 | rm /tmp/test1file && 53 | rm /tmp/test2file && 54 | rm /tmp/test2pipe && 55 | rm /tmp/test2pipesamefile && 56 | rm /tmp/test3simplefile && 57 | rm /tmp/test3-release 58 | ' 59 | 60 | test_done 61 | 62 | --------------------------------------------------------------------------------