summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--debian/ceph.NEWS180
1 files changed, 0 insertions, 180 deletions
diff --git a/debian/ceph.NEWS b/debian/ceph.NEWS
deleted file mode 100644
index ee9db2f22..000000000
--- a/debian/ceph.NEWS
+++ /dev/null
@@ -1,180 +0,0 @@
-ceph (10.2.5-1) unstable; urgency=medium
-
- ## Upgrades from Debian Jessie
-
- Online upgrades from Ceph versions prior to Hammer (0.94.x) are not
- supported by upstream. As Debian Jessie has Ceph Firefly (0.80.x) an
- online upgrade from Jessie to Stretch is not possible. You have to first
- shutdown all Ceph daemons on all nodes, upgrade everything to the new
- version and start all daemons again.
-
- Ceph daemons are not automatically restarted on upgrade to minimize
- disruption. You have to manually restart them after the upgrade.
-
- -- Gaudenz Steinlin <gaudenz@debian.org> Sun, 08 Jan 2017 14:57:35 +0100
-
-ceph (9.2.0-1) experimental; urgency=medium
-
- ## systemd Enablement
-
- For all distributions that support systemd (Debian Jessie 8.x,
- Ubuntu >= 16.04), Ceph daemons are now managed using upstream provided
- systemd files instead of the legacy sysvinit scripts or distro provided
- systemd files. For example:
-
- systemctl start ceph.target # start all daemons
- systemctl status ceph-osd@12 # check status of osd.12
-
- To upgrade existing deployments that use the older systemd service
- configurations (Ubuntu >= 15.04, Debian >= Jessie), you need to switch
- to using the new ceph-mon@ service:
-
- systemctl stop ceph-mon
- systemctl disable ceph-mon
-
- systemctl start ceph-mon@`hostname`
- systemctl enable ceph-mon@`hostname`
-
- and also enable the ceph target post upgrade:
-
- systemctl enable ceph.target
-
- The main notable distro that is *not* using systemd is Ubuntu 14.04
- (The next Ubuntu LTS, 16.04, will use systemd instead of upstart).
-
- ## Ceph daemons no longer run as root
-
- Ceph daemons now run as user and group 'ceph' by default. The
- ceph user has a static UID assigned by Debian to ensure consistency
- across servers within a Ceph deployment.
-
- If your systems already have a ceph user, upgrading the package will cause
- problems. We suggest you first remove or rename the existing 'ceph' user
- and 'ceph' group before upgrading.
-
- When upgrading, administrators have two options:
-
- 1. Add the following line to 'ceph.conf' on all hosts:
-
- setuser match path = /var/lib/ceph/$type/$cluster-$id
-
- This will make the Ceph daemons run as root (i.e., not drop
- privileges and switch to user ceph) if the daemon's data
- directory is still owned by root. Newly deployed daemons will
- be created with data owned by user ceph and will run with
- reduced privileges, but upgraded daemons will continue to run as
- root.
-
- 2. Fix the data ownership during the upgrade. This is the
- preferred option, but it is more work and can be very time
- consuming. The process for each host is to:
-
- 1. Upgrade the ceph package. This creates the ceph user and group. For
- example:
-
- apt-get install ceph
-
- NOTE: the permissions on /var/lib/ceph/mon will be set to ceph:ceph
- as part of the package upgrade process on existing *systemd*
- based installations; the ceph-mon systemd service will be
- automatically restarted as part of the upgrade. All other
- filesystem permissions on systemd based installs will
- remain unmodified by the upgrade.
-
- 2. Stop the daemon(s):
-
- systemctl stop ceph-osd@* # debian, ubuntu >= 15.04
- stop ceph-all # ubuntu 14.04
-
- 3. Fix the ownership:
-
- chown -R ceph:ceph /var/lib/ceph
-
- 4. Restart the daemon(s):
-
- start ceph-all # ubuntu 14.04
- systemctl start ceph.target # debian, ubuntu >= 15.04
-
- Alternatively, the same process can be done with a single daemon
- type, for example by stopping only monitors and chowning only
- '/var/lib/ceph/osd'.
-
- ## KeyValueStore OSD on-disk format changes
-
- The on-disk format for the experimental KeyValueStore OSD backend has
- changed. You will need to remove any OSDs using that backend before you
- upgrade any test clusters that use it.
-
- ## Deprecated commands
-
- 'ceph scrub', 'ceph compact' and 'ceph sync force' are now DEPRECATED.
- Users should instead use 'ceph mon scrub', 'ceph mon compact' and
- 'ceph mon sync force'.
-
- ## Full pool behaviour
-
- When a pool quota is reached, librados operations now block indefinitely,
- the same way they do when the cluster fills up. (Previously they would
- return -ENOSPC). By default, a full cluster or pool will now block. If
- your librados application can handle ENOSPC or EDQUOT errors gracefully,
- you can get error returns instead by using the new librados
- OPERATION_FULL_TRY flag.
-
- -- James Page <james.page@ubuntu.com> Mon, 30 Nov 2015 09:23:09 +0000
-
-ceph (0.80.9-2) unstable; urgency=medium
-
- ## CRUSH fixes in 0.80.9
-
- The 0.80.9 point release fixes several issues with CRUSH that trigger excessive
- data migration when adjusting OSD weights. These are most obvious when a very
- small weight change (e.g., a change from 0 to .01) triggers a large amount of
- movement, but the same set of bugs can also lead to excessive (though less
- noticeable) movement in other cases.
-
- However, because the bug may already have affected your cluster, fixing it
- may trigger movement back to the more correct location. For this reason, you
- must manually opt-in to the fixed behavior.
-
- In order to set the new tunable to correct the behavior:
-
- ceph osd crush set-tunable straw_calc_version 1
-
- Note that this change will have no immediate effect. However, from this
- point forward, any ‘straw’ bucket in your CRUSH map that is adjusted will get
- non-buggy internal weights, and that transition may trigger some rebalancing.
-
- You can estimate how much rebalancing will eventually be necessary on your
- cluster with:
-
- ceph osd getcrushmap -o /tmp/cm
- crushtool -i /tmp/cm --num-rep 3 --test --show-mappings > /tmp/a 2>&1
- crushtool -i /tmp/cm --set-straw-calc-version 1 -o /tmp/cm2
- crushtool -i /tmp/cm2 --reweight -o /tmp/cm2
- crushtool -i /tmp/cm2 --num-rep 3 --test --show-mappings > /tmp/b 2>&1
- wc -l /tmp/a # num total mappings
- diff -u /tmp/a /tmp/b | grep -c ^+ # num changed mappings
-
- Divide the total number of lines in /tmp/a with the number of lines
- changed. We've found that most clusters are under 10%.
-
- You can force all of this rebalancing to happen at once with:
-
- ceph osd crush reweight-all
-
- Otherwise, it will happen at some unknown point in the future when
- CRUSH weights are next adjusted.
-
- ## Mapping rbd devices with rbdmap on systemd systems
-
- If you have setup rbd mappings in /etc/ceph/rbdmap and corresponding mounts
- in /etc/fstab things might break with systemd because systemd waits for the
- rbd device to appear before the legacy rbdmap init file has a chance to run
- and drops into emergency mode if it times out.
-
- This can be fixed by adding the nofail option in /etc/fstab to all rbd
- backed mount points. With this systemd does not wait for the device and
- proceeds with the boot process. After rbdmap mapped the device, systemd
- detects the new device and mounts the file system.
-
- -- Gaudenz Steinlin <gaudenz@debian.org> Mon, 04 May 2015 22:49:48 +0200