- This is 2.1 final release, we made a lot of tests, especially on importing cluster and infernalis supporting.
- 8 new features are added, and 27 bugs are fixed in this release.
- Spotlights:
- Import existing cluster (see "Cluster Management\Import Cluster"), user could import existing cluster, and monitor its status, or manipulate it by adding/removing servers or devices.
- Manage storage group (see "Cluster Management\Manage Storage group"), user could see the crushmap hierarchy, and understand storage group coverage on the topology.
- Manage zone (see "Cluster Management\Manage zone"), user could add new zones then associates servers to those new zones.
- Key Summary
- VSM-53 Allow VSM to attach to and manage an existing Ceph cluster
- VSM-185 show SMART information for storage devices
- VSM-409 add "storage group" column on OSD list page
- VSM-418 Add support to Ceph infernalis
- VSM-385 Change PG count even after ceph cluster deployed
- VSM-337 Upgrade: expect to upgrade from 1.0 to 2.0
- VSM-407 suggest to split disk status and capacity utilization in different column in device list page
- VSM-396 Support DNS lookup beside current /etc/hosts lookup
- Key Summary
- VSM-427 when updating storage group, it's required to add a hour glass.
- VSM-428 if storage group is referenced by some pools, it still allows to update its marker.
- VSM-423 when creating new pool, after spinner is disappeared, the pool is not showing up
- VSM-419 Stop Server error in 2.1 beta
- VSM-172 cluster.manifest - request for improvement from Marcin
- VSM-261 when executing vsm-controller, a TypeError raises
- VSM-330 When upgrading from Firefly to Giant, the upgrade complains missing python-rados package
- VSM-344 after a few times install, clean-up, reinstall, we have seem some "eaddrinuse" errors on starting rabbitmq
- VSM-354 when 'create cluster' page opened,other pages still can be opened
- VSM-371 When installing with dependent packages pre-prepared, the installer will stop with complains
- VSM-395 Remove then restore an OSD, the OSD can't hit to in-up state again.
- VSM-422 stop/start servers should display an "hour glass" during operation
- VSM-425 Dashboard Monitor widget: boxes too small for hostname font
- VSM-426 blocked on process indicator
- VSM-388 VSM fails to create a Monitor daemon (after clicking OK in a popup warning window)
- VSM-429 VSM agent fail to save ceph.conf when size is above 127 KB.
- VSM-225 VSM Creates Very Small pg_num and pg._num Size for EC Pool
- VSM-420 VSM 2.0 is_server_manifest_error
- VSM-421 During cluster import, vsm agents don't have sufficient rights to /var/lib/ceph
- VSM-380 it would be nice if install.sh script should take care of agent-tokens, even at the case restart install.sh with prior errors.
- VSM-406 servers can't be stopped after cluster imported
- VSM-410 The scrollbar of manage server is missing now.
- VSM-413 Manage device, click new, popup “internal ERROR” on the top right corner.
- VSM-416 at import cluster page, the scroll bar for crushmap and ceph.conf can scroll
- VSM-417 when creating cluster, the monitor status is incorrect
- VSM-387 ceph version conflicts when reinstall vsm at the case the previous one made ceph upgrade
- VSM-377 VSM installation failure
- This is 2.1 beta 1 release, only very limited tests on it.
- 27 new features are added, and 51 bugs are fixed in this release.
- VSM-66 VSM interoperates with Ceph update
- VSM-226 Need Documentation for VSM REST API
- VSM-78 Integration with Cinder capacity filter
- VSM-58 Calamari Integration
- VSM-56 Report total RBD Image commit
- VSM-32 support to define storage group with storage node or rack as unit
- VSM-376 support to label device with by-uuid
- VSM-382 undeploy/redeploy ceph cluster
- VSM-389 Need to automatically retrieve the osd info from existing cluster.
- VSM-386 batch add osds
- VSM-355 there are some used disk paths list in the data path field and journal dik path field on page 'add new osd'
- VSM-90 Monitor Status page improvements.
- VSM-96 UI responsiveness
- VSM-98 Server Configuration Change.
- VSM-341 expect to have some utilties to help make automation tests.
- VSM-352 servermgmt page autorefrush too frequent
- VSM-372 narrow the attack surface from VSM to Openstack cluster
- VSM-373 at adding new OSD page, expect to list device name like /dev/sdd1 instead of pci bus address.
- VSM-140 Ceph Development (Epic 6): Prototype Calamari/VSM dashboard implementation
- VSM-88 On monitor status page, report what server each monitor is running on.
- VSM-242 Allow user to modify ceph.conf outside VSM
- VSM-15 VSM-backup prompt info not correct
- VSM-124 [CDEK-1852] VSM | adding possibility to manipulate ceph values in cluster.manifest file
- VSM-4 Average Response Time" missing in dashboard Overview panel "VSM Status" section.
- VSM-184 add automated script to help deploy VSM on multiple nodes
- VSM-159 add issue reporting tool
- VSM-156 add sanity check tool to help identify potential issues before or after deployment
- VSM-349 click 'create cluster',I got the tip:there are some zones no monitor created
- VSM-411 When create cluster, there are four servers and choose all servers as storage node but only three as monitors, the cluster can not be created sucessfully.
- VSM-329 Remove Monitors button in Manage Servers hangs when Monitor node also has MDS daemon
- VSM-400 the UI of all server operator pages will appear "loading" without any operation
- VSM-356 I got warning info means that the number of pg in each osd is too large after upgrade ceph from a lower version to hammar
- VSM-397 mysqld takes up 100% CPU on one core and cause VSM dashboard to become unusable
- VSM-412 After remove server or remove monitor, failed to add the monitor back.
- VSM-391 the "ceph df" number is not consistent with pool quota
- VSM-399 the UI messy of the manage servers page
- VSM-402 after stop server, then start server, the osd tree changes
- VSM-392 Have removed the volume from the openstack, but from the vsm rbd status page, the rbd list still include the volume
- VSM-384 stuck at restart all ceph servers after stopped them all from UI
- VSM-394 present more than one pool to openstack cinder, it always creates volumes on a pool
- VSM-321 no upstart mechanism used for ubuntu when controlling ceph service
- VSM-336 On Dashboard, even no cluster is created, the VSM version and uptime should be displayed
- VSM-24 [CDEK-1661] VSM Dashboard | Manage Servers | Reset server status - works not correctly.
- VSM-365 Creating Cluster stucks at ceph.conf creation when running VSM on CentOS 7.1
- VSM-19 CDEK-1613] VSM | Reset Server Status button - return Error:Network error
- VSM-379 Trace back in browser when using reset server status action buttons
- VSM-381 run diamond through service instead of current process launching
- VSM-378 Performance data is retrieved from outside nodes
- VSM-374 the down server is not reflected in VSM
- VSM-375 Malformed JSON in 'Integrate Cluster' function
- VSM-366 the password for openstack access is shown as plain text
- VSM-312 vsm-node sets node status="Need more IP" if a Monitor only node does not have a cluster IP address..
- VSM-367 can't create cluster at public cloud environment
- VSM-368 The default password is not following the same password policy to include uppercase and digitals.
- VSM-369 Change password: "!" doesnt' support in password even prompt message says OK
- VSM-244 Internal server error when installing v1.1
- VSM-224 Controller node error in /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-239 with automatic deployment, the execution is blocked at asking if start mysql service
- VSM-193 hard-coded cluster id
- VSM-179 keep ceph.conf up to date when executing "remove server" operations.
- VSM-176 SSL certificate password is stored in a plain text file
- VSM-177 wrong /etc/fstab entry for osd device mount point
- VSM-166 cluster_manifest sanity check program gives incorrect advice for auth_keys
- VSM-171 [CDEK1672] VSM_CLI | list shows Admin network in Public IP section
- VSM-168 [CDEK1800] VSM_CLI | remove mds - doesn't update vsm database
- VSM-121 Storage node unable to connect to controller although network is OK and all setting correct
- VSM-123 Storage node will not be able to contact controller node to install if http proxy set
- VSM-260 the check_network in server_manifest will be wrong when it has a single network card
- VSM-236 no way to check manifest correctness after editing them
- VSM-233 console blocks when running automatic installation procedure
- VSM-33 negative update time in RBD list
- VSM-216 Add storage group requires at least 3 nodes
- VSM-113 [CDEK-1835] VSM | /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-51 Install Fails for VSM 0.8.0 Engineering Build Release
- VSM-29 vsm-agent process causes high i/o on os disk
- VSM-230 when presenting pool to openstack, cache tiering pools should be listed.
- VSM-207 can't assume eth0 device name
- VSM-26 [CDEK-1664] VSM | Not possible to replace node if ceph contain only 3 nodes.
- This is 2.1 alpha 1 release, only very limited tests on it.
- Some spotlights:
- Manage existing cluster through "Cluster Management\Import Cluster"
0. applicable to 2.1 alpha 1.
- deploy a ceph cluster external (maybe there is already one running)
- deploy VSM with agents deployed on each ceph node with appropriate role defined.
- in VSM web ui:
- navigate to "Cluster Management\Import Cluster"
- you should see all ceph nodes listed.
- click "import cluster", then we will see an new page where it asks for crushmap and ceph.conf
- there are two approaches to get crushmap:
- paste crushmap directly into the text box under "Crushmap"
- click "Auto Detect", then a dialog will pop up, you'd select one monitor host, and point where is the monitor keyring file on the monitor host.
- for ceph.conf, we see some ceph.conf doesn't include osd information, if this is the case, you'd manually add them into ceph.conf.
- when crushmap and ceph.conf are ready, click "validate" to see if all can pass. If yes, you could click "submit", if succeed, the message tip will show up to say the import is successful.
- wait for a few minutes before the status gets synced, then you will see the cluster status.
- Crushmap management 0. only draft UI, not ready.
- Manage existing cluster through "Cluster Management\Import Cluster"
0. applicable to 2.1 alpha 1.
- 16 new features are added, and 39 bugs are fixed till this release.
- VSM-355 there are some used disk paths list in the data path field and journal dik path field on page 'add new osd'
- VSM-90 Monitor Status page improvements.
- VSM-96 UI responsiveness
- VSM-98 Server Configuration Change.
- VSM-352 servermgmt page autorefrush too frequent
- VSM-372 narrow the attack surface from VSM to Openstack cluster
- VSM-373 at adding new OSD page, expect to list device name like /dev/sdd1 instead of pci bus address.
- VSM-140 Ceph Development (Epic 6): Prototype Calamari/VSM dashboard implementation
- VSM-88 On monitor status page, report what server each monitor is running on.
- VSM-15 VSM-backup prompt info not correct
- VSM-242 Allow user to modify ceph.conf outside VSM
- VSM-124 [CDEK-1852] VSM | adding possibility to manipulate ceph values in cluster.manifest file
- VSM-4 Average Response Time" missing in dashboard Overview panel "VSM Status" section.
- VSM-159 add issue reporting tool
- VSM-184 add automated script to help deploy VSM on multiple nodes
- VSM-156 add sanity check tool to help identify potential issues before or after deployment
- Key Summary
- VSM-321 no upstart mechanism used for ubuntu when controlling ceph service
- VSM-336 On Dashboard, even no cluster is created, the VSM version and uptime should be displayed
- VSM-24 [CDEK-1661] VSM Dashboard | Manage Servers | Reset server status - works not correctly.
- VSM-365 Creating Cluster stucks at ceph.conf creation when running VSM on CentOS 7.1
- VSM-19 CDEK-1613] VSM | Reset Server Status button - return Error:Network error
- VSM-379 Trace back in browser when using reset server status action buttons
- VSM-381 run diamond through service instead of current process launching
- VSM-378 Performance data is retrieved from outside nodes
- VSM-374 the down server is not reflected in VSM
- VSM-375 Malformed JSON in 'Integrate Cluster' function
- VSM-366 the password for openstack access is shown as plain text
- VSM-312 vsm-node sets node status="Need more IP" if a Monitor only node does not have a cluster IP address..
- VSM-367 can't create cluster at public cloud environment
- VSM-368 The default password is not following the same password policy to include uppercase and digitals.
- VSM-369 Change password: "!" doesnt' support in password even prompt message says OK
- VSM-224 Controller node error in /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-244 Internal server error when installing v1.1
- VSM-179 keep ceph.conf up to date when executing "remove server" operations.
- VSM-176 SSL certificate password is stored in a plain text file
- VSM-193 hard-coded cluster id
- VSM-177 wrong /etc/fstab entry for osd device mount point
- VSM-239 with automatic deployment, the execution is blocked at asking if start mysql service
- VSM-166 cluster_manifest sanity check program gives incorrect advice for auth_keys
- VSM-121 Storage node unable to connect to controller although network is OK and all setting correct
- VSM-123 Storage node will not be able to contact controller node to install if http proxy set
- VSM-171 [CDEK1672] VSM_CLI | list shows Admin network in Public IP section
- VSM-168 [CDEK1800] VSM_CLI | remove mds - doesn't update vsm database
- VSM-236 no way to check manifest correctness after editing them
- VSM-233 console blocks when running automatic installation procedure
- VSM-260 the check_network in server_manifest will be wrong when it has a single network card
- VSM-33 negative update time in RBD list
- VSM-51 Install Fails for VSM 0.8.0 Engineering Build Release
- VSM-113 [CDEK-1835] VSM | /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-29 vsm-agent process causes high i/o on os disk
- VSM-216 Add storage group requires at least 3 nodes
- VSM-26 [CDEK-1664] VSM | Not possible to replace node if ceph contain only 3 nodes.
- VSM-207 can't assume eth0 device name
- VSM-230 when presenting pool to openstack, cache tiering pools should be listed.
- VSM-393 When importing cluster, if there is no OSd and MON sections, the import will fail
- This is the 2.0 final release, which includes both Ubuntu 14 and CentOS 7 packages.
- 16 new features are added, and 33 bugs are fixed till this release.
- VSM-342 support ceph upgrade on centos 7
- VSM-88 On monitor status page, report what server each monitor is running on.
- VSM-162 need to support typical openstack deployment with multiple cinder/nova/glance instances beside currrent all-in-one mode
- VSM-246 a separate tool to help make a node ready for addition into cluster
- VSM-307 the operator expects to upgrade ceph cluster
- VSM-309 Disk identification: the operator expects to identify physical location of failed disks for replacement
- VSM-327 Dashboard: it's required to have summary for device capacity status to show if full
- VSM-358 Support normal user instead of root to install vsm
- VSM-355 there are some used disk paths list in the data path field and journal dik path field on page 'add new osd'
- VSM-352 servermgmt page autorefrush too frequent
- VSM-346 Add complete set of Juno features in Horizon; pluggable settings, etc.
- VSM-341 expect to have some utilties to help make automation tests.
- VSM-291 The operator expects to present pools to multiple openstack clusters or regions
- VSM-331 it's necesary to monitor device utilization from dashboard
- VSM-339 Configuration Management: as more and more features are added, and different scenarios require different settings, configuration is requried.
- VSM-337 Upgrade: expect to upgrade from 1.0 to 2.0
- VSM-312 vsm-node sets node status="Need more IP" if a Monitor only node does not have a cluster IP address..
- VSM-17 [CDEK-1608] User can remove himself, but after command execution he is not automatically log out
- VSM-329 Remove Monitors button in Manage Servers hangs when Monitor node also has MDS daemon
- VSM-274 a new user logged in vsm and can delete himself
- VSM-362 there are a lot of "in-Up" log in apache error log
- VSM-364 when remove one monitor without MDS co-located, the operation keeps running and no end.
- VSM-359 get smart info inaccurate
- VSM-363 if removing a monitor where the MDS is co-located, after finished, the status has changed to "available" instead of "active".
- VSM-319 performance monitor error when restart vsm processes on controller
- VSM-322 there are exceptions when request vsm dashboard
- VSM-330 When upgrading from Firefly to Giant, the upgrade complains missing python-rados package
- VSM-360 No servers list after newly installed with errors " Permission denied: '/tmp/crushmap'"
- VSM-332 sometimes diamond service start failed
- VSM-333 Error: "Activation of org.freedesktop.secrets timed out", when installing with 2.0 beta 1
- VSM-345 remove server error:no pemssion to stop osd service
- VSM-348 deploy agent node after vsm cluster has been created, the token in .token file is invalid
- VSM-361 when one osd is full, the dashboard doesn't mark one osd is full
- VSM-347 When clicking on "add new osd" at device management, a server error page pops up
- VSM-353 changed the password of current user,an error page pops up
- VSM-336 On Dashboard, even no cluster is created, the VSM version and uptime should be displayed
- VSM-343 can't create cache tier with latest code
- VSM-344 after a few times install, clean-up, reinstall, we have seem some "eaddrinuse" errors on starting rabbitmq
- VSM-357 'add new disk':if data disk and journal disk is in the different partitions of the same disk, the opearation will be unsuccessful
- VSM-350 no response when get latency performance on dashboard
- VSM-351 ceph upgrade on ubuntu unsuccess
- VSM-356 I got warning info means that the number of pg in each osd is too large after upgrade ceph from a lower version to hammar
- VSM-354 when 'create cluster' page opened,other pages still can be opened
- VSM-349 click 'create cluster',I got the tip:there are some zones no monitor created
- VSM-340 Manifest: even using the same token generated from agent-token, server_manifest still complains connection error
- VSM-328 Dashboard: when one disk is full, the warning message is not showing correctly
- VSM-338 Upgrade: current ceph upgrade requires all agent nodes have internet connections, which normally is not true.
- VSM-334 upgrade: when set hammer url, then click "submit", the final ceph version is still firefly.
- VSM-335 The version on UI only shows 2.0.0, but no build number
- this is a 2.0 beta 1 release, still plenty of changes in this release, and we did some initial tests on it.
- 15 new features are added, and 41 bugs are fixed till this release.
- VSM-135 CentOS 6.5 and 7.0 Support
- VSM-296 Show performance on the dashboard
- VSM-4 Average Response Time" missing in dashboard Overview panel "VSM Status" section.
- VSM-159 add issue reporting tool
- VSM-156 add sanity check tool to help identify potential issues before or after deployment
- VSM-15 VSM-backup prompt info not correct
- VSM-124 [CDEK-1852] VSM | adding possibility to manipulate ceph values in cluster.manifest file
- VSM-242 Allow user to modify ceph.conf outside VSM
- VSM-184 add automated script to help deploy VSM on multiple nodes
- VSM-103 Suport a Single Keystone Instance in OpenStack Environments
- VSM-303 Ajust the UI like openstack dashboard
- VSM-54 On OSD Status page: Provide filter option to only display OSDs that are not up and in
- VSM-190 Add disks to cluster
- VSM-220 Operator expects to know what's happending inside ceph cluster through performance counters
- VSM-134 Ubuntu 14.X Support
- VSM-318 "remove osd" doesn't work
- VSM-325 Monitor: negative update time shown on monitor status when removing an monitor.
- VSM-269 Smart info doesn't work in Device mgmt.
- VSM-324 Monitor: inconsistent monitor number on Monitor status page when removing one monitor
- VSM-326 UI: the progress indicator is not showing at the center of screen but the top-left
- VSM-323 Installer: when installing diamond on each agents, there are a lot of File or Folder not found issues
- VSM-306 when executing buildvsm on ubuntu, the script complains no vsm user created
- VSM-263 The pie chart in "Storage Group Status" page is disappeared.
- VSM-320 if installing from a node instead of the controller, the installation will be stopped at missing deployrc
- VSM-176 SSL certificate password is stored in a plain text file
- VSM-230 when presenting pool to openstack, cache tiering pools should be listed.
- VSM-26 [CDEK-1664] VSM | Not possible to replace node if ceph contain only 3 nodes.
- VSM-51 Install Fails for VSM 0.8.0 Engineering Build Release
- VSM-236 no way to check manifest correctness after editing them
- VSM-29 VSM-agent process causes high i/o on os disk
- VSM-216 Add storage group requires at least 3 nodes
- VSM-33 negative update time in RBD list
- VSM-113 [CDEK-1835] VSM | /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-179 keep ceph.conf up to date when executing "remove server" operations.
- VSM-177 wrong /etc/fstab entry for osd device mount point
- VSM-233 console blocks when running automatic installation procedure
- VSM-239 with automatic deployment, the execution is blocked at asking if start mysql service
- VSM-166 cluster_manifest sanity check program gives incorrect advice for auth_keys
- VSM-207 can't assume eth0 device name
- VSM-224 Controller node error in /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-121 Storage node unable to connect to controller although network is OK and all setting correct
- VSM-244 Internal server error when installing v1.1
- VSM-123 Storage node will not be able to contact controller node to install if http proxy set
- VSM-171 [CDEK1672] VSM_CLI | list shows Admin network in Public IP section
- VSM-168 [CDEK1800] VSM_CLI | remove mds - doesn't update vsm database
- VSM-193 hard-coded cluster id
- VSM-260 the check_network in server_manifest will be wrong when it has a single network card
- VSM-317 the update time in MDS block is empty
- VSM-249 can not open dashboard and get the debug infomation "Invalid block tag: 'horizon_nav'"
- VSM-299 when after installed vsm, but no ceph cluster deployed, there is still show 3 storage groups
- VSM-310 with Firefox, for all tables, the vertical separator between the first and second column gets disappered
- VSM-313 the logo is disappeared when running on IE 11
- VSM-314 there is no progress indicator when an operation is not returned.
- VSM-316 the button style in "device management" page is different from others.
- VSM-311 The blocks in dashboard are exceed out of the width of heading
- VSM-315 "add server" doesn't work in alpha 2
- VSM-332 sometimes diamond service start failed
- VSM-330 When upgrading from Firefly to Giant, the upgrade complains missing python-rados package
- this is a 2.0 alpha 2 release, we did some initial tests on it.
- 29 new features are added, and 47 bugs are fixed till this release.
- VSM-63 Filter/Identify OSDs that are not up and in
- VSM-61 Openstack Prod release alignment with VSM
- VSM-55 Paginate OSD pages
- VSM-10 VSM1.0 -[CDEK-1190] VSM "Create cluster" cannot select storage nodes or monito
- VSM-304 Support to present pool to Openstack Juno All in one
- VSM-282 current installer scripts only support ip address based, it's also required to support host name based
- VSM-256 current logic enforces 3 replicas, somethings user may require to loose or tighten the enforcement.
- VSM-257 reorganize functions inside installer to make it's possible to install controller or agent separately
- VSM-243 the "create cluster" button doesn't take in effect after click on ubuntu 14
- VSM-235 make scripts can support both rpm and deb packages
- VSM-222 It's necessary to import a VSM created cluster by VSM itself.
- VSM-221 upgrade vsm dependent openstack packages to juno
- VSM-212 request to support ceph hammer release
- VSM-209 support multiple subnets
- VSM-189 “Add Server” progress
- VSM-163 server_manifest does not check for local host in /etc/hosts
- VSM-154 Request to have a new column on storage-group-status page to have number of OSD's in each storage group.
- VSM-143 On OSD Status page, limit to 100 OSDs per page
- VSM-144 On OSD Status page, sort OSDs by status, server, % capacity used
- VSM-130 OpenStack Juno Support
- VSM-94 Implement Pass-through Parameters
- VSM-85 Storage Group Status improvements
- VSM-89 Full OS disk indication
- VSM-84 Improved server status reporting
- VSM-83 Restart stopped monitor
- VSM-69 Override default PG per OSD value
- VSM-73 Zone monitor warning
- VSM-46 overlay vsm with existing ceph cluster
- VSM-23 Create User UI should prompt more information
- VSM-305 presenting pool should keep backward comatibility, say working with juno, icehouse, havana
- VSM-300 if no cluster is created, when open "server management""manage servers", the Ui becomes messy
- VSM-298 on dashboard, when adding new monitors, the monitor block doesn't update accordingly
- VSM-297 on dashboard, the storage group balls have no data
- VSM-293 wrong version of vsm displayed in the dashboard
- VSM-295 the installation fails at downloading dependent packages if http proxy is required
- VSM-292 the background of tool tip is not clear to view
- VSM-286 present pools to openstack again after you have presented pools, the attach status is always starting
- VSM-285 present multi pools to openstack cinder, the cinder.conf is not complete
- VSM-287 messy login window if accessing https:///dashboard
- VSM-284 at cluster status monitor, it's expected to provide complete summary instead the partial information from dashboardard
- VSM-283 after restart ceph cluster, the osd tree becomes messy
- VSM-280 "Internal Server Error" when click on "Data Device Status"
- VSM-278 after I add a server by clicking 'add server',the osd tree become messy
- VSM-279 click 'restart osd' or 'restore osd',the osd tree will become messy
- VSM-281 on dashboard, the cluster status like "warning" shouldn't be clickable
- VSM-276 Allow agent to be deployed on nodes where no ceph role assigned
- VSM-277 click on "add storage group" will tigger "something went wrong" error on UI
- VSM-272 on OS with newer rabbitmq version like 3.4, message queue can't be accessed
- VSM-273 "Import Error name Login" received when try to login to web console
- VSM-275 the overview page is messy with Firefox
- VSM-270 reinstall the keystone when deploy the vsm
- VSM-266 Dashboard UI correction
- VSM-265 Click Manage VSM"Add/Remove User""Change Password", no page shows up
- VSM-264 on "PG Status" page, the text layout is messy.
- VSM-267 The logo is not fully shown.
- VSM-268 Can't create EC pool
- VSM-262 when clicking on "device management", the page shows "something went wrong"
- VSM-258 wrong dependencies url
- VSM-259 script 'vsm-agent' missing LSB tags and overrides
- VSM-253 A lot of buttons is invalid such as “addmonitor”,“removemonitor”,“addserver”,“removeserver”,“restore osd”,“remove osd” and so on
- VSM-254 the successful tip boxes can't shut down such as "Success: Successfully created storage pool: "
- VSM-255 when creating ceph cluster with one zone defined, the ceph cluster keeps in health_warn status
- VSM-252 Most of the interface's auto refresh function is invalid
- VSM-251 after clicked clusterManagement->create cluster->create cluster,there is not mds created.
- VSM-250 Error when packing
- VSM-245 error creating cluster
- VSM-231 VSM | Not possible to choose the machines during cluster creation
- VSM-229 Error state displayed on green color
- VSM-225 VSM Creates Very Small pg_num and pg._num Size for EC Pool
- VSM-218 on centos 7, the web layout is messy
- VSM-195 if no cluster initialized, the UI shouldn’t knock at ‘manage servers’.
- VSM-194 User experience: OSD Status and Manage Devices.
- VSM-180 When a cluster is destroyed and recreated, VSM controller increments the cluster id (aka database rowid)
- VSM-153 after clicked vsm->settings->update,there is no successful tip
- VSM-68 Veriffy there are enough discrete servers and/or zones to support a VSM requested replication level
- VSM-5 KB used in "Pool status" panel not consistent with "ceph df" in backend
- VSM-313 the logo is disappeared when running on IE 11
- this is alpha 1 release for 2.0, which is not well verified yet, just for early preview.
- this is a bugfix release for 1.1, which fixed the issues found in VSM-242 ("Allow user to modify ceph.conf outside VSM"). Also adding an new script uninstall.sh to help uninstall VSM in the case user expects to restart installation again.
- VSM-209 support multiple subnets
- VSM-260 the check_network in server_manifest will be wrong when it has a single network card
- starting from v1.1, the vsm dependencies are maintained on vsm-dependencies repository for tracking and trouble shooting.
- starting from v1.1, an automatic deployment tool is provided to deploy whole vsm controller and agents from one placement.
- VSM-156 add sanity check tool to help identify potential issues before or after deployment
- VSM-159 add issue reporting tool
- VSM-184 add automated script to help deploy VSM on multiple nodes
- VSM-242 Allow user to modify ceph.conf outside VSM
- VSM-4 Average Response Time" missing in dashboard Overview panel "VSM Status" section.
- VSM-15 VSM-backup prompt info not correct
- VSM-25 VSM Dashboard | Capacity of hard drives is wrong and percentage used capacity is not correct.
- VSM-26 [CDEK-1664] VSM | Not possible to replace node if ceph contain only 3 nodes.
- VSM-29 vsm-agent process causes high i/o on os disk
- VSM-33 negative update time in RBD list
- VSM-51 Install Fails for VSM 0.8.0 Engineering Build Release
- VSM-113 [CDEK-1835] VSM | /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-121 Storage node unable to connect to controller although network is OK and all setting correct
- VSM-123 Storage node will not be able to contact controller node to install if http proxy set
- VSM-124 [CDEK-1852] VSM | adding possibility to manipulate ceph values in cluster.manifest file
- VSM-166 cluster_manifest sanity check program gives incorrect advice for auth_keys
- VSM-168 [CDEK1800] VSM_CLI | remove mds - doesn't update vsm database
- VSM-171 [CDEK1672] VSM_CLI | list shows Admin network in Public IP section
- VSM-176 SSL certificate password is stored in a plain text file
- VSM-177 wrong /etc/fstab entry for osd device mount point
- VSM-179 keep ceph.conf up to date when executing "remove server" operations.
- VSM-193 hard-coded cluster id
- VSM-207 can't assume eth0 device name
- VSM-216 Add storage group requires at least 3 nodes
- VSM-217 Problem with replication size on "pool" created on newly added storage group.
- VSM-224 Controller node error in /var/log/httpd/error_log - constantly ongoing messages [error] <Response [200]>
- VSM-230 when presenting pool to openstack, cache tiering pools should be listed.
- VSM-233 console blocks when running automatic installation procedure
- VSM-236 no way to check manifest correctness after editing them
- VSM-239 with automatic deployment, the execution is blocked at asking if start mysql service
- VSM-244 Internal server error when installing v1.1
Special Notes
For installation, please check the INSTALL.md on current release branch instead of the one on master branch.
New Features
Openstack Icehouse support
Fixed bugs
VSM-187 GUI shows always replication 3 when replicated pool is other than "Same as Primary"
VSM-183 ct_target_max_mem_mb and ct_target_max_capacity_gb sets the same value in create cache tier
VSM-182 add osd_heartbeat_grace and osd_heartbeat_interval to cluster.manifest
VSM-173 Can select same storage group in create replicated pool -- which should not
Known issues
VSM-191 VSM does not display the replica storage group.
VSM-192 force nonempty.
VSM-188 Missing column in manage pool page.