There are different public cloud use cases. Here at The Virtualization Practice we moved our datacenter from the north to the south part of the country and utilized the cloud to host the workloads during the transition. Edward Haletky, yesterday posted about Evaluating the Cloud: Keeping your Cloud Presence and presented the question and his thoughts of is it worth staying in the cloud or bringing the data home.
The Cloud Management layer of the Software Defined Data Center is where the flexibility of the SDDC is translated into tangible benefits for the business constituents of the virtualized data center, the private clouds, the hybrid clouds and the public clouds. Without a robust Cloud Management layer, the IT Operations flexibility of the SDDC cannot translate into the relevant level of Business Agility for the business.
By far, the lowest hanging fruit of virtualization and cloud environment security is the segregation of your management control from your workloads. Separation of data and control planes have been recommended for everything from storage (EMC ViPR) up to the workloads running within virtual machines. The same holds true for cloud and virtual environment management tools, tasks, and functions. Up to now there have been very few choices in how such segregation could occur using properly placed firewalls or by using some form of proxy and the only proxy available was HyTrust. But this has changed. There are some other tools that will help with this segregation of data from control and do they give the level of auditing we require to solve the delegate user problem?
Software Defined Data Center Analytics will play a critical role in keeping your SDDC up and running with acceptable performance. Multiple real time data feeds used across a variety of management solutions represent the only viable method to keeping up with the rate of change in the SDDC.
Tintri is now shipping version 2.0 of their array software, Tintri OS, which further automates administration of their arrays and adds ReplicateVM remote replication features. ReplicateVM is an incredibly flexible asynchronous, IP-based replication technology that operates per-VM, allowing each VM to have custom RPOs.
At EMCworld 2013, one of the big stories was Pivotal and it’s importance to the EMC2 family and the future of computing. Pivotal is geared to provide the next generation of computing. According to EMC2 have gone past the Client-Server style to a scale-out, scale-up, big data, fast data Internet of Things form of computing. The real question however, is how can we move traditional business critical applications to this new model, or should we? Is there migration path one can take?