InfoWorld’s 2018 Technology of the Year Award winners

InfoWorld’s 2018 Technology of the Year Award winners

The Open Compute Project’s open hardware standards have done much to push forward the development of cloud-scale hardware. By sharing designs for connectors, racks, servers, switches, and storage hardware, the OCP has defined a new generation of data center technologies and made them widely available – and able to be manufactured at the scale the big public clouds need.

Project Olympus is one of Microsoft’s open hardware designs, shared with OCP members and driven as an open design project with multiple forks of Microsoft’s initial spec. Built around the OCP’s Universal Motherboard specification, Project Olympus is a flexible compute server, with support for Intel, AMD, and ARM64 processors as well as FPGA, GPGPU, and other specialized silicon to add features as needed.

The initial Project Olympus hardware has been joined by a second, deeper chassis design, the Project Olympus Hyperscale GPU Accelerator. The “HGX-1” hosts eight Pascal-class Nvidia GPUs for machine learning workloads. Four HGX-1 servers can be linked together via Nvidia’s NVLink to give you up to 32 GPUS, ideal for complex workloads.

Cloud data centers need lots of hardware, but workloads are moving away from the one-size-fits-all x86 server. The flexible Project Olympus design allows the same chassis to support different motherboards and thus handle the varied workloads running on modern cloud infrastructures. And as it’s open hardware, it can be produced by any manufacturer, ensuring wide availability and low prices.

— Simon Bisson

Source: InfoWorld Big Data

Technology of the Year 2018: The best hardware, software, and cloud services

Technology of the Year 2018: The best hardware, software, and cloud services

Was 2017 the year that every product under the sun was marketed as being cognitive, having machine learning, or being artificially intelligent? Well, yes. But don’t hate all of them. In many cases, machine learning actually did improve the functionality of products, sometimes in surprising ways.

Our reviewers didn’t give any prizes for incorporating AI, but did pick out the most prominent tools for building and training models. These include the deep learning frameworks Tensor­Flow and PyTorch, the automated model-building package Driverless AI, and the solid machine learning toolbox Scikit-learn.

The MLlib portion of Apache Spark fits into this group as well, as does the 25-year-old(!) R programming language, of which our reviewer says, “No matter what the machine learning problem, there is likely a solution in CPAN, the comprehensive repository for R code, and in all likelihood it was written by an expert in the domain.”

2017 was also the year when you could pick a database without making huge compromises. Do you need SQL, geographic distribution, horizontal scalability, and strong consistency? Both Google Cloud Spanner and CockroachDB have all of that. Do you need a distributed NoSQL database with a choice of APIs and consistency models? That would be Microsoft’s Azure Cosmos DB.

Are you serving data from multiple endpoints? You’ll probably want to use GraphQL to query them, and you might use Apollo Server as a driver if your client is a Node.js application. Taking a more graph-oriented view of data, a GraphQL query looks something like a JSON structure with the data left out.

As for graph database servers, consider Neo4j, which offers highly available clusters, ACID transactions, and causal consistency. Are you looking for an in-memory GPU-based SQL database that can update geospatial displays of billions of locations in milliseconds? MapD is what you need.

Two up-and-coming programming languages made the cut, for completely different domains. Kotlin looks like a streamlined version of object-oriented Java, but it is also a full-blown functional programming language, and most importantly eliminates the danger of null pointer references and eases the handling of null values. Rust, on the other hand, offers memory safety in an alternative to C and C++ that is designed for bare-metal and systems-level programming.

Speaking of safety, we also salute two security products—one for making it easier for developers to build secure applications, the other for extending security defenses to modern application environments. GitHub security alerts notify you when GitHub detects a vulnerability in one of your GitHub project dependencies, and suggest known fixes from the GitHub community. Signal Sciences protects against threats to your web applications and APIs. 

If you’ve started deploying Docker containers, sooner or later you’re going to want to orchestrate and manage clusters of them. For that, you’ll most likely want Kubernetes, either by itself, or as a service in the AWS, Azure, or Google clouds. Honeycomb goes beyond monitoring and logging to give your distributed systems observability.

Recently, the heavyweight Angular and React frameworks have dominated the discussion of JavaScript web applications. There’s a simpler framework that is gaining mindshare, however: Vue.js. Vue.js still builds a virtual DOM, but it doesn’t make you learn non-standard syntax or install a specialized tool chain just to deploy a site.

Microsoft’s relationship with Linux has been troubled over the years, to say the least. For example, in 2001 Steve Ballmer called Linux a “cancer.” The need for Linux in the Azure cloud changed all that, and the Windows Subsystem for Linux allows you to run a for-real Ubuntu or Suse Bash shell in Windows 10, allowing you to install and run legitimate Linux binary apps from the standard repositories, including the Azure Bash command line.

Read about all of these winning products, and many more, in our tour of 2018 Technology of the Year Award winners.

Source: InfoWorld Big Data

IDG Contributor Network: Harmonizing big data with an enterprise knowledge graph

IDG Contributor Network: Harmonizing big data with an enterprise knowledge graph

One of the most significant results of the big data era is the broadening diversity of data types required to solidify data as an enterprise asset. The maturation of technologies addressing scale and speed has done little to decrease the difficulties associated with complexity, schema transformation and integration of data necessary for informed action.

The influence of cloud computing, mobile technologies, and distributed computing environments contribute to today’s variegated IT landscape for big data. Conventional approaches to master data management and data lakes lack critical requirements to unite data—regardless of location—across the enterprise for singular control over multiple sources.

The enterprise knowledge graph concept directly addresses these limitations, heralding an evolutionary leap forward in big data management. It provides singular access for data across the enterprise in any form, harmonizes those data in a standardized format, and assists with the facilitation of action required to repeatedly leverage them for use cases spanning organizations and verticals.

Enterprise-spanning connections and data representation

An enterprise data fabric delivers these benefits by successfully extending the notion of master data management and data lakes. The former is predominantly a means of describing the underlying data, typically via unified schema. In their original inception data lakes grant universal access to data in their native formats, yet lack the necessary metadata and semantic consistency for long term sustainability.  

Enterprise knowledge graphs, however, include the metadata and semantic benefits of MDM hubs but link all data together in adherence to semantic standards. The combination of enterprise-wide ontologies, taxonomies, and terminology delivers data in a representation (in terms of meaning and schema) immediately identifiable to the user. These linked data approaches connect all data uniformly.

Health care providers, for example, can connect the voluminous types of data relevant to their industry by creating an exhaustive list of events such as diagnostics, patient outcomes, operations, billing codes, and others, describing them with standardized models and fortifying them with uniform terminology across the data spectrum.

Regardless of where data is—whether in the cloud, a cache, or awaiting computation—users can link them in the same consistent format that has meaning to their business purposes. The standardized ontologies, which are malleable to incorporate new events, and unified terminology align all data to the knowledge graph’s schema regardless of their origination or other points of distinctions.

Active automation

The true value of an enterprise knowledge graph’s access layer is in the automated action it facilitates. With data stemming from any number of disparate source systems, the automatic generation of code for analytics or transformation is invaluable. Such automation is one of the crucial advantages of an enterprise knowledge graph that reduces the difficulty in not only accessing data, but also applying them to necessary action.

The use cases provisioned by this combination are innumerable. A health care organization attempting to predict the event of respiratory failure for patients in multiple locations could use knowledge graph applications to monitor the blood pressure of all hospital incumbents. The graph would enable the organization to create an abstract description of the blood pressure data related to this event (respiratory failure), then automatically compile that description into code which obtains the prediction data.

The overarching value proposition of this approach is that the user simply issues a query for the data he or she needs regardless of where data originated. The automation capabilities of an enterprise knowledge graph create the action whereby the data that pertains to the query is attained. The key difference is the user need not necessarily know the source system or the particularities of its schema to get the data. Access is much more expedient since all of the data engineering work of cleansing and transforming data is done upfront prior to issuing queries.

In addition, the data relevant to the query can stem from multiple source systems, but is still accessed from a single place (the knowledge graph). The user is not responsible for personally accessing those systems; instead, the query mechanism is able to cull the appropriate data from the varying systems accessible from the centralized knowledge graph.


An enterprise knowledge graph effectively unifies several aspects of the considerable variations intrinsic to big data. It unifies the means of accessing, representing, automating and even moving data from an array of source systems and architectural complexities. In addition to streamlining how users retrieve diverse data via automation capabilities, the knowledge graph standardizes those data according to relevant business terms and models. The result is a homogenized data set wrought from any number of data types and sources.

This article is published as part of the IDG Contributor Network. Want to Join?

Source: InfoWorld Big Data

IDG Contributor Network: AI disruptions and the power of intelligent data

IDG Contributor Network: AI disruptions and the power of intelligent data

Machine learning and Artificial Intelligence brings in transformations through practically every domain and industry in an unprecedented way. Fueled by ever increasing computing resources, evolution of faster algorithms, developments in machine learning backed by vast amounts of data—AI is bringing rapid changes the existing business processes.

It is important that an AI system is engineered to interpret and demonstrate a general intelligence as humans, demonstrate a level of intelligence that is not specific to one category of tasks or at least be able to generalize those specifics, and relate those understandings in the context of real world tasks, issues and cases.

Ability to balance this interpretation in the right manner enables an AI system to deal with new situations which are very different that the ones system has encountered earlier.

The “intelligence” in the data

Companies are striving to bring transformations, areas such as operations optimization, fraud detection and prevention, and financial performance improvements are becoming more and more focused, and one of the key factors that’s drives these initiatives to success is the capability to drive them with intelligent data from trusted source of repositories. As for any other digital strategies organizations build, data is pivotal to success of artificial intelligence strategy. Said differently, AI systems can only be as intelligent as the data they deal with.

TensorFlow, an open-source library and part of Google brain project for machine learning, performs language modeling for sentiment analysis and predicting the next words in sentences given the history of previous words. Language modeling is one of the core tasks of natural language processing and is widely used in the areas of image captioning, machine translation, and speech recognition. TensorFlow creates predictive models by training machine learning algorithms with large data sets, and the algorithm iteratively makes predictions based on the training data, but desired result is not to create a model predicting the test data sets but to create a model with a good generalization. TensorFlow is flexible and can be used to express a wide variety of algorithms, including training and inference algorithms for deep neural networks, large datasets generated from YouTube videos, and photos repositories, to name a few.

Why is enterprise data so critical for AI?

Data is the primary source that builds knowledge base with timely and meaningful insights. Data velocity and volume, both play a crucial role in defining an accurate and information based decision. The logic applies to machine learning. AI systems modeled with data from wide variety of sources are expected to produce results with better accuracy, in a timely manner. This is important for any organization in any industry that wants to build AI capabilities.

Machine learning algorithms do the job to fit a specific AI model to data. First a model is built and then it gets trained by presenting large volumes of data sets to it. The model adjusts itself to tune to various datasets, however, affinity to one data set or training data may bring more challenges when a new scenario is encountered.

Generalization and specialization

People learn many things from examples. Teachers teach many things to children by citing examples. They learn to differentiate between two similar objects by knowing more smaller attributes about those objects. Barring the exception of some superkids, it is difficult for most two-year-old kids to differentiate cars models. Human intelligence at that age forms a model with the knowledge based on inductive learning. This technique is referred as generalization. Inductive learning requires the model to grow by exposing more cars to it, and build a belief based on the developing cognitive abilities and from observing more cars. Based on the initial belief, it can be assumed that a car has four wheels.

The specialization technique is exactly opposite of generalization. Human intelligence grows with age, and human cognition recognizes more attributes in people’s model of understanding to differentiate things based on those attributes. Likewise, AI agents need specialization when inductive learning on a very large data make the model overgeneralized. For example, after observing restaurants and their menu, we may build a general belief that all restaurants serve food. But not all restaurants serve kosher food. Here, we must use specialization to restrict the generalization of the belief to find special restaurants that serve kosher food. In simple words, specialization can be achieved by adding extra conditions to the predefined model.

AI strategies to teach machines to think like people do

AI based systems utilize generalization and specialization to build their core knowledge base and train their model. The idea is to maintain a single model based on the training data and to adjust it as new scenarios arrive to maintain consistency. This requires them to process data or information about a subject area in different manner based on the need. To create better predictive models that are capable of generalizing, systems also need to know when to stop training the model so that it doesn’t overfit.

Overfitting can happen when the model becomes very large and the AI agent starts making decisions based on training data. If the training data sets are too narrow and do not contain many scenarios, it may cause underfitting problem as well. In the opposite strategy, AI system needs to evaluate many scenarios and build the model in a way that it can represent some sort of specialization.

Role of quality enterprise data

Not all the data sets are equally useful though, consider the AI agent in a self-driving car trained only on straight highways will have a hard time navigating in a desert unless it was trained by using similar models. Good AI agents need simple data sets to build a generalized model first and then also need diverse data sets to address tasks where specialization is needed. This requires a constant inflow and wide variety of metadata and data to train the machine learning system, and provide a trusted and effective artificial intelligence.

Enterprises can build great AI systems by using data across repositories and by building rich training data sets and catalogues and further train AI agents using those training data sets. Meta data can also be integrated from range of systems like mainframes, HR, CRM, and social media and can be gathered in to the training data sets and catalogs.

This article is published as part of the IDG Contributor Network. Want to Join?

Source: InfoWorld Big Data

TensorFlow review: The best deep learning library gets better

TensorFlow review: The best deep learning library gets better

If you looked at TensorFlow as a deep learning framework last year and decided that it was too hard or too immature to use, it might be time to give it another look.

Since I reviewed TensorFlow r0.10 in October 2016, Google’s open source framework for deep learning has become more mature, implemented more algorithms and deployment options, and become easier to program. TensorFlow is now up to version r1.4.1 (stable version and web documentation), r1.5 (release candidate), and pre-release r1.6 (master branch and daily builds).

The TensorFlow project has been quite active. As a crude measure, the TensorFlow repository on GitHub currently has about 27 thousand commits, 85 thousand stars, and 42 thousand forks. These are impressive numbers reflecting high activity and interest, exceeding even the activity on the Node.js repo. A comparable framework, MXNet, which is strongly supported by Amazon, has considerably lower activity metrics: less than 7 thousand commits, about 13 thousand stars, and less than 5 thousand forks. Another statistic of note, from the TensorFlow r1.0 release in February 2017, is that people were using TensorFlow in more than 6,000 open source repositories online.

Much of the information in my TensorFlow r0.10 review and my November 2016 TensorFlow tutorial is still relevant. In this review I will concentrate on the current state of TensorFlow as of January 2018, and bring out the important features added in the last year or so.

Source: InfoWorld Big Data

Comcast Selects AWS As Its Preferred Public Cloud Provider

Comcast Selects AWS As Its Preferred Public Cloud Provider

Amazon Web Services, Inc. (AWS) has announced that Comcast Cable has selected AWS as its preferred public cloud infrastructure provider. Comcast Cable will expand its use of AWS by migrating material workloads and building new applications on AWS.

AWS provides important infrastructure and services to Comcast as it focuses on building cloud-native products and services on AWS that adapt and evolve to meet the needs of customers. That focus is reflected in the class-defining X1 Platform, award-winning voice-control technology, and Xfinity xFi, Comcast’s personalized Wi-Fi experience that gives customers pinpoint control over their home networks. Comcast’s primary businesses, Comcast Cable and NBCUniversal, are currently running workloads on AWS which has enabled these businesses to become more nimble and launch new, revenue-generating initiatives in the competitive entertainment industry.

“We have deepened our strategic relationship with AWS, making the industry’s leading cloud our preferred public cloud provider,” said Jan Hofmeyr, chief network and operations officer and senior vice president at Comcast Cable. “Over the years, we have leveraged the breadth and depth of AWS’s services, including compute, storage, and analytics. In that process, we’ve found AWS to be extremely customer focused.”

“For industry leaders like Comcast Cable, the quest to anticipate and exceed consumers’ media and entertainment needs is never ending,” said Mike Clayville, vice president, worldwide commercial sales at AWS. “Comcast Cable’s goal has always been to stay a step ahead of the competition. In order to do that, they wanted solutions that were agile, flexible and ready for what’s next. Together, AWS and Comcast Cable collaborated to enable them to confidently move core business workloads, build new applications with ease, and gain the agility they required by using AWS.”

Source: CloudStrategyMag

Cloud Standards Customer Council Updates White Paper

Cloud Standards Customer Council Updates White Paper

The Cloud Standards Customer Council™ (CSCC™) has published version 2.0 of its white paper, Interoperability and Portability for Cloud Computing: A Guide. Authors of the paper will present a complimentary webinar on January 24, 2018, from 11:00am – 12:00pm ET to introduce the paper.

Today, organizations are employing a wide range of cloud services and transitioning data and applications to cloud computing environments. The topics of “interoperability” and “portability” are significant considerations in relation to the use of cloud services, but there is also confusion and misunderstanding of exactly what this entails. The aim of this guide is to provide a clear definition of interoperability and of portability and how these relate to various aspects of cloud computing and to cloud services.

Version 2.0 has been updated to reflect the new ISO/IEC 19941 Cloud Computing Interoperability and Portability standard and its facet models of interoperability, data portability, and application portability. The model of an application and the process of porting an application have been updated to reflect new thinking contained in ISO/IEC 19941. Additionally, containers and the role of automation have been addressed, as these have become dominant trends for cloud users.

The paper provides guidance for avoiding vendor-lock in, which allows customers to make the best use of multiple diverse cloud services that can cooperate and interoperate with each other, which is critical to future cloud service adoption and the realization of the benefits of computing as a utility.

Source: CloudStrategyMag

IDG Contributor Network: Dawn of intelligent applications

IDG Contributor Network: Dawn of intelligent applications

Data remains a foundational element of computing. Recently, Hadoop and big data have been a central part of data progression, allowing you to capture data at scale. But companies now look to the expanding use of cloud computing and machine learning to create more intelligent applications.

This new generation of applications use all the data they can, including incoming real-time data, to respond in the moment to changing circumstances and formulate advantageous outcomes. This includes delivering on the digital transformation promise sought by companies to deliver rich customer experiences.

Intelligent applications can converge database and data warehouse workloads, allowing companies to respond and react to changing conditions in real time.

This builds on a theme covered by nearly every large industry analyst firm regarding the merging of transactional and analytical functions. Gartner refers to this convergence as hybrid transaction analytical processing, or HTAP; 451 Research refers to it as hybrid operational analytical processing, or HOAP; and Forrester refers to it as translytical data platforms.

According to Forrester:

Analytics at the speed of transactions has become an important agenda item for organizations. Translytical data platforms, an emerging technology, deliver faster access to business data to support various workloads and use cases. Enterprise architecture pros can use them to drive new business initiatives.

451 Research also calls out the idea of seizing the moment:

Organizations are zeroing in on the so-called “transaction window” and realizing that it presents a significant opportunity—and once it’s gone, it’s gone for good.

Intelligent applications in finance, media, and energy sectors

The largest industry sectors are using these converged technologies for their intelligent applications. These applications collect and process data from a variety of sources, provide experiences in real time, and make use of the latest techniques in machine learning and artificial intelligence to push their usefulness forward.

Consider the following examples.


A popular intelligent application in finance is the new frontier of digital wealth management, including real-time portfolio analytics for clients across any platform. As one example, JP Morgan Chase highlighted its investment in digital wealth management in an investor presentation last year.

Behind the scenes, many of these digital wealth services are powered by digital startups such as InvestCloud, which states that its wealth management products “allow wealth managers to get a whole view of their clients—instantaneously and at scale.” Other companies in this space include SigFig, which showcases “experience integrating our platform with TD Ameritrade, Charles Schwab, Vanguard, E*Trade, among others.”


In the energy sector, intelligent applications include IoT data pipelines using sensor data. Real-time capture and analysis of this data, with machine learning model scoring, provides helpful downtime mitigation and savings for global companies. Shell describes its advanced analytics for sensor collection on its website:

Digital sensors installed in our operations around the world—from production fields to manufacturing complexes—produce a constant flow of data which we analyze to improve processes and take better business decisions.

The technology can optimize the performance of plants by predicting when maintenance will be needed, to avoid unplanned downtime and lost productivity.

More than 5,000 machines globally connect to the system, which is thought to have saved more than 3.5 million barrels in lost production since its introduction.


Perhaps no transformation is more visible in media than the shift from broadcast and cable television to real-time streaming. This change drives media companies to seek end user analytics and advertising opportunities tied to streaming, and key intelligent applications to drive revenue. This technology race led Disney to acquire a majority stake in BAMTech. Disney said the following in a news release:

The media landscape is increasingly defined by direct relationships between content creators and consumers, and our control of BAMTech’s full array of innovative technology will give us the power to forge those connections, along with the flexibility to quickly adapt to shifts in the market.

4 steps towards intelligent applications

Nearly every company is charting its own digital transformation blueprint. Adding intelligent applications to the mix can jumpstart digital results. Here are four quick steps to get started:

1. Identify corporate growth areas

Projects aligned with company objectives often get faster funding and resources.

2. Capture new and existing data sources

Showcase both to enhance existing value and demonstrate new data.

3. Design an architecture that supports a real-time feedback loop

Merge transactions and analytics where possible to incorporate real-time intelligence, including machine learning model scoring.

4. Build intelligent applications using the latest best practices across industries

Track primary applications in high growth areas of finance, media, and energy to see how companies are putting technology to use.

New intelligent applications have the power to transform businesses, drive new revenue, deepen customer engagement, and optimize operations. With a simple action plan and examples from industry leaders you can set your own company on a path to success.

This article is published as part of the IDG Contributor Network. Want to Join?

Source: InfoWorld Big Data

CloudHealth Technologies And ParkMyCloud Partner

CloudHealth Technologies And ParkMyCloud Partner

CloudHealth Technologies and ParkMyCloud have announced they are partnering to marry the hybrid cloud governance of CloudHealth with the automated cost control of ParkMyCloud.

Customers leveraging the integrated solution will experience greater return on their cloud investments. They will be able to automate cloud cost control, simplify management, and consequently free up teams to focus on driving more strategic projects within their organizations.

Public cloud provides agility, efficiency, and flexibility; however, as organizations ramp up in the public cloud, they often find consumption growing rapidly, leading to overspending and inefficient resource utilization. For many, this rising cost is unaccounted for. According to Gartner, “With increasing business unit IT spending on cloud services, IT leaders must prevent new risks, sprawl, cost overruns and missed SLAs. Dynamic optimization technology can help balance the benefits of agility with required governance controls for cloud services and virtualized infrastructure.”1

With CloudHealth’s expertise in cloud governance and management, and ParkMyCloud’s specialization in cloud cost control, they can together provide complete visibility and control over multi-cloud environments, enabling customers to drive better business value. Joint customers will experience a seamless, integrated solution, including: 

Improved Cloud ROI: Users will realize immediate cost savings with non-disruptive, policy-driven automation that helps to eliminate tedious tasks and enable teams to roll out new offerings and updates faster to market.

Ability to Drive More Strategic Projects: Rather than focusing on “keeping the lights on,” technical experts can shift their efforts to continuously innovate and thereby maintain a competitive advantage—not just business as usual.

Simplified Hybrid Cloud Governance: Broken down by environment, department, application, resource and more, this integrated offering empowers teams to implement better resource management through simple and easy-to-use, customizable dashboards for different personas such as CFO and CTO, among others.

Better Analytics: Users gain unparalleled insight and can use this visibility to make smarter business decisions.

“There are huge business gains to be reaped in the public cloud, but business transformation also brings complexity,” said Tom Axbey, CEO and president, CloudHealth Technologies. “In partnering with ParkMyCloud, we’re eliminating cloud management inefficiency and bridging the divide between cost and utilization. We’re uniting disparate cloud environments, business ops and DevOps teams by equipping them with the tools they need to be agile and productive.”

“Our goal at ParkMyCloud has always been to help our customers do more with their cloud by saving time and money,” said Jay Chapel, CEO and founder, ParkMyCloud. “That’s why we provide them with automated cost control, which we do by finding and turning off idle resources in AWS, Azure, and Google Cloud. By collaborating with CloudHealth Technologies, we are providing customers with end-to-end visibility and control over their environments for optimized cloud usage and spend.”

“With ParkMyCloud, anyone in your organization can be responsible for their own cloud costs,” said Reed Savory, Connotate. “That’s immensely valuable and brings real cost savings. CloudHealth offers unparalleled rightsizing capabilities, among other things, so we know we aren’t leaving money on the table. Integrating these solutions will only further the simplicity and efficiency with which our team operates – even as we manage a rapidly scaling cloud environment.”

1. Gartner, Innovation Insight for Dynamic Optimization Technology for Infrastructure Resources and Cloud Services, Donna Scott and Milind Govekar, Refreshed: 6 February 2017 | Published: 29 February 2016.

Source: CloudStrategyMag

Rootstock Acquires Kenandy Inc.

Rootstock Acquires Kenandy Inc.

Rootstock Software has announced the acquisition of Kenandy Inc., a developer of cloud ERP software. The acquisition solidifies Rootstock Software’s standing as a leading developer of cloud ERP applications focused on the needs of manufacturers, distributors and supply chain organizations utilizing the Salesforce Platform.

“By combining the talent, skills and intellectual property of both companies, Rootstock will achieve greater economies of scale to compete with the likes of Oracle-NetSuite, Microsoft and SAP, while giving us the ability to create more cutting-edge capabilities for Kenandy and Rootstock customers,” said Patrick Garrehy, CEO of Rootstock. “Both Rootstock and Kenandy are known for taking a customer-centric approach to ERP, and this furthers Rootstock’s recent focus on advancing personalized manufacturing with capabilities that deliver a more individualized customer experience across all customer touch points in an organization.”

The combination of Rootstock Cloud ERP, Salesforce Sales Cloud, Salesforce Service Cloud, and the underlying Salesforce Platform technologies, such as Salesforce IoT and Einstein Analytics, form a compelling choice for manufacturers, distributors and supply chain organizations aiming to compete more effectively and better serve their customers. “As the cloud marketplace has matured, buyers are increasingly making CRM and ERP decisions in the context of an overall cloud platform strategy,” said Garrehy. “This acquisition creates a bigger, stronger entity focused on building cloud ERP applications for those who choose the Salesforce Platform as their underlying technology.”

Rootstock, which recently announced a strategic win replacing SAP at Mipox, a global producer of polishing films for the high-tech industry, to provide cloud ERP applications for more than 400 users in 12 countries, aims to not only advance ERP technology with the combined companies, but also continue developing best practices that aid manufacturers in migrating from legacy ERP. “When it comes to cloud ERP implementations, customer success is often determined by how you implement, not just what you implement,” said Garrehy. “Our combined company is dedicated to making the transition from legacy ERP easier for our customers. We welcome Kenandy customers into the Rootstock fold.”

Source: CloudStrategyMag