Our client has over 40 data centers around the world, managing hundreds of thousands of customers' virtual machines. Every few minutes they receive data from each piece of equipment (servers, routers, switches, etc.), such as CPU usage, memory and packet throughputs, etc. The client used a third-party tool to aggregate all of this information and manually create MSRs for their customers. To make this process more efficient and cost-effective, the client decided to replace the third-party tool with their own solution. The new solution would utilize GCP since the client is an official Google partner.
First, our team investigated the workflow of the third-party tool to learn how the data is retrieved, interpreted, and transformed. After a thorough analysis, we built a new solution that uses Google Cloud Dataflow to transfer data from Kafka to BigQuery. Since the client used different vendors on their servers to collect information (Cisco, IBM, etc.), we also needed to standardize and unify all data. Therefore, instead of using the standard ETL (Extract-Transform-Load) approach, we used the ELT method, where data is first extracted from Kafka, loaded into BigQuery, and only then transformed. This has enabled the solution to create comprehensible reporting tables with information about the performance of each server instance.
Our team has also helped the client streamline the generation of MSRs. The process involved lots of manual tasks and consumed nearly 17,000 work hours per year. N-iX has helped the client simplify this process by consolidating information to BigQuerry as a single location. Once the data was consistently available across the customer base, our team automated the manipulation of raw data into MSRs.