We use it as a scheduling tool. We use it for infrastructure backups and running scheduled tasks, but nothing in regards to data analytics.
It is an infrastructure process behind the scenes: custom backups and custom file migrations.
We use it as a scheduling tool. We use it for infrastructure backups and running scheduled tasks, but nothing in regards to data analytics.
It is an infrastructure process behind the scenes: custom backups and custom file migrations.
We leverage Control-M for backups. That would be a critical process that we have integrated. This allows teams that rely on the backups to have a single pane of glass through the dashboard to determine if their backup is successful. It allows email alerts or triggers, if something fails or we need to do manual intervention.
My organization has been able to script scheduled jobs in Control-M to potentially replace legacy products that are at end of life or end of service. The previous backup applications that were being used for specific files, folders, or applications were no longer being supported, therefore being able to use Control-M to replace that has been very valuable.
We rely on Control-M for automation. Anything that would have been a manual effort previously or legacy, Control-M has been able to replace.
The scheduler allows you to pretty much run anything from anywhere. It is very convenient. The sensor reporting that the scheduler gives you can monitor hundreds of jobs that could potentially be running in a given hour.
All the scheduled tasks are available in a dashboard or workflow view that different teams leverage. This is important and great. Having the ability to have a dashboard or workflow allows for easier troubleshooting. We also have alerting set up through email triggers, which are very helpful.
We leverage it for file transfer. We don't necessarily have application workflows dependent on those, but we do have Control-M for the migration of files. The visibility of a successful transfer is very useful, e.g., the ability to report on that or view whether that job succeeded or failed in the dashboard. You have an alert that would trigger on a failure. So, failure is automated. The Control-M job could retry that file migration a number of times based on logic that you have programmed into the job, and having to avoid manual intervention is useful.
The alerts are helpful and can contribute to faster issue resolution in the event of an issue.
The infrastructure updates could use improvement. Some of the previous updates that we have run to get to version nineteen were troublesome. So, a more seamless upgrade path for the infrastructure components would be useful. I don't know if they have replaced that in version 20 or if version 20 has an easier path, but I would like to see the upgrade from one version to the next version be a little smoother.
I have been using it for about five years.
The platform has been great. I don't think we have had any downtime besides our upgrade process.
The scheduling process has been able to handle almost everything that we have asked it to do. It seems to be able to run pretty much anything from anywhere within our environment.
This solution was a new integration/installation done before my involvement.
The application was a part of the infrastructure when I joined. We have been able to add automations for components that were otherwise manual.
The upgrades are a bit complex. The last time we did an upgrade, it took several hours.
The upgrade was planned. We ran into an issue, then we had to reach out to support. They were quick to respond, but the resolution did take several hours. They did a good job. The issue was resolved in a timely manner during our upgrade window. Their service was an eight or nine out of 10, as far as issue resolution. To be a 10 out of 10, I would like something prescheduled. If we could have had support personnel available for the upgrade procedure, it would have been helpful. So, it was just the time element.
The product is helpful for its automation components.
It is worth evaluating.
Control-M is mainly an infrastructure tool that we use for scheduled tasks. The IT teams and most of the operations teams are the ones who use it. I would estimate about 10 people, but the management of the application is centralized.
The big lesson learnt: Reach out to support when using the product and do something that you could reimagine.
We don't have any data analytics in Control-M.
We don't have developer integration with Control-M at this point.
Control-M is doing a fantastic job for what we use it for. The product is a nine out of 10.
I am working with a beauty products company and we are dealing with supply chain issues. Most of the jobs in Control-M are through SAP.
Right now we have it deployed on-prem, but we are planning to move to the cloud very soon. We are using Control-M Workload Archiving, Control-M Enterprise Manager, Control-M servers, agents, APIs, REST APIs, and Control-M Forecast. We use all the services Control-M provides except Control-M Workload Change Manager.
Since moving to Control-M we have seen a lot of reductions in the manpower needed. For notification, ticketing, and integration, we have different teams. We have Azure teams and some Windows teams. Previously, they were using and managing their own scripts and manually running them. After the migration to Control-M, there were no limitations. Where there are different protocols we can use the APIs and integrate things with Control-M. There are no worries about integrations with Control-M. In UC4 there were lots of limitations because we needed the same protocols to integrate things. With Control-M, there are no such limitations.
In our current environment, there are three sets of applications. The first, an online application, is dependent on some 45 files that have to be generated on Saturday. Our middleware job is supposed to run once all the 45 files have been generated by SAP jobs. There are sequences running through Control-M: First are the SAP jobs that generate the files in a certain location. Once those files are there, the sequence initiates the middleware that moves the files to the proper IT server. All these process flow dependencies go through Control-M very easily.
We have also automated daily audit reports through the solution's reporting facility. Through scripting, we get an alert when anything happens in the Control-M environment. An issue might occur with the agent, the process, or the Control-M server. We have everything reported via email. We can easily see what happened on a given day and sort out any issues.
As a result of using Control-M we have also seen an improvement in Service Level Operations performance. We have some monitoring tools in Control-M and our service SLAs have definitely improved. We have a ticketing system integrated with it and we can easily monitor the SLAs for tickets generated through Control-M. If the person responsible for a ticket will not handle it in the right amount of time, the ticket will pop up with a message saying it's in danger of breaching the SLA. Our service levels are much higher with Control-M, when compared to other tools.
Control-M Workload Archiving is a very good feature for us. It helps with our customer requirements in terms of reporting and auditing. We have internal audits every quarter, and every six months we have external audits. During these audits, the auditors get historical data through Control-M. Previously, when we didn't have any archive server, we managed the data in Control-M with man-made scripts, and we would pull the data for the last 365 days, or three or four months back. Since we installed the archiving, we have been able to pull the data, anytime and anywhere, with just one click.
Control-M gives us a unified view where we can easily define, orchestrate, and monitor all our application workflows and data pipelines. We are mostly using SAP and other business warehouse jobs, and we can easily see the systems through Control-M. It gives us a very good view of geographical data. If I go through the Web Services to show things to my customers, they are very satisfied with the Control-M views. They can check historical data and they can see the current view. They can easily pull these up. We are satisfied with the fact that, with one click, we can see all the applications within one view.
Our line-of-business personnel use Control-M’s web interface to support their business initiatives. One of our big applications is our JG application, where a user needs a data pipeline and Power BI jobs with refreshed data. Instead of the user having to send a request to our Control-M team, they can use the Web Services directly to check their data. If they're using an iPad or a desktop, they can easily check on it themselves. They're not dependent on the Control-M team directly. We educate users on how to check things and how to pull the reports. It is very easy to use. Also, we don't have 24/7 support within our company. Suppose a user needs something at midnight. They don't have to wait for the Control-M admin team to give them the report. They can directly pull the details.
With the current version update, I'm not sure why we needed a separate database upgrade. Why not put it all in one package? Previously, you could do it either via a manual upgrade or an in-place upgrade but it wasn't separate. But for the current version, we needed to upgrade the database separately. It meant doubling our tasks to do the upgrade. That is something that needs to be improved.
I've been using Control-M for the last 17 years. My specialization is in Control-M and I'm very happy and very comfortable with it.
I would rate its stability highly, compared to other tools in the market right now, such as UC4, and AutoSys. In the past, I have worked with many banks. All these financial companies are using Control-M, and there is a reason: It's due to the stability.
I would give the scalability a nine out of 10.
In our environment right now, out of 20,000 jobs in Control-M, 15,000 are SAP. We are planning to expand our usage of Control-M to Power BI, Business Warehouse, PeopleSoft, and Azure. Those are in our pipeline right now.
We have about 25,000 users of Control-M on different projects, in the U.S., Japan, India, and Asia Pacific. Some are monitoring programs through Control-M, some are only doing scheduling. Some are responsible for designing, others for the implementation before the licensing. And once this transition team is done, the operations team comes into the picture for monitoring. We have a separate team for integration, as well.
The number of people we require for day-to-day administration of Control-M depends on the job size and the user requirements. We work in an offshore and onsite model. We have a key administrator over the 20,000 jobs, seven schedulers, and nine people on the monitoring team, and that work is done 24/7. The schedulers and admin work 24/5.
In a case where we fail to understand an issue by collecting data on our own through our audit reports, we open a case with BMC. BMC always gives us a fast resolution. Their support is very good.
The setup of the current version of Control-M, overall, is very easy. The upgrade is in-place. With one click the agent upgrades, the server upgrades. The only point, as I mentioned, with upgrading, is that we needed a separate database update. When we upgrade our Control-M server, the database server should be upgraded at the same time.
The initial implementation in my current environment was in 2006. When we took over we just upgraded it. After that, we implemented two more Control-M Servers in this environment, as a PoC.
The amount of time required to implement it depends on the environment we are working with. In this environment, we have two production servers, four QA servers, and two testing servers. We have eight Control-M servers, three Control-M Enterprise Manager servers, and more than 400 agents. It depends on the change process. In our change process, we first need to upgrade our QA and test environments. Once that is done, we can go for the production environment the next day. After that, over the next seven days, we update our Control-M agents. Some of the upgrades require downtime. In four to five hours, we could easily update everything, but it's dependent on the downtime and the customer requirements.
When we upgraded to version 20, first we implemented it in our QA environment and we tested the new version in our test environment for three to four months. Once we see there are no bugs, we implement it in our production environment. We've seen a lot of bugs and BMC has had to produce some patches that we have had to apply in our environment. That is why we approach it the way we do in a QA environment, and wait for three months, and then go to production.
When we moved to Control-M, we used the Control-M Conversion Tool. It's a very important tool. It gives us an idea of where we stand. If I'm going to move an old environment to a new environment, it helps us with any errors so that we can rectify them.
Back in 2016, when I was working with version 7, I opened a case with BMC and they helped me to upgrade everything. It was a very good experience. They dedicated a resource to us. We gave them our implementation plan, they reviewed it, and they suggested how to remediate some missing steps. We followed their approach and, at the time of cut-over, they assigned a dedicated resource. If there was an issue, we could open a ticket and they would come online and sort it out. The BMC Assisted MIGration Offering (AMIGO) is a very good program.
You must accept that BMC licensing can be very confusing. No one can easily understand how they calculate things, whether it is user-based, job-based, or server-based. The calculation is quite tough. How BMC calculates licensing is not easily available anywhere. It's a very tough part for the client at times.
But BMC is a market leader, so users don't easily go for different vendors. If there's an option to go with Control-M, they will always choose BMC. But for people who find the licensing challenging, they will go with a different vendor.
For us, the licensing part is managed by a team in the U.S. But what I deal with is that we have to manage our Control-M jobs to a maximum of 30,000, because we have 30,000 licenses. We have 20,000 with fraud detection and 10,000 in non-fraud. There is a BMC utility that can guide you and alert you if the forecast is for an increase beyond the licensing. It will notify us: "Hey, you have a license for 20,000 and the Control-M forecast shows you might need to increase that number in the coming days." So we do some cleanup, some internal housekeeping to remove things and remain under the threshold. Those are some of the things we do as administrators. We try to manage under whatever licensing we have. Through the BMC reporting tool, we can see our peak number of users in a month. BMC charges if you go over a certain peak.
Control-M is very robust. There is no harm to the customer if you choose Control-M every time. But when it comes to licensing, it's very expensive, and sometimes users think twice.
Previously we were using UC4 for more than 20,000 jobs. But our customers were not very comfortable with the user interface in UC4. Certain things were not appropriate in that tool. Since our decision to migrate to Control-M, our customers have been very satisfied.
Integration is very easy. When I'm thinking about integrating Control-M with anything I'm not worried about it. I know Control-M will definitely have a way to integrate easily. I have used UC4, AutoSys, and Dollar Universe. But when the requirements include integration, I always think of Control-M, because I know the integration will be very easy. I will never go for any of those other tools.
Control-M is very critical for anyone who is using it.
Overall, we have a great visual of all of our key business processes, and it gives us a secure way of transferring everything in and out of the business so that if anything were to be intercepted, it would be secure and not compromised.
We transfer financial files between Google cloud. We use it for the I series. We have a lot of automated jobs, around 3,000 jobs per day, that we load that range between just regular commands for our planning allocations, finance, or data warehouse along with Google cloud. We're starting to implement a lot of that, but a lot of it has been automated and it allows us to process everything in a timely manner.
We are in the process of implementing the managed file transfer which gives us the dashboard, but we are still fine-tuning that. Overall, it does give us a great picture and helps everything. If there's something delayed, it gives us the opportunity to send out a notification to a team to say that their process is delayed. We get tickets created and have everything sorted in a timely manner.
We use Control-M's web. It makes it very easy for us to show them what they need to see and what they don't need to see. They mainly can just view the tasks that they have, but it's pretty divvied up permission-wise.
Control-M integrates file transfers within our application workflows. It has made everything a lot quicker. We've been able to get files transferred to vendors and we've been able to retrieve files from vendors rapidly and securely.
It also streamlines our data and analytics project. Mainly developers will create either different types of processes that we will implement within Control-M to make it automated and that definitely, I would imagine, helps streamline and format certain projects and reports that we send out to executives that helps out a lot. I don't know the exact extent of it, but I would imagine that it has helped our business service delivery.
It has helped to achieve faster issue resolution. With the shouts and notifications that we get, we're able to create tickets as soon as a problem surfaces. So as soon as we do get a job failure, we get an email notification that prompts us to create a ticket, page out the team, and get it resolved in a matter of our terms of our SLA.
The most valuable features are the Advanced File Transfer and the managed file transfer. They make transferring files securely seamless. It's very easy to set up, get deployed, and have it transferred to and from vendors. As long as we can get our firewall rules implemented at a decent time, it's very easy and seamless to get important files transferred in a secure manner.
Control-M has automated critical processes. We run a lot of our backups through Control-M, daily sales reporting, and warehouse initiatives with shipping and planning. There are a bunch of finance processes that go through here that are time-critical. It's made everything more streamlined and secure and it comes through much quicker than doing it manually.
We have had a few small bugs with the configuration of the different types of jobs where it is the order of operations if it's doing a statement, we've noticed that if you try and do a little bit of both, it may cause one of them not to work.
We've also had a few database bugs within our organization. I think we are migrating to OpenJDK rather than just regular Java and that has since shown some issues with our Control-M instance, timing out and causing our jobs to stop running. We are still working with BMC to fine-tune that and get that resolved.
I believe the file transfer process does everything that it needs to do. I don't believe that there's anything that would need to be changed there with all the features that it has, it's pretty robust. But overall I don't really see many changes that we would need.
I have been using Control-M for three to four years.
Other than the database connections that we've had and as of, I believe when we upgraded or moved away from Java using OpenJDK, it's been hit or miss. I know that we've had a few instances where our jobs just stopped processing, but we're not sure if that's related to the application itself or if that's something in our environment, but overall I am personally okay with the way that it runs.
We run it on windows as well as Linux, and we are still trying to work on getting it to our DR site. But, I believe we are able to process quite a bit through there.
We use it for our I series AS 400. We also use it for Google Cloud, Cognos, ADP, many custom applications that we run as well, but we do a lot of I series.
I do not plan to expand it to other applications in the future.
My department consists of eight people, and we are mainly data center analysts. I'm their manager. We also have developers with a select few developers that are able to get in and view it, but they cannot actually create anything. They can just view and see what is running.
Between five to 10 users are responsible for the day-to-day administration of Control-M.
I've never used Control-M before, prior to being here and all I had to use were the help guides from the web, as well as the user interface that we have. The help administration guide has been the only way that we are able to get questions resolved and to go through support.
Their support is hit or miss. We have had successful sessions with them. And then we have other ones where there are fingers being pointed and it doesn't really solve anything. We have a rep that my manager goes through, but we seem to usually get issues resolved in a timely manner.
We have seen ROI. We were able to have fewer people manually running tasks. We're able to put them right into here and we're able to scale and move a lot of file transfers through here.
It is a little bit expensive. I believe that however we are set up, it might be per job that we load or the highest number of jobs that are loaded monthly and I believe it is quite expensive.
My advice would be to try and utilize as many features as you can. Don't get overly creative with things because that can just confuse other people. If there are other users getting in there, you want to definitely have a standard workflow on how jobs should be created, organized, and make sure that you keep track of what's being changed so that if something were to fail it's easily trackable.
It's a very robust application and there is a lot that can be sent to it and sent out of it and you do not want it to get into the wrong hands because you can do quite a bit with it.
I would rate Control-M an eight out of ten.
We use Control-M to automate scripts that we use in banking and automotive use cases. These are our two big applications. We have a total of 18 applications running in Control-M now, and we want to move over approximately 13 more.
Control-M is running in a virtual machine.
Before Control-M, we had a lot of applications running under different operating systems, including Windows Server and Linux Server. We had a lot of scripts and a lot of programs that were running on the servers. When we implemented Control-M, we were able to automate a lot of those scripts. We have a lot of bank applications and processes and to this point, we have automated about 30% of the ones that we have to do.
We have automated some of our critical processes in core banking. Many of them are now being handled by Control-M. However, we have not yet finished all of the scripts.
Control-M gives us good visibility of our applications and processes. For example, in the morning we can see the results of all of the scripts, whereas, in the past, we could not do that. Our goal is to move the execution of the scripts from the server to Control-M. At this point, the scripts are controlled from Control-M but the execution is done on the server.
We have four domains in Control-M. We have planning, monitoring, history, and forecast. We do not perform data analytics yet.
Our clients use the web-based interface to interact with Control-M.
When a new team member or a new client wants to use Control-M, we have to install a client on their machine. After that is done, there are three options. The first is called Workload, and it is used for observing or monitoring the workload and execution of the jobs. The second one is called the configuration control manager, and it's for configuration administration. The third is reporting, which is another important one. We use the reporting module to generate our reports that concern the execution of the jobs.
We use Control-M to integrate file transfers with our workflows. It is called Advanced File Transfer (AFT) and is used by our financial team. We have another technology for file transfers but the problem with it is that it provides no transparency. There is no interface to see the transfers between applications. With Control-M, we can monitor the transfer between applications and it's great because we can see everything that happened throughout the day.
AFT allows us to configure actions. For example, if a file transfer does not complete successfully then we can send a notification to the destination about the problem.
Control-M has helped to improve our data transfers because it allows us to monitor the execution of the process. With other technologies, we cannot do that. Also, it allows us to configure the notifications, which is very important for us because it will automatically tell the other team when there is a problem with the transfer.
The most valuable feature is the monitoring, which allows me to see the execution and results of each of the scripts.
Being able to view the history is very important because if we have a problem then that is where we search for the details.
From an administrative perspective, the planning domain is very important when we want to add a new feature or a new script.
The forecast domain is what we used to ensure that the implementation is working and that the configuration is okay.
Compared to similar technologies, AFT takes a lot of time when transferring a large file from server to server.
The history module only contains a maximum of 10 days, but we would like to have access to more. For example, it would be helpful to have 30 days or two months of history available.
There should be more granular control available for monitoring applications and sub-applications. For example, when we want to monitor a job, we can specify the application, but we want to have the option to only specify sub-applications that are related to it. As it is now, all of the sub-applications are monitored.
We implemented Control-M for our clients approximately four years ago.
Control-M has been stable for us since we implemented it, four or five years ago. We have not had any problems with the database, file system, or scheduling component.
The scalability is excellent.
We have about 13 people who work regularly with Control-M. We are all engineers and IT managers, and I am the main administrator. The other administrators are in charge of their specific applications, and they need access to Control-M because they need to see the execution plans for the applications that they are in charge of.
I have worked with BMC technical support and I would rate them a nine out of ten.
They respond very quickly, according to the severity of the problem. Also, the responses that they give are really clear and assist us with finding the problem, as well as the root cause.
I have worked with Dollar Universe and AutoSys in the past, before working with Control-M. I find that overall, Control-M is the best one for several reasons.
First, with Control-M, it's easy for someone to be an administrator. All of the documentation is available online, which is important. The second point is that the interface is easy to use. The third is that the solution is really stable compared to other products, such as AutoSys or Dollar Universe. These solutions were not stable in our environment. Part of the reason was that we had trouble finding any documentation online.
This is an expensive product compared to other solutions, although I think that it is a good one. We are in a good position with licensing, as we can run 10,000 jobs. To this point, we have 3,000 jobs that are running, which gives us room to integrate the remainder of our applications.
My advice for anybody who is looking to use Control-M is to have a lot of money. It is a good solution but it is expensive compared to others.
I would rate this solution a ten out of ten.
It is controlling our workflows, ingesting data, and then putting it up into our database platforms. In turn, those are consumed by our internal clients.
We do integrate Control-M Python Client and cloud data service integrations with some of our cloud providers. We have pipelines going out to the public cloud and some pipelines that are internal.
We have public and private cloud channels as well as on-prem. The expectation for most large financial institutions is that we will get 99.9% to the public cloud eventually. We want everything to be in OpEx as opposed to CapEx. We don't want data centers. We just want access to our data and to be able to turn it into information, which in turn, turns it into actionable items. Ideally, we would love to not support any on-prem or hybrid solutions, having everything be public.
Control-M has improved our visibility and streamlining. We have better clarity into data flows. We can resolve issues faster by not trying to reverse engineer what pipeline the infraction may have come through. We are not completely there yet, but we have better clarity and visibility.
We have a better picture of our auditability. When someone comes to us, and asks for sources, "How did the deltas occur?" We can provide answers quickly, or at least quicker than what we used to. We are actually sure of the information that we provide, where before it was like, "Hmm, I think it comes from over there. Let me double check, but it gets really convoluted over here and I think that is where it comes from." Now, if it is within the Control-M environment, it has a straightforward answer that we can provide with confidence.
The speed of our audit preparation process is faster. When questions come in about flow, data, or sources, we don't have to try to reverse engineer anything anymore. We are able to go straight to Control-M and find out what the flow is or what happened. The visibility is there. We see the endpoint on this, such as, "What is the reverse flow on it? Where did it come in? Where did that data flow come from?" So, it is not a spaghetti mess anymore. This makes auditability easier. We are able to provide answers more quickly, which in turn, makes the audit process quicker.
Control-M has improved our business service delivery speed. It is more reliable and has increased the release schedules. We are also working on testing standards, and it has shortened the window of getting things to us. It has shortened the window, not to market, but basically getting them live.
Control-M is critical to our business. If the support ends, we are at risk in some of our critical flows. We have redundancy around it that has been purposely built. We do that with all of our solutions. That way, we are not tied into one specific vendor, then if something happens tomorrow, we don't have a fire drill. We have things in place, but to a certain extent, there is heavy reliance on this solution.
The most valuable feature is the Self Service tool. They have metrics in place almost all across the pipeline, which is really nice.
We are not yet really a power user of it. You can take as many training classes as you need, but it is not until you are working with a subject-matter expert (SME) on it that you can find out how you can really make this tool sing. My engineers know how to work Control-M. However, if I ask them, "Oh, is this the most efficient way of doing it?" They may not be able to say, "Yes." It is doing what we want it to do. That is nice and okay, but is it the most efficient, effective way? So, we are not there yet.
I have been using it for about four years.
The platform is good. We haven't had any major outages. The stability is there.
We really haven't pushed it to any of its limits. No scalability concerns have come up for what we are doing.
If you came to me, saying, "Hey, I was looking at Control-M, but it has some issues." I am going to sit there, and go, "Tell me what the issue is." Right now, we are not using the far reaches of whatever cloud providers are out there. Control-M does well with the major providers.
The community is not as robust as some of our other tools that were replaced. The problem was the other tools that we were using didn't do everything that Control-M is now able to do, like monitoring and the entire pipeline flow.
The community and the networking that goes on within that community need improvement. We want to be able to reach out to an SME, and say, "Hey, we are doing it this way. Does that make sense?" Ideally, they come back. and say, "Yes, it does make sense to do it that way. However, if you want to do it this way, then it is a little more efficient." We understand that one solution framework doesn't fit everybody. Depending on the breadth of the data and how broad it is, you may have different models for one over the other.
I would rate the technical support as seven or eight out of 10.
Positive
We had a patchwork set of solutions in place that were getting the job done. The problem with that was we had a lot of SMEs within certain verticals. Therefore, there wasn't one overall picture. Every time we went from one step to another step, we had to start talking to another person to figure out what was going on. So, we were trying to bring everything together under one solution with Control-M.
We are able to have a better picture of our data consumption, e.g., what files or data is brought in. Previously, we would ingest data at different points. The question that would always come back to us would be, "Where did this data come from?" Then, we would always have to reverse engineer and have some documentation on it, but the documentation would be outdated. Someone would change the pipeline and forget to change the documentation. With Control-M, we can see everything in one location. To a certain extent, it is not documentation.
I am an engineer by trade. I have been doing this for over 30 years. I know that it is nice that someone puts together a document describing the environment, but as soon as that document is saved that document is outdated.
We don't throw another tool into the toolbox just because it is a nice pretty tool. We try to figure out what the benefits are. Ideally, in our world, we try to reduce the number of tools because I don't need 50 different screwdrivers in my tool kit. I make sure that I have a flathead and a Phillips, but I don't need 50 screwdrivers. Here, we brought in this solution and it replaced some existing solutions. Now, my engineers don't need to know X number of products. They only need to know half of X number of products.
The tool was vetted by another group before making it available to the organization and putting it into our toolbox. Then, when it was available, we looked to leverage it.
One of the restrictions that we had was with some of the licensing, and not having any insight on the financials part of the product. I don't know what the licensing on the product is, but we don't have an unlimited enterprise license. So, there might be a limitation on either the cost of the licensing or the number of seats.
There is always a learning curve any time you are using a new product. Our engineers who are using Control-M are kind of happy with it. There really are no negatives on its learning curve. I am always weary with new products since it is another thing that someone needs to learn, but now there are other products that we don't use because of Control-M. What I would not be open to is bringing in another product, where we need our engineers to know how to work it and make it efficient as well as support other products already in our environment. So, I like that we can get rid of three or four products and replace them with a single product. As long as the learning curve is not too steep, that is an advantage to me.
We are looking into using Control-M to deliver analytics for complex data. So, the solution is doing either machine learning or complex analytics on top of the data flow. While we do some analytics, it is not to the extent that we really want to.
I would rate this solution as a high seven or low eight (out of 10).
We use Control-M for managed file transfer in the enterprise manager automating database workflows. We're using Oracle Business Intelligence with a generic database like Microsoft SQL Server. Next year, we plan to use Control-M for AWS Natural and upgrade Control-M to Helix Control-M. We need managed file transfers between our servers in Key West and Orlando. Vast amounts of data are routinely transferred between servers for backup and snapshots.
We will deploy it on AWS and Azure next year. We will also integrate Control-M with other solutions like Salesforce and COS, which is challenging because COS is a complicated legacy IBM OS. Some of our customers require Control-M to do managed file transfer and also COS conversions between IBM jobs. Anything that involves COS or OS/400 is complicated but doable.
With Managed File Transfer, we gain speed and reliability because a managed transfer continuously checks the CRC of the data packages. That's a God send for those data transfers. Though we are migrating to the cloud, we still use some physical servers for sensitive data from our customers protected with NDAs.
We use Control-M for the maintenance of our Oracle and SQL Server databases. It automates maintenance in packages, including standard procedures on the databases themselves, snapshots, checking integrity, verifying the RDBMS of the databases, etc. It ensures that they aren't clogged, that they run smoothly, and that there aren't any jobs stuck, eating up the performance of the server or any of the CPU cores.
In the past, we had some troubles, and we needed a database admin to keep an eye on it almost 24/7 using the OES. It's essential to ensure everything inside the OES runs smoothly, and there are no stuck jobs or queries eating up table spaces. An admin is still required, but most jobs are now automated. It has had a significant impact on staffing. In the past, we had a couple of DBAs exclusively assigned to Oracle that we were able to reassign to other jobs.
We reassigned them to other tests and outsourced one to work with our customers. Once we delegated DBA tasks to Control-M for our Oracle databases, we could reassign that DBA as a resource to our client in Puerto Rico. He became a source of income for the company. Also, with time saved by automating all the critical internal business processes, we could dedicate more time and resources to other projects that require human attention. We could devote more resources to projects that advance the company's strategic vision instead of monitoring an Oracle RDBMS 24/7.
If I had to rate how critical Control-M is to our business, I would say it is an eight out of ten. I won't give it a nine because we still rely on older applications, such as Oracle databases, but an orchestrator will always get at least an eight on our book. For speed of process execution, I would say it is a nine out of ten. Previously, it was a four, and now it's a nine.
Control-M is intuitive, and BMC has tutorials for every application to help you with the basics. Once you know what you're doing, everything falls into place. The graphical interface is drag and drop. There are plenty of objects to drag and drop inside. You need to study them, but once you know how it works, it's just dragging and dropping like you are playing with a Lego set.
You drag two actions to the workspace and connect them to establish a relationship, schedules, and subtasks inside each one of them. It seems complex initially, but it becomes intuitive the longer you use it.
You can almost reach out intuitively into every nook and cranny of the entire UI. It's user-friendly for the initiate, but you could be lost if you've never used an orchestrator or an enterprise-grade software like Control-M. However, Control-M has built-in tutorials that help you with the first steps. The tutorial isn't comprehensive, but at least you will learn the first steps, so you can advance and learn more.
The performance could be better. Control-M Enterprise Manager tends to slow the system down even on a server with a six-core processor and 32 gigabytes of RAM. The console is Java-based, so maybe OpenJDK 16 or 17 would be a performance improvement.
I've been using Control-M for five years.
The scalability is quite good. The framework lets you start with Control-M Enterprise Manager and add other products as you see fit. We added MFT, then Control-M for databases, and Oracle Business Intelligence. One of our customers added Control-M, including the agent for IBMI and another for Azure.
I rate BMC support a solid nine out of ten. I say nine because I never give a ten to anyone.
Positive
Before Control-M, we had a traditional solution using an FTP server. Even with a T1 line that provided almost gigabyte speed, we still had artifacts during the transfer that corrupted the data. It caused serious problems when transferring 30 gigabytes of a necessary backup overnight that failed on gigabyte 28 because it was corrupted. Still, the mirror server rejected the mage because of corruption when we tried to restore it.
We had to rely on traditional monitoring tools like SolarWinds and IBM solutions, which are pretty expensive. These tools only monitor, so they're typically not reactive or able to orchestrate the steps of a workflow. They don't follow up on each step inside the workflow, notify you when a step completes, or send alerts when something gets stuck and requires action.
I'm the senior services manager, and overseeing the deployment of Control-M is part of my job. I did not install it, but I supervised the team. It was straightforward because we all got our BMC certification before the deployment. Our team included me and two technicians. We also had a DBA around to integrate the database.
We did everything ourselves with some occasional help from BMC support. We emailed them a couple of times to check something, but so far, everything has gone smoothly.
We recovered our initial investment in six months and were ready to commit more, so we could recover more. We saw an ROI with Control-M in the first two years because we could take a DBA off monitoring databases and loan them out to another company while saving time by speeding up these processes.
Control-M gives us a lot of flexibility to automate our time-sensitive and data-critical processes. This is essential for enterprises, but Control-M isn't for everyone. Small and medium-sized businesses can use Control-M, but its power can only be leveraged by large enterprises because of the complexity of their business processes and the sheer size of data they handle. I think enterprise companies are the only ones that see an effective ROI from purchasing a tool like BMC Control-M for automating their business processes.
Small companies that purchase Control-M often cancel the contract after six months because it's too expensive, and they can get the job done using traditional methods. That's okay. It's about the business processes' complexity, depth, and maturity.
The price is right because of the licensing schema, which is based on nodes and processes. You purchase what you use, no more and no less, and you can grow with time.
There are few options like Control-M in the market, and the closest competitors are far more expensive.
I rate Control-M a nine out of ten. Control-M is flexible. You can use it in Azure, and they have a generic option for the cloud. You can deploy it in your own private cloud or on other cloud solutions like Kubernetes. You can use Control-M for big data applications like IBM InfoSphere. There's a Control-M solution for almost any situation.
There is so much to learn on the backend of the business processes. Typically when you see a business process, you only see a workflow, like a flow chart, arrows, boxes, etc. However, there's a whole new world under the hood. It's crucial to dig deeper and learn how to improve the processes. It's like you become the mechanic of your own car. The more you understand the engine, the more you can tweak it to get more speed, gas mileage, performance, strength, horsepower, etc. Control-M almost compels you to learn about that.
It's user-friendly, but you need some training. We have a certification from BMC. You need some prior training specifically in Control-M Enterprise Manager to know what you're doing because it's delicate. There are so many ways to customize job creation, automation, monitoring, etc. that you need at least a crash course on creating a job, monitoring, setting up alarms, and building workflows.
It should take you no more than a week to get the hang of it, and there's BMC University, where you can get free training to use Control-M. Once you know the basics, Control-M practically handles itself. It's intuitive once you figure it out.
We primarily use Control-M for financial services. We do file transfers, payroll, HR, and other related tasks.
The top three processes that we have automated with Control-M are payroll, HR reports, and time reports. This automation gives HR, and the business in general, a clearer picture of what is happening as far as the payroll timesheets go, including who's punching in and punching out. Essentially, it improves transparency.
Control-M provides us with a unified view where we can define and monitor applications, and this is very important to us, especially in the audit process. We have auditors and they request certain information; using Control-M, we can log in and create the report according to the parameters they're asking for. It makes life much easier.
Our developers use the web-based interface to monitor their jobs. They do not have access to do anything else but they can tell if their jobs have run, or not.
Our developers leverage the “as-code” interfaces and it makes it very easy to roll out new applications and application updates. Everything is automated as far as transferring files in and out to certain people. This is helpful because it doesn't have to be done manually. It also generates reports automatically for us. Control-M jobs produce the reports so we don't have to create them every day.
We just started using Control-M to streamline our data and analytics projects. So far, it has given us some actionable insights. The streamlining has improved our business service delivery because we can tell if something is running behind, and why. We know if there's an issue before anybody notices.
Control-M has improved our data transfers because it is much easier to do encryption back and forth when sending files.
This product has helped us to achieve faster issue resolution. I estimate that issues are now normally resolved within 10 minutes. It's very quick.
Control-M has helped us to improve service-level operations performance. We have a critical job stream and because we're an institution, we have to have certain data out at a certain time for the federal reserve. If we can detect when something is running behind, and why, then we can notify them ahead of time so that they know the reports are going to be late. It helps them on their end, too. This way, they don't have to call and ask us where their report is.
The most valuable feature is the reporting function. It allows us to pull up reports for specific information that the end-users are looking for.
The planning and monitoring features in Control-M 20 help us because we can forecast to assist with network maintenance. For example, if we have something major going on with the network and there is going to be downtime, we can do a forecast to see what's going to be running at that particular time and adjust things accordingly.
I would like to see more audit report templates added, and perhaps more customizability in terms of reporting.
I have been using Control-M for 15 years.
Control-M is very stable and we've had no major issues with it.
Scalability is not a problem. It absolutely extends with our needs and the jobs that it needs to run in. At this time, it is running payroll reports and other payroll jobs. We are looking at expanding this to other applications in the future, although there is nothing definite yet.
Technical support is very responsive and we have never had any issues with them. Generally, if we have a problem or question, we can open a ticket with BMC and we usually get a response back within an hour, or no later than two hours.
I would rate their support a ten out of ten.
We work with BMC for upgrades and support. We are part of the AMIGO program.
We have absolutely seen a return on investment with Control-M. It is centralized and it's made everything easier for the business end, in particular for getting their reports on time.
The pricing of Control-M is reasonable.
The developers in our organization are responsible for creating the scripts. There are 20 of them and they monitor their jobs. With respect to operations including creating, running, modifying, and killing the jobs, there is a group of six staff in charge of that. This group also creates the schedule and the calendars, so essentially, they take care of the day-to-day administration.
My advice for anybody who is considering Control-M is that if they have questions then BMC is great to talk to, but there is also a BMC community and if they have questions or want to know how it's running or working for other organizations, they can post and generally get a response back. There are user groups specifically within Control-M.
I would rate this solution a ten out of ten.
I work for the second largest chain of supermarkets in the UK. We are running about 90% of our jobs through Control-M. This applies for jobs and scripts on-premises and in the cloud.
When we used Control-M version 7, we were just doing scheduling. When we moved to Control-M version 9 six months ago, we started using the cloud plugins, like AWS.
Control-M is business-critical for our operations. It does all our monitoring and tracking.
Our command center people watch the Control-M job status and alerts. Since the pandemic started, and we are working from home, we have been providing them with Self Service. We started this two or three months back. Now, they can watch the jobs and alerts through their mobile and iPads instead of logging into their laptops.
We set up a file transfer mechanism because this will be easier for Control-M to track end-to-end.
We use Control-M as part of our DevOps automation toolchains. We have a four-person team for Control-M. We help the DevOps team create new jobs. We assign a dedicated resource to understand their requirements and how they can be integrated with other jobs. Because Control-M works end-to-end, it is critical for our DevOps daily jobs.
We use Control-M to streamline our data and analytics projects. Control-M has helped improve our data transfers. If there are no security concerns, the data can directly link to downstream systems. We use Control-M to watch all the transfers of files to their targets.
All our Control-M alerts go to our internal automation.
It has two-way integration. We now have a ServiceNow integration.
Control-M reporting isn't that good. It is very limited. We would like the ability to create our own reports as well as the ability to publish dashboards in the cloud, which would help us. Improved reporting will help us determine statuses and get the answers that we need. However, I personally think BMC is not focusing on the reporting. I have even visited the BMC office in India, and asked, "Why haven't you improved the reporting?"
There are some latency issues with jobs between on-premises and the cloud. BMC is helping a lot to check the imports and exports from version 7 to version 9, including the EM server and the mainframe.
Control-M could improve agentless connectivity a little more. We are using it almost 100% with agents, but when we start using agentless, Sainsbury's Bank has different security mechanisms and we cannot install Control-M. For example, the agentless connection fluctuates a lot, which triggers alerts.
I have worked with Control-M for almost 10 years, since 2010.
The stability is very good.
The scalability of the latest version is a drastic improvement compared to version 7.
We are getting good help from them. When I use Support Central, I can also see tickets that have been created by my colleagues.
We currently have IBM TWS as a job scheduler, but they don't automate their ticketing. Whereas, Control-M has automatic ticketing.
We are using TWS for mainframe data. We are looking to start moving all our TWS jobs to Control-M now that Control-M is in the cloud. We are looking at moving these jobs around September or October, then we will have 200,000 jobs daily in Control-M.
We are trying to import from Control-M version 7 to Control-M version 9, but have experienced a major problem with its new features (database-related stuff). We are slowly fixing this as we go, with the help of BMC. Right now, we are doing this process step-by-step, but we can't upgrade everything to the latest version. We can only move everything to Control-M version 9.5.
Initially, we were first-timers doing the cloud. We had so many trials and errors. For importing, we created virtual machines in AWS and set up a lot of automation. However, we needed a static IP address for Control-M. So, we had to start from scratch to create new virtual machines with static IP addresses.
We are currently importing step-by-step. We still have two mainframe servers that we need to do and should be done by the end of August.
We have 70,000 jobs running daily. Control-M has reduced the number of jobs that we are running daily. We used to have more than 500,000 jobs running daily. This is very important to us.
Control-M has helped us resolve issues 70% to 80% faster. It provides us with alerts instead of having someone go to that particular server and check the logs to determine where the issue is. We can simply click on the alert information, then everything is in front of us. This provides us with time savings, human effort savings, and process savings.
You can't compare other tools to Control-M, because Control-M is further ahead of any other tool.
Once a year, as part of our disaster recovery, we restart Control-M and see what happens. Next, we will run those jobs through Control-M. Then, we will show management, "This is what happens if you use Control-M and if you don't use Control-M."
There are some areas of our business where we don't have Control-M. When we start doing those areas through Control-M, it will be an end-to-end solution.
We don't use Control-M for file transfers. We have proposed using Control-M for file transfer with version 9, which is in the cloud.
In the future, we will give control to the DevOps team through BMC AMI Change Manager. They will create the jobs, then send them to our BMC Control-M team for review, testing, and promotion to production. However, adopting this will take some time.
I would rate Control-M as a nine out of 10.